commit
stringlengths
40
40
old_file
stringlengths
4
118
new_file
stringlengths
4
118
old_contents
stringlengths
0
2.94k
new_contents
stringlengths
1
4.43k
subject
stringlengths
15
444
message
stringlengths
16
3.45k
lang
stringclasses
1 value
license
stringclasses
13 values
repos
stringlengths
5
43.2k
prompt
stringlengths
17
4.58k
response
stringlengths
1
4.43k
prompt_tagged
stringlengths
58
4.62k
response_tagged
stringlengths
1
4.43k
text
stringlengths
132
7.29k
text_tagged
stringlengths
173
7.33k
b9b7fb6f4a7c334f56a9a6cd34c50124008adb26
apps/homepage/model.py
apps/homepage/model.py
from django.utils.timezone import now from apps.banners.models import Slide from apps.campaigns.models import Campaign from apps.fundraisers.models import FundRaiser from apps.projects.models import Project from apps.quotes.models import Quote from apps.statistics.models import Statistic # Instead of serving all the objects separately we combine Slide, Quote and Stats into a dummy object class HomePage(object): def get(self, language): self.id = 1 self.quotes= Quote.objects.published().filter(language=language) self.slides = Slide.objects.published().filter(language=language) stats = Statistic.objects.order_by('-creation_date').all() if len(stats) > 0: self.stats = stats[0] else: self.stats = None projects = Project.objects.filter(phase='campaign').order_by('?') if len(projects) > 4: self.projects = projects[0:4] elif len(projects) > 0: self.projects = projects[0:len(projects)] else: self.projects = None try: self.campaign = Campaign.objects.get(start__lte=now(), end__gte=now()) # NOTE: MultipleObjectsReturned is not caught yet! self.fundraisers = FundRaiser.objects.filter(project__is_campaign=True).order_by('?') except Campaign.DoesNotExist: self.campaign, self.fundraisers = None, None return self
from django.utils.timezone import now from apps.banners.models import Slide from apps.campaigns.models import Campaign from apps.fundraisers.models import FundRaiser from apps.projects.models import Project from apps.quotes.models import Quote from apps.statistics.models import Statistic # Instead of serving all the objects separately we combine Slide, Quote and Stats into a dummy object class HomePage(object): def get(self, language): self.id = 1 self.quotes= Quote.objects.published().filter(language=language) self.slides = Slide.objects.published().filter(language=language) stats = Statistic.objects.order_by('-creation_date').all() if len(stats) > 0: self.stats = stats[0] else: self.stats = None projects = Project.objects.filter(phase='campaign').order_by('?') if len(projects) > 4: self.projects = projects[0:4] elif len(projects) > 0: self.projects = projects[0:len(projects)] else: self.projects = None try: self.campaign = Campaign.objects.get(start__lte=now(), end__gte=now()) # NOTE: MultipleObjectsReturned is not caught yet! self.fundraisers = FundRaiser.objects.filter(project__is_campaign=True).order_by('-created') except Campaign.DoesNotExist: self.campaign, self.fundraisers = None, None return self
Sort fundraiser on newest on hp
Sort fundraiser on newest on hp
Python
bsd-3-clause
onepercentclub/onepercentclub-site,onepercentclub/onepercentclub-site,onepercentclub/onepercentclub-site,onepercentclub/onepercentclub-site,onepercentclub/onepercentclub-site
from django.utils.timezone import now from apps.banners.models import Slide from apps.campaigns.models import Campaign from apps.fundraisers.models import FundRaiser from apps.projects.models import Project from apps.quotes.models import Quote from apps.statistics.models import Statistic # Instead of serving all the objects separately we combine Slide, Quote and Stats into a dummy object class HomePage(object): def get(self, language): self.id = 1 self.quotes= Quote.objects.published().filter(language=language) self.slides = Slide.objects.published().filter(language=language) stats = Statistic.objects.order_by('-creation_date').all() if len(stats) > 0: self.stats = stats[0] else: self.stats = None projects = Project.objects.filter(phase='campaign').order_by('?') if len(projects) > 4: self.projects = projects[0:4] elif len(projects) > 0: self.projects = projects[0:len(projects)] else: self.projects = None try: self.campaign = Campaign.objects.get(start__lte=now(), end__gte=now()) # NOTE: MultipleObjectsReturned is not caught yet! self.fundraisers = FundRaiser.objects.filter(project__is_campaign=True).order_by('?') except Campaign.DoesNotExist: self.campaign, self.fundraisers = None, None return selfSort fundraiser on newest on hp
from django.utils.timezone import now from apps.banners.models import Slide from apps.campaigns.models import Campaign from apps.fundraisers.models import FundRaiser from apps.projects.models import Project from apps.quotes.models import Quote from apps.statistics.models import Statistic # Instead of serving all the objects separately we combine Slide, Quote and Stats into a dummy object class HomePage(object): def get(self, language): self.id = 1 self.quotes= Quote.objects.published().filter(language=language) self.slides = Slide.objects.published().filter(language=language) stats = Statistic.objects.order_by('-creation_date').all() if len(stats) > 0: self.stats = stats[0] else: self.stats = None projects = Project.objects.filter(phase='campaign').order_by('?') if len(projects) > 4: self.projects = projects[0:4] elif len(projects) > 0: self.projects = projects[0:len(projects)] else: self.projects = None try: self.campaign = Campaign.objects.get(start__lte=now(), end__gte=now()) # NOTE: MultipleObjectsReturned is not caught yet! self.fundraisers = FundRaiser.objects.filter(project__is_campaign=True).order_by('-created') except Campaign.DoesNotExist: self.campaign, self.fundraisers = None, None return self
<commit_before>from django.utils.timezone import now from apps.banners.models import Slide from apps.campaigns.models import Campaign from apps.fundraisers.models import FundRaiser from apps.projects.models import Project from apps.quotes.models import Quote from apps.statistics.models import Statistic # Instead of serving all the objects separately we combine Slide, Quote and Stats into a dummy object class HomePage(object): def get(self, language): self.id = 1 self.quotes= Quote.objects.published().filter(language=language) self.slides = Slide.objects.published().filter(language=language) stats = Statistic.objects.order_by('-creation_date').all() if len(stats) > 0: self.stats = stats[0] else: self.stats = None projects = Project.objects.filter(phase='campaign').order_by('?') if len(projects) > 4: self.projects = projects[0:4] elif len(projects) > 0: self.projects = projects[0:len(projects)] else: self.projects = None try: self.campaign = Campaign.objects.get(start__lte=now(), end__gte=now()) # NOTE: MultipleObjectsReturned is not caught yet! self.fundraisers = FundRaiser.objects.filter(project__is_campaign=True).order_by('?') except Campaign.DoesNotExist: self.campaign, self.fundraisers = None, None return self<commit_msg>Sort fundraiser on newest on hp<commit_after>
from django.utils.timezone import now from apps.banners.models import Slide from apps.campaigns.models import Campaign from apps.fundraisers.models import FundRaiser from apps.projects.models import Project from apps.quotes.models import Quote from apps.statistics.models import Statistic # Instead of serving all the objects separately we combine Slide, Quote and Stats into a dummy object class HomePage(object): def get(self, language): self.id = 1 self.quotes= Quote.objects.published().filter(language=language) self.slides = Slide.objects.published().filter(language=language) stats = Statistic.objects.order_by('-creation_date').all() if len(stats) > 0: self.stats = stats[0] else: self.stats = None projects = Project.objects.filter(phase='campaign').order_by('?') if len(projects) > 4: self.projects = projects[0:4] elif len(projects) > 0: self.projects = projects[0:len(projects)] else: self.projects = None try: self.campaign = Campaign.objects.get(start__lte=now(), end__gte=now()) # NOTE: MultipleObjectsReturned is not caught yet! self.fundraisers = FundRaiser.objects.filter(project__is_campaign=True).order_by('-created') except Campaign.DoesNotExist: self.campaign, self.fundraisers = None, None return self
from django.utils.timezone import now from apps.banners.models import Slide from apps.campaigns.models import Campaign from apps.fundraisers.models import FundRaiser from apps.projects.models import Project from apps.quotes.models import Quote from apps.statistics.models import Statistic # Instead of serving all the objects separately we combine Slide, Quote and Stats into a dummy object class HomePage(object): def get(self, language): self.id = 1 self.quotes= Quote.objects.published().filter(language=language) self.slides = Slide.objects.published().filter(language=language) stats = Statistic.objects.order_by('-creation_date').all() if len(stats) > 0: self.stats = stats[0] else: self.stats = None projects = Project.objects.filter(phase='campaign').order_by('?') if len(projects) > 4: self.projects = projects[0:4] elif len(projects) > 0: self.projects = projects[0:len(projects)] else: self.projects = None try: self.campaign = Campaign.objects.get(start__lte=now(), end__gte=now()) # NOTE: MultipleObjectsReturned is not caught yet! self.fundraisers = FundRaiser.objects.filter(project__is_campaign=True).order_by('?') except Campaign.DoesNotExist: self.campaign, self.fundraisers = None, None return selfSort fundraiser on newest on hpfrom django.utils.timezone import now from apps.banners.models import Slide from apps.campaigns.models import Campaign from apps.fundraisers.models import FundRaiser from apps.projects.models import Project from apps.quotes.models import Quote from apps.statistics.models import Statistic # Instead of serving all the objects separately we combine Slide, Quote and Stats into a dummy object class HomePage(object): def get(self, language): self.id = 1 self.quotes= Quote.objects.published().filter(language=language) self.slides = Slide.objects.published().filter(language=language) stats = Statistic.objects.order_by('-creation_date').all() if len(stats) > 0: self.stats = stats[0] else: self.stats = None projects = Project.objects.filter(phase='campaign').order_by('?') if len(projects) > 4: self.projects = projects[0:4] elif len(projects) > 0: self.projects = projects[0:len(projects)] else: self.projects = None try: self.campaign = Campaign.objects.get(start__lte=now(), end__gte=now()) # NOTE: MultipleObjectsReturned is not caught yet! self.fundraisers = FundRaiser.objects.filter(project__is_campaign=True).order_by('-created') except Campaign.DoesNotExist: self.campaign, self.fundraisers = None, None return self
<commit_before>from django.utils.timezone import now from apps.banners.models import Slide from apps.campaigns.models import Campaign from apps.fundraisers.models import FundRaiser from apps.projects.models import Project from apps.quotes.models import Quote from apps.statistics.models import Statistic # Instead of serving all the objects separately we combine Slide, Quote and Stats into a dummy object class HomePage(object): def get(self, language): self.id = 1 self.quotes= Quote.objects.published().filter(language=language) self.slides = Slide.objects.published().filter(language=language) stats = Statistic.objects.order_by('-creation_date').all() if len(stats) > 0: self.stats = stats[0] else: self.stats = None projects = Project.objects.filter(phase='campaign').order_by('?') if len(projects) > 4: self.projects = projects[0:4] elif len(projects) > 0: self.projects = projects[0:len(projects)] else: self.projects = None try: self.campaign = Campaign.objects.get(start__lte=now(), end__gte=now()) # NOTE: MultipleObjectsReturned is not caught yet! self.fundraisers = FundRaiser.objects.filter(project__is_campaign=True).order_by('?') except Campaign.DoesNotExist: self.campaign, self.fundraisers = None, None return self<commit_msg>Sort fundraiser on newest on hp<commit_after>from django.utils.timezone import now from apps.banners.models import Slide from apps.campaigns.models import Campaign from apps.fundraisers.models import FundRaiser from apps.projects.models import Project from apps.quotes.models import Quote from apps.statistics.models import Statistic # Instead of serving all the objects separately we combine Slide, Quote and Stats into a dummy object class HomePage(object): def get(self, language): self.id = 1 self.quotes= Quote.objects.published().filter(language=language) self.slides = Slide.objects.published().filter(language=language) stats = Statistic.objects.order_by('-creation_date').all() if len(stats) > 0: self.stats = stats[0] else: self.stats = None projects = Project.objects.filter(phase='campaign').order_by('?') if len(projects) > 4: self.projects = projects[0:4] elif len(projects) > 0: self.projects = projects[0:len(projects)] else: self.projects = None try: self.campaign = Campaign.objects.get(start__lte=now(), end__gte=now()) # NOTE: MultipleObjectsReturned is not caught yet! self.fundraisers = FundRaiser.objects.filter(project__is_campaign=True).order_by('-created') except Campaign.DoesNotExist: self.campaign, self.fundraisers = None, None return self
8bd47ec3983981d0a2ac8d9f9c17f4c1c9c8fbd3
apps/profiles/tests.py
apps/profiles/tests.py
""" This file demonstrates writing tests using the unittest module. These will pass when you run "manage.py test". Replace this with more appropriate tests for your application. """ from django.core.urlresolvers import reverse from django.test import TestCase from django_dynamic_fixture import G from rest_framework import status from apps.authentication.models import OnlineUser as User class ProfilesURLTestCase(TestCase): def test_user_search(self): user = G(User) url = reverse('profiles_user_search') self.client.force_login(user) response = self.client.get(url) self.assertEqual(response.status_code, status.HTTP_200_OK)
from django.core.urlresolvers import reverse from django.test import TestCase from django_dynamic_fixture import G from rest_framework import status from apps.authentication.models import OnlineUser as User from apps.profiles.forms import ProfileForm class ProfilesURLTestCase(TestCase): def test_user_search(self): user = G(User) url = reverse('profiles_user_search') self.client.force_login(user) response = self.client.get(url) self.assertEqual(response.status_code, status.HTTP_200_OK) class ProfileViewEditTestCase(TestCase): @classmethod def setUpClass(cls): super().setUpClass() cls._url = reverse('profile_edit') cls._user = G(User) def setUp(self): self.client.force_login(self._user) def test_profile_retrieve(self): response = self.client.get(self._url) self.assertEqual(200, response.status_code) def test_profile_save(self): response = self.client.post(self._url) self.assertEqual(200, response.status_code) def test_profile_save_valid_zip(self): data = { 'zip_code': 7030 } response = self.client.post(self._url, data) self.assertEqual(200, response.status_code) class ProfileEditFormTestCase(TestCase): def test_profile_form_valid_zip(self): data = { 'gender': 'male', 'zip_code': 7030 } form = ProfileForm(data=data) self.assertTrue(form.is_valid()) def test_profile_form_invalid_zip(self): data = { 'gender': 'male', 'zip_code': 123 } form = ProfileForm(data=data) self.assertFalse(form.is_valid())
Test that saving ProfileForm works
Test that saving ProfileForm works
Python
mit
dotKom/onlineweb4,dotKom/onlineweb4,dotKom/onlineweb4,dotKom/onlineweb4
""" This file demonstrates writing tests using the unittest module. These will pass when you run "manage.py test". Replace this with more appropriate tests for your application. """ from django.core.urlresolvers import reverse from django.test import TestCase from django_dynamic_fixture import G from rest_framework import status from apps.authentication.models import OnlineUser as User class ProfilesURLTestCase(TestCase): def test_user_search(self): user = G(User) url = reverse('profiles_user_search') self.client.force_login(user) response = self.client.get(url) self.assertEqual(response.status_code, status.HTTP_200_OK) Test that saving ProfileForm works
from django.core.urlresolvers import reverse from django.test import TestCase from django_dynamic_fixture import G from rest_framework import status from apps.authentication.models import OnlineUser as User from apps.profiles.forms import ProfileForm class ProfilesURLTestCase(TestCase): def test_user_search(self): user = G(User) url = reverse('profiles_user_search') self.client.force_login(user) response = self.client.get(url) self.assertEqual(response.status_code, status.HTTP_200_OK) class ProfileViewEditTestCase(TestCase): @classmethod def setUpClass(cls): super().setUpClass() cls._url = reverse('profile_edit') cls._user = G(User) def setUp(self): self.client.force_login(self._user) def test_profile_retrieve(self): response = self.client.get(self._url) self.assertEqual(200, response.status_code) def test_profile_save(self): response = self.client.post(self._url) self.assertEqual(200, response.status_code) def test_profile_save_valid_zip(self): data = { 'zip_code': 7030 } response = self.client.post(self._url, data) self.assertEqual(200, response.status_code) class ProfileEditFormTestCase(TestCase): def test_profile_form_valid_zip(self): data = { 'gender': 'male', 'zip_code': 7030 } form = ProfileForm(data=data) self.assertTrue(form.is_valid()) def test_profile_form_invalid_zip(self): data = { 'gender': 'male', 'zip_code': 123 } form = ProfileForm(data=data) self.assertFalse(form.is_valid())
<commit_before>""" This file demonstrates writing tests using the unittest module. These will pass when you run "manage.py test". Replace this with more appropriate tests for your application. """ from django.core.urlresolvers import reverse from django.test import TestCase from django_dynamic_fixture import G from rest_framework import status from apps.authentication.models import OnlineUser as User class ProfilesURLTestCase(TestCase): def test_user_search(self): user = G(User) url = reverse('profiles_user_search') self.client.force_login(user) response = self.client.get(url) self.assertEqual(response.status_code, status.HTTP_200_OK) <commit_msg>Test that saving ProfileForm works<commit_after>
from django.core.urlresolvers import reverse from django.test import TestCase from django_dynamic_fixture import G from rest_framework import status from apps.authentication.models import OnlineUser as User from apps.profiles.forms import ProfileForm class ProfilesURLTestCase(TestCase): def test_user_search(self): user = G(User) url = reverse('profiles_user_search') self.client.force_login(user) response = self.client.get(url) self.assertEqual(response.status_code, status.HTTP_200_OK) class ProfileViewEditTestCase(TestCase): @classmethod def setUpClass(cls): super().setUpClass() cls._url = reverse('profile_edit') cls._user = G(User) def setUp(self): self.client.force_login(self._user) def test_profile_retrieve(self): response = self.client.get(self._url) self.assertEqual(200, response.status_code) def test_profile_save(self): response = self.client.post(self._url) self.assertEqual(200, response.status_code) def test_profile_save_valid_zip(self): data = { 'zip_code': 7030 } response = self.client.post(self._url, data) self.assertEqual(200, response.status_code) class ProfileEditFormTestCase(TestCase): def test_profile_form_valid_zip(self): data = { 'gender': 'male', 'zip_code': 7030 } form = ProfileForm(data=data) self.assertTrue(form.is_valid()) def test_profile_form_invalid_zip(self): data = { 'gender': 'male', 'zip_code': 123 } form = ProfileForm(data=data) self.assertFalse(form.is_valid())
""" This file demonstrates writing tests using the unittest module. These will pass when you run "manage.py test". Replace this with more appropriate tests for your application. """ from django.core.urlresolvers import reverse from django.test import TestCase from django_dynamic_fixture import G from rest_framework import status from apps.authentication.models import OnlineUser as User class ProfilesURLTestCase(TestCase): def test_user_search(self): user = G(User) url = reverse('profiles_user_search') self.client.force_login(user) response = self.client.get(url) self.assertEqual(response.status_code, status.HTTP_200_OK) Test that saving ProfileForm worksfrom django.core.urlresolvers import reverse from django.test import TestCase from django_dynamic_fixture import G from rest_framework import status from apps.authentication.models import OnlineUser as User from apps.profiles.forms import ProfileForm class ProfilesURLTestCase(TestCase): def test_user_search(self): user = G(User) url = reverse('profiles_user_search') self.client.force_login(user) response = self.client.get(url) self.assertEqual(response.status_code, status.HTTP_200_OK) class ProfileViewEditTestCase(TestCase): @classmethod def setUpClass(cls): super().setUpClass() cls._url = reverse('profile_edit') cls._user = G(User) def setUp(self): self.client.force_login(self._user) def test_profile_retrieve(self): response = self.client.get(self._url) self.assertEqual(200, response.status_code) def test_profile_save(self): response = self.client.post(self._url) self.assertEqual(200, response.status_code) def test_profile_save_valid_zip(self): data = { 'zip_code': 7030 } response = self.client.post(self._url, data) self.assertEqual(200, response.status_code) class ProfileEditFormTestCase(TestCase): def test_profile_form_valid_zip(self): data = { 'gender': 'male', 'zip_code': 7030 } form = ProfileForm(data=data) self.assertTrue(form.is_valid()) def test_profile_form_invalid_zip(self): data = { 'gender': 'male', 'zip_code': 123 } form = ProfileForm(data=data) self.assertFalse(form.is_valid())
<commit_before>""" This file demonstrates writing tests using the unittest module. These will pass when you run "manage.py test". Replace this with more appropriate tests for your application. """ from django.core.urlresolvers import reverse from django.test import TestCase from django_dynamic_fixture import G from rest_framework import status from apps.authentication.models import OnlineUser as User class ProfilesURLTestCase(TestCase): def test_user_search(self): user = G(User) url = reverse('profiles_user_search') self.client.force_login(user) response = self.client.get(url) self.assertEqual(response.status_code, status.HTTP_200_OK) <commit_msg>Test that saving ProfileForm works<commit_after>from django.core.urlresolvers import reverse from django.test import TestCase from django_dynamic_fixture import G from rest_framework import status from apps.authentication.models import OnlineUser as User from apps.profiles.forms import ProfileForm class ProfilesURLTestCase(TestCase): def test_user_search(self): user = G(User) url = reverse('profiles_user_search') self.client.force_login(user) response = self.client.get(url) self.assertEqual(response.status_code, status.HTTP_200_OK) class ProfileViewEditTestCase(TestCase): @classmethod def setUpClass(cls): super().setUpClass() cls._url = reverse('profile_edit') cls._user = G(User) def setUp(self): self.client.force_login(self._user) def test_profile_retrieve(self): response = self.client.get(self._url) self.assertEqual(200, response.status_code) def test_profile_save(self): response = self.client.post(self._url) self.assertEqual(200, response.status_code) def test_profile_save_valid_zip(self): data = { 'zip_code': 7030 } response = self.client.post(self._url, data) self.assertEqual(200, response.status_code) class ProfileEditFormTestCase(TestCase): def test_profile_form_valid_zip(self): data = { 'gender': 'male', 'zip_code': 7030 } form = ProfileForm(data=data) self.assertTrue(form.is_valid()) def test_profile_form_invalid_zip(self): data = { 'gender': 'male', 'zip_code': 123 } form = ProfileForm(data=data) self.assertFalse(form.is_valid())
308cbf1f62e254643a0ad47db8ad55eb63e1c888
argonauts/testutils.py
argonauts/testutils.py
import json import functools from django.conf import settings from django.test import Client, TestCase __all__ = ['JsonTestClient', 'JsonTestCase'] class JsonTestClient(Client): def _json_request(self, method, url, data=None, *args, **kwargs): method_func = getattr(super(JsonTestClient, self), method) if method == 'get': encode = lambda x: x else: encode = json.dumps if data is not None: resp = method_func(url, encode(data), content_type='application/json', *args, **kwargs) else: resp = method_func(url, content_type='application/json', *args, **kwargs) if resp['Content-Type'].startswith('application/json') and resp.content: charset = resp.charset or settings.DEFAULT_CHARSET resp.json = json.loads(resp.content.decode(charset)) return resp def __getattribute__(self, attr): if attr in ('get', 'post', 'put', 'delete', 'trace', 'head', 'patch', 'options'): return functools.partial(self._json_request, attr) else: return super(JsonTestClient, self).__getattribute__(attr) class JsonTestCase(TestCase): client_class = JsonTestClient
import json import functools from django.conf import settings from django.test import Client, TestCase __all__ = ['JsonTestClient', 'JsonTestMixin', 'JsonTestCase'] class JsonTestClient(Client): def _json_request(self, method, url, data=None, *args, **kwargs): method_func = getattr(super(JsonTestClient, self), method) if method == 'get': encode = lambda x: x else: encode = json.dumps if data is not None: resp = method_func(url, encode(data), content_type='application/json', *args, **kwargs) else: resp = method_func(url, content_type='application/json', *args, **kwargs) if resp['Content-Type'].startswith('application/json') and resp.content: charset = resp.charset or settings.DEFAULT_CHARSET resp.json = json.loads(resp.content.decode(charset)) return resp def __getattribute__(self, attr): if attr in ('get', 'post', 'put', 'delete', 'trace', 'head', 'patch', 'options'): return functools.partial(self._json_request, attr) else: return super(JsonTestClient, self).__getattribute__(attr) class JsonTestMixin(object): client_class = JsonTestClient class JsonTestCase(JsonTestMixin, TestCase): pass
Make the TestCase a mixin
Make the TestCase a mixin
Python
bsd-2-clause
fusionbox/django-argonauts
import json import functools from django.conf import settings from django.test import Client, TestCase __all__ = ['JsonTestClient', 'JsonTestCase'] class JsonTestClient(Client): def _json_request(self, method, url, data=None, *args, **kwargs): method_func = getattr(super(JsonTestClient, self), method) if method == 'get': encode = lambda x: x else: encode = json.dumps if data is not None: resp = method_func(url, encode(data), content_type='application/json', *args, **kwargs) else: resp = method_func(url, content_type='application/json', *args, **kwargs) if resp['Content-Type'].startswith('application/json') and resp.content: charset = resp.charset or settings.DEFAULT_CHARSET resp.json = json.loads(resp.content.decode(charset)) return resp def __getattribute__(self, attr): if attr in ('get', 'post', 'put', 'delete', 'trace', 'head', 'patch', 'options'): return functools.partial(self._json_request, attr) else: return super(JsonTestClient, self).__getattribute__(attr) class JsonTestCase(TestCase): client_class = JsonTestClient Make the TestCase a mixin
import json import functools from django.conf import settings from django.test import Client, TestCase __all__ = ['JsonTestClient', 'JsonTestMixin', 'JsonTestCase'] class JsonTestClient(Client): def _json_request(self, method, url, data=None, *args, **kwargs): method_func = getattr(super(JsonTestClient, self), method) if method == 'get': encode = lambda x: x else: encode = json.dumps if data is not None: resp = method_func(url, encode(data), content_type='application/json', *args, **kwargs) else: resp = method_func(url, content_type='application/json', *args, **kwargs) if resp['Content-Type'].startswith('application/json') and resp.content: charset = resp.charset or settings.DEFAULT_CHARSET resp.json = json.loads(resp.content.decode(charset)) return resp def __getattribute__(self, attr): if attr in ('get', 'post', 'put', 'delete', 'trace', 'head', 'patch', 'options'): return functools.partial(self._json_request, attr) else: return super(JsonTestClient, self).__getattribute__(attr) class JsonTestMixin(object): client_class = JsonTestClient class JsonTestCase(JsonTestMixin, TestCase): pass
<commit_before>import json import functools from django.conf import settings from django.test import Client, TestCase __all__ = ['JsonTestClient', 'JsonTestCase'] class JsonTestClient(Client): def _json_request(self, method, url, data=None, *args, **kwargs): method_func = getattr(super(JsonTestClient, self), method) if method == 'get': encode = lambda x: x else: encode = json.dumps if data is not None: resp = method_func(url, encode(data), content_type='application/json', *args, **kwargs) else: resp = method_func(url, content_type='application/json', *args, **kwargs) if resp['Content-Type'].startswith('application/json') and resp.content: charset = resp.charset or settings.DEFAULT_CHARSET resp.json = json.loads(resp.content.decode(charset)) return resp def __getattribute__(self, attr): if attr in ('get', 'post', 'put', 'delete', 'trace', 'head', 'patch', 'options'): return functools.partial(self._json_request, attr) else: return super(JsonTestClient, self).__getattribute__(attr) class JsonTestCase(TestCase): client_class = JsonTestClient <commit_msg>Make the TestCase a mixin<commit_after>
import json import functools from django.conf import settings from django.test import Client, TestCase __all__ = ['JsonTestClient', 'JsonTestMixin', 'JsonTestCase'] class JsonTestClient(Client): def _json_request(self, method, url, data=None, *args, **kwargs): method_func = getattr(super(JsonTestClient, self), method) if method == 'get': encode = lambda x: x else: encode = json.dumps if data is not None: resp = method_func(url, encode(data), content_type='application/json', *args, **kwargs) else: resp = method_func(url, content_type='application/json', *args, **kwargs) if resp['Content-Type'].startswith('application/json') and resp.content: charset = resp.charset or settings.DEFAULT_CHARSET resp.json = json.loads(resp.content.decode(charset)) return resp def __getattribute__(self, attr): if attr in ('get', 'post', 'put', 'delete', 'trace', 'head', 'patch', 'options'): return functools.partial(self._json_request, attr) else: return super(JsonTestClient, self).__getattribute__(attr) class JsonTestMixin(object): client_class = JsonTestClient class JsonTestCase(JsonTestMixin, TestCase): pass
import json import functools from django.conf import settings from django.test import Client, TestCase __all__ = ['JsonTestClient', 'JsonTestCase'] class JsonTestClient(Client): def _json_request(self, method, url, data=None, *args, **kwargs): method_func = getattr(super(JsonTestClient, self), method) if method == 'get': encode = lambda x: x else: encode = json.dumps if data is not None: resp = method_func(url, encode(data), content_type='application/json', *args, **kwargs) else: resp = method_func(url, content_type='application/json', *args, **kwargs) if resp['Content-Type'].startswith('application/json') and resp.content: charset = resp.charset or settings.DEFAULT_CHARSET resp.json = json.loads(resp.content.decode(charset)) return resp def __getattribute__(self, attr): if attr in ('get', 'post', 'put', 'delete', 'trace', 'head', 'patch', 'options'): return functools.partial(self._json_request, attr) else: return super(JsonTestClient, self).__getattribute__(attr) class JsonTestCase(TestCase): client_class = JsonTestClient Make the TestCase a mixinimport json import functools from django.conf import settings from django.test import Client, TestCase __all__ = ['JsonTestClient', 'JsonTestMixin', 'JsonTestCase'] class JsonTestClient(Client): def _json_request(self, method, url, data=None, *args, **kwargs): method_func = getattr(super(JsonTestClient, self), method) if method == 'get': encode = lambda x: x else: encode = json.dumps if data is not None: resp = method_func(url, encode(data), content_type='application/json', *args, **kwargs) else: resp = method_func(url, content_type='application/json', *args, **kwargs) if resp['Content-Type'].startswith('application/json') and resp.content: charset = resp.charset or settings.DEFAULT_CHARSET resp.json = json.loads(resp.content.decode(charset)) return resp def __getattribute__(self, attr): if attr in ('get', 'post', 'put', 'delete', 'trace', 'head', 'patch', 'options'): return functools.partial(self._json_request, attr) else: return super(JsonTestClient, self).__getattribute__(attr) class JsonTestMixin(object): client_class = JsonTestClient class JsonTestCase(JsonTestMixin, TestCase): pass
<commit_before>import json import functools from django.conf import settings from django.test import Client, TestCase __all__ = ['JsonTestClient', 'JsonTestCase'] class JsonTestClient(Client): def _json_request(self, method, url, data=None, *args, **kwargs): method_func = getattr(super(JsonTestClient, self), method) if method == 'get': encode = lambda x: x else: encode = json.dumps if data is not None: resp = method_func(url, encode(data), content_type='application/json', *args, **kwargs) else: resp = method_func(url, content_type='application/json', *args, **kwargs) if resp['Content-Type'].startswith('application/json') and resp.content: charset = resp.charset or settings.DEFAULT_CHARSET resp.json = json.loads(resp.content.decode(charset)) return resp def __getattribute__(self, attr): if attr in ('get', 'post', 'put', 'delete', 'trace', 'head', 'patch', 'options'): return functools.partial(self._json_request, attr) else: return super(JsonTestClient, self).__getattribute__(attr) class JsonTestCase(TestCase): client_class = JsonTestClient <commit_msg>Make the TestCase a mixin<commit_after>import json import functools from django.conf import settings from django.test import Client, TestCase __all__ = ['JsonTestClient', 'JsonTestMixin', 'JsonTestCase'] class JsonTestClient(Client): def _json_request(self, method, url, data=None, *args, **kwargs): method_func = getattr(super(JsonTestClient, self), method) if method == 'get': encode = lambda x: x else: encode = json.dumps if data is not None: resp = method_func(url, encode(data), content_type='application/json', *args, **kwargs) else: resp = method_func(url, content_type='application/json', *args, **kwargs) if resp['Content-Type'].startswith('application/json') and resp.content: charset = resp.charset or settings.DEFAULT_CHARSET resp.json = json.loads(resp.content.decode(charset)) return resp def __getattribute__(self, attr): if attr in ('get', 'post', 'put', 'delete', 'trace', 'head', 'patch', 'options'): return functools.partial(self._json_request, attr) else: return super(JsonTestClient, self).__getattribute__(attr) class JsonTestMixin(object): client_class = JsonTestClient class JsonTestCase(JsonTestMixin, TestCase): pass
456e5a63333e683b7167bf151b97a49a5cf5c5fe
app/models/job.py
app/models/job.py
# Copyright (C) 2014 Linaro Ltd. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from base import BaseDocument JOB_COLLECTION = 'job' class JobDocument(BaseDocument): JOB_ID_FORMAT = '%s-%s' def __init__(self, name): super(JobDocument, self).__init__(name) @property def collection(self): return JOB_COLLECTION
# Copyright (C) 2014 Linaro Ltd. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from base import BaseDocument JOB_COLLECTION = 'job' class JobDocument(BaseDocument): JOB_ID_FORMAT = '%s-%s' def __init__(self, name, job=None, kernel=None): super(JobDocument, self).__init__(name) self._private = False self._job = job self._kernel = kernel self._created = None @property def collection(self): return JOB_COLLECTION @property def private(self): return self._private @private.setter def private(self, value): self._private = value @property def job(self): return self._job @job.setter def job(self, value): self._job = value @property def kernel(self): return self._kernel @kernel.setter def kernel(self, value): self._kernel = value @property def created(self): return self._created @created.setter def created(self, value): self._created = value def to_dict(self): job_dict = super(JobDocument, self).to_dict() job_dict['private'] = self._private job_dict['job'] = self._job job_dict['kernel'] = self._kernel job_dict['created'] = str(self._created) return job_dict
Rework the jod document model.
Rework the jod document model. * Add the created field that will store a datetime object. * Add reference to the kernel and the job inside the document, without relying on the Jod document name itself. Since we use the dash as a separator, and other job names can have dash in them, we cannot separate job from kernel easily.
Python
agpl-3.0
joyxu/kernelci-backend,joyxu/kernelci-backend,kernelci/kernelci-backend,kernelci/kernelci-backend,joyxu/kernelci-backend
# Copyright (C) 2014 Linaro Ltd. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from base import BaseDocument JOB_COLLECTION = 'job' class JobDocument(BaseDocument): JOB_ID_FORMAT = '%s-%s' def __init__(self, name): super(JobDocument, self).__init__(name) @property def collection(self): return JOB_COLLECTION Rework the jod document model. * Add the created field that will store a datetime object. * Add reference to the kernel and the job inside the document, without relying on the Jod document name itself. Since we use the dash as a separator, and other job names can have dash in them, we cannot separate job from kernel easily.
# Copyright (C) 2014 Linaro Ltd. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from base import BaseDocument JOB_COLLECTION = 'job' class JobDocument(BaseDocument): JOB_ID_FORMAT = '%s-%s' def __init__(self, name, job=None, kernel=None): super(JobDocument, self).__init__(name) self._private = False self._job = job self._kernel = kernel self._created = None @property def collection(self): return JOB_COLLECTION @property def private(self): return self._private @private.setter def private(self, value): self._private = value @property def job(self): return self._job @job.setter def job(self, value): self._job = value @property def kernel(self): return self._kernel @kernel.setter def kernel(self, value): self._kernel = value @property def created(self): return self._created @created.setter def created(self, value): self._created = value def to_dict(self): job_dict = super(JobDocument, self).to_dict() job_dict['private'] = self._private job_dict['job'] = self._job job_dict['kernel'] = self._kernel job_dict['created'] = str(self._created) return job_dict
<commit_before># Copyright (C) 2014 Linaro Ltd. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from base import BaseDocument JOB_COLLECTION = 'job' class JobDocument(BaseDocument): JOB_ID_FORMAT = '%s-%s' def __init__(self, name): super(JobDocument, self).__init__(name) @property def collection(self): return JOB_COLLECTION <commit_msg>Rework the jod document model. * Add the created field that will store a datetime object. * Add reference to the kernel and the job inside the document, without relying on the Jod document name itself. Since we use the dash as a separator, and other job names can have dash in them, we cannot separate job from kernel easily.<commit_after>
# Copyright (C) 2014 Linaro Ltd. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from base import BaseDocument JOB_COLLECTION = 'job' class JobDocument(BaseDocument): JOB_ID_FORMAT = '%s-%s' def __init__(self, name, job=None, kernel=None): super(JobDocument, self).__init__(name) self._private = False self._job = job self._kernel = kernel self._created = None @property def collection(self): return JOB_COLLECTION @property def private(self): return self._private @private.setter def private(self, value): self._private = value @property def job(self): return self._job @job.setter def job(self, value): self._job = value @property def kernel(self): return self._kernel @kernel.setter def kernel(self, value): self._kernel = value @property def created(self): return self._created @created.setter def created(self, value): self._created = value def to_dict(self): job_dict = super(JobDocument, self).to_dict() job_dict['private'] = self._private job_dict['job'] = self._job job_dict['kernel'] = self._kernel job_dict['created'] = str(self._created) return job_dict
# Copyright (C) 2014 Linaro Ltd. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from base import BaseDocument JOB_COLLECTION = 'job' class JobDocument(BaseDocument): JOB_ID_FORMAT = '%s-%s' def __init__(self, name): super(JobDocument, self).__init__(name) @property def collection(self): return JOB_COLLECTION Rework the jod document model. * Add the created field that will store a datetime object. * Add reference to the kernel and the job inside the document, without relying on the Jod document name itself. Since we use the dash as a separator, and other job names can have dash in them, we cannot separate job from kernel easily.# Copyright (C) 2014 Linaro Ltd. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from base import BaseDocument JOB_COLLECTION = 'job' class JobDocument(BaseDocument): JOB_ID_FORMAT = '%s-%s' def __init__(self, name, job=None, kernel=None): super(JobDocument, self).__init__(name) self._private = False self._job = job self._kernel = kernel self._created = None @property def collection(self): return JOB_COLLECTION @property def private(self): return self._private @private.setter def private(self, value): self._private = value @property def job(self): return self._job @job.setter def job(self, value): self._job = value @property def kernel(self): return self._kernel @kernel.setter def kernel(self, value): self._kernel = value @property def created(self): return self._created @created.setter def created(self, value): self._created = value def to_dict(self): job_dict = super(JobDocument, self).to_dict() job_dict['private'] = self._private job_dict['job'] = self._job job_dict['kernel'] = self._kernel job_dict['created'] = str(self._created) return job_dict
<commit_before># Copyright (C) 2014 Linaro Ltd. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from base import BaseDocument JOB_COLLECTION = 'job' class JobDocument(BaseDocument): JOB_ID_FORMAT = '%s-%s' def __init__(self, name): super(JobDocument, self).__init__(name) @property def collection(self): return JOB_COLLECTION <commit_msg>Rework the jod document model. * Add the created field that will store a datetime object. * Add reference to the kernel and the job inside the document, without relying on the Jod document name itself. Since we use the dash as a separator, and other job names can have dash in them, we cannot separate job from kernel easily.<commit_after># Copyright (C) 2014 Linaro Ltd. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from base import BaseDocument JOB_COLLECTION = 'job' class JobDocument(BaseDocument): JOB_ID_FORMAT = '%s-%s' def __init__(self, name, job=None, kernel=None): super(JobDocument, self).__init__(name) self._private = False self._job = job self._kernel = kernel self._created = None @property def collection(self): return JOB_COLLECTION @property def private(self): return self._private @private.setter def private(self, value): self._private = value @property def job(self): return self._job @job.setter def job(self, value): self._job = value @property def kernel(self): return self._kernel @kernel.setter def kernel(self, value): self._kernel = value @property def created(self): return self._created @created.setter def created(self, value): self._created = value def to_dict(self): job_dict = super(JobDocument, self).to_dict() job_dict['private'] = self._private job_dict['job'] = self._job job_dict['kernel'] = self._kernel job_dict['created'] = str(self._created) return job_dict
4b7713a1891aa86c0f16fafdea8770495070bfcb
html_snapshots/utils.py
html_snapshots/utils.py
import os import rmc.shared.constants as c import rmc.models as m FILE_DIR = os.path.dirname(os.path.realpath(__file__)) HTML_DIR = os.path.join(c.SHARED_DATA_DIR, 'html_snapshots') def write(file_path, content): ensure_dir(file_path) with open(file_path, 'w') as f: f.write(content) def ensure_dir(file_path): d = os.path.dirname(file_path) if not os.path.exists(d): os.makedirs(d) def generate_urls(): urls = [] # Home page urls.append('') # Course pages for course in m.Course.objects: course_id = course.id urls.append('course/' + course_id) return urls
import os import mongoengine as me import rmc.shared.constants as c import rmc.models as m FILE_DIR = os.path.dirname(os.path.realpath(__file__)) HTML_DIR = os.path.join(c.SHARED_DATA_DIR, 'html_snapshots') me.connect(c.MONGO_DB_RMC, host=c.MONGO_HOST, port=c.MONGO_PORT) def write(file_path, content): ensure_dir(file_path) with open(file_path, 'w') as f: f.write(content) def ensure_dir(file_path): d = os.path.dirname(file_path) if not os.path.exists(d): os.makedirs(d) def generate_urls(): urls = [] # Home page urls.append('') # Course pages for course in m.Course.objects: course_id = course.id urls.append('course/' + course_id) return urls
Create mongoengine connection when taking phantom snapshots
Create mongoengine connection when taking phantom snapshots
Python
mit
ccqi/rmc,JGulbronson/rmc,JGulbronson/rmc,UWFlow/rmc,shakilkanji/rmc,ccqi/rmc,UWFlow/rmc,sachdevs/rmc,sachdevs/rmc,MichalKononenko/rmc,UWFlow/rmc,MichalKononenko/rmc,sachdevs/rmc,duaayousif/rmc,MichalKononenko/rmc,MichalKononenko/rmc,JGulbronson/rmc,ccqi/rmc,JGulbronson/rmc,sachdevs/rmc,UWFlow/rmc,UWFlow/rmc,duaayousif/rmc,ccqi/rmc,shakilkanji/rmc,ccqi/rmc,rageandqq/rmc,JGulbronson/rmc,rageandqq/rmc,rageandqq/rmc,MichalKononenko/rmc,shakilkanji/rmc,duaayousif/rmc,shakilkanji/rmc,duaayousif/rmc,rageandqq/rmc,rageandqq/rmc,sachdevs/rmc,duaayousif/rmc,shakilkanji/rmc
import os import rmc.shared.constants as c import rmc.models as m FILE_DIR = os.path.dirname(os.path.realpath(__file__)) HTML_DIR = os.path.join(c.SHARED_DATA_DIR, 'html_snapshots') def write(file_path, content): ensure_dir(file_path) with open(file_path, 'w') as f: f.write(content) def ensure_dir(file_path): d = os.path.dirname(file_path) if not os.path.exists(d): os.makedirs(d) def generate_urls(): urls = [] # Home page urls.append('') # Course pages for course in m.Course.objects: course_id = course.id urls.append('course/' + course_id) return urls Create mongoengine connection when taking phantom snapshots
import os import mongoengine as me import rmc.shared.constants as c import rmc.models as m FILE_DIR = os.path.dirname(os.path.realpath(__file__)) HTML_DIR = os.path.join(c.SHARED_DATA_DIR, 'html_snapshots') me.connect(c.MONGO_DB_RMC, host=c.MONGO_HOST, port=c.MONGO_PORT) def write(file_path, content): ensure_dir(file_path) with open(file_path, 'w') as f: f.write(content) def ensure_dir(file_path): d = os.path.dirname(file_path) if not os.path.exists(d): os.makedirs(d) def generate_urls(): urls = [] # Home page urls.append('') # Course pages for course in m.Course.objects: course_id = course.id urls.append('course/' + course_id) return urls
<commit_before>import os import rmc.shared.constants as c import rmc.models as m FILE_DIR = os.path.dirname(os.path.realpath(__file__)) HTML_DIR = os.path.join(c.SHARED_DATA_DIR, 'html_snapshots') def write(file_path, content): ensure_dir(file_path) with open(file_path, 'w') as f: f.write(content) def ensure_dir(file_path): d = os.path.dirname(file_path) if not os.path.exists(d): os.makedirs(d) def generate_urls(): urls = [] # Home page urls.append('') # Course pages for course in m.Course.objects: course_id = course.id urls.append('course/' + course_id) return urls <commit_msg>Create mongoengine connection when taking phantom snapshots<commit_after>
import os import mongoengine as me import rmc.shared.constants as c import rmc.models as m FILE_DIR = os.path.dirname(os.path.realpath(__file__)) HTML_DIR = os.path.join(c.SHARED_DATA_DIR, 'html_snapshots') me.connect(c.MONGO_DB_RMC, host=c.MONGO_HOST, port=c.MONGO_PORT) def write(file_path, content): ensure_dir(file_path) with open(file_path, 'w') as f: f.write(content) def ensure_dir(file_path): d = os.path.dirname(file_path) if not os.path.exists(d): os.makedirs(d) def generate_urls(): urls = [] # Home page urls.append('') # Course pages for course in m.Course.objects: course_id = course.id urls.append('course/' + course_id) return urls
import os import rmc.shared.constants as c import rmc.models as m FILE_DIR = os.path.dirname(os.path.realpath(__file__)) HTML_DIR = os.path.join(c.SHARED_DATA_DIR, 'html_snapshots') def write(file_path, content): ensure_dir(file_path) with open(file_path, 'w') as f: f.write(content) def ensure_dir(file_path): d = os.path.dirname(file_path) if not os.path.exists(d): os.makedirs(d) def generate_urls(): urls = [] # Home page urls.append('') # Course pages for course in m.Course.objects: course_id = course.id urls.append('course/' + course_id) return urls Create mongoengine connection when taking phantom snapshotsimport os import mongoengine as me import rmc.shared.constants as c import rmc.models as m FILE_DIR = os.path.dirname(os.path.realpath(__file__)) HTML_DIR = os.path.join(c.SHARED_DATA_DIR, 'html_snapshots') me.connect(c.MONGO_DB_RMC, host=c.MONGO_HOST, port=c.MONGO_PORT) def write(file_path, content): ensure_dir(file_path) with open(file_path, 'w') as f: f.write(content) def ensure_dir(file_path): d = os.path.dirname(file_path) if not os.path.exists(d): os.makedirs(d) def generate_urls(): urls = [] # Home page urls.append('') # Course pages for course in m.Course.objects: course_id = course.id urls.append('course/' + course_id) return urls
<commit_before>import os import rmc.shared.constants as c import rmc.models as m FILE_DIR = os.path.dirname(os.path.realpath(__file__)) HTML_DIR = os.path.join(c.SHARED_DATA_DIR, 'html_snapshots') def write(file_path, content): ensure_dir(file_path) with open(file_path, 'w') as f: f.write(content) def ensure_dir(file_path): d = os.path.dirname(file_path) if not os.path.exists(d): os.makedirs(d) def generate_urls(): urls = [] # Home page urls.append('') # Course pages for course in m.Course.objects: course_id = course.id urls.append('course/' + course_id) return urls <commit_msg>Create mongoengine connection when taking phantom snapshots<commit_after>import os import mongoengine as me import rmc.shared.constants as c import rmc.models as m FILE_DIR = os.path.dirname(os.path.realpath(__file__)) HTML_DIR = os.path.join(c.SHARED_DATA_DIR, 'html_snapshots') me.connect(c.MONGO_DB_RMC, host=c.MONGO_HOST, port=c.MONGO_PORT) def write(file_path, content): ensure_dir(file_path) with open(file_path, 'w') as f: f.write(content) def ensure_dir(file_path): d = os.path.dirname(file_path) if not os.path.exists(d): os.makedirs(d) def generate_urls(): urls = [] # Home page urls.append('') # Course pages for course in m.Course.objects: course_id = course.id urls.append('course/' + course_id) return urls
e3dcbe5fb142b7ce564a90cf127de418d0a62db3
src/sentry/runner/hacks.py
src/sentry/runner/hacks.py
""" sentry.runner.hacks ~~~~~~~~~~~~~~~~~~~ :copyright: (c) 2015 by the Sentry Team, see AUTHORS for more details. :license: BSD, see LICENSE for more details. """ from __future__ import absolute_import, print_function from sentry.http import get_server_hostname class AllowedHosts(object): # HACK: This is a fake stub for settings.ALLOWED_HOSTS # This is needing since ALLOWED_HOSTS is engrained # in Django internals, so we want this "tuple" to respond # to runtime changes based on our system.url-prefix Option def __iter__(self): yield get_server_hostname() or '*'
""" sentry.runner.hacks ~~~~~~~~~~~~~~~~~~~ :copyright: (c) 2015 by the Sentry Team, see AUTHORS for more details. :license: BSD, see LICENSE for more details. """ from __future__ import absolute_import, print_function from sentry.http import get_server_hostname class AllowedHosts(object): # HACK: This is a fake stub for settings.ALLOWED_HOSTS # This is needing since ALLOWED_HOSTS is engrained # in Django internals, so we want this "tuple" to respond # to runtime changes based on our system.url-prefix Option def __iter__(self): yield get_server_hostname() or '*' def __repr__(self): return repr(tuple(self))
Add a nice repr for AllowedHosts object so the admin makes sense
Add a nice repr for AllowedHosts object so the admin makes sense
Python
bsd-3-clause
fotinakis/sentry,daevaorn/sentry,beeftornado/sentry,daevaorn/sentry,ifduyue/sentry,alexm92/sentry,ifduyue/sentry,JamesMura/sentry,JamesMura/sentry,alexm92/sentry,jean/sentry,BuildingLink/sentry,JackDanger/sentry,gencer/sentry,nicholasserra/sentry,jean/sentry,ifduyue/sentry,mvaled/sentry,ifduyue/sentry,zenefits/sentry,zenefits/sentry,fotinakis/sentry,JamesMura/sentry,gencer/sentry,nicholasserra/sentry,gencer/sentry,mvaled/sentry,looker/sentry,looker/sentry,BuildingLink/sentry,mvaled/sentry,jean/sentry,beeftornado/sentry,mvaled/sentry,JackDanger/sentry,fotinakis/sentry,JamesMura/sentry,mitsuhiko/sentry,looker/sentry,daevaorn/sentry,JamesMura/sentry,mvaled/sentry,mitsuhiko/sentry,ifduyue/sentry,daevaorn/sentry,jean/sentry,nicholasserra/sentry,BuildingLink/sentry,zenefits/sentry,jean/sentry,zenefits/sentry,BuildingLink/sentry,zenefits/sentry,gencer/sentry,looker/sentry,fotinakis/sentry,mvaled/sentry,looker/sentry,JackDanger/sentry,BuildingLink/sentry,gencer/sentry,beeftornado/sentry,alexm92/sentry
""" sentry.runner.hacks ~~~~~~~~~~~~~~~~~~~ :copyright: (c) 2015 by the Sentry Team, see AUTHORS for more details. :license: BSD, see LICENSE for more details. """ from __future__ import absolute_import, print_function from sentry.http import get_server_hostname class AllowedHosts(object): # HACK: This is a fake stub for settings.ALLOWED_HOSTS # This is needing since ALLOWED_HOSTS is engrained # in Django internals, so we want this "tuple" to respond # to runtime changes based on our system.url-prefix Option def __iter__(self): yield get_server_hostname() or '*' Add a nice repr for AllowedHosts object so the admin makes sense
""" sentry.runner.hacks ~~~~~~~~~~~~~~~~~~~ :copyright: (c) 2015 by the Sentry Team, see AUTHORS for more details. :license: BSD, see LICENSE for more details. """ from __future__ import absolute_import, print_function from sentry.http import get_server_hostname class AllowedHosts(object): # HACK: This is a fake stub for settings.ALLOWED_HOSTS # This is needing since ALLOWED_HOSTS is engrained # in Django internals, so we want this "tuple" to respond # to runtime changes based on our system.url-prefix Option def __iter__(self): yield get_server_hostname() or '*' def __repr__(self): return repr(tuple(self))
<commit_before>""" sentry.runner.hacks ~~~~~~~~~~~~~~~~~~~ :copyright: (c) 2015 by the Sentry Team, see AUTHORS for more details. :license: BSD, see LICENSE for more details. """ from __future__ import absolute_import, print_function from sentry.http import get_server_hostname class AllowedHosts(object): # HACK: This is a fake stub for settings.ALLOWED_HOSTS # This is needing since ALLOWED_HOSTS is engrained # in Django internals, so we want this "tuple" to respond # to runtime changes based on our system.url-prefix Option def __iter__(self): yield get_server_hostname() or '*' <commit_msg>Add a nice repr for AllowedHosts object so the admin makes sense<commit_after>
""" sentry.runner.hacks ~~~~~~~~~~~~~~~~~~~ :copyright: (c) 2015 by the Sentry Team, see AUTHORS for more details. :license: BSD, see LICENSE for more details. """ from __future__ import absolute_import, print_function from sentry.http import get_server_hostname class AllowedHosts(object): # HACK: This is a fake stub for settings.ALLOWED_HOSTS # This is needing since ALLOWED_HOSTS is engrained # in Django internals, so we want this "tuple" to respond # to runtime changes based on our system.url-prefix Option def __iter__(self): yield get_server_hostname() or '*' def __repr__(self): return repr(tuple(self))
""" sentry.runner.hacks ~~~~~~~~~~~~~~~~~~~ :copyright: (c) 2015 by the Sentry Team, see AUTHORS for more details. :license: BSD, see LICENSE for more details. """ from __future__ import absolute_import, print_function from sentry.http import get_server_hostname class AllowedHosts(object): # HACK: This is a fake stub for settings.ALLOWED_HOSTS # This is needing since ALLOWED_HOSTS is engrained # in Django internals, so we want this "tuple" to respond # to runtime changes based on our system.url-prefix Option def __iter__(self): yield get_server_hostname() or '*' Add a nice repr for AllowedHosts object so the admin makes sense""" sentry.runner.hacks ~~~~~~~~~~~~~~~~~~~ :copyright: (c) 2015 by the Sentry Team, see AUTHORS for more details. :license: BSD, see LICENSE for more details. """ from __future__ import absolute_import, print_function from sentry.http import get_server_hostname class AllowedHosts(object): # HACK: This is a fake stub for settings.ALLOWED_HOSTS # This is needing since ALLOWED_HOSTS is engrained # in Django internals, so we want this "tuple" to respond # to runtime changes based on our system.url-prefix Option def __iter__(self): yield get_server_hostname() or '*' def __repr__(self): return repr(tuple(self))
<commit_before>""" sentry.runner.hacks ~~~~~~~~~~~~~~~~~~~ :copyright: (c) 2015 by the Sentry Team, see AUTHORS for more details. :license: BSD, see LICENSE for more details. """ from __future__ import absolute_import, print_function from sentry.http import get_server_hostname class AllowedHosts(object): # HACK: This is a fake stub for settings.ALLOWED_HOSTS # This is needing since ALLOWED_HOSTS is engrained # in Django internals, so we want this "tuple" to respond # to runtime changes based on our system.url-prefix Option def __iter__(self): yield get_server_hostname() or '*' <commit_msg>Add a nice repr for AllowedHosts object so the admin makes sense<commit_after>""" sentry.runner.hacks ~~~~~~~~~~~~~~~~~~~ :copyright: (c) 2015 by the Sentry Team, see AUTHORS for more details. :license: BSD, see LICENSE for more details. """ from __future__ import absolute_import, print_function from sentry.http import get_server_hostname class AllowedHosts(object): # HACK: This is a fake stub for settings.ALLOWED_HOSTS # This is needing since ALLOWED_HOSTS is engrained # in Django internals, so we want this "tuple" to respond # to runtime changes based on our system.url-prefix Option def __iter__(self): yield get_server_hostname() or '*' def __repr__(self): return repr(tuple(self))
1ef9c5c76e1c1110803d83833fc69ba201d33582
pavement.py
pavement.py
# -*- coding: utf-8 -*- from paver.easy import task, sh, needs, path from paver.setuputils import setup setup(name='ldapom', version='0.9.4', description='A simple ldap object mapper for python', author='Florian Richter', author_email='mail@f1ori.de', url='https://github.com/HaDiNet/ldapom', license='MIT', keywords = "ldap object mapper", long_description=path('README').text(), py_modules=['ldapom'], ) @task def docs(options): sh('doxygen') @task def test(options): sh('python tests.py') @task def coverage(options): sh('coverage run --source ldapom.py ./tests.py') sh('coverage xml')
# -*- coding: utf-8 -*- from paver.easy import * @task def test(options): info("Running tests for Python 2") sh('python2 tests.py') info("Running tests for Python 3") sh('python3 tests.py') @task def coverage(options): info("Running coverage for Python 2") sh('coverage2 run --source ldapom.py ./tests.py') sh('coverage2 report') info("Running coverage for Python 3") sh('coverage3 run --source ldapom.py ./tests.py') sh('coverage3 report')
Update paver to test and report coverage for both Python 2 and 3
Update paver to test and report coverage for both Python 2 and 3
Python
mit
HaDiNet/ldapom
# -*- coding: utf-8 -*- from paver.easy import task, sh, needs, path from paver.setuputils import setup setup(name='ldapom', version='0.9.4', description='A simple ldap object mapper for python', author='Florian Richter', author_email='mail@f1ori.de', url='https://github.com/HaDiNet/ldapom', license='MIT', keywords = "ldap object mapper", long_description=path('README').text(), py_modules=['ldapom'], ) @task def docs(options): sh('doxygen') @task def test(options): sh('python tests.py') @task def coverage(options): sh('coverage run --source ldapom.py ./tests.py') sh('coverage xml') Update paver to test and report coverage for both Python 2 and 3
# -*- coding: utf-8 -*- from paver.easy import * @task def test(options): info("Running tests for Python 2") sh('python2 tests.py') info("Running tests for Python 3") sh('python3 tests.py') @task def coverage(options): info("Running coverage for Python 2") sh('coverage2 run --source ldapom.py ./tests.py') sh('coverage2 report') info("Running coverage for Python 3") sh('coverage3 run --source ldapom.py ./tests.py') sh('coverage3 report')
<commit_before># -*- coding: utf-8 -*- from paver.easy import task, sh, needs, path from paver.setuputils import setup setup(name='ldapom', version='0.9.4', description='A simple ldap object mapper for python', author='Florian Richter', author_email='mail@f1ori.de', url='https://github.com/HaDiNet/ldapom', license='MIT', keywords = "ldap object mapper", long_description=path('README').text(), py_modules=['ldapom'], ) @task def docs(options): sh('doxygen') @task def test(options): sh('python tests.py') @task def coverage(options): sh('coverage run --source ldapom.py ./tests.py') sh('coverage xml') <commit_msg>Update paver to test and report coverage for both Python 2 and 3<commit_after>
# -*- coding: utf-8 -*- from paver.easy import * @task def test(options): info("Running tests for Python 2") sh('python2 tests.py') info("Running tests for Python 3") sh('python3 tests.py') @task def coverage(options): info("Running coverage for Python 2") sh('coverage2 run --source ldapom.py ./tests.py') sh('coverage2 report') info("Running coverage for Python 3") sh('coverage3 run --source ldapom.py ./tests.py') sh('coverage3 report')
# -*- coding: utf-8 -*- from paver.easy import task, sh, needs, path from paver.setuputils import setup setup(name='ldapom', version='0.9.4', description='A simple ldap object mapper for python', author='Florian Richter', author_email='mail@f1ori.de', url='https://github.com/HaDiNet/ldapom', license='MIT', keywords = "ldap object mapper", long_description=path('README').text(), py_modules=['ldapom'], ) @task def docs(options): sh('doxygen') @task def test(options): sh('python tests.py') @task def coverage(options): sh('coverage run --source ldapom.py ./tests.py') sh('coverage xml') Update paver to test and report coverage for both Python 2 and 3# -*- coding: utf-8 -*- from paver.easy import * @task def test(options): info("Running tests for Python 2") sh('python2 tests.py') info("Running tests for Python 3") sh('python3 tests.py') @task def coverage(options): info("Running coverage for Python 2") sh('coverage2 run --source ldapom.py ./tests.py') sh('coverage2 report') info("Running coverage for Python 3") sh('coverage3 run --source ldapom.py ./tests.py') sh('coverage3 report')
<commit_before># -*- coding: utf-8 -*- from paver.easy import task, sh, needs, path from paver.setuputils import setup setup(name='ldapom', version='0.9.4', description='A simple ldap object mapper for python', author='Florian Richter', author_email='mail@f1ori.de', url='https://github.com/HaDiNet/ldapom', license='MIT', keywords = "ldap object mapper", long_description=path('README').text(), py_modules=['ldapom'], ) @task def docs(options): sh('doxygen') @task def test(options): sh('python tests.py') @task def coverage(options): sh('coverage run --source ldapom.py ./tests.py') sh('coverage xml') <commit_msg>Update paver to test and report coverage for both Python 2 and 3<commit_after># -*- coding: utf-8 -*- from paver.easy import * @task def test(options): info("Running tests for Python 2") sh('python2 tests.py') info("Running tests for Python 3") sh('python3 tests.py') @task def coverage(options): info("Running coverage for Python 2") sh('coverage2 run --source ldapom.py ./tests.py') sh('coverage2 report') info("Running coverage for Python 3") sh('coverage3 run --source ldapom.py ./tests.py') sh('coverage3 report')
05742ef5fb1f163750679ae56ab5ea97f7050811
tests/test_apps_up.py
tests/test_apps_up.py
import os import pytest import requests REGISTER_TITLE_URL = os.environ['DIGITAL_REGISTER_URL'] USERNAME = os.environ['SMOKE_USERNAME'] PASSWORD = os.environ['SMOKE_PASSWORD'] TITLE_NUMBER = os.environ['SMOKE_TITLE_NUMBER'] PARTIAL_ADDRESS = os.environ['SMOKE_PARTIAL_ADDRESS'] POSTCODE = os.environ['SMOKE_POSTCODE'] def test_frontend_up(): # login stuff response = requests.post('{}/login?next=titles'.format(REGISTER_TITLE_URL), data={'username': USERNAME, 'password': PASSWORD}, follow_redirects=False) import pdb; pdb.set_trace()
import os import pytest import requests REGISTER_TITLE_URL = os.environ['DIGITAL_REGISTER_URL'] USERNAME = os.environ['SMOKE_USERNAME'] PASSWORD = os.environ['SMOKE_PASSWORD'] TITLE_NUMBER = os.environ['SMOKE_TITLE_NUMBER'] PARTIAL_ADDRESS = os.environ['SMOKE_PARTIAL_ADDRESS'] POSTCODE = os.environ['SMOKE_POSTCODE'] def test_frontend_up(): # login stuff response = requests.post('{}/login?next=titles'.format(REGISTER_TITLE_URL), data={'username': USERNAME, 'password': PASSWORD}, allow_redirects=False) import pdb; pdb.set_trace()
Use correct kwarg for requests.post()
Use correct kwarg for requests.post()
Python
mit
LandRegistry/digital-register-smoke-tests,LandRegistry/digital-register-smoke-tests
import os import pytest import requests REGISTER_TITLE_URL = os.environ['DIGITAL_REGISTER_URL'] USERNAME = os.environ['SMOKE_USERNAME'] PASSWORD = os.environ['SMOKE_PASSWORD'] TITLE_NUMBER = os.environ['SMOKE_TITLE_NUMBER'] PARTIAL_ADDRESS = os.environ['SMOKE_PARTIAL_ADDRESS'] POSTCODE = os.environ['SMOKE_POSTCODE'] def test_frontend_up(): # login stuff response = requests.post('{}/login?next=titles'.format(REGISTER_TITLE_URL), data={'username': USERNAME, 'password': PASSWORD}, follow_redirects=False) import pdb; pdb.set_trace() Use correct kwarg for requests.post()
import os import pytest import requests REGISTER_TITLE_URL = os.environ['DIGITAL_REGISTER_URL'] USERNAME = os.environ['SMOKE_USERNAME'] PASSWORD = os.environ['SMOKE_PASSWORD'] TITLE_NUMBER = os.environ['SMOKE_TITLE_NUMBER'] PARTIAL_ADDRESS = os.environ['SMOKE_PARTIAL_ADDRESS'] POSTCODE = os.environ['SMOKE_POSTCODE'] def test_frontend_up(): # login stuff response = requests.post('{}/login?next=titles'.format(REGISTER_TITLE_URL), data={'username': USERNAME, 'password': PASSWORD}, allow_redirects=False) import pdb; pdb.set_trace()
<commit_before>import os import pytest import requests REGISTER_TITLE_URL = os.environ['DIGITAL_REGISTER_URL'] USERNAME = os.environ['SMOKE_USERNAME'] PASSWORD = os.environ['SMOKE_PASSWORD'] TITLE_NUMBER = os.environ['SMOKE_TITLE_NUMBER'] PARTIAL_ADDRESS = os.environ['SMOKE_PARTIAL_ADDRESS'] POSTCODE = os.environ['SMOKE_POSTCODE'] def test_frontend_up(): # login stuff response = requests.post('{}/login?next=titles'.format(REGISTER_TITLE_URL), data={'username': USERNAME, 'password': PASSWORD}, follow_redirects=False) import pdb; pdb.set_trace() <commit_msg>Use correct kwarg for requests.post()<commit_after>
import os import pytest import requests REGISTER_TITLE_URL = os.environ['DIGITAL_REGISTER_URL'] USERNAME = os.environ['SMOKE_USERNAME'] PASSWORD = os.environ['SMOKE_PASSWORD'] TITLE_NUMBER = os.environ['SMOKE_TITLE_NUMBER'] PARTIAL_ADDRESS = os.environ['SMOKE_PARTIAL_ADDRESS'] POSTCODE = os.environ['SMOKE_POSTCODE'] def test_frontend_up(): # login stuff response = requests.post('{}/login?next=titles'.format(REGISTER_TITLE_URL), data={'username': USERNAME, 'password': PASSWORD}, allow_redirects=False) import pdb; pdb.set_trace()
import os import pytest import requests REGISTER_TITLE_URL = os.environ['DIGITAL_REGISTER_URL'] USERNAME = os.environ['SMOKE_USERNAME'] PASSWORD = os.environ['SMOKE_PASSWORD'] TITLE_NUMBER = os.environ['SMOKE_TITLE_NUMBER'] PARTIAL_ADDRESS = os.environ['SMOKE_PARTIAL_ADDRESS'] POSTCODE = os.environ['SMOKE_POSTCODE'] def test_frontend_up(): # login stuff response = requests.post('{}/login?next=titles'.format(REGISTER_TITLE_URL), data={'username': USERNAME, 'password': PASSWORD}, follow_redirects=False) import pdb; pdb.set_trace() Use correct kwarg for requests.post()import os import pytest import requests REGISTER_TITLE_URL = os.environ['DIGITAL_REGISTER_URL'] USERNAME = os.environ['SMOKE_USERNAME'] PASSWORD = os.environ['SMOKE_PASSWORD'] TITLE_NUMBER = os.environ['SMOKE_TITLE_NUMBER'] PARTIAL_ADDRESS = os.environ['SMOKE_PARTIAL_ADDRESS'] POSTCODE = os.environ['SMOKE_POSTCODE'] def test_frontend_up(): # login stuff response = requests.post('{}/login?next=titles'.format(REGISTER_TITLE_URL), data={'username': USERNAME, 'password': PASSWORD}, allow_redirects=False) import pdb; pdb.set_trace()
<commit_before>import os import pytest import requests REGISTER_TITLE_URL = os.environ['DIGITAL_REGISTER_URL'] USERNAME = os.environ['SMOKE_USERNAME'] PASSWORD = os.environ['SMOKE_PASSWORD'] TITLE_NUMBER = os.environ['SMOKE_TITLE_NUMBER'] PARTIAL_ADDRESS = os.environ['SMOKE_PARTIAL_ADDRESS'] POSTCODE = os.environ['SMOKE_POSTCODE'] def test_frontend_up(): # login stuff response = requests.post('{}/login?next=titles'.format(REGISTER_TITLE_URL), data={'username': USERNAME, 'password': PASSWORD}, follow_redirects=False) import pdb; pdb.set_trace() <commit_msg>Use correct kwarg for requests.post()<commit_after>import os import pytest import requests REGISTER_TITLE_URL = os.environ['DIGITAL_REGISTER_URL'] USERNAME = os.environ['SMOKE_USERNAME'] PASSWORD = os.environ['SMOKE_PASSWORD'] TITLE_NUMBER = os.environ['SMOKE_TITLE_NUMBER'] PARTIAL_ADDRESS = os.environ['SMOKE_PARTIAL_ADDRESS'] POSTCODE = os.environ['SMOKE_POSTCODE'] def test_frontend_up(): # login stuff response = requests.post('{}/login?next=titles'.format(REGISTER_TITLE_URL), data={'username': USERNAME, 'password': PASSWORD}, allow_redirects=False) import pdb; pdb.set_trace()
abe744c5a099fd988ff3fe5eb1d50cca7a633d74
var/spack/repos/builtin/packages/parallel-netcdf/package.py
var/spack/repos/builtin/packages/parallel-netcdf/package.py
from spack import * class ParallelNetcdf(Package): """Parallel netCDF (PnetCDF) is a library providing high-performance parallel I/O while still maintaining file-format compatibility with Unidata's NetCDF.""" homepage = "https://trac.mcs.anl.gov/projects/parallel-netcdf" url = "http://cucis.ece.northwestern.edu/projects/PnetCDF/Release/parallel-netcdf-1.6.1.tar.gz" version('1.6.1', '62a094eb952f9d1e15f07d56e535052604f1ac34') depends_on("m4") depends_on("mpi") def install(self, spec, prefix): configure("--prefix=%s" % prefix, "--with-mpi=%s" % spec['mpi'].prefix) make() make("install")
from spack import * class ParallelNetcdf(Package): """Parallel netCDF (PnetCDF) is a library providing high-performance parallel I/O while still maintaining file-format compatibility with Unidata's NetCDF.""" homepage = "https://trac.mcs.anl.gov/projects/parallel-netcdf" url = "http://cucis.ece.northwestern.edu/projects/PnetCDF/Release/parallel-netcdf-1.6.1.tar.gz" version('1.7.0', '267eab7b6f9dc78c4d0e6def2def3aea4bc7c9f0') version('1.6.1', '62a094eb952f9d1e15f07d56e535052604f1ac34') depends_on("m4") depends_on("mpi") def install(self, spec, prefix): configure("--prefix=%s" % prefix, "--with-mpi=%s" % spec['mpi'].prefix) make() make("install")
Add latest version of PnetCDF
Add latest version of PnetCDF
Python
lgpl-2.1
mfherbst/spack,EmreAtes/spack,lgarren/spack,tmerrick1/spack,matthiasdiener/spack,skosukhin/spack,tmerrick1/spack,iulian787/spack,skosukhin/spack,LLNL/spack,matthiasdiener/spack,iulian787/spack,krafczyk/spack,TheTimmy/spack,mfherbst/spack,LLNL/spack,mfherbst/spack,EmreAtes/spack,tmerrick1/spack,mfherbst/spack,TheTimmy/spack,krafczyk/spack,EmreAtes/spack,iulian787/spack,iulian787/spack,lgarren/spack,LLNL/spack,lgarren/spack,lgarren/spack,matthiasdiener/spack,iulian787/spack,skosukhin/spack,mfherbst/spack,lgarren/spack,matthiasdiener/spack,tmerrick1/spack,matthiasdiener/spack,LLNL/spack,skosukhin/spack,krafczyk/spack,EmreAtes/spack,LLNL/spack,TheTimmy/spack,TheTimmy/spack,skosukhin/spack,EmreAtes/spack,tmerrick1/spack,krafczyk/spack,TheTimmy/spack,krafczyk/spack
from spack import * class ParallelNetcdf(Package): """Parallel netCDF (PnetCDF) is a library providing high-performance parallel I/O while still maintaining file-format compatibility with Unidata's NetCDF.""" homepage = "https://trac.mcs.anl.gov/projects/parallel-netcdf" url = "http://cucis.ece.northwestern.edu/projects/PnetCDF/Release/parallel-netcdf-1.6.1.tar.gz" version('1.6.1', '62a094eb952f9d1e15f07d56e535052604f1ac34') depends_on("m4") depends_on("mpi") def install(self, spec, prefix): configure("--prefix=%s" % prefix, "--with-mpi=%s" % spec['mpi'].prefix) make() make("install") Add latest version of PnetCDF
from spack import * class ParallelNetcdf(Package): """Parallel netCDF (PnetCDF) is a library providing high-performance parallel I/O while still maintaining file-format compatibility with Unidata's NetCDF.""" homepage = "https://trac.mcs.anl.gov/projects/parallel-netcdf" url = "http://cucis.ece.northwestern.edu/projects/PnetCDF/Release/parallel-netcdf-1.6.1.tar.gz" version('1.7.0', '267eab7b6f9dc78c4d0e6def2def3aea4bc7c9f0') version('1.6.1', '62a094eb952f9d1e15f07d56e535052604f1ac34') depends_on("m4") depends_on("mpi") def install(self, spec, prefix): configure("--prefix=%s" % prefix, "--with-mpi=%s" % spec['mpi'].prefix) make() make("install")
<commit_before>from spack import * class ParallelNetcdf(Package): """Parallel netCDF (PnetCDF) is a library providing high-performance parallel I/O while still maintaining file-format compatibility with Unidata's NetCDF.""" homepage = "https://trac.mcs.anl.gov/projects/parallel-netcdf" url = "http://cucis.ece.northwestern.edu/projects/PnetCDF/Release/parallel-netcdf-1.6.1.tar.gz" version('1.6.1', '62a094eb952f9d1e15f07d56e535052604f1ac34') depends_on("m4") depends_on("mpi") def install(self, spec, prefix): configure("--prefix=%s" % prefix, "--with-mpi=%s" % spec['mpi'].prefix) make() make("install") <commit_msg>Add latest version of PnetCDF<commit_after>
from spack import * class ParallelNetcdf(Package): """Parallel netCDF (PnetCDF) is a library providing high-performance parallel I/O while still maintaining file-format compatibility with Unidata's NetCDF.""" homepage = "https://trac.mcs.anl.gov/projects/parallel-netcdf" url = "http://cucis.ece.northwestern.edu/projects/PnetCDF/Release/parallel-netcdf-1.6.1.tar.gz" version('1.7.0', '267eab7b6f9dc78c4d0e6def2def3aea4bc7c9f0') version('1.6.1', '62a094eb952f9d1e15f07d56e535052604f1ac34') depends_on("m4") depends_on("mpi") def install(self, spec, prefix): configure("--prefix=%s" % prefix, "--with-mpi=%s" % spec['mpi'].prefix) make() make("install")
from spack import * class ParallelNetcdf(Package): """Parallel netCDF (PnetCDF) is a library providing high-performance parallel I/O while still maintaining file-format compatibility with Unidata's NetCDF.""" homepage = "https://trac.mcs.anl.gov/projects/parallel-netcdf" url = "http://cucis.ece.northwestern.edu/projects/PnetCDF/Release/parallel-netcdf-1.6.1.tar.gz" version('1.6.1', '62a094eb952f9d1e15f07d56e535052604f1ac34') depends_on("m4") depends_on("mpi") def install(self, spec, prefix): configure("--prefix=%s" % prefix, "--with-mpi=%s" % spec['mpi'].prefix) make() make("install") Add latest version of PnetCDFfrom spack import * class ParallelNetcdf(Package): """Parallel netCDF (PnetCDF) is a library providing high-performance parallel I/O while still maintaining file-format compatibility with Unidata's NetCDF.""" homepage = "https://trac.mcs.anl.gov/projects/parallel-netcdf" url = "http://cucis.ece.northwestern.edu/projects/PnetCDF/Release/parallel-netcdf-1.6.1.tar.gz" version('1.7.0', '267eab7b6f9dc78c4d0e6def2def3aea4bc7c9f0') version('1.6.1', '62a094eb952f9d1e15f07d56e535052604f1ac34') depends_on("m4") depends_on("mpi") def install(self, spec, prefix): configure("--prefix=%s" % prefix, "--with-mpi=%s" % spec['mpi'].prefix) make() make("install")
<commit_before>from spack import * class ParallelNetcdf(Package): """Parallel netCDF (PnetCDF) is a library providing high-performance parallel I/O while still maintaining file-format compatibility with Unidata's NetCDF.""" homepage = "https://trac.mcs.anl.gov/projects/parallel-netcdf" url = "http://cucis.ece.northwestern.edu/projects/PnetCDF/Release/parallel-netcdf-1.6.1.tar.gz" version('1.6.1', '62a094eb952f9d1e15f07d56e535052604f1ac34') depends_on("m4") depends_on("mpi") def install(self, spec, prefix): configure("--prefix=%s" % prefix, "--with-mpi=%s" % spec['mpi'].prefix) make() make("install") <commit_msg>Add latest version of PnetCDF<commit_after>from spack import * class ParallelNetcdf(Package): """Parallel netCDF (PnetCDF) is a library providing high-performance parallel I/O while still maintaining file-format compatibility with Unidata's NetCDF.""" homepage = "https://trac.mcs.anl.gov/projects/parallel-netcdf" url = "http://cucis.ece.northwestern.edu/projects/PnetCDF/Release/parallel-netcdf-1.6.1.tar.gz" version('1.7.0', '267eab7b6f9dc78c4d0e6def2def3aea4bc7c9f0') version('1.6.1', '62a094eb952f9d1e15f07d56e535052604f1ac34') depends_on("m4") depends_on("mpi") def install(self, spec, prefix): configure("--prefix=%s" % prefix, "--with-mpi=%s" % spec['mpi'].prefix) make() make("install")
8297d4a8650b94bee6aa2c9a83b699f443596ce6
stagecraft/tools/txex-migration.py
stagecraft/tools/txex-migration.py
#!/usr/bin/env python import os import sys try: username = os.environ['GOOGLE_USERNAME'] password = os.environ['GOOGLE_PASSWORD'] except KeyError: print("Please supply username (GOOGLE_USERNAME)" "and password (GOOGLE_PASSWORD) as environment variables") sys.exit(1) column_positions = { 'names_name': 7, 'names_slug': 8, 'names_service_name': 5, 'names_service_slug': 6, 'names_tx_id_column': 17 } from spreadsheets import SpreadsheetMunger munger = SpreadsheetMunger(column_positions) print munger.load(username, password)
#!/usr/bin/env python import os import sys try: username = os.environ['GOOGLE_USERNAME'] password = os.environ['GOOGLE_PASSWORD'] except KeyError: print("Please supply username (GOOGLE_USERNAME)" "and password (GOOGLE_PASSWORD) as environment variables") sys.exit(1) column_positions = { 'names_name': 8, 'names_slug': 9, 'names_service_name': 6, 'names_service_slug': 7, 'names_tx_id_column': 18 } from spreadsheets import SpreadsheetMunger munger = SpreadsheetMunger(column_positions) print munger.load(username, password)
Fix column positions to reflect current spreadsheet.
Fix column positions to reflect current spreadsheet.
Python
mit
alphagov/stagecraft,alphagov/stagecraft,alphagov/stagecraft,alphagov/stagecraft
#!/usr/bin/env python import os import sys try: username = os.environ['GOOGLE_USERNAME'] password = os.environ['GOOGLE_PASSWORD'] except KeyError: print("Please supply username (GOOGLE_USERNAME)" "and password (GOOGLE_PASSWORD) as environment variables") sys.exit(1) column_positions = { 'names_name': 7, 'names_slug': 8, 'names_service_name': 5, 'names_service_slug': 6, 'names_tx_id_column': 17 } from spreadsheets import SpreadsheetMunger munger = SpreadsheetMunger(column_positions) print munger.load(username, password) Fix column positions to reflect current spreadsheet.
#!/usr/bin/env python import os import sys try: username = os.environ['GOOGLE_USERNAME'] password = os.environ['GOOGLE_PASSWORD'] except KeyError: print("Please supply username (GOOGLE_USERNAME)" "and password (GOOGLE_PASSWORD) as environment variables") sys.exit(1) column_positions = { 'names_name': 8, 'names_slug': 9, 'names_service_name': 6, 'names_service_slug': 7, 'names_tx_id_column': 18 } from spreadsheets import SpreadsheetMunger munger = SpreadsheetMunger(column_positions) print munger.load(username, password)
<commit_before>#!/usr/bin/env python import os import sys try: username = os.environ['GOOGLE_USERNAME'] password = os.environ['GOOGLE_PASSWORD'] except KeyError: print("Please supply username (GOOGLE_USERNAME)" "and password (GOOGLE_PASSWORD) as environment variables") sys.exit(1) column_positions = { 'names_name': 7, 'names_slug': 8, 'names_service_name': 5, 'names_service_slug': 6, 'names_tx_id_column': 17 } from spreadsheets import SpreadsheetMunger munger = SpreadsheetMunger(column_positions) print munger.load(username, password) <commit_msg>Fix column positions to reflect current spreadsheet.<commit_after>
#!/usr/bin/env python import os import sys try: username = os.environ['GOOGLE_USERNAME'] password = os.environ['GOOGLE_PASSWORD'] except KeyError: print("Please supply username (GOOGLE_USERNAME)" "and password (GOOGLE_PASSWORD) as environment variables") sys.exit(1) column_positions = { 'names_name': 8, 'names_slug': 9, 'names_service_name': 6, 'names_service_slug': 7, 'names_tx_id_column': 18 } from spreadsheets import SpreadsheetMunger munger = SpreadsheetMunger(column_positions) print munger.load(username, password)
#!/usr/bin/env python import os import sys try: username = os.environ['GOOGLE_USERNAME'] password = os.environ['GOOGLE_PASSWORD'] except KeyError: print("Please supply username (GOOGLE_USERNAME)" "and password (GOOGLE_PASSWORD) as environment variables") sys.exit(1) column_positions = { 'names_name': 7, 'names_slug': 8, 'names_service_name': 5, 'names_service_slug': 6, 'names_tx_id_column': 17 } from spreadsheets import SpreadsheetMunger munger = SpreadsheetMunger(column_positions) print munger.load(username, password) Fix column positions to reflect current spreadsheet.#!/usr/bin/env python import os import sys try: username = os.environ['GOOGLE_USERNAME'] password = os.environ['GOOGLE_PASSWORD'] except KeyError: print("Please supply username (GOOGLE_USERNAME)" "and password (GOOGLE_PASSWORD) as environment variables") sys.exit(1) column_positions = { 'names_name': 8, 'names_slug': 9, 'names_service_name': 6, 'names_service_slug': 7, 'names_tx_id_column': 18 } from spreadsheets import SpreadsheetMunger munger = SpreadsheetMunger(column_positions) print munger.load(username, password)
<commit_before>#!/usr/bin/env python import os import sys try: username = os.environ['GOOGLE_USERNAME'] password = os.environ['GOOGLE_PASSWORD'] except KeyError: print("Please supply username (GOOGLE_USERNAME)" "and password (GOOGLE_PASSWORD) as environment variables") sys.exit(1) column_positions = { 'names_name': 7, 'names_slug': 8, 'names_service_name': 5, 'names_service_slug': 6, 'names_tx_id_column': 17 } from spreadsheets import SpreadsheetMunger munger = SpreadsheetMunger(column_positions) print munger.load(username, password) <commit_msg>Fix column positions to reflect current spreadsheet.<commit_after>#!/usr/bin/env python import os import sys try: username = os.environ['GOOGLE_USERNAME'] password = os.environ['GOOGLE_PASSWORD'] except KeyError: print("Please supply username (GOOGLE_USERNAME)" "and password (GOOGLE_PASSWORD) as environment variables") sys.exit(1) column_positions = { 'names_name': 8, 'names_slug': 9, 'names_service_name': 6, 'names_service_slug': 7, 'names_tx_id_column': 18 } from spreadsheets import SpreadsheetMunger munger = SpreadsheetMunger(column_positions) print munger.load(username, password)
b9bfabe2648dbc7604488b7bda5c53e460155072
glharvest/tests/test_void.py
glharvest/tests/test_void.py
"""test_void.py Test the parsing of VoID dump files. """ try: import RDF except ImportError: import sys sys.path.append('/usr/lib/python2.7/dist-packages/') import RDF from glharvest import util, void def test_returns_none_if_the_registry_file_is_not_found(): m = util.load_file_into_model("nonexistantvoidfile.ttl") assert m is None def test_can_load_a_simple_void_file(): m = util.load_file_into_model('tests/data/simple-void.ttl', 'turtle') p = void.parse_void_model(m) assert p == { 'http://lod.dataone.org/test': { 'dataDump': 'http://lod.dataone.org/test.ttl', 'features': [ 'http://lod.dataone.org/fulldump' ] } }
"""test_void.py Test the parsing of VoID dump files. """ try: import RDF except ImportError: import sys sys.path.append('/usr/lib/python2.7/dist-packages/') import RDF from glharvest import util, void def test_returns_none_if_the_registry_file_is_not_found(): m = util.load_file_into_model("nonexistantvoidfile.ttl") assert m is None def test_can_load_a_simple_void_file(): m = util.load_file_into_model('tests/data/simple-void.ttl', 'turtle') p = void.parse_void_model(m) assert p == { 'http://lod.dataone.org/test': { 'dumps': ['http://lod.dataone.org/test.ttl'], 'features': [ 'http://lod.dataone.org/fulldump' ] } }
Update void test to reflect the new void model structure
Update void test to reflect the new void model structure The first version I coded up only allowed for a single dump in each dataset. The most current version expects an array of dumps so I needed to update the test.
Python
apache-2.0
ec-geolink/glharvest,ec-geolink/glharvest,ec-geolink/glharvest
"""test_void.py Test the parsing of VoID dump files. """ try: import RDF except ImportError: import sys sys.path.append('/usr/lib/python2.7/dist-packages/') import RDF from glharvest import util, void def test_returns_none_if_the_registry_file_is_not_found(): m = util.load_file_into_model("nonexistantvoidfile.ttl") assert m is None def test_can_load_a_simple_void_file(): m = util.load_file_into_model('tests/data/simple-void.ttl', 'turtle') p = void.parse_void_model(m) assert p == { 'http://lod.dataone.org/test': { 'dataDump': 'http://lod.dataone.org/test.ttl', 'features': [ 'http://lod.dataone.org/fulldump' ] } } Update void test to reflect the new void model structure The first version I coded up only allowed for a single dump in each dataset. The most current version expects an array of dumps so I needed to update the test.
"""test_void.py Test the parsing of VoID dump files. """ try: import RDF except ImportError: import sys sys.path.append('/usr/lib/python2.7/dist-packages/') import RDF from glharvest import util, void def test_returns_none_if_the_registry_file_is_not_found(): m = util.load_file_into_model("nonexistantvoidfile.ttl") assert m is None def test_can_load_a_simple_void_file(): m = util.load_file_into_model('tests/data/simple-void.ttl', 'turtle') p = void.parse_void_model(m) assert p == { 'http://lod.dataone.org/test': { 'dumps': ['http://lod.dataone.org/test.ttl'], 'features': [ 'http://lod.dataone.org/fulldump' ] } }
<commit_before>"""test_void.py Test the parsing of VoID dump files. """ try: import RDF except ImportError: import sys sys.path.append('/usr/lib/python2.7/dist-packages/') import RDF from glharvest import util, void def test_returns_none_if_the_registry_file_is_not_found(): m = util.load_file_into_model("nonexistantvoidfile.ttl") assert m is None def test_can_load_a_simple_void_file(): m = util.load_file_into_model('tests/data/simple-void.ttl', 'turtle') p = void.parse_void_model(m) assert p == { 'http://lod.dataone.org/test': { 'dataDump': 'http://lod.dataone.org/test.ttl', 'features': [ 'http://lod.dataone.org/fulldump' ] } } <commit_msg>Update void test to reflect the new void model structure The first version I coded up only allowed for a single dump in each dataset. The most current version expects an array of dumps so I needed to update the test.<commit_after>
"""test_void.py Test the parsing of VoID dump files. """ try: import RDF except ImportError: import sys sys.path.append('/usr/lib/python2.7/dist-packages/') import RDF from glharvest import util, void def test_returns_none_if_the_registry_file_is_not_found(): m = util.load_file_into_model("nonexistantvoidfile.ttl") assert m is None def test_can_load_a_simple_void_file(): m = util.load_file_into_model('tests/data/simple-void.ttl', 'turtle') p = void.parse_void_model(m) assert p == { 'http://lod.dataone.org/test': { 'dumps': ['http://lod.dataone.org/test.ttl'], 'features': [ 'http://lod.dataone.org/fulldump' ] } }
"""test_void.py Test the parsing of VoID dump files. """ try: import RDF except ImportError: import sys sys.path.append('/usr/lib/python2.7/dist-packages/') import RDF from glharvest import util, void def test_returns_none_if_the_registry_file_is_not_found(): m = util.load_file_into_model("nonexistantvoidfile.ttl") assert m is None def test_can_load_a_simple_void_file(): m = util.load_file_into_model('tests/data/simple-void.ttl', 'turtle') p = void.parse_void_model(m) assert p == { 'http://lod.dataone.org/test': { 'dataDump': 'http://lod.dataone.org/test.ttl', 'features': [ 'http://lod.dataone.org/fulldump' ] } } Update void test to reflect the new void model structure The first version I coded up only allowed for a single dump in each dataset. The most current version expects an array of dumps so I needed to update the test."""test_void.py Test the parsing of VoID dump files. """ try: import RDF except ImportError: import sys sys.path.append('/usr/lib/python2.7/dist-packages/') import RDF from glharvest import util, void def test_returns_none_if_the_registry_file_is_not_found(): m = util.load_file_into_model("nonexistantvoidfile.ttl") assert m is None def test_can_load_a_simple_void_file(): m = util.load_file_into_model('tests/data/simple-void.ttl', 'turtle') p = void.parse_void_model(m) assert p == { 'http://lod.dataone.org/test': { 'dumps': ['http://lod.dataone.org/test.ttl'], 'features': [ 'http://lod.dataone.org/fulldump' ] } }
<commit_before>"""test_void.py Test the parsing of VoID dump files. """ try: import RDF except ImportError: import sys sys.path.append('/usr/lib/python2.7/dist-packages/') import RDF from glharvest import util, void def test_returns_none_if_the_registry_file_is_not_found(): m = util.load_file_into_model("nonexistantvoidfile.ttl") assert m is None def test_can_load_a_simple_void_file(): m = util.load_file_into_model('tests/data/simple-void.ttl', 'turtle') p = void.parse_void_model(m) assert p == { 'http://lod.dataone.org/test': { 'dataDump': 'http://lod.dataone.org/test.ttl', 'features': [ 'http://lod.dataone.org/fulldump' ] } } <commit_msg>Update void test to reflect the new void model structure The first version I coded up only allowed for a single dump in each dataset. The most current version expects an array of dumps so I needed to update the test.<commit_after>"""test_void.py Test the parsing of VoID dump files. """ try: import RDF except ImportError: import sys sys.path.append('/usr/lib/python2.7/dist-packages/') import RDF from glharvest import util, void def test_returns_none_if_the_registry_file_is_not_found(): m = util.load_file_into_model("nonexistantvoidfile.ttl") assert m is None def test_can_load_a_simple_void_file(): m = util.load_file_into_model('tests/data/simple-void.ttl', 'turtle') p = void.parse_void_model(m) assert p == { 'http://lod.dataone.org/test': { 'dumps': ['http://lod.dataone.org/test.ttl'], 'features': [ 'http://lod.dataone.org/fulldump' ] } }
cdc5627cfad3b4fb413bed86d76dbe083e6727a7
hnotebook/notebooks/admin.py
hnotebook/notebooks/admin.py
from django.contrib import admin # Register your models here.
from django.contrib import admin from .models import Notebook class NotebookAdmin(admin.ModelAdmin): model = Notebook admin.site.register(Notebook, NotebookAdmin)
Add Admin for Notebook model
Add Admin for Notebook model
Python
mit
marcwebbie/hnotebook
from django.contrib import admin # Register your models here. Add Admin for Notebook model
from django.contrib import admin from .models import Notebook class NotebookAdmin(admin.ModelAdmin): model = Notebook admin.site.register(Notebook, NotebookAdmin)
<commit_before>from django.contrib import admin # Register your models here. <commit_msg>Add Admin for Notebook model<commit_after>
from django.contrib import admin from .models import Notebook class NotebookAdmin(admin.ModelAdmin): model = Notebook admin.site.register(Notebook, NotebookAdmin)
from django.contrib import admin # Register your models here. Add Admin for Notebook modelfrom django.contrib import admin from .models import Notebook class NotebookAdmin(admin.ModelAdmin): model = Notebook admin.site.register(Notebook, NotebookAdmin)
<commit_before>from django.contrib import admin # Register your models here. <commit_msg>Add Admin for Notebook model<commit_after>from django.contrib import admin from .models import Notebook class NotebookAdmin(admin.ModelAdmin): model = Notebook admin.site.register(Notebook, NotebookAdmin)
f09208c047de7c31b9a76d903aa996bf74a3159c
bin/targetselection.py
bin/targetselection.py
import numpy from desitarget.io import read_tractor, write_targets from desitarget.cuts import LRG, ELG, BGS, QSO from desitarget import targetmask from argparse import ArgumentParser ap = ArgumentParser() ap.add_argument("--type", choices=["tractor"], default="tractor", help="Assume a type for src files") ap.add_argument("src", help="File that stores Candidates/Objects") ap.add_argument("dest", help="File that stores targets") TYPES = { 'LRG': LRG, 'ELG': ELG, 'BGS': BGS, 'QSO': QSO, } def main(): ns = ap.parse_args() candidates = read_tractor(ns.src) # FIXME: fits doesn't like u8; there must be a workaround # but lets stick with i8 for now. tsbits = numpy.zeros(len(candidates), dtype='i8') for t in TYPES.keys(): cut = TYPES[t] bitfield = targetmask.mask(t) with numpy.errstate(all='ignore'): mask = cut.apply(candidates) tsbits[mask] |= bitfield assert ((tsbits & bitfield) != 0).sum() == mask.sum() print (t, 'selected', mask.sum()) write_targets(ns.dest, candidates, tsbits) print ('written to', ns.dest) if __name__ == "__main__": main()
from __future__ import print_function import numpy from desitarget.io import read_tractor, write_targets from desitarget.cuts import LRG, ELG, BGS, QSO from desitarget import targetmask from argparse import ArgumentParser ap = ArgumentParser() ap.add_argument("--type", choices=["tractor"], default="tractor", help="Assume a type for src files") ap.add_argument("src", help="File that stores Candidates/Objects") ap.add_argument("dest", help="File that stores targets") TYPES = { 'LRG': LRG, 'ELG': ELG, 'BGS': BGS, 'QSO': QSO, } def main(): ns = ap.parse_args() candidates = read_tractor(ns.src) # FIXME: fits doesn't like u8; there must be a workaround # but lets stick with i8 for now. tsbits = numpy.zeros(len(candidates), dtype='i8') for t in TYPES.keys(): cut = TYPES[t] bitfield = targetmask.mask(t) with numpy.errstate(all='ignore'): mask = cut.apply(candidates) tsbits[mask] |= bitfield assert ((tsbits & bitfield) != 0).sum() == mask.sum() print (t, 'selected', mask.sum()) write_targets(ns.dest, candidates, tsbits) print ('written to', ns.dest) if __name__ == "__main__": main()
Use print_function in toplevel script.
Use print_function in toplevel script. (Avoid logging hassle for now)
Python
bsd-3-clause
desihub/desitarget,desihub/desitarget
import numpy from desitarget.io import read_tractor, write_targets from desitarget.cuts import LRG, ELG, BGS, QSO from desitarget import targetmask from argparse import ArgumentParser ap = ArgumentParser() ap.add_argument("--type", choices=["tractor"], default="tractor", help="Assume a type for src files") ap.add_argument("src", help="File that stores Candidates/Objects") ap.add_argument("dest", help="File that stores targets") TYPES = { 'LRG': LRG, 'ELG': ELG, 'BGS': BGS, 'QSO': QSO, } def main(): ns = ap.parse_args() candidates = read_tractor(ns.src) # FIXME: fits doesn't like u8; there must be a workaround # but lets stick with i8 for now. tsbits = numpy.zeros(len(candidates), dtype='i8') for t in TYPES.keys(): cut = TYPES[t] bitfield = targetmask.mask(t) with numpy.errstate(all='ignore'): mask = cut.apply(candidates) tsbits[mask] |= bitfield assert ((tsbits & bitfield) != 0).sum() == mask.sum() print (t, 'selected', mask.sum()) write_targets(ns.dest, candidates, tsbits) print ('written to', ns.dest) if __name__ == "__main__": main() Use print_function in toplevel script. (Avoid logging hassle for now)
from __future__ import print_function import numpy from desitarget.io import read_tractor, write_targets from desitarget.cuts import LRG, ELG, BGS, QSO from desitarget import targetmask from argparse import ArgumentParser ap = ArgumentParser() ap.add_argument("--type", choices=["tractor"], default="tractor", help="Assume a type for src files") ap.add_argument("src", help="File that stores Candidates/Objects") ap.add_argument("dest", help="File that stores targets") TYPES = { 'LRG': LRG, 'ELG': ELG, 'BGS': BGS, 'QSO': QSO, } def main(): ns = ap.parse_args() candidates = read_tractor(ns.src) # FIXME: fits doesn't like u8; there must be a workaround # but lets stick with i8 for now. tsbits = numpy.zeros(len(candidates), dtype='i8') for t in TYPES.keys(): cut = TYPES[t] bitfield = targetmask.mask(t) with numpy.errstate(all='ignore'): mask = cut.apply(candidates) tsbits[mask] |= bitfield assert ((tsbits & bitfield) != 0).sum() == mask.sum() print (t, 'selected', mask.sum()) write_targets(ns.dest, candidates, tsbits) print ('written to', ns.dest) if __name__ == "__main__": main()
<commit_before>import numpy from desitarget.io import read_tractor, write_targets from desitarget.cuts import LRG, ELG, BGS, QSO from desitarget import targetmask from argparse import ArgumentParser ap = ArgumentParser() ap.add_argument("--type", choices=["tractor"], default="tractor", help="Assume a type for src files") ap.add_argument("src", help="File that stores Candidates/Objects") ap.add_argument("dest", help="File that stores targets") TYPES = { 'LRG': LRG, 'ELG': ELG, 'BGS': BGS, 'QSO': QSO, } def main(): ns = ap.parse_args() candidates = read_tractor(ns.src) # FIXME: fits doesn't like u8; there must be a workaround # but lets stick with i8 for now. tsbits = numpy.zeros(len(candidates), dtype='i8') for t in TYPES.keys(): cut = TYPES[t] bitfield = targetmask.mask(t) with numpy.errstate(all='ignore'): mask = cut.apply(candidates) tsbits[mask] |= bitfield assert ((tsbits & bitfield) != 0).sum() == mask.sum() print (t, 'selected', mask.sum()) write_targets(ns.dest, candidates, tsbits) print ('written to', ns.dest) if __name__ == "__main__": main() <commit_msg>Use print_function in toplevel script. (Avoid logging hassle for now)<commit_after>
from __future__ import print_function import numpy from desitarget.io import read_tractor, write_targets from desitarget.cuts import LRG, ELG, BGS, QSO from desitarget import targetmask from argparse import ArgumentParser ap = ArgumentParser() ap.add_argument("--type", choices=["tractor"], default="tractor", help="Assume a type for src files") ap.add_argument("src", help="File that stores Candidates/Objects") ap.add_argument("dest", help="File that stores targets") TYPES = { 'LRG': LRG, 'ELG': ELG, 'BGS': BGS, 'QSO': QSO, } def main(): ns = ap.parse_args() candidates = read_tractor(ns.src) # FIXME: fits doesn't like u8; there must be a workaround # but lets stick with i8 for now. tsbits = numpy.zeros(len(candidates), dtype='i8') for t in TYPES.keys(): cut = TYPES[t] bitfield = targetmask.mask(t) with numpy.errstate(all='ignore'): mask = cut.apply(candidates) tsbits[mask] |= bitfield assert ((tsbits & bitfield) != 0).sum() == mask.sum() print (t, 'selected', mask.sum()) write_targets(ns.dest, candidates, tsbits) print ('written to', ns.dest) if __name__ == "__main__": main()
import numpy from desitarget.io import read_tractor, write_targets from desitarget.cuts import LRG, ELG, BGS, QSO from desitarget import targetmask from argparse import ArgumentParser ap = ArgumentParser() ap.add_argument("--type", choices=["tractor"], default="tractor", help="Assume a type for src files") ap.add_argument("src", help="File that stores Candidates/Objects") ap.add_argument("dest", help="File that stores targets") TYPES = { 'LRG': LRG, 'ELG': ELG, 'BGS': BGS, 'QSO': QSO, } def main(): ns = ap.parse_args() candidates = read_tractor(ns.src) # FIXME: fits doesn't like u8; there must be a workaround # but lets stick with i8 for now. tsbits = numpy.zeros(len(candidates), dtype='i8') for t in TYPES.keys(): cut = TYPES[t] bitfield = targetmask.mask(t) with numpy.errstate(all='ignore'): mask = cut.apply(candidates) tsbits[mask] |= bitfield assert ((tsbits & bitfield) != 0).sum() == mask.sum() print (t, 'selected', mask.sum()) write_targets(ns.dest, candidates, tsbits) print ('written to', ns.dest) if __name__ == "__main__": main() Use print_function in toplevel script. (Avoid logging hassle for now)from __future__ import print_function import numpy from desitarget.io import read_tractor, write_targets from desitarget.cuts import LRG, ELG, BGS, QSO from desitarget import targetmask from argparse import ArgumentParser ap = ArgumentParser() ap.add_argument("--type", choices=["tractor"], default="tractor", help="Assume a type for src files") ap.add_argument("src", help="File that stores Candidates/Objects") ap.add_argument("dest", help="File that stores targets") TYPES = { 'LRG': LRG, 'ELG': ELG, 'BGS': BGS, 'QSO': QSO, } def main(): ns = ap.parse_args() candidates = read_tractor(ns.src) # FIXME: fits doesn't like u8; there must be a workaround # but lets stick with i8 for now. tsbits = numpy.zeros(len(candidates), dtype='i8') for t in TYPES.keys(): cut = TYPES[t] bitfield = targetmask.mask(t) with numpy.errstate(all='ignore'): mask = cut.apply(candidates) tsbits[mask] |= bitfield assert ((tsbits & bitfield) != 0).sum() == mask.sum() print (t, 'selected', mask.sum()) write_targets(ns.dest, candidates, tsbits) print ('written to', ns.dest) if __name__ == "__main__": main()
<commit_before>import numpy from desitarget.io import read_tractor, write_targets from desitarget.cuts import LRG, ELG, BGS, QSO from desitarget import targetmask from argparse import ArgumentParser ap = ArgumentParser() ap.add_argument("--type", choices=["tractor"], default="tractor", help="Assume a type for src files") ap.add_argument("src", help="File that stores Candidates/Objects") ap.add_argument("dest", help="File that stores targets") TYPES = { 'LRG': LRG, 'ELG': ELG, 'BGS': BGS, 'QSO': QSO, } def main(): ns = ap.parse_args() candidates = read_tractor(ns.src) # FIXME: fits doesn't like u8; there must be a workaround # but lets stick with i8 for now. tsbits = numpy.zeros(len(candidates), dtype='i8') for t in TYPES.keys(): cut = TYPES[t] bitfield = targetmask.mask(t) with numpy.errstate(all='ignore'): mask = cut.apply(candidates) tsbits[mask] |= bitfield assert ((tsbits & bitfield) != 0).sum() == mask.sum() print (t, 'selected', mask.sum()) write_targets(ns.dest, candidates, tsbits) print ('written to', ns.dest) if __name__ == "__main__": main() <commit_msg>Use print_function in toplevel script. (Avoid logging hassle for now)<commit_after>from __future__ import print_function import numpy from desitarget.io import read_tractor, write_targets from desitarget.cuts import LRG, ELG, BGS, QSO from desitarget import targetmask from argparse import ArgumentParser ap = ArgumentParser() ap.add_argument("--type", choices=["tractor"], default="tractor", help="Assume a type for src files") ap.add_argument("src", help="File that stores Candidates/Objects") ap.add_argument("dest", help="File that stores targets") TYPES = { 'LRG': LRG, 'ELG': ELG, 'BGS': BGS, 'QSO': QSO, } def main(): ns = ap.parse_args() candidates = read_tractor(ns.src) # FIXME: fits doesn't like u8; there must be a workaround # but lets stick with i8 for now. tsbits = numpy.zeros(len(candidates), dtype='i8') for t in TYPES.keys(): cut = TYPES[t] bitfield = targetmask.mask(t) with numpy.errstate(all='ignore'): mask = cut.apply(candidates) tsbits[mask] |= bitfield assert ((tsbits & bitfield) != 0).sum() == mask.sum() print (t, 'selected', mask.sum()) write_targets(ns.dest, candidates, tsbits) print ('written to', ns.dest) if __name__ == "__main__": main()
75a59409410a8f264e7d56ddd853002ffbb28600
corehq/tests/noseplugins/patches.py
corehq/tests/noseplugins/patches.py
from nose.plugins import Plugin from corehq.form_processor.tests.utils import patch_testcase_databases from corehq.util.es.testing import patch_es_user_signals from corehq.util.test_utils import patch_foreign_value_caches class PatchesPlugin(Plugin): """Patches various things before tests are run""" name = "patches" enabled = True def options(self, parser, env): """Do not call super (always enabled)""" def begin(self): patch_assertItemsEqual() patch_testcase_databases() fix_freezegun_bugs() patch_es_user_signals() patch_foreign_value_caches() def patch_assertItemsEqual(): import unittest unittest.TestCase.assertItemsEqual = unittest.TestCase.assertCountEqual GLOBAL_FREEZEGUN_IGNORE_LIST = ["kafka."] def fix_freezegun_bugs(): """Fix error in freezegun.api.freeze_time This error occurs in a background thread that is either triggered by a test using freezegun or becomes active while freezegun patches are in place. More complete error details: ``` Exception in thread cchq-producer-network-thread: Traceback (most recent call last): ... freezegun/api.py", line 151, in _should_use_real_time if not ignore_lists[-1]: IndexError: list index out of range ``` """ import freezegun.api as api def freeze_time(*args, **kw): kw["ignore"] = kw.get("ignore", []) + GLOBAL_FREEZEGUN_IGNORE_LIST return real_freeze_time(*args, **kw) # add base ignore list to avoid index error assert not api.ignore_lists, f"expected empty list, got {api.ignore_lists}" api.ignore_lists.append(tuple(GLOBAL_FREEZEGUN_IGNORE_LIST)) # patch freeze_time so it always ignores kafka real_freeze_time = api.freeze_time api.freeze_time = freeze_time
from nose.plugins import Plugin from corehq.form_processor.tests.utils import patch_testcase_databases from corehq.util.es.testing import patch_es_user_signals from corehq.util.test_utils import patch_foreign_value_caches class PatchesPlugin(Plugin): """Patches various things before tests are run""" name = "patches" enabled = True def options(self, parser, env): """Do not call super (always enabled)""" def begin(self): patch_assertItemsEqual() patch_testcase_databases() extend_freezegun_ignore_list() patch_es_user_signals() patch_foreign_value_caches() def patch_assertItemsEqual(): import unittest unittest.TestCase.assertItemsEqual = unittest.TestCase.assertCountEqual GLOBAL_FREEZEGUN_IGNORE_LIST = ["kafka."] def extend_freezegun_ignore_list(): """Extend the freezegun ignore list""" import freezegun freezegun.configure(extend_ignore_list=GLOBAL_FREEZEGUN_IGNORE_LIST)
Update freezegun ignore list patch
Update freezegun ignore list patch As of v1.1.0, freezegun supports configuring the ignore list.
Python
bsd-3-clause
dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq
from nose.plugins import Plugin from corehq.form_processor.tests.utils import patch_testcase_databases from corehq.util.es.testing import patch_es_user_signals from corehq.util.test_utils import patch_foreign_value_caches class PatchesPlugin(Plugin): """Patches various things before tests are run""" name = "patches" enabled = True def options(self, parser, env): """Do not call super (always enabled)""" def begin(self): patch_assertItemsEqual() patch_testcase_databases() fix_freezegun_bugs() patch_es_user_signals() patch_foreign_value_caches() def patch_assertItemsEqual(): import unittest unittest.TestCase.assertItemsEqual = unittest.TestCase.assertCountEqual GLOBAL_FREEZEGUN_IGNORE_LIST = ["kafka."] def fix_freezegun_bugs(): """Fix error in freezegun.api.freeze_time This error occurs in a background thread that is either triggered by a test using freezegun or becomes active while freezegun patches are in place. More complete error details: ``` Exception in thread cchq-producer-network-thread: Traceback (most recent call last): ... freezegun/api.py", line 151, in _should_use_real_time if not ignore_lists[-1]: IndexError: list index out of range ``` """ import freezegun.api as api def freeze_time(*args, **kw): kw["ignore"] = kw.get("ignore", []) + GLOBAL_FREEZEGUN_IGNORE_LIST return real_freeze_time(*args, **kw) # add base ignore list to avoid index error assert not api.ignore_lists, f"expected empty list, got {api.ignore_lists}" api.ignore_lists.append(tuple(GLOBAL_FREEZEGUN_IGNORE_LIST)) # patch freeze_time so it always ignores kafka real_freeze_time = api.freeze_time api.freeze_time = freeze_time Update freezegun ignore list patch As of v1.1.0, freezegun supports configuring the ignore list.
from nose.plugins import Plugin from corehq.form_processor.tests.utils import patch_testcase_databases from corehq.util.es.testing import patch_es_user_signals from corehq.util.test_utils import patch_foreign_value_caches class PatchesPlugin(Plugin): """Patches various things before tests are run""" name = "patches" enabled = True def options(self, parser, env): """Do not call super (always enabled)""" def begin(self): patch_assertItemsEqual() patch_testcase_databases() extend_freezegun_ignore_list() patch_es_user_signals() patch_foreign_value_caches() def patch_assertItemsEqual(): import unittest unittest.TestCase.assertItemsEqual = unittest.TestCase.assertCountEqual GLOBAL_FREEZEGUN_IGNORE_LIST = ["kafka."] def extend_freezegun_ignore_list(): """Extend the freezegun ignore list""" import freezegun freezegun.configure(extend_ignore_list=GLOBAL_FREEZEGUN_IGNORE_LIST)
<commit_before>from nose.plugins import Plugin from corehq.form_processor.tests.utils import patch_testcase_databases from corehq.util.es.testing import patch_es_user_signals from corehq.util.test_utils import patch_foreign_value_caches class PatchesPlugin(Plugin): """Patches various things before tests are run""" name = "patches" enabled = True def options(self, parser, env): """Do not call super (always enabled)""" def begin(self): patch_assertItemsEqual() patch_testcase_databases() fix_freezegun_bugs() patch_es_user_signals() patch_foreign_value_caches() def patch_assertItemsEqual(): import unittest unittest.TestCase.assertItemsEqual = unittest.TestCase.assertCountEqual GLOBAL_FREEZEGUN_IGNORE_LIST = ["kafka."] def fix_freezegun_bugs(): """Fix error in freezegun.api.freeze_time This error occurs in a background thread that is either triggered by a test using freezegun or becomes active while freezegun patches are in place. More complete error details: ``` Exception in thread cchq-producer-network-thread: Traceback (most recent call last): ... freezegun/api.py", line 151, in _should_use_real_time if not ignore_lists[-1]: IndexError: list index out of range ``` """ import freezegun.api as api def freeze_time(*args, **kw): kw["ignore"] = kw.get("ignore", []) + GLOBAL_FREEZEGUN_IGNORE_LIST return real_freeze_time(*args, **kw) # add base ignore list to avoid index error assert not api.ignore_lists, f"expected empty list, got {api.ignore_lists}" api.ignore_lists.append(tuple(GLOBAL_FREEZEGUN_IGNORE_LIST)) # patch freeze_time so it always ignores kafka real_freeze_time = api.freeze_time api.freeze_time = freeze_time <commit_msg>Update freezegun ignore list patch As of v1.1.0, freezegun supports configuring the ignore list.<commit_after>
from nose.plugins import Plugin from corehq.form_processor.tests.utils import patch_testcase_databases from corehq.util.es.testing import patch_es_user_signals from corehq.util.test_utils import patch_foreign_value_caches class PatchesPlugin(Plugin): """Patches various things before tests are run""" name = "patches" enabled = True def options(self, parser, env): """Do not call super (always enabled)""" def begin(self): patch_assertItemsEqual() patch_testcase_databases() extend_freezegun_ignore_list() patch_es_user_signals() patch_foreign_value_caches() def patch_assertItemsEqual(): import unittest unittest.TestCase.assertItemsEqual = unittest.TestCase.assertCountEqual GLOBAL_FREEZEGUN_IGNORE_LIST = ["kafka."] def extend_freezegun_ignore_list(): """Extend the freezegun ignore list""" import freezegun freezegun.configure(extend_ignore_list=GLOBAL_FREEZEGUN_IGNORE_LIST)
from nose.plugins import Plugin from corehq.form_processor.tests.utils import patch_testcase_databases from corehq.util.es.testing import patch_es_user_signals from corehq.util.test_utils import patch_foreign_value_caches class PatchesPlugin(Plugin): """Patches various things before tests are run""" name = "patches" enabled = True def options(self, parser, env): """Do not call super (always enabled)""" def begin(self): patch_assertItemsEqual() patch_testcase_databases() fix_freezegun_bugs() patch_es_user_signals() patch_foreign_value_caches() def patch_assertItemsEqual(): import unittest unittest.TestCase.assertItemsEqual = unittest.TestCase.assertCountEqual GLOBAL_FREEZEGUN_IGNORE_LIST = ["kafka."] def fix_freezegun_bugs(): """Fix error in freezegun.api.freeze_time This error occurs in a background thread that is either triggered by a test using freezegun or becomes active while freezegun patches are in place. More complete error details: ``` Exception in thread cchq-producer-network-thread: Traceback (most recent call last): ... freezegun/api.py", line 151, in _should_use_real_time if not ignore_lists[-1]: IndexError: list index out of range ``` """ import freezegun.api as api def freeze_time(*args, **kw): kw["ignore"] = kw.get("ignore", []) + GLOBAL_FREEZEGUN_IGNORE_LIST return real_freeze_time(*args, **kw) # add base ignore list to avoid index error assert not api.ignore_lists, f"expected empty list, got {api.ignore_lists}" api.ignore_lists.append(tuple(GLOBAL_FREEZEGUN_IGNORE_LIST)) # patch freeze_time so it always ignores kafka real_freeze_time = api.freeze_time api.freeze_time = freeze_time Update freezegun ignore list patch As of v1.1.0, freezegun supports configuring the ignore list.from nose.plugins import Plugin from corehq.form_processor.tests.utils import patch_testcase_databases from corehq.util.es.testing import patch_es_user_signals from corehq.util.test_utils import patch_foreign_value_caches class PatchesPlugin(Plugin): """Patches various things before tests are run""" name = "patches" enabled = True def options(self, parser, env): """Do not call super (always enabled)""" def begin(self): patch_assertItemsEqual() patch_testcase_databases() extend_freezegun_ignore_list() patch_es_user_signals() patch_foreign_value_caches() def patch_assertItemsEqual(): import unittest unittest.TestCase.assertItemsEqual = unittest.TestCase.assertCountEqual GLOBAL_FREEZEGUN_IGNORE_LIST = ["kafka."] def extend_freezegun_ignore_list(): """Extend the freezegun ignore list""" import freezegun freezegun.configure(extend_ignore_list=GLOBAL_FREEZEGUN_IGNORE_LIST)
<commit_before>from nose.plugins import Plugin from corehq.form_processor.tests.utils import patch_testcase_databases from corehq.util.es.testing import patch_es_user_signals from corehq.util.test_utils import patch_foreign_value_caches class PatchesPlugin(Plugin): """Patches various things before tests are run""" name = "patches" enabled = True def options(self, parser, env): """Do not call super (always enabled)""" def begin(self): patch_assertItemsEqual() patch_testcase_databases() fix_freezegun_bugs() patch_es_user_signals() patch_foreign_value_caches() def patch_assertItemsEqual(): import unittest unittest.TestCase.assertItemsEqual = unittest.TestCase.assertCountEqual GLOBAL_FREEZEGUN_IGNORE_LIST = ["kafka."] def fix_freezegun_bugs(): """Fix error in freezegun.api.freeze_time This error occurs in a background thread that is either triggered by a test using freezegun or becomes active while freezegun patches are in place. More complete error details: ``` Exception in thread cchq-producer-network-thread: Traceback (most recent call last): ... freezegun/api.py", line 151, in _should_use_real_time if not ignore_lists[-1]: IndexError: list index out of range ``` """ import freezegun.api as api def freeze_time(*args, **kw): kw["ignore"] = kw.get("ignore", []) + GLOBAL_FREEZEGUN_IGNORE_LIST return real_freeze_time(*args, **kw) # add base ignore list to avoid index error assert not api.ignore_lists, f"expected empty list, got {api.ignore_lists}" api.ignore_lists.append(tuple(GLOBAL_FREEZEGUN_IGNORE_LIST)) # patch freeze_time so it always ignores kafka real_freeze_time = api.freeze_time api.freeze_time = freeze_time <commit_msg>Update freezegun ignore list patch As of v1.1.0, freezegun supports configuring the ignore list.<commit_after>from nose.plugins import Plugin from corehq.form_processor.tests.utils import patch_testcase_databases from corehq.util.es.testing import patch_es_user_signals from corehq.util.test_utils import patch_foreign_value_caches class PatchesPlugin(Plugin): """Patches various things before tests are run""" name = "patches" enabled = True def options(self, parser, env): """Do not call super (always enabled)""" def begin(self): patch_assertItemsEqual() patch_testcase_databases() extend_freezegun_ignore_list() patch_es_user_signals() patch_foreign_value_caches() def patch_assertItemsEqual(): import unittest unittest.TestCase.assertItemsEqual = unittest.TestCase.assertCountEqual GLOBAL_FREEZEGUN_IGNORE_LIST = ["kafka."] def extend_freezegun_ignore_list(): """Extend the freezegun ignore list""" import freezegun freezegun.configure(extend_ignore_list=GLOBAL_FREEZEGUN_IGNORE_LIST)
9f3cf22575d9d71136bea0282ac0e0420211d9c9
test/test_util/test_StopWatch.py
test/test_util/test_StopWatch.py
# -*- encoding: utf-8 -*- """Created on Dec 16, 2014. @author: Katharina Eggensperger @projekt: AutoML2015 """ from __future__ import print_function import time import unittest from autosklearn.util import StopWatch class Test(unittest.TestCase): _multiprocess_can_split_ = True def test_stopwatch_overhead(self): # CPU overhead start = time.clock() watch = StopWatch() for i in range(1, 100000): watch.start_task('task_%d' % i) watch.stop_task('task_%d' % i) stop = time.clock() dur = stop - start cpu_overhead = dur - watch.cpu_sum() self.assertLess(cpu_overhead, 1.5) # Wall Overhead start = time.time() watch = StopWatch() for i in range(1, 100000): watch.start_task('task_%d' % i) watch.stop_task('task_%d' % i) stop = time.time() dur = stop - start wall_overhead = dur - watch.wall_sum() self.assertLess(wall_overhead, 2) self.assertLess(cpu_overhead, 2*wall_overhead) if __name__ == '__main__': # import sys;sys.argv = ['', 'Test.testName'] unittest.main()
# -*- encoding: utf-8 -*- """Created on Dec 16, 2014. @author: Katharina Eggensperger @projekt: AutoML2015 """ from __future__ import print_function import time import unittest from autosklearn.util import StopWatch class Test(unittest.TestCase): _multiprocess_can_split_ = True def test_stopwatch_overhead(self): # CPU overhead start = time.clock() watch = StopWatch() for i in range(1, 1000): watch.start_task('task_%d' % i) watch.stop_task('task_%d' % i) stop = time.clock() dur = stop - start cpu_overhead = dur - watch.cpu_sum() self.assertLess(cpu_overhead, 1.5) # Wall Overhead start = time.time() watch = StopWatch() for i in range(1, 1000): watch.start_task('task_%d' % i) watch.stop_task('task_%d' % i) stop = time.time() dur = stop - start wall_overhead = dur - watch.wall_sum() self.assertLess(wall_overhead, 2) self.assertLess(cpu_overhead, 2*wall_overhead) if __name__ == '__main__': # import sys;sys.argv = ['', 'Test.testName'] unittest.main()
TEST reduce number of iterations for stopwatch test
TEST reduce number of iterations for stopwatch test
Python
bsd-3-clause
automl/auto-sklearn,automl/auto-sklearn
# -*- encoding: utf-8 -*- """Created on Dec 16, 2014. @author: Katharina Eggensperger @projekt: AutoML2015 """ from __future__ import print_function import time import unittest from autosklearn.util import StopWatch class Test(unittest.TestCase): _multiprocess_can_split_ = True def test_stopwatch_overhead(self): # CPU overhead start = time.clock() watch = StopWatch() for i in range(1, 100000): watch.start_task('task_%d' % i) watch.stop_task('task_%d' % i) stop = time.clock() dur = stop - start cpu_overhead = dur - watch.cpu_sum() self.assertLess(cpu_overhead, 1.5) # Wall Overhead start = time.time() watch = StopWatch() for i in range(1, 100000): watch.start_task('task_%d' % i) watch.stop_task('task_%d' % i) stop = time.time() dur = stop - start wall_overhead = dur - watch.wall_sum() self.assertLess(wall_overhead, 2) self.assertLess(cpu_overhead, 2*wall_overhead) if __name__ == '__main__': # import sys;sys.argv = ['', 'Test.testName'] unittest.main() TEST reduce number of iterations for stopwatch test
# -*- encoding: utf-8 -*- """Created on Dec 16, 2014. @author: Katharina Eggensperger @projekt: AutoML2015 """ from __future__ import print_function import time import unittest from autosklearn.util import StopWatch class Test(unittest.TestCase): _multiprocess_can_split_ = True def test_stopwatch_overhead(self): # CPU overhead start = time.clock() watch = StopWatch() for i in range(1, 1000): watch.start_task('task_%d' % i) watch.stop_task('task_%d' % i) stop = time.clock() dur = stop - start cpu_overhead = dur - watch.cpu_sum() self.assertLess(cpu_overhead, 1.5) # Wall Overhead start = time.time() watch = StopWatch() for i in range(1, 1000): watch.start_task('task_%d' % i) watch.stop_task('task_%d' % i) stop = time.time() dur = stop - start wall_overhead = dur - watch.wall_sum() self.assertLess(wall_overhead, 2) self.assertLess(cpu_overhead, 2*wall_overhead) if __name__ == '__main__': # import sys;sys.argv = ['', 'Test.testName'] unittest.main()
<commit_before># -*- encoding: utf-8 -*- """Created on Dec 16, 2014. @author: Katharina Eggensperger @projekt: AutoML2015 """ from __future__ import print_function import time import unittest from autosklearn.util import StopWatch class Test(unittest.TestCase): _multiprocess_can_split_ = True def test_stopwatch_overhead(self): # CPU overhead start = time.clock() watch = StopWatch() for i in range(1, 100000): watch.start_task('task_%d' % i) watch.stop_task('task_%d' % i) stop = time.clock() dur = stop - start cpu_overhead = dur - watch.cpu_sum() self.assertLess(cpu_overhead, 1.5) # Wall Overhead start = time.time() watch = StopWatch() for i in range(1, 100000): watch.start_task('task_%d' % i) watch.stop_task('task_%d' % i) stop = time.time() dur = stop - start wall_overhead = dur - watch.wall_sum() self.assertLess(wall_overhead, 2) self.assertLess(cpu_overhead, 2*wall_overhead) if __name__ == '__main__': # import sys;sys.argv = ['', 'Test.testName'] unittest.main() <commit_msg>TEST reduce number of iterations for stopwatch test<commit_after>
# -*- encoding: utf-8 -*- """Created on Dec 16, 2014. @author: Katharina Eggensperger @projekt: AutoML2015 """ from __future__ import print_function import time import unittest from autosklearn.util import StopWatch class Test(unittest.TestCase): _multiprocess_can_split_ = True def test_stopwatch_overhead(self): # CPU overhead start = time.clock() watch = StopWatch() for i in range(1, 1000): watch.start_task('task_%d' % i) watch.stop_task('task_%d' % i) stop = time.clock() dur = stop - start cpu_overhead = dur - watch.cpu_sum() self.assertLess(cpu_overhead, 1.5) # Wall Overhead start = time.time() watch = StopWatch() for i in range(1, 1000): watch.start_task('task_%d' % i) watch.stop_task('task_%d' % i) stop = time.time() dur = stop - start wall_overhead = dur - watch.wall_sum() self.assertLess(wall_overhead, 2) self.assertLess(cpu_overhead, 2*wall_overhead) if __name__ == '__main__': # import sys;sys.argv = ['', 'Test.testName'] unittest.main()
# -*- encoding: utf-8 -*- """Created on Dec 16, 2014. @author: Katharina Eggensperger @projekt: AutoML2015 """ from __future__ import print_function import time import unittest from autosklearn.util import StopWatch class Test(unittest.TestCase): _multiprocess_can_split_ = True def test_stopwatch_overhead(self): # CPU overhead start = time.clock() watch = StopWatch() for i in range(1, 100000): watch.start_task('task_%d' % i) watch.stop_task('task_%d' % i) stop = time.clock() dur = stop - start cpu_overhead = dur - watch.cpu_sum() self.assertLess(cpu_overhead, 1.5) # Wall Overhead start = time.time() watch = StopWatch() for i in range(1, 100000): watch.start_task('task_%d' % i) watch.stop_task('task_%d' % i) stop = time.time() dur = stop - start wall_overhead = dur - watch.wall_sum() self.assertLess(wall_overhead, 2) self.assertLess(cpu_overhead, 2*wall_overhead) if __name__ == '__main__': # import sys;sys.argv = ['', 'Test.testName'] unittest.main() TEST reduce number of iterations for stopwatch test# -*- encoding: utf-8 -*- """Created on Dec 16, 2014. @author: Katharina Eggensperger @projekt: AutoML2015 """ from __future__ import print_function import time import unittest from autosklearn.util import StopWatch class Test(unittest.TestCase): _multiprocess_can_split_ = True def test_stopwatch_overhead(self): # CPU overhead start = time.clock() watch = StopWatch() for i in range(1, 1000): watch.start_task('task_%d' % i) watch.stop_task('task_%d' % i) stop = time.clock() dur = stop - start cpu_overhead = dur - watch.cpu_sum() self.assertLess(cpu_overhead, 1.5) # Wall Overhead start = time.time() watch = StopWatch() for i in range(1, 1000): watch.start_task('task_%d' % i) watch.stop_task('task_%d' % i) stop = time.time() dur = stop - start wall_overhead = dur - watch.wall_sum() self.assertLess(wall_overhead, 2) self.assertLess(cpu_overhead, 2*wall_overhead) if __name__ == '__main__': # import sys;sys.argv = ['', 'Test.testName'] unittest.main()
<commit_before># -*- encoding: utf-8 -*- """Created on Dec 16, 2014. @author: Katharina Eggensperger @projekt: AutoML2015 """ from __future__ import print_function import time import unittest from autosklearn.util import StopWatch class Test(unittest.TestCase): _multiprocess_can_split_ = True def test_stopwatch_overhead(self): # CPU overhead start = time.clock() watch = StopWatch() for i in range(1, 100000): watch.start_task('task_%d' % i) watch.stop_task('task_%d' % i) stop = time.clock() dur = stop - start cpu_overhead = dur - watch.cpu_sum() self.assertLess(cpu_overhead, 1.5) # Wall Overhead start = time.time() watch = StopWatch() for i in range(1, 100000): watch.start_task('task_%d' % i) watch.stop_task('task_%d' % i) stop = time.time() dur = stop - start wall_overhead = dur - watch.wall_sum() self.assertLess(wall_overhead, 2) self.assertLess(cpu_overhead, 2*wall_overhead) if __name__ == '__main__': # import sys;sys.argv = ['', 'Test.testName'] unittest.main() <commit_msg>TEST reduce number of iterations for stopwatch test<commit_after># -*- encoding: utf-8 -*- """Created on Dec 16, 2014. @author: Katharina Eggensperger @projekt: AutoML2015 """ from __future__ import print_function import time import unittest from autosklearn.util import StopWatch class Test(unittest.TestCase): _multiprocess_can_split_ = True def test_stopwatch_overhead(self): # CPU overhead start = time.clock() watch = StopWatch() for i in range(1, 1000): watch.start_task('task_%d' % i) watch.stop_task('task_%d' % i) stop = time.clock() dur = stop - start cpu_overhead = dur - watch.cpu_sum() self.assertLess(cpu_overhead, 1.5) # Wall Overhead start = time.time() watch = StopWatch() for i in range(1, 1000): watch.start_task('task_%d' % i) watch.stop_task('task_%d' % i) stop = time.time() dur = stop - start wall_overhead = dur - watch.wall_sum() self.assertLess(wall_overhead, 2) self.assertLess(cpu_overhead, 2*wall_overhead) if __name__ == '__main__': # import sys;sys.argv = ['', 'Test.testName'] unittest.main()
29f8aacde96007976b0aa0cde6d6d37b37e517a9
app/status/views.py
app/status/views.py
from flask import jsonify, current_app from . import status from . import utils from ..main.helpers.service import ServiceLoader from ..main import main @status.route('/_status') def status(): # ServiceLoader is the only thing that actually connects to the API service_loader = ServiceLoader( main.config['API_URL'], main.config['API_AUTH_TOKEN'] ) api_response = utils.return_response_from_api_status_call( service_loader.status ) apis_wot_got_errors = [] if api_response is None or api_response.status_code is not 200: apis_wot_got_errors.append("(Data) API") # if no errors found, return everything if not apis_wot_got_errors: return jsonify( status="ok", version=utils.get_version_label(), api_status=api_response.json(), ) message = "Error connecting to the " \ + (" and the ".join(apis_wot_got_errors)) \ + "." current_app.logger.error(message) return jsonify( status="error", version=utils.get_version_label(), api_status=utils.return_json_or_none(api_response), message=message, ), 500
from flask import jsonify, current_app from . import status from . import utils from ..main.helpers.service import ServiceLoader from ..main import main @status.route('/_status') def status(): # ServiceLoader is the only thing that actually connects to the API service_loader = ServiceLoader( main.config['API_URL'], main.config['API_AUTH_TOKEN'] ) api_response = utils.return_response_from_api_status_call( service_loader.status ) apis_with_errors = [] if api_response is None or api_response.status_code != 200: apis_with_errors.append("(Data) API") # if no errors found, return everything if not apis_with_errors: return jsonify( status="ok", version=utils.get_version_label(), api_status=api_response.json(), ) message = "Error connecting to the " \ + (" and the ".join(apis_with_errors)) \ + "." current_app.logger.error(message) return jsonify( status="error", version=utils.get_version_label(), api_status=utils.return_json_or_none(api_response), message=message, ), 500
Change variable name & int comparison.
Change variable name & int comparison.
Python
mit
mtekel/digitalmarketplace-admin-frontend,mtekel/digitalmarketplace-admin-frontend,alphagov/digitalmarketplace-admin-frontend,alphagov/digitalmarketplace-admin-frontend,alphagov/digitalmarketplace-admin-frontend,mtekel/digitalmarketplace-admin-frontend,mtekel/digitalmarketplace-admin-frontend,alphagov/digitalmarketplace-admin-frontend
from flask import jsonify, current_app from . import status from . import utils from ..main.helpers.service import ServiceLoader from ..main import main @status.route('/_status') def status(): # ServiceLoader is the only thing that actually connects to the API service_loader = ServiceLoader( main.config['API_URL'], main.config['API_AUTH_TOKEN'] ) api_response = utils.return_response_from_api_status_call( service_loader.status ) apis_wot_got_errors = [] if api_response is None or api_response.status_code is not 200: apis_wot_got_errors.append("(Data) API") # if no errors found, return everything if not apis_wot_got_errors: return jsonify( status="ok", version=utils.get_version_label(), api_status=api_response.json(), ) message = "Error connecting to the " \ + (" and the ".join(apis_wot_got_errors)) \ + "." current_app.logger.error(message) return jsonify( status="error", version=utils.get_version_label(), api_status=utils.return_json_or_none(api_response), message=message, ), 500 Change variable name & int comparison.
from flask import jsonify, current_app from . import status from . import utils from ..main.helpers.service import ServiceLoader from ..main import main @status.route('/_status') def status(): # ServiceLoader is the only thing that actually connects to the API service_loader = ServiceLoader( main.config['API_URL'], main.config['API_AUTH_TOKEN'] ) api_response = utils.return_response_from_api_status_call( service_loader.status ) apis_with_errors = [] if api_response is None or api_response.status_code != 200: apis_with_errors.append("(Data) API") # if no errors found, return everything if not apis_with_errors: return jsonify( status="ok", version=utils.get_version_label(), api_status=api_response.json(), ) message = "Error connecting to the " \ + (" and the ".join(apis_with_errors)) \ + "." current_app.logger.error(message) return jsonify( status="error", version=utils.get_version_label(), api_status=utils.return_json_or_none(api_response), message=message, ), 500
<commit_before>from flask import jsonify, current_app from . import status from . import utils from ..main.helpers.service import ServiceLoader from ..main import main @status.route('/_status') def status(): # ServiceLoader is the only thing that actually connects to the API service_loader = ServiceLoader( main.config['API_URL'], main.config['API_AUTH_TOKEN'] ) api_response = utils.return_response_from_api_status_call( service_loader.status ) apis_wot_got_errors = [] if api_response is None or api_response.status_code is not 200: apis_wot_got_errors.append("(Data) API") # if no errors found, return everything if not apis_wot_got_errors: return jsonify( status="ok", version=utils.get_version_label(), api_status=api_response.json(), ) message = "Error connecting to the " \ + (" and the ".join(apis_wot_got_errors)) \ + "." current_app.logger.error(message) return jsonify( status="error", version=utils.get_version_label(), api_status=utils.return_json_or_none(api_response), message=message, ), 500 <commit_msg>Change variable name & int comparison.<commit_after>
from flask import jsonify, current_app from . import status from . import utils from ..main.helpers.service import ServiceLoader from ..main import main @status.route('/_status') def status(): # ServiceLoader is the only thing that actually connects to the API service_loader = ServiceLoader( main.config['API_URL'], main.config['API_AUTH_TOKEN'] ) api_response = utils.return_response_from_api_status_call( service_loader.status ) apis_with_errors = [] if api_response is None or api_response.status_code != 200: apis_with_errors.append("(Data) API") # if no errors found, return everything if not apis_with_errors: return jsonify( status="ok", version=utils.get_version_label(), api_status=api_response.json(), ) message = "Error connecting to the " \ + (" and the ".join(apis_with_errors)) \ + "." current_app.logger.error(message) return jsonify( status="error", version=utils.get_version_label(), api_status=utils.return_json_or_none(api_response), message=message, ), 500
from flask import jsonify, current_app from . import status from . import utils from ..main.helpers.service import ServiceLoader from ..main import main @status.route('/_status') def status(): # ServiceLoader is the only thing that actually connects to the API service_loader = ServiceLoader( main.config['API_URL'], main.config['API_AUTH_TOKEN'] ) api_response = utils.return_response_from_api_status_call( service_loader.status ) apis_wot_got_errors = [] if api_response is None or api_response.status_code is not 200: apis_wot_got_errors.append("(Data) API") # if no errors found, return everything if not apis_wot_got_errors: return jsonify( status="ok", version=utils.get_version_label(), api_status=api_response.json(), ) message = "Error connecting to the " \ + (" and the ".join(apis_wot_got_errors)) \ + "." current_app.logger.error(message) return jsonify( status="error", version=utils.get_version_label(), api_status=utils.return_json_or_none(api_response), message=message, ), 500 Change variable name & int comparison.from flask import jsonify, current_app from . import status from . import utils from ..main.helpers.service import ServiceLoader from ..main import main @status.route('/_status') def status(): # ServiceLoader is the only thing that actually connects to the API service_loader = ServiceLoader( main.config['API_URL'], main.config['API_AUTH_TOKEN'] ) api_response = utils.return_response_from_api_status_call( service_loader.status ) apis_with_errors = [] if api_response is None or api_response.status_code != 200: apis_with_errors.append("(Data) API") # if no errors found, return everything if not apis_with_errors: return jsonify( status="ok", version=utils.get_version_label(), api_status=api_response.json(), ) message = "Error connecting to the " \ + (" and the ".join(apis_with_errors)) \ + "." current_app.logger.error(message) return jsonify( status="error", version=utils.get_version_label(), api_status=utils.return_json_or_none(api_response), message=message, ), 500
<commit_before>from flask import jsonify, current_app from . import status from . import utils from ..main.helpers.service import ServiceLoader from ..main import main @status.route('/_status') def status(): # ServiceLoader is the only thing that actually connects to the API service_loader = ServiceLoader( main.config['API_URL'], main.config['API_AUTH_TOKEN'] ) api_response = utils.return_response_from_api_status_call( service_loader.status ) apis_wot_got_errors = [] if api_response is None or api_response.status_code is not 200: apis_wot_got_errors.append("(Data) API") # if no errors found, return everything if not apis_wot_got_errors: return jsonify( status="ok", version=utils.get_version_label(), api_status=api_response.json(), ) message = "Error connecting to the " \ + (" and the ".join(apis_wot_got_errors)) \ + "." current_app.logger.error(message) return jsonify( status="error", version=utils.get_version_label(), api_status=utils.return_json_or_none(api_response), message=message, ), 500 <commit_msg>Change variable name & int comparison.<commit_after>from flask import jsonify, current_app from . import status from . import utils from ..main.helpers.service import ServiceLoader from ..main import main @status.route('/_status') def status(): # ServiceLoader is the only thing that actually connects to the API service_loader = ServiceLoader( main.config['API_URL'], main.config['API_AUTH_TOKEN'] ) api_response = utils.return_response_from_api_status_call( service_loader.status ) apis_with_errors = [] if api_response is None or api_response.status_code != 200: apis_with_errors.append("(Data) API") # if no errors found, return everything if not apis_with_errors: return jsonify( status="ok", version=utils.get_version_label(), api_status=api_response.json(), ) message = "Error connecting to the " \ + (" and the ".join(apis_with_errors)) \ + "." current_app.logger.error(message) return jsonify( status="error", version=utils.get_version_label(), api_status=utils.return_json_or_none(api_response), message=message, ), 500
637f6c09bf3ac558e6e30d748dfb0838e4a3720f
classes/person.py
classes/person.py
class Person(object): def __init__(self, iden, person_type, person_name, person_surname="", wants_accommodation="N"): self.person_name = person_name self.person_surname = person_surname self.person_type = person_type self.wants_accommodation = wants_accommodation self.iden = iden def full_name(self): self.full_name = self.person_name + " " + self.person_surname return self.full_name
class Person(object): def __init__(self, iden, person_type, person_name, person_surname="", wants_accommodation="N"): self.person_name = person_name self.person_surname = person_surname self.person_type = person_type self.wants_accommodation = wants_accommodation self.iden = iden
Remove redundant full name method
Remove redundant full name method
Python
mit
peterpaints/room-allocator
class Person(object): def __init__(self, iden, person_type, person_name, person_surname="", wants_accommodation="N"): self.person_name = person_name self.person_surname = person_surname self.person_type = person_type self.wants_accommodation = wants_accommodation self.iden = iden def full_name(self): self.full_name = self.person_name + " " + self.person_surname return self.full_name Remove redundant full name method
class Person(object): def __init__(self, iden, person_type, person_name, person_surname="", wants_accommodation="N"): self.person_name = person_name self.person_surname = person_surname self.person_type = person_type self.wants_accommodation = wants_accommodation self.iden = iden
<commit_before>class Person(object): def __init__(self, iden, person_type, person_name, person_surname="", wants_accommodation="N"): self.person_name = person_name self.person_surname = person_surname self.person_type = person_type self.wants_accommodation = wants_accommodation self.iden = iden def full_name(self): self.full_name = self.person_name + " " + self.person_surname return self.full_name <commit_msg>Remove redundant full name method<commit_after>
class Person(object): def __init__(self, iden, person_type, person_name, person_surname="", wants_accommodation="N"): self.person_name = person_name self.person_surname = person_surname self.person_type = person_type self.wants_accommodation = wants_accommodation self.iden = iden
class Person(object): def __init__(self, iden, person_type, person_name, person_surname="", wants_accommodation="N"): self.person_name = person_name self.person_surname = person_surname self.person_type = person_type self.wants_accommodation = wants_accommodation self.iden = iden def full_name(self): self.full_name = self.person_name + " " + self.person_surname return self.full_name Remove redundant full name methodclass Person(object): def __init__(self, iden, person_type, person_name, person_surname="", wants_accommodation="N"): self.person_name = person_name self.person_surname = person_surname self.person_type = person_type self.wants_accommodation = wants_accommodation self.iden = iden
<commit_before>class Person(object): def __init__(self, iden, person_type, person_name, person_surname="", wants_accommodation="N"): self.person_name = person_name self.person_surname = person_surname self.person_type = person_type self.wants_accommodation = wants_accommodation self.iden = iden def full_name(self): self.full_name = self.person_name + " " + self.person_surname return self.full_name <commit_msg>Remove redundant full name method<commit_after>class Person(object): def __init__(self, iden, person_type, person_name, person_surname="", wants_accommodation="N"): self.person_name = person_name self.person_surname = person_surname self.person_type = person_type self.wants_accommodation = wants_accommodation self.iden = iden
53f967cf2cb1159e75b4bac267e163cd5a3ba156
wsgi/openshift/urls.py
wsgi/openshift/urls.py
from django.conf.urls import patterns, include, url # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Examples: url(r'^$', 'openshift.views.home', name='home'), # url(r'^openshift/', include('openshift.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: # url(r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: url(r'^admin/', include(admin.site.urls)), )
from django.conf.urls.defaults import patterns, include, url # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Examples: url(r'^$', 'openshift.views.home', name='home'), # url(r'^openshift/', include('openshift.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: # url(r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: url(r'^admin/', include(admin.site.urls)), )
Revert "Change to get Django 1.5 to work."
Revert "Change to get Django 1.5 to work." This reverts commit 92b9b557eef77f7ea4c05c74c1c229a2b508e640. It didn't resolve all the problems, and I'm out of time.
Python
agpl-3.0
esplinr/foodcheck,esplinr/foodcheck,esplinr/foodcheck,esplinr/foodcheck
from django.conf.urls import patterns, include, url # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Examples: url(r'^$', 'openshift.views.home', name='home'), # url(r'^openshift/', include('openshift.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: # url(r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: url(r'^admin/', include(admin.site.urls)), ) Revert "Change to get Django 1.5 to work." This reverts commit 92b9b557eef77f7ea4c05c74c1c229a2b508e640. It didn't resolve all the problems, and I'm out of time.
from django.conf.urls.defaults import patterns, include, url # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Examples: url(r'^$', 'openshift.views.home', name='home'), # url(r'^openshift/', include('openshift.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: # url(r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: url(r'^admin/', include(admin.site.urls)), )
<commit_before>from django.conf.urls import patterns, include, url # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Examples: url(r'^$', 'openshift.views.home', name='home'), # url(r'^openshift/', include('openshift.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: # url(r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: url(r'^admin/', include(admin.site.urls)), ) <commit_msg>Revert "Change to get Django 1.5 to work." This reverts commit 92b9b557eef77f7ea4c05c74c1c229a2b508e640. It didn't resolve all the problems, and I'm out of time.<commit_after>
from django.conf.urls.defaults import patterns, include, url # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Examples: url(r'^$', 'openshift.views.home', name='home'), # url(r'^openshift/', include('openshift.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: # url(r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: url(r'^admin/', include(admin.site.urls)), )
from django.conf.urls import patterns, include, url # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Examples: url(r'^$', 'openshift.views.home', name='home'), # url(r'^openshift/', include('openshift.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: # url(r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: url(r'^admin/', include(admin.site.urls)), ) Revert "Change to get Django 1.5 to work." This reverts commit 92b9b557eef77f7ea4c05c74c1c229a2b508e640. It didn't resolve all the problems, and I'm out of time.from django.conf.urls.defaults import patterns, include, url # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Examples: url(r'^$', 'openshift.views.home', name='home'), # url(r'^openshift/', include('openshift.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: # url(r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: url(r'^admin/', include(admin.site.urls)), )
<commit_before>from django.conf.urls import patterns, include, url # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Examples: url(r'^$', 'openshift.views.home', name='home'), # url(r'^openshift/', include('openshift.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: # url(r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: url(r'^admin/', include(admin.site.urls)), ) <commit_msg>Revert "Change to get Django 1.5 to work." This reverts commit 92b9b557eef77f7ea4c05c74c1c229a2b508e640. It didn't resolve all the problems, and I'm out of time.<commit_after>from django.conf.urls.defaults import patterns, include, url # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Examples: url(r'^$', 'openshift.views.home', name='home'), # url(r'^openshift/', include('openshift.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: # url(r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: url(r'^admin/', include(admin.site.urls)), )
aa360309f387f19f6566d08325cd1aa1131768da
bulbs/utils/filters.py
bulbs/utils/filters.py
from rest_framework import filters class CaseInsensitiveBooleanFilter(filters.BaseFilterBackend): """Set a boolean_fields tuple on the viewset and set this class as a filter_backend to filter listed fields through a case-insensitive transformation to be used for filtering. i.e. query params such as 'true' become boolean True, and params with a value 'false' become boolean False.""" def filter_queryset(self, request, queryset, view): boolean_fields = getattr(view, 'boolean_fields', None) if not boolean_fields: return queryset boolean_filters = {} for field in boolean_fields: if field in request.QUERY_PARAMS: val = request.QUERY_PARAMS[field] if val in ['true', 'True']: boolean_filters[field] = True elif val in ['false', 'False']: boolean_filters[field] = False if len(boolean_filters) > 0: return queryset.filter(**boolean_filters) return queryset
from rest_framework import filters class CaseInsensitiveBooleanFilter(filters.BaseFilterBackend): """Set a boolean_fields tuple on the viewset and set this class as a filter_backend to filter listed fields through a case-insensitive transformation to be used for filtering. i.e. query params such as 'true' become boolean True, and params with a value 'false' become boolean False.""" def filter_queryset(self, request, queryset, view): boolean_fields = getattr(view, 'boolean_fields', None) if not boolean_fields: return queryset boolean_filters = {} for field in boolean_fields: if field in request.QUERY_PARAMS: val = request.QUERY_PARAMS[field].lower() if val == 'true': boolean_filters[field] = True elif val == 'false': boolean_filters[field] = False if len(boolean_filters) > 0: return queryset.filter(**boolean_filters) return queryset
Cover every case for CaseInsensitiveBooleanFilter
Cover every case for CaseInsensitiveBooleanFilter
Python
mit
pombredanne/django-bulbs,theonion/django-bulbs,theonion/django-bulbs,theonion/django-bulbs,theonion/django-bulbs,theonion/django-bulbs,pombredanne/django-bulbs
from rest_framework import filters class CaseInsensitiveBooleanFilter(filters.BaseFilterBackend): """Set a boolean_fields tuple on the viewset and set this class as a filter_backend to filter listed fields through a case-insensitive transformation to be used for filtering. i.e. query params such as 'true' become boolean True, and params with a value 'false' become boolean False.""" def filter_queryset(self, request, queryset, view): boolean_fields = getattr(view, 'boolean_fields', None) if not boolean_fields: return queryset boolean_filters = {} for field in boolean_fields: if field in request.QUERY_PARAMS: val = request.QUERY_PARAMS[field] if val in ['true', 'True']: boolean_filters[field] = True elif val in ['false', 'False']: boolean_filters[field] = False if len(boolean_filters) > 0: return queryset.filter(**boolean_filters) return queryset Cover every case for CaseInsensitiveBooleanFilter
from rest_framework import filters class CaseInsensitiveBooleanFilter(filters.BaseFilterBackend): """Set a boolean_fields tuple on the viewset and set this class as a filter_backend to filter listed fields through a case-insensitive transformation to be used for filtering. i.e. query params such as 'true' become boolean True, and params with a value 'false' become boolean False.""" def filter_queryset(self, request, queryset, view): boolean_fields = getattr(view, 'boolean_fields', None) if not boolean_fields: return queryset boolean_filters = {} for field in boolean_fields: if field in request.QUERY_PARAMS: val = request.QUERY_PARAMS[field].lower() if val == 'true': boolean_filters[field] = True elif val == 'false': boolean_filters[field] = False if len(boolean_filters) > 0: return queryset.filter(**boolean_filters) return queryset
<commit_before>from rest_framework import filters class CaseInsensitiveBooleanFilter(filters.BaseFilterBackend): """Set a boolean_fields tuple on the viewset and set this class as a filter_backend to filter listed fields through a case-insensitive transformation to be used for filtering. i.e. query params such as 'true' become boolean True, and params with a value 'false' become boolean False.""" def filter_queryset(self, request, queryset, view): boolean_fields = getattr(view, 'boolean_fields', None) if not boolean_fields: return queryset boolean_filters = {} for field in boolean_fields: if field in request.QUERY_PARAMS: val = request.QUERY_PARAMS[field] if val in ['true', 'True']: boolean_filters[field] = True elif val in ['false', 'False']: boolean_filters[field] = False if len(boolean_filters) > 0: return queryset.filter(**boolean_filters) return queryset <commit_msg>Cover every case for CaseInsensitiveBooleanFilter<commit_after>
from rest_framework import filters class CaseInsensitiveBooleanFilter(filters.BaseFilterBackend): """Set a boolean_fields tuple on the viewset and set this class as a filter_backend to filter listed fields through a case-insensitive transformation to be used for filtering. i.e. query params such as 'true' become boolean True, and params with a value 'false' become boolean False.""" def filter_queryset(self, request, queryset, view): boolean_fields = getattr(view, 'boolean_fields', None) if not boolean_fields: return queryset boolean_filters = {} for field in boolean_fields: if field in request.QUERY_PARAMS: val = request.QUERY_PARAMS[field].lower() if val == 'true': boolean_filters[field] = True elif val == 'false': boolean_filters[field] = False if len(boolean_filters) > 0: return queryset.filter(**boolean_filters) return queryset
from rest_framework import filters class CaseInsensitiveBooleanFilter(filters.BaseFilterBackend): """Set a boolean_fields tuple on the viewset and set this class as a filter_backend to filter listed fields through a case-insensitive transformation to be used for filtering. i.e. query params such as 'true' become boolean True, and params with a value 'false' become boolean False.""" def filter_queryset(self, request, queryset, view): boolean_fields = getattr(view, 'boolean_fields', None) if not boolean_fields: return queryset boolean_filters = {} for field in boolean_fields: if field in request.QUERY_PARAMS: val = request.QUERY_PARAMS[field] if val in ['true', 'True']: boolean_filters[field] = True elif val in ['false', 'False']: boolean_filters[field] = False if len(boolean_filters) > 0: return queryset.filter(**boolean_filters) return queryset Cover every case for CaseInsensitiveBooleanFilterfrom rest_framework import filters class CaseInsensitiveBooleanFilter(filters.BaseFilterBackend): """Set a boolean_fields tuple on the viewset and set this class as a filter_backend to filter listed fields through a case-insensitive transformation to be used for filtering. i.e. query params such as 'true' become boolean True, and params with a value 'false' become boolean False.""" def filter_queryset(self, request, queryset, view): boolean_fields = getattr(view, 'boolean_fields', None) if not boolean_fields: return queryset boolean_filters = {} for field in boolean_fields: if field in request.QUERY_PARAMS: val = request.QUERY_PARAMS[field].lower() if val == 'true': boolean_filters[field] = True elif val == 'false': boolean_filters[field] = False if len(boolean_filters) > 0: return queryset.filter(**boolean_filters) return queryset
<commit_before>from rest_framework import filters class CaseInsensitiveBooleanFilter(filters.BaseFilterBackend): """Set a boolean_fields tuple on the viewset and set this class as a filter_backend to filter listed fields through a case-insensitive transformation to be used for filtering. i.e. query params such as 'true' become boolean True, and params with a value 'false' become boolean False.""" def filter_queryset(self, request, queryset, view): boolean_fields = getattr(view, 'boolean_fields', None) if not boolean_fields: return queryset boolean_filters = {} for field in boolean_fields: if field in request.QUERY_PARAMS: val = request.QUERY_PARAMS[field] if val in ['true', 'True']: boolean_filters[field] = True elif val in ['false', 'False']: boolean_filters[field] = False if len(boolean_filters) > 0: return queryset.filter(**boolean_filters) return queryset <commit_msg>Cover every case for CaseInsensitiveBooleanFilter<commit_after>from rest_framework import filters class CaseInsensitiveBooleanFilter(filters.BaseFilterBackend): """Set a boolean_fields tuple on the viewset and set this class as a filter_backend to filter listed fields through a case-insensitive transformation to be used for filtering. i.e. query params such as 'true' become boolean True, and params with a value 'false' become boolean False.""" def filter_queryset(self, request, queryset, view): boolean_fields = getattr(view, 'boolean_fields', None) if not boolean_fields: return queryset boolean_filters = {} for field in boolean_fields: if field in request.QUERY_PARAMS: val = request.QUERY_PARAMS[field].lower() if val == 'true': boolean_filters[field] = True elif val == 'false': boolean_filters[field] = False if len(boolean_filters) > 0: return queryset.filter(**boolean_filters) return queryset
74f915946c346238b6badb7c494cefa356cf6f84
busineme/core/views.py
busineme/core/views.py
"""Busine-me API Universidade de Brasilia - FGA Técnicas de Programação, 2/2015 @file views.py Views (on classic MVC, controllers) with methods that control the requisitions for the user authentication and manipulation. """ from django.views.generic import View from core.serializers import serialize_objects from .models import Busline from django.http import JsonResponse STATUS_OK = 200 STATUS_NOT_FOUND = 404 STATUS_CREATED = 201 STATUS_SERVER_ERROR = 500 class BuslineSearchResultView(View): http_method_names = [u'get', u'post'] def get(self, request): """Returns all users.""" json_data = serialize_objects(Busline.objects.all()) return JsonResponse(json_data, content_type='application/json') def getbusline(self, line_number): busline = Busline.api_filter_startswith(line_number) json_data = serialize_objects(busline) return JsonResponse(json_data, content_type='application/json')
"""Busine-me API Universidade de Brasilia - FGA Técnicas de Programação, 2/2015 @file views.py Views (on classic MVC, controllers) with methods that control the requisitions for the user authentication and manipulation. """ from django.views.generic import View from core.serializers import serialize_objects from .models import Busline from django.http import JsonResponse STATUS_OK = 200 STATUS_NOT_FOUND = 404 STATUS_CREATED = 201 STATUS_SERVER_ERROR = 500 class BuslineSearchResultView(View): http_method_names = [u'get', u'post'] def get(self, request): """Returns all users.""" json_data = serialize_objects(Busline.objects.all()) return JsonResponse(json_data, content_type='application/json') def get_busline(self, line_number): busline = Busline.api_filter_startswith(line_number) json_data = serialize_objects(busline) return JsonResponse(json_data, content_type='application/json')
Change name getbusline name method
Change name getbusline name method
Python
agpl-3.0
aldarionsevero/busine.meAPI,aldarionsevero/busine.meAPI
"""Busine-me API Universidade de Brasilia - FGA Técnicas de Programação, 2/2015 @file views.py Views (on classic MVC, controllers) with methods that control the requisitions for the user authentication and manipulation. """ from django.views.generic import View from core.serializers import serialize_objects from .models import Busline from django.http import JsonResponse STATUS_OK = 200 STATUS_NOT_FOUND = 404 STATUS_CREATED = 201 STATUS_SERVER_ERROR = 500 class BuslineSearchResultView(View): http_method_names = [u'get', u'post'] def get(self, request): """Returns all users.""" json_data = serialize_objects(Busline.objects.all()) return JsonResponse(json_data, content_type='application/json') def getbusline(self, line_number): busline = Busline.api_filter_startswith(line_number) json_data = serialize_objects(busline) return JsonResponse(json_data, content_type='application/json') Change name getbusline name method
"""Busine-me API Universidade de Brasilia - FGA Técnicas de Programação, 2/2015 @file views.py Views (on classic MVC, controllers) with methods that control the requisitions for the user authentication and manipulation. """ from django.views.generic import View from core.serializers import serialize_objects from .models import Busline from django.http import JsonResponse STATUS_OK = 200 STATUS_NOT_FOUND = 404 STATUS_CREATED = 201 STATUS_SERVER_ERROR = 500 class BuslineSearchResultView(View): http_method_names = [u'get', u'post'] def get(self, request): """Returns all users.""" json_data = serialize_objects(Busline.objects.all()) return JsonResponse(json_data, content_type='application/json') def get_busline(self, line_number): busline = Busline.api_filter_startswith(line_number) json_data = serialize_objects(busline) return JsonResponse(json_data, content_type='application/json')
<commit_before>"""Busine-me API Universidade de Brasilia - FGA Técnicas de Programação, 2/2015 @file views.py Views (on classic MVC, controllers) with methods that control the requisitions for the user authentication and manipulation. """ from django.views.generic import View from core.serializers import serialize_objects from .models import Busline from django.http import JsonResponse STATUS_OK = 200 STATUS_NOT_FOUND = 404 STATUS_CREATED = 201 STATUS_SERVER_ERROR = 500 class BuslineSearchResultView(View): http_method_names = [u'get', u'post'] def get(self, request): """Returns all users.""" json_data = serialize_objects(Busline.objects.all()) return JsonResponse(json_data, content_type='application/json') def getbusline(self, line_number): busline = Busline.api_filter_startswith(line_number) json_data = serialize_objects(busline) return JsonResponse(json_data, content_type='application/json') <commit_msg>Change name getbusline name method<commit_after>
"""Busine-me API Universidade de Brasilia - FGA Técnicas de Programação, 2/2015 @file views.py Views (on classic MVC, controllers) with methods that control the requisitions for the user authentication and manipulation. """ from django.views.generic import View from core.serializers import serialize_objects from .models import Busline from django.http import JsonResponse STATUS_OK = 200 STATUS_NOT_FOUND = 404 STATUS_CREATED = 201 STATUS_SERVER_ERROR = 500 class BuslineSearchResultView(View): http_method_names = [u'get', u'post'] def get(self, request): """Returns all users.""" json_data = serialize_objects(Busline.objects.all()) return JsonResponse(json_data, content_type='application/json') def get_busline(self, line_number): busline = Busline.api_filter_startswith(line_number) json_data = serialize_objects(busline) return JsonResponse(json_data, content_type='application/json')
"""Busine-me API Universidade de Brasilia - FGA Técnicas de Programação, 2/2015 @file views.py Views (on classic MVC, controllers) with methods that control the requisitions for the user authentication and manipulation. """ from django.views.generic import View from core.serializers import serialize_objects from .models import Busline from django.http import JsonResponse STATUS_OK = 200 STATUS_NOT_FOUND = 404 STATUS_CREATED = 201 STATUS_SERVER_ERROR = 500 class BuslineSearchResultView(View): http_method_names = [u'get', u'post'] def get(self, request): """Returns all users.""" json_data = serialize_objects(Busline.objects.all()) return JsonResponse(json_data, content_type='application/json') def getbusline(self, line_number): busline = Busline.api_filter_startswith(line_number) json_data = serialize_objects(busline) return JsonResponse(json_data, content_type='application/json') Change name getbusline name method"""Busine-me API Universidade de Brasilia - FGA Técnicas de Programação, 2/2015 @file views.py Views (on classic MVC, controllers) with methods that control the requisitions for the user authentication and manipulation. """ from django.views.generic import View from core.serializers import serialize_objects from .models import Busline from django.http import JsonResponse STATUS_OK = 200 STATUS_NOT_FOUND = 404 STATUS_CREATED = 201 STATUS_SERVER_ERROR = 500 class BuslineSearchResultView(View): http_method_names = [u'get', u'post'] def get(self, request): """Returns all users.""" json_data = serialize_objects(Busline.objects.all()) return JsonResponse(json_data, content_type='application/json') def get_busline(self, line_number): busline = Busline.api_filter_startswith(line_number) json_data = serialize_objects(busline) return JsonResponse(json_data, content_type='application/json')
<commit_before>"""Busine-me API Universidade de Brasilia - FGA Técnicas de Programação, 2/2015 @file views.py Views (on classic MVC, controllers) with methods that control the requisitions for the user authentication and manipulation. """ from django.views.generic import View from core.serializers import serialize_objects from .models import Busline from django.http import JsonResponse STATUS_OK = 200 STATUS_NOT_FOUND = 404 STATUS_CREATED = 201 STATUS_SERVER_ERROR = 500 class BuslineSearchResultView(View): http_method_names = [u'get', u'post'] def get(self, request): """Returns all users.""" json_data = serialize_objects(Busline.objects.all()) return JsonResponse(json_data, content_type='application/json') def getbusline(self, line_number): busline = Busline.api_filter_startswith(line_number) json_data = serialize_objects(busline) return JsonResponse(json_data, content_type='application/json') <commit_msg>Change name getbusline name method<commit_after>"""Busine-me API Universidade de Brasilia - FGA Técnicas de Programação, 2/2015 @file views.py Views (on classic MVC, controllers) with methods that control the requisitions for the user authentication and manipulation. """ from django.views.generic import View from core.serializers import serialize_objects from .models import Busline from django.http import JsonResponse STATUS_OK = 200 STATUS_NOT_FOUND = 404 STATUS_CREATED = 201 STATUS_SERVER_ERROR = 500 class BuslineSearchResultView(View): http_method_names = [u'get', u'post'] def get(self, request): """Returns all users.""" json_data = serialize_objects(Busline.objects.all()) return JsonResponse(json_data, content_type='application/json') def get_busline(self, line_number): busline = Busline.api_filter_startswith(line_number) json_data = serialize_objects(busline) return JsonResponse(json_data, content_type='application/json')
72bbd1a5e356b57842b07aa3a58d1e314228091d
tests/pytests/unit/pillar/test_pillar.py
tests/pytests/unit/pillar/test_pillar.py
import pytest import salt.loader import salt.pillar from salt.utils.odict import OrderedDict @pytest.mark.parametrize( "envs", ( ["a", "b", "c"], ["c", "b", "a"], ["b", "a", "c"], ), ) def test_pillar_envs_order(envs, temp_salt_minion, tmp_path): opts = temp_salt_minion.config.copy() # Stop using OrderedDict once we drop Py3.5 support opts["pillar_roots"] = OrderedDict() for envname in envs: opts["pillar_roots"][envname] = [str(tmp_path / envname)] grains = salt.loader.grains(opts) pillar = salt.pillar.Pillar( opts, grains, temp_salt_minion.id, "base", ) # The base environment is always present and as the first environment name assert pillar._get_envs() == ["base"] + envs
import pytest import salt.loader import salt.pillar from salt.utils.odict import OrderedDict @pytest.mark.parametrize( "envs", ( ["a", "b", "c"], ["c", "b", "a"], ["b", "a", "c"], ), ) def test_pillar_envs_order(envs, temp_salt_minion, tmp_path): opts = temp_salt_minion.config.copy() # Stop using OrderedDict once we drop Py3.5 support opts["pillar_roots"] = OrderedDict() for envname in envs: opts["pillar_roots"][envname] = [str(tmp_path / envname)] grains = salt.loader.grains(opts) pillar = salt.pillar.Pillar( opts, grains, temp_salt_minion.id, "base", ) # The base environment is always present and as the first environment name assert pillar._get_envs() == ["base"] + envs def test_pillar_get_tops_should_not_error_when_merging_strategy_is_none_and_no_pillarenv( temp_salt_minion, ): opts = temp_salt_minion.config.copy() opts["pillarenv"] = None opts["pillar_source_merging_strategy"] = "none" pillar = salt.pillar.Pillar( opts=opts, grains=salt.loader.grains(opts), minion_id=temp_salt_minion.id, saltenv="base", ) tops, errors = pillar.get_tops() assert not errors
Add testcase written by @waynew
Add testcase written by @waynew
Python
apache-2.0
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
import pytest import salt.loader import salt.pillar from salt.utils.odict import OrderedDict @pytest.mark.parametrize( "envs", ( ["a", "b", "c"], ["c", "b", "a"], ["b", "a", "c"], ), ) def test_pillar_envs_order(envs, temp_salt_minion, tmp_path): opts = temp_salt_minion.config.copy() # Stop using OrderedDict once we drop Py3.5 support opts["pillar_roots"] = OrderedDict() for envname in envs: opts["pillar_roots"][envname] = [str(tmp_path / envname)] grains = salt.loader.grains(opts) pillar = salt.pillar.Pillar( opts, grains, temp_salt_minion.id, "base", ) # The base environment is always present and as the first environment name assert pillar._get_envs() == ["base"] + envs Add testcase written by @waynew
import pytest import salt.loader import salt.pillar from salt.utils.odict import OrderedDict @pytest.mark.parametrize( "envs", ( ["a", "b", "c"], ["c", "b", "a"], ["b", "a", "c"], ), ) def test_pillar_envs_order(envs, temp_salt_minion, tmp_path): opts = temp_salt_minion.config.copy() # Stop using OrderedDict once we drop Py3.5 support opts["pillar_roots"] = OrderedDict() for envname in envs: opts["pillar_roots"][envname] = [str(tmp_path / envname)] grains = salt.loader.grains(opts) pillar = salt.pillar.Pillar( opts, grains, temp_salt_minion.id, "base", ) # The base environment is always present and as the first environment name assert pillar._get_envs() == ["base"] + envs def test_pillar_get_tops_should_not_error_when_merging_strategy_is_none_and_no_pillarenv( temp_salt_minion, ): opts = temp_salt_minion.config.copy() opts["pillarenv"] = None opts["pillar_source_merging_strategy"] = "none" pillar = salt.pillar.Pillar( opts=opts, grains=salt.loader.grains(opts), minion_id=temp_salt_minion.id, saltenv="base", ) tops, errors = pillar.get_tops() assert not errors
<commit_before>import pytest import salt.loader import salt.pillar from salt.utils.odict import OrderedDict @pytest.mark.parametrize( "envs", ( ["a", "b", "c"], ["c", "b", "a"], ["b", "a", "c"], ), ) def test_pillar_envs_order(envs, temp_salt_minion, tmp_path): opts = temp_salt_minion.config.copy() # Stop using OrderedDict once we drop Py3.5 support opts["pillar_roots"] = OrderedDict() for envname in envs: opts["pillar_roots"][envname] = [str(tmp_path / envname)] grains = salt.loader.grains(opts) pillar = salt.pillar.Pillar( opts, grains, temp_salt_minion.id, "base", ) # The base environment is always present and as the first environment name assert pillar._get_envs() == ["base"] + envs <commit_msg>Add testcase written by @waynew<commit_after>
import pytest import salt.loader import salt.pillar from salt.utils.odict import OrderedDict @pytest.mark.parametrize( "envs", ( ["a", "b", "c"], ["c", "b", "a"], ["b", "a", "c"], ), ) def test_pillar_envs_order(envs, temp_salt_minion, tmp_path): opts = temp_salt_minion.config.copy() # Stop using OrderedDict once we drop Py3.5 support opts["pillar_roots"] = OrderedDict() for envname in envs: opts["pillar_roots"][envname] = [str(tmp_path / envname)] grains = salt.loader.grains(opts) pillar = salt.pillar.Pillar( opts, grains, temp_salt_minion.id, "base", ) # The base environment is always present and as the first environment name assert pillar._get_envs() == ["base"] + envs def test_pillar_get_tops_should_not_error_when_merging_strategy_is_none_and_no_pillarenv( temp_salt_minion, ): opts = temp_salt_minion.config.copy() opts["pillarenv"] = None opts["pillar_source_merging_strategy"] = "none" pillar = salt.pillar.Pillar( opts=opts, grains=salt.loader.grains(opts), minion_id=temp_salt_minion.id, saltenv="base", ) tops, errors = pillar.get_tops() assert not errors
import pytest import salt.loader import salt.pillar from salt.utils.odict import OrderedDict @pytest.mark.parametrize( "envs", ( ["a", "b", "c"], ["c", "b", "a"], ["b", "a", "c"], ), ) def test_pillar_envs_order(envs, temp_salt_minion, tmp_path): opts = temp_salt_minion.config.copy() # Stop using OrderedDict once we drop Py3.5 support opts["pillar_roots"] = OrderedDict() for envname in envs: opts["pillar_roots"][envname] = [str(tmp_path / envname)] grains = salt.loader.grains(opts) pillar = salt.pillar.Pillar( opts, grains, temp_salt_minion.id, "base", ) # The base environment is always present and as the first environment name assert pillar._get_envs() == ["base"] + envs Add testcase written by @waynewimport pytest import salt.loader import salt.pillar from salt.utils.odict import OrderedDict @pytest.mark.parametrize( "envs", ( ["a", "b", "c"], ["c", "b", "a"], ["b", "a", "c"], ), ) def test_pillar_envs_order(envs, temp_salt_minion, tmp_path): opts = temp_salt_minion.config.copy() # Stop using OrderedDict once we drop Py3.5 support opts["pillar_roots"] = OrderedDict() for envname in envs: opts["pillar_roots"][envname] = [str(tmp_path / envname)] grains = salt.loader.grains(opts) pillar = salt.pillar.Pillar( opts, grains, temp_salt_minion.id, "base", ) # The base environment is always present and as the first environment name assert pillar._get_envs() == ["base"] + envs def test_pillar_get_tops_should_not_error_when_merging_strategy_is_none_and_no_pillarenv( temp_salt_minion, ): opts = temp_salt_minion.config.copy() opts["pillarenv"] = None opts["pillar_source_merging_strategy"] = "none" pillar = salt.pillar.Pillar( opts=opts, grains=salt.loader.grains(opts), minion_id=temp_salt_minion.id, saltenv="base", ) tops, errors = pillar.get_tops() assert not errors
<commit_before>import pytest import salt.loader import salt.pillar from salt.utils.odict import OrderedDict @pytest.mark.parametrize( "envs", ( ["a", "b", "c"], ["c", "b", "a"], ["b", "a", "c"], ), ) def test_pillar_envs_order(envs, temp_salt_minion, tmp_path): opts = temp_salt_minion.config.copy() # Stop using OrderedDict once we drop Py3.5 support opts["pillar_roots"] = OrderedDict() for envname in envs: opts["pillar_roots"][envname] = [str(tmp_path / envname)] grains = salt.loader.grains(opts) pillar = salt.pillar.Pillar( opts, grains, temp_salt_minion.id, "base", ) # The base environment is always present and as the first environment name assert pillar._get_envs() == ["base"] + envs <commit_msg>Add testcase written by @waynew<commit_after>import pytest import salt.loader import salt.pillar from salt.utils.odict import OrderedDict @pytest.mark.parametrize( "envs", ( ["a", "b", "c"], ["c", "b", "a"], ["b", "a", "c"], ), ) def test_pillar_envs_order(envs, temp_salt_minion, tmp_path): opts = temp_salt_minion.config.copy() # Stop using OrderedDict once we drop Py3.5 support opts["pillar_roots"] = OrderedDict() for envname in envs: opts["pillar_roots"][envname] = [str(tmp_path / envname)] grains = salt.loader.grains(opts) pillar = salt.pillar.Pillar( opts, grains, temp_salt_minion.id, "base", ) # The base environment is always present and as the first environment name assert pillar._get_envs() == ["base"] + envs def test_pillar_get_tops_should_not_error_when_merging_strategy_is_none_and_no_pillarenv( temp_salt_minion, ): opts = temp_salt_minion.config.copy() opts["pillarenv"] = None opts["pillar_source_merging_strategy"] = "none" pillar = salt.pillar.Pillar( opts=opts, grains=salt.loader.grains(opts), minion_id=temp_salt_minion.id, saltenv="base", ) tops, errors = pillar.get_tops() assert not errors
1f0195c1d1695119287e3693360ec44564ed0a09
app/main/views/sub_navigation_dictionaries.py
app/main/views/sub_navigation_dictionaries.py
def features_nav(): return [ { "name": "Features", "link": "main.features", "sub_navigation_items": [ { "name": "Emails", "link": "main.features_email", }, { "name": "Text messages", "link": "main.features_sms", }, { "name": "Letters", "link": "main.features_letters", }, ] }, { "name": "Roadmap", "link": "main.roadmap", }, { "name": "Trial mode", "link": "main.trial_mode_new", }, { "name": "Message status", "link": "main.message_status", }, { "name": "Security", "link": "main.security", }, { "name": "Terms of use", "link": "main.terms", }, { "name": "Using Notify", "link": "main.using_notify", }, ]
def features_nav(): return [ { "name": "Features", "link": "main.features", "sub_navigation_items": [ { "name": "Emails", "link": "main.features_email", }, { "name": "Text messages", "link": "main.features_sms", }, { "name": "Letters", "link": "main.features_letters", }, ] }, { "name": "Roadmap", "link": "main.roadmap", }, { "name": "Trial mode", "link": "main.trial_mode_new", }, { "name": "Message status", "link": "main.message_status", }, { "name": "Security", "link": "main.security", }, { "name": "Terms of use", "link": "main.terms", }, ]
Remove ‘using notify’ from nav (it’s a redirect now)
Remove ‘using notify’ from nav (it’s a redirect now)
Python
mit
alphagov/notifications-admin,alphagov/notifications-admin,alphagov/notifications-admin,alphagov/notifications-admin
def features_nav(): return [ { "name": "Features", "link": "main.features", "sub_navigation_items": [ { "name": "Emails", "link": "main.features_email", }, { "name": "Text messages", "link": "main.features_sms", }, { "name": "Letters", "link": "main.features_letters", }, ] }, { "name": "Roadmap", "link": "main.roadmap", }, { "name": "Trial mode", "link": "main.trial_mode_new", }, { "name": "Message status", "link": "main.message_status", }, { "name": "Security", "link": "main.security", }, { "name": "Terms of use", "link": "main.terms", }, { "name": "Using Notify", "link": "main.using_notify", }, ] Remove ‘using notify’ from nav (it’s a redirect now)
def features_nav(): return [ { "name": "Features", "link": "main.features", "sub_navigation_items": [ { "name": "Emails", "link": "main.features_email", }, { "name": "Text messages", "link": "main.features_sms", }, { "name": "Letters", "link": "main.features_letters", }, ] }, { "name": "Roadmap", "link": "main.roadmap", }, { "name": "Trial mode", "link": "main.trial_mode_new", }, { "name": "Message status", "link": "main.message_status", }, { "name": "Security", "link": "main.security", }, { "name": "Terms of use", "link": "main.terms", }, ]
<commit_before>def features_nav(): return [ { "name": "Features", "link": "main.features", "sub_navigation_items": [ { "name": "Emails", "link": "main.features_email", }, { "name": "Text messages", "link": "main.features_sms", }, { "name": "Letters", "link": "main.features_letters", }, ] }, { "name": "Roadmap", "link": "main.roadmap", }, { "name": "Trial mode", "link": "main.trial_mode_new", }, { "name": "Message status", "link": "main.message_status", }, { "name": "Security", "link": "main.security", }, { "name": "Terms of use", "link": "main.terms", }, { "name": "Using Notify", "link": "main.using_notify", }, ] <commit_msg>Remove ‘using notify’ from nav (it’s a redirect now)<commit_after>
def features_nav(): return [ { "name": "Features", "link": "main.features", "sub_navigation_items": [ { "name": "Emails", "link": "main.features_email", }, { "name": "Text messages", "link": "main.features_sms", }, { "name": "Letters", "link": "main.features_letters", }, ] }, { "name": "Roadmap", "link": "main.roadmap", }, { "name": "Trial mode", "link": "main.trial_mode_new", }, { "name": "Message status", "link": "main.message_status", }, { "name": "Security", "link": "main.security", }, { "name": "Terms of use", "link": "main.terms", }, ]
def features_nav(): return [ { "name": "Features", "link": "main.features", "sub_navigation_items": [ { "name": "Emails", "link": "main.features_email", }, { "name": "Text messages", "link": "main.features_sms", }, { "name": "Letters", "link": "main.features_letters", }, ] }, { "name": "Roadmap", "link": "main.roadmap", }, { "name": "Trial mode", "link": "main.trial_mode_new", }, { "name": "Message status", "link": "main.message_status", }, { "name": "Security", "link": "main.security", }, { "name": "Terms of use", "link": "main.terms", }, { "name": "Using Notify", "link": "main.using_notify", }, ] Remove ‘using notify’ from nav (it’s a redirect now)def features_nav(): return [ { "name": "Features", "link": "main.features", "sub_navigation_items": [ { "name": "Emails", "link": "main.features_email", }, { "name": "Text messages", "link": "main.features_sms", }, { "name": "Letters", "link": "main.features_letters", }, ] }, { "name": "Roadmap", "link": "main.roadmap", }, { "name": "Trial mode", "link": "main.trial_mode_new", }, { "name": "Message status", "link": "main.message_status", }, { "name": "Security", "link": "main.security", }, { "name": "Terms of use", "link": "main.terms", }, ]
<commit_before>def features_nav(): return [ { "name": "Features", "link": "main.features", "sub_navigation_items": [ { "name": "Emails", "link": "main.features_email", }, { "name": "Text messages", "link": "main.features_sms", }, { "name": "Letters", "link": "main.features_letters", }, ] }, { "name": "Roadmap", "link": "main.roadmap", }, { "name": "Trial mode", "link": "main.trial_mode_new", }, { "name": "Message status", "link": "main.message_status", }, { "name": "Security", "link": "main.security", }, { "name": "Terms of use", "link": "main.terms", }, { "name": "Using Notify", "link": "main.using_notify", }, ] <commit_msg>Remove ‘using notify’ from nav (it’s a redirect now)<commit_after>def features_nav(): return [ { "name": "Features", "link": "main.features", "sub_navigation_items": [ { "name": "Emails", "link": "main.features_email", }, { "name": "Text messages", "link": "main.features_sms", }, { "name": "Letters", "link": "main.features_letters", }, ] }, { "name": "Roadmap", "link": "main.roadmap", }, { "name": "Trial mode", "link": "main.trial_mode_new", }, { "name": "Message status", "link": "main.message_status", }, { "name": "Security", "link": "main.security", }, { "name": "Terms of use", "link": "main.terms", }, ]
d06d65cea4ae9efa547af43a551e24a459e0627e
tbmodels/_legacy_decode.py
tbmodels/_legacy_decode.py
from ._tb_model import Model def _decode(hf): if 'tb_model' in hf or 'hop' in hf: return _decode_model(hf) elif 'val' in hf: return _decode_val(hf) elif '0' in hf: return _decode_iterable(hf) else: raise ValueError('File structure not understood.') def _decode_iterable(hf): return [_decode(hf[key]) for key in sorted(hf, key=int)] def _decode_model(hf): return Model.from_hdf5(hf) def _decode_val(hf): return hf['val'].value
""" Defines decoding for the legacy (pre fsc.hdf5-io) HDF5 format. """ from ._tb_model import Model def _decode(hdf5_handle): """ Decode the object at the given HDF5 node. """ if 'tb_model' in hdf5_handle or 'hop' in hdf5_handle: return _decode_model(hdf5_handle) elif 'val' in hdf5_handle: return _decode_val(hdf5_handle) elif '0' in hdf5_handle: return _decode_iterable(hdf5_handle) else: raise ValueError('File structure not understood.') def _decode_iterable(hdf5_handle): return [_decode(hdf5_handle[key]) for key in sorted(hdf5_handle, key=int)] def _decode_model(hdf5_handle): return Model.from_hdf5(hdf5_handle) def _decode_val(hdf5_handle): return hdf5_handle['val'].value
Fix pylint issues in legacy_decode.
Fix pylint issues in legacy_decode.
Python
apache-2.0
Z2PackDev/TBmodels,Z2PackDev/TBmodels
from ._tb_model import Model def _decode(hf): if 'tb_model' in hf or 'hop' in hf: return _decode_model(hf) elif 'val' in hf: return _decode_val(hf) elif '0' in hf: return _decode_iterable(hf) else: raise ValueError('File structure not understood.') def _decode_iterable(hf): return [_decode(hf[key]) for key in sorted(hf, key=int)] def _decode_model(hf): return Model.from_hdf5(hf) def _decode_val(hf): return hf['val'].value Fix pylint issues in legacy_decode.
""" Defines decoding for the legacy (pre fsc.hdf5-io) HDF5 format. """ from ._tb_model import Model def _decode(hdf5_handle): """ Decode the object at the given HDF5 node. """ if 'tb_model' in hdf5_handle or 'hop' in hdf5_handle: return _decode_model(hdf5_handle) elif 'val' in hdf5_handle: return _decode_val(hdf5_handle) elif '0' in hdf5_handle: return _decode_iterable(hdf5_handle) else: raise ValueError('File structure not understood.') def _decode_iterable(hdf5_handle): return [_decode(hdf5_handle[key]) for key in sorted(hdf5_handle, key=int)] def _decode_model(hdf5_handle): return Model.from_hdf5(hdf5_handle) def _decode_val(hdf5_handle): return hdf5_handle['val'].value
<commit_before>from ._tb_model import Model def _decode(hf): if 'tb_model' in hf or 'hop' in hf: return _decode_model(hf) elif 'val' in hf: return _decode_val(hf) elif '0' in hf: return _decode_iterable(hf) else: raise ValueError('File structure not understood.') def _decode_iterable(hf): return [_decode(hf[key]) for key in sorted(hf, key=int)] def _decode_model(hf): return Model.from_hdf5(hf) def _decode_val(hf): return hf['val'].value <commit_msg>Fix pylint issues in legacy_decode.<commit_after>
""" Defines decoding for the legacy (pre fsc.hdf5-io) HDF5 format. """ from ._tb_model import Model def _decode(hdf5_handle): """ Decode the object at the given HDF5 node. """ if 'tb_model' in hdf5_handle or 'hop' in hdf5_handle: return _decode_model(hdf5_handle) elif 'val' in hdf5_handle: return _decode_val(hdf5_handle) elif '0' in hdf5_handle: return _decode_iterable(hdf5_handle) else: raise ValueError('File structure not understood.') def _decode_iterable(hdf5_handle): return [_decode(hdf5_handle[key]) for key in sorted(hdf5_handle, key=int)] def _decode_model(hdf5_handle): return Model.from_hdf5(hdf5_handle) def _decode_val(hdf5_handle): return hdf5_handle['val'].value
from ._tb_model import Model def _decode(hf): if 'tb_model' in hf or 'hop' in hf: return _decode_model(hf) elif 'val' in hf: return _decode_val(hf) elif '0' in hf: return _decode_iterable(hf) else: raise ValueError('File structure not understood.') def _decode_iterable(hf): return [_decode(hf[key]) for key in sorted(hf, key=int)] def _decode_model(hf): return Model.from_hdf5(hf) def _decode_val(hf): return hf['val'].value Fix pylint issues in legacy_decode.""" Defines decoding for the legacy (pre fsc.hdf5-io) HDF5 format. """ from ._tb_model import Model def _decode(hdf5_handle): """ Decode the object at the given HDF5 node. """ if 'tb_model' in hdf5_handle or 'hop' in hdf5_handle: return _decode_model(hdf5_handle) elif 'val' in hdf5_handle: return _decode_val(hdf5_handle) elif '0' in hdf5_handle: return _decode_iterable(hdf5_handle) else: raise ValueError('File structure not understood.') def _decode_iterable(hdf5_handle): return [_decode(hdf5_handle[key]) for key in sorted(hdf5_handle, key=int)] def _decode_model(hdf5_handle): return Model.from_hdf5(hdf5_handle) def _decode_val(hdf5_handle): return hdf5_handle['val'].value
<commit_before>from ._tb_model import Model def _decode(hf): if 'tb_model' in hf or 'hop' in hf: return _decode_model(hf) elif 'val' in hf: return _decode_val(hf) elif '0' in hf: return _decode_iterable(hf) else: raise ValueError('File structure not understood.') def _decode_iterable(hf): return [_decode(hf[key]) for key in sorted(hf, key=int)] def _decode_model(hf): return Model.from_hdf5(hf) def _decode_val(hf): return hf['val'].value <commit_msg>Fix pylint issues in legacy_decode.<commit_after>""" Defines decoding for the legacy (pre fsc.hdf5-io) HDF5 format. """ from ._tb_model import Model def _decode(hdf5_handle): """ Decode the object at the given HDF5 node. """ if 'tb_model' in hdf5_handle or 'hop' in hdf5_handle: return _decode_model(hdf5_handle) elif 'val' in hdf5_handle: return _decode_val(hdf5_handle) elif '0' in hdf5_handle: return _decode_iterable(hdf5_handle) else: raise ValueError('File structure not understood.') def _decode_iterable(hdf5_handle): return [_decode(hdf5_handle[key]) for key in sorted(hdf5_handle, key=int)] def _decode_model(hdf5_handle): return Model.from_hdf5(hdf5_handle) def _decode_val(hdf5_handle): return hdf5_handle['val'].value
3a312dc4134d28d8c8fb0020444a1bbf1277a4cb
lib/automatic_timestamps/models.py
lib/automatic_timestamps/models.py
from django.db import models import datetime class TimestampModel(models.Model): """ Extend the default Django model to add timestamps to all objects. """ class Meta: abstract = True # Timestamps! created_at = models.DateTimeField() updated_at = models.DateTimeField() def save(self): """ Override the save method to automatically set the created_at and updated_at fields with current date info. """ if self.created_at == None: self.created_at = datetime.datetime.now() self.updated_at = datetime.datetime.now() super(TimestampModel, self).save()
from django.db import models import datetime class TimestampModel(models.Model): """ Extend the default Django model to add timestamps to all objects. """ class Meta: abstract = True # Timestamps! created_at = models.DateTimeField() updated_at = models.DateTimeField() def save(self, *args, **kwargs): """ Override the save method to automatically set the created_at and updated_at fields with current date info. """ if self.created_at == None: self.created_at = datetime.datetime.now() self.updated_at = datetime.datetime.now() super(TimestampModel, self).save()
Add *args/**kwargs to save() as per the Django docs
Add *args/**kwargs to save() as per the Django docs
Python
mit
tofumatt/quotes,tofumatt/quotes
from django.db import models import datetime class TimestampModel(models.Model): """ Extend the default Django model to add timestamps to all objects. """ class Meta: abstract = True # Timestamps! created_at = models.DateTimeField() updated_at = models.DateTimeField() def save(self): """ Override the save method to automatically set the created_at and updated_at fields with current date info. """ if self.created_at == None: self.created_at = datetime.datetime.now() self.updated_at = datetime.datetime.now() super(TimestampModel, self).save() Add *args/**kwargs to save() as per the Django docs
from django.db import models import datetime class TimestampModel(models.Model): """ Extend the default Django model to add timestamps to all objects. """ class Meta: abstract = True # Timestamps! created_at = models.DateTimeField() updated_at = models.DateTimeField() def save(self, *args, **kwargs): """ Override the save method to automatically set the created_at and updated_at fields with current date info. """ if self.created_at == None: self.created_at = datetime.datetime.now() self.updated_at = datetime.datetime.now() super(TimestampModel, self).save()
<commit_before>from django.db import models import datetime class TimestampModel(models.Model): """ Extend the default Django model to add timestamps to all objects. """ class Meta: abstract = True # Timestamps! created_at = models.DateTimeField() updated_at = models.DateTimeField() def save(self): """ Override the save method to automatically set the created_at and updated_at fields with current date info. """ if self.created_at == None: self.created_at = datetime.datetime.now() self.updated_at = datetime.datetime.now() super(TimestampModel, self).save() <commit_msg>Add *args/**kwargs to save() as per the Django docs<commit_after>
from django.db import models import datetime class TimestampModel(models.Model): """ Extend the default Django model to add timestamps to all objects. """ class Meta: abstract = True # Timestamps! created_at = models.DateTimeField() updated_at = models.DateTimeField() def save(self, *args, **kwargs): """ Override the save method to automatically set the created_at and updated_at fields with current date info. """ if self.created_at == None: self.created_at = datetime.datetime.now() self.updated_at = datetime.datetime.now() super(TimestampModel, self).save()
from django.db import models import datetime class TimestampModel(models.Model): """ Extend the default Django model to add timestamps to all objects. """ class Meta: abstract = True # Timestamps! created_at = models.DateTimeField() updated_at = models.DateTimeField() def save(self): """ Override the save method to automatically set the created_at and updated_at fields with current date info. """ if self.created_at == None: self.created_at = datetime.datetime.now() self.updated_at = datetime.datetime.now() super(TimestampModel, self).save() Add *args/**kwargs to save() as per the Django docsfrom django.db import models import datetime class TimestampModel(models.Model): """ Extend the default Django model to add timestamps to all objects. """ class Meta: abstract = True # Timestamps! created_at = models.DateTimeField() updated_at = models.DateTimeField() def save(self, *args, **kwargs): """ Override the save method to automatically set the created_at and updated_at fields with current date info. """ if self.created_at == None: self.created_at = datetime.datetime.now() self.updated_at = datetime.datetime.now() super(TimestampModel, self).save()
<commit_before>from django.db import models import datetime class TimestampModel(models.Model): """ Extend the default Django model to add timestamps to all objects. """ class Meta: abstract = True # Timestamps! created_at = models.DateTimeField() updated_at = models.DateTimeField() def save(self): """ Override the save method to automatically set the created_at and updated_at fields with current date info. """ if self.created_at == None: self.created_at = datetime.datetime.now() self.updated_at = datetime.datetime.now() super(TimestampModel, self).save() <commit_msg>Add *args/**kwargs to save() as per the Django docs<commit_after>from django.db import models import datetime class TimestampModel(models.Model): """ Extend the default Django model to add timestamps to all objects. """ class Meta: abstract = True # Timestamps! created_at = models.DateTimeField() updated_at = models.DateTimeField() def save(self, *args, **kwargs): """ Override the save method to automatically set the created_at and updated_at fields with current date info. """ if self.created_at == None: self.created_at = datetime.datetime.now() self.updated_at = datetime.datetime.now() super(TimestampModel, self).save()
4db2d879cb8ee7d8ddd1543e6aed50f40e44ca66
src/pi/scanning_proxy.py
src/pi/scanning_proxy.py
"""Philips hue proxy code.""" import abc import logging import sys import threading from pi import proxy class ScanningProxy(proxy.Proxy): """A proxy object with a background scan thread.""" __metaclass__ = abc.ABCMeta def __init__(self, refresh_period): self._refresh_period = refresh_period self._exiting = False self._scan_thread_condition = threading.Condition() self._scan_thread = threading.Thread(target=self._scan) self._scan_thread.daemon = True self._scan_thread.start() @proxy.command def scan(self): with self._scan_thread_condition: self._scan_thread_condition.notify() def _scan(self): """Loop thread for scanning.""" while not self._exiting: # We always do a scan on start up. try: self._scan_once() except: logging.error('Error during %s scan', self.__class__.__name__, exc_info=sys.exc_info()) with self._scan_thread_condition: self._scan_thread_condition.wait(self._refresh_period) if self._exiting: break logging.info('Exiting %s scan thread', self.__class__.__name__) @abc.abstractmethod def _scan_once(self): pass def stop(self): self._exiting = True self.scan() self._scan_thread.join()
"""Philips hue proxy code.""" import abc import logging import sys import threading from pi import proxy class ScanningProxy(proxy.Proxy): """A proxy object with a background scan thread.""" __metaclass__ = abc.ABCMeta def __init__(self, refresh_period): self._refresh_period = refresh_period self._exiting = False self._scan_thread_condition = threading.Condition() self._scan_thread = threading.Thread(target=self._scan) self._scan_thread.daemon = True self._scan_thread.start() @proxy.command def scan(self): with self._scan_thread_condition: self._scan_thread_condition.notify() def _scan(self): """Loop thread for scanning.""" while not self._exiting: # We always do a scan on start up. try: self._scan_once() except: logging.error('Error during %s scan', self.__class__.__name__, exc_info=sys.exc_info()) with self._scan_thread_condition: if not self._exiting: self._scan_thread_condition.wait(self._refresh_period) if self._exiting: break logging.info('Exiting %s scan thread', self.__class__.__name__) @abc.abstractmethod def _scan_once(self): pass def stop(self): with self._scan_thread_condition: self._exiting = True self._scan_thread_condition.notify() self._scan_thread.join()
Fix race on exit in scanning proxy.
Fix race on exit in scanning proxy.
Python
mit
tomwilkie/awesomation,tomwilkie/awesomation,tomwilkie/awesomation,tomwilkie/awesomation,tomwilkie/awesomation
"""Philips hue proxy code.""" import abc import logging import sys import threading from pi import proxy class ScanningProxy(proxy.Proxy): """A proxy object with a background scan thread.""" __metaclass__ = abc.ABCMeta def __init__(self, refresh_period): self._refresh_period = refresh_period self._exiting = False self._scan_thread_condition = threading.Condition() self._scan_thread = threading.Thread(target=self._scan) self._scan_thread.daemon = True self._scan_thread.start() @proxy.command def scan(self): with self._scan_thread_condition: self._scan_thread_condition.notify() def _scan(self): """Loop thread for scanning.""" while not self._exiting: # We always do a scan on start up. try: self._scan_once() except: logging.error('Error during %s scan', self.__class__.__name__, exc_info=sys.exc_info()) with self._scan_thread_condition: self._scan_thread_condition.wait(self._refresh_period) if self._exiting: break logging.info('Exiting %s scan thread', self.__class__.__name__) @abc.abstractmethod def _scan_once(self): pass def stop(self): self._exiting = True self.scan() self._scan_thread.join() Fix race on exit in scanning proxy.
"""Philips hue proxy code.""" import abc import logging import sys import threading from pi import proxy class ScanningProxy(proxy.Proxy): """A proxy object with a background scan thread.""" __metaclass__ = abc.ABCMeta def __init__(self, refresh_period): self._refresh_period = refresh_period self._exiting = False self._scan_thread_condition = threading.Condition() self._scan_thread = threading.Thread(target=self._scan) self._scan_thread.daemon = True self._scan_thread.start() @proxy.command def scan(self): with self._scan_thread_condition: self._scan_thread_condition.notify() def _scan(self): """Loop thread for scanning.""" while not self._exiting: # We always do a scan on start up. try: self._scan_once() except: logging.error('Error during %s scan', self.__class__.__name__, exc_info=sys.exc_info()) with self._scan_thread_condition: if not self._exiting: self._scan_thread_condition.wait(self._refresh_period) if self._exiting: break logging.info('Exiting %s scan thread', self.__class__.__name__) @abc.abstractmethod def _scan_once(self): pass def stop(self): with self._scan_thread_condition: self._exiting = True self._scan_thread_condition.notify() self._scan_thread.join()
<commit_before>"""Philips hue proxy code.""" import abc import logging import sys import threading from pi import proxy class ScanningProxy(proxy.Proxy): """A proxy object with a background scan thread.""" __metaclass__ = abc.ABCMeta def __init__(self, refresh_period): self._refresh_period = refresh_period self._exiting = False self._scan_thread_condition = threading.Condition() self._scan_thread = threading.Thread(target=self._scan) self._scan_thread.daemon = True self._scan_thread.start() @proxy.command def scan(self): with self._scan_thread_condition: self._scan_thread_condition.notify() def _scan(self): """Loop thread for scanning.""" while not self._exiting: # We always do a scan on start up. try: self._scan_once() except: logging.error('Error during %s scan', self.__class__.__name__, exc_info=sys.exc_info()) with self._scan_thread_condition: self._scan_thread_condition.wait(self._refresh_period) if self._exiting: break logging.info('Exiting %s scan thread', self.__class__.__name__) @abc.abstractmethod def _scan_once(self): pass def stop(self): self._exiting = True self.scan() self._scan_thread.join() <commit_msg>Fix race on exit in scanning proxy.<commit_after>
"""Philips hue proxy code.""" import abc import logging import sys import threading from pi import proxy class ScanningProxy(proxy.Proxy): """A proxy object with a background scan thread.""" __metaclass__ = abc.ABCMeta def __init__(self, refresh_period): self._refresh_period = refresh_period self._exiting = False self._scan_thread_condition = threading.Condition() self._scan_thread = threading.Thread(target=self._scan) self._scan_thread.daemon = True self._scan_thread.start() @proxy.command def scan(self): with self._scan_thread_condition: self._scan_thread_condition.notify() def _scan(self): """Loop thread for scanning.""" while not self._exiting: # We always do a scan on start up. try: self._scan_once() except: logging.error('Error during %s scan', self.__class__.__name__, exc_info=sys.exc_info()) with self._scan_thread_condition: if not self._exiting: self._scan_thread_condition.wait(self._refresh_period) if self._exiting: break logging.info('Exiting %s scan thread', self.__class__.__name__) @abc.abstractmethod def _scan_once(self): pass def stop(self): with self._scan_thread_condition: self._exiting = True self._scan_thread_condition.notify() self._scan_thread.join()
"""Philips hue proxy code.""" import abc import logging import sys import threading from pi import proxy class ScanningProxy(proxy.Proxy): """A proxy object with a background scan thread.""" __metaclass__ = abc.ABCMeta def __init__(self, refresh_period): self._refresh_period = refresh_period self._exiting = False self._scan_thread_condition = threading.Condition() self._scan_thread = threading.Thread(target=self._scan) self._scan_thread.daemon = True self._scan_thread.start() @proxy.command def scan(self): with self._scan_thread_condition: self._scan_thread_condition.notify() def _scan(self): """Loop thread for scanning.""" while not self._exiting: # We always do a scan on start up. try: self._scan_once() except: logging.error('Error during %s scan', self.__class__.__name__, exc_info=sys.exc_info()) with self._scan_thread_condition: self._scan_thread_condition.wait(self._refresh_period) if self._exiting: break logging.info('Exiting %s scan thread', self.__class__.__name__) @abc.abstractmethod def _scan_once(self): pass def stop(self): self._exiting = True self.scan() self._scan_thread.join() Fix race on exit in scanning proxy."""Philips hue proxy code.""" import abc import logging import sys import threading from pi import proxy class ScanningProxy(proxy.Proxy): """A proxy object with a background scan thread.""" __metaclass__ = abc.ABCMeta def __init__(self, refresh_period): self._refresh_period = refresh_period self._exiting = False self._scan_thread_condition = threading.Condition() self._scan_thread = threading.Thread(target=self._scan) self._scan_thread.daemon = True self._scan_thread.start() @proxy.command def scan(self): with self._scan_thread_condition: self._scan_thread_condition.notify() def _scan(self): """Loop thread for scanning.""" while not self._exiting: # We always do a scan on start up. try: self._scan_once() except: logging.error('Error during %s scan', self.__class__.__name__, exc_info=sys.exc_info()) with self._scan_thread_condition: if not self._exiting: self._scan_thread_condition.wait(self._refresh_period) if self._exiting: break logging.info('Exiting %s scan thread', self.__class__.__name__) @abc.abstractmethod def _scan_once(self): pass def stop(self): with self._scan_thread_condition: self._exiting = True self._scan_thread_condition.notify() self._scan_thread.join()
<commit_before>"""Philips hue proxy code.""" import abc import logging import sys import threading from pi import proxy class ScanningProxy(proxy.Proxy): """A proxy object with a background scan thread.""" __metaclass__ = abc.ABCMeta def __init__(self, refresh_period): self._refresh_period = refresh_period self._exiting = False self._scan_thread_condition = threading.Condition() self._scan_thread = threading.Thread(target=self._scan) self._scan_thread.daemon = True self._scan_thread.start() @proxy.command def scan(self): with self._scan_thread_condition: self._scan_thread_condition.notify() def _scan(self): """Loop thread for scanning.""" while not self._exiting: # We always do a scan on start up. try: self._scan_once() except: logging.error('Error during %s scan', self.__class__.__name__, exc_info=sys.exc_info()) with self._scan_thread_condition: self._scan_thread_condition.wait(self._refresh_period) if self._exiting: break logging.info('Exiting %s scan thread', self.__class__.__name__) @abc.abstractmethod def _scan_once(self): pass def stop(self): self._exiting = True self.scan() self._scan_thread.join() <commit_msg>Fix race on exit in scanning proxy.<commit_after>"""Philips hue proxy code.""" import abc import logging import sys import threading from pi import proxy class ScanningProxy(proxy.Proxy): """A proxy object with a background scan thread.""" __metaclass__ = abc.ABCMeta def __init__(self, refresh_period): self._refresh_period = refresh_period self._exiting = False self._scan_thread_condition = threading.Condition() self._scan_thread = threading.Thread(target=self._scan) self._scan_thread.daemon = True self._scan_thread.start() @proxy.command def scan(self): with self._scan_thread_condition: self._scan_thread_condition.notify() def _scan(self): """Loop thread for scanning.""" while not self._exiting: # We always do a scan on start up. try: self._scan_once() except: logging.error('Error during %s scan', self.__class__.__name__, exc_info=sys.exc_info()) with self._scan_thread_condition: if not self._exiting: self._scan_thread_condition.wait(self._refresh_period) if self._exiting: break logging.info('Exiting %s scan thread', self.__class__.__name__) @abc.abstractmethod def _scan_once(self): pass def stop(self): with self._scan_thread_condition: self._exiting = True self._scan_thread_condition.notify() self._scan_thread.join()
058a3e918cccde7dcace79df2b257bd29277bcd0
tests/builtins/test_abs.py
tests/builtins/test_abs.py
from .. utils import TranspileTestCase, BuiltinFunctionTestCase class AbsTests(TranspileTestCase): pass class BuiltinAbsFunctionTests(BuiltinFunctionTestCase, TranspileTestCase): functions = ["abs"] not_implemented = [ 'test_bool', 'test_bytearray', 'test_bytes', 'test_class', 'test_complex', 'test_frozenset', 'test_set', ]
from unittest import expectedFailure from .. utils import TranspileTestCase, BuiltinFunctionTestCase class AbsTests(TranspileTestCase): @expectedFailure def test_abs_not_implemented(self): self.assertCodeExecution(""" class NotAbsLike: pass x = NotAbsLike() print(abs(x)) """) class BuiltinAbsFunctionTests(BuiltinFunctionTestCase, TranspileTestCase): functions = ["abs"] not_implemented = [ 'test_bool', 'test_bytearray', 'test_bytes', 'test_class', 'test_complex', 'test_frozenset', 'test_set', ]
Add failing test for builtin abs() on objects without __abs__()
Add failing test for builtin abs() on objects without __abs__()
Python
bsd-3-clause
pombredanne/voc,ASP1234/voc,freakboy3742/voc,gEt-rIgHt-jR/voc,cflee/voc,cflee/voc,Felix5721/voc,glasnt/voc,glasnt/voc,pombredanne/voc,freakboy3742/voc,gEt-rIgHt-jR/voc,ASP1234/voc,Felix5721/voc
from .. utils import TranspileTestCase, BuiltinFunctionTestCase class AbsTests(TranspileTestCase): pass class BuiltinAbsFunctionTests(BuiltinFunctionTestCase, TranspileTestCase): functions = ["abs"] not_implemented = [ 'test_bool', 'test_bytearray', 'test_bytes', 'test_class', 'test_complex', 'test_frozenset', 'test_set', ] Add failing test for builtin abs() on objects without __abs__()
from unittest import expectedFailure from .. utils import TranspileTestCase, BuiltinFunctionTestCase class AbsTests(TranspileTestCase): @expectedFailure def test_abs_not_implemented(self): self.assertCodeExecution(""" class NotAbsLike: pass x = NotAbsLike() print(abs(x)) """) class BuiltinAbsFunctionTests(BuiltinFunctionTestCase, TranspileTestCase): functions = ["abs"] not_implemented = [ 'test_bool', 'test_bytearray', 'test_bytes', 'test_class', 'test_complex', 'test_frozenset', 'test_set', ]
<commit_before>from .. utils import TranspileTestCase, BuiltinFunctionTestCase class AbsTests(TranspileTestCase): pass class BuiltinAbsFunctionTests(BuiltinFunctionTestCase, TranspileTestCase): functions = ["abs"] not_implemented = [ 'test_bool', 'test_bytearray', 'test_bytes', 'test_class', 'test_complex', 'test_frozenset', 'test_set', ] <commit_msg>Add failing test for builtin abs() on objects without __abs__()<commit_after>
from unittest import expectedFailure from .. utils import TranspileTestCase, BuiltinFunctionTestCase class AbsTests(TranspileTestCase): @expectedFailure def test_abs_not_implemented(self): self.assertCodeExecution(""" class NotAbsLike: pass x = NotAbsLike() print(abs(x)) """) class BuiltinAbsFunctionTests(BuiltinFunctionTestCase, TranspileTestCase): functions = ["abs"] not_implemented = [ 'test_bool', 'test_bytearray', 'test_bytes', 'test_class', 'test_complex', 'test_frozenset', 'test_set', ]
from .. utils import TranspileTestCase, BuiltinFunctionTestCase class AbsTests(TranspileTestCase): pass class BuiltinAbsFunctionTests(BuiltinFunctionTestCase, TranspileTestCase): functions = ["abs"] not_implemented = [ 'test_bool', 'test_bytearray', 'test_bytes', 'test_class', 'test_complex', 'test_frozenset', 'test_set', ] Add failing test for builtin abs() on objects without __abs__()from unittest import expectedFailure from .. utils import TranspileTestCase, BuiltinFunctionTestCase class AbsTests(TranspileTestCase): @expectedFailure def test_abs_not_implemented(self): self.assertCodeExecution(""" class NotAbsLike: pass x = NotAbsLike() print(abs(x)) """) class BuiltinAbsFunctionTests(BuiltinFunctionTestCase, TranspileTestCase): functions = ["abs"] not_implemented = [ 'test_bool', 'test_bytearray', 'test_bytes', 'test_class', 'test_complex', 'test_frozenset', 'test_set', ]
<commit_before>from .. utils import TranspileTestCase, BuiltinFunctionTestCase class AbsTests(TranspileTestCase): pass class BuiltinAbsFunctionTests(BuiltinFunctionTestCase, TranspileTestCase): functions = ["abs"] not_implemented = [ 'test_bool', 'test_bytearray', 'test_bytes', 'test_class', 'test_complex', 'test_frozenset', 'test_set', ] <commit_msg>Add failing test for builtin abs() on objects without __abs__()<commit_after>from unittest import expectedFailure from .. utils import TranspileTestCase, BuiltinFunctionTestCase class AbsTests(TranspileTestCase): @expectedFailure def test_abs_not_implemented(self): self.assertCodeExecution(""" class NotAbsLike: pass x = NotAbsLike() print(abs(x)) """) class BuiltinAbsFunctionTests(BuiltinFunctionTestCase, TranspileTestCase): functions = ["abs"] not_implemented = [ 'test_bool', 'test_bytearray', 'test_bytes', 'test_class', 'test_complex', 'test_frozenset', 'test_set', ]
bc4fb65f76aa011e44bbe01b7965bc99eff5d85e
tests/test_recalcitrant.py
tests/test_recalcitrant.py
"Test for recalcitrant and obtuse graphs to describe" from wordgraph.points import Point import wordgraph import random from utilities import EPOCH_START, time_values def test_time_goes_backwards(): "A valid time series where time changes linearly backwards" values = [1.0] * 10 times = (EPOCH_START-i for i in range(10)) datapoints = [Point(x=t, y=v) for (v, t) in zip(values, time)] features = wordgraph.describe(datapoints) assert features is None def test_random_data(): "A time series of 50 data points where every value is random" rng = random.Random(0) values = [rng.random() for i in range(50)] datapoints = time_values(values) features = wordgraph.describe(datapoints) assert features is None
"Test for recalcitrant and obtuse graphs to describe" from wordgraph.points import Point import wordgraph import random import pytest from utilities import EPOCH_START, time_values def test_time_goes_backwards(): "A valid time series where time changes linearly backwards" values = [1.0] * 10 times = (EPOCH_START-i for i in range(10)) datapoints = [Point(x=t, y=v) for (v, t) in zip(values, time)] features = wordgraph.describe(datapoints) assert features is None def test_random_data(): "A time series of 50 data points where every value is random" rng = random.Random(0) values = [rng.random() for i in range(50)] datapoints = time_values(values) features = wordgraph.describe(datapoints) assert features is None def test_too_few_points(): """A time series with too few data points to be analysed. Expected to raise an exception. """ with pytest.raises(ValueError): features = wordgraph.describe([Point(x=0, y=0)]) def test_nonuniform_time_periods(): """A time series where time periods are wildly different. Expected to raise an exception. """ times = [1, 3, 4, 6, 7, 9, 10] datapoints = [Point(x=t, y=1.0) for t in times] with pytest.raises(ValueError): features = wordgraph.describe(datapoints)
Test expected failures of the anlayzer
Test expected failures of the anlayzer The analyzer is not expected to cope with too few data points for time series with greatly varying time ranges. It should raise an exception in these cases.
Python
apache-2.0
tleeuwenburg/wordgraph,tleeuwenburg/wordgraph
"Test for recalcitrant and obtuse graphs to describe" from wordgraph.points import Point import wordgraph import random from utilities import EPOCH_START, time_values def test_time_goes_backwards(): "A valid time series where time changes linearly backwards" values = [1.0] * 10 times = (EPOCH_START-i for i in range(10)) datapoints = [Point(x=t, y=v) for (v, t) in zip(values, time)] features = wordgraph.describe(datapoints) assert features is None def test_random_data(): "A time series of 50 data points where every value is random" rng = random.Random(0) values = [rng.random() for i in range(50)] datapoints = time_values(values) features = wordgraph.describe(datapoints) assert features is None Test expected failures of the anlayzer The analyzer is not expected to cope with too few data points for time series with greatly varying time ranges. It should raise an exception in these cases.
"Test for recalcitrant and obtuse graphs to describe" from wordgraph.points import Point import wordgraph import random import pytest from utilities import EPOCH_START, time_values def test_time_goes_backwards(): "A valid time series where time changes linearly backwards" values = [1.0] * 10 times = (EPOCH_START-i for i in range(10)) datapoints = [Point(x=t, y=v) for (v, t) in zip(values, time)] features = wordgraph.describe(datapoints) assert features is None def test_random_data(): "A time series of 50 data points where every value is random" rng = random.Random(0) values = [rng.random() for i in range(50)] datapoints = time_values(values) features = wordgraph.describe(datapoints) assert features is None def test_too_few_points(): """A time series with too few data points to be analysed. Expected to raise an exception. """ with pytest.raises(ValueError): features = wordgraph.describe([Point(x=0, y=0)]) def test_nonuniform_time_periods(): """A time series where time periods are wildly different. Expected to raise an exception. """ times = [1, 3, 4, 6, 7, 9, 10] datapoints = [Point(x=t, y=1.0) for t in times] with pytest.raises(ValueError): features = wordgraph.describe(datapoints)
<commit_before>"Test for recalcitrant and obtuse graphs to describe" from wordgraph.points import Point import wordgraph import random from utilities import EPOCH_START, time_values def test_time_goes_backwards(): "A valid time series where time changes linearly backwards" values = [1.0] * 10 times = (EPOCH_START-i for i in range(10)) datapoints = [Point(x=t, y=v) for (v, t) in zip(values, time)] features = wordgraph.describe(datapoints) assert features is None def test_random_data(): "A time series of 50 data points where every value is random" rng = random.Random(0) values = [rng.random() for i in range(50)] datapoints = time_values(values) features = wordgraph.describe(datapoints) assert features is None <commit_msg>Test expected failures of the anlayzer The analyzer is not expected to cope with too few data points for time series with greatly varying time ranges. It should raise an exception in these cases.<commit_after>
"Test for recalcitrant and obtuse graphs to describe" from wordgraph.points import Point import wordgraph import random import pytest from utilities import EPOCH_START, time_values def test_time_goes_backwards(): "A valid time series where time changes linearly backwards" values = [1.0] * 10 times = (EPOCH_START-i for i in range(10)) datapoints = [Point(x=t, y=v) for (v, t) in zip(values, time)] features = wordgraph.describe(datapoints) assert features is None def test_random_data(): "A time series of 50 data points where every value is random" rng = random.Random(0) values = [rng.random() for i in range(50)] datapoints = time_values(values) features = wordgraph.describe(datapoints) assert features is None def test_too_few_points(): """A time series with too few data points to be analysed. Expected to raise an exception. """ with pytest.raises(ValueError): features = wordgraph.describe([Point(x=0, y=0)]) def test_nonuniform_time_periods(): """A time series where time periods are wildly different. Expected to raise an exception. """ times = [1, 3, 4, 6, 7, 9, 10] datapoints = [Point(x=t, y=1.0) for t in times] with pytest.raises(ValueError): features = wordgraph.describe(datapoints)
"Test for recalcitrant and obtuse graphs to describe" from wordgraph.points import Point import wordgraph import random from utilities import EPOCH_START, time_values def test_time_goes_backwards(): "A valid time series where time changes linearly backwards" values = [1.0] * 10 times = (EPOCH_START-i for i in range(10)) datapoints = [Point(x=t, y=v) for (v, t) in zip(values, time)] features = wordgraph.describe(datapoints) assert features is None def test_random_data(): "A time series of 50 data points where every value is random" rng = random.Random(0) values = [rng.random() for i in range(50)] datapoints = time_values(values) features = wordgraph.describe(datapoints) assert features is None Test expected failures of the anlayzer The analyzer is not expected to cope with too few data points for time series with greatly varying time ranges. It should raise an exception in these cases."Test for recalcitrant and obtuse graphs to describe" from wordgraph.points import Point import wordgraph import random import pytest from utilities import EPOCH_START, time_values def test_time_goes_backwards(): "A valid time series where time changes linearly backwards" values = [1.0] * 10 times = (EPOCH_START-i for i in range(10)) datapoints = [Point(x=t, y=v) for (v, t) in zip(values, time)] features = wordgraph.describe(datapoints) assert features is None def test_random_data(): "A time series of 50 data points where every value is random" rng = random.Random(0) values = [rng.random() for i in range(50)] datapoints = time_values(values) features = wordgraph.describe(datapoints) assert features is None def test_too_few_points(): """A time series with too few data points to be analysed. Expected to raise an exception. """ with pytest.raises(ValueError): features = wordgraph.describe([Point(x=0, y=0)]) def test_nonuniform_time_periods(): """A time series where time periods are wildly different. Expected to raise an exception. """ times = [1, 3, 4, 6, 7, 9, 10] datapoints = [Point(x=t, y=1.0) for t in times] with pytest.raises(ValueError): features = wordgraph.describe(datapoints)
<commit_before>"Test for recalcitrant and obtuse graphs to describe" from wordgraph.points import Point import wordgraph import random from utilities import EPOCH_START, time_values def test_time_goes_backwards(): "A valid time series where time changes linearly backwards" values = [1.0] * 10 times = (EPOCH_START-i for i in range(10)) datapoints = [Point(x=t, y=v) for (v, t) in zip(values, time)] features = wordgraph.describe(datapoints) assert features is None def test_random_data(): "A time series of 50 data points where every value is random" rng = random.Random(0) values = [rng.random() for i in range(50)] datapoints = time_values(values) features = wordgraph.describe(datapoints) assert features is None <commit_msg>Test expected failures of the anlayzer The analyzer is not expected to cope with too few data points for time series with greatly varying time ranges. It should raise an exception in these cases.<commit_after>"Test for recalcitrant and obtuse graphs to describe" from wordgraph.points import Point import wordgraph import random import pytest from utilities import EPOCH_START, time_values def test_time_goes_backwards(): "A valid time series where time changes linearly backwards" values = [1.0] * 10 times = (EPOCH_START-i for i in range(10)) datapoints = [Point(x=t, y=v) for (v, t) in zip(values, time)] features = wordgraph.describe(datapoints) assert features is None def test_random_data(): "A time series of 50 data points where every value is random" rng = random.Random(0) values = [rng.random() for i in range(50)] datapoints = time_values(values) features = wordgraph.describe(datapoints) assert features is None def test_too_few_points(): """A time series with too few data points to be analysed. Expected to raise an exception. """ with pytest.raises(ValueError): features = wordgraph.describe([Point(x=0, y=0)]) def test_nonuniform_time_periods(): """A time series where time periods are wildly different. Expected to raise an exception. """ times = [1, 3, 4, 6, 7, 9, 10] datapoints = [Point(x=t, y=1.0) for t in times] with pytest.raises(ValueError): features = wordgraph.describe(datapoints)
e562223cd45c89b9ce4d5a70ef4f4ad5daea968b
deflect/urls.py
deflect/urls.py
from django.conf.urls import patterns from django.conf.urls import url from .views import redirect urlpatterns = patterns('', url(r'^(?P<key>[a-zA-Z0-9]+)$', redirect, name='deflect-redirect'), )
from django.conf.urls import patterns from django.conf.urls import url from .views import redirect urlpatterns = patterns('', url(r'^(?P<key>[a-zA-Z0-9-]+)$', redirect, name='deflect-redirect'), )
Allow dashes in URL path regexp
Allow dashes in URL path regexp
Python
bsd-3-clause
jbittel/django-deflect
from django.conf.urls import patterns from django.conf.urls import url from .views import redirect urlpatterns = patterns('', url(r'^(?P<key>[a-zA-Z0-9]+)$', redirect, name='deflect-redirect'), ) Allow dashes in URL path regexp
from django.conf.urls import patterns from django.conf.urls import url from .views import redirect urlpatterns = patterns('', url(r'^(?P<key>[a-zA-Z0-9-]+)$', redirect, name='deflect-redirect'), )
<commit_before>from django.conf.urls import patterns from django.conf.urls import url from .views import redirect urlpatterns = patterns('', url(r'^(?P<key>[a-zA-Z0-9]+)$', redirect, name='deflect-redirect'), ) <commit_msg>Allow dashes in URL path regexp<commit_after>
from django.conf.urls import patterns from django.conf.urls import url from .views import redirect urlpatterns = patterns('', url(r'^(?P<key>[a-zA-Z0-9-]+)$', redirect, name='deflect-redirect'), )
from django.conf.urls import patterns from django.conf.urls import url from .views import redirect urlpatterns = patterns('', url(r'^(?P<key>[a-zA-Z0-9]+)$', redirect, name='deflect-redirect'), ) Allow dashes in URL path regexpfrom django.conf.urls import patterns from django.conf.urls import url from .views import redirect urlpatterns = patterns('', url(r'^(?P<key>[a-zA-Z0-9-]+)$', redirect, name='deflect-redirect'), )
<commit_before>from django.conf.urls import patterns from django.conf.urls import url from .views import redirect urlpatterns = patterns('', url(r'^(?P<key>[a-zA-Z0-9]+)$', redirect, name='deflect-redirect'), ) <commit_msg>Allow dashes in URL path regexp<commit_after>from django.conf.urls import patterns from django.conf.urls import url from .views import redirect urlpatterns = patterns('', url(r'^(?P<key>[a-zA-Z0-9-]+)$', redirect, name='deflect-redirect'), )
48da1258ddaa7b8c740eba67fc82edb11b163b64
server_env_ebill_paynet/__manifest__.py
server_env_ebill_paynet/__manifest__.py
# Copyright 2020 Camptocamp SA # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html) { "name": "Server environment for Ebill Paynet", "version": "13.0.1.0.0", "author": "Camptocamp,Odoo Community Association (OCA)", "license": "AGPL-3", "category": "Tools", "depends": ["server_environment", "ebill_paynet"], "website": "https://www.camptocamp.com", "installable": True, }
# Copyright 2020 Camptocamp SA # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html) { "name": "Server environment for Ebill Paynet", "version": "13.0.1.0.0", "author": "Camptocamp,Odoo Community Association (OCA)", "license": "AGPL-3", "category": "Tools", "depends": ["server_environment", "ebill_paynet"], "website": "https://github.com/OCA/l10n-switzerland", "installable": True, }
Fix linting and add info in roadmap
Fix linting and add info in roadmap
Python
agpl-3.0
OCA/l10n-switzerland,OCA/l10n-switzerland
# Copyright 2020 Camptocamp SA # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html) { "name": "Server environment for Ebill Paynet", "version": "13.0.1.0.0", "author": "Camptocamp,Odoo Community Association (OCA)", "license": "AGPL-3", "category": "Tools", "depends": ["server_environment", "ebill_paynet"], "website": "https://www.camptocamp.com", "installable": True, } Fix linting and add info in roadmap
# Copyright 2020 Camptocamp SA # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html) { "name": "Server environment for Ebill Paynet", "version": "13.0.1.0.0", "author": "Camptocamp,Odoo Community Association (OCA)", "license": "AGPL-3", "category": "Tools", "depends": ["server_environment", "ebill_paynet"], "website": "https://github.com/OCA/l10n-switzerland", "installable": True, }
<commit_before># Copyright 2020 Camptocamp SA # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html) { "name": "Server environment for Ebill Paynet", "version": "13.0.1.0.0", "author": "Camptocamp,Odoo Community Association (OCA)", "license": "AGPL-3", "category": "Tools", "depends": ["server_environment", "ebill_paynet"], "website": "https://www.camptocamp.com", "installable": True, } <commit_msg>Fix linting and add info in roadmap<commit_after>
# Copyright 2020 Camptocamp SA # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html) { "name": "Server environment for Ebill Paynet", "version": "13.0.1.0.0", "author": "Camptocamp,Odoo Community Association (OCA)", "license": "AGPL-3", "category": "Tools", "depends": ["server_environment", "ebill_paynet"], "website": "https://github.com/OCA/l10n-switzerland", "installable": True, }
# Copyright 2020 Camptocamp SA # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html) { "name": "Server environment for Ebill Paynet", "version": "13.0.1.0.0", "author": "Camptocamp,Odoo Community Association (OCA)", "license": "AGPL-3", "category": "Tools", "depends": ["server_environment", "ebill_paynet"], "website": "https://www.camptocamp.com", "installable": True, } Fix linting and add info in roadmap# Copyright 2020 Camptocamp SA # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html) { "name": "Server environment for Ebill Paynet", "version": "13.0.1.0.0", "author": "Camptocamp,Odoo Community Association (OCA)", "license": "AGPL-3", "category": "Tools", "depends": ["server_environment", "ebill_paynet"], "website": "https://github.com/OCA/l10n-switzerland", "installable": True, }
<commit_before># Copyright 2020 Camptocamp SA # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html) { "name": "Server environment for Ebill Paynet", "version": "13.0.1.0.0", "author": "Camptocamp,Odoo Community Association (OCA)", "license": "AGPL-3", "category": "Tools", "depends": ["server_environment", "ebill_paynet"], "website": "https://www.camptocamp.com", "installable": True, } <commit_msg>Fix linting and add info in roadmap<commit_after># Copyright 2020 Camptocamp SA # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html) { "name": "Server environment for Ebill Paynet", "version": "13.0.1.0.0", "author": "Camptocamp,Odoo Community Association (OCA)", "license": "AGPL-3", "category": "Tools", "depends": ["server_environment", "ebill_paynet"], "website": "https://github.com/OCA/l10n-switzerland", "installable": True, }
c70a409c7717fa62517b69f8a5f20f10d5325751
test/common/memcached_workload_common.py
test/common/memcached_workload_common.py
# This is a (hopefully temporary) shim that uses the rdb protocol to # implement part of the memcache API import contextlib import rdb_workload_common @contextlib.contextmanager def make_memcache_connection(opts): with rdb_workload_common.make_table_and_connection(opts) as (table, conn): yield MemcacheRdbShim(table, conn) class MemcacheRdbShim(object): def __init__(self, table, conn): self.table = table self.conn = conn def get(self, key): response = self.table.get(key).run(self.conn) if response: return response['val'] def set(self, key, val): response = self.table.insert({ 'id': key, 'val': val }, upsert=True ).run(self.conn) error = response.get('first_error') if error: raise Exception(error) return response['inserted'] | response['replaced'] | response['unchanged'] def delete(self, key): response = self.table.get(key).delete().run(self.conn) error = response.get('first_error') if error: raise Exception(error) return response['deleted'] def option_parser_for_memcache(): return rdb_workload_common.option_parser_for_connect()
# This is a (hopefully temporary) shim that uses the rdb protocol to # implement part of the memcache API import contextlib import rdb_workload_common @contextlib.contextmanager def make_memcache_connection(opts): with rdb_workload_common.make_table_and_connection(opts) as (table, conn): yield MemcacheRdbShim(table, conn) class MemcacheRdbShim(object): def __init__(self, table, conn): self.table = table self.conn = conn def get(self, key): response = self.table.get(key).run(self.conn) if response: return response['val'] def set(self, key, val): response = self.table.insert({ 'id': key, 'val': val }, conflict='replace' ).run(self.conn) error = response.get('first_error') if error: raise Exception(error) return response['inserted'] | response['replaced'] | response['unchanged'] def delete(self, key): response = self.table.get(key).delete().run(self.conn) error = response.get('first_error') if error: raise Exception(error) return response['deleted'] def option_parser_for_memcache(): return rdb_workload_common.option_parser_for_connect()
Replace upsert=True with conflict='replace' in tests
Replace upsert=True with conflict='replace' in tests Review 1804 by @gchpaco Related to #2733
Python
apache-2.0
jesseditson/rethinkdb,lenstr/rethinkdb,gdi2290/rethinkdb,grandquista/rethinkdb,catroot/rethinkdb,bpradipt/rethinkdb,gavioto/rethinkdb,jesseditson/rethinkdb,sbusso/rethinkdb,sontek/rethinkdb,bchavez/rethinkdb,wujf/rethinkdb,greyhwndz/rethinkdb,Qinusty/rethinkdb,gdi2290/rethinkdb,urandu/rethinkdb,jmptrader/rethinkdb,bchavez/rethinkdb,pap/rethinkdb,grandquista/rethinkdb,AntouanK/rethinkdb,elkingtonmcb/rethinkdb,alash3al/rethinkdb,losywee/rethinkdb,jmptrader/rethinkdb,wkennington/rethinkdb,mbroadst/rethinkdb,robertjpayne/rethinkdb,matthaywardwebdesign/rethinkdb,losywee/rethinkdb,lenstr/rethinkdb,sbusso/rethinkdb,Qinusty/rethinkdb,niieani/rethinkdb,scripni/rethinkdb,bpradipt/rethinkdb,KSanthanam/rethinkdb,jmptrader/rethinkdb,yakovenkodenis/rethinkdb,gavioto/rethinkdb,dparnell/rethinkdb,pap/rethinkdb,catroot/rethinkdb,rrampage/rethinkdb,rrampage/rethinkdb,eliangidoni/rethinkdb,sebadiaz/rethinkdb,catroot/rethinkdb,marshall007/rethinkdb,RubenKelevra/rethinkdb,Qinusty/rethinkdb,4talesa/rethinkdb,niieani/rethinkdb,scripni/rethinkdb,wujf/rethinkdb,yaolinz/rethinkdb,dparnell/rethinkdb,sontek/rethinkdb,gavioto/rethinkdb,mquandalle/rethinkdb,alash3al/rethinkdb,JackieXie168/rethinkdb,rrampage/rethinkdb,AntouanK/rethinkdb,wujf/rethinkdb,niieani/rethinkdb,JackieXie168/rethinkdb,marshall007/rethinkdb,yaolinz/rethinkdb,wojons/rethinkdb,tempbottle/rethinkdb,ajose01/rethinkdb,sbusso/rethinkdb,robertjpayne/rethinkdb,matthaywardwebdesign/rethinkdb,AntouanK/rethinkdb,yaolinz/rethinkdb,mquandalle/rethinkdb,tempbottle/rethinkdb,tempbottle/rethinkdb,marshall007/rethinkdb,catroot/rethinkdb,Wilbeibi/rethinkdb,jesseditson/rethinkdb,bpradipt/rethinkdb,4talesa/rethinkdb,rrampage/rethinkdb,yaolinz/rethinkdb,urandu/rethinkdb,bchavez/rethinkdb,sebadiaz/rethinkdb,wojons/rethinkdb,captainpete/rethinkdb,4talesa/rethinkdb,sbusso/rethinkdb,eliangidoni/rethinkdb,grandquista/rethinkdb,sbusso/rethinkdb,RubenKelevra/rethinkdb,spblightadv/rethinkdb,mcanthony/rethinkdb,captainpete/rethinkdb,gavioto/rethinkdb,ayumilong/rethinkdb,catroot/rethinkdb,Qinusty/rethinkdb,mbroadst/rethinkdb,spblightadv/rethinkdb,jesseditson/rethinkdb,sontek/rethinkdb,bchavez/rethinkdb,RubenKelevra/rethinkdb,catroot/rethinkdb,JackieXie168/rethinkdb,dparnell/rethinkdb,victorbriz/rethinkdb,niieani/rethinkdb,sontek/rethinkdb,elkingtonmcb/rethinkdb,bpradipt/rethinkdb,RubenKelevra/rethinkdb,robertjpayne/rethinkdb,ajose01/rethinkdb,dparnell/rethinkdb,lenstr/rethinkdb,KSanthanam/rethinkdb,catroot/rethinkdb,eliangidoni/rethinkdb,captainpete/rethinkdb,grandquista/rethinkdb,eliangidoni/rethinkdb,greyhwndz/rethinkdb,grandquista/rethinkdb,gdi2290/rethinkdb,mquandalle/rethinkdb,alash3al/rethinkdb,victorbriz/rethinkdb,wujf/rethinkdb,matthaywardwebdesign/rethinkdb,RubenKelevra/rethinkdb,sebadiaz/rethinkdb,lenstr/rethinkdb,alash3al/rethinkdb,tempbottle/rethinkdb,lenstr/rethinkdb,jmptrader/rethinkdb,ajose01/rethinkdb,scripni/rethinkdb,sontek/rethinkdb,yaolinz/rethinkdb,tempbottle/rethinkdb,tempbottle/rethinkdb,victorbriz/rethinkdb,mcanthony/rethinkdb,rrampage/rethinkdb,dparnell/rethinkdb,scripni/rethinkdb,yakovenkodenis/rethinkdb,spblightadv/rethinkdb,matthaywardwebdesign/rethinkdb,wojons/rethinkdb,losywee/rethinkdb,Qinusty/rethinkdb,KSanthanam/rethinkdb,gdi2290/rethinkdb,grandquista/rethinkdb,urandu/rethinkdb,scripni/rethinkdb,wujf/rethinkdb,4talesa/rethinkdb,bpradipt/rethinkdb,victorbriz/rethinkdb,JackieXie168/rethinkdb,AntouanK/rethinkdb,mquandalle/rethinkdb,dparnell/rethinkdb,victorbriz/rethinkdb,elkingtonmcb/rethinkdb,grandquista/rethinkdb,wujf/rethinkdb,niieani/rethinkdb,alash3al/rethinkdb,captainpete/rethinkdb,wojons/rethinkdb,eliangidoni/rethinkdb,spblightadv/rethinkdb,Wilbeibi/rethinkdb,sebadiaz/rethinkdb,ayumilong/rethinkdb,bpradipt/rethinkdb,lenstr/rethinkdb,gdi2290/rethinkdb,catroot/rethinkdb,pap/rethinkdb,matthaywardwebdesign/rethinkdb,yakovenkodenis/rethinkdb,jesseditson/rethinkdb,Wilbeibi/rethinkdb,urandu/rethinkdb,captainpete/rethinkdb,ayumilong/rethinkdb,wkennington/rethinkdb,yakovenkodenis/rethinkdb,mcanthony/rethinkdb,jesseditson/rethinkdb,robertjpayne/rethinkdb,marshall007/rethinkdb,bchavez/rethinkdb,marshall007/rethinkdb,elkingtonmcb/rethinkdb,sbusso/rethinkdb,robertjpayne/rethinkdb,marshall007/rethinkdb,eliangidoni/rethinkdb,gdi2290/rethinkdb,JackieXie168/rethinkdb,elkingtonmcb/rethinkdb,losywee/rethinkdb,KSanthanam/rethinkdb,KSanthanam/rethinkdb,gavioto/rethinkdb,AntouanK/rethinkdb,marshall007/rethinkdb,jesseditson/rethinkdb,losywee/rethinkdb,bchavez/rethinkdb,spblightadv/rethinkdb,Qinusty/rethinkdb,elkingtonmcb/rethinkdb,yaolinz/rethinkdb,mbroadst/rethinkdb,wojons/rethinkdb,gdi2290/rethinkdb,ajose01/rethinkdb,mbroadst/rethinkdb,urandu/rethinkdb,wkennington/rethinkdb,rrampage/rethinkdb,ayumilong/rethinkdb,sontek/rethinkdb,sebadiaz/rethinkdb,alash3al/rethinkdb,urandu/rethinkdb,Wilbeibi/rethinkdb,mbroadst/rethinkdb,mquandalle/rethinkdb,Qinusty/rethinkdb,spblightadv/rethinkdb,rrampage/rethinkdb,KSanthanam/rethinkdb,matthaywardwebdesign/rethinkdb,tempbottle/rethinkdb,gavioto/rethinkdb,greyhwndz/rethinkdb,Qinusty/rethinkdb,eliangidoni/rethinkdb,jmptrader/rethinkdb,yaolinz/rethinkdb,losywee/rethinkdb,losywee/rethinkdb,yakovenkodenis/rethinkdb,yakovenkodenis/rethinkdb,bchavez/rethinkdb,yakovenkodenis/rethinkdb,4talesa/rethinkdb,sontek/rethinkdb,dparnell/rethinkdb,bpradipt/rethinkdb,matthaywardwebdesign/rethinkdb,RubenKelevra/rethinkdb,mcanthony/rethinkdb,eliangidoni/rethinkdb,greyhwndz/rethinkdb,matthaywardwebdesign/rethinkdb,lenstr/rethinkdb,sbusso/rethinkdb,captainpete/rethinkdb,Wilbeibi/rethinkdb,4talesa/rethinkdb,ajose01/rethinkdb,AntouanK/rethinkdb,robertjpayne/rethinkdb,ayumilong/rethinkdb,mquandalle/rethinkdb,mcanthony/rethinkdb,victorbriz/rethinkdb,bchavez/rethinkdb,yakovenkodenis/rethinkdb,wkennington/rethinkdb,gavioto/rethinkdb,jesseditson/rethinkdb,mquandalle/rethinkdb,wujf/rethinkdb,sebadiaz/rethinkdb,mbroadst/rethinkdb,Qinusty/rethinkdb,AntouanK/rethinkdb,grandquista/rethinkdb,JackieXie168/rethinkdb,spblightadv/rethinkdb,Wilbeibi/rethinkdb,wkennington/rethinkdb,urandu/rethinkdb,bchavez/rethinkdb,mcanthony/rethinkdb,KSanthanam/rethinkdb,ayumilong/rethinkdb,jmptrader/rethinkdb,wojons/rethinkdb,greyhwndz/rethinkdb,victorbriz/rethinkdb,niieani/rethinkdb,mbroadst/rethinkdb,dparnell/rethinkdb,ayumilong/rethinkdb,mcanthony/rethinkdb,urandu/rethinkdb,wkennington/rethinkdb,ayumilong/rethinkdb,victorbriz/rethinkdb,spblightadv/rethinkdb,niieani/rethinkdb,pap/rethinkdb,wkennington/rethinkdb,sebadiaz/rethinkdb,niieani/rethinkdb,ajose01/rethinkdb,captainpete/rethinkdb,wojons/rethinkdb,Wilbeibi/rethinkdb,mbroadst/rethinkdb,AntouanK/rethinkdb,RubenKelevra/rethinkdb,greyhwndz/rethinkdb,robertjpayne/rethinkdb,JackieXie168/rethinkdb,RubenKelevra/rethinkdb,elkingtonmcb/rethinkdb,robertjpayne/rethinkdb,jmptrader/rethinkdb,mcanthony/rethinkdb,ajose01/rethinkdb,marshall007/rethinkdb,robertjpayne/rethinkdb,bpradipt/rethinkdb,jmptrader/rethinkdb,scripni/rethinkdb,eliangidoni/rethinkdb,ajose01/rethinkdb,captainpete/rethinkdb,alash3al/rethinkdb,4talesa/rethinkdb,pap/rethinkdb,JackieXie168/rethinkdb,Wilbeibi/rethinkdb,elkingtonmcb/rethinkdb,alash3al/rethinkdb,pap/rethinkdb,lenstr/rethinkdb,pap/rethinkdb,4talesa/rethinkdb,tempbottle/rethinkdb,mquandalle/rethinkdb,wojons/rethinkdb,grandquista/rethinkdb,rrampage/rethinkdb,KSanthanam/rethinkdb,gavioto/rethinkdb,sontek/rethinkdb,greyhwndz/rethinkdb,scripni/rethinkdb,sbusso/rethinkdb,JackieXie168/rethinkdb,pap/rethinkdb,losywee/rethinkdb,bpradipt/rethinkdb,mbroadst/rethinkdb,dparnell/rethinkdb,wkennington/rethinkdb,yaolinz/rethinkdb,sebadiaz/rethinkdb,scripni/rethinkdb,greyhwndz/rethinkdb
# This is a (hopefully temporary) shim that uses the rdb protocol to # implement part of the memcache API import contextlib import rdb_workload_common @contextlib.contextmanager def make_memcache_connection(opts): with rdb_workload_common.make_table_and_connection(opts) as (table, conn): yield MemcacheRdbShim(table, conn) class MemcacheRdbShim(object): def __init__(self, table, conn): self.table = table self.conn = conn def get(self, key): response = self.table.get(key).run(self.conn) if response: return response['val'] def set(self, key, val): response = self.table.insert({ 'id': key, 'val': val }, upsert=True ).run(self.conn) error = response.get('first_error') if error: raise Exception(error) return response['inserted'] | response['replaced'] | response['unchanged'] def delete(self, key): response = self.table.get(key).delete().run(self.conn) error = response.get('first_error') if error: raise Exception(error) return response['deleted'] def option_parser_for_memcache(): return rdb_workload_common.option_parser_for_connect() Replace upsert=True with conflict='replace' in tests Review 1804 by @gchpaco Related to #2733
# This is a (hopefully temporary) shim that uses the rdb protocol to # implement part of the memcache API import contextlib import rdb_workload_common @contextlib.contextmanager def make_memcache_connection(opts): with rdb_workload_common.make_table_and_connection(opts) as (table, conn): yield MemcacheRdbShim(table, conn) class MemcacheRdbShim(object): def __init__(self, table, conn): self.table = table self.conn = conn def get(self, key): response = self.table.get(key).run(self.conn) if response: return response['val'] def set(self, key, val): response = self.table.insert({ 'id': key, 'val': val }, conflict='replace' ).run(self.conn) error = response.get('first_error') if error: raise Exception(error) return response['inserted'] | response['replaced'] | response['unchanged'] def delete(self, key): response = self.table.get(key).delete().run(self.conn) error = response.get('first_error') if error: raise Exception(error) return response['deleted'] def option_parser_for_memcache(): return rdb_workload_common.option_parser_for_connect()
<commit_before># This is a (hopefully temporary) shim that uses the rdb protocol to # implement part of the memcache API import contextlib import rdb_workload_common @contextlib.contextmanager def make_memcache_connection(opts): with rdb_workload_common.make_table_and_connection(opts) as (table, conn): yield MemcacheRdbShim(table, conn) class MemcacheRdbShim(object): def __init__(self, table, conn): self.table = table self.conn = conn def get(self, key): response = self.table.get(key).run(self.conn) if response: return response['val'] def set(self, key, val): response = self.table.insert({ 'id': key, 'val': val }, upsert=True ).run(self.conn) error = response.get('first_error') if error: raise Exception(error) return response['inserted'] | response['replaced'] | response['unchanged'] def delete(self, key): response = self.table.get(key).delete().run(self.conn) error = response.get('first_error') if error: raise Exception(error) return response['deleted'] def option_parser_for_memcache(): return rdb_workload_common.option_parser_for_connect() <commit_msg>Replace upsert=True with conflict='replace' in tests Review 1804 by @gchpaco Related to #2733<commit_after>
# This is a (hopefully temporary) shim that uses the rdb protocol to # implement part of the memcache API import contextlib import rdb_workload_common @contextlib.contextmanager def make_memcache_connection(opts): with rdb_workload_common.make_table_and_connection(opts) as (table, conn): yield MemcacheRdbShim(table, conn) class MemcacheRdbShim(object): def __init__(self, table, conn): self.table = table self.conn = conn def get(self, key): response = self.table.get(key).run(self.conn) if response: return response['val'] def set(self, key, val): response = self.table.insert({ 'id': key, 'val': val }, conflict='replace' ).run(self.conn) error = response.get('first_error') if error: raise Exception(error) return response['inserted'] | response['replaced'] | response['unchanged'] def delete(self, key): response = self.table.get(key).delete().run(self.conn) error = response.get('first_error') if error: raise Exception(error) return response['deleted'] def option_parser_for_memcache(): return rdb_workload_common.option_parser_for_connect()
# This is a (hopefully temporary) shim that uses the rdb protocol to # implement part of the memcache API import contextlib import rdb_workload_common @contextlib.contextmanager def make_memcache_connection(opts): with rdb_workload_common.make_table_and_connection(opts) as (table, conn): yield MemcacheRdbShim(table, conn) class MemcacheRdbShim(object): def __init__(self, table, conn): self.table = table self.conn = conn def get(self, key): response = self.table.get(key).run(self.conn) if response: return response['val'] def set(self, key, val): response = self.table.insert({ 'id': key, 'val': val }, upsert=True ).run(self.conn) error = response.get('first_error') if error: raise Exception(error) return response['inserted'] | response['replaced'] | response['unchanged'] def delete(self, key): response = self.table.get(key).delete().run(self.conn) error = response.get('first_error') if error: raise Exception(error) return response['deleted'] def option_parser_for_memcache(): return rdb_workload_common.option_parser_for_connect() Replace upsert=True with conflict='replace' in tests Review 1804 by @gchpaco Related to #2733# This is a (hopefully temporary) shim that uses the rdb protocol to # implement part of the memcache API import contextlib import rdb_workload_common @contextlib.contextmanager def make_memcache_connection(opts): with rdb_workload_common.make_table_and_connection(opts) as (table, conn): yield MemcacheRdbShim(table, conn) class MemcacheRdbShim(object): def __init__(self, table, conn): self.table = table self.conn = conn def get(self, key): response = self.table.get(key).run(self.conn) if response: return response['val'] def set(self, key, val): response = self.table.insert({ 'id': key, 'val': val }, conflict='replace' ).run(self.conn) error = response.get('first_error') if error: raise Exception(error) return response['inserted'] | response['replaced'] | response['unchanged'] def delete(self, key): response = self.table.get(key).delete().run(self.conn) error = response.get('first_error') if error: raise Exception(error) return response['deleted'] def option_parser_for_memcache(): return rdb_workload_common.option_parser_for_connect()
<commit_before># This is a (hopefully temporary) shim that uses the rdb protocol to # implement part of the memcache API import contextlib import rdb_workload_common @contextlib.contextmanager def make_memcache_connection(opts): with rdb_workload_common.make_table_and_connection(opts) as (table, conn): yield MemcacheRdbShim(table, conn) class MemcacheRdbShim(object): def __init__(self, table, conn): self.table = table self.conn = conn def get(self, key): response = self.table.get(key).run(self.conn) if response: return response['val'] def set(self, key, val): response = self.table.insert({ 'id': key, 'val': val }, upsert=True ).run(self.conn) error = response.get('first_error') if error: raise Exception(error) return response['inserted'] | response['replaced'] | response['unchanged'] def delete(self, key): response = self.table.get(key).delete().run(self.conn) error = response.get('first_error') if error: raise Exception(error) return response['deleted'] def option_parser_for_memcache(): return rdb_workload_common.option_parser_for_connect() <commit_msg>Replace upsert=True with conflict='replace' in tests Review 1804 by @gchpaco Related to #2733<commit_after># This is a (hopefully temporary) shim that uses the rdb protocol to # implement part of the memcache API import contextlib import rdb_workload_common @contextlib.contextmanager def make_memcache_connection(opts): with rdb_workload_common.make_table_and_connection(opts) as (table, conn): yield MemcacheRdbShim(table, conn) class MemcacheRdbShim(object): def __init__(self, table, conn): self.table = table self.conn = conn def get(self, key): response = self.table.get(key).run(self.conn) if response: return response['val'] def set(self, key, val): response = self.table.insert({ 'id': key, 'val': val }, conflict='replace' ).run(self.conn) error = response.get('first_error') if error: raise Exception(error) return response['inserted'] | response['replaced'] | response['unchanged'] def delete(self, key): response = self.table.get(key).delete().run(self.conn) error = response.get('first_error') if error: raise Exception(error) return response['deleted'] def option_parser_for_memcache(): return rdb_workload_common.option_parser_for_connect()
db30a8044f790dec8dd18255786d39e31021c7df
imbox/utils.py
imbox/utils.py
import logging logger = logging.getLogger(__name__) def str_encode(value='', encoding=None, errors='strict'): logger.debug("Encode str {} with and errors {}".format(value, encoding, errors)) return str(value, encoding, errors) def str_decode(value='', encoding=None, errors='strict'): if isinstance(value, str): return bytes(value, encoding, errors).decode('utf-8') elif isinstance(value, bytes): return value.decode(encoding or 'utf-8', errors=errors) else: raise TypeError("Cannot decode '{}' object".format(value.__class__))
import logging logger = logging.getLogger(__name__) def str_encode(value='', encoding=None, errors='strict'): logger.debug("Encode str {value} with encoding {encoding} and errors {errors}".format( value=value, encoding=encoding, errors=errors)) return str(value, encoding, errors) def str_decode(value='', encoding=None, errors='strict'): if isinstance(value, str): return bytes(value, encoding, errors).decode('utf-8') elif isinstance(value, bytes): return value.decode(encoding or 'utf-8', errors=errors) else: raise TypeError("Cannot decode '{}' object".format(value.__class__))
Add a forgotten parameter in a log line
Add a forgotten parameter in a log line
Python
mit
martinrusev/imbox
import logging logger = logging.getLogger(__name__) def str_encode(value='', encoding=None, errors='strict'): logger.debug("Encode str {} with and errors {}".format(value, encoding, errors)) return str(value, encoding, errors) def str_decode(value='', encoding=None, errors='strict'): if isinstance(value, str): return bytes(value, encoding, errors).decode('utf-8') elif isinstance(value, bytes): return value.decode(encoding or 'utf-8', errors=errors) else: raise TypeError("Cannot decode '{}' object".format(value.__class__)) Add a forgotten parameter in a log line
import logging logger = logging.getLogger(__name__) def str_encode(value='', encoding=None, errors='strict'): logger.debug("Encode str {value} with encoding {encoding} and errors {errors}".format( value=value, encoding=encoding, errors=errors)) return str(value, encoding, errors) def str_decode(value='', encoding=None, errors='strict'): if isinstance(value, str): return bytes(value, encoding, errors).decode('utf-8') elif isinstance(value, bytes): return value.decode(encoding or 'utf-8', errors=errors) else: raise TypeError("Cannot decode '{}' object".format(value.__class__))
<commit_before>import logging logger = logging.getLogger(__name__) def str_encode(value='', encoding=None, errors='strict'): logger.debug("Encode str {} with and errors {}".format(value, encoding, errors)) return str(value, encoding, errors) def str_decode(value='', encoding=None, errors='strict'): if isinstance(value, str): return bytes(value, encoding, errors).decode('utf-8') elif isinstance(value, bytes): return value.decode(encoding or 'utf-8', errors=errors) else: raise TypeError("Cannot decode '{}' object".format(value.__class__)) <commit_msg>Add a forgotten parameter in a log line<commit_after>
import logging logger = logging.getLogger(__name__) def str_encode(value='', encoding=None, errors='strict'): logger.debug("Encode str {value} with encoding {encoding} and errors {errors}".format( value=value, encoding=encoding, errors=errors)) return str(value, encoding, errors) def str_decode(value='', encoding=None, errors='strict'): if isinstance(value, str): return bytes(value, encoding, errors).decode('utf-8') elif isinstance(value, bytes): return value.decode(encoding or 'utf-8', errors=errors) else: raise TypeError("Cannot decode '{}' object".format(value.__class__))
import logging logger = logging.getLogger(__name__) def str_encode(value='', encoding=None, errors='strict'): logger.debug("Encode str {} with and errors {}".format(value, encoding, errors)) return str(value, encoding, errors) def str_decode(value='', encoding=None, errors='strict'): if isinstance(value, str): return bytes(value, encoding, errors).decode('utf-8') elif isinstance(value, bytes): return value.decode(encoding or 'utf-8', errors=errors) else: raise TypeError("Cannot decode '{}' object".format(value.__class__)) Add a forgotten parameter in a log lineimport logging logger = logging.getLogger(__name__) def str_encode(value='', encoding=None, errors='strict'): logger.debug("Encode str {value} with encoding {encoding} and errors {errors}".format( value=value, encoding=encoding, errors=errors)) return str(value, encoding, errors) def str_decode(value='', encoding=None, errors='strict'): if isinstance(value, str): return bytes(value, encoding, errors).decode('utf-8') elif isinstance(value, bytes): return value.decode(encoding or 'utf-8', errors=errors) else: raise TypeError("Cannot decode '{}' object".format(value.__class__))
<commit_before>import logging logger = logging.getLogger(__name__) def str_encode(value='', encoding=None, errors='strict'): logger.debug("Encode str {} with and errors {}".format(value, encoding, errors)) return str(value, encoding, errors) def str_decode(value='', encoding=None, errors='strict'): if isinstance(value, str): return bytes(value, encoding, errors).decode('utf-8') elif isinstance(value, bytes): return value.decode(encoding or 'utf-8', errors=errors) else: raise TypeError("Cannot decode '{}' object".format(value.__class__)) <commit_msg>Add a forgotten parameter in a log line<commit_after>import logging logger = logging.getLogger(__name__) def str_encode(value='', encoding=None, errors='strict'): logger.debug("Encode str {value} with encoding {encoding} and errors {errors}".format( value=value, encoding=encoding, errors=errors)) return str(value, encoding, errors) def str_decode(value='', encoding=None, errors='strict'): if isinstance(value, str): return bytes(value, encoding, errors).decode('utf-8') elif isinstance(value, bytes): return value.decode(encoding or 'utf-8', errors=errors) else: raise TypeError("Cannot decode '{}' object".format(value.__class__))
440295122eec2a73012c6cb6c75ba384cc25b17b
formapi/calls.py
formapi/calls.py
from django.forms import forms class APICall(forms.Form): def __init__(self, api_key=None, *args, **kwargs): super(APICall, self).__init__(*args, **kwargs) self.api_key = api_key def add_error(self, error_msg): errors = self.non_field_errors() errors.append(error_msg) self._errors[forms.NON_FIELD_ERRORS] = errors def clean(self): for name, data in self.cleaned_data.iteritems(): setattr(self, name, data) return super(APICall, self).clean() def action(self, test): raise NotImplementedError('APIForms must implement action(self, test)')
# coding=utf-8 from django.forms import forms class APICall(forms.Form): def __init__(self, api_key=None, *args, **kwargs): super(APICall, self).__init__(*args, **kwargs) self.api_key = api_key def add_error(self, error_msg, field_name=forms.NON_FIELD_ERRORS): # TODO: with Django master you would just raise ValidationError({field_name: error_msg}) self._errors.setdefault(field_name, self.error_class()).append(error_msg) def clean(self): for name, data in self.cleaned_data.iteritems(): setattr(self, name, data) return super(APICall, self).clean() def action(self, test): raise NotImplementedError('APIForms must implement action(self, test)')
Allow adding error message for specified field
Allow adding error message for specified field
Python
mit
5monkeys/django-formapi,andreif/django-formapi,andreif/django-formapi,5monkeys/django-formapi
from django.forms import forms class APICall(forms.Form): def __init__(self, api_key=None, *args, **kwargs): super(APICall, self).__init__(*args, **kwargs) self.api_key = api_key def add_error(self, error_msg): errors = self.non_field_errors() errors.append(error_msg) self._errors[forms.NON_FIELD_ERRORS] = errors def clean(self): for name, data in self.cleaned_data.iteritems(): setattr(self, name, data) return super(APICall, self).clean() def action(self, test): raise NotImplementedError('APIForms must implement action(self, test)') Allow adding error message for specified field
# coding=utf-8 from django.forms import forms class APICall(forms.Form): def __init__(self, api_key=None, *args, **kwargs): super(APICall, self).__init__(*args, **kwargs) self.api_key = api_key def add_error(self, error_msg, field_name=forms.NON_FIELD_ERRORS): # TODO: with Django master you would just raise ValidationError({field_name: error_msg}) self._errors.setdefault(field_name, self.error_class()).append(error_msg) def clean(self): for name, data in self.cleaned_data.iteritems(): setattr(self, name, data) return super(APICall, self).clean() def action(self, test): raise NotImplementedError('APIForms must implement action(self, test)')
<commit_before>from django.forms import forms class APICall(forms.Form): def __init__(self, api_key=None, *args, **kwargs): super(APICall, self).__init__(*args, **kwargs) self.api_key = api_key def add_error(self, error_msg): errors = self.non_field_errors() errors.append(error_msg) self._errors[forms.NON_FIELD_ERRORS] = errors def clean(self): for name, data in self.cleaned_data.iteritems(): setattr(self, name, data) return super(APICall, self).clean() def action(self, test): raise NotImplementedError('APIForms must implement action(self, test)') <commit_msg>Allow adding error message for specified field<commit_after>
# coding=utf-8 from django.forms import forms class APICall(forms.Form): def __init__(self, api_key=None, *args, **kwargs): super(APICall, self).__init__(*args, **kwargs) self.api_key = api_key def add_error(self, error_msg, field_name=forms.NON_FIELD_ERRORS): # TODO: with Django master you would just raise ValidationError({field_name: error_msg}) self._errors.setdefault(field_name, self.error_class()).append(error_msg) def clean(self): for name, data in self.cleaned_data.iteritems(): setattr(self, name, data) return super(APICall, self).clean() def action(self, test): raise NotImplementedError('APIForms must implement action(self, test)')
from django.forms import forms class APICall(forms.Form): def __init__(self, api_key=None, *args, **kwargs): super(APICall, self).__init__(*args, **kwargs) self.api_key = api_key def add_error(self, error_msg): errors = self.non_field_errors() errors.append(error_msg) self._errors[forms.NON_FIELD_ERRORS] = errors def clean(self): for name, data in self.cleaned_data.iteritems(): setattr(self, name, data) return super(APICall, self).clean() def action(self, test): raise NotImplementedError('APIForms must implement action(self, test)') Allow adding error message for specified field# coding=utf-8 from django.forms import forms class APICall(forms.Form): def __init__(self, api_key=None, *args, **kwargs): super(APICall, self).__init__(*args, **kwargs) self.api_key = api_key def add_error(self, error_msg, field_name=forms.NON_FIELD_ERRORS): # TODO: with Django master you would just raise ValidationError({field_name: error_msg}) self._errors.setdefault(field_name, self.error_class()).append(error_msg) def clean(self): for name, data in self.cleaned_data.iteritems(): setattr(self, name, data) return super(APICall, self).clean() def action(self, test): raise NotImplementedError('APIForms must implement action(self, test)')
<commit_before>from django.forms import forms class APICall(forms.Form): def __init__(self, api_key=None, *args, **kwargs): super(APICall, self).__init__(*args, **kwargs) self.api_key = api_key def add_error(self, error_msg): errors = self.non_field_errors() errors.append(error_msg) self._errors[forms.NON_FIELD_ERRORS] = errors def clean(self): for name, data in self.cleaned_data.iteritems(): setattr(self, name, data) return super(APICall, self).clean() def action(self, test): raise NotImplementedError('APIForms must implement action(self, test)') <commit_msg>Allow adding error message for specified field<commit_after># coding=utf-8 from django.forms import forms class APICall(forms.Form): def __init__(self, api_key=None, *args, **kwargs): super(APICall, self).__init__(*args, **kwargs) self.api_key = api_key def add_error(self, error_msg, field_name=forms.NON_FIELD_ERRORS): # TODO: with Django master you would just raise ValidationError({field_name: error_msg}) self._errors.setdefault(field_name, self.error_class()).append(error_msg) def clean(self): for name, data in self.cleaned_data.iteritems(): setattr(self, name, data) return super(APICall, self).clean() def action(self, test): raise NotImplementedError('APIForms must implement action(self, test)')
5b600e32a05d041facd64db79ea91e928d37f300
tests/test_postgres_processor.py
tests/test_postgres_processor.py
import pytest # from sqlalchemy import create_engine # from sqlalchemy.orm import sessionmaker from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document from . import utils test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.postgres def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID']
import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from . import utils from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.postgres def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID']
Use newly configured conftest in postgres processor test
Use newly configured conftest in postgres processor test
Python
apache-2.0
fabianvf/scrapi,felliott/scrapi,mehanig/scrapi,erinspace/scrapi,erinspace/scrapi,CenterForOpenScience/scrapi,CenterForOpenScience/scrapi,mehanig/scrapi,fabianvf/scrapi,felliott/scrapi
import pytest # from sqlalchemy import create_engine # from sqlalchemy.orm import sessionmaker from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document from . import utils test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.postgres def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID'] Use newly configured conftest in postgres processor test
import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from . import utils from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.postgres def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID']
<commit_before>import pytest # from sqlalchemy import create_engine # from sqlalchemy.orm import sessionmaker from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document from . import utils test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.postgres def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID'] <commit_msg>Use newly configured conftest in postgres processor test<commit_after>
import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from . import utils from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.postgres def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID']
import pytest # from sqlalchemy import create_engine # from sqlalchemy.orm import sessionmaker from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document from . import utils test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.postgres def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID'] Use newly configured conftest in postgres processor testimport pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from . import utils from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.postgres def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID']
<commit_before>import pytest # from sqlalchemy import create_engine # from sqlalchemy.orm import sessionmaker from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document from . import utils test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.postgres def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID'] <commit_msg>Use newly configured conftest in postgres processor test<commit_after>import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from . import utils from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.postgres def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID']
73154aa2498b6826612fae3b287c528e3406bec4
jacquard/service/commands.py
jacquard/service/commands.py
"""Command-line utilities for HTTP service subsystem.""" import werkzeug.debug import werkzeug.serving from jacquard.commands import BaseCommand from jacquard.service import get_wsgi_app class RunServer(BaseCommand): """ Run a debug server. **This is for debug, local use only, not production.** This command is named to mirror its equivalent in Django. It configures the WSGI app and serves it through Werkzeug's simple serving mechanism, with a debugger attached, and auto-reloading. """ help = "run a (local, debug) server" def add_arguments(self, parser): """Add argparse arguments.""" parser.add_argument( '-p', '--port', type=int, default=1212, help="port to bind to", ) parser.add_argument( '-b', '--bind', type=str, default='::1', help="address to bind to", ) def handle(self, config, options): """Run command.""" app = get_wsgi_app(config) werkzeug.serving.run_simple( options.bind, options.port, app, use_reloader=True, use_debugger=True, use_evalex=True, threaded=False, processes=1, )
"""Command-line utilities for HTTP service subsystem.""" import werkzeug.debug import werkzeug.serving from jacquard.commands import BaseCommand from jacquard.service import get_wsgi_app class RunServer(BaseCommand): """ Run a debug server. **This is for debug, local use only, not production.** This command is named to mirror its equivalent in Django. It configures the WSGI app and serves it through Werkzeug's simple serving mechanism, with a debugger attached, and auto-reloading. """ plumbing = True help = "run a (local, debug) server" def add_arguments(self, parser): """Add argparse arguments.""" parser.add_argument( '-p', '--port', type=int, default=1212, help="port to bind to", ) parser.add_argument( '-b', '--bind', type=str, default='::1', help="address to bind to", ) def handle(self, config, options): """Run command.""" app = get_wsgi_app(config) werkzeug.serving.run_simple( options.bind, options.port, app, use_reloader=True, use_debugger=True, use_evalex=True, threaded=False, processes=1, )
Mark runserver as a plumbing command
Mark runserver as a plumbing command
Python
mit
prophile/jacquard,prophile/jacquard
"""Command-line utilities for HTTP service subsystem.""" import werkzeug.debug import werkzeug.serving from jacquard.commands import BaseCommand from jacquard.service import get_wsgi_app class RunServer(BaseCommand): """ Run a debug server. **This is for debug, local use only, not production.** This command is named to mirror its equivalent in Django. It configures the WSGI app and serves it through Werkzeug's simple serving mechanism, with a debugger attached, and auto-reloading. """ help = "run a (local, debug) server" def add_arguments(self, parser): """Add argparse arguments.""" parser.add_argument( '-p', '--port', type=int, default=1212, help="port to bind to", ) parser.add_argument( '-b', '--bind', type=str, default='::1', help="address to bind to", ) def handle(self, config, options): """Run command.""" app = get_wsgi_app(config) werkzeug.serving.run_simple( options.bind, options.port, app, use_reloader=True, use_debugger=True, use_evalex=True, threaded=False, processes=1, ) Mark runserver as a plumbing command
"""Command-line utilities for HTTP service subsystem.""" import werkzeug.debug import werkzeug.serving from jacquard.commands import BaseCommand from jacquard.service import get_wsgi_app class RunServer(BaseCommand): """ Run a debug server. **This is for debug, local use only, not production.** This command is named to mirror its equivalent in Django. It configures the WSGI app and serves it through Werkzeug's simple serving mechanism, with a debugger attached, and auto-reloading. """ plumbing = True help = "run a (local, debug) server" def add_arguments(self, parser): """Add argparse arguments.""" parser.add_argument( '-p', '--port', type=int, default=1212, help="port to bind to", ) parser.add_argument( '-b', '--bind', type=str, default='::1', help="address to bind to", ) def handle(self, config, options): """Run command.""" app = get_wsgi_app(config) werkzeug.serving.run_simple( options.bind, options.port, app, use_reloader=True, use_debugger=True, use_evalex=True, threaded=False, processes=1, )
<commit_before>"""Command-line utilities for HTTP service subsystem.""" import werkzeug.debug import werkzeug.serving from jacquard.commands import BaseCommand from jacquard.service import get_wsgi_app class RunServer(BaseCommand): """ Run a debug server. **This is for debug, local use only, not production.** This command is named to mirror its equivalent in Django. It configures the WSGI app and serves it through Werkzeug's simple serving mechanism, with a debugger attached, and auto-reloading. """ help = "run a (local, debug) server" def add_arguments(self, parser): """Add argparse arguments.""" parser.add_argument( '-p', '--port', type=int, default=1212, help="port to bind to", ) parser.add_argument( '-b', '--bind', type=str, default='::1', help="address to bind to", ) def handle(self, config, options): """Run command.""" app = get_wsgi_app(config) werkzeug.serving.run_simple( options.bind, options.port, app, use_reloader=True, use_debugger=True, use_evalex=True, threaded=False, processes=1, ) <commit_msg>Mark runserver as a plumbing command<commit_after>
"""Command-line utilities for HTTP service subsystem.""" import werkzeug.debug import werkzeug.serving from jacquard.commands import BaseCommand from jacquard.service import get_wsgi_app class RunServer(BaseCommand): """ Run a debug server. **This is for debug, local use only, not production.** This command is named to mirror its equivalent in Django. It configures the WSGI app and serves it through Werkzeug's simple serving mechanism, with a debugger attached, and auto-reloading. """ plumbing = True help = "run a (local, debug) server" def add_arguments(self, parser): """Add argparse arguments.""" parser.add_argument( '-p', '--port', type=int, default=1212, help="port to bind to", ) parser.add_argument( '-b', '--bind', type=str, default='::1', help="address to bind to", ) def handle(self, config, options): """Run command.""" app = get_wsgi_app(config) werkzeug.serving.run_simple( options.bind, options.port, app, use_reloader=True, use_debugger=True, use_evalex=True, threaded=False, processes=1, )
"""Command-line utilities for HTTP service subsystem.""" import werkzeug.debug import werkzeug.serving from jacquard.commands import BaseCommand from jacquard.service import get_wsgi_app class RunServer(BaseCommand): """ Run a debug server. **This is for debug, local use only, not production.** This command is named to mirror its equivalent in Django. It configures the WSGI app and serves it through Werkzeug's simple serving mechanism, with a debugger attached, and auto-reloading. """ help = "run a (local, debug) server" def add_arguments(self, parser): """Add argparse arguments.""" parser.add_argument( '-p', '--port', type=int, default=1212, help="port to bind to", ) parser.add_argument( '-b', '--bind', type=str, default='::1', help="address to bind to", ) def handle(self, config, options): """Run command.""" app = get_wsgi_app(config) werkzeug.serving.run_simple( options.bind, options.port, app, use_reloader=True, use_debugger=True, use_evalex=True, threaded=False, processes=1, ) Mark runserver as a plumbing command"""Command-line utilities for HTTP service subsystem.""" import werkzeug.debug import werkzeug.serving from jacquard.commands import BaseCommand from jacquard.service import get_wsgi_app class RunServer(BaseCommand): """ Run a debug server. **This is for debug, local use only, not production.** This command is named to mirror its equivalent in Django. It configures the WSGI app and serves it through Werkzeug's simple serving mechanism, with a debugger attached, and auto-reloading. """ plumbing = True help = "run a (local, debug) server" def add_arguments(self, parser): """Add argparse arguments.""" parser.add_argument( '-p', '--port', type=int, default=1212, help="port to bind to", ) parser.add_argument( '-b', '--bind', type=str, default='::1', help="address to bind to", ) def handle(self, config, options): """Run command.""" app = get_wsgi_app(config) werkzeug.serving.run_simple( options.bind, options.port, app, use_reloader=True, use_debugger=True, use_evalex=True, threaded=False, processes=1, )
<commit_before>"""Command-line utilities for HTTP service subsystem.""" import werkzeug.debug import werkzeug.serving from jacquard.commands import BaseCommand from jacquard.service import get_wsgi_app class RunServer(BaseCommand): """ Run a debug server. **This is for debug, local use only, not production.** This command is named to mirror its equivalent in Django. It configures the WSGI app and serves it through Werkzeug's simple serving mechanism, with a debugger attached, and auto-reloading. """ help = "run a (local, debug) server" def add_arguments(self, parser): """Add argparse arguments.""" parser.add_argument( '-p', '--port', type=int, default=1212, help="port to bind to", ) parser.add_argument( '-b', '--bind', type=str, default='::1', help="address to bind to", ) def handle(self, config, options): """Run command.""" app = get_wsgi_app(config) werkzeug.serving.run_simple( options.bind, options.port, app, use_reloader=True, use_debugger=True, use_evalex=True, threaded=False, processes=1, ) <commit_msg>Mark runserver as a plumbing command<commit_after>"""Command-line utilities for HTTP service subsystem.""" import werkzeug.debug import werkzeug.serving from jacquard.commands import BaseCommand from jacquard.service import get_wsgi_app class RunServer(BaseCommand): """ Run a debug server. **This is for debug, local use only, not production.** This command is named to mirror its equivalent in Django. It configures the WSGI app and serves it through Werkzeug's simple serving mechanism, with a debugger attached, and auto-reloading. """ plumbing = True help = "run a (local, debug) server" def add_arguments(self, parser): """Add argparse arguments.""" parser.add_argument( '-p', '--port', type=int, default=1212, help="port to bind to", ) parser.add_argument( '-b', '--bind', type=str, default='::1', help="address to bind to", ) def handle(self, config, options): """Run command.""" app = get_wsgi_app(config) werkzeug.serving.run_simple( options.bind, options.port, app, use_reloader=True, use_debugger=True, use_evalex=True, threaded=False, processes=1, )
3e58c3707b8451bb053b2465a6a68438219fd348
python/script_device_PIR_hallway.py
python/script_device_PIR_hallway.py
#!/usr/bin/python # -*- coding: utf-8 -*- import domoticz try: execfile("/etc/domoticz/scripts.conf") except: exec(open("/etc/domoticz/scripts.conf").read()) debug = True if changed_device.name == pir: if debug: domoticz.log("Start " + pir) dev = domoticz.devices[atSleep] dev.off() if debug: domoticz.log("End " + pir)
#!/usr/bin/python # -*- coding: utf-8 -*- import domoticz try: execfile("/etc/domoticz/scripts.conf") except: exec(open("/etc/domoticz/scripts.conf").read()) debug = True if changed_device.name == pir and changed_device.is_on(): if debug: domoticz.log("Start " + pir) dev = domoticz.devices[atSleep] dev.off() if debug: domoticz.log("End " + pir)
Make pir in hallway only trigger at on events
Make pir in hallway only trigger at on events
Python
mit
tomhur/domoticz-scripts,tomhur/domoticz-scripts
#!/usr/bin/python # -*- coding: utf-8 -*- import domoticz try: execfile("/etc/domoticz/scripts.conf") except: exec(open("/etc/domoticz/scripts.conf").read()) debug = True if changed_device.name == pir: if debug: domoticz.log("Start " + pir) dev = domoticz.devices[atSleep] dev.off() if debug: domoticz.log("End " + pir) Make pir in hallway only trigger at on events
#!/usr/bin/python # -*- coding: utf-8 -*- import domoticz try: execfile("/etc/domoticz/scripts.conf") except: exec(open("/etc/domoticz/scripts.conf").read()) debug = True if changed_device.name == pir and changed_device.is_on(): if debug: domoticz.log("Start " + pir) dev = domoticz.devices[atSleep] dev.off() if debug: domoticz.log("End " + pir)
<commit_before>#!/usr/bin/python # -*- coding: utf-8 -*- import domoticz try: execfile("/etc/domoticz/scripts.conf") except: exec(open("/etc/domoticz/scripts.conf").read()) debug = True if changed_device.name == pir: if debug: domoticz.log("Start " + pir) dev = domoticz.devices[atSleep] dev.off() if debug: domoticz.log("End " + pir) <commit_msg>Make pir in hallway only trigger at on events<commit_after>
#!/usr/bin/python # -*- coding: utf-8 -*- import domoticz try: execfile("/etc/domoticz/scripts.conf") except: exec(open("/etc/domoticz/scripts.conf").read()) debug = True if changed_device.name == pir and changed_device.is_on(): if debug: domoticz.log("Start " + pir) dev = domoticz.devices[atSleep] dev.off() if debug: domoticz.log("End " + pir)
#!/usr/bin/python # -*- coding: utf-8 -*- import domoticz try: execfile("/etc/domoticz/scripts.conf") except: exec(open("/etc/domoticz/scripts.conf").read()) debug = True if changed_device.name == pir: if debug: domoticz.log("Start " + pir) dev = domoticz.devices[atSleep] dev.off() if debug: domoticz.log("End " + pir) Make pir in hallway only trigger at on events#!/usr/bin/python # -*- coding: utf-8 -*- import domoticz try: execfile("/etc/domoticz/scripts.conf") except: exec(open("/etc/domoticz/scripts.conf").read()) debug = True if changed_device.name == pir and changed_device.is_on(): if debug: domoticz.log("Start " + pir) dev = domoticz.devices[atSleep] dev.off() if debug: domoticz.log("End " + pir)
<commit_before>#!/usr/bin/python # -*- coding: utf-8 -*- import domoticz try: execfile("/etc/domoticz/scripts.conf") except: exec(open("/etc/domoticz/scripts.conf").read()) debug = True if changed_device.name == pir: if debug: domoticz.log("Start " + pir) dev = domoticz.devices[atSleep] dev.off() if debug: domoticz.log("End " + pir) <commit_msg>Make pir in hallway only trigger at on events<commit_after>#!/usr/bin/python # -*- coding: utf-8 -*- import domoticz try: execfile("/etc/domoticz/scripts.conf") except: exec(open("/etc/domoticz/scripts.conf").read()) debug = True if changed_device.name == pir and changed_device.is_on(): if debug: domoticz.log("Start " + pir) dev = domoticz.devices[atSleep] dev.off() if debug: domoticz.log("End " + pir)
70df4ca8235b3ae29ef2843169f9119d29bda44a
tracker/models/__init__.py
tracker/models/__init__.py
from BasicInfo import BasicInfo from BloodExam import BloodExam from BloodExam import BloodParasite from Child import Child from ChildForm import ChildForm from DentalExam import DentalExam from DentalExam import DentalExamDiagnosis from MedicalExamPart1Info import MedicalExamPart1Info from MedicalExamPart2Info import MedicalExamPart2Info from MedicalExamPart2Info import MedicalExamDiagnosis from PsychologicalExamInfo import PsychologicalExamInfo from PsychologicalExamInfo import PsychologicalExamDiagnosis from Residence import Residence from SocialExamInfo import SocialExamInfo
from BasicInfo import BasicInfo from BloodExam import BloodExam from BloodExam import BloodParasite from Child import Child from ChildForm import ChildForm from DentalExam import DentalExam from DentalExam import DentalExamDiagnosis from MedicalExamPart1Info import MedicalExamPart1Info from MedicalExamPart2Info import MedicalExamPart2Info from MedicalExamPart2Info import MedicalExamDiagnosis from PsychologicalExamInfo import PsychologicalExamInfo from PsychologicalExamInfo import PsychologicalExamDiagnosis from Residence import Residence from SocialExamInfo import SocialExamInfo from ConsultationHistory import ConsultationHistory from DiseaseHistory import DiseaseHistory from OperationHistory import OperationHistory
Include history in the models.
Include history in the models.
Python
mit
sscalpone/HBI,sscalpone/HBI,sscalpone/HBI,sscalpone/HBI
from BasicInfo import BasicInfo from BloodExam import BloodExam from BloodExam import BloodParasite from Child import Child from ChildForm import ChildForm from DentalExam import DentalExam from DentalExam import DentalExamDiagnosis from MedicalExamPart1Info import MedicalExamPart1Info from MedicalExamPart2Info import MedicalExamPart2Info from MedicalExamPart2Info import MedicalExamDiagnosis from PsychologicalExamInfo import PsychologicalExamInfo from PsychologicalExamInfo import PsychologicalExamDiagnosis from Residence import Residence from SocialExamInfo import SocialExamInfo Include history in the models.
from BasicInfo import BasicInfo from BloodExam import BloodExam from BloodExam import BloodParasite from Child import Child from ChildForm import ChildForm from DentalExam import DentalExam from DentalExam import DentalExamDiagnosis from MedicalExamPart1Info import MedicalExamPart1Info from MedicalExamPart2Info import MedicalExamPart2Info from MedicalExamPart2Info import MedicalExamDiagnosis from PsychologicalExamInfo import PsychologicalExamInfo from PsychologicalExamInfo import PsychologicalExamDiagnosis from Residence import Residence from SocialExamInfo import SocialExamInfo from ConsultationHistory import ConsultationHistory from DiseaseHistory import DiseaseHistory from OperationHistory import OperationHistory
<commit_before>from BasicInfo import BasicInfo from BloodExam import BloodExam from BloodExam import BloodParasite from Child import Child from ChildForm import ChildForm from DentalExam import DentalExam from DentalExam import DentalExamDiagnosis from MedicalExamPart1Info import MedicalExamPart1Info from MedicalExamPart2Info import MedicalExamPart2Info from MedicalExamPart2Info import MedicalExamDiagnosis from PsychologicalExamInfo import PsychologicalExamInfo from PsychologicalExamInfo import PsychologicalExamDiagnosis from Residence import Residence from SocialExamInfo import SocialExamInfo <commit_msg>Include history in the models.<commit_after>
from BasicInfo import BasicInfo from BloodExam import BloodExam from BloodExam import BloodParasite from Child import Child from ChildForm import ChildForm from DentalExam import DentalExam from DentalExam import DentalExamDiagnosis from MedicalExamPart1Info import MedicalExamPart1Info from MedicalExamPart2Info import MedicalExamPart2Info from MedicalExamPart2Info import MedicalExamDiagnosis from PsychologicalExamInfo import PsychologicalExamInfo from PsychologicalExamInfo import PsychologicalExamDiagnosis from Residence import Residence from SocialExamInfo import SocialExamInfo from ConsultationHistory import ConsultationHistory from DiseaseHistory import DiseaseHistory from OperationHistory import OperationHistory
from BasicInfo import BasicInfo from BloodExam import BloodExam from BloodExam import BloodParasite from Child import Child from ChildForm import ChildForm from DentalExam import DentalExam from DentalExam import DentalExamDiagnosis from MedicalExamPart1Info import MedicalExamPart1Info from MedicalExamPart2Info import MedicalExamPart2Info from MedicalExamPart2Info import MedicalExamDiagnosis from PsychologicalExamInfo import PsychologicalExamInfo from PsychologicalExamInfo import PsychologicalExamDiagnosis from Residence import Residence from SocialExamInfo import SocialExamInfo Include history in the models.from BasicInfo import BasicInfo from BloodExam import BloodExam from BloodExam import BloodParasite from Child import Child from ChildForm import ChildForm from DentalExam import DentalExam from DentalExam import DentalExamDiagnosis from MedicalExamPart1Info import MedicalExamPart1Info from MedicalExamPart2Info import MedicalExamPart2Info from MedicalExamPart2Info import MedicalExamDiagnosis from PsychologicalExamInfo import PsychologicalExamInfo from PsychologicalExamInfo import PsychologicalExamDiagnosis from Residence import Residence from SocialExamInfo import SocialExamInfo from ConsultationHistory import ConsultationHistory from DiseaseHistory import DiseaseHistory from OperationHistory import OperationHistory
<commit_before>from BasicInfo import BasicInfo from BloodExam import BloodExam from BloodExam import BloodParasite from Child import Child from ChildForm import ChildForm from DentalExam import DentalExam from DentalExam import DentalExamDiagnosis from MedicalExamPart1Info import MedicalExamPart1Info from MedicalExamPart2Info import MedicalExamPart2Info from MedicalExamPart2Info import MedicalExamDiagnosis from PsychologicalExamInfo import PsychologicalExamInfo from PsychologicalExamInfo import PsychologicalExamDiagnosis from Residence import Residence from SocialExamInfo import SocialExamInfo <commit_msg>Include history in the models.<commit_after>from BasicInfo import BasicInfo from BloodExam import BloodExam from BloodExam import BloodParasite from Child import Child from ChildForm import ChildForm from DentalExam import DentalExam from DentalExam import DentalExamDiagnosis from MedicalExamPart1Info import MedicalExamPart1Info from MedicalExamPart2Info import MedicalExamPart2Info from MedicalExamPart2Info import MedicalExamDiagnosis from PsychologicalExamInfo import PsychologicalExamInfo from PsychologicalExamInfo import PsychologicalExamDiagnosis from Residence import Residence from SocialExamInfo import SocialExamInfo from ConsultationHistory import ConsultationHistory from DiseaseHistory import DiseaseHistory from OperationHistory import OperationHistory
5368e0ad7be4cdf7df2da392fdaabb89c3a4ad55
test_settings.py
test_settings.py
SECRET_KEY = "lorem ipsum" INSTALLED_APPS = ( 'tango_shared', )
SECRET_KEY = "lorem ipsum" INSTALLED_APPS = ( 'tango_shared', ) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': ':memory:', } } SITE_ID = 1
Add missing test settings (in-memory sqlite3 db + SITE_ID)
Add missing test settings (in-memory sqlite3 db + SITE_ID)
Python
mit
tBaxter/tango-shared-core,tBaxter/tango-shared-core,tBaxter/tango-shared-core
SECRET_KEY = "lorem ipsum" INSTALLED_APPS = ( 'tango_shared', ) Add missing test settings (in-memory sqlite3 db + SITE_ID)
SECRET_KEY = "lorem ipsum" INSTALLED_APPS = ( 'tango_shared', ) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': ':memory:', } } SITE_ID = 1
<commit_before>SECRET_KEY = "lorem ipsum" INSTALLED_APPS = ( 'tango_shared', ) <commit_msg>Add missing test settings (in-memory sqlite3 db + SITE_ID)<commit_after>
SECRET_KEY = "lorem ipsum" INSTALLED_APPS = ( 'tango_shared', ) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': ':memory:', } } SITE_ID = 1
SECRET_KEY = "lorem ipsum" INSTALLED_APPS = ( 'tango_shared', ) Add missing test settings (in-memory sqlite3 db + SITE_ID)SECRET_KEY = "lorem ipsum" INSTALLED_APPS = ( 'tango_shared', ) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': ':memory:', } } SITE_ID = 1
<commit_before>SECRET_KEY = "lorem ipsum" INSTALLED_APPS = ( 'tango_shared', ) <commit_msg>Add missing test settings (in-memory sqlite3 db + SITE_ID)<commit_after>SECRET_KEY = "lorem ipsum" INSTALLED_APPS = ( 'tango_shared', ) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': ':memory:', } } SITE_ID = 1
f629e0ea3dc06e91d6d666b1a8bfefdd925287ff
rafem/__init__.py
rafem/__init__.py
"""River Avulsion Module.""" from .riverbmi import BmiRiverModule from .rivermodule import RiverModule __all__ = ['BmiRiverModule', 'RiverModule']
"""River Avulsion Module.""" from .riverbmi import BmiRiverModule from .rivermodule import rivermodule __all__ = ['BmiRiverModule', 'rivermodule']
Rename package from avulsion to rafem.
Rename package from avulsion to rafem.
Python
mit
katmratliff/avulsion-bmi,mcflugen/avulsion-bmi
"""River Avulsion Module.""" from .riverbmi import BmiRiverModule from .rivermodule import RiverModule __all__ = ['BmiRiverModule', 'RiverModule'] Rename package from avulsion to rafem.
"""River Avulsion Module.""" from .riverbmi import BmiRiverModule from .rivermodule import rivermodule __all__ = ['BmiRiverModule', 'rivermodule']
<commit_before>"""River Avulsion Module.""" from .riverbmi import BmiRiverModule from .rivermodule import RiverModule __all__ = ['BmiRiverModule', 'RiverModule'] <commit_msg>Rename package from avulsion to rafem.<commit_after>
"""River Avulsion Module.""" from .riverbmi import BmiRiverModule from .rivermodule import rivermodule __all__ = ['BmiRiverModule', 'rivermodule']
"""River Avulsion Module.""" from .riverbmi import BmiRiverModule from .rivermodule import RiverModule __all__ = ['BmiRiverModule', 'RiverModule'] Rename package from avulsion to rafem."""River Avulsion Module.""" from .riverbmi import BmiRiverModule from .rivermodule import rivermodule __all__ = ['BmiRiverModule', 'rivermodule']
<commit_before>"""River Avulsion Module.""" from .riverbmi import BmiRiverModule from .rivermodule import RiverModule __all__ = ['BmiRiverModule', 'RiverModule'] <commit_msg>Rename package from avulsion to rafem.<commit_after>"""River Avulsion Module.""" from .riverbmi import BmiRiverModule from .rivermodule import rivermodule __all__ = ['BmiRiverModule', 'rivermodule']
46956660b1c4533e7a69fe2aa0dc17b73ce490ac
transporter/transporter.py
transporter/transporter.py
from urlparse import urlparse import os import adapters try: import paramiko except ImportError: pass """The following protocals are supported ftp, ftps, http and https. sftp and ssh require paramiko to be installed """ class Transporter(object): availible_adapters = { "ftp": adapters.FtpAdapter, "ftps": adapters.FtpAdapter, "file": adapters.LocalFileAdapter } adapter = None def __init__(self, uri): uri = urlparse(uri) if uri.scheme not in self.availible_adapters: msg = u"{0} is not a support scheme. Availible schemes: {1}".format( uri.scheme, [s for s in self.availible_adapters]) raise NotImplemented(msg) self.adapter = self.availible_adapters[uri.scheme](uri) def __getattr__(self, attr): return getattr(self.adapter, attr) def __repr__(self): return u'<Transporter {0}>'.format(self.adapter.__repr__()) def download(uri): f = os.path.basename(uri) uri = os.path.dirname(uri) uri = urlparse(uri) return Transporter(uri).download(f) def upload(f, uri): return Transporter(uri).upload(f) def transport(source, destination): f = download(source) return upload(destination, f)
from urlparse import urlparse import os import adapters try: import paramiko except ImportError: pass """The following protocals are supported ftp, ftps, http and https. sftp and ssh require paramiko to be installed """ class Transporter(object): availible_adapters = { "ftp": adapters.FtpAdapter, "ftps": adapters.FtpAdapter, "file": adapters.LocalFileAdapter } default_scheme = "file" adapter = None def __init__(self, uri): uri = urlparse(uri) scheme = uri.scheme or self.default_scheme if scheme not in self.availible_adapters: msg = u"{0} is not a support scheme. Availible schemes: {1}".format( scheme, [s for s in self.availible_adapters]) raise NotImplemented(msg) self.adapter = self.availible_adapters[scheme](uri) def __getattr__(self, attr): return getattr(self.adapter, attr) def __repr__(self): return u'<Transporter {0}>'.format(self.adapter.__repr__()) def download(uri): f = os.path.basename(uri) uri = os.path.dirname(uri) uri = urlparse(uri) return Transporter(uri).download(f) def upload(f, uri): return Transporter(uri).upload(f) def transport(source, destination): f = download(source) return upload(destination, f)
Use LocalFileAdapter when no scheme is given
Use LocalFileAdapter when no scheme is given >>> t1 = Transporter('/file/path') >>> t2 = Transporter('file:/file/path') >>> type(t1.adapter) == type(t2.adapter) True >>> t1.pwd() == t2.pwd() True
Python
bsd-2-clause
joshmaker/transporter
from urlparse import urlparse import os import adapters try: import paramiko except ImportError: pass """The following protocals are supported ftp, ftps, http and https. sftp and ssh require paramiko to be installed """ class Transporter(object): availible_adapters = { "ftp": adapters.FtpAdapter, "ftps": adapters.FtpAdapter, "file": adapters.LocalFileAdapter } adapter = None def __init__(self, uri): uri = urlparse(uri) if uri.scheme not in self.availible_adapters: msg = u"{0} is not a support scheme. Availible schemes: {1}".format( uri.scheme, [s for s in self.availible_adapters]) raise NotImplemented(msg) self.adapter = self.availible_adapters[uri.scheme](uri) def __getattr__(self, attr): return getattr(self.adapter, attr) def __repr__(self): return u'<Transporter {0}>'.format(self.adapter.__repr__()) def download(uri): f = os.path.basename(uri) uri = os.path.dirname(uri) uri = urlparse(uri) return Transporter(uri).download(f) def upload(f, uri): return Transporter(uri).upload(f) def transport(source, destination): f = download(source) return upload(destination, f) Use LocalFileAdapter when no scheme is given >>> t1 = Transporter('/file/path') >>> t2 = Transporter('file:/file/path') >>> type(t1.adapter) == type(t2.adapter) True >>> t1.pwd() == t2.pwd() True
from urlparse import urlparse import os import adapters try: import paramiko except ImportError: pass """The following protocals are supported ftp, ftps, http and https. sftp and ssh require paramiko to be installed """ class Transporter(object): availible_adapters = { "ftp": adapters.FtpAdapter, "ftps": adapters.FtpAdapter, "file": adapters.LocalFileAdapter } default_scheme = "file" adapter = None def __init__(self, uri): uri = urlparse(uri) scheme = uri.scheme or self.default_scheme if scheme not in self.availible_adapters: msg = u"{0} is not a support scheme. Availible schemes: {1}".format( scheme, [s for s in self.availible_adapters]) raise NotImplemented(msg) self.adapter = self.availible_adapters[scheme](uri) def __getattr__(self, attr): return getattr(self.adapter, attr) def __repr__(self): return u'<Transporter {0}>'.format(self.adapter.__repr__()) def download(uri): f = os.path.basename(uri) uri = os.path.dirname(uri) uri = urlparse(uri) return Transporter(uri).download(f) def upload(f, uri): return Transporter(uri).upload(f) def transport(source, destination): f = download(source) return upload(destination, f)
<commit_before>from urlparse import urlparse import os import adapters try: import paramiko except ImportError: pass """The following protocals are supported ftp, ftps, http and https. sftp and ssh require paramiko to be installed """ class Transporter(object): availible_adapters = { "ftp": adapters.FtpAdapter, "ftps": adapters.FtpAdapter, "file": adapters.LocalFileAdapter } adapter = None def __init__(self, uri): uri = urlparse(uri) if uri.scheme not in self.availible_adapters: msg = u"{0} is not a support scheme. Availible schemes: {1}".format( uri.scheme, [s for s in self.availible_adapters]) raise NotImplemented(msg) self.adapter = self.availible_adapters[uri.scheme](uri) def __getattr__(self, attr): return getattr(self.adapter, attr) def __repr__(self): return u'<Transporter {0}>'.format(self.adapter.__repr__()) def download(uri): f = os.path.basename(uri) uri = os.path.dirname(uri) uri = urlparse(uri) return Transporter(uri).download(f) def upload(f, uri): return Transporter(uri).upload(f) def transport(source, destination): f = download(source) return upload(destination, f) <commit_msg>Use LocalFileAdapter when no scheme is given >>> t1 = Transporter('/file/path') >>> t2 = Transporter('file:/file/path') >>> type(t1.adapter) == type(t2.adapter) True >>> t1.pwd() == t2.pwd() True<commit_after>
from urlparse import urlparse import os import adapters try: import paramiko except ImportError: pass """The following protocals are supported ftp, ftps, http and https. sftp and ssh require paramiko to be installed """ class Transporter(object): availible_adapters = { "ftp": adapters.FtpAdapter, "ftps": adapters.FtpAdapter, "file": adapters.LocalFileAdapter } default_scheme = "file" adapter = None def __init__(self, uri): uri = urlparse(uri) scheme = uri.scheme or self.default_scheme if scheme not in self.availible_adapters: msg = u"{0} is not a support scheme. Availible schemes: {1}".format( scheme, [s for s in self.availible_adapters]) raise NotImplemented(msg) self.adapter = self.availible_adapters[scheme](uri) def __getattr__(self, attr): return getattr(self.adapter, attr) def __repr__(self): return u'<Transporter {0}>'.format(self.adapter.__repr__()) def download(uri): f = os.path.basename(uri) uri = os.path.dirname(uri) uri = urlparse(uri) return Transporter(uri).download(f) def upload(f, uri): return Transporter(uri).upload(f) def transport(source, destination): f = download(source) return upload(destination, f)
from urlparse import urlparse import os import adapters try: import paramiko except ImportError: pass """The following protocals are supported ftp, ftps, http and https. sftp and ssh require paramiko to be installed """ class Transporter(object): availible_adapters = { "ftp": adapters.FtpAdapter, "ftps": adapters.FtpAdapter, "file": adapters.LocalFileAdapter } adapter = None def __init__(self, uri): uri = urlparse(uri) if uri.scheme not in self.availible_adapters: msg = u"{0} is not a support scheme. Availible schemes: {1}".format( uri.scheme, [s for s in self.availible_adapters]) raise NotImplemented(msg) self.adapter = self.availible_adapters[uri.scheme](uri) def __getattr__(self, attr): return getattr(self.adapter, attr) def __repr__(self): return u'<Transporter {0}>'.format(self.adapter.__repr__()) def download(uri): f = os.path.basename(uri) uri = os.path.dirname(uri) uri = urlparse(uri) return Transporter(uri).download(f) def upload(f, uri): return Transporter(uri).upload(f) def transport(source, destination): f = download(source) return upload(destination, f) Use LocalFileAdapter when no scheme is given >>> t1 = Transporter('/file/path') >>> t2 = Transporter('file:/file/path') >>> type(t1.adapter) == type(t2.adapter) True >>> t1.pwd() == t2.pwd() Truefrom urlparse import urlparse import os import adapters try: import paramiko except ImportError: pass """The following protocals are supported ftp, ftps, http and https. sftp and ssh require paramiko to be installed """ class Transporter(object): availible_adapters = { "ftp": adapters.FtpAdapter, "ftps": adapters.FtpAdapter, "file": adapters.LocalFileAdapter } default_scheme = "file" adapter = None def __init__(self, uri): uri = urlparse(uri) scheme = uri.scheme or self.default_scheme if scheme not in self.availible_adapters: msg = u"{0} is not a support scheme. Availible schemes: {1}".format( scheme, [s for s in self.availible_adapters]) raise NotImplemented(msg) self.adapter = self.availible_adapters[scheme](uri) def __getattr__(self, attr): return getattr(self.adapter, attr) def __repr__(self): return u'<Transporter {0}>'.format(self.adapter.__repr__()) def download(uri): f = os.path.basename(uri) uri = os.path.dirname(uri) uri = urlparse(uri) return Transporter(uri).download(f) def upload(f, uri): return Transporter(uri).upload(f) def transport(source, destination): f = download(source) return upload(destination, f)
<commit_before>from urlparse import urlparse import os import adapters try: import paramiko except ImportError: pass """The following protocals are supported ftp, ftps, http and https. sftp and ssh require paramiko to be installed """ class Transporter(object): availible_adapters = { "ftp": adapters.FtpAdapter, "ftps": adapters.FtpAdapter, "file": adapters.LocalFileAdapter } adapter = None def __init__(self, uri): uri = urlparse(uri) if uri.scheme not in self.availible_adapters: msg = u"{0} is not a support scheme. Availible schemes: {1}".format( uri.scheme, [s for s in self.availible_adapters]) raise NotImplemented(msg) self.adapter = self.availible_adapters[uri.scheme](uri) def __getattr__(self, attr): return getattr(self.adapter, attr) def __repr__(self): return u'<Transporter {0}>'.format(self.adapter.__repr__()) def download(uri): f = os.path.basename(uri) uri = os.path.dirname(uri) uri = urlparse(uri) return Transporter(uri).download(f) def upload(f, uri): return Transporter(uri).upload(f) def transport(source, destination): f = download(source) return upload(destination, f) <commit_msg>Use LocalFileAdapter when no scheme is given >>> t1 = Transporter('/file/path') >>> t2 = Transporter('file:/file/path') >>> type(t1.adapter) == type(t2.adapter) True >>> t1.pwd() == t2.pwd() True<commit_after>from urlparse import urlparse import os import adapters try: import paramiko except ImportError: pass """The following protocals are supported ftp, ftps, http and https. sftp and ssh require paramiko to be installed """ class Transporter(object): availible_adapters = { "ftp": adapters.FtpAdapter, "ftps": adapters.FtpAdapter, "file": adapters.LocalFileAdapter } default_scheme = "file" adapter = None def __init__(self, uri): uri = urlparse(uri) scheme = uri.scheme or self.default_scheme if scheme not in self.availible_adapters: msg = u"{0} is not a support scheme. Availible schemes: {1}".format( scheme, [s for s in self.availible_adapters]) raise NotImplemented(msg) self.adapter = self.availible_adapters[scheme](uri) def __getattr__(self, attr): return getattr(self.adapter, attr) def __repr__(self): return u'<Transporter {0}>'.format(self.adapter.__repr__()) def download(uri): f = os.path.basename(uri) uri = os.path.dirname(uri) uri = urlparse(uri) return Transporter(uri).download(f) def upload(f, uri): return Transporter(uri).upload(f) def transport(source, destination): f = download(source) return upload(destination, f)
a62343a536bf6a8b655ace66e09a17ea483e6fbe
txircd/modules/cmd_user.py
txircd/modules/cmd_user.py
from twisted.words.protocols import irc from txircd.modbase import Command import string class UserCommand(Command): def onUse(self, user, data): if not user.username: user.registered -= 1 user.username = data["ident"] user.realname = data["gecos"] if user.registered == 0: user.register() def processParams(self, user, params): if user.registered == 0: user.sendMessage(irc.ERR_ALREADYREGISTRED, ":Unauthorized command (already registered)") return {} if params and len(params) < 4: user.sendMessage(irc.ERR_NEEDMOREPARAMS, "USER", ":Not enough parameters") return {} ident = filter(lambda x: x in string.ascii_letters + string.digits + "-_", params[0][:12]) if not ident: user.sendMessage(irc.ERR_NEEDMOREPARAMS, "USER", ":Your username is not valid") return {} return { "user": user, "ident": ident, "gecos": params[3] } class Spawner(object): def __init__(self, ircd): self.ircd = ircd def spawn(self): return { "commands": { "USER": UserCommand() } } def cleanup(self): del self.ircd.commands["USER"]
from twisted.words.protocols import irc from txircd.modbase import Command import string class UserCommand(Command): def onUse(self, user, data): if not user.username: user.registered -= 1 user.username = data["ident"] user.realname = data["gecos"] if user.registered == 0: user.register() def processParams(self, user, params): if user.registered == 0: user.sendMessage(irc.ERR_ALREADYREGISTRED, ":You may not reregister") return {} if params and len(params) < 4: user.sendMessage(irc.ERR_NEEDMOREPARAMS, "USER", ":Not enough parameters") return {} ident = filter(lambda x: x in string.ascii_letters + string.digits + "-_", params[0][:12]) if not ident: user.sendMessage(irc.ERR_NEEDMOREPARAMS, "USER", ":Your username is not valid") return {} return { "user": user, "ident": ident, "gecos": params[3] } class Spawner(object): def __init__(self, ircd): self.ircd = ircd def spawn(self): return { "commands": { "USER": UserCommand() } } def cleanup(self): del self.ircd.commands["USER"]
Fix message with 462 numeric
Fix message with 462 numeric
Python
bsd-3-clause
ElementalAlchemist/txircd,DesertBus/txircd,Heufneutje/txircd
from twisted.words.protocols import irc from txircd.modbase import Command import string class UserCommand(Command): def onUse(self, user, data): if not user.username: user.registered -= 1 user.username = data["ident"] user.realname = data["gecos"] if user.registered == 0: user.register() def processParams(self, user, params): if user.registered == 0: user.sendMessage(irc.ERR_ALREADYREGISTRED, ":Unauthorized command (already registered)") return {} if params and len(params) < 4: user.sendMessage(irc.ERR_NEEDMOREPARAMS, "USER", ":Not enough parameters") return {} ident = filter(lambda x: x in string.ascii_letters + string.digits + "-_", params[0][:12]) if not ident: user.sendMessage(irc.ERR_NEEDMOREPARAMS, "USER", ":Your username is not valid") return {} return { "user": user, "ident": ident, "gecos": params[3] } class Spawner(object): def __init__(self, ircd): self.ircd = ircd def spawn(self): return { "commands": { "USER": UserCommand() } } def cleanup(self): del self.ircd.commands["USER"]Fix message with 462 numeric
from twisted.words.protocols import irc from txircd.modbase import Command import string class UserCommand(Command): def onUse(self, user, data): if not user.username: user.registered -= 1 user.username = data["ident"] user.realname = data["gecos"] if user.registered == 0: user.register() def processParams(self, user, params): if user.registered == 0: user.sendMessage(irc.ERR_ALREADYREGISTRED, ":You may not reregister") return {} if params and len(params) < 4: user.sendMessage(irc.ERR_NEEDMOREPARAMS, "USER", ":Not enough parameters") return {} ident = filter(lambda x: x in string.ascii_letters + string.digits + "-_", params[0][:12]) if not ident: user.sendMessage(irc.ERR_NEEDMOREPARAMS, "USER", ":Your username is not valid") return {} return { "user": user, "ident": ident, "gecos": params[3] } class Spawner(object): def __init__(self, ircd): self.ircd = ircd def spawn(self): return { "commands": { "USER": UserCommand() } } def cleanup(self): del self.ircd.commands["USER"]
<commit_before>from twisted.words.protocols import irc from txircd.modbase import Command import string class UserCommand(Command): def onUse(self, user, data): if not user.username: user.registered -= 1 user.username = data["ident"] user.realname = data["gecos"] if user.registered == 0: user.register() def processParams(self, user, params): if user.registered == 0: user.sendMessage(irc.ERR_ALREADYREGISTRED, ":Unauthorized command (already registered)") return {} if params and len(params) < 4: user.sendMessage(irc.ERR_NEEDMOREPARAMS, "USER", ":Not enough parameters") return {} ident = filter(lambda x: x in string.ascii_letters + string.digits + "-_", params[0][:12]) if not ident: user.sendMessage(irc.ERR_NEEDMOREPARAMS, "USER", ":Your username is not valid") return {} return { "user": user, "ident": ident, "gecos": params[3] } class Spawner(object): def __init__(self, ircd): self.ircd = ircd def spawn(self): return { "commands": { "USER": UserCommand() } } def cleanup(self): del self.ircd.commands["USER"]<commit_msg>Fix message with 462 numeric<commit_after>
from twisted.words.protocols import irc from txircd.modbase import Command import string class UserCommand(Command): def onUse(self, user, data): if not user.username: user.registered -= 1 user.username = data["ident"] user.realname = data["gecos"] if user.registered == 0: user.register() def processParams(self, user, params): if user.registered == 0: user.sendMessage(irc.ERR_ALREADYREGISTRED, ":You may not reregister") return {} if params and len(params) < 4: user.sendMessage(irc.ERR_NEEDMOREPARAMS, "USER", ":Not enough parameters") return {} ident = filter(lambda x: x in string.ascii_letters + string.digits + "-_", params[0][:12]) if not ident: user.sendMessage(irc.ERR_NEEDMOREPARAMS, "USER", ":Your username is not valid") return {} return { "user": user, "ident": ident, "gecos": params[3] } class Spawner(object): def __init__(self, ircd): self.ircd = ircd def spawn(self): return { "commands": { "USER": UserCommand() } } def cleanup(self): del self.ircd.commands["USER"]
from twisted.words.protocols import irc from txircd.modbase import Command import string class UserCommand(Command): def onUse(self, user, data): if not user.username: user.registered -= 1 user.username = data["ident"] user.realname = data["gecos"] if user.registered == 0: user.register() def processParams(self, user, params): if user.registered == 0: user.sendMessage(irc.ERR_ALREADYREGISTRED, ":Unauthorized command (already registered)") return {} if params and len(params) < 4: user.sendMessage(irc.ERR_NEEDMOREPARAMS, "USER", ":Not enough parameters") return {} ident = filter(lambda x: x in string.ascii_letters + string.digits + "-_", params[0][:12]) if not ident: user.sendMessage(irc.ERR_NEEDMOREPARAMS, "USER", ":Your username is not valid") return {} return { "user": user, "ident": ident, "gecos": params[3] } class Spawner(object): def __init__(self, ircd): self.ircd = ircd def spawn(self): return { "commands": { "USER": UserCommand() } } def cleanup(self): del self.ircd.commands["USER"]Fix message with 462 numericfrom twisted.words.protocols import irc from txircd.modbase import Command import string class UserCommand(Command): def onUse(self, user, data): if not user.username: user.registered -= 1 user.username = data["ident"] user.realname = data["gecos"] if user.registered == 0: user.register() def processParams(self, user, params): if user.registered == 0: user.sendMessage(irc.ERR_ALREADYREGISTRED, ":You may not reregister") return {} if params and len(params) < 4: user.sendMessage(irc.ERR_NEEDMOREPARAMS, "USER", ":Not enough parameters") return {} ident = filter(lambda x: x in string.ascii_letters + string.digits + "-_", params[0][:12]) if not ident: user.sendMessage(irc.ERR_NEEDMOREPARAMS, "USER", ":Your username is not valid") return {} return { "user": user, "ident": ident, "gecos": params[3] } class Spawner(object): def __init__(self, ircd): self.ircd = ircd def spawn(self): return { "commands": { "USER": UserCommand() } } def cleanup(self): del self.ircd.commands["USER"]
<commit_before>from twisted.words.protocols import irc from txircd.modbase import Command import string class UserCommand(Command): def onUse(self, user, data): if not user.username: user.registered -= 1 user.username = data["ident"] user.realname = data["gecos"] if user.registered == 0: user.register() def processParams(self, user, params): if user.registered == 0: user.sendMessage(irc.ERR_ALREADYREGISTRED, ":Unauthorized command (already registered)") return {} if params and len(params) < 4: user.sendMessage(irc.ERR_NEEDMOREPARAMS, "USER", ":Not enough parameters") return {} ident = filter(lambda x: x in string.ascii_letters + string.digits + "-_", params[0][:12]) if not ident: user.sendMessage(irc.ERR_NEEDMOREPARAMS, "USER", ":Your username is not valid") return {} return { "user": user, "ident": ident, "gecos": params[3] } class Spawner(object): def __init__(self, ircd): self.ircd = ircd def spawn(self): return { "commands": { "USER": UserCommand() } } def cleanup(self): del self.ircd.commands["USER"]<commit_msg>Fix message with 462 numeric<commit_after>from twisted.words.protocols import irc from txircd.modbase import Command import string class UserCommand(Command): def onUse(self, user, data): if not user.username: user.registered -= 1 user.username = data["ident"] user.realname = data["gecos"] if user.registered == 0: user.register() def processParams(self, user, params): if user.registered == 0: user.sendMessage(irc.ERR_ALREADYREGISTRED, ":You may not reregister") return {} if params and len(params) < 4: user.sendMessage(irc.ERR_NEEDMOREPARAMS, "USER", ":Not enough parameters") return {} ident = filter(lambda x: x in string.ascii_letters + string.digits + "-_", params[0][:12]) if not ident: user.sendMessage(irc.ERR_NEEDMOREPARAMS, "USER", ":Your username is not valid") return {} return { "user": user, "ident": ident, "gecos": params[3] } class Spawner(object): def __init__(self, ircd): self.ircd = ircd def spawn(self): return { "commands": { "USER": UserCommand() } } def cleanup(self): del self.ircd.commands["USER"]
dfdd9b2a8bad5d61f5bf166d9c72b19f4c639383
mock/buildbot_secret.py
mock/buildbot_secret.py
GITHUB_WEBHOOK_SECRET="nothing to see here" GITHUB_OAUTH_CLIENT_ID="nothing to see here" GITHUB_OAUTH_CLIENT_SECRET="nothing to see here" GITHUB_STATUS_OAUTH_TOKEN="nothing to see here" COVERALLS_REPO_TOKEN="nothing to see here" CODECOV_REPO_TOKEN="nothing to see here" FREEBSDCI_OAUTH_TOKEN="nothing to see here" FQDN="buildog.julialang.org" BUILDBOT_BRANCH="master" db_user="nothing to see here" db_password="nothing to see here" DOCUMENTER_KEY="nothing to see here"
GITHUB_WEBHOOK_SECRET="nothing to see here" GITHUB_OAUTH_CLIENT_ID="nothing to see here" GITHUB_OAUTH_CLIENT_SECRET="nothing to see here" GITHUB_STATUS_OAUTH_TOKEN="nothing to see here" COVERALLS_REPO_TOKEN="nothing to see here" CODECOV_REPO_TOKEN="nothing to see here" FREEBSDCI_OAUTH_TOKEN="nothing to see here" FQDN="buildog.julialang.org" BUILDBOT_BRANCH="master" db_user="nothing to see here" db_password="nothing to see here" DOCUMENTER_KEY="nothing to see here" MACOS_CODESIGN_IDENTITY="nothing to see here"
Add mock value for `MACOS_CODESIGN_IDENTITY`
Add mock value for `MACOS_CODESIGN_IDENTITY`
Python
mit
staticfloat/julia-buildbot,staticfloat/julia-buildbot
GITHUB_WEBHOOK_SECRET="nothing to see here" GITHUB_OAUTH_CLIENT_ID="nothing to see here" GITHUB_OAUTH_CLIENT_SECRET="nothing to see here" GITHUB_STATUS_OAUTH_TOKEN="nothing to see here" COVERALLS_REPO_TOKEN="nothing to see here" CODECOV_REPO_TOKEN="nothing to see here" FREEBSDCI_OAUTH_TOKEN="nothing to see here" FQDN="buildog.julialang.org" BUILDBOT_BRANCH="master" db_user="nothing to see here" db_password="nothing to see here" DOCUMENTER_KEY="nothing to see here" Add mock value for `MACOS_CODESIGN_IDENTITY`
GITHUB_WEBHOOK_SECRET="nothing to see here" GITHUB_OAUTH_CLIENT_ID="nothing to see here" GITHUB_OAUTH_CLIENT_SECRET="nothing to see here" GITHUB_STATUS_OAUTH_TOKEN="nothing to see here" COVERALLS_REPO_TOKEN="nothing to see here" CODECOV_REPO_TOKEN="nothing to see here" FREEBSDCI_OAUTH_TOKEN="nothing to see here" FQDN="buildog.julialang.org" BUILDBOT_BRANCH="master" db_user="nothing to see here" db_password="nothing to see here" DOCUMENTER_KEY="nothing to see here" MACOS_CODESIGN_IDENTITY="nothing to see here"
<commit_before>GITHUB_WEBHOOK_SECRET="nothing to see here" GITHUB_OAUTH_CLIENT_ID="nothing to see here" GITHUB_OAUTH_CLIENT_SECRET="nothing to see here" GITHUB_STATUS_OAUTH_TOKEN="nothing to see here" COVERALLS_REPO_TOKEN="nothing to see here" CODECOV_REPO_TOKEN="nothing to see here" FREEBSDCI_OAUTH_TOKEN="nothing to see here" FQDN="buildog.julialang.org" BUILDBOT_BRANCH="master" db_user="nothing to see here" db_password="nothing to see here" DOCUMENTER_KEY="nothing to see here" <commit_msg>Add mock value for `MACOS_CODESIGN_IDENTITY`<commit_after>
GITHUB_WEBHOOK_SECRET="nothing to see here" GITHUB_OAUTH_CLIENT_ID="nothing to see here" GITHUB_OAUTH_CLIENT_SECRET="nothing to see here" GITHUB_STATUS_OAUTH_TOKEN="nothing to see here" COVERALLS_REPO_TOKEN="nothing to see here" CODECOV_REPO_TOKEN="nothing to see here" FREEBSDCI_OAUTH_TOKEN="nothing to see here" FQDN="buildog.julialang.org" BUILDBOT_BRANCH="master" db_user="nothing to see here" db_password="nothing to see here" DOCUMENTER_KEY="nothing to see here" MACOS_CODESIGN_IDENTITY="nothing to see here"
GITHUB_WEBHOOK_SECRET="nothing to see here" GITHUB_OAUTH_CLIENT_ID="nothing to see here" GITHUB_OAUTH_CLIENT_SECRET="nothing to see here" GITHUB_STATUS_OAUTH_TOKEN="nothing to see here" COVERALLS_REPO_TOKEN="nothing to see here" CODECOV_REPO_TOKEN="nothing to see here" FREEBSDCI_OAUTH_TOKEN="nothing to see here" FQDN="buildog.julialang.org" BUILDBOT_BRANCH="master" db_user="nothing to see here" db_password="nothing to see here" DOCUMENTER_KEY="nothing to see here" Add mock value for `MACOS_CODESIGN_IDENTITY`GITHUB_WEBHOOK_SECRET="nothing to see here" GITHUB_OAUTH_CLIENT_ID="nothing to see here" GITHUB_OAUTH_CLIENT_SECRET="nothing to see here" GITHUB_STATUS_OAUTH_TOKEN="nothing to see here" COVERALLS_REPO_TOKEN="nothing to see here" CODECOV_REPO_TOKEN="nothing to see here" FREEBSDCI_OAUTH_TOKEN="nothing to see here" FQDN="buildog.julialang.org" BUILDBOT_BRANCH="master" db_user="nothing to see here" db_password="nothing to see here" DOCUMENTER_KEY="nothing to see here" MACOS_CODESIGN_IDENTITY="nothing to see here"
<commit_before>GITHUB_WEBHOOK_SECRET="nothing to see here" GITHUB_OAUTH_CLIENT_ID="nothing to see here" GITHUB_OAUTH_CLIENT_SECRET="nothing to see here" GITHUB_STATUS_OAUTH_TOKEN="nothing to see here" COVERALLS_REPO_TOKEN="nothing to see here" CODECOV_REPO_TOKEN="nothing to see here" FREEBSDCI_OAUTH_TOKEN="nothing to see here" FQDN="buildog.julialang.org" BUILDBOT_BRANCH="master" db_user="nothing to see here" db_password="nothing to see here" DOCUMENTER_KEY="nothing to see here" <commit_msg>Add mock value for `MACOS_CODESIGN_IDENTITY`<commit_after>GITHUB_WEBHOOK_SECRET="nothing to see here" GITHUB_OAUTH_CLIENT_ID="nothing to see here" GITHUB_OAUTH_CLIENT_SECRET="nothing to see here" GITHUB_STATUS_OAUTH_TOKEN="nothing to see here" COVERALLS_REPO_TOKEN="nothing to see here" CODECOV_REPO_TOKEN="nothing to see here" FREEBSDCI_OAUTH_TOKEN="nothing to see here" FQDN="buildog.julialang.org" BUILDBOT_BRANCH="master" db_user="nothing to see here" db_password="nothing to see here" DOCUMENTER_KEY="nothing to see here" MACOS_CODESIGN_IDENTITY="nothing to see here"
886c2b92d8dcc40577341245f7973d4a2d31aa90
tests/core/test_mixer.py
tests/core/test_mixer.py
from __future__ import absolute_import, unicode_literals import unittest from mopidy import core class CoreMixerTest(unittest.TestCase): def setUp(self): # noqa: N802 self.core = core.Core(mixer=None, backends=[]) def test_volume(self): self.assertEqual(self.core.mixer.get_volume(), None) self.core.mixer.set_volume(30) self.assertEqual(self.core.mixer.get_volume(), 30) self.core.mixer.set_volume(70) self.assertEqual(self.core.mixer.get_volume(), 70) def test_mute(self): self.assertEqual(self.core.mixer.get_mute(), False) self.core.mixer.set_mute(True) self.assertEqual(self.core.mixer.get_mute(), True)
from __future__ import absolute_import, unicode_literals import unittest import mock from mopidy import core, mixer class CoreMixerTest(unittest.TestCase): def setUp(self): # noqa: N802 self.mixer = mock.Mock(spec=mixer.Mixer) self.core = core.Core(mixer=self.mixer, backends=[]) def test_get_volume(self): self.mixer.get_volume.return_value.get.return_value = 30 self.assertEqual(self.core.mixer.get_volume(), 30) self.mixer.get_volume.assert_called_once_with() def test_set_volume(self): self.core.mixer.set_volume(30) self.mixer.set_volume.assert_called_once_with(30) def test_get_mute(self): self.mixer.get_mute.return_value.get.return_value = True self.assertEqual(self.core.mixer.get_mute(), True) self.mixer.get_mute.assert_called_once_with() def test_set_mute(self): self.core.mixer.set_mute(True) self.mixer.set_mute.assert_called_once_with(True)
Use a mixer mock in tests
core: Use a mixer mock in tests
Python
apache-2.0
ali/mopidy,tkem/mopidy,dbrgn/mopidy,ali/mopidy,bacontext/mopidy,SuperStarPL/mopidy,mokieyue/mopidy,SuperStarPL/mopidy,swak/mopidy,mopidy/mopidy,jodal/mopidy,ali/mopidy,swak/mopidy,dbrgn/mopidy,dbrgn/mopidy,ZenithDK/mopidy,rawdlite/mopidy,hkariti/mopidy,quartz55/mopidy,jodal/mopidy,jodal/mopidy,kingosticks/mopidy,vrs01/mopidy,rawdlite/mopidy,diandiankan/mopidy,hkariti/mopidy,SuperStarPL/mopidy,dbrgn/mopidy,jcass77/mopidy,adamcik/mopidy,pacificIT/mopidy,jmarsik/mopidy,tkem/mopidy,jmarsik/mopidy,mopidy/mopidy,diandiankan/mopidy,tkem/mopidy,mopidy/mopidy,vrs01/mopidy,rawdlite/mopidy,vrs01/mopidy,ali/mopidy,pacificIT/mopidy,jmarsik/mopidy,glogiotatidis/mopidy,ZenithDK/mopidy,jcass77/mopidy,bacontext/mopidy,ZenithDK/mopidy,glogiotatidis/mopidy,adamcik/mopidy,SuperStarPL/mopidy,glogiotatidis/mopidy,pacificIT/mopidy,bacontext/mopidy,bencevans/mopidy,vrs01/mopidy,rawdlite/mopidy,jcass77/mopidy,quartz55/mopidy,ZenithDK/mopidy,swak/mopidy,bencevans/mopidy,quartz55/mopidy,mokieyue/mopidy,kingosticks/mopidy,bencevans/mopidy,kingosticks/mopidy,diandiankan/mopidy,bencevans/mopidy,glogiotatidis/mopidy,tkem/mopidy,quartz55/mopidy,diandiankan/mopidy,mokieyue/mopidy,mokieyue/mopidy,jmarsik/mopidy,bacontext/mopidy,swak/mopidy,adamcik/mopidy,hkariti/mopidy,hkariti/mopidy,pacificIT/mopidy
from __future__ import absolute_import, unicode_literals import unittest from mopidy import core class CoreMixerTest(unittest.TestCase): def setUp(self): # noqa: N802 self.core = core.Core(mixer=None, backends=[]) def test_volume(self): self.assertEqual(self.core.mixer.get_volume(), None) self.core.mixer.set_volume(30) self.assertEqual(self.core.mixer.get_volume(), 30) self.core.mixer.set_volume(70) self.assertEqual(self.core.mixer.get_volume(), 70) def test_mute(self): self.assertEqual(self.core.mixer.get_mute(), False) self.core.mixer.set_mute(True) self.assertEqual(self.core.mixer.get_mute(), True) core: Use a mixer mock in tests
from __future__ import absolute_import, unicode_literals import unittest import mock from mopidy import core, mixer class CoreMixerTest(unittest.TestCase): def setUp(self): # noqa: N802 self.mixer = mock.Mock(spec=mixer.Mixer) self.core = core.Core(mixer=self.mixer, backends=[]) def test_get_volume(self): self.mixer.get_volume.return_value.get.return_value = 30 self.assertEqual(self.core.mixer.get_volume(), 30) self.mixer.get_volume.assert_called_once_with() def test_set_volume(self): self.core.mixer.set_volume(30) self.mixer.set_volume.assert_called_once_with(30) def test_get_mute(self): self.mixer.get_mute.return_value.get.return_value = True self.assertEqual(self.core.mixer.get_mute(), True) self.mixer.get_mute.assert_called_once_with() def test_set_mute(self): self.core.mixer.set_mute(True) self.mixer.set_mute.assert_called_once_with(True)
<commit_before>from __future__ import absolute_import, unicode_literals import unittest from mopidy import core class CoreMixerTest(unittest.TestCase): def setUp(self): # noqa: N802 self.core = core.Core(mixer=None, backends=[]) def test_volume(self): self.assertEqual(self.core.mixer.get_volume(), None) self.core.mixer.set_volume(30) self.assertEqual(self.core.mixer.get_volume(), 30) self.core.mixer.set_volume(70) self.assertEqual(self.core.mixer.get_volume(), 70) def test_mute(self): self.assertEqual(self.core.mixer.get_mute(), False) self.core.mixer.set_mute(True) self.assertEqual(self.core.mixer.get_mute(), True) <commit_msg>core: Use a mixer mock in tests<commit_after>
from __future__ import absolute_import, unicode_literals import unittest import mock from mopidy import core, mixer class CoreMixerTest(unittest.TestCase): def setUp(self): # noqa: N802 self.mixer = mock.Mock(spec=mixer.Mixer) self.core = core.Core(mixer=self.mixer, backends=[]) def test_get_volume(self): self.mixer.get_volume.return_value.get.return_value = 30 self.assertEqual(self.core.mixer.get_volume(), 30) self.mixer.get_volume.assert_called_once_with() def test_set_volume(self): self.core.mixer.set_volume(30) self.mixer.set_volume.assert_called_once_with(30) def test_get_mute(self): self.mixer.get_mute.return_value.get.return_value = True self.assertEqual(self.core.mixer.get_mute(), True) self.mixer.get_mute.assert_called_once_with() def test_set_mute(self): self.core.mixer.set_mute(True) self.mixer.set_mute.assert_called_once_with(True)
from __future__ import absolute_import, unicode_literals import unittest from mopidy import core class CoreMixerTest(unittest.TestCase): def setUp(self): # noqa: N802 self.core = core.Core(mixer=None, backends=[]) def test_volume(self): self.assertEqual(self.core.mixer.get_volume(), None) self.core.mixer.set_volume(30) self.assertEqual(self.core.mixer.get_volume(), 30) self.core.mixer.set_volume(70) self.assertEqual(self.core.mixer.get_volume(), 70) def test_mute(self): self.assertEqual(self.core.mixer.get_mute(), False) self.core.mixer.set_mute(True) self.assertEqual(self.core.mixer.get_mute(), True) core: Use a mixer mock in testsfrom __future__ import absolute_import, unicode_literals import unittest import mock from mopidy import core, mixer class CoreMixerTest(unittest.TestCase): def setUp(self): # noqa: N802 self.mixer = mock.Mock(spec=mixer.Mixer) self.core = core.Core(mixer=self.mixer, backends=[]) def test_get_volume(self): self.mixer.get_volume.return_value.get.return_value = 30 self.assertEqual(self.core.mixer.get_volume(), 30) self.mixer.get_volume.assert_called_once_with() def test_set_volume(self): self.core.mixer.set_volume(30) self.mixer.set_volume.assert_called_once_with(30) def test_get_mute(self): self.mixer.get_mute.return_value.get.return_value = True self.assertEqual(self.core.mixer.get_mute(), True) self.mixer.get_mute.assert_called_once_with() def test_set_mute(self): self.core.mixer.set_mute(True) self.mixer.set_mute.assert_called_once_with(True)
<commit_before>from __future__ import absolute_import, unicode_literals import unittest from mopidy import core class CoreMixerTest(unittest.TestCase): def setUp(self): # noqa: N802 self.core = core.Core(mixer=None, backends=[]) def test_volume(self): self.assertEqual(self.core.mixer.get_volume(), None) self.core.mixer.set_volume(30) self.assertEqual(self.core.mixer.get_volume(), 30) self.core.mixer.set_volume(70) self.assertEqual(self.core.mixer.get_volume(), 70) def test_mute(self): self.assertEqual(self.core.mixer.get_mute(), False) self.core.mixer.set_mute(True) self.assertEqual(self.core.mixer.get_mute(), True) <commit_msg>core: Use a mixer mock in tests<commit_after>from __future__ import absolute_import, unicode_literals import unittest import mock from mopidy import core, mixer class CoreMixerTest(unittest.TestCase): def setUp(self): # noqa: N802 self.mixer = mock.Mock(spec=mixer.Mixer) self.core = core.Core(mixer=self.mixer, backends=[]) def test_get_volume(self): self.mixer.get_volume.return_value.get.return_value = 30 self.assertEqual(self.core.mixer.get_volume(), 30) self.mixer.get_volume.assert_called_once_with() def test_set_volume(self): self.core.mixer.set_volume(30) self.mixer.set_volume.assert_called_once_with(30) def test_get_mute(self): self.mixer.get_mute.return_value.get.return_value = True self.assertEqual(self.core.mixer.get_mute(), True) self.mixer.get_mute.assert_called_once_with() def test_set_mute(self): self.core.mixer.set_mute(True) self.mixer.set_mute.assert_called_once_with(True)
c45e00924fbe90fb6ff9465202a77d390c685dc4
tests/test_cli_update.py
tests/test_cli_update.py
# -*- coding: utf-8 -*- import pathlib import json def test_store_template_data_to_json(cli_runner, tmp_rc, tmp_templates_file): result = cli_runner([ '-c', tmp_rc, 'update' ]) assert result.exit_code == 0 templates = pathlib.Path(tmp_templates_file) assert templates.exists() with templates.open('r', encoding='utf8') as fh: template_data = json.load(fh) fetched_templates = [template['name'] for template in template_data] expected_templates = [ 'cookiecutter-pypackage', 'cookiecutter-pylibrary', 'cookiecutter-pytest-plugin', 'cookiecutter-tapioca', 'cookiecutter-django', ] assert fetched_templates == expected_templates
# -*- coding: utf-8 -*- import pathlib import json from configparser import RawConfigParser import pytest def test_store_template_data_to_json(cli_runner, tmp_rc, tmp_templates_file): result = cli_runner([ '-c', tmp_rc, 'update' ]) assert result.exit_code == 0 templates = pathlib.Path(tmp_templates_file) assert templates.exists() with templates.open('r', encoding='utf8') as fh: template_data = json.load(fh) fetched_templates = [template['name'] for template in template_data] expected_templates = [ 'cookiecutter-pypackage', 'cookiecutter-pylibrary', 'cookiecutter-pytest-plugin', 'cookiecutter-tapioca', 'cookiecutter-django', ] assert fetched_templates == expected_templates @pytest.fixture def incomplete_rc(tmpdir): rc_file = str(tmpdir / 'noperc') config = RawConfigParser() config['nope'] = {'foo': 'bar'} with open(rc_file, 'w', encoding='utf-8') as fh: config.write(fh) return rc_file def test_fail_missing_username(cli_runner, incomplete_rc, tmp_templates_file): result = cli_runner([ '-c', incomplete_rc, 'update', '-t', '1234', '-d', tmp_templates_file ]) assert result.exit_code == 2 assert 'Error: Missing option "-u" / "--username".' in result.output
Implement a test for missing username in update cmd
Implement a test for missing username in update cmd
Python
bsd-3-clause
hackebrot/cibopath
# -*- coding: utf-8 -*- import pathlib import json def test_store_template_data_to_json(cli_runner, tmp_rc, tmp_templates_file): result = cli_runner([ '-c', tmp_rc, 'update' ]) assert result.exit_code == 0 templates = pathlib.Path(tmp_templates_file) assert templates.exists() with templates.open('r', encoding='utf8') as fh: template_data = json.load(fh) fetched_templates = [template['name'] for template in template_data] expected_templates = [ 'cookiecutter-pypackage', 'cookiecutter-pylibrary', 'cookiecutter-pytest-plugin', 'cookiecutter-tapioca', 'cookiecutter-django', ] assert fetched_templates == expected_templates Implement a test for missing username in update cmd
# -*- coding: utf-8 -*- import pathlib import json from configparser import RawConfigParser import pytest def test_store_template_data_to_json(cli_runner, tmp_rc, tmp_templates_file): result = cli_runner([ '-c', tmp_rc, 'update' ]) assert result.exit_code == 0 templates = pathlib.Path(tmp_templates_file) assert templates.exists() with templates.open('r', encoding='utf8') as fh: template_data = json.load(fh) fetched_templates = [template['name'] for template in template_data] expected_templates = [ 'cookiecutter-pypackage', 'cookiecutter-pylibrary', 'cookiecutter-pytest-plugin', 'cookiecutter-tapioca', 'cookiecutter-django', ] assert fetched_templates == expected_templates @pytest.fixture def incomplete_rc(tmpdir): rc_file = str(tmpdir / 'noperc') config = RawConfigParser() config['nope'] = {'foo': 'bar'} with open(rc_file, 'w', encoding='utf-8') as fh: config.write(fh) return rc_file def test_fail_missing_username(cli_runner, incomplete_rc, tmp_templates_file): result = cli_runner([ '-c', incomplete_rc, 'update', '-t', '1234', '-d', tmp_templates_file ]) assert result.exit_code == 2 assert 'Error: Missing option "-u" / "--username".' in result.output
<commit_before># -*- coding: utf-8 -*- import pathlib import json def test_store_template_data_to_json(cli_runner, tmp_rc, tmp_templates_file): result = cli_runner([ '-c', tmp_rc, 'update' ]) assert result.exit_code == 0 templates = pathlib.Path(tmp_templates_file) assert templates.exists() with templates.open('r', encoding='utf8') as fh: template_data = json.load(fh) fetched_templates = [template['name'] for template in template_data] expected_templates = [ 'cookiecutter-pypackage', 'cookiecutter-pylibrary', 'cookiecutter-pytest-plugin', 'cookiecutter-tapioca', 'cookiecutter-django', ] assert fetched_templates == expected_templates <commit_msg>Implement a test for missing username in update cmd<commit_after>
# -*- coding: utf-8 -*- import pathlib import json from configparser import RawConfigParser import pytest def test_store_template_data_to_json(cli_runner, tmp_rc, tmp_templates_file): result = cli_runner([ '-c', tmp_rc, 'update' ]) assert result.exit_code == 0 templates = pathlib.Path(tmp_templates_file) assert templates.exists() with templates.open('r', encoding='utf8') as fh: template_data = json.load(fh) fetched_templates = [template['name'] for template in template_data] expected_templates = [ 'cookiecutter-pypackage', 'cookiecutter-pylibrary', 'cookiecutter-pytest-plugin', 'cookiecutter-tapioca', 'cookiecutter-django', ] assert fetched_templates == expected_templates @pytest.fixture def incomplete_rc(tmpdir): rc_file = str(tmpdir / 'noperc') config = RawConfigParser() config['nope'] = {'foo': 'bar'} with open(rc_file, 'w', encoding='utf-8') as fh: config.write(fh) return rc_file def test_fail_missing_username(cli_runner, incomplete_rc, tmp_templates_file): result = cli_runner([ '-c', incomplete_rc, 'update', '-t', '1234', '-d', tmp_templates_file ]) assert result.exit_code == 2 assert 'Error: Missing option "-u" / "--username".' in result.output
# -*- coding: utf-8 -*- import pathlib import json def test_store_template_data_to_json(cli_runner, tmp_rc, tmp_templates_file): result = cli_runner([ '-c', tmp_rc, 'update' ]) assert result.exit_code == 0 templates = pathlib.Path(tmp_templates_file) assert templates.exists() with templates.open('r', encoding='utf8') as fh: template_data = json.load(fh) fetched_templates = [template['name'] for template in template_data] expected_templates = [ 'cookiecutter-pypackage', 'cookiecutter-pylibrary', 'cookiecutter-pytest-plugin', 'cookiecutter-tapioca', 'cookiecutter-django', ] assert fetched_templates == expected_templates Implement a test for missing username in update cmd# -*- coding: utf-8 -*- import pathlib import json from configparser import RawConfigParser import pytest def test_store_template_data_to_json(cli_runner, tmp_rc, tmp_templates_file): result = cli_runner([ '-c', tmp_rc, 'update' ]) assert result.exit_code == 0 templates = pathlib.Path(tmp_templates_file) assert templates.exists() with templates.open('r', encoding='utf8') as fh: template_data = json.load(fh) fetched_templates = [template['name'] for template in template_data] expected_templates = [ 'cookiecutter-pypackage', 'cookiecutter-pylibrary', 'cookiecutter-pytest-plugin', 'cookiecutter-tapioca', 'cookiecutter-django', ] assert fetched_templates == expected_templates @pytest.fixture def incomplete_rc(tmpdir): rc_file = str(tmpdir / 'noperc') config = RawConfigParser() config['nope'] = {'foo': 'bar'} with open(rc_file, 'w', encoding='utf-8') as fh: config.write(fh) return rc_file def test_fail_missing_username(cli_runner, incomplete_rc, tmp_templates_file): result = cli_runner([ '-c', incomplete_rc, 'update', '-t', '1234', '-d', tmp_templates_file ]) assert result.exit_code == 2 assert 'Error: Missing option "-u" / "--username".' in result.output
<commit_before># -*- coding: utf-8 -*- import pathlib import json def test_store_template_data_to_json(cli_runner, tmp_rc, tmp_templates_file): result = cli_runner([ '-c', tmp_rc, 'update' ]) assert result.exit_code == 0 templates = pathlib.Path(tmp_templates_file) assert templates.exists() with templates.open('r', encoding='utf8') as fh: template_data = json.load(fh) fetched_templates = [template['name'] for template in template_data] expected_templates = [ 'cookiecutter-pypackage', 'cookiecutter-pylibrary', 'cookiecutter-pytest-plugin', 'cookiecutter-tapioca', 'cookiecutter-django', ] assert fetched_templates == expected_templates <commit_msg>Implement a test for missing username in update cmd<commit_after># -*- coding: utf-8 -*- import pathlib import json from configparser import RawConfigParser import pytest def test_store_template_data_to_json(cli_runner, tmp_rc, tmp_templates_file): result = cli_runner([ '-c', tmp_rc, 'update' ]) assert result.exit_code == 0 templates = pathlib.Path(tmp_templates_file) assert templates.exists() with templates.open('r', encoding='utf8') as fh: template_data = json.load(fh) fetched_templates = [template['name'] for template in template_data] expected_templates = [ 'cookiecutter-pypackage', 'cookiecutter-pylibrary', 'cookiecutter-pytest-plugin', 'cookiecutter-tapioca', 'cookiecutter-django', ] assert fetched_templates == expected_templates @pytest.fixture def incomplete_rc(tmpdir): rc_file = str(tmpdir / 'noperc') config = RawConfigParser() config['nope'] = {'foo': 'bar'} with open(rc_file, 'w', encoding='utf-8') as fh: config.write(fh) return rc_file def test_fail_missing_username(cli_runner, incomplete_rc, tmp_templates_file): result = cli_runner([ '-c', incomplete_rc, 'update', '-t', '1234', '-d', tmp_templates_file ]) assert result.exit_code == 2 assert 'Error: Missing option "-u" / "--username".' in result.output
a4104dea137b8fa2aedc38ac3bda53c559c1f45a
tests/test_opentransf.py
tests/test_opentransf.py
import pymorph import numpy as np def test_opentransf(): f = np.array([ [0,0,0,0,0,0,0,0], [0,0,1,1,1,1,0,0], [0,1,0,1,1,1,0,0], [0,0,1,1,1,1,0,0], [1,1,0,0,0,0,0,0]], bool) ot = pymorph.opentransf( f, 'city-block') for y in xrange(ot.shape[0]): for x in xrange(ot.shape[1]): r = ot[y,x] t = f.copy() for k in xrange(1, r+1): assert t[y,x] t = pymorph.open(f, pymorph.sedisk(k, 2, 'city-block')) assert not t[y,x] def test_all(): f = np.arange(9).reshape((3,3)) % 3 > 0 # linear-h crashed in 0.95 # and was underlying cause of crash in patsec(f, 'linear-h') g = pymorph.opentransf(f, 'linear-h')
import pymorph import numpy as np def test_opentransf(): f = np.array([ [0,0,0,0,0,0,0,0], [0,0,1,1,1,1,0,0], [0,1,0,1,1,1,0,0], [0,0,1,1,1,1,0,0], [1,1,0,0,0,0,0,0]], bool) ot = pymorph.opentransf( f, 'city-block') for y in xrange(ot.shape[0]): for x in xrange(ot.shape[1]): r = ot[y,x] t = f.copy() for k in xrange(1, r+1): assert t[y,x] t = pymorph.open(f, pymorph.sedisk(k, 2, 'city-block')) assert not t[y,x] def test_all_types(): f = np.arange(9).reshape((3,3)) % 3 > 0 # linear-h crashed in 0.95 # and was underlying cause of crash in patsec(f, 'linear-h') def test_type(type, Buser): g = pymorph.opentransf(f, type, Buser=Buser) yield test_type, 'linear-h', None yield test_type, 'octagon', None yield test_type, 'chessboard', None yield test_type, 'city-block', None yield test_type, 'linear-v', None yield test_type, 'linear-45r', None yield test_type, 'linear-45l', None Buser = np.ones((3,3),bool) Buser[2,2] = 0 yield test_type, 'user', Buser
Test all cases of type for opentransf
TST: Test all cases of type for opentransf
Python
bsd-3-clause
luispedro/pymorph
import pymorph import numpy as np def test_opentransf(): f = np.array([ [0,0,0,0,0,0,0,0], [0,0,1,1,1,1,0,0], [0,1,0,1,1,1,0,0], [0,0,1,1,1,1,0,0], [1,1,0,0,0,0,0,0]], bool) ot = pymorph.opentransf( f, 'city-block') for y in xrange(ot.shape[0]): for x in xrange(ot.shape[1]): r = ot[y,x] t = f.copy() for k in xrange(1, r+1): assert t[y,x] t = pymorph.open(f, pymorph.sedisk(k, 2, 'city-block')) assert not t[y,x] def test_all(): f = np.arange(9).reshape((3,3)) % 3 > 0 # linear-h crashed in 0.95 # and was underlying cause of crash in patsec(f, 'linear-h') g = pymorph.opentransf(f, 'linear-h') TST: Test all cases of type for opentransf
import pymorph import numpy as np def test_opentransf(): f = np.array([ [0,0,0,0,0,0,0,0], [0,0,1,1,1,1,0,0], [0,1,0,1,1,1,0,0], [0,0,1,1,1,1,0,0], [1,1,0,0,0,0,0,0]], bool) ot = pymorph.opentransf( f, 'city-block') for y in xrange(ot.shape[0]): for x in xrange(ot.shape[1]): r = ot[y,x] t = f.copy() for k in xrange(1, r+1): assert t[y,x] t = pymorph.open(f, pymorph.sedisk(k, 2, 'city-block')) assert not t[y,x] def test_all_types(): f = np.arange(9).reshape((3,3)) % 3 > 0 # linear-h crashed in 0.95 # and was underlying cause of crash in patsec(f, 'linear-h') def test_type(type, Buser): g = pymorph.opentransf(f, type, Buser=Buser) yield test_type, 'linear-h', None yield test_type, 'octagon', None yield test_type, 'chessboard', None yield test_type, 'city-block', None yield test_type, 'linear-v', None yield test_type, 'linear-45r', None yield test_type, 'linear-45l', None Buser = np.ones((3,3),bool) Buser[2,2] = 0 yield test_type, 'user', Buser
<commit_before>import pymorph import numpy as np def test_opentransf(): f = np.array([ [0,0,0,0,0,0,0,0], [0,0,1,1,1,1,0,0], [0,1,0,1,1,1,0,0], [0,0,1,1,1,1,0,0], [1,1,0,0,0,0,0,0]], bool) ot = pymorph.opentransf( f, 'city-block') for y in xrange(ot.shape[0]): for x in xrange(ot.shape[1]): r = ot[y,x] t = f.copy() for k in xrange(1, r+1): assert t[y,x] t = pymorph.open(f, pymorph.sedisk(k, 2, 'city-block')) assert not t[y,x] def test_all(): f = np.arange(9).reshape((3,3)) % 3 > 0 # linear-h crashed in 0.95 # and was underlying cause of crash in patsec(f, 'linear-h') g = pymorph.opentransf(f, 'linear-h') <commit_msg>TST: Test all cases of type for opentransf<commit_after>
import pymorph import numpy as np def test_opentransf(): f = np.array([ [0,0,0,0,0,0,0,0], [0,0,1,1,1,1,0,0], [0,1,0,1,1,1,0,0], [0,0,1,1,1,1,0,0], [1,1,0,0,0,0,0,0]], bool) ot = pymorph.opentransf( f, 'city-block') for y in xrange(ot.shape[0]): for x in xrange(ot.shape[1]): r = ot[y,x] t = f.copy() for k in xrange(1, r+1): assert t[y,x] t = pymorph.open(f, pymorph.sedisk(k, 2, 'city-block')) assert not t[y,x] def test_all_types(): f = np.arange(9).reshape((3,3)) % 3 > 0 # linear-h crashed in 0.95 # and was underlying cause of crash in patsec(f, 'linear-h') def test_type(type, Buser): g = pymorph.opentransf(f, type, Buser=Buser) yield test_type, 'linear-h', None yield test_type, 'octagon', None yield test_type, 'chessboard', None yield test_type, 'city-block', None yield test_type, 'linear-v', None yield test_type, 'linear-45r', None yield test_type, 'linear-45l', None Buser = np.ones((3,3),bool) Buser[2,2] = 0 yield test_type, 'user', Buser
import pymorph import numpy as np def test_opentransf(): f = np.array([ [0,0,0,0,0,0,0,0], [0,0,1,1,1,1,0,0], [0,1,0,1,1,1,0,0], [0,0,1,1,1,1,0,0], [1,1,0,0,0,0,0,0]], bool) ot = pymorph.opentransf( f, 'city-block') for y in xrange(ot.shape[0]): for x in xrange(ot.shape[1]): r = ot[y,x] t = f.copy() for k in xrange(1, r+1): assert t[y,x] t = pymorph.open(f, pymorph.sedisk(k, 2, 'city-block')) assert not t[y,x] def test_all(): f = np.arange(9).reshape((3,3)) % 3 > 0 # linear-h crashed in 0.95 # and was underlying cause of crash in patsec(f, 'linear-h') g = pymorph.opentransf(f, 'linear-h') TST: Test all cases of type for opentransfimport pymorph import numpy as np def test_opentransf(): f = np.array([ [0,0,0,0,0,0,0,0], [0,0,1,1,1,1,0,0], [0,1,0,1,1,1,0,0], [0,0,1,1,1,1,0,0], [1,1,0,0,0,0,0,0]], bool) ot = pymorph.opentransf( f, 'city-block') for y in xrange(ot.shape[0]): for x in xrange(ot.shape[1]): r = ot[y,x] t = f.copy() for k in xrange(1, r+1): assert t[y,x] t = pymorph.open(f, pymorph.sedisk(k, 2, 'city-block')) assert not t[y,x] def test_all_types(): f = np.arange(9).reshape((3,3)) % 3 > 0 # linear-h crashed in 0.95 # and was underlying cause of crash in patsec(f, 'linear-h') def test_type(type, Buser): g = pymorph.opentransf(f, type, Buser=Buser) yield test_type, 'linear-h', None yield test_type, 'octagon', None yield test_type, 'chessboard', None yield test_type, 'city-block', None yield test_type, 'linear-v', None yield test_type, 'linear-45r', None yield test_type, 'linear-45l', None Buser = np.ones((3,3),bool) Buser[2,2] = 0 yield test_type, 'user', Buser
<commit_before>import pymorph import numpy as np def test_opentransf(): f = np.array([ [0,0,0,0,0,0,0,0], [0,0,1,1,1,1,0,0], [0,1,0,1,1,1,0,0], [0,0,1,1,1,1,0,0], [1,1,0,0,0,0,0,0]], bool) ot = pymorph.opentransf( f, 'city-block') for y in xrange(ot.shape[0]): for x in xrange(ot.shape[1]): r = ot[y,x] t = f.copy() for k in xrange(1, r+1): assert t[y,x] t = pymorph.open(f, pymorph.sedisk(k, 2, 'city-block')) assert not t[y,x] def test_all(): f = np.arange(9).reshape((3,3)) % 3 > 0 # linear-h crashed in 0.95 # and was underlying cause of crash in patsec(f, 'linear-h') g = pymorph.opentransf(f, 'linear-h') <commit_msg>TST: Test all cases of type for opentransf<commit_after>import pymorph import numpy as np def test_opentransf(): f = np.array([ [0,0,0,0,0,0,0,0], [0,0,1,1,1,1,0,0], [0,1,0,1,1,1,0,0], [0,0,1,1,1,1,0,0], [1,1,0,0,0,0,0,0]], bool) ot = pymorph.opentransf( f, 'city-block') for y in xrange(ot.shape[0]): for x in xrange(ot.shape[1]): r = ot[y,x] t = f.copy() for k in xrange(1, r+1): assert t[y,x] t = pymorph.open(f, pymorph.sedisk(k, 2, 'city-block')) assert not t[y,x] def test_all_types(): f = np.arange(9).reshape((3,3)) % 3 > 0 # linear-h crashed in 0.95 # and was underlying cause of crash in patsec(f, 'linear-h') def test_type(type, Buser): g = pymorph.opentransf(f, type, Buser=Buser) yield test_type, 'linear-h', None yield test_type, 'octagon', None yield test_type, 'chessboard', None yield test_type, 'city-block', None yield test_type, 'linear-v', None yield test_type, 'linear-45r', None yield test_type, 'linear-45l', None Buser = np.ones((3,3),bool) Buser[2,2] = 0 yield test_type, 'user', Buser
95e1d4c2ec42f09fddf48c5a32f0fe409132380b
lab/monitors/nova_service_list.py
lab/monitors/nova_service_list.py
def start(lab, log, args): import time from fabric.context_managers import shell_env grep_host = args.get('grep_host', 'overcloud-') duration = args['duration'] period = args['period'] statuses = {'up': 1, 'down': 0} server = lab.director() start_time = time.time() while start_time + duration > time.time(): with shell_env(OS_AUTH_URL=lab.cloud.end_point, OS_USERNAME=lab.cloud.user, OS_PASSWORD=lab.cloud.password, OS_TENANT_NAME=lab.cloud.tenant): res = server.run("nova service-list | grep {0} | awk '{{print $4 \" \" $6 \" \" $12}}'".format(grep_host), warn_only=True) results = [line.split() for line in res.split('\n')] msg = ' '.join(['{1}:{0}={2}'.format(r[0], r[1], statuses[r[2]]) for r in results]) log.info('{1}'.format(grep_host, msg)) time.sleep(period)
def start(lab, log, args): from fabric.context_managers import shell_env grep_host = args.get('grep_host', 'overcloud-') statuses = {'up': 1, 'down': 0} server = lab.director() with shell_env(OS_AUTH_URL=lab.cloud.end_point, OS_USERNAME=lab.cloud.user, OS_PASSWORD=lab.cloud.password, OS_TENANT_NAME=lab.cloud.tenant): res = server.run("nova service-list | grep {0} | awk '{{print $4 \" \" $6 \" \" $12}}'".format(grep_host), warn_only=True) results = [line.split() for line in res.split('\n')] msg = ' '.join(['{1}:{0}={2}'.format(r[0], r[1], statuses[r[2]]) for r in results]) log.info('{1}'.format(grep_host, msg))
Verify services status if FI is rebooted
Verify services status if FI is rebooted Change-Id: Ia02ef16d53fbb7b55a8de884ff16a4bef345a1f2
Python
apache-2.0
CiscoSystems/os-sqe,CiscoSystems/os-sqe,CiscoSystems/os-sqe
def start(lab, log, args): import time from fabric.context_managers import shell_env grep_host = args.get('grep_host', 'overcloud-') duration = args['duration'] period = args['period'] statuses = {'up': 1, 'down': 0} server = lab.director() start_time = time.time() while start_time + duration > time.time(): with shell_env(OS_AUTH_URL=lab.cloud.end_point, OS_USERNAME=lab.cloud.user, OS_PASSWORD=lab.cloud.password, OS_TENANT_NAME=lab.cloud.tenant): res = server.run("nova service-list | grep {0} | awk '{{print $4 \" \" $6 \" \" $12}}'".format(grep_host), warn_only=True) results = [line.split() for line in res.split('\n')] msg = ' '.join(['{1}:{0}={2}'.format(r[0], r[1], statuses[r[2]]) for r in results]) log.info('{1}'.format(grep_host, msg)) time.sleep(period) Verify services status if FI is rebooted Change-Id: Ia02ef16d53fbb7b55a8de884ff16a4bef345a1f2
def start(lab, log, args): from fabric.context_managers import shell_env grep_host = args.get('grep_host', 'overcloud-') statuses = {'up': 1, 'down': 0} server = lab.director() with shell_env(OS_AUTH_URL=lab.cloud.end_point, OS_USERNAME=lab.cloud.user, OS_PASSWORD=lab.cloud.password, OS_TENANT_NAME=lab.cloud.tenant): res = server.run("nova service-list | grep {0} | awk '{{print $4 \" \" $6 \" \" $12}}'".format(grep_host), warn_only=True) results = [line.split() for line in res.split('\n')] msg = ' '.join(['{1}:{0}={2}'.format(r[0], r[1], statuses[r[2]]) for r in results]) log.info('{1}'.format(grep_host, msg))
<commit_before>def start(lab, log, args): import time from fabric.context_managers import shell_env grep_host = args.get('grep_host', 'overcloud-') duration = args['duration'] period = args['period'] statuses = {'up': 1, 'down': 0} server = lab.director() start_time = time.time() while start_time + duration > time.time(): with shell_env(OS_AUTH_URL=lab.cloud.end_point, OS_USERNAME=lab.cloud.user, OS_PASSWORD=lab.cloud.password, OS_TENANT_NAME=lab.cloud.tenant): res = server.run("nova service-list | grep {0} | awk '{{print $4 \" \" $6 \" \" $12}}'".format(grep_host), warn_only=True) results = [line.split() for line in res.split('\n')] msg = ' '.join(['{1}:{0}={2}'.format(r[0], r[1], statuses[r[2]]) for r in results]) log.info('{1}'.format(grep_host, msg)) time.sleep(period) <commit_msg>Verify services status if FI is rebooted Change-Id: Ia02ef16d53fbb7b55a8de884ff16a4bef345a1f2<commit_after>
def start(lab, log, args): from fabric.context_managers import shell_env grep_host = args.get('grep_host', 'overcloud-') statuses = {'up': 1, 'down': 0} server = lab.director() with shell_env(OS_AUTH_URL=lab.cloud.end_point, OS_USERNAME=lab.cloud.user, OS_PASSWORD=lab.cloud.password, OS_TENANT_NAME=lab.cloud.tenant): res = server.run("nova service-list | grep {0} | awk '{{print $4 \" \" $6 \" \" $12}}'".format(grep_host), warn_only=True) results = [line.split() for line in res.split('\n')] msg = ' '.join(['{1}:{0}={2}'.format(r[0], r[1], statuses[r[2]]) for r in results]) log.info('{1}'.format(grep_host, msg))
def start(lab, log, args): import time from fabric.context_managers import shell_env grep_host = args.get('grep_host', 'overcloud-') duration = args['duration'] period = args['period'] statuses = {'up': 1, 'down': 0} server = lab.director() start_time = time.time() while start_time + duration > time.time(): with shell_env(OS_AUTH_URL=lab.cloud.end_point, OS_USERNAME=lab.cloud.user, OS_PASSWORD=lab.cloud.password, OS_TENANT_NAME=lab.cloud.tenant): res = server.run("nova service-list | grep {0} | awk '{{print $4 \" \" $6 \" \" $12}}'".format(grep_host), warn_only=True) results = [line.split() for line in res.split('\n')] msg = ' '.join(['{1}:{0}={2}'.format(r[0], r[1], statuses[r[2]]) for r in results]) log.info('{1}'.format(grep_host, msg)) time.sleep(period) Verify services status if FI is rebooted Change-Id: Ia02ef16d53fbb7b55a8de884ff16a4bef345a1f2def start(lab, log, args): from fabric.context_managers import shell_env grep_host = args.get('grep_host', 'overcloud-') statuses = {'up': 1, 'down': 0} server = lab.director() with shell_env(OS_AUTH_URL=lab.cloud.end_point, OS_USERNAME=lab.cloud.user, OS_PASSWORD=lab.cloud.password, OS_TENANT_NAME=lab.cloud.tenant): res = server.run("nova service-list | grep {0} | awk '{{print $4 \" \" $6 \" \" $12}}'".format(grep_host), warn_only=True) results = [line.split() for line in res.split('\n')] msg = ' '.join(['{1}:{0}={2}'.format(r[0], r[1], statuses[r[2]]) for r in results]) log.info('{1}'.format(grep_host, msg))
<commit_before>def start(lab, log, args): import time from fabric.context_managers import shell_env grep_host = args.get('grep_host', 'overcloud-') duration = args['duration'] period = args['period'] statuses = {'up': 1, 'down': 0} server = lab.director() start_time = time.time() while start_time + duration > time.time(): with shell_env(OS_AUTH_URL=lab.cloud.end_point, OS_USERNAME=lab.cloud.user, OS_PASSWORD=lab.cloud.password, OS_TENANT_NAME=lab.cloud.tenant): res = server.run("nova service-list | grep {0} | awk '{{print $4 \" \" $6 \" \" $12}}'".format(grep_host), warn_only=True) results = [line.split() for line in res.split('\n')] msg = ' '.join(['{1}:{0}={2}'.format(r[0], r[1], statuses[r[2]]) for r in results]) log.info('{1}'.format(grep_host, msg)) time.sleep(period) <commit_msg>Verify services status if FI is rebooted Change-Id: Ia02ef16d53fbb7b55a8de884ff16a4bef345a1f2<commit_after>def start(lab, log, args): from fabric.context_managers import shell_env grep_host = args.get('grep_host', 'overcloud-') statuses = {'up': 1, 'down': 0} server = lab.director() with shell_env(OS_AUTH_URL=lab.cloud.end_point, OS_USERNAME=lab.cloud.user, OS_PASSWORD=lab.cloud.password, OS_TENANT_NAME=lab.cloud.tenant): res = server.run("nova service-list | grep {0} | awk '{{print $4 \" \" $6 \" \" $12}}'".format(grep_host), warn_only=True) results = [line.split() for line in res.split('\n')] msg = ' '.join(['{1}:{0}={2}'.format(r[0], r[1], statuses[r[2]]) for r in results]) log.info('{1}'.format(grep_host, msg))
26bae1f6094550939b1ed2ded3885e5d7befc39d
rply/token.py
rply/token.py
class BaseBox(object): pass class Token(BaseBox): def __init__(self, name, value, source_pos=None): BaseBox.__init__(self) self.name = name self.value = value self.source_pos = source_pos def __eq__(self, other): return self.name == other.name and self.value == other.value def gettokentype(self): return self.name def getsourcepos(self): return self.source_pos def getstr(self): return self.value class SourcePosition(object): def __init__(self, idx, lineno, colno): self.idx = idx self.lineno = lineno self.colno = colno
class BaseBox(object): pass class Token(BaseBox): def __init__(self, name, value, source_pos=None): self.name = name self.value = value self.source_pos = source_pos def __eq__(self, other): return self.name == other.name and self.value == other.value def gettokentype(self): return self.name def getsourcepos(self): return self.source_pos def getstr(self): return self.value class SourcePosition(object): def __init__(self, idx, lineno, colno): self.idx = idx self.lineno = lineno self.colno = colno
Drop the __init__ call to object.__init__, RPython doesn't like it and it doesn't doa nything
Drop the __init__ call to object.__init__, RPython doesn't like it and it doesn't doa nything
Python
bsd-3-clause
agamdua/rply,agamdua/rply
class BaseBox(object): pass class Token(BaseBox): def __init__(self, name, value, source_pos=None): BaseBox.__init__(self) self.name = name self.value = value self.source_pos = source_pos def __eq__(self, other): return self.name == other.name and self.value == other.value def gettokentype(self): return self.name def getsourcepos(self): return self.source_pos def getstr(self): return self.value class SourcePosition(object): def __init__(self, idx, lineno, colno): self.idx = idx self.lineno = lineno self.colno = colno Drop the __init__ call to object.__init__, RPython doesn't like it and it doesn't doa nything
class BaseBox(object): pass class Token(BaseBox): def __init__(self, name, value, source_pos=None): self.name = name self.value = value self.source_pos = source_pos def __eq__(self, other): return self.name == other.name and self.value == other.value def gettokentype(self): return self.name def getsourcepos(self): return self.source_pos def getstr(self): return self.value class SourcePosition(object): def __init__(self, idx, lineno, colno): self.idx = idx self.lineno = lineno self.colno = colno
<commit_before>class BaseBox(object): pass class Token(BaseBox): def __init__(self, name, value, source_pos=None): BaseBox.__init__(self) self.name = name self.value = value self.source_pos = source_pos def __eq__(self, other): return self.name == other.name and self.value == other.value def gettokentype(self): return self.name def getsourcepos(self): return self.source_pos def getstr(self): return self.value class SourcePosition(object): def __init__(self, idx, lineno, colno): self.idx = idx self.lineno = lineno self.colno = colno <commit_msg>Drop the __init__ call to object.__init__, RPython doesn't like it and it doesn't doa nything<commit_after>
class BaseBox(object): pass class Token(BaseBox): def __init__(self, name, value, source_pos=None): self.name = name self.value = value self.source_pos = source_pos def __eq__(self, other): return self.name == other.name and self.value == other.value def gettokentype(self): return self.name def getsourcepos(self): return self.source_pos def getstr(self): return self.value class SourcePosition(object): def __init__(self, idx, lineno, colno): self.idx = idx self.lineno = lineno self.colno = colno
class BaseBox(object): pass class Token(BaseBox): def __init__(self, name, value, source_pos=None): BaseBox.__init__(self) self.name = name self.value = value self.source_pos = source_pos def __eq__(self, other): return self.name == other.name and self.value == other.value def gettokentype(self): return self.name def getsourcepos(self): return self.source_pos def getstr(self): return self.value class SourcePosition(object): def __init__(self, idx, lineno, colno): self.idx = idx self.lineno = lineno self.colno = colno Drop the __init__ call to object.__init__, RPython doesn't like it and it doesn't doa nythingclass BaseBox(object): pass class Token(BaseBox): def __init__(self, name, value, source_pos=None): self.name = name self.value = value self.source_pos = source_pos def __eq__(self, other): return self.name == other.name and self.value == other.value def gettokentype(self): return self.name def getsourcepos(self): return self.source_pos def getstr(self): return self.value class SourcePosition(object): def __init__(self, idx, lineno, colno): self.idx = idx self.lineno = lineno self.colno = colno
<commit_before>class BaseBox(object): pass class Token(BaseBox): def __init__(self, name, value, source_pos=None): BaseBox.__init__(self) self.name = name self.value = value self.source_pos = source_pos def __eq__(self, other): return self.name == other.name and self.value == other.value def gettokentype(self): return self.name def getsourcepos(self): return self.source_pos def getstr(self): return self.value class SourcePosition(object): def __init__(self, idx, lineno, colno): self.idx = idx self.lineno = lineno self.colno = colno <commit_msg>Drop the __init__ call to object.__init__, RPython doesn't like it and it doesn't doa nything<commit_after>class BaseBox(object): pass class Token(BaseBox): def __init__(self, name, value, source_pos=None): self.name = name self.value = value self.source_pos = source_pos def __eq__(self, other): return self.name == other.name and self.value == other.value def gettokentype(self): return self.name def getsourcepos(self): return self.source_pos def getstr(self): return self.value class SourcePosition(object): def __init__(self, idx, lineno, colno): self.idx = idx self.lineno = lineno self.colno = colno
be53f1234bec0bca4c35f020905e24d0637b91e3
tests/run/coroutines.py
tests/run/coroutines.py
# cython: language_level=3 # mode: run # tag: pep492, pure3.5 async def test_coroutine_frame(awaitable): """ >>> class Awaitable(object): ... def __await__(self): ... return iter([2]) >>> coro = test_coroutine_frame(Awaitable()) >>> import types >>> isinstance(coro.cr_frame, types.FrameType) or coro.cr_frame True >>> coro.cr_frame is coro.cr_frame # assert that it's cached True >>> coro.cr_frame.f_code is not None True >>> code_obj = coro.cr_frame.f_code >>> code_obj.co_argcount 1 >>> code_obj.co_varnames ('awaitable', 'b') >>> next(coro.__await__()) # avoid "not awaited" warning 2 """ b = await awaitable return b
# cython: language_level=3 # mode: run # tag: pep492, pure3.5, gh1462, async, await async def test_coroutine_frame(awaitable): """ >>> class Awaitable(object): ... def __await__(self): ... return iter([2]) >>> coro = test_coroutine_frame(Awaitable()) >>> import types >>> isinstance(coro.cr_frame, types.FrameType) or coro.cr_frame True >>> coro.cr_frame is coro.cr_frame # assert that it's cached True >>> coro.cr_frame.f_code is not None True >>> code_obj = coro.cr_frame.f_code >>> code_obj.co_argcount 1 >>> code_obj.co_varnames ('awaitable', 'b') >>> next(coro.__await__()) # avoid "not awaited" warning 2 """ b = await awaitable return b # gh1462: Using decorators on coroutines. def pass_through(func): return func @pass_through async def test_pass_through(): """ >>> t = test_pass_through() >>> try: t.send(None) ... except StopIteration as ex: ... print(ex.args[0] if ex.args else None) ... else: print("NOT STOPPED!") None """ @pass_through(pass_through) async def test_pass_through_with_args(): """ >>> t = test_pass_through_with_args() >>> try: t.send(None) ... except StopIteration as ex: ... print(ex.args[0] if ex.args else None) ... else: print("NOT STOPPED!") None """
Add an explicit test for async-def functions with decorators. Closes GH-1462.
Add an explicit test for async-def functions with decorators. Closes GH-1462.
Python
apache-2.0
scoder/cython,da-woods/cython,scoder/cython,scoder/cython,cython/cython,da-woods/cython,cython/cython,cython/cython,da-woods/cython,scoder/cython,cython/cython,da-woods/cython
# cython: language_level=3 # mode: run # tag: pep492, pure3.5 async def test_coroutine_frame(awaitable): """ >>> class Awaitable(object): ... def __await__(self): ... return iter([2]) >>> coro = test_coroutine_frame(Awaitable()) >>> import types >>> isinstance(coro.cr_frame, types.FrameType) or coro.cr_frame True >>> coro.cr_frame is coro.cr_frame # assert that it's cached True >>> coro.cr_frame.f_code is not None True >>> code_obj = coro.cr_frame.f_code >>> code_obj.co_argcount 1 >>> code_obj.co_varnames ('awaitable', 'b') >>> next(coro.__await__()) # avoid "not awaited" warning 2 """ b = await awaitable return b Add an explicit test for async-def functions with decorators. Closes GH-1462.
# cython: language_level=3 # mode: run # tag: pep492, pure3.5, gh1462, async, await async def test_coroutine_frame(awaitable): """ >>> class Awaitable(object): ... def __await__(self): ... return iter([2]) >>> coro = test_coroutine_frame(Awaitable()) >>> import types >>> isinstance(coro.cr_frame, types.FrameType) or coro.cr_frame True >>> coro.cr_frame is coro.cr_frame # assert that it's cached True >>> coro.cr_frame.f_code is not None True >>> code_obj = coro.cr_frame.f_code >>> code_obj.co_argcount 1 >>> code_obj.co_varnames ('awaitable', 'b') >>> next(coro.__await__()) # avoid "not awaited" warning 2 """ b = await awaitable return b # gh1462: Using decorators on coroutines. def pass_through(func): return func @pass_through async def test_pass_through(): """ >>> t = test_pass_through() >>> try: t.send(None) ... except StopIteration as ex: ... print(ex.args[0] if ex.args else None) ... else: print("NOT STOPPED!") None """ @pass_through(pass_through) async def test_pass_through_with_args(): """ >>> t = test_pass_through_with_args() >>> try: t.send(None) ... except StopIteration as ex: ... print(ex.args[0] if ex.args else None) ... else: print("NOT STOPPED!") None """
<commit_before># cython: language_level=3 # mode: run # tag: pep492, pure3.5 async def test_coroutine_frame(awaitable): """ >>> class Awaitable(object): ... def __await__(self): ... return iter([2]) >>> coro = test_coroutine_frame(Awaitable()) >>> import types >>> isinstance(coro.cr_frame, types.FrameType) or coro.cr_frame True >>> coro.cr_frame is coro.cr_frame # assert that it's cached True >>> coro.cr_frame.f_code is not None True >>> code_obj = coro.cr_frame.f_code >>> code_obj.co_argcount 1 >>> code_obj.co_varnames ('awaitable', 'b') >>> next(coro.__await__()) # avoid "not awaited" warning 2 """ b = await awaitable return b <commit_msg>Add an explicit test for async-def functions with decorators. Closes GH-1462.<commit_after>
# cython: language_level=3 # mode: run # tag: pep492, pure3.5, gh1462, async, await async def test_coroutine_frame(awaitable): """ >>> class Awaitable(object): ... def __await__(self): ... return iter([2]) >>> coro = test_coroutine_frame(Awaitable()) >>> import types >>> isinstance(coro.cr_frame, types.FrameType) or coro.cr_frame True >>> coro.cr_frame is coro.cr_frame # assert that it's cached True >>> coro.cr_frame.f_code is not None True >>> code_obj = coro.cr_frame.f_code >>> code_obj.co_argcount 1 >>> code_obj.co_varnames ('awaitable', 'b') >>> next(coro.__await__()) # avoid "not awaited" warning 2 """ b = await awaitable return b # gh1462: Using decorators on coroutines. def pass_through(func): return func @pass_through async def test_pass_through(): """ >>> t = test_pass_through() >>> try: t.send(None) ... except StopIteration as ex: ... print(ex.args[0] if ex.args else None) ... else: print("NOT STOPPED!") None """ @pass_through(pass_through) async def test_pass_through_with_args(): """ >>> t = test_pass_through_with_args() >>> try: t.send(None) ... except StopIteration as ex: ... print(ex.args[0] if ex.args else None) ... else: print("NOT STOPPED!") None """
# cython: language_level=3 # mode: run # tag: pep492, pure3.5 async def test_coroutine_frame(awaitable): """ >>> class Awaitable(object): ... def __await__(self): ... return iter([2]) >>> coro = test_coroutine_frame(Awaitable()) >>> import types >>> isinstance(coro.cr_frame, types.FrameType) or coro.cr_frame True >>> coro.cr_frame is coro.cr_frame # assert that it's cached True >>> coro.cr_frame.f_code is not None True >>> code_obj = coro.cr_frame.f_code >>> code_obj.co_argcount 1 >>> code_obj.co_varnames ('awaitable', 'b') >>> next(coro.__await__()) # avoid "not awaited" warning 2 """ b = await awaitable return b Add an explicit test for async-def functions with decorators. Closes GH-1462.# cython: language_level=3 # mode: run # tag: pep492, pure3.5, gh1462, async, await async def test_coroutine_frame(awaitable): """ >>> class Awaitable(object): ... def __await__(self): ... return iter([2]) >>> coro = test_coroutine_frame(Awaitable()) >>> import types >>> isinstance(coro.cr_frame, types.FrameType) or coro.cr_frame True >>> coro.cr_frame is coro.cr_frame # assert that it's cached True >>> coro.cr_frame.f_code is not None True >>> code_obj = coro.cr_frame.f_code >>> code_obj.co_argcount 1 >>> code_obj.co_varnames ('awaitable', 'b') >>> next(coro.__await__()) # avoid "not awaited" warning 2 """ b = await awaitable return b # gh1462: Using decorators on coroutines. def pass_through(func): return func @pass_through async def test_pass_through(): """ >>> t = test_pass_through() >>> try: t.send(None) ... except StopIteration as ex: ... print(ex.args[0] if ex.args else None) ... else: print("NOT STOPPED!") None """ @pass_through(pass_through) async def test_pass_through_with_args(): """ >>> t = test_pass_through_with_args() >>> try: t.send(None) ... except StopIteration as ex: ... print(ex.args[0] if ex.args else None) ... else: print("NOT STOPPED!") None """
<commit_before># cython: language_level=3 # mode: run # tag: pep492, pure3.5 async def test_coroutine_frame(awaitable): """ >>> class Awaitable(object): ... def __await__(self): ... return iter([2]) >>> coro = test_coroutine_frame(Awaitable()) >>> import types >>> isinstance(coro.cr_frame, types.FrameType) or coro.cr_frame True >>> coro.cr_frame is coro.cr_frame # assert that it's cached True >>> coro.cr_frame.f_code is not None True >>> code_obj = coro.cr_frame.f_code >>> code_obj.co_argcount 1 >>> code_obj.co_varnames ('awaitable', 'b') >>> next(coro.__await__()) # avoid "not awaited" warning 2 """ b = await awaitable return b <commit_msg>Add an explicit test for async-def functions with decorators. Closes GH-1462.<commit_after># cython: language_level=3 # mode: run # tag: pep492, pure3.5, gh1462, async, await async def test_coroutine_frame(awaitable): """ >>> class Awaitable(object): ... def __await__(self): ... return iter([2]) >>> coro = test_coroutine_frame(Awaitable()) >>> import types >>> isinstance(coro.cr_frame, types.FrameType) or coro.cr_frame True >>> coro.cr_frame is coro.cr_frame # assert that it's cached True >>> coro.cr_frame.f_code is not None True >>> code_obj = coro.cr_frame.f_code >>> code_obj.co_argcount 1 >>> code_obj.co_varnames ('awaitable', 'b') >>> next(coro.__await__()) # avoid "not awaited" warning 2 """ b = await awaitable return b # gh1462: Using decorators on coroutines. def pass_through(func): return func @pass_through async def test_pass_through(): """ >>> t = test_pass_through() >>> try: t.send(None) ... except StopIteration as ex: ... print(ex.args[0] if ex.args else None) ... else: print("NOT STOPPED!") None """ @pass_through(pass_through) async def test_pass_through_with_args(): """ >>> t = test_pass_through_with_args() >>> try: t.send(None) ... except StopIteration as ex: ... print(ex.args[0] if ex.args else None) ... else: print("NOT STOPPED!") None """
fa78cd1b3aa29cfe2846f4a999b4bb7436b339ea
tests/test_responses.py
tests/test_responses.py
from jsonrpcclient.responses import Ok, parse, parse_json def test_parse(): assert parse({"jsonrpc": "2.0", "result": "pong", "id": 1}) == Ok("pong", 1) def test_parse_json(): assert parse_json('{"jsonrpc": "2.0", "result": "pong", "id": 1}') == Ok("pong", 1)
from jsonrpcclient.responses import Error, Ok, parse, parse_json def test_Ok(): assert repr(Ok("foo", 1)) == "Ok(result='foo', id=1)" def test_Error(): assert ( repr(Error(1, "foo", "bar", 2)) == "Error(code=1, message='foo', data='bar', id=2)" ) def test_parse(): assert parse({"jsonrpc": "2.0", "result": "pong", "id": 1}) == Ok("pong", 1) def test_parse_json(): assert parse_json('{"jsonrpc": "2.0", "result": "pong", "id": 1}') == Ok("pong", 1)
Bring code coverage to 100%
Bring code coverage to 100%
Python
mit
bcb/jsonrpcclient
from jsonrpcclient.responses import Ok, parse, parse_json def test_parse(): assert parse({"jsonrpc": "2.0", "result": "pong", "id": 1}) == Ok("pong", 1) def test_parse_json(): assert parse_json('{"jsonrpc": "2.0", "result": "pong", "id": 1}') == Ok("pong", 1) Bring code coverage to 100%
from jsonrpcclient.responses import Error, Ok, parse, parse_json def test_Ok(): assert repr(Ok("foo", 1)) == "Ok(result='foo', id=1)" def test_Error(): assert ( repr(Error(1, "foo", "bar", 2)) == "Error(code=1, message='foo', data='bar', id=2)" ) def test_parse(): assert parse({"jsonrpc": "2.0", "result": "pong", "id": 1}) == Ok("pong", 1) def test_parse_json(): assert parse_json('{"jsonrpc": "2.0", "result": "pong", "id": 1}') == Ok("pong", 1)
<commit_before>from jsonrpcclient.responses import Ok, parse, parse_json def test_parse(): assert parse({"jsonrpc": "2.0", "result": "pong", "id": 1}) == Ok("pong", 1) def test_parse_json(): assert parse_json('{"jsonrpc": "2.0", "result": "pong", "id": 1}') == Ok("pong", 1) <commit_msg>Bring code coverage to 100%<commit_after>
from jsonrpcclient.responses import Error, Ok, parse, parse_json def test_Ok(): assert repr(Ok("foo", 1)) == "Ok(result='foo', id=1)" def test_Error(): assert ( repr(Error(1, "foo", "bar", 2)) == "Error(code=1, message='foo', data='bar', id=2)" ) def test_parse(): assert parse({"jsonrpc": "2.0", "result": "pong", "id": 1}) == Ok("pong", 1) def test_parse_json(): assert parse_json('{"jsonrpc": "2.0", "result": "pong", "id": 1}') == Ok("pong", 1)
from jsonrpcclient.responses import Ok, parse, parse_json def test_parse(): assert parse({"jsonrpc": "2.0", "result": "pong", "id": 1}) == Ok("pong", 1) def test_parse_json(): assert parse_json('{"jsonrpc": "2.0", "result": "pong", "id": 1}') == Ok("pong", 1) Bring code coverage to 100%from jsonrpcclient.responses import Error, Ok, parse, parse_json def test_Ok(): assert repr(Ok("foo", 1)) == "Ok(result='foo', id=1)" def test_Error(): assert ( repr(Error(1, "foo", "bar", 2)) == "Error(code=1, message='foo', data='bar', id=2)" ) def test_parse(): assert parse({"jsonrpc": "2.0", "result": "pong", "id": 1}) == Ok("pong", 1) def test_parse_json(): assert parse_json('{"jsonrpc": "2.0", "result": "pong", "id": 1}') == Ok("pong", 1)
<commit_before>from jsonrpcclient.responses import Ok, parse, parse_json def test_parse(): assert parse({"jsonrpc": "2.0", "result": "pong", "id": 1}) == Ok("pong", 1) def test_parse_json(): assert parse_json('{"jsonrpc": "2.0", "result": "pong", "id": 1}') == Ok("pong", 1) <commit_msg>Bring code coverage to 100%<commit_after>from jsonrpcclient.responses import Error, Ok, parse, parse_json def test_Ok(): assert repr(Ok("foo", 1)) == "Ok(result='foo', id=1)" def test_Error(): assert ( repr(Error(1, "foo", "bar", 2)) == "Error(code=1, message='foo', data='bar', id=2)" ) def test_parse(): assert parse({"jsonrpc": "2.0", "result": "pong", "id": 1}) == Ok("pong", 1) def test_parse_json(): assert parse_json('{"jsonrpc": "2.0", "result": "pong", "id": 1}') == Ok("pong", 1)
9dfe31f52d1cf4dfb11a1ffd8c14274e4b9ec135
tests/test_tokenizer.py
tests/test_tokenizer.py
import unittest from halng.tokenizer import MegaHALTokenizer class testMegaHALTokenizer(unittest.TestCase): def setUp(self): self.tokenizer = MegaHALTokenizer() def testSplitEmpty(self): self.assertEquals(len(self.tokenizer.split("")), 0) def testSplitSentence(self): words = self.tokenizer.split("hi.") self.assertEquals(len(words), 2) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ".") def testSplitImplicitStop(self): words = self.tokenizer.split("hi") self.assertEquals(len(words), 2) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ".") def testSplitUrl(self): words = self.tokenizer.split("http://www.google.com/") self.assertEquals(len(words), 8) self.assertEquals(words[0], "HTTP") self.assertEquals(words[1], "://") self.assertEquals(words[2], "WWW") self.assertEquals(words[3], ".") self.assertEquals(words[4], "GOOGLE") self.assertEquals(words[5], ".") self.assertEquals(words[6], "COM") self.assertEquals(words[7], "/.") if __name__ == '__main__': unittest.main()
import unittest from halng.tokenizer import MegaHALTokenizer class testMegaHALTokenizer(unittest.TestCase): def setUp(self): self.tokenizer = MegaHALTokenizer() def testSplitEmpty(self): self.assertEquals(len(self.tokenizer.split("")), 0) def testSplitSentence(self): words = self.tokenizer.split("hi.") self.assertEquals(len(words), 2) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ".") def testSplitComma(self): words = self.tokenizer.split("hi, hal") self.assertEquals(len(words), 4) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ", ") self.assertEquals(words[2], "HAL") self.assertEquals(words[3], ".") def testSplitImplicitStop(self): words = self.tokenizer.split("hi") self.assertEquals(len(words), 2) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ".") def testSplitUrl(self): words = self.tokenizer.split("http://www.google.com/") self.assertEquals(len(words), 8) self.assertEquals(words[0], "HTTP") self.assertEquals(words[1], "://") self.assertEquals(words[2], "WWW") self.assertEquals(words[3], ".") self.assertEquals(words[4], "GOOGLE") self.assertEquals(words[5], ".") self.assertEquals(words[6], "COM") self.assertEquals(words[7], "/.") if __name__ == '__main__': unittest.main()
Add a test that ensures commas are part of non-word runs.
Add a test that ensures commas are part of non-word runs.
Python
mit
meska/cobe,wodim/cobe-ng,meska/cobe,tiagochiavericosta/cobe,DarkMio/cobe,LeMagnesium/cobe,tiagochiavericosta/cobe,pteichman/cobe,pteichman/cobe,LeMagnesium/cobe,DarkMio/cobe,wodim/cobe-ng
import unittest from halng.tokenizer import MegaHALTokenizer class testMegaHALTokenizer(unittest.TestCase): def setUp(self): self.tokenizer = MegaHALTokenizer() def testSplitEmpty(self): self.assertEquals(len(self.tokenizer.split("")), 0) def testSplitSentence(self): words = self.tokenizer.split("hi.") self.assertEquals(len(words), 2) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ".") def testSplitImplicitStop(self): words = self.tokenizer.split("hi") self.assertEquals(len(words), 2) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ".") def testSplitUrl(self): words = self.tokenizer.split("http://www.google.com/") self.assertEquals(len(words), 8) self.assertEquals(words[0], "HTTP") self.assertEquals(words[1], "://") self.assertEquals(words[2], "WWW") self.assertEquals(words[3], ".") self.assertEquals(words[4], "GOOGLE") self.assertEquals(words[5], ".") self.assertEquals(words[6], "COM") self.assertEquals(words[7], "/.") if __name__ == '__main__': unittest.main() Add a test that ensures commas are part of non-word runs.
import unittest from halng.tokenizer import MegaHALTokenizer class testMegaHALTokenizer(unittest.TestCase): def setUp(self): self.tokenizer = MegaHALTokenizer() def testSplitEmpty(self): self.assertEquals(len(self.tokenizer.split("")), 0) def testSplitSentence(self): words = self.tokenizer.split("hi.") self.assertEquals(len(words), 2) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ".") def testSplitComma(self): words = self.tokenizer.split("hi, hal") self.assertEquals(len(words), 4) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ", ") self.assertEquals(words[2], "HAL") self.assertEquals(words[3], ".") def testSplitImplicitStop(self): words = self.tokenizer.split("hi") self.assertEquals(len(words), 2) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ".") def testSplitUrl(self): words = self.tokenizer.split("http://www.google.com/") self.assertEquals(len(words), 8) self.assertEquals(words[0], "HTTP") self.assertEquals(words[1], "://") self.assertEquals(words[2], "WWW") self.assertEquals(words[3], ".") self.assertEquals(words[4], "GOOGLE") self.assertEquals(words[5], ".") self.assertEquals(words[6], "COM") self.assertEquals(words[7], "/.") if __name__ == '__main__': unittest.main()
<commit_before>import unittest from halng.tokenizer import MegaHALTokenizer class testMegaHALTokenizer(unittest.TestCase): def setUp(self): self.tokenizer = MegaHALTokenizer() def testSplitEmpty(self): self.assertEquals(len(self.tokenizer.split("")), 0) def testSplitSentence(self): words = self.tokenizer.split("hi.") self.assertEquals(len(words), 2) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ".") def testSplitImplicitStop(self): words = self.tokenizer.split("hi") self.assertEquals(len(words), 2) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ".") def testSplitUrl(self): words = self.tokenizer.split("http://www.google.com/") self.assertEquals(len(words), 8) self.assertEquals(words[0], "HTTP") self.assertEquals(words[1], "://") self.assertEquals(words[2], "WWW") self.assertEquals(words[3], ".") self.assertEquals(words[4], "GOOGLE") self.assertEquals(words[5], ".") self.assertEquals(words[6], "COM") self.assertEquals(words[7], "/.") if __name__ == '__main__': unittest.main() <commit_msg>Add a test that ensures commas are part of non-word runs.<commit_after>
import unittest from halng.tokenizer import MegaHALTokenizer class testMegaHALTokenizer(unittest.TestCase): def setUp(self): self.tokenizer = MegaHALTokenizer() def testSplitEmpty(self): self.assertEquals(len(self.tokenizer.split("")), 0) def testSplitSentence(self): words = self.tokenizer.split("hi.") self.assertEquals(len(words), 2) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ".") def testSplitComma(self): words = self.tokenizer.split("hi, hal") self.assertEquals(len(words), 4) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ", ") self.assertEquals(words[2], "HAL") self.assertEquals(words[3], ".") def testSplitImplicitStop(self): words = self.tokenizer.split("hi") self.assertEquals(len(words), 2) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ".") def testSplitUrl(self): words = self.tokenizer.split("http://www.google.com/") self.assertEquals(len(words), 8) self.assertEquals(words[0], "HTTP") self.assertEquals(words[1], "://") self.assertEquals(words[2], "WWW") self.assertEquals(words[3], ".") self.assertEquals(words[4], "GOOGLE") self.assertEquals(words[5], ".") self.assertEquals(words[6], "COM") self.assertEquals(words[7], "/.") if __name__ == '__main__': unittest.main()
import unittest from halng.tokenizer import MegaHALTokenizer class testMegaHALTokenizer(unittest.TestCase): def setUp(self): self.tokenizer = MegaHALTokenizer() def testSplitEmpty(self): self.assertEquals(len(self.tokenizer.split("")), 0) def testSplitSentence(self): words = self.tokenizer.split("hi.") self.assertEquals(len(words), 2) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ".") def testSplitImplicitStop(self): words = self.tokenizer.split("hi") self.assertEquals(len(words), 2) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ".") def testSplitUrl(self): words = self.tokenizer.split("http://www.google.com/") self.assertEquals(len(words), 8) self.assertEquals(words[0], "HTTP") self.assertEquals(words[1], "://") self.assertEquals(words[2], "WWW") self.assertEquals(words[3], ".") self.assertEquals(words[4], "GOOGLE") self.assertEquals(words[5], ".") self.assertEquals(words[6], "COM") self.assertEquals(words[7], "/.") if __name__ == '__main__': unittest.main() Add a test that ensures commas are part of non-word runs.import unittest from halng.tokenizer import MegaHALTokenizer class testMegaHALTokenizer(unittest.TestCase): def setUp(self): self.tokenizer = MegaHALTokenizer() def testSplitEmpty(self): self.assertEquals(len(self.tokenizer.split("")), 0) def testSplitSentence(self): words = self.tokenizer.split("hi.") self.assertEquals(len(words), 2) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ".") def testSplitComma(self): words = self.tokenizer.split("hi, hal") self.assertEquals(len(words), 4) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ", ") self.assertEquals(words[2], "HAL") self.assertEquals(words[3], ".") def testSplitImplicitStop(self): words = self.tokenizer.split("hi") self.assertEquals(len(words), 2) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ".") def testSplitUrl(self): words = self.tokenizer.split("http://www.google.com/") self.assertEquals(len(words), 8) self.assertEquals(words[0], "HTTP") self.assertEquals(words[1], "://") self.assertEquals(words[2], "WWW") self.assertEquals(words[3], ".") self.assertEquals(words[4], "GOOGLE") self.assertEquals(words[5], ".") self.assertEquals(words[6], "COM") self.assertEquals(words[7], "/.") if __name__ == '__main__': unittest.main()
<commit_before>import unittest from halng.tokenizer import MegaHALTokenizer class testMegaHALTokenizer(unittest.TestCase): def setUp(self): self.tokenizer = MegaHALTokenizer() def testSplitEmpty(self): self.assertEquals(len(self.tokenizer.split("")), 0) def testSplitSentence(self): words = self.tokenizer.split("hi.") self.assertEquals(len(words), 2) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ".") def testSplitImplicitStop(self): words = self.tokenizer.split("hi") self.assertEquals(len(words), 2) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ".") def testSplitUrl(self): words = self.tokenizer.split("http://www.google.com/") self.assertEquals(len(words), 8) self.assertEquals(words[0], "HTTP") self.assertEquals(words[1], "://") self.assertEquals(words[2], "WWW") self.assertEquals(words[3], ".") self.assertEquals(words[4], "GOOGLE") self.assertEquals(words[5], ".") self.assertEquals(words[6], "COM") self.assertEquals(words[7], "/.") if __name__ == '__main__': unittest.main() <commit_msg>Add a test that ensures commas are part of non-word runs.<commit_after>import unittest from halng.tokenizer import MegaHALTokenizer class testMegaHALTokenizer(unittest.TestCase): def setUp(self): self.tokenizer = MegaHALTokenizer() def testSplitEmpty(self): self.assertEquals(len(self.tokenizer.split("")), 0) def testSplitSentence(self): words = self.tokenizer.split("hi.") self.assertEquals(len(words), 2) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ".") def testSplitComma(self): words = self.tokenizer.split("hi, hal") self.assertEquals(len(words), 4) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ", ") self.assertEquals(words[2], "HAL") self.assertEquals(words[3], ".") def testSplitImplicitStop(self): words = self.tokenizer.split("hi") self.assertEquals(len(words), 2) self.assertEquals(words[0], "HI") self.assertEquals(words[1], ".") def testSplitUrl(self): words = self.tokenizer.split("http://www.google.com/") self.assertEquals(len(words), 8) self.assertEquals(words[0], "HTTP") self.assertEquals(words[1], "://") self.assertEquals(words[2], "WWW") self.assertEquals(words[3], ".") self.assertEquals(words[4], "GOOGLE") self.assertEquals(words[5], ".") self.assertEquals(words[6], "COM") self.assertEquals(words[7], "/.") if __name__ == '__main__': unittest.main()
9ffcae90963ed97136142bdd1443f633f11a5837
settings.py
settings.py
from fabric.api import env env.hosts = ['zygal@kontar.kattare.com'] env.local_dir = 'public' env.remote_dir = 'temp'
from fabric.api import env env.hosts = ['myuser@mysite.com'] env.local_dir = 'public' env.remote_dir = 'temp'
Set some default hosts environment
Set some default hosts environment
Python
mit
nfletton/froid,nfletton/froid
from fabric.api import env env.hosts = ['zygal@kontar.kattare.com'] env.local_dir = 'public' env.remote_dir = 'temp' Set some default hosts environment
from fabric.api import env env.hosts = ['myuser@mysite.com'] env.local_dir = 'public' env.remote_dir = 'temp'
<commit_before>from fabric.api import env env.hosts = ['zygal@kontar.kattare.com'] env.local_dir = 'public' env.remote_dir = 'temp' <commit_msg>Set some default hosts environment<commit_after>
from fabric.api import env env.hosts = ['myuser@mysite.com'] env.local_dir = 'public' env.remote_dir = 'temp'
from fabric.api import env env.hosts = ['zygal@kontar.kattare.com'] env.local_dir = 'public' env.remote_dir = 'temp' Set some default hosts environmentfrom fabric.api import env env.hosts = ['myuser@mysite.com'] env.local_dir = 'public' env.remote_dir = 'temp'
<commit_before>from fabric.api import env env.hosts = ['zygal@kontar.kattare.com'] env.local_dir = 'public' env.remote_dir = 'temp' <commit_msg>Set some default hosts environment<commit_after>from fabric.api import env env.hosts = ['myuser@mysite.com'] env.local_dir = 'public' env.remote_dir = 'temp'
03c7f149ac0162a78892593d33b5866a1a9b72df
tests/test_settings.py
tests/test_settings.py
from __future__ import unicode_literals from django.test import TestCase from rest_framework.settings import APISettings class TestSettings(TestCase): def test_import_error_message_maintained(self): """ Make sure import errors are captured and raised sensibly. """ settings = APISettings({ 'DEFAULT_RENDERER_CLASSES': [ 'tests.invalid_module.InvalidClassName' ] }) with self.assertRaises(ImportError): settings.DEFAULT_RENDERER_CLASSES class TestSettingTypes(TestCase): def test_settings_consistently_coerced_to_list(self): settings = APISettings({ 'DEFAULT_THROTTLE_CLASSES': ('rest_framework.throttling.BaseThrottle',) }) self.assertTrue(isinstance(settings.DEFAULT_THROTTLE_CLASSES, list)) settings = APISettings({ 'DEFAULT_THROTTLE_CLASSES': () }) self.assertTrue(isinstance(settings.DEFAULT_THROTTLE_CLASSES, list))
from __future__ import unicode_literals from django.test import TestCase from rest_framework.settings import APISettings class TestSettings(TestCase): def test_import_error_message_maintained(self): """ Make sure import errors are captured and raised sensibly. """ settings = APISettings({ 'DEFAULT_RENDERER_CLASSES': [ 'tests.invalid_module.InvalidClassName' ] }) with self.assertRaises(ImportError): settings.DEFAULT_RENDERER_CLASSES def test_loud_error_raised_on_removed_setting(self): """ Make sure user is alerted with an error when a removed setting is set. """ with self.asserRaise(AttributeError): APISettings({ 'MAX_PAGINATE_BY': 100 }) class TestSettingTypes(TestCase): def test_settings_consistently_coerced_to_list(self): settings = APISettings({ 'DEFAULT_THROTTLE_CLASSES': ('rest_framework.throttling.BaseThrottle',) }) self.assertTrue(isinstance(settings.DEFAULT_THROTTLE_CLASSES, list)) settings = APISettings({ 'DEFAULT_THROTTLE_CLASSES': () }) self.assertTrue(isinstance(settings.DEFAULT_THROTTLE_CLASSES, list))
Test case for settings check
Test case for settings check
Python
bsd-2-clause
davesque/django-rest-framework,dmwyatt/django-rest-framework,jpadilla/django-rest-framework,kgeorgy/django-rest-framework,atombrella/django-rest-framework,davesque/django-rest-framework,pombredanne/django-rest-framework,cyberj/django-rest-framework,ossanna16/django-rest-framework,dmwyatt/django-rest-framework,edx/django-rest-framework,johnraz/django-rest-framework,agconti/django-rest-framework,davesque/django-rest-framework,callorico/django-rest-framework,tomchristie/django-rest-framework,edx/django-rest-framework,uploadcare/django-rest-framework,pombredanne/django-rest-framework,pombredanne/django-rest-framework,sheppard/django-rest-framework,tomchristie/django-rest-framework,kgeorgy/django-rest-framework,callorico/django-rest-framework,linovia/django-rest-framework,edx/django-rest-framework,jpadilla/django-rest-framework,kgeorgy/django-rest-framework,agconti/django-rest-framework,tomchristie/django-rest-framework,jpadilla/django-rest-framework,johnraz/django-rest-framework,ossanna16/django-rest-framework,cyberj/django-rest-framework,rhblind/django-rest-framework,atombrella/django-rest-framework,werthen/django-rest-framework,linovia/django-rest-framework,uploadcare/django-rest-framework,rhblind/django-rest-framework,cyberj/django-rest-framework,werthen/django-rest-framework,ossanna16/django-rest-framework,sheppard/django-rest-framework,dmwyatt/django-rest-framework,linovia/django-rest-framework,rhblind/django-rest-framework,werthen/django-rest-framework,sheppard/django-rest-framework,callorico/django-rest-framework,atombrella/django-rest-framework,johnraz/django-rest-framework,agconti/django-rest-framework,uploadcare/django-rest-framework
from __future__ import unicode_literals from django.test import TestCase from rest_framework.settings import APISettings class TestSettings(TestCase): def test_import_error_message_maintained(self): """ Make sure import errors are captured and raised sensibly. """ settings = APISettings({ 'DEFAULT_RENDERER_CLASSES': [ 'tests.invalid_module.InvalidClassName' ] }) with self.assertRaises(ImportError): settings.DEFAULT_RENDERER_CLASSES class TestSettingTypes(TestCase): def test_settings_consistently_coerced_to_list(self): settings = APISettings({ 'DEFAULT_THROTTLE_CLASSES': ('rest_framework.throttling.BaseThrottle',) }) self.assertTrue(isinstance(settings.DEFAULT_THROTTLE_CLASSES, list)) settings = APISettings({ 'DEFAULT_THROTTLE_CLASSES': () }) self.assertTrue(isinstance(settings.DEFAULT_THROTTLE_CLASSES, list)) Test case for settings check
from __future__ import unicode_literals from django.test import TestCase from rest_framework.settings import APISettings class TestSettings(TestCase): def test_import_error_message_maintained(self): """ Make sure import errors are captured and raised sensibly. """ settings = APISettings({ 'DEFAULT_RENDERER_CLASSES': [ 'tests.invalid_module.InvalidClassName' ] }) with self.assertRaises(ImportError): settings.DEFAULT_RENDERER_CLASSES def test_loud_error_raised_on_removed_setting(self): """ Make sure user is alerted with an error when a removed setting is set. """ with self.asserRaise(AttributeError): APISettings({ 'MAX_PAGINATE_BY': 100 }) class TestSettingTypes(TestCase): def test_settings_consistently_coerced_to_list(self): settings = APISettings({ 'DEFAULT_THROTTLE_CLASSES': ('rest_framework.throttling.BaseThrottle',) }) self.assertTrue(isinstance(settings.DEFAULT_THROTTLE_CLASSES, list)) settings = APISettings({ 'DEFAULT_THROTTLE_CLASSES': () }) self.assertTrue(isinstance(settings.DEFAULT_THROTTLE_CLASSES, list))
<commit_before>from __future__ import unicode_literals from django.test import TestCase from rest_framework.settings import APISettings class TestSettings(TestCase): def test_import_error_message_maintained(self): """ Make sure import errors are captured and raised sensibly. """ settings = APISettings({ 'DEFAULT_RENDERER_CLASSES': [ 'tests.invalid_module.InvalidClassName' ] }) with self.assertRaises(ImportError): settings.DEFAULT_RENDERER_CLASSES class TestSettingTypes(TestCase): def test_settings_consistently_coerced_to_list(self): settings = APISettings({ 'DEFAULT_THROTTLE_CLASSES': ('rest_framework.throttling.BaseThrottle',) }) self.assertTrue(isinstance(settings.DEFAULT_THROTTLE_CLASSES, list)) settings = APISettings({ 'DEFAULT_THROTTLE_CLASSES': () }) self.assertTrue(isinstance(settings.DEFAULT_THROTTLE_CLASSES, list)) <commit_msg>Test case for settings check<commit_after>
from __future__ import unicode_literals from django.test import TestCase from rest_framework.settings import APISettings class TestSettings(TestCase): def test_import_error_message_maintained(self): """ Make sure import errors are captured and raised sensibly. """ settings = APISettings({ 'DEFAULT_RENDERER_CLASSES': [ 'tests.invalid_module.InvalidClassName' ] }) with self.assertRaises(ImportError): settings.DEFAULT_RENDERER_CLASSES def test_loud_error_raised_on_removed_setting(self): """ Make sure user is alerted with an error when a removed setting is set. """ with self.asserRaise(AttributeError): APISettings({ 'MAX_PAGINATE_BY': 100 }) class TestSettingTypes(TestCase): def test_settings_consistently_coerced_to_list(self): settings = APISettings({ 'DEFAULT_THROTTLE_CLASSES': ('rest_framework.throttling.BaseThrottle',) }) self.assertTrue(isinstance(settings.DEFAULT_THROTTLE_CLASSES, list)) settings = APISettings({ 'DEFAULT_THROTTLE_CLASSES': () }) self.assertTrue(isinstance(settings.DEFAULT_THROTTLE_CLASSES, list))
from __future__ import unicode_literals from django.test import TestCase from rest_framework.settings import APISettings class TestSettings(TestCase): def test_import_error_message_maintained(self): """ Make sure import errors are captured and raised sensibly. """ settings = APISettings({ 'DEFAULT_RENDERER_CLASSES': [ 'tests.invalid_module.InvalidClassName' ] }) with self.assertRaises(ImportError): settings.DEFAULT_RENDERER_CLASSES class TestSettingTypes(TestCase): def test_settings_consistently_coerced_to_list(self): settings = APISettings({ 'DEFAULT_THROTTLE_CLASSES': ('rest_framework.throttling.BaseThrottle',) }) self.assertTrue(isinstance(settings.DEFAULT_THROTTLE_CLASSES, list)) settings = APISettings({ 'DEFAULT_THROTTLE_CLASSES': () }) self.assertTrue(isinstance(settings.DEFAULT_THROTTLE_CLASSES, list)) Test case for settings checkfrom __future__ import unicode_literals from django.test import TestCase from rest_framework.settings import APISettings class TestSettings(TestCase): def test_import_error_message_maintained(self): """ Make sure import errors are captured and raised sensibly. """ settings = APISettings({ 'DEFAULT_RENDERER_CLASSES': [ 'tests.invalid_module.InvalidClassName' ] }) with self.assertRaises(ImportError): settings.DEFAULT_RENDERER_CLASSES def test_loud_error_raised_on_removed_setting(self): """ Make sure user is alerted with an error when a removed setting is set. """ with self.asserRaise(AttributeError): APISettings({ 'MAX_PAGINATE_BY': 100 }) class TestSettingTypes(TestCase): def test_settings_consistently_coerced_to_list(self): settings = APISettings({ 'DEFAULT_THROTTLE_CLASSES': ('rest_framework.throttling.BaseThrottle',) }) self.assertTrue(isinstance(settings.DEFAULT_THROTTLE_CLASSES, list)) settings = APISettings({ 'DEFAULT_THROTTLE_CLASSES': () }) self.assertTrue(isinstance(settings.DEFAULT_THROTTLE_CLASSES, list))
<commit_before>from __future__ import unicode_literals from django.test import TestCase from rest_framework.settings import APISettings class TestSettings(TestCase): def test_import_error_message_maintained(self): """ Make sure import errors are captured and raised sensibly. """ settings = APISettings({ 'DEFAULT_RENDERER_CLASSES': [ 'tests.invalid_module.InvalidClassName' ] }) with self.assertRaises(ImportError): settings.DEFAULT_RENDERER_CLASSES class TestSettingTypes(TestCase): def test_settings_consistently_coerced_to_list(self): settings = APISettings({ 'DEFAULT_THROTTLE_CLASSES': ('rest_framework.throttling.BaseThrottle',) }) self.assertTrue(isinstance(settings.DEFAULT_THROTTLE_CLASSES, list)) settings = APISettings({ 'DEFAULT_THROTTLE_CLASSES': () }) self.assertTrue(isinstance(settings.DEFAULT_THROTTLE_CLASSES, list)) <commit_msg>Test case for settings check<commit_after>from __future__ import unicode_literals from django.test import TestCase from rest_framework.settings import APISettings class TestSettings(TestCase): def test_import_error_message_maintained(self): """ Make sure import errors are captured and raised sensibly. """ settings = APISettings({ 'DEFAULT_RENDERER_CLASSES': [ 'tests.invalid_module.InvalidClassName' ] }) with self.assertRaises(ImportError): settings.DEFAULT_RENDERER_CLASSES def test_loud_error_raised_on_removed_setting(self): """ Make sure user is alerted with an error when a removed setting is set. """ with self.asserRaise(AttributeError): APISettings({ 'MAX_PAGINATE_BY': 100 }) class TestSettingTypes(TestCase): def test_settings_consistently_coerced_to_list(self): settings = APISettings({ 'DEFAULT_THROTTLE_CLASSES': ('rest_framework.throttling.BaseThrottle',) }) self.assertTrue(isinstance(settings.DEFAULT_THROTTLE_CLASSES, list)) settings = APISettings({ 'DEFAULT_THROTTLE_CLASSES': () }) self.assertTrue(isinstance(settings.DEFAULT_THROTTLE_CLASSES, list))
fabc8a9cd7f2f23cac3b6cab12eb08fc83875045
vtki/_version.py
vtki/_version.py
""" version info for vtki """ # major, minor, patch version_info = 0, 16, 5 # Nice string for the version __version__ = '.'.join(map(str, version_info))
""" version info for vtki """ # major, minor, patch version_info = 0, 16, 6 # Nice string for the version __version__ = '.'.join(map(str, version_info))
Bump version: 0.16.5 → 0.16.6
Bump version: 0.16.5 → 0.16.6
Python
mit
akaszynski/vtkInterface
""" version info for vtki """ # major, minor, patch version_info = 0, 16, 5 # Nice string for the version __version__ = '.'.join(map(str, version_info)) Bump version: 0.16.5 → 0.16.6
""" version info for vtki """ # major, minor, patch version_info = 0, 16, 6 # Nice string for the version __version__ = '.'.join(map(str, version_info))
<commit_before>""" version info for vtki """ # major, minor, patch version_info = 0, 16, 5 # Nice string for the version __version__ = '.'.join(map(str, version_info)) <commit_msg>Bump version: 0.16.5 → 0.16.6<commit_after>
""" version info for vtki """ # major, minor, patch version_info = 0, 16, 6 # Nice string for the version __version__ = '.'.join(map(str, version_info))
""" version info for vtki """ # major, minor, patch version_info = 0, 16, 5 # Nice string for the version __version__ = '.'.join(map(str, version_info)) Bump version: 0.16.5 → 0.16.6""" version info for vtki """ # major, minor, patch version_info = 0, 16, 6 # Nice string for the version __version__ = '.'.join(map(str, version_info))
<commit_before>""" version info for vtki """ # major, minor, patch version_info = 0, 16, 5 # Nice string for the version __version__ = '.'.join(map(str, version_info)) <commit_msg>Bump version: 0.16.5 → 0.16.6<commit_after>""" version info for vtki """ # major, minor, patch version_info = 0, 16, 6 # Nice string for the version __version__ = '.'.join(map(str, version_info))
4f5e20cc85395312ad66ef3731f2bf7e09987976
commands/alias.py
commands/alias.py
from devbot import chat def call(message, name, protocol, cfg, commands): if message == '': chat.say('/r ' + commands['help']['alias'].format('alias'), protocol) return aliases = '' for tupl in commands['regex'].items(): if tupl[1] == message: aliases = aliases + tupl[0] + ', ' if aliases == '': chat.say('Sorry, there is no command by that name.', protocol) else: chat.say('/r ' + aliases, protocol)
from devbot import chat def call(message, name, protocol, cfg, commands): if message == '': chat.say('/r ' + commands['help']['alias'].format('alias'), protocol) return aliases = '' for tupl in commands['regex'].items(): if tupl[1] == message: aliases = aliases + tupl[0] + ', ' if aliases == '': chat.say('/r Sorry, there is no command by that name.', protocol) else: chat.say('/r ' + aliases, protocol)
Fix bug with DevotedBot speaking error in chat
Fix bug with DevotedBot speaking error in chat
Python
mit
Ameliorate/DevotedBot,Ameliorate/DevotedBot
from devbot import chat def call(message, name, protocol, cfg, commands): if message == '': chat.say('/r ' + commands['help']['alias'].format('alias'), protocol) return aliases = '' for tupl in commands['regex'].items(): if tupl[1] == message: aliases = aliases + tupl[0] + ', ' if aliases == '': chat.say('Sorry, there is no command by that name.', protocol) else: chat.say('/r ' + aliases, protocol) Fix bug with DevotedBot speaking error in chat
from devbot import chat def call(message, name, protocol, cfg, commands): if message == '': chat.say('/r ' + commands['help']['alias'].format('alias'), protocol) return aliases = '' for tupl in commands['regex'].items(): if tupl[1] == message: aliases = aliases + tupl[0] + ', ' if aliases == '': chat.say('/r Sorry, there is no command by that name.', protocol) else: chat.say('/r ' + aliases, protocol)
<commit_before>from devbot import chat def call(message, name, protocol, cfg, commands): if message == '': chat.say('/r ' + commands['help']['alias'].format('alias'), protocol) return aliases = '' for tupl in commands['regex'].items(): if tupl[1] == message: aliases = aliases + tupl[0] + ', ' if aliases == '': chat.say('Sorry, there is no command by that name.', protocol) else: chat.say('/r ' + aliases, protocol) <commit_msg>Fix bug with DevotedBot speaking error in chat<commit_after>
from devbot import chat def call(message, name, protocol, cfg, commands): if message == '': chat.say('/r ' + commands['help']['alias'].format('alias'), protocol) return aliases = '' for tupl in commands['regex'].items(): if tupl[1] == message: aliases = aliases + tupl[0] + ', ' if aliases == '': chat.say('/r Sorry, there is no command by that name.', protocol) else: chat.say('/r ' + aliases, protocol)
from devbot import chat def call(message, name, protocol, cfg, commands): if message == '': chat.say('/r ' + commands['help']['alias'].format('alias'), protocol) return aliases = '' for tupl in commands['regex'].items(): if tupl[1] == message: aliases = aliases + tupl[0] + ', ' if aliases == '': chat.say('Sorry, there is no command by that name.', protocol) else: chat.say('/r ' + aliases, protocol) Fix bug with DevotedBot speaking error in chatfrom devbot import chat def call(message, name, protocol, cfg, commands): if message == '': chat.say('/r ' + commands['help']['alias'].format('alias'), protocol) return aliases = '' for tupl in commands['regex'].items(): if tupl[1] == message: aliases = aliases + tupl[0] + ', ' if aliases == '': chat.say('/r Sorry, there is no command by that name.', protocol) else: chat.say('/r ' + aliases, protocol)
<commit_before>from devbot import chat def call(message, name, protocol, cfg, commands): if message == '': chat.say('/r ' + commands['help']['alias'].format('alias'), protocol) return aliases = '' for tupl in commands['regex'].items(): if tupl[1] == message: aliases = aliases + tupl[0] + ', ' if aliases == '': chat.say('Sorry, there is no command by that name.', protocol) else: chat.say('/r ' + aliases, protocol) <commit_msg>Fix bug with DevotedBot speaking error in chat<commit_after>from devbot import chat def call(message, name, protocol, cfg, commands): if message == '': chat.say('/r ' + commands['help']['alias'].format('alias'), protocol) return aliases = '' for tupl in commands['regex'].items(): if tupl[1] == message: aliases = aliases + tupl[0] + ', ' if aliases == '': chat.say('/r Sorry, there is no command by that name.', protocol) else: chat.say('/r ' + aliases, protocol)
033a9c6e8b704eb92a7b1a9a82235fee7374f6be
logger/utilities.py
logger/utilities.py
#!/usr/bin/env python3 """Small utility functions for use in various places.""" __all__ = ["pick", "is_dunder", "convert_to_od", "counter_to_iterable"] import collections import itertools def pick(arg, default): """Handler for default versus given argument.""" return default if arg is None else arg def is_dunder(name): """Return True if a __dunder__ name, False otherwise.""" return name[:2] == name[-2:] == "__" and "_" not in (name[2:3], name[-3:-2]) def convert_to_od(mapping, order): """Convert mapping to an OrderedDict instance using order.""" return collections.OrderedDict([(i, mapping[i]) for i in order]) def counter_to_iterable(counter): """Convert a counter to an iterable / iterator.""" for item in itertools.starmap(itertools.repeat, counter): yield from item
#!/usr/bin/env python3 """Small utility functions for use in various places.""" __all__ = ["pick", "is_dunder", "convert_to_od", "counter_to_iterable", "count"] import collections import itertools def pick(arg, default): """Handler for default versus given argument.""" return default if arg is None else arg def is_dunder(name): """Return True if a __dunder__ name, False otherwise.""" return name[:2] == name[-2:] == "__" and "_" not in (name[2:3], name[-3:-2]) def convert_to_od(mapping, order): """Convert mapping to an OrderedDict instance using order.""" return collections.OrderedDict([(i, mapping[i]) for i in order]) def counter_to_iterable(counter): """Convert a counter to an iterable / iterator.""" for item in itertools.starmap(itertools.repeat, counter): yield from item def count(iterable): """Yield (item, count) two-tuples of the iterable.""" seen = [] full = list(iterable) for item in full: if item in seen: continue seen.append(item) yield (item, full.count(item))
Add a new 'count' utility function
Add a new 'count' utility function
Python
bsd-2-clause
Vgr255/logging
#!/usr/bin/env python3 """Small utility functions for use in various places.""" __all__ = ["pick", "is_dunder", "convert_to_od", "counter_to_iterable"] import collections import itertools def pick(arg, default): """Handler for default versus given argument.""" return default if arg is None else arg def is_dunder(name): """Return True if a __dunder__ name, False otherwise.""" return name[:2] == name[-2:] == "__" and "_" not in (name[2:3], name[-3:-2]) def convert_to_od(mapping, order): """Convert mapping to an OrderedDict instance using order.""" return collections.OrderedDict([(i, mapping[i]) for i in order]) def counter_to_iterable(counter): """Convert a counter to an iterable / iterator.""" for item in itertools.starmap(itertools.repeat, counter): yield from item Add a new 'count' utility function
#!/usr/bin/env python3 """Small utility functions for use in various places.""" __all__ = ["pick", "is_dunder", "convert_to_od", "counter_to_iterable", "count"] import collections import itertools def pick(arg, default): """Handler for default versus given argument.""" return default if arg is None else arg def is_dunder(name): """Return True if a __dunder__ name, False otherwise.""" return name[:2] == name[-2:] == "__" and "_" not in (name[2:3], name[-3:-2]) def convert_to_od(mapping, order): """Convert mapping to an OrderedDict instance using order.""" return collections.OrderedDict([(i, mapping[i]) for i in order]) def counter_to_iterable(counter): """Convert a counter to an iterable / iterator.""" for item in itertools.starmap(itertools.repeat, counter): yield from item def count(iterable): """Yield (item, count) two-tuples of the iterable.""" seen = [] full = list(iterable) for item in full: if item in seen: continue seen.append(item) yield (item, full.count(item))
<commit_before>#!/usr/bin/env python3 """Small utility functions for use in various places.""" __all__ = ["pick", "is_dunder", "convert_to_od", "counter_to_iterable"] import collections import itertools def pick(arg, default): """Handler for default versus given argument.""" return default if arg is None else arg def is_dunder(name): """Return True if a __dunder__ name, False otherwise.""" return name[:2] == name[-2:] == "__" and "_" not in (name[2:3], name[-3:-2]) def convert_to_od(mapping, order): """Convert mapping to an OrderedDict instance using order.""" return collections.OrderedDict([(i, mapping[i]) for i in order]) def counter_to_iterable(counter): """Convert a counter to an iterable / iterator.""" for item in itertools.starmap(itertools.repeat, counter): yield from item <commit_msg>Add a new 'count' utility function<commit_after>
#!/usr/bin/env python3 """Small utility functions for use in various places.""" __all__ = ["pick", "is_dunder", "convert_to_od", "counter_to_iterable", "count"] import collections import itertools def pick(arg, default): """Handler for default versus given argument.""" return default if arg is None else arg def is_dunder(name): """Return True if a __dunder__ name, False otherwise.""" return name[:2] == name[-2:] == "__" and "_" not in (name[2:3], name[-3:-2]) def convert_to_od(mapping, order): """Convert mapping to an OrderedDict instance using order.""" return collections.OrderedDict([(i, mapping[i]) for i in order]) def counter_to_iterable(counter): """Convert a counter to an iterable / iterator.""" for item in itertools.starmap(itertools.repeat, counter): yield from item def count(iterable): """Yield (item, count) two-tuples of the iterable.""" seen = [] full = list(iterable) for item in full: if item in seen: continue seen.append(item) yield (item, full.count(item))
#!/usr/bin/env python3 """Small utility functions for use in various places.""" __all__ = ["pick", "is_dunder", "convert_to_od", "counter_to_iterable"] import collections import itertools def pick(arg, default): """Handler for default versus given argument.""" return default if arg is None else arg def is_dunder(name): """Return True if a __dunder__ name, False otherwise.""" return name[:2] == name[-2:] == "__" and "_" not in (name[2:3], name[-3:-2]) def convert_to_od(mapping, order): """Convert mapping to an OrderedDict instance using order.""" return collections.OrderedDict([(i, mapping[i]) for i in order]) def counter_to_iterable(counter): """Convert a counter to an iterable / iterator.""" for item in itertools.starmap(itertools.repeat, counter): yield from item Add a new 'count' utility function#!/usr/bin/env python3 """Small utility functions for use in various places.""" __all__ = ["pick", "is_dunder", "convert_to_od", "counter_to_iterable", "count"] import collections import itertools def pick(arg, default): """Handler for default versus given argument.""" return default if arg is None else arg def is_dunder(name): """Return True if a __dunder__ name, False otherwise.""" return name[:2] == name[-2:] == "__" and "_" not in (name[2:3], name[-3:-2]) def convert_to_od(mapping, order): """Convert mapping to an OrderedDict instance using order.""" return collections.OrderedDict([(i, mapping[i]) for i in order]) def counter_to_iterable(counter): """Convert a counter to an iterable / iterator.""" for item in itertools.starmap(itertools.repeat, counter): yield from item def count(iterable): """Yield (item, count) two-tuples of the iterable.""" seen = [] full = list(iterable) for item in full: if item in seen: continue seen.append(item) yield (item, full.count(item))
<commit_before>#!/usr/bin/env python3 """Small utility functions for use in various places.""" __all__ = ["pick", "is_dunder", "convert_to_od", "counter_to_iterable"] import collections import itertools def pick(arg, default): """Handler for default versus given argument.""" return default if arg is None else arg def is_dunder(name): """Return True if a __dunder__ name, False otherwise.""" return name[:2] == name[-2:] == "__" and "_" not in (name[2:3], name[-3:-2]) def convert_to_od(mapping, order): """Convert mapping to an OrderedDict instance using order.""" return collections.OrderedDict([(i, mapping[i]) for i in order]) def counter_to_iterable(counter): """Convert a counter to an iterable / iterator.""" for item in itertools.starmap(itertools.repeat, counter): yield from item <commit_msg>Add a new 'count' utility function<commit_after>#!/usr/bin/env python3 """Small utility functions for use in various places.""" __all__ = ["pick", "is_dunder", "convert_to_od", "counter_to_iterable", "count"] import collections import itertools def pick(arg, default): """Handler for default versus given argument.""" return default if arg is None else arg def is_dunder(name): """Return True if a __dunder__ name, False otherwise.""" return name[:2] == name[-2:] == "__" and "_" not in (name[2:3], name[-3:-2]) def convert_to_od(mapping, order): """Convert mapping to an OrderedDict instance using order.""" return collections.OrderedDict([(i, mapping[i]) for i in order]) def counter_to_iterable(counter): """Convert a counter to an iterable / iterator.""" for item in itertools.starmap(itertools.repeat, counter): yield from item def count(iterable): """Yield (item, count) two-tuples of the iterable.""" seen = [] full = list(iterable) for item in full: if item in seen: continue seen.append(item) yield (item, full.count(item))
e75e741770d1735c52770900b1cf59f207f2264e
asp/__init__.py
asp/__init__.py
# From http://stackoverflow.com/questions/458550/standard-way-to-embed-version-into-python-package # Author: James Antill (http://stackoverflow.com/users/10314/james-antill) __version__ = '0.1' __version_info__ = tuple([ int(num) for num in __version__.split('.')])
# From http://stackoverflow.com/questions/458550/standard-way-to-embed-version-into-python-package # Author: James Antill (http://stackoverflow.com/users/10314/james-antill) __version__ = '0.1' __version_info__ = tuple([ int(num) for num in __version__.split('.')]) class SpecializationError(Exception): """ Exception that caused specialization not to occur. Attributes: msg -- the message/explanation to the user phase -- which phase of specialization caused the error """ def __init__(self, msg, phase="Unknown phase"): self.value = value
Add asp.SpecializationError class for specialization-related exceptions.
Add asp.SpecializationError class for specialization-related exceptions.
Python
bsd-3-clause
shoaibkamil/asp,mbdriscoll/asp-old,mbdriscoll/asp-old,shoaibkamil/asp,richardxia/asp-multilevel-debug,richardxia/asp-multilevel-debug,pbirsinger/aspNew,pbirsinger/aspNew,mbdriscoll/asp-old,mbdriscoll/asp-old,shoaibkamil/asp,richardxia/asp-multilevel-debug,pbirsinger/aspNew
# From http://stackoverflow.com/questions/458550/standard-way-to-embed-version-into-python-package # Author: James Antill (http://stackoverflow.com/users/10314/james-antill) __version__ = '0.1' __version_info__ = tuple([ int(num) for num in __version__.split('.')]) Add asp.SpecializationError class for specialization-related exceptions.
# From http://stackoverflow.com/questions/458550/standard-way-to-embed-version-into-python-package # Author: James Antill (http://stackoverflow.com/users/10314/james-antill) __version__ = '0.1' __version_info__ = tuple([ int(num) for num in __version__.split('.')]) class SpecializationError(Exception): """ Exception that caused specialization not to occur. Attributes: msg -- the message/explanation to the user phase -- which phase of specialization caused the error """ def __init__(self, msg, phase="Unknown phase"): self.value = value
<commit_before># From http://stackoverflow.com/questions/458550/standard-way-to-embed-version-into-python-package # Author: James Antill (http://stackoverflow.com/users/10314/james-antill) __version__ = '0.1' __version_info__ = tuple([ int(num) for num in __version__.split('.')]) <commit_msg>Add asp.SpecializationError class for specialization-related exceptions.<commit_after>
# From http://stackoverflow.com/questions/458550/standard-way-to-embed-version-into-python-package # Author: James Antill (http://stackoverflow.com/users/10314/james-antill) __version__ = '0.1' __version_info__ = tuple([ int(num) for num in __version__.split('.')]) class SpecializationError(Exception): """ Exception that caused specialization not to occur. Attributes: msg -- the message/explanation to the user phase -- which phase of specialization caused the error """ def __init__(self, msg, phase="Unknown phase"): self.value = value
# From http://stackoverflow.com/questions/458550/standard-way-to-embed-version-into-python-package # Author: James Antill (http://stackoverflow.com/users/10314/james-antill) __version__ = '0.1' __version_info__ = tuple([ int(num) for num in __version__.split('.')]) Add asp.SpecializationError class for specialization-related exceptions.# From http://stackoverflow.com/questions/458550/standard-way-to-embed-version-into-python-package # Author: James Antill (http://stackoverflow.com/users/10314/james-antill) __version__ = '0.1' __version_info__ = tuple([ int(num) for num in __version__.split('.')]) class SpecializationError(Exception): """ Exception that caused specialization not to occur. Attributes: msg -- the message/explanation to the user phase -- which phase of specialization caused the error """ def __init__(self, msg, phase="Unknown phase"): self.value = value
<commit_before># From http://stackoverflow.com/questions/458550/standard-way-to-embed-version-into-python-package # Author: James Antill (http://stackoverflow.com/users/10314/james-antill) __version__ = '0.1' __version_info__ = tuple([ int(num) for num in __version__.split('.')]) <commit_msg>Add asp.SpecializationError class for specialization-related exceptions.<commit_after># From http://stackoverflow.com/questions/458550/standard-way-to-embed-version-into-python-package # Author: James Antill (http://stackoverflow.com/users/10314/james-antill) __version__ = '0.1' __version_info__ = tuple([ int(num) for num in __version__.split('.')]) class SpecializationError(Exception): """ Exception that caused specialization not to occur. Attributes: msg -- the message/explanation to the user phase -- which phase of specialization caused the error """ def __init__(self, msg, phase="Unknown phase"): self.value = value
64fdf3a83b7b8649de6216cd67fb1a8ae0d3f1a0
bin/receive.py
bin/receive.py
#!/usr/bin/env python import pika import subprocess import json import os from pymongo import MongoClient dbcon = MongoClient() NETSNIFF_UTIL = os.path.join(os.path.dirname(os.path.realpath(__file__)), 'tools', 'netsniff.js') queuecon = pika.BlockingConnection(pika.ConnectionParameters( 'localhost')) channel = queuecon.channel() channel.exchange_declare(exchange='perfmonitor', type='direct') result = channel.queue_declare('perf') queue_name = result.method.queue channel.queue_bind(exchange='perfmonitor', queue=queue_name, routing_key='perftest') print ' [*] Waiting for messages. To exit press CTRL+C' def callback(ch, method, properties, body): print " [x] Received %r" % (body,) content = json.loads(body) print ' [x] Executing command phantomjs', content['url'] harcontent = subprocess.check_output(['phantomjs', NETSNIFF_UTIL, content['url']]) jscontent = json.loads(harcontent) jscontent['site'] = content['site'] dbcon.perfmonitor.har.insert(jscontent) ch.basic_ack(delivery_tag = method.delivery_tag) print " [x] Done" channel.basic_consume(callback, queue=queue_name) channel.start_consuming()
#!/usr/bin/env python import pika import subprocess import json import os from pymongo import MongoClient dbcon = MongoClient() NETSNIFF_UTIL = os.path.join(os.path.dirname(os.path.realpath(__file__)), 'tools', 'netsniff.js') queuecon = pika.BlockingConnection(pika.ConnectionParameters( 'localhost')) channel = queuecon.channel() channel.exchange_declare(exchange='perfmonitor', type='direct') result = channel.queue_declare('perf') queue_name = result.method.queue channel.queue_bind(exchange='perfmonitor', queue=queue_name, routing_key='perftest') print ' [*] Waiting for messages. To exit press CTRL+C' def callback(ch, method, properties, body): print " [x] Received %r" % (body,) content = json.loads(body) print ' [x] Executing command phantomjs', content['url'] harcontent = subprocess.check_output(['phantomjs', NETSNIFF_UTIL, content['url']]) try: jscontent = json.loads(harcontent) jscontent['site'] = content['site'] dbcon.perfmonitor.har.insert(jscontent) except: print ' [x] Unable to parse JSON, ignoring request' ch.basic_ack(delivery_tag = method.delivery_tag) print " [x] Done" channel.basic_consume(callback, queue=queue_name) channel.start_consuming()
Handle errors in queue processing
Handle errors in queue processing
Python
mit
leibowitz/perfmonitor,leibowitz/perfmonitor,leibowitz/perfmonitor
#!/usr/bin/env python import pika import subprocess import json import os from pymongo import MongoClient dbcon = MongoClient() NETSNIFF_UTIL = os.path.join(os.path.dirname(os.path.realpath(__file__)), 'tools', 'netsniff.js') queuecon = pika.BlockingConnection(pika.ConnectionParameters( 'localhost')) channel = queuecon.channel() channel.exchange_declare(exchange='perfmonitor', type='direct') result = channel.queue_declare('perf') queue_name = result.method.queue channel.queue_bind(exchange='perfmonitor', queue=queue_name, routing_key='perftest') print ' [*] Waiting for messages. To exit press CTRL+C' def callback(ch, method, properties, body): print " [x] Received %r" % (body,) content = json.loads(body) print ' [x] Executing command phantomjs', content['url'] harcontent = subprocess.check_output(['phantomjs', NETSNIFF_UTIL, content['url']]) jscontent = json.loads(harcontent) jscontent['site'] = content['site'] dbcon.perfmonitor.har.insert(jscontent) ch.basic_ack(delivery_tag = method.delivery_tag) print " [x] Done" channel.basic_consume(callback, queue=queue_name) channel.start_consuming() Handle errors in queue processing
#!/usr/bin/env python import pika import subprocess import json import os from pymongo import MongoClient dbcon = MongoClient() NETSNIFF_UTIL = os.path.join(os.path.dirname(os.path.realpath(__file__)), 'tools', 'netsniff.js') queuecon = pika.BlockingConnection(pika.ConnectionParameters( 'localhost')) channel = queuecon.channel() channel.exchange_declare(exchange='perfmonitor', type='direct') result = channel.queue_declare('perf') queue_name = result.method.queue channel.queue_bind(exchange='perfmonitor', queue=queue_name, routing_key='perftest') print ' [*] Waiting for messages. To exit press CTRL+C' def callback(ch, method, properties, body): print " [x] Received %r" % (body,) content = json.loads(body) print ' [x] Executing command phantomjs', content['url'] harcontent = subprocess.check_output(['phantomjs', NETSNIFF_UTIL, content['url']]) try: jscontent = json.loads(harcontent) jscontent['site'] = content['site'] dbcon.perfmonitor.har.insert(jscontent) except: print ' [x] Unable to parse JSON, ignoring request' ch.basic_ack(delivery_tag = method.delivery_tag) print " [x] Done" channel.basic_consume(callback, queue=queue_name) channel.start_consuming()
<commit_before>#!/usr/bin/env python import pika import subprocess import json import os from pymongo import MongoClient dbcon = MongoClient() NETSNIFF_UTIL = os.path.join(os.path.dirname(os.path.realpath(__file__)), 'tools', 'netsniff.js') queuecon = pika.BlockingConnection(pika.ConnectionParameters( 'localhost')) channel = queuecon.channel() channel.exchange_declare(exchange='perfmonitor', type='direct') result = channel.queue_declare('perf') queue_name = result.method.queue channel.queue_bind(exchange='perfmonitor', queue=queue_name, routing_key='perftest') print ' [*] Waiting for messages. To exit press CTRL+C' def callback(ch, method, properties, body): print " [x] Received %r" % (body,) content = json.loads(body) print ' [x] Executing command phantomjs', content['url'] harcontent = subprocess.check_output(['phantomjs', NETSNIFF_UTIL, content['url']]) jscontent = json.loads(harcontent) jscontent['site'] = content['site'] dbcon.perfmonitor.har.insert(jscontent) ch.basic_ack(delivery_tag = method.delivery_tag) print " [x] Done" channel.basic_consume(callback, queue=queue_name) channel.start_consuming() <commit_msg>Handle errors in queue processing<commit_after>
#!/usr/bin/env python import pika import subprocess import json import os from pymongo import MongoClient dbcon = MongoClient() NETSNIFF_UTIL = os.path.join(os.path.dirname(os.path.realpath(__file__)), 'tools', 'netsniff.js') queuecon = pika.BlockingConnection(pika.ConnectionParameters( 'localhost')) channel = queuecon.channel() channel.exchange_declare(exchange='perfmonitor', type='direct') result = channel.queue_declare('perf') queue_name = result.method.queue channel.queue_bind(exchange='perfmonitor', queue=queue_name, routing_key='perftest') print ' [*] Waiting for messages. To exit press CTRL+C' def callback(ch, method, properties, body): print " [x] Received %r" % (body,) content = json.loads(body) print ' [x] Executing command phantomjs', content['url'] harcontent = subprocess.check_output(['phantomjs', NETSNIFF_UTIL, content['url']]) try: jscontent = json.loads(harcontent) jscontent['site'] = content['site'] dbcon.perfmonitor.har.insert(jscontent) except: print ' [x] Unable to parse JSON, ignoring request' ch.basic_ack(delivery_tag = method.delivery_tag) print " [x] Done" channel.basic_consume(callback, queue=queue_name) channel.start_consuming()
#!/usr/bin/env python import pika import subprocess import json import os from pymongo import MongoClient dbcon = MongoClient() NETSNIFF_UTIL = os.path.join(os.path.dirname(os.path.realpath(__file__)), 'tools', 'netsniff.js') queuecon = pika.BlockingConnection(pika.ConnectionParameters( 'localhost')) channel = queuecon.channel() channel.exchange_declare(exchange='perfmonitor', type='direct') result = channel.queue_declare('perf') queue_name = result.method.queue channel.queue_bind(exchange='perfmonitor', queue=queue_name, routing_key='perftest') print ' [*] Waiting for messages. To exit press CTRL+C' def callback(ch, method, properties, body): print " [x] Received %r" % (body,) content = json.loads(body) print ' [x] Executing command phantomjs', content['url'] harcontent = subprocess.check_output(['phantomjs', NETSNIFF_UTIL, content['url']]) jscontent = json.loads(harcontent) jscontent['site'] = content['site'] dbcon.perfmonitor.har.insert(jscontent) ch.basic_ack(delivery_tag = method.delivery_tag) print " [x] Done" channel.basic_consume(callback, queue=queue_name) channel.start_consuming() Handle errors in queue processing#!/usr/bin/env python import pika import subprocess import json import os from pymongo import MongoClient dbcon = MongoClient() NETSNIFF_UTIL = os.path.join(os.path.dirname(os.path.realpath(__file__)), 'tools', 'netsniff.js') queuecon = pika.BlockingConnection(pika.ConnectionParameters( 'localhost')) channel = queuecon.channel() channel.exchange_declare(exchange='perfmonitor', type='direct') result = channel.queue_declare('perf') queue_name = result.method.queue channel.queue_bind(exchange='perfmonitor', queue=queue_name, routing_key='perftest') print ' [*] Waiting for messages. To exit press CTRL+C' def callback(ch, method, properties, body): print " [x] Received %r" % (body,) content = json.loads(body) print ' [x] Executing command phantomjs', content['url'] harcontent = subprocess.check_output(['phantomjs', NETSNIFF_UTIL, content['url']]) try: jscontent = json.loads(harcontent) jscontent['site'] = content['site'] dbcon.perfmonitor.har.insert(jscontent) except: print ' [x] Unable to parse JSON, ignoring request' ch.basic_ack(delivery_tag = method.delivery_tag) print " [x] Done" channel.basic_consume(callback, queue=queue_name) channel.start_consuming()
<commit_before>#!/usr/bin/env python import pika import subprocess import json import os from pymongo import MongoClient dbcon = MongoClient() NETSNIFF_UTIL = os.path.join(os.path.dirname(os.path.realpath(__file__)), 'tools', 'netsniff.js') queuecon = pika.BlockingConnection(pika.ConnectionParameters( 'localhost')) channel = queuecon.channel() channel.exchange_declare(exchange='perfmonitor', type='direct') result = channel.queue_declare('perf') queue_name = result.method.queue channel.queue_bind(exchange='perfmonitor', queue=queue_name, routing_key='perftest') print ' [*] Waiting for messages. To exit press CTRL+C' def callback(ch, method, properties, body): print " [x] Received %r" % (body,) content = json.loads(body) print ' [x] Executing command phantomjs', content['url'] harcontent = subprocess.check_output(['phantomjs', NETSNIFF_UTIL, content['url']]) jscontent = json.loads(harcontent) jscontent['site'] = content['site'] dbcon.perfmonitor.har.insert(jscontent) ch.basic_ack(delivery_tag = method.delivery_tag) print " [x] Done" channel.basic_consume(callback, queue=queue_name) channel.start_consuming() <commit_msg>Handle errors in queue processing<commit_after>#!/usr/bin/env python import pika import subprocess import json import os from pymongo import MongoClient dbcon = MongoClient() NETSNIFF_UTIL = os.path.join(os.path.dirname(os.path.realpath(__file__)), 'tools', 'netsniff.js') queuecon = pika.BlockingConnection(pika.ConnectionParameters( 'localhost')) channel = queuecon.channel() channel.exchange_declare(exchange='perfmonitor', type='direct') result = channel.queue_declare('perf') queue_name = result.method.queue channel.queue_bind(exchange='perfmonitor', queue=queue_name, routing_key='perftest') print ' [*] Waiting for messages. To exit press CTRL+C' def callback(ch, method, properties, body): print " [x] Received %r" % (body,) content = json.loads(body) print ' [x] Executing command phantomjs', content['url'] harcontent = subprocess.check_output(['phantomjs', NETSNIFF_UTIL, content['url']]) try: jscontent = json.loads(harcontent) jscontent['site'] = content['site'] dbcon.perfmonitor.har.insert(jscontent) except: print ' [x] Unable to parse JSON, ignoring request' ch.basic_ack(delivery_tag = method.delivery_tag) print " [x] Done" channel.basic_consume(callback, queue=queue_name) channel.start_consuming()
4708751802258c724bafe843845e91d88599df8b
marmoset/installimage/__init__.py
marmoset/installimage/__init__.py
from .installimage_config import InstallimageConfig def create(args): install_config = InstallimageConfig(args.mac) for var in args.var: install_config.add_or_set(var[0], var[1]) install_config.create() msg = 'Created %s with following Options:\n%s' % (install_config.file_path(), install_config.get_content()) print(msg) def list(args): for install_config in InstallimageConfig.all(): print('%s' % install_config.mac) def remove(args): install_config = InstallimageConfig(args.mac) if install_config.remove(): print('Removed', install_config.file_path()) else: print('No entry found for', install_config.mac)
Add cli handling for installimage
Add cli handling for installimage
Python
agpl-3.0
aibor/marmoset
Add cli handling for installimage
from .installimage_config import InstallimageConfig def create(args): install_config = InstallimageConfig(args.mac) for var in args.var: install_config.add_or_set(var[0], var[1]) install_config.create() msg = 'Created %s with following Options:\n%s' % (install_config.file_path(), install_config.get_content()) print(msg) def list(args): for install_config in InstallimageConfig.all(): print('%s' % install_config.mac) def remove(args): install_config = InstallimageConfig(args.mac) if install_config.remove(): print('Removed', install_config.file_path()) else: print('No entry found for', install_config.mac)
<commit_before><commit_msg>Add cli handling for installimage<commit_after>
from .installimage_config import InstallimageConfig def create(args): install_config = InstallimageConfig(args.mac) for var in args.var: install_config.add_or_set(var[0], var[1]) install_config.create() msg = 'Created %s with following Options:\n%s' % (install_config.file_path(), install_config.get_content()) print(msg) def list(args): for install_config in InstallimageConfig.all(): print('%s' % install_config.mac) def remove(args): install_config = InstallimageConfig(args.mac) if install_config.remove(): print('Removed', install_config.file_path()) else: print('No entry found for', install_config.mac)
Add cli handling for installimagefrom .installimage_config import InstallimageConfig def create(args): install_config = InstallimageConfig(args.mac) for var in args.var: install_config.add_or_set(var[0], var[1]) install_config.create() msg = 'Created %s with following Options:\n%s' % (install_config.file_path(), install_config.get_content()) print(msg) def list(args): for install_config in InstallimageConfig.all(): print('%s' % install_config.mac) def remove(args): install_config = InstallimageConfig(args.mac) if install_config.remove(): print('Removed', install_config.file_path()) else: print('No entry found for', install_config.mac)
<commit_before><commit_msg>Add cli handling for installimage<commit_after>from .installimage_config import InstallimageConfig def create(args): install_config = InstallimageConfig(args.mac) for var in args.var: install_config.add_or_set(var[0], var[1]) install_config.create() msg = 'Created %s with following Options:\n%s' % (install_config.file_path(), install_config.get_content()) print(msg) def list(args): for install_config in InstallimageConfig.all(): print('%s' % install_config.mac) def remove(args): install_config = InstallimageConfig(args.mac) if install_config.remove(): print('Removed', install_config.file_path()) else: print('No entry found for', install_config.mac)
6711d68999e5a9b0ea72a9a4f33cfc86b4230012
pattern_matcher/pattern_matcher.py
pattern_matcher/pattern_matcher.py
from .regex import RegexFactory from .patterns import Patterns class Matcher(object): NO_MATCH = 'NO MATCH' def __init__(self, raw_patterns, path, re_factory=RegexFactory): self.raw_patterns = raw_patterns self.path = path self.re = re_factory().create(self.path) self.patterns = Patterns(self.re.findall(self.raw_patterns)) class Matcher(object): def __init__(self, patterns, re_factory=RegexFactory): self.patterns self.re_factory = re_factory def _find_matches(self, path): regex = self.re_factory.new(path) return regex.findall(self.patterns) def _get_best_match(self, matches): pass def match(self, path): """Matches a path to a path pattern.""" matches = self._find_matches(path) return self._get_best_match(matches) class PathMatcher(object): """Matches a Path to the Path Pattern.""" def __init__(self, input, output): self.input = InputManager(input) self.output = OutputManager(output) self.matcher = Matcher() def match(self): for path in self.input.stream: self.matcher.match(path.strip()) # send to stdout class InputManager(object): """Manages the input to the matcher.""" pass class OutputManager(object): """Manages the output of the matcher.""" pass if __name__ == '__main__': import sys main = PathMatcher(sys.stdin, sys.stdout) main.match()
from .regex import RegexFactory from .patterns import Patterns class Matcher(object): NO_MATCH = 'NO MATCH' def __init__(self, raw_patterns, path, re_factory=RegexFactory): self.raw_patterns = raw_patterns self.path = path self.re = re_factory().create(self.path) self.patterns = Patterns(self.re.findall(self.raw_patterns)) def match(self): matches = self.patterns.get_best_patterns() if len(matches) != 1: return self.NO_MATCH return str(matches.pop()) class PathMatcher(object): """Matches a Path to the Path Pattern.""" def __init__(self, input, output): self.input = InputManager(input) self.output = OutputManager(output) self.matcher = Matcher def match(self): for path in self.input.stream: matcher = self.Matcher(path.strip()) print(matcher.match()) # send to stdout class InputManager(object): """Manages the input to the matcher.""" pass class OutputManager(object): """Manages the output of the matcher.""" pass if __name__ == '__main__': import sys main = PathMatcher(sys.stdin, sys.stdout) main.match()
Change the way the Matcher is initialized and how matches are retrieved.
Change the way the Matcher is initialized and how matches are retrieved.
Python
mit
damonkelley/pattern-matcher
from .regex import RegexFactory from .patterns import Patterns class Matcher(object): NO_MATCH = 'NO MATCH' def __init__(self, raw_patterns, path, re_factory=RegexFactory): self.raw_patterns = raw_patterns self.path = path self.re = re_factory().create(self.path) self.patterns = Patterns(self.re.findall(self.raw_patterns)) class Matcher(object): def __init__(self, patterns, re_factory=RegexFactory): self.patterns self.re_factory = re_factory def _find_matches(self, path): regex = self.re_factory.new(path) return regex.findall(self.patterns) def _get_best_match(self, matches): pass def match(self, path): """Matches a path to a path pattern.""" matches = self._find_matches(path) return self._get_best_match(matches) class PathMatcher(object): """Matches a Path to the Path Pattern.""" def __init__(self, input, output): self.input = InputManager(input) self.output = OutputManager(output) self.matcher = Matcher() def match(self): for path in self.input.stream: self.matcher.match(path.strip()) # send to stdout class InputManager(object): """Manages the input to the matcher.""" pass class OutputManager(object): """Manages the output of the matcher.""" pass if __name__ == '__main__': import sys main = PathMatcher(sys.stdin, sys.stdout) main.match() Change the way the Matcher is initialized and how matches are retrieved.
from .regex import RegexFactory from .patterns import Patterns class Matcher(object): NO_MATCH = 'NO MATCH' def __init__(self, raw_patterns, path, re_factory=RegexFactory): self.raw_patterns = raw_patterns self.path = path self.re = re_factory().create(self.path) self.patterns = Patterns(self.re.findall(self.raw_patterns)) def match(self): matches = self.patterns.get_best_patterns() if len(matches) != 1: return self.NO_MATCH return str(matches.pop()) class PathMatcher(object): """Matches a Path to the Path Pattern.""" def __init__(self, input, output): self.input = InputManager(input) self.output = OutputManager(output) self.matcher = Matcher def match(self): for path in self.input.stream: matcher = self.Matcher(path.strip()) print(matcher.match()) # send to stdout class InputManager(object): """Manages the input to the matcher.""" pass class OutputManager(object): """Manages the output of the matcher.""" pass if __name__ == '__main__': import sys main = PathMatcher(sys.stdin, sys.stdout) main.match()
<commit_before>from .regex import RegexFactory from .patterns import Patterns class Matcher(object): NO_MATCH = 'NO MATCH' def __init__(self, raw_patterns, path, re_factory=RegexFactory): self.raw_patterns = raw_patterns self.path = path self.re = re_factory().create(self.path) self.patterns = Patterns(self.re.findall(self.raw_patterns)) class Matcher(object): def __init__(self, patterns, re_factory=RegexFactory): self.patterns self.re_factory = re_factory def _find_matches(self, path): regex = self.re_factory.new(path) return regex.findall(self.patterns) def _get_best_match(self, matches): pass def match(self, path): """Matches a path to a path pattern.""" matches = self._find_matches(path) return self._get_best_match(matches) class PathMatcher(object): """Matches a Path to the Path Pattern.""" def __init__(self, input, output): self.input = InputManager(input) self.output = OutputManager(output) self.matcher = Matcher() def match(self): for path in self.input.stream: self.matcher.match(path.strip()) # send to stdout class InputManager(object): """Manages the input to the matcher.""" pass class OutputManager(object): """Manages the output of the matcher.""" pass if __name__ == '__main__': import sys main = PathMatcher(sys.stdin, sys.stdout) main.match() <commit_msg>Change the way the Matcher is initialized and how matches are retrieved.<commit_after>
from .regex import RegexFactory from .patterns import Patterns class Matcher(object): NO_MATCH = 'NO MATCH' def __init__(self, raw_patterns, path, re_factory=RegexFactory): self.raw_patterns = raw_patterns self.path = path self.re = re_factory().create(self.path) self.patterns = Patterns(self.re.findall(self.raw_patterns)) def match(self): matches = self.patterns.get_best_patterns() if len(matches) != 1: return self.NO_MATCH return str(matches.pop()) class PathMatcher(object): """Matches a Path to the Path Pattern.""" def __init__(self, input, output): self.input = InputManager(input) self.output = OutputManager(output) self.matcher = Matcher def match(self): for path in self.input.stream: matcher = self.Matcher(path.strip()) print(matcher.match()) # send to stdout class InputManager(object): """Manages the input to the matcher.""" pass class OutputManager(object): """Manages the output of the matcher.""" pass if __name__ == '__main__': import sys main = PathMatcher(sys.stdin, sys.stdout) main.match()
from .regex import RegexFactory from .patterns import Patterns class Matcher(object): NO_MATCH = 'NO MATCH' def __init__(self, raw_patterns, path, re_factory=RegexFactory): self.raw_patterns = raw_patterns self.path = path self.re = re_factory().create(self.path) self.patterns = Patterns(self.re.findall(self.raw_patterns)) class Matcher(object): def __init__(self, patterns, re_factory=RegexFactory): self.patterns self.re_factory = re_factory def _find_matches(self, path): regex = self.re_factory.new(path) return regex.findall(self.patterns) def _get_best_match(self, matches): pass def match(self, path): """Matches a path to a path pattern.""" matches = self._find_matches(path) return self._get_best_match(matches) class PathMatcher(object): """Matches a Path to the Path Pattern.""" def __init__(self, input, output): self.input = InputManager(input) self.output = OutputManager(output) self.matcher = Matcher() def match(self): for path in self.input.stream: self.matcher.match(path.strip()) # send to stdout class InputManager(object): """Manages the input to the matcher.""" pass class OutputManager(object): """Manages the output of the matcher.""" pass if __name__ == '__main__': import sys main = PathMatcher(sys.stdin, sys.stdout) main.match() Change the way the Matcher is initialized and how matches are retrieved.from .regex import RegexFactory from .patterns import Patterns class Matcher(object): NO_MATCH = 'NO MATCH' def __init__(self, raw_patterns, path, re_factory=RegexFactory): self.raw_patterns = raw_patterns self.path = path self.re = re_factory().create(self.path) self.patterns = Patterns(self.re.findall(self.raw_patterns)) def match(self): matches = self.patterns.get_best_patterns() if len(matches) != 1: return self.NO_MATCH return str(matches.pop()) class PathMatcher(object): """Matches a Path to the Path Pattern.""" def __init__(self, input, output): self.input = InputManager(input) self.output = OutputManager(output) self.matcher = Matcher def match(self): for path in self.input.stream: matcher = self.Matcher(path.strip()) print(matcher.match()) # send to stdout class InputManager(object): """Manages the input to the matcher.""" pass class OutputManager(object): """Manages the output of the matcher.""" pass if __name__ == '__main__': import sys main = PathMatcher(sys.stdin, sys.stdout) main.match()
<commit_before>from .regex import RegexFactory from .patterns import Patterns class Matcher(object): NO_MATCH = 'NO MATCH' def __init__(self, raw_patterns, path, re_factory=RegexFactory): self.raw_patterns = raw_patterns self.path = path self.re = re_factory().create(self.path) self.patterns = Patterns(self.re.findall(self.raw_patterns)) class Matcher(object): def __init__(self, patterns, re_factory=RegexFactory): self.patterns self.re_factory = re_factory def _find_matches(self, path): regex = self.re_factory.new(path) return regex.findall(self.patterns) def _get_best_match(self, matches): pass def match(self, path): """Matches a path to a path pattern.""" matches = self._find_matches(path) return self._get_best_match(matches) class PathMatcher(object): """Matches a Path to the Path Pattern.""" def __init__(self, input, output): self.input = InputManager(input) self.output = OutputManager(output) self.matcher = Matcher() def match(self): for path in self.input.stream: self.matcher.match(path.strip()) # send to stdout class InputManager(object): """Manages the input to the matcher.""" pass class OutputManager(object): """Manages the output of the matcher.""" pass if __name__ == '__main__': import sys main = PathMatcher(sys.stdin, sys.stdout) main.match() <commit_msg>Change the way the Matcher is initialized and how matches are retrieved.<commit_after>from .regex import RegexFactory from .patterns import Patterns class Matcher(object): NO_MATCH = 'NO MATCH' def __init__(self, raw_patterns, path, re_factory=RegexFactory): self.raw_patterns = raw_patterns self.path = path self.re = re_factory().create(self.path) self.patterns = Patterns(self.re.findall(self.raw_patterns)) def match(self): matches = self.patterns.get_best_patterns() if len(matches) != 1: return self.NO_MATCH return str(matches.pop()) class PathMatcher(object): """Matches a Path to the Path Pattern.""" def __init__(self, input, output): self.input = InputManager(input) self.output = OutputManager(output) self.matcher = Matcher def match(self): for path in self.input.stream: matcher = self.Matcher(path.strip()) print(matcher.match()) # send to stdout class InputManager(object): """Manages the input to the matcher.""" pass class OutputManager(object): """Manages the output of the matcher.""" pass if __name__ == '__main__': import sys main = PathMatcher(sys.stdin, sys.stdout) main.match()
34b736645e45126c6cabb6d3f5427a697ebe74ff
tutorial/polls/views.py
tutorial/polls/views.py
from django.shortcuts import render # Create your views here.
from django.shortcuts import render from django.http import HttpResponse # Create your views here. def index(request): return HttpResponse("Hello world you are at the index")
Create a simple Index view
Create a simple Index view
Python
mit
ikosenn/django_reignited,ikosenn/django_reignited
from django.shortcuts import render # Create your views here. Create a simple Index view
from django.shortcuts import render from django.http import HttpResponse # Create your views here. def index(request): return HttpResponse("Hello world you are at the index")
<commit_before>from django.shortcuts import render # Create your views here. <commit_msg>Create a simple Index view<commit_after>
from django.shortcuts import render from django.http import HttpResponse # Create your views here. def index(request): return HttpResponse("Hello world you are at the index")
from django.shortcuts import render # Create your views here. Create a simple Index viewfrom django.shortcuts import render from django.http import HttpResponse # Create your views here. def index(request): return HttpResponse("Hello world you are at the index")
<commit_before>from django.shortcuts import render # Create your views here. <commit_msg>Create a simple Index view<commit_after>from django.shortcuts import render from django.http import HttpResponse # Create your views here. def index(request): return HttpResponse("Hello world you are at the index")
1bc95e2e2a2d4d0daf6cfcdbf7e4803b13262a49
etcd3/__init__.py
etcd3/__init__.py
from __future__ import absolute_import __author__ = 'Louis Taylor' __email__ = 'louis@kragniz.eu' __version__ = '0.1.0' import grpc from etcdrpc import rpc_pb2 as etcdrpc channel = grpc.insecure_channel('localhost:2379') stub = etcdrpc.KVStub(channel) put_request = etcdrpc.PutRequest() put_request.key = 'doot'.encode('utf-8') put_request.value = 'toottoot'.encode('utf-8') print(stub.Put(put_request)) class Etcd3Client(object): def __init__(self): pass def get(self, key): pass def put(self, key, value): pass
from __future__ import absolute_import __author__ = 'Louis Taylor' __email__ = 'louis@kragniz.eu' __version__ = '0.1.0' import grpc from etcdrpc import rpc_pb2 as etcdrpc class Etcd3Client(object): def __init__(self): self.channel = grpc.insecure_channel('localhost:2379') self.kvstub = etcdrpc.KVStub(channel) def get(self, key): pass def put(self, key, value): put_request = etcdrpc.PutRequest() put_request.key = key.encode('utf-8') put_request.value = value.encode('utf-8') self.kvstub.Put(put_request)
Move put code into method
Move put code into method
Python
apache-2.0
kragniz/python-etcd3
from __future__ import absolute_import __author__ = 'Louis Taylor' __email__ = 'louis@kragniz.eu' __version__ = '0.1.0' import grpc from etcdrpc import rpc_pb2 as etcdrpc channel = grpc.insecure_channel('localhost:2379') stub = etcdrpc.KVStub(channel) put_request = etcdrpc.PutRequest() put_request.key = 'doot'.encode('utf-8') put_request.value = 'toottoot'.encode('utf-8') print(stub.Put(put_request)) class Etcd3Client(object): def __init__(self): pass def get(self, key): pass def put(self, key, value): pass Move put code into method
from __future__ import absolute_import __author__ = 'Louis Taylor' __email__ = 'louis@kragniz.eu' __version__ = '0.1.0' import grpc from etcdrpc import rpc_pb2 as etcdrpc class Etcd3Client(object): def __init__(self): self.channel = grpc.insecure_channel('localhost:2379') self.kvstub = etcdrpc.KVStub(channel) def get(self, key): pass def put(self, key, value): put_request = etcdrpc.PutRequest() put_request.key = key.encode('utf-8') put_request.value = value.encode('utf-8') self.kvstub.Put(put_request)
<commit_before>from __future__ import absolute_import __author__ = 'Louis Taylor' __email__ = 'louis@kragniz.eu' __version__ = '0.1.0' import grpc from etcdrpc import rpc_pb2 as etcdrpc channel = grpc.insecure_channel('localhost:2379') stub = etcdrpc.KVStub(channel) put_request = etcdrpc.PutRequest() put_request.key = 'doot'.encode('utf-8') put_request.value = 'toottoot'.encode('utf-8') print(stub.Put(put_request)) class Etcd3Client(object): def __init__(self): pass def get(self, key): pass def put(self, key, value): pass <commit_msg>Move put code into method<commit_after>
from __future__ import absolute_import __author__ = 'Louis Taylor' __email__ = 'louis@kragniz.eu' __version__ = '0.1.0' import grpc from etcdrpc import rpc_pb2 as etcdrpc class Etcd3Client(object): def __init__(self): self.channel = grpc.insecure_channel('localhost:2379') self.kvstub = etcdrpc.KVStub(channel) def get(self, key): pass def put(self, key, value): put_request = etcdrpc.PutRequest() put_request.key = key.encode('utf-8') put_request.value = value.encode('utf-8') self.kvstub.Put(put_request)
from __future__ import absolute_import __author__ = 'Louis Taylor' __email__ = 'louis@kragniz.eu' __version__ = '0.1.0' import grpc from etcdrpc import rpc_pb2 as etcdrpc channel = grpc.insecure_channel('localhost:2379') stub = etcdrpc.KVStub(channel) put_request = etcdrpc.PutRequest() put_request.key = 'doot'.encode('utf-8') put_request.value = 'toottoot'.encode('utf-8') print(stub.Put(put_request)) class Etcd3Client(object): def __init__(self): pass def get(self, key): pass def put(self, key, value): pass Move put code into methodfrom __future__ import absolute_import __author__ = 'Louis Taylor' __email__ = 'louis@kragniz.eu' __version__ = '0.1.0' import grpc from etcdrpc import rpc_pb2 as etcdrpc class Etcd3Client(object): def __init__(self): self.channel = grpc.insecure_channel('localhost:2379') self.kvstub = etcdrpc.KVStub(channel) def get(self, key): pass def put(self, key, value): put_request = etcdrpc.PutRequest() put_request.key = key.encode('utf-8') put_request.value = value.encode('utf-8') self.kvstub.Put(put_request)
<commit_before>from __future__ import absolute_import __author__ = 'Louis Taylor' __email__ = 'louis@kragniz.eu' __version__ = '0.1.0' import grpc from etcdrpc import rpc_pb2 as etcdrpc channel = grpc.insecure_channel('localhost:2379') stub = etcdrpc.KVStub(channel) put_request = etcdrpc.PutRequest() put_request.key = 'doot'.encode('utf-8') put_request.value = 'toottoot'.encode('utf-8') print(stub.Put(put_request)) class Etcd3Client(object): def __init__(self): pass def get(self, key): pass def put(self, key, value): pass <commit_msg>Move put code into method<commit_after>from __future__ import absolute_import __author__ = 'Louis Taylor' __email__ = 'louis@kragniz.eu' __version__ = '0.1.0' import grpc from etcdrpc import rpc_pb2 as etcdrpc class Etcd3Client(object): def __init__(self): self.channel = grpc.insecure_channel('localhost:2379') self.kvstub = etcdrpc.KVStub(channel) def get(self, key): pass def put(self, key, value): put_request = etcdrpc.PutRequest() put_request.key = key.encode('utf-8') put_request.value = value.encode('utf-8') self.kvstub.Put(put_request)
536211012be24a20c34ef0af1fcc555672129354
byceps/util/system.py
byceps/util/system.py
# -*- coding: utf-8 -*- """ byceps.util.system ~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2016 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ import os CONFIG_ENV_VAR_NAME = 'BYCEPS_CONFIG' def get_config_env_name_from_env(*, default=None): """Return the configuration environment name set via environment variable. Raise an exception if it isn't set. """ env = os.environ.get(CONFIG_ENV_VAR_NAME) if env is None: if default is None: raise Exception( "No configuration environment was specified via the '{}' " "environment variable.".format(CONFIG_ENV_VAR_NAME)) env = default return env
# -*- coding: utf-8 -*- """ byceps.util.system ~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2016 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ import os CONFIG_ENV_VAR_NAME = 'BYCEPS_CONFIG' def get_config_env_name_from_env(): """Return the configuration environment name set via environment variable. Raise an exception if it isn't set. """ env = os.environ.get(CONFIG_ENV_VAR_NAME) if not env: raise Exception( "No configuration environment was specified via the '{}' " "environment variable.".format(CONFIG_ENV_VAR_NAME)) return env
Remove default argument from function that reads the configuration name from the environment
Remove default argument from function that reads the configuration name from the environment
Python
bsd-3-clause
homeworkprod/byceps,m-ober/byceps,m-ober/byceps,m-ober/byceps,homeworkprod/byceps,homeworkprod/byceps
# -*- coding: utf-8 -*- """ byceps.util.system ~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2016 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ import os CONFIG_ENV_VAR_NAME = 'BYCEPS_CONFIG' def get_config_env_name_from_env(*, default=None): """Return the configuration environment name set via environment variable. Raise an exception if it isn't set. """ env = os.environ.get(CONFIG_ENV_VAR_NAME) if env is None: if default is None: raise Exception( "No configuration environment was specified via the '{}' " "environment variable.".format(CONFIG_ENV_VAR_NAME)) env = default return env Remove default argument from function that reads the configuration name from the environment
# -*- coding: utf-8 -*- """ byceps.util.system ~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2016 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ import os CONFIG_ENV_VAR_NAME = 'BYCEPS_CONFIG' def get_config_env_name_from_env(): """Return the configuration environment name set via environment variable. Raise an exception if it isn't set. """ env = os.environ.get(CONFIG_ENV_VAR_NAME) if not env: raise Exception( "No configuration environment was specified via the '{}' " "environment variable.".format(CONFIG_ENV_VAR_NAME)) return env
<commit_before># -*- coding: utf-8 -*- """ byceps.util.system ~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2016 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ import os CONFIG_ENV_VAR_NAME = 'BYCEPS_CONFIG' def get_config_env_name_from_env(*, default=None): """Return the configuration environment name set via environment variable. Raise an exception if it isn't set. """ env = os.environ.get(CONFIG_ENV_VAR_NAME) if env is None: if default is None: raise Exception( "No configuration environment was specified via the '{}' " "environment variable.".format(CONFIG_ENV_VAR_NAME)) env = default return env <commit_msg>Remove default argument from function that reads the configuration name from the environment<commit_after>
# -*- coding: utf-8 -*- """ byceps.util.system ~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2016 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ import os CONFIG_ENV_VAR_NAME = 'BYCEPS_CONFIG' def get_config_env_name_from_env(): """Return the configuration environment name set via environment variable. Raise an exception if it isn't set. """ env = os.environ.get(CONFIG_ENV_VAR_NAME) if not env: raise Exception( "No configuration environment was specified via the '{}' " "environment variable.".format(CONFIG_ENV_VAR_NAME)) return env
# -*- coding: utf-8 -*- """ byceps.util.system ~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2016 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ import os CONFIG_ENV_VAR_NAME = 'BYCEPS_CONFIG' def get_config_env_name_from_env(*, default=None): """Return the configuration environment name set via environment variable. Raise an exception if it isn't set. """ env = os.environ.get(CONFIG_ENV_VAR_NAME) if env is None: if default is None: raise Exception( "No configuration environment was specified via the '{}' " "environment variable.".format(CONFIG_ENV_VAR_NAME)) env = default return env Remove default argument from function that reads the configuration name from the environment# -*- coding: utf-8 -*- """ byceps.util.system ~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2016 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ import os CONFIG_ENV_VAR_NAME = 'BYCEPS_CONFIG' def get_config_env_name_from_env(): """Return the configuration environment name set via environment variable. Raise an exception if it isn't set. """ env = os.environ.get(CONFIG_ENV_VAR_NAME) if not env: raise Exception( "No configuration environment was specified via the '{}' " "environment variable.".format(CONFIG_ENV_VAR_NAME)) return env
<commit_before># -*- coding: utf-8 -*- """ byceps.util.system ~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2016 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ import os CONFIG_ENV_VAR_NAME = 'BYCEPS_CONFIG' def get_config_env_name_from_env(*, default=None): """Return the configuration environment name set via environment variable. Raise an exception if it isn't set. """ env = os.environ.get(CONFIG_ENV_VAR_NAME) if env is None: if default is None: raise Exception( "No configuration environment was specified via the '{}' " "environment variable.".format(CONFIG_ENV_VAR_NAME)) env = default return env <commit_msg>Remove default argument from function that reads the configuration name from the environment<commit_after># -*- coding: utf-8 -*- """ byceps.util.system ~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2016 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ import os CONFIG_ENV_VAR_NAME = 'BYCEPS_CONFIG' def get_config_env_name_from_env(): """Return the configuration environment name set via environment variable. Raise an exception if it isn't set. """ env = os.environ.get(CONFIG_ENV_VAR_NAME) if not env: raise Exception( "No configuration environment was specified via the '{}' " "environment variable.".format(CONFIG_ENV_VAR_NAME)) return env
703ba21925f2e7ad9c5bb970ac89e834676962a8
api/rest.py
api/rest.py
#!/usr/bin/env python from mcapi.mcapp import app from mcapi import tservices, public, utils, private, access from mcapi.user import account, datadirs, datafiles, reviews, ud import sys if __name__ == '__main__': if len(sys.argv) >= 2: debug = True else: debug = False if len(sys.argv) == 3: app.run(debug=debug, host='0.0.0.0') else: app.run(debug=debug)
#!/usr/bin/env python from mcapi.mcapp import app from mcapi import tservices, public, utils, private, access, process, machine, template from mcapi.user import account, datadirs, datafiles, reviews, ud import sys if __name__ == '__main__': if len(sys.argv) >= 2: debug = True else: debug = False if len(sys.argv) == 3: app.run(debug=debug, host='0.0.0.0') else: app.run(debug=debug)
Add in references to new services.
Add in references to new services.
Python
mit
materials-commons/materialscommons.org,materials-commons/materialscommons.org,materials-commons/materialscommons.org,materials-commons/materialscommons.org,materials-commons/materialscommons.org
#!/usr/bin/env python from mcapi.mcapp import app from mcapi import tservices, public, utils, private, access from mcapi.user import account, datadirs, datafiles, reviews, ud import sys if __name__ == '__main__': if len(sys.argv) >= 2: debug = True else: debug = False if len(sys.argv) == 3: app.run(debug=debug, host='0.0.0.0') else: app.run(debug=debug) Add in references to new services.
#!/usr/bin/env python from mcapi.mcapp import app from mcapi import tservices, public, utils, private, access, process, machine, template from mcapi.user import account, datadirs, datafiles, reviews, ud import sys if __name__ == '__main__': if len(sys.argv) >= 2: debug = True else: debug = False if len(sys.argv) == 3: app.run(debug=debug, host='0.0.0.0') else: app.run(debug=debug)
<commit_before>#!/usr/bin/env python from mcapi.mcapp import app from mcapi import tservices, public, utils, private, access from mcapi.user import account, datadirs, datafiles, reviews, ud import sys if __name__ == '__main__': if len(sys.argv) >= 2: debug = True else: debug = False if len(sys.argv) == 3: app.run(debug=debug, host='0.0.0.0') else: app.run(debug=debug) <commit_msg>Add in references to new services.<commit_after>
#!/usr/bin/env python from mcapi.mcapp import app from mcapi import tservices, public, utils, private, access, process, machine, template from mcapi.user import account, datadirs, datafiles, reviews, ud import sys if __name__ == '__main__': if len(sys.argv) >= 2: debug = True else: debug = False if len(sys.argv) == 3: app.run(debug=debug, host='0.0.0.0') else: app.run(debug=debug)
#!/usr/bin/env python from mcapi.mcapp import app from mcapi import tservices, public, utils, private, access from mcapi.user import account, datadirs, datafiles, reviews, ud import sys if __name__ == '__main__': if len(sys.argv) >= 2: debug = True else: debug = False if len(sys.argv) == 3: app.run(debug=debug, host='0.0.0.0') else: app.run(debug=debug) Add in references to new services.#!/usr/bin/env python from mcapi.mcapp import app from mcapi import tservices, public, utils, private, access, process, machine, template from mcapi.user import account, datadirs, datafiles, reviews, ud import sys if __name__ == '__main__': if len(sys.argv) >= 2: debug = True else: debug = False if len(sys.argv) == 3: app.run(debug=debug, host='0.0.0.0') else: app.run(debug=debug)
<commit_before>#!/usr/bin/env python from mcapi.mcapp import app from mcapi import tservices, public, utils, private, access from mcapi.user import account, datadirs, datafiles, reviews, ud import sys if __name__ == '__main__': if len(sys.argv) >= 2: debug = True else: debug = False if len(sys.argv) == 3: app.run(debug=debug, host='0.0.0.0') else: app.run(debug=debug) <commit_msg>Add in references to new services.<commit_after>#!/usr/bin/env python from mcapi.mcapp import app from mcapi import tservices, public, utils, private, access, process, machine, template from mcapi.user import account, datadirs, datafiles, reviews, ud import sys if __name__ == '__main__': if len(sys.argv) >= 2: debug = True else: debug = False if len(sys.argv) == 3: app.run(debug=debug, host='0.0.0.0') else: app.run(debug=debug)
87153cb1a9727d17d31f3aabb28affddca3191bf
sqltocpp.py
sqltocpp.py
import click from sqltocpp import convert @click.command() @click.option('--sql', help='schema file name') @click.option('--target', default='schema.hpp', help='hpp file name') def execute(sql, target): convert.schema_to_struct(sql, target) if __name__ == '__main__': try: execute() except: execute("--help")
import click from sqltocpp import convert @click.command() @click.argument('sql_schema_file') @click.option('--target', default='schema.hpp', help='hpp file name') def execute(sql_schema_file, target): convert.schema_to_struct(sql_schema_file, target) if __name__ == '__main__': execute()
Add click based commandline interface script
Add click based commandline interface script sqltocpp is intended to be a CLI command. This enables it to be so
Python
mit
banjocat/SqlToCpp,banjocat/SqlToCpp
import click from sqltocpp import convert @click.command() @click.option('--sql', help='schema file name') @click.option('--target', default='schema.hpp', help='hpp file name') def execute(sql, target): convert.schema_to_struct(sql, target) if __name__ == '__main__': try: execute() except: execute("--help") Add click based commandline interface script sqltocpp is intended to be a CLI command. This enables it to be so
import click from sqltocpp import convert @click.command() @click.argument('sql_schema_file') @click.option('--target', default='schema.hpp', help='hpp file name') def execute(sql_schema_file, target): convert.schema_to_struct(sql_schema_file, target) if __name__ == '__main__': execute()
<commit_before>import click from sqltocpp import convert @click.command() @click.option('--sql', help='schema file name') @click.option('--target', default='schema.hpp', help='hpp file name') def execute(sql, target): convert.schema_to_struct(sql, target) if __name__ == '__main__': try: execute() except: execute("--help") <commit_msg>Add click based commandline interface script sqltocpp is intended to be a CLI command. This enables it to be so<commit_after>
import click from sqltocpp import convert @click.command() @click.argument('sql_schema_file') @click.option('--target', default='schema.hpp', help='hpp file name') def execute(sql_schema_file, target): convert.schema_to_struct(sql_schema_file, target) if __name__ == '__main__': execute()
import click from sqltocpp import convert @click.command() @click.option('--sql', help='schema file name') @click.option('--target', default='schema.hpp', help='hpp file name') def execute(sql, target): convert.schema_to_struct(sql, target) if __name__ == '__main__': try: execute() except: execute("--help") Add click based commandline interface script sqltocpp is intended to be a CLI command. This enables it to be soimport click from sqltocpp import convert @click.command() @click.argument('sql_schema_file') @click.option('--target', default='schema.hpp', help='hpp file name') def execute(sql_schema_file, target): convert.schema_to_struct(sql_schema_file, target) if __name__ == '__main__': execute()
<commit_before>import click from sqltocpp import convert @click.command() @click.option('--sql', help='schema file name') @click.option('--target', default='schema.hpp', help='hpp file name') def execute(sql, target): convert.schema_to_struct(sql, target) if __name__ == '__main__': try: execute() except: execute("--help") <commit_msg>Add click based commandline interface script sqltocpp is intended to be a CLI command. This enables it to be so<commit_after>import click from sqltocpp import convert @click.command() @click.argument('sql_schema_file') @click.option('--target', default='schema.hpp', help='hpp file name') def execute(sql_schema_file, target): convert.schema_to_struct(sql_schema_file, target) if __name__ == '__main__': execute()
fe2300d809d863b63f3100ff72757f48cb11789b
nn_patterns/explainer/__init__.py
nn_patterns/explainer/__init__.py
from .base import * from .gradient_based import * from .misc import * from .pattern_based import * from .relevance_based import * def create_explainer(name, output_layer, patterns=None, to_layer=None, **kwargs): return { # Utility. "input": InputExplainer, "random": RandomExplainer, # Gradient based "gradient": GradientExplainer, "deconvnet": DeConvNetExplainer, "guided": GuidedBackpropExplainer, "gradient.alt": AlternativeGradientExplainer, # Relevance based "lrp.z": LRPZExplainer, "lrp.eps": LRPEpsExplainer, # Pattern based "patternnet": PatternNetExplainer, "patternnet.guided": GuidedPatternNetExplainer, "patternlrp": PatternLRPExplainer, }[name](output_layer, patterns=patterns, to_layer=to_layer, **kwargs)
from .base import * from .gradient_based import * from .misc import * from .pattern_based import * from .relevance_based import * def create_explainer(name, output_layer, patterns=None, **kwargs): return { # Utility. "input": InputExplainer, "random": RandomExplainer, # Gradient based "gradient": GradientExplainer, "deconvnet": DeConvNetExplainer, "guided": GuidedBackpropExplainer, "gradient.alt": AlternativeGradientExplainer, # Relevance based "lrp.z": LRPZExplainer, "lrp.eps": LRPEpsExplainer, # Pattern based "patternnet": PatternNetExplainer, "patternnet.guided": GuidedPatternNetExplainer, "patternlrp": PatternLRPExplainer, }[name](output_layer, patterns=patterns, **kwargs)
Fix interface. to_layer parameter is obsolete.
Fix interface. to_layer parameter is obsolete.
Python
mit
pikinder/nn-patterns
from .base import * from .gradient_based import * from .misc import * from .pattern_based import * from .relevance_based import * def create_explainer(name, output_layer, patterns=None, to_layer=None, **kwargs): return { # Utility. "input": InputExplainer, "random": RandomExplainer, # Gradient based "gradient": GradientExplainer, "deconvnet": DeConvNetExplainer, "guided": GuidedBackpropExplainer, "gradient.alt": AlternativeGradientExplainer, # Relevance based "lrp.z": LRPZExplainer, "lrp.eps": LRPEpsExplainer, # Pattern based "patternnet": PatternNetExplainer, "patternnet.guided": GuidedPatternNetExplainer, "patternlrp": PatternLRPExplainer, }[name](output_layer, patterns=patterns, to_layer=to_layer, **kwargs) Fix interface. to_layer parameter is obsolete.
from .base import * from .gradient_based import * from .misc import * from .pattern_based import * from .relevance_based import * def create_explainer(name, output_layer, patterns=None, **kwargs): return { # Utility. "input": InputExplainer, "random": RandomExplainer, # Gradient based "gradient": GradientExplainer, "deconvnet": DeConvNetExplainer, "guided": GuidedBackpropExplainer, "gradient.alt": AlternativeGradientExplainer, # Relevance based "lrp.z": LRPZExplainer, "lrp.eps": LRPEpsExplainer, # Pattern based "patternnet": PatternNetExplainer, "patternnet.guided": GuidedPatternNetExplainer, "patternlrp": PatternLRPExplainer, }[name](output_layer, patterns=patterns, **kwargs)
<commit_before> from .base import * from .gradient_based import * from .misc import * from .pattern_based import * from .relevance_based import * def create_explainer(name, output_layer, patterns=None, to_layer=None, **kwargs): return { # Utility. "input": InputExplainer, "random": RandomExplainer, # Gradient based "gradient": GradientExplainer, "deconvnet": DeConvNetExplainer, "guided": GuidedBackpropExplainer, "gradient.alt": AlternativeGradientExplainer, # Relevance based "lrp.z": LRPZExplainer, "lrp.eps": LRPEpsExplainer, # Pattern based "patternnet": PatternNetExplainer, "patternnet.guided": GuidedPatternNetExplainer, "patternlrp": PatternLRPExplainer, }[name](output_layer, patterns=patterns, to_layer=to_layer, **kwargs) <commit_msg>Fix interface. to_layer parameter is obsolete.<commit_after>
from .base import * from .gradient_based import * from .misc import * from .pattern_based import * from .relevance_based import * def create_explainer(name, output_layer, patterns=None, **kwargs): return { # Utility. "input": InputExplainer, "random": RandomExplainer, # Gradient based "gradient": GradientExplainer, "deconvnet": DeConvNetExplainer, "guided": GuidedBackpropExplainer, "gradient.alt": AlternativeGradientExplainer, # Relevance based "lrp.z": LRPZExplainer, "lrp.eps": LRPEpsExplainer, # Pattern based "patternnet": PatternNetExplainer, "patternnet.guided": GuidedPatternNetExplainer, "patternlrp": PatternLRPExplainer, }[name](output_layer, patterns=patterns, **kwargs)
from .base import * from .gradient_based import * from .misc import * from .pattern_based import * from .relevance_based import * def create_explainer(name, output_layer, patterns=None, to_layer=None, **kwargs): return { # Utility. "input": InputExplainer, "random": RandomExplainer, # Gradient based "gradient": GradientExplainer, "deconvnet": DeConvNetExplainer, "guided": GuidedBackpropExplainer, "gradient.alt": AlternativeGradientExplainer, # Relevance based "lrp.z": LRPZExplainer, "lrp.eps": LRPEpsExplainer, # Pattern based "patternnet": PatternNetExplainer, "patternnet.guided": GuidedPatternNetExplainer, "patternlrp": PatternLRPExplainer, }[name](output_layer, patterns=patterns, to_layer=to_layer, **kwargs) Fix interface. to_layer parameter is obsolete. from .base import * from .gradient_based import * from .misc import * from .pattern_based import * from .relevance_based import * def create_explainer(name, output_layer, patterns=None, **kwargs): return { # Utility. "input": InputExplainer, "random": RandomExplainer, # Gradient based "gradient": GradientExplainer, "deconvnet": DeConvNetExplainer, "guided": GuidedBackpropExplainer, "gradient.alt": AlternativeGradientExplainer, # Relevance based "lrp.z": LRPZExplainer, "lrp.eps": LRPEpsExplainer, # Pattern based "patternnet": PatternNetExplainer, "patternnet.guided": GuidedPatternNetExplainer, "patternlrp": PatternLRPExplainer, }[name](output_layer, patterns=patterns, **kwargs)
<commit_before> from .base import * from .gradient_based import * from .misc import * from .pattern_based import * from .relevance_based import * def create_explainer(name, output_layer, patterns=None, to_layer=None, **kwargs): return { # Utility. "input": InputExplainer, "random": RandomExplainer, # Gradient based "gradient": GradientExplainer, "deconvnet": DeConvNetExplainer, "guided": GuidedBackpropExplainer, "gradient.alt": AlternativeGradientExplainer, # Relevance based "lrp.z": LRPZExplainer, "lrp.eps": LRPEpsExplainer, # Pattern based "patternnet": PatternNetExplainer, "patternnet.guided": GuidedPatternNetExplainer, "patternlrp": PatternLRPExplainer, }[name](output_layer, patterns=patterns, to_layer=to_layer, **kwargs) <commit_msg>Fix interface. to_layer parameter is obsolete.<commit_after> from .base import * from .gradient_based import * from .misc import * from .pattern_based import * from .relevance_based import * def create_explainer(name, output_layer, patterns=None, **kwargs): return { # Utility. "input": InputExplainer, "random": RandomExplainer, # Gradient based "gradient": GradientExplainer, "deconvnet": DeConvNetExplainer, "guided": GuidedBackpropExplainer, "gradient.alt": AlternativeGradientExplainer, # Relevance based "lrp.z": LRPZExplainer, "lrp.eps": LRPEpsExplainer, # Pattern based "patternnet": PatternNetExplainer, "patternnet.guided": GuidedPatternNetExplainer, "patternlrp": PatternLRPExplainer, }[name](output_layer, patterns=patterns, **kwargs)
267c17ce984952d16623b0305975626397529ca8
tests/config_test.py
tests/config_test.py
import pytest from timewreport.config import TimeWarriorConfig def test_get_value_should_return_value_if_key_available(): config = TimeWarriorConfig({'FOO': 'foo'}) assert config.get_value('FOO', 'bar') == 'foo' def test_get_value_should_return_default_if_key_not_available(): config = TimeWarriorConfig({'BAR': 'foo'}) assert config.get_value('FOO', 'bar') == 'bar' @pytest.fixture(scope='function', params=['on', 1, 'yes', 'y', 'true']) def trueish_value(request): return request.param def test_get_boolean_should_return_true_on_trueish_values(trueish_value): config = TimeWarriorConfig({'KEY': trueish_value}) assert config.get_boolean('KEY', False) is True def test_get_boolean_should_return_false_on_falseish_values(): config = TimeWarriorConfig({'KEY': 'foo'}) assert config.get_boolean('KEY', True) is False
import pytest from timewreport.config import TimeWarriorConfig def test_get_value_should_return_value_if_key_available(): config = TimeWarriorConfig({'FOO': 'foo'}) assert config.get_value('FOO', 'bar') == 'foo' def test_get_value_should_return_default_if_key_not_available(): config = TimeWarriorConfig({'BAR': 'foo'}) assert config.get_value('FOO', 'bar') == 'bar' @pytest.fixture(scope='function', params=['on', 1, 'yes', 'y', 'true']) def trueish_value(request): return request.param def test_get_boolean_should_return_true_on_trueish_values(trueish_value): config = TimeWarriorConfig({'KEY': trueish_value}) assert config.get_boolean('KEY', False) is True @pytest.fixture(scope='function', params=['off', 0, 'no', 'n', 'false']) def falseish_value(request): return request.param def test_get_boolean_should_return_false_on_falseish_values(falseish_value): config = TimeWarriorConfig({'KEY': falseish_value}) assert config.get_boolean('KEY', True) is False
Add tests for falseish config values
Add tests for falseish config values
Python
mit
lauft/timew-report
import pytest from timewreport.config import TimeWarriorConfig def test_get_value_should_return_value_if_key_available(): config = TimeWarriorConfig({'FOO': 'foo'}) assert config.get_value('FOO', 'bar') == 'foo' def test_get_value_should_return_default_if_key_not_available(): config = TimeWarriorConfig({'BAR': 'foo'}) assert config.get_value('FOO', 'bar') == 'bar' @pytest.fixture(scope='function', params=['on', 1, 'yes', 'y', 'true']) def trueish_value(request): return request.param def test_get_boolean_should_return_true_on_trueish_values(trueish_value): config = TimeWarriorConfig({'KEY': trueish_value}) assert config.get_boolean('KEY', False) is True def test_get_boolean_should_return_false_on_falseish_values(): config = TimeWarriorConfig({'KEY': 'foo'}) assert config.get_boolean('KEY', True) is False Add tests for falseish config values
import pytest from timewreport.config import TimeWarriorConfig def test_get_value_should_return_value_if_key_available(): config = TimeWarriorConfig({'FOO': 'foo'}) assert config.get_value('FOO', 'bar') == 'foo' def test_get_value_should_return_default_if_key_not_available(): config = TimeWarriorConfig({'BAR': 'foo'}) assert config.get_value('FOO', 'bar') == 'bar' @pytest.fixture(scope='function', params=['on', 1, 'yes', 'y', 'true']) def trueish_value(request): return request.param def test_get_boolean_should_return_true_on_trueish_values(trueish_value): config = TimeWarriorConfig({'KEY': trueish_value}) assert config.get_boolean('KEY', False) is True @pytest.fixture(scope='function', params=['off', 0, 'no', 'n', 'false']) def falseish_value(request): return request.param def test_get_boolean_should_return_false_on_falseish_values(falseish_value): config = TimeWarriorConfig({'KEY': falseish_value}) assert config.get_boolean('KEY', True) is False
<commit_before>import pytest from timewreport.config import TimeWarriorConfig def test_get_value_should_return_value_if_key_available(): config = TimeWarriorConfig({'FOO': 'foo'}) assert config.get_value('FOO', 'bar') == 'foo' def test_get_value_should_return_default_if_key_not_available(): config = TimeWarriorConfig({'BAR': 'foo'}) assert config.get_value('FOO', 'bar') == 'bar' @pytest.fixture(scope='function', params=['on', 1, 'yes', 'y', 'true']) def trueish_value(request): return request.param def test_get_boolean_should_return_true_on_trueish_values(trueish_value): config = TimeWarriorConfig({'KEY': trueish_value}) assert config.get_boolean('KEY', False) is True def test_get_boolean_should_return_false_on_falseish_values(): config = TimeWarriorConfig({'KEY': 'foo'}) assert config.get_boolean('KEY', True) is False <commit_msg>Add tests for falseish config values<commit_after>
import pytest from timewreport.config import TimeWarriorConfig def test_get_value_should_return_value_if_key_available(): config = TimeWarriorConfig({'FOO': 'foo'}) assert config.get_value('FOO', 'bar') == 'foo' def test_get_value_should_return_default_if_key_not_available(): config = TimeWarriorConfig({'BAR': 'foo'}) assert config.get_value('FOO', 'bar') == 'bar' @pytest.fixture(scope='function', params=['on', 1, 'yes', 'y', 'true']) def trueish_value(request): return request.param def test_get_boolean_should_return_true_on_trueish_values(trueish_value): config = TimeWarriorConfig({'KEY': trueish_value}) assert config.get_boolean('KEY', False) is True @pytest.fixture(scope='function', params=['off', 0, 'no', 'n', 'false']) def falseish_value(request): return request.param def test_get_boolean_should_return_false_on_falseish_values(falseish_value): config = TimeWarriorConfig({'KEY': falseish_value}) assert config.get_boolean('KEY', True) is False
import pytest from timewreport.config import TimeWarriorConfig def test_get_value_should_return_value_if_key_available(): config = TimeWarriorConfig({'FOO': 'foo'}) assert config.get_value('FOO', 'bar') == 'foo' def test_get_value_should_return_default_if_key_not_available(): config = TimeWarriorConfig({'BAR': 'foo'}) assert config.get_value('FOO', 'bar') == 'bar' @pytest.fixture(scope='function', params=['on', 1, 'yes', 'y', 'true']) def trueish_value(request): return request.param def test_get_boolean_should_return_true_on_trueish_values(trueish_value): config = TimeWarriorConfig({'KEY': trueish_value}) assert config.get_boolean('KEY', False) is True def test_get_boolean_should_return_false_on_falseish_values(): config = TimeWarriorConfig({'KEY': 'foo'}) assert config.get_boolean('KEY', True) is False Add tests for falseish config valuesimport pytest from timewreport.config import TimeWarriorConfig def test_get_value_should_return_value_if_key_available(): config = TimeWarriorConfig({'FOO': 'foo'}) assert config.get_value('FOO', 'bar') == 'foo' def test_get_value_should_return_default_if_key_not_available(): config = TimeWarriorConfig({'BAR': 'foo'}) assert config.get_value('FOO', 'bar') == 'bar' @pytest.fixture(scope='function', params=['on', 1, 'yes', 'y', 'true']) def trueish_value(request): return request.param def test_get_boolean_should_return_true_on_trueish_values(trueish_value): config = TimeWarriorConfig({'KEY': trueish_value}) assert config.get_boolean('KEY', False) is True @pytest.fixture(scope='function', params=['off', 0, 'no', 'n', 'false']) def falseish_value(request): return request.param def test_get_boolean_should_return_false_on_falseish_values(falseish_value): config = TimeWarriorConfig({'KEY': falseish_value}) assert config.get_boolean('KEY', True) is False
<commit_before>import pytest from timewreport.config import TimeWarriorConfig def test_get_value_should_return_value_if_key_available(): config = TimeWarriorConfig({'FOO': 'foo'}) assert config.get_value('FOO', 'bar') == 'foo' def test_get_value_should_return_default_if_key_not_available(): config = TimeWarriorConfig({'BAR': 'foo'}) assert config.get_value('FOO', 'bar') == 'bar' @pytest.fixture(scope='function', params=['on', 1, 'yes', 'y', 'true']) def trueish_value(request): return request.param def test_get_boolean_should_return_true_on_trueish_values(trueish_value): config = TimeWarriorConfig({'KEY': trueish_value}) assert config.get_boolean('KEY', False) is True def test_get_boolean_should_return_false_on_falseish_values(): config = TimeWarriorConfig({'KEY': 'foo'}) assert config.get_boolean('KEY', True) is False <commit_msg>Add tests for falseish config values<commit_after>import pytest from timewreport.config import TimeWarriorConfig def test_get_value_should_return_value_if_key_available(): config = TimeWarriorConfig({'FOO': 'foo'}) assert config.get_value('FOO', 'bar') == 'foo' def test_get_value_should_return_default_if_key_not_available(): config = TimeWarriorConfig({'BAR': 'foo'}) assert config.get_value('FOO', 'bar') == 'bar' @pytest.fixture(scope='function', params=['on', 1, 'yes', 'y', 'true']) def trueish_value(request): return request.param def test_get_boolean_should_return_true_on_trueish_values(trueish_value): config = TimeWarriorConfig({'KEY': trueish_value}) assert config.get_boolean('KEY', False) is True @pytest.fixture(scope='function', params=['off', 0, 'no', 'n', 'false']) def falseish_value(request): return request.param def test_get_boolean_should_return_false_on_falseish_values(falseish_value): config = TimeWarriorConfig({'KEY': falseish_value}) assert config.get_boolean('KEY', True) is False
6c157525bc32f1e6005be69bd6fde61d0d002ad3
wizard/post_function.py
wizard/post_function.py
from openerp import pooler def call_post_function(cr, uid, context): """This functionality allows users of module account.move.reversal to call a function of the desired openerp model, after the reversal of the move. The call automatically sends at least the database cursor (cr) and the user id (uid) for security reasons. Two key parameters are required in the context to do so: - 'post_function_obj': the osv model where the function is defined, - 'post_function_name': the name of the function to call, And two optional key parameters: - 'post_function_args': an iterable object listing the required arguments to pass after 'cr, uid', - 'post_function_kwargs': a dictionary object listing the optionnal keyword args to pass. """ if 'post_function_obj' in context: # We get the function addr by its name, # and call it with (cr, uid, *args, **kwargs) getattr( pooler.get_pool(cr.dbname)[context['post_function_obj']], context['post_function_name'] )( cr, uid, *context['post_function_args'], **context['post_function_kwargs'] ) # We clean the context to avoid multiple calls of the function. context.pop('post_function_obj') context.pop('post_function_name') context.pop('post_function_args') context.pop('post_function_kwargs')
from openerp import pooler def call_post_function(cr, uid, context): """This functionality allows users of module account.move.reversal to call a function of the desired openerp model, after the reversal of the move. The call automatically sends at least the database cursor (cr) and the user id (uid) for security reasons. Two key parameters are required in the context to do so: - 'post_function_obj': the osv model where the function is defined, - 'post_function_name': the name of the function to call, And two optional key parameters: - 'post_function_args': an iterable object listing the required arguments to pass after 'cr, uid', - 'post_function_kwargs': a dictionary object listing the optionnal keyword args to pass. """ if 'post_function_obj' in context: # We get the function addr by its name, # and call it with (cr, uid, *args, **kwargs) getattr( pooler.get_pool(cr.dbname)[context['post_function_obj']], context['post_function_name'] )( cr, uid, *context.get('post_function_args', []), **context.get('post_function_kwargs', {}) ) # We clean the context to avoid multiple calls of the function. context.pop('post_function_obj') context.pop('post_function_name') context.pop('post_function_args') context.pop('post_function_kwargs')
Remove some required arguments in post function context
Remove some required arguments in post function context
Python
agpl-3.0
xcgd/account_move_reversal
from openerp import pooler def call_post_function(cr, uid, context): """This functionality allows users of module account.move.reversal to call a function of the desired openerp model, after the reversal of the move. The call automatically sends at least the database cursor (cr) and the user id (uid) for security reasons. Two key parameters are required in the context to do so: - 'post_function_obj': the osv model where the function is defined, - 'post_function_name': the name of the function to call, And two optional key parameters: - 'post_function_args': an iterable object listing the required arguments to pass after 'cr, uid', - 'post_function_kwargs': a dictionary object listing the optionnal keyword args to pass. """ if 'post_function_obj' in context: # We get the function addr by its name, # and call it with (cr, uid, *args, **kwargs) getattr( pooler.get_pool(cr.dbname)[context['post_function_obj']], context['post_function_name'] )( cr, uid, *context['post_function_args'], **context['post_function_kwargs'] ) # We clean the context to avoid multiple calls of the function. context.pop('post_function_obj') context.pop('post_function_name') context.pop('post_function_args') context.pop('post_function_kwargs') Remove some required arguments in post function context
from openerp import pooler def call_post_function(cr, uid, context): """This functionality allows users of module account.move.reversal to call a function of the desired openerp model, after the reversal of the move. The call automatically sends at least the database cursor (cr) and the user id (uid) for security reasons. Two key parameters are required in the context to do so: - 'post_function_obj': the osv model where the function is defined, - 'post_function_name': the name of the function to call, And two optional key parameters: - 'post_function_args': an iterable object listing the required arguments to pass after 'cr, uid', - 'post_function_kwargs': a dictionary object listing the optionnal keyword args to pass. """ if 'post_function_obj' in context: # We get the function addr by its name, # and call it with (cr, uid, *args, **kwargs) getattr( pooler.get_pool(cr.dbname)[context['post_function_obj']], context['post_function_name'] )( cr, uid, *context.get('post_function_args', []), **context.get('post_function_kwargs', {}) ) # We clean the context to avoid multiple calls of the function. context.pop('post_function_obj') context.pop('post_function_name') context.pop('post_function_args') context.pop('post_function_kwargs')
<commit_before>from openerp import pooler def call_post_function(cr, uid, context): """This functionality allows users of module account.move.reversal to call a function of the desired openerp model, after the reversal of the move. The call automatically sends at least the database cursor (cr) and the user id (uid) for security reasons. Two key parameters are required in the context to do so: - 'post_function_obj': the osv model where the function is defined, - 'post_function_name': the name of the function to call, And two optional key parameters: - 'post_function_args': an iterable object listing the required arguments to pass after 'cr, uid', - 'post_function_kwargs': a dictionary object listing the optionnal keyword args to pass. """ if 'post_function_obj' in context: # We get the function addr by its name, # and call it with (cr, uid, *args, **kwargs) getattr( pooler.get_pool(cr.dbname)[context['post_function_obj']], context['post_function_name'] )( cr, uid, *context['post_function_args'], **context['post_function_kwargs'] ) # We clean the context to avoid multiple calls of the function. context.pop('post_function_obj') context.pop('post_function_name') context.pop('post_function_args') context.pop('post_function_kwargs') <commit_msg>Remove some required arguments in post function context<commit_after>
from openerp import pooler def call_post_function(cr, uid, context): """This functionality allows users of module account.move.reversal to call a function of the desired openerp model, after the reversal of the move. The call automatically sends at least the database cursor (cr) and the user id (uid) for security reasons. Two key parameters are required in the context to do so: - 'post_function_obj': the osv model where the function is defined, - 'post_function_name': the name of the function to call, And two optional key parameters: - 'post_function_args': an iterable object listing the required arguments to pass after 'cr, uid', - 'post_function_kwargs': a dictionary object listing the optionnal keyword args to pass. """ if 'post_function_obj' in context: # We get the function addr by its name, # and call it with (cr, uid, *args, **kwargs) getattr( pooler.get_pool(cr.dbname)[context['post_function_obj']], context['post_function_name'] )( cr, uid, *context.get('post_function_args', []), **context.get('post_function_kwargs', {}) ) # We clean the context to avoid multiple calls of the function. context.pop('post_function_obj') context.pop('post_function_name') context.pop('post_function_args') context.pop('post_function_kwargs')
from openerp import pooler def call_post_function(cr, uid, context): """This functionality allows users of module account.move.reversal to call a function of the desired openerp model, after the reversal of the move. The call automatically sends at least the database cursor (cr) and the user id (uid) for security reasons. Two key parameters are required in the context to do so: - 'post_function_obj': the osv model where the function is defined, - 'post_function_name': the name of the function to call, And two optional key parameters: - 'post_function_args': an iterable object listing the required arguments to pass after 'cr, uid', - 'post_function_kwargs': a dictionary object listing the optionnal keyword args to pass. """ if 'post_function_obj' in context: # We get the function addr by its name, # and call it with (cr, uid, *args, **kwargs) getattr( pooler.get_pool(cr.dbname)[context['post_function_obj']], context['post_function_name'] )( cr, uid, *context['post_function_args'], **context['post_function_kwargs'] ) # We clean the context to avoid multiple calls of the function. context.pop('post_function_obj') context.pop('post_function_name') context.pop('post_function_args') context.pop('post_function_kwargs') Remove some required arguments in post function contextfrom openerp import pooler def call_post_function(cr, uid, context): """This functionality allows users of module account.move.reversal to call a function of the desired openerp model, after the reversal of the move. The call automatically sends at least the database cursor (cr) and the user id (uid) for security reasons. Two key parameters are required in the context to do so: - 'post_function_obj': the osv model where the function is defined, - 'post_function_name': the name of the function to call, And two optional key parameters: - 'post_function_args': an iterable object listing the required arguments to pass after 'cr, uid', - 'post_function_kwargs': a dictionary object listing the optionnal keyword args to pass. """ if 'post_function_obj' in context: # We get the function addr by its name, # and call it with (cr, uid, *args, **kwargs) getattr( pooler.get_pool(cr.dbname)[context['post_function_obj']], context['post_function_name'] )( cr, uid, *context.get('post_function_args', []), **context.get('post_function_kwargs', {}) ) # We clean the context to avoid multiple calls of the function. context.pop('post_function_obj') context.pop('post_function_name') context.pop('post_function_args') context.pop('post_function_kwargs')
<commit_before>from openerp import pooler def call_post_function(cr, uid, context): """This functionality allows users of module account.move.reversal to call a function of the desired openerp model, after the reversal of the move. The call automatically sends at least the database cursor (cr) and the user id (uid) for security reasons. Two key parameters are required in the context to do so: - 'post_function_obj': the osv model where the function is defined, - 'post_function_name': the name of the function to call, And two optional key parameters: - 'post_function_args': an iterable object listing the required arguments to pass after 'cr, uid', - 'post_function_kwargs': a dictionary object listing the optionnal keyword args to pass. """ if 'post_function_obj' in context: # We get the function addr by its name, # and call it with (cr, uid, *args, **kwargs) getattr( pooler.get_pool(cr.dbname)[context['post_function_obj']], context['post_function_name'] )( cr, uid, *context['post_function_args'], **context['post_function_kwargs'] ) # We clean the context to avoid multiple calls of the function. context.pop('post_function_obj') context.pop('post_function_name') context.pop('post_function_args') context.pop('post_function_kwargs') <commit_msg>Remove some required arguments in post function context<commit_after>from openerp import pooler def call_post_function(cr, uid, context): """This functionality allows users of module account.move.reversal to call a function of the desired openerp model, after the reversal of the move. The call automatically sends at least the database cursor (cr) and the user id (uid) for security reasons. Two key parameters are required in the context to do so: - 'post_function_obj': the osv model where the function is defined, - 'post_function_name': the name of the function to call, And two optional key parameters: - 'post_function_args': an iterable object listing the required arguments to pass after 'cr, uid', - 'post_function_kwargs': a dictionary object listing the optionnal keyword args to pass. """ if 'post_function_obj' in context: # We get the function addr by its name, # and call it with (cr, uid, *args, **kwargs) getattr( pooler.get_pool(cr.dbname)[context['post_function_obj']], context['post_function_name'] )( cr, uid, *context.get('post_function_args', []), **context.get('post_function_kwargs', {}) ) # We clean the context to avoid multiple calls of the function. context.pop('post_function_obj') context.pop('post_function_name') context.pop('post_function_args') context.pop('post_function_kwargs')
3271722d3905a7727c20989fa98d804cb4df1b82
mysite/urls.py
mysite/urls.py
"""vote URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.11/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import include, url from django.contrib import admin urlpatterns = [ url(r'^polls/', include('polls.urls')), url(r'^admin/', admin.site.urls), ]
"""vote URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.11/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import include, url from django.contrib import admin from django.http.response import HttpResponseRedirect urlpatterns = [ url(r'^polls/', include('polls.urls')), url(r'^admin/', admin.site.urls), url(r'^$', lambda r: HttpResponseRedirect('polls/')), ]
Add redirect from / to /polls
Add redirect from / to /polls
Python
apache-2.0
gerard-/votingapp,gerard-/votingapp
"""vote URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.11/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import include, url from django.contrib import admin urlpatterns = [ url(r'^polls/', include('polls.urls')), url(r'^admin/', admin.site.urls), ] Add redirect from / to /polls
"""vote URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.11/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import include, url from django.contrib import admin from django.http.response import HttpResponseRedirect urlpatterns = [ url(r'^polls/', include('polls.urls')), url(r'^admin/', admin.site.urls), url(r'^$', lambda r: HttpResponseRedirect('polls/')), ]
<commit_before>"""vote URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.11/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import include, url from django.contrib import admin urlpatterns = [ url(r'^polls/', include('polls.urls')), url(r'^admin/', admin.site.urls), ] <commit_msg>Add redirect from / to /polls<commit_after>
"""vote URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.11/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import include, url from django.contrib import admin from django.http.response import HttpResponseRedirect urlpatterns = [ url(r'^polls/', include('polls.urls')), url(r'^admin/', admin.site.urls), url(r'^$', lambda r: HttpResponseRedirect('polls/')), ]
"""vote URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.11/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import include, url from django.contrib import admin urlpatterns = [ url(r'^polls/', include('polls.urls')), url(r'^admin/', admin.site.urls), ] Add redirect from / to /polls"""vote URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.11/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import include, url from django.contrib import admin from django.http.response import HttpResponseRedirect urlpatterns = [ url(r'^polls/', include('polls.urls')), url(r'^admin/', admin.site.urls), url(r'^$', lambda r: HttpResponseRedirect('polls/')), ]
<commit_before>"""vote URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.11/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import include, url from django.contrib import admin urlpatterns = [ url(r'^polls/', include('polls.urls')), url(r'^admin/', admin.site.urls), ] <commit_msg>Add redirect from / to /polls<commit_after>"""vote URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.11/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import include, url from django.contrib import admin from django.http.response import HttpResponseRedirect urlpatterns = [ url(r'^polls/', include('polls.urls')), url(r'^admin/', admin.site.urls), url(r'^$', lambda r: HttpResponseRedirect('polls/')), ]
f405ee262ab6f87df8e83e024d3c615842fa37ce
fandjango/urls.py
fandjango/urls.py
from django.conf.urls.defaults import * from views import * urlpatterns = patterns('', url(r'^authorize_application.html$', authorize_application, name='authorize_application'), url(r'^deauthorize_application.html$', deauthorize_application) )
from django.conf.urls.defaults import * from views import * urlpatterns = patterns('', url(r'^authorize_application.html$', authorize_application, name='authorize_application'), url(r'^deauthorize_application.html$', deauthorize_application, name='deauthorize_application') )
Add a name to the URL configuration for 'deauthorize_application'
Add a name to the URL configuration for 'deauthorize_application'
Python
mit
jgorset/fandjango,jgorset/fandjango
from django.conf.urls.defaults import * from views import * urlpatterns = patterns('', url(r'^authorize_application.html$', authorize_application, name='authorize_application'), url(r'^deauthorize_application.html$', deauthorize_application) )Add a name to the URL configuration for 'deauthorize_application'
from django.conf.urls.defaults import * from views import * urlpatterns = patterns('', url(r'^authorize_application.html$', authorize_application, name='authorize_application'), url(r'^deauthorize_application.html$', deauthorize_application, name='deauthorize_application') )
<commit_before>from django.conf.urls.defaults import * from views import * urlpatterns = patterns('', url(r'^authorize_application.html$', authorize_application, name='authorize_application'), url(r'^deauthorize_application.html$', deauthorize_application) )<commit_msg>Add a name to the URL configuration for 'deauthorize_application'<commit_after>
from django.conf.urls.defaults import * from views import * urlpatterns = patterns('', url(r'^authorize_application.html$', authorize_application, name='authorize_application'), url(r'^deauthorize_application.html$', deauthorize_application, name='deauthorize_application') )
from django.conf.urls.defaults import * from views import * urlpatterns = patterns('', url(r'^authorize_application.html$', authorize_application, name='authorize_application'), url(r'^deauthorize_application.html$', deauthorize_application) )Add a name to the URL configuration for 'deauthorize_application'from django.conf.urls.defaults import * from views import * urlpatterns = patterns('', url(r'^authorize_application.html$', authorize_application, name='authorize_application'), url(r'^deauthorize_application.html$', deauthorize_application, name='deauthorize_application') )
<commit_before>from django.conf.urls.defaults import * from views import * urlpatterns = patterns('', url(r'^authorize_application.html$', authorize_application, name='authorize_application'), url(r'^deauthorize_application.html$', deauthorize_application) )<commit_msg>Add a name to the URL configuration for 'deauthorize_application'<commit_after>from django.conf.urls.defaults import * from views import * urlpatterns = patterns('', url(r'^authorize_application.html$', authorize_application, name='authorize_application'), url(r'^deauthorize_application.html$', deauthorize_application, name='deauthorize_application') )
a2ea31300e614c27c3f99e079293728bee9fbcf4
paws_config.py
paws_config.py
import tornado from tornado import gen import json import os BASE_PATH = '.' base_url = '/paws/public/' @gen.coroutine def uid_for_user(user): url = 'https://meta.wikimedia.org/w/api.php?' + \ 'action=query&meta=globaluserinfo' + \ '&format=json&formatversion=2' + \ '&guiuser={}'.format(user) client = tornado.httpclient.AsyncHTTPClient() resp = yield client.fetch(url) parsed = json.loads(resp.body.decode('utf-8')) if 'missing' in parsed['query']['globaluserinfo']: return None return parsed['query']['globaluserinfo']['id'] cached_uids = {} @gen.coroutine def path_for_url_segment(url): """ Takes a URL segment and returns a full filesystem path Example: input: YuviPanda/Something.ipynb output: 43/public/Something.ipynb """ splits = url.split('/') username = splits[0] path = '/'.join(splits[1:]) if username in cached_uids: uid = cached_uids[username] else: uid = yield uid_for_user(username) if uid is None: raise tornado.web.HTTPError(404) cached_uids[username] = uid return os.path.join(BASE_PATH, str(uid), 'public', path)
import tornado from tornado import gen import json import os BASE_PATH = '.' base_url = '/paws/public/' @gen.coroutine def uid_for_user(user): url = 'https://meta.wikimedia.org/w/api.php?' + \ 'action=query&meta=globaluserinfo' + \ '&format=json&formatversion=2' + \ '&guiuser={}'.format(user) client = tornado.httpclient.AsyncHTTPClient() resp = yield client.fetch(url) parsed = json.loads(resp.body.decode('utf-8')) if 'missing' in parsed['query']['globaluserinfo']: return None return parsed['query']['globaluserinfo']['id'] # These do not change, so let's cache them in memory cached_uids = {} @gen.coroutine def path_for_url_segment(url): """ Takes a URL segment and returns a full filesystem path Example: input: YuviPanda/Something.ipynb output: 43/public/Something.ipynb """ splits = url.split('/') username = splits[0] path = '/'.join(splits[1:]) if username in cached_uids: uid = cached_uids[username] else: uid = yield uid_for_user(username) if uid is None: raise tornado.web.HTTPError(404) cached_uids[username] = uid return os.path.join(BASE_PATH, str(uid), 'public', path)
Add clarifying(?) comment about uid caching
Add clarifying(?) comment about uid caching
Python
bsd-3-clause
yuvipanda/nbserve
import tornado from tornado import gen import json import os BASE_PATH = '.' base_url = '/paws/public/' @gen.coroutine def uid_for_user(user): url = 'https://meta.wikimedia.org/w/api.php?' + \ 'action=query&meta=globaluserinfo' + \ '&format=json&formatversion=2' + \ '&guiuser={}'.format(user) client = tornado.httpclient.AsyncHTTPClient() resp = yield client.fetch(url) parsed = json.loads(resp.body.decode('utf-8')) if 'missing' in parsed['query']['globaluserinfo']: return None return parsed['query']['globaluserinfo']['id'] cached_uids = {} @gen.coroutine def path_for_url_segment(url): """ Takes a URL segment and returns a full filesystem path Example: input: YuviPanda/Something.ipynb output: 43/public/Something.ipynb """ splits = url.split('/') username = splits[0] path = '/'.join(splits[1:]) if username in cached_uids: uid = cached_uids[username] else: uid = yield uid_for_user(username) if uid is None: raise tornado.web.HTTPError(404) cached_uids[username] = uid return os.path.join(BASE_PATH, str(uid), 'public', path) Add clarifying(?) comment about uid caching
import tornado from tornado import gen import json import os BASE_PATH = '.' base_url = '/paws/public/' @gen.coroutine def uid_for_user(user): url = 'https://meta.wikimedia.org/w/api.php?' + \ 'action=query&meta=globaluserinfo' + \ '&format=json&formatversion=2' + \ '&guiuser={}'.format(user) client = tornado.httpclient.AsyncHTTPClient() resp = yield client.fetch(url) parsed = json.loads(resp.body.decode('utf-8')) if 'missing' in parsed['query']['globaluserinfo']: return None return parsed['query']['globaluserinfo']['id'] # These do not change, so let's cache them in memory cached_uids = {} @gen.coroutine def path_for_url_segment(url): """ Takes a URL segment and returns a full filesystem path Example: input: YuviPanda/Something.ipynb output: 43/public/Something.ipynb """ splits = url.split('/') username = splits[0] path = '/'.join(splits[1:]) if username in cached_uids: uid = cached_uids[username] else: uid = yield uid_for_user(username) if uid is None: raise tornado.web.HTTPError(404) cached_uids[username] = uid return os.path.join(BASE_PATH, str(uid), 'public', path)
<commit_before>import tornado from tornado import gen import json import os BASE_PATH = '.' base_url = '/paws/public/' @gen.coroutine def uid_for_user(user): url = 'https://meta.wikimedia.org/w/api.php?' + \ 'action=query&meta=globaluserinfo' + \ '&format=json&formatversion=2' + \ '&guiuser={}'.format(user) client = tornado.httpclient.AsyncHTTPClient() resp = yield client.fetch(url) parsed = json.loads(resp.body.decode('utf-8')) if 'missing' in parsed['query']['globaluserinfo']: return None return parsed['query']['globaluserinfo']['id'] cached_uids = {} @gen.coroutine def path_for_url_segment(url): """ Takes a URL segment and returns a full filesystem path Example: input: YuviPanda/Something.ipynb output: 43/public/Something.ipynb """ splits = url.split('/') username = splits[0] path = '/'.join(splits[1:]) if username in cached_uids: uid = cached_uids[username] else: uid = yield uid_for_user(username) if uid is None: raise tornado.web.HTTPError(404) cached_uids[username] = uid return os.path.join(BASE_PATH, str(uid), 'public', path) <commit_msg>Add clarifying(?) comment about uid caching<commit_after>
import tornado from tornado import gen import json import os BASE_PATH = '.' base_url = '/paws/public/' @gen.coroutine def uid_for_user(user): url = 'https://meta.wikimedia.org/w/api.php?' + \ 'action=query&meta=globaluserinfo' + \ '&format=json&formatversion=2' + \ '&guiuser={}'.format(user) client = tornado.httpclient.AsyncHTTPClient() resp = yield client.fetch(url) parsed = json.loads(resp.body.decode('utf-8')) if 'missing' in parsed['query']['globaluserinfo']: return None return parsed['query']['globaluserinfo']['id'] # These do not change, so let's cache them in memory cached_uids = {} @gen.coroutine def path_for_url_segment(url): """ Takes a URL segment and returns a full filesystem path Example: input: YuviPanda/Something.ipynb output: 43/public/Something.ipynb """ splits = url.split('/') username = splits[0] path = '/'.join(splits[1:]) if username in cached_uids: uid = cached_uids[username] else: uid = yield uid_for_user(username) if uid is None: raise tornado.web.HTTPError(404) cached_uids[username] = uid return os.path.join(BASE_PATH, str(uid), 'public', path)
import tornado from tornado import gen import json import os BASE_PATH = '.' base_url = '/paws/public/' @gen.coroutine def uid_for_user(user): url = 'https://meta.wikimedia.org/w/api.php?' + \ 'action=query&meta=globaluserinfo' + \ '&format=json&formatversion=2' + \ '&guiuser={}'.format(user) client = tornado.httpclient.AsyncHTTPClient() resp = yield client.fetch(url) parsed = json.loads(resp.body.decode('utf-8')) if 'missing' in parsed['query']['globaluserinfo']: return None return parsed['query']['globaluserinfo']['id'] cached_uids = {} @gen.coroutine def path_for_url_segment(url): """ Takes a URL segment and returns a full filesystem path Example: input: YuviPanda/Something.ipynb output: 43/public/Something.ipynb """ splits = url.split('/') username = splits[0] path = '/'.join(splits[1:]) if username in cached_uids: uid = cached_uids[username] else: uid = yield uid_for_user(username) if uid is None: raise tornado.web.HTTPError(404) cached_uids[username] = uid return os.path.join(BASE_PATH, str(uid), 'public', path) Add clarifying(?) comment about uid cachingimport tornado from tornado import gen import json import os BASE_PATH = '.' base_url = '/paws/public/' @gen.coroutine def uid_for_user(user): url = 'https://meta.wikimedia.org/w/api.php?' + \ 'action=query&meta=globaluserinfo' + \ '&format=json&formatversion=2' + \ '&guiuser={}'.format(user) client = tornado.httpclient.AsyncHTTPClient() resp = yield client.fetch(url) parsed = json.loads(resp.body.decode('utf-8')) if 'missing' in parsed['query']['globaluserinfo']: return None return parsed['query']['globaluserinfo']['id'] # These do not change, so let's cache them in memory cached_uids = {} @gen.coroutine def path_for_url_segment(url): """ Takes a URL segment and returns a full filesystem path Example: input: YuviPanda/Something.ipynb output: 43/public/Something.ipynb """ splits = url.split('/') username = splits[0] path = '/'.join(splits[1:]) if username in cached_uids: uid = cached_uids[username] else: uid = yield uid_for_user(username) if uid is None: raise tornado.web.HTTPError(404) cached_uids[username] = uid return os.path.join(BASE_PATH, str(uid), 'public', path)
<commit_before>import tornado from tornado import gen import json import os BASE_PATH = '.' base_url = '/paws/public/' @gen.coroutine def uid_for_user(user): url = 'https://meta.wikimedia.org/w/api.php?' + \ 'action=query&meta=globaluserinfo' + \ '&format=json&formatversion=2' + \ '&guiuser={}'.format(user) client = tornado.httpclient.AsyncHTTPClient() resp = yield client.fetch(url) parsed = json.loads(resp.body.decode('utf-8')) if 'missing' in parsed['query']['globaluserinfo']: return None return parsed['query']['globaluserinfo']['id'] cached_uids = {} @gen.coroutine def path_for_url_segment(url): """ Takes a URL segment and returns a full filesystem path Example: input: YuviPanda/Something.ipynb output: 43/public/Something.ipynb """ splits = url.split('/') username = splits[0] path = '/'.join(splits[1:]) if username in cached_uids: uid = cached_uids[username] else: uid = yield uid_for_user(username) if uid is None: raise tornado.web.HTTPError(404) cached_uids[username] = uid return os.path.join(BASE_PATH, str(uid), 'public', path) <commit_msg>Add clarifying(?) comment about uid caching<commit_after>import tornado from tornado import gen import json import os BASE_PATH = '.' base_url = '/paws/public/' @gen.coroutine def uid_for_user(user): url = 'https://meta.wikimedia.org/w/api.php?' + \ 'action=query&meta=globaluserinfo' + \ '&format=json&formatversion=2' + \ '&guiuser={}'.format(user) client = tornado.httpclient.AsyncHTTPClient() resp = yield client.fetch(url) parsed = json.loads(resp.body.decode('utf-8')) if 'missing' in parsed['query']['globaluserinfo']: return None return parsed['query']['globaluserinfo']['id'] # These do not change, so let's cache them in memory cached_uids = {} @gen.coroutine def path_for_url_segment(url): """ Takes a URL segment and returns a full filesystem path Example: input: YuviPanda/Something.ipynb output: 43/public/Something.ipynb """ splits = url.split('/') username = splits[0] path = '/'.join(splits[1:]) if username in cached_uids: uid = cached_uids[username] else: uid = yield uid_for_user(username) if uid is None: raise tornado.web.HTTPError(404) cached_uids[username] = uid return os.path.join(BASE_PATH, str(uid), 'public', path)
20f67ed02cd31d88a897b6e7ac3c93c482ea1684
awslimits/data_setup.py
awslimits/data_setup.py
from .support import load_tickets, load_default_limits import settings def update_data(): print 'loading default limits...' load_default_limits() if settings.PREMIUM_ACCOUNT: print 'loading tickets...' load_tickets() print 'done' if __name__ == "__main__": update_data()
from .support import load_tickets, load_default_limits import settings def update_data(): print('loading default limits...') load_default_limits() if settings.PREMIUM_ACCOUNT: print('loading tickets...') load_tickets() print('done') if __name__ == "__main__": update_data()
Fix print statements for Python 3
Fix print statements for Python 3
Python
apache-2.0
spulec/awslimits,Yipit/awslimits,Yipit/awslimits,spulec/awslimits
from .support import load_tickets, load_default_limits import settings def update_data(): print 'loading default limits...' load_default_limits() if settings.PREMIUM_ACCOUNT: print 'loading tickets...' load_tickets() print 'done' if __name__ == "__main__": update_data() Fix print statements for Python 3
from .support import load_tickets, load_default_limits import settings def update_data(): print('loading default limits...') load_default_limits() if settings.PREMIUM_ACCOUNT: print('loading tickets...') load_tickets() print('done') if __name__ == "__main__": update_data()
<commit_before>from .support import load_tickets, load_default_limits import settings def update_data(): print 'loading default limits...' load_default_limits() if settings.PREMIUM_ACCOUNT: print 'loading tickets...' load_tickets() print 'done' if __name__ == "__main__": update_data() <commit_msg>Fix print statements for Python 3<commit_after>
from .support import load_tickets, load_default_limits import settings def update_data(): print('loading default limits...') load_default_limits() if settings.PREMIUM_ACCOUNT: print('loading tickets...') load_tickets() print('done') if __name__ == "__main__": update_data()
from .support import load_tickets, load_default_limits import settings def update_data(): print 'loading default limits...' load_default_limits() if settings.PREMIUM_ACCOUNT: print 'loading tickets...' load_tickets() print 'done' if __name__ == "__main__": update_data() Fix print statements for Python 3from .support import load_tickets, load_default_limits import settings def update_data(): print('loading default limits...') load_default_limits() if settings.PREMIUM_ACCOUNT: print('loading tickets...') load_tickets() print('done') if __name__ == "__main__": update_data()
<commit_before>from .support import load_tickets, load_default_limits import settings def update_data(): print 'loading default limits...' load_default_limits() if settings.PREMIUM_ACCOUNT: print 'loading tickets...' load_tickets() print 'done' if __name__ == "__main__": update_data() <commit_msg>Fix print statements for Python 3<commit_after>from .support import load_tickets, load_default_limits import settings def update_data(): print('loading default limits...') load_default_limits() if settings.PREMIUM_ACCOUNT: print('loading tickets...') load_tickets() print('done') if __name__ == "__main__": update_data()
bd20dbda918cdec93ab6d1fe5bba0ce064a60103
fairseq/scoring/wer.py
fairseq/scoring/wer.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import editdistance from fairseq.scoring import register_scoring @register_scoring("wer") class WerScorer(object): def __init__(self, *unused): self.reset() def reset(self): self.distance = 0 self.target_length = 0 def add_string(self, ref, pred): pred_items = ref.split() targ_items = pred.split() self.distance += editdistance.eval(pred_items, targ_items) self.target_length += len(targ_items) def result_string(self): return f"WER: {self.score()}" def score(self): return ( 100.0 * self.distance / self.target_length if self.target_length > 0 else 0 )
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import editdistance from fairseq.scoring import register_scoring @register_scoring("wer") class WerScorer(object): def __init__(self, *unused): self.reset() def reset(self): self.distance = 0 self.ref_length = 0 def add_string(self, ref, pred): ref_items = ref.split() pred_items = pred.split() self.distance += editdistance.eval(ref_items, pred_items) self.ref_length += len(ref_items) def result_string(self): return f"WER: {self.score()}" def score(self): return ( 100.0 * self.distance / self.ref_length if self.ref_length > 0 else 0 )
Fix typos in WER scorer
Fix typos in WER scorer Summary: Fix typos in WER scorer - The typos lead to using prediction length as the denominator in the formula, which is wrong. Reviewed By: alexeib Differential Revision: D23139261 fbshipit-source-id: d1bba0044365813603ce358388e880c1b3f9ec6b
Python
mit
pytorch/fairseq,pytorch/fairseq,pytorch/fairseq
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import editdistance from fairseq.scoring import register_scoring @register_scoring("wer") class WerScorer(object): def __init__(self, *unused): self.reset() def reset(self): self.distance = 0 self.target_length = 0 def add_string(self, ref, pred): pred_items = ref.split() targ_items = pred.split() self.distance += editdistance.eval(pred_items, targ_items) self.target_length += len(targ_items) def result_string(self): return f"WER: {self.score()}" def score(self): return ( 100.0 * self.distance / self.target_length if self.target_length > 0 else 0 ) Fix typos in WER scorer Summary: Fix typos in WER scorer - The typos lead to using prediction length as the denominator in the formula, which is wrong. Reviewed By: alexeib Differential Revision: D23139261 fbshipit-source-id: d1bba0044365813603ce358388e880c1b3f9ec6b
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import editdistance from fairseq.scoring import register_scoring @register_scoring("wer") class WerScorer(object): def __init__(self, *unused): self.reset() def reset(self): self.distance = 0 self.ref_length = 0 def add_string(self, ref, pred): ref_items = ref.split() pred_items = pred.split() self.distance += editdistance.eval(ref_items, pred_items) self.ref_length += len(ref_items) def result_string(self): return f"WER: {self.score()}" def score(self): return ( 100.0 * self.distance / self.ref_length if self.ref_length > 0 else 0 )
<commit_before># Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import editdistance from fairseq.scoring import register_scoring @register_scoring("wer") class WerScorer(object): def __init__(self, *unused): self.reset() def reset(self): self.distance = 0 self.target_length = 0 def add_string(self, ref, pred): pred_items = ref.split() targ_items = pred.split() self.distance += editdistance.eval(pred_items, targ_items) self.target_length += len(targ_items) def result_string(self): return f"WER: {self.score()}" def score(self): return ( 100.0 * self.distance / self.target_length if self.target_length > 0 else 0 ) <commit_msg>Fix typos in WER scorer Summary: Fix typos in WER scorer - The typos lead to using prediction length as the denominator in the formula, which is wrong. Reviewed By: alexeib Differential Revision: D23139261 fbshipit-source-id: d1bba0044365813603ce358388e880c1b3f9ec6b<commit_after>
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import editdistance from fairseq.scoring import register_scoring @register_scoring("wer") class WerScorer(object): def __init__(self, *unused): self.reset() def reset(self): self.distance = 0 self.ref_length = 0 def add_string(self, ref, pred): ref_items = ref.split() pred_items = pred.split() self.distance += editdistance.eval(ref_items, pred_items) self.ref_length += len(ref_items) def result_string(self): return f"WER: {self.score()}" def score(self): return ( 100.0 * self.distance / self.ref_length if self.ref_length > 0 else 0 )
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import editdistance from fairseq.scoring import register_scoring @register_scoring("wer") class WerScorer(object): def __init__(self, *unused): self.reset() def reset(self): self.distance = 0 self.target_length = 0 def add_string(self, ref, pred): pred_items = ref.split() targ_items = pred.split() self.distance += editdistance.eval(pred_items, targ_items) self.target_length += len(targ_items) def result_string(self): return f"WER: {self.score()}" def score(self): return ( 100.0 * self.distance / self.target_length if self.target_length > 0 else 0 ) Fix typos in WER scorer Summary: Fix typos in WER scorer - The typos lead to using prediction length as the denominator in the formula, which is wrong. Reviewed By: alexeib Differential Revision: D23139261 fbshipit-source-id: d1bba0044365813603ce358388e880c1b3f9ec6b# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import editdistance from fairseq.scoring import register_scoring @register_scoring("wer") class WerScorer(object): def __init__(self, *unused): self.reset() def reset(self): self.distance = 0 self.ref_length = 0 def add_string(self, ref, pred): ref_items = ref.split() pred_items = pred.split() self.distance += editdistance.eval(ref_items, pred_items) self.ref_length += len(ref_items) def result_string(self): return f"WER: {self.score()}" def score(self): return ( 100.0 * self.distance / self.ref_length if self.ref_length > 0 else 0 )
<commit_before># Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import editdistance from fairseq.scoring import register_scoring @register_scoring("wer") class WerScorer(object): def __init__(self, *unused): self.reset() def reset(self): self.distance = 0 self.target_length = 0 def add_string(self, ref, pred): pred_items = ref.split() targ_items = pred.split() self.distance += editdistance.eval(pred_items, targ_items) self.target_length += len(targ_items) def result_string(self): return f"WER: {self.score()}" def score(self): return ( 100.0 * self.distance / self.target_length if self.target_length > 0 else 0 ) <commit_msg>Fix typos in WER scorer Summary: Fix typos in WER scorer - The typos lead to using prediction length as the denominator in the formula, which is wrong. Reviewed By: alexeib Differential Revision: D23139261 fbshipit-source-id: d1bba0044365813603ce358388e880c1b3f9ec6b<commit_after># Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import editdistance from fairseq.scoring import register_scoring @register_scoring("wer") class WerScorer(object): def __init__(self, *unused): self.reset() def reset(self): self.distance = 0 self.ref_length = 0 def add_string(self, ref, pred): ref_items = ref.split() pred_items = pred.split() self.distance += editdistance.eval(ref_items, pred_items) self.ref_length += len(ref_items) def result_string(self): return f"WER: {self.score()}" def score(self): return ( 100.0 * self.distance / self.ref_length if self.ref_length > 0 else 0 )
6a219ca1451be9a68b567288fbd014624cc2417b
murano/tests/unit/api/middleware/test_version_negotiation.py
murano/tests/unit/api/middleware/test_version_negotiation.py
# Copyright 2016 AT&T Corp # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import webob from murano.api import versions from murano.api.middleware import version_negotiation from murano.tests.unit import base class MiddlewareVersionNegotiationTest(base.MuranoTestCase): def test_middleware_version_negotiation_default(self): middleware_vn = version_negotiation.VersionNegotiationFilter(None) request = webob.Request.blank('/environments') result = middleware_vn.process_request(request) self.assertTrue(isinstance(result, versions.Controller))
# Copyright 2016 AT&T Corp # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import webob from murano.api import versions from murano.api.middleware import version_negotiation from murano.tests.unit import base class MiddlewareVersionNegotiationTest(base.MuranoTestCase): def test_middleware_version_negotiation_default(self): middleware_vn = version_negotiation.VersionNegotiationFilter(None) request = webob.Request.blank('/environments') result = middleware_vn.process_request(request) self.assertIsInstance(result, versions.Controller)
Change assertTrue(isinstance()) by optimal assert
Change assertTrue(isinstance()) by optimal assert Some of tests use different method of assertTrue(isinstance(A, B)) or assertEqual(type(A), B). The correct way is to use assertIsInstance(A, B) provided by testtools Change-Id: I3231db85708b36d2092c14ad45b42f42a0151027
Python
apache-2.0
openstack/murano,openstack/murano
# Copyright 2016 AT&T Corp # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import webob from murano.api import versions from murano.api.middleware import version_negotiation from murano.tests.unit import base class MiddlewareVersionNegotiationTest(base.MuranoTestCase): def test_middleware_version_negotiation_default(self): middleware_vn = version_negotiation.VersionNegotiationFilter(None) request = webob.Request.blank('/environments') result = middleware_vn.process_request(request) self.assertTrue(isinstance(result, versions.Controller)) Change assertTrue(isinstance()) by optimal assert Some of tests use different method of assertTrue(isinstance(A, B)) or assertEqual(type(A), B). The correct way is to use assertIsInstance(A, B) provided by testtools Change-Id: I3231db85708b36d2092c14ad45b42f42a0151027
# Copyright 2016 AT&T Corp # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import webob from murano.api import versions from murano.api.middleware import version_negotiation from murano.tests.unit import base class MiddlewareVersionNegotiationTest(base.MuranoTestCase): def test_middleware_version_negotiation_default(self): middleware_vn = version_negotiation.VersionNegotiationFilter(None) request = webob.Request.blank('/environments') result = middleware_vn.process_request(request) self.assertIsInstance(result, versions.Controller)
<commit_before># Copyright 2016 AT&T Corp # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import webob from murano.api import versions from murano.api.middleware import version_negotiation from murano.tests.unit import base class MiddlewareVersionNegotiationTest(base.MuranoTestCase): def test_middleware_version_negotiation_default(self): middleware_vn = version_negotiation.VersionNegotiationFilter(None) request = webob.Request.blank('/environments') result = middleware_vn.process_request(request) self.assertTrue(isinstance(result, versions.Controller)) <commit_msg>Change assertTrue(isinstance()) by optimal assert Some of tests use different method of assertTrue(isinstance(A, B)) or assertEqual(type(A), B). The correct way is to use assertIsInstance(A, B) provided by testtools Change-Id: I3231db85708b36d2092c14ad45b42f42a0151027<commit_after>
# Copyright 2016 AT&T Corp # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import webob from murano.api import versions from murano.api.middleware import version_negotiation from murano.tests.unit import base class MiddlewareVersionNegotiationTest(base.MuranoTestCase): def test_middleware_version_negotiation_default(self): middleware_vn = version_negotiation.VersionNegotiationFilter(None) request = webob.Request.blank('/environments') result = middleware_vn.process_request(request) self.assertIsInstance(result, versions.Controller)
# Copyright 2016 AT&T Corp # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import webob from murano.api import versions from murano.api.middleware import version_negotiation from murano.tests.unit import base class MiddlewareVersionNegotiationTest(base.MuranoTestCase): def test_middleware_version_negotiation_default(self): middleware_vn = version_negotiation.VersionNegotiationFilter(None) request = webob.Request.blank('/environments') result = middleware_vn.process_request(request) self.assertTrue(isinstance(result, versions.Controller)) Change assertTrue(isinstance()) by optimal assert Some of tests use different method of assertTrue(isinstance(A, B)) or assertEqual(type(A), B). The correct way is to use assertIsInstance(A, B) provided by testtools Change-Id: I3231db85708b36d2092c14ad45b42f42a0151027# Copyright 2016 AT&T Corp # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import webob from murano.api import versions from murano.api.middleware import version_negotiation from murano.tests.unit import base class MiddlewareVersionNegotiationTest(base.MuranoTestCase): def test_middleware_version_negotiation_default(self): middleware_vn = version_negotiation.VersionNegotiationFilter(None) request = webob.Request.blank('/environments') result = middleware_vn.process_request(request) self.assertIsInstance(result, versions.Controller)
<commit_before># Copyright 2016 AT&T Corp # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import webob from murano.api import versions from murano.api.middleware import version_negotiation from murano.tests.unit import base class MiddlewareVersionNegotiationTest(base.MuranoTestCase): def test_middleware_version_negotiation_default(self): middleware_vn = version_negotiation.VersionNegotiationFilter(None) request = webob.Request.blank('/environments') result = middleware_vn.process_request(request) self.assertTrue(isinstance(result, versions.Controller)) <commit_msg>Change assertTrue(isinstance()) by optimal assert Some of tests use different method of assertTrue(isinstance(A, B)) or assertEqual(type(A), B). The correct way is to use assertIsInstance(A, B) provided by testtools Change-Id: I3231db85708b36d2092c14ad45b42f42a0151027<commit_after># Copyright 2016 AT&T Corp # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import webob from murano.api import versions from murano.api.middleware import version_negotiation from murano.tests.unit import base class MiddlewareVersionNegotiationTest(base.MuranoTestCase): def test_middleware_version_negotiation_default(self): middleware_vn = version_negotiation.VersionNegotiationFilter(None) request = webob.Request.blank('/environments') result = middleware_vn.process_request(request) self.assertIsInstance(result, versions.Controller)
7a7b42495fda7e7fabbf982e1194ea2fff6fdf15
code/DataSet.py
code/DataSet.py
# Neurotopics/code/DataSet.py import neurosynth.analysis.reduce as nsar class DataSet: """ A DataSet takes a NeuroSynth dataset and extracts the DOI's, and word subset of interest. It uses reduce.average_within_regions to get the average activation in the regions of interest (ROIs) of the img. """ def __init__(self, neurosynth_dataset, img, word_counts_file): self.word_subset = neurosynth_dataset.get_feature_names() self.dois = list(neurosynth_dataset.feature_table.ids) self.average_activation = nsar.average_within_regions(neurosynth_dataset, img) self.get_word_subset_counts(self.word_subset, word_counts_file) def get_word_subset_counts(self, word_subset, word_counts_file): pass
# Neurotopics/code/DataSet.py import neurosynth.analysis.reduce as nsar class DataSet: """ A DataSet takes a NeuroSynth dataset and extracts the DOI's, and word subset of interest. It uses reduce.average_within_regions to get the average activation in the regions of interest (ROIs) of the img. """ def __init__(self, neurosynth_dataset, image_file_name, word_counts_file): self.word_subset = neurosynth_dataset.get_feature_names() self.dois = list(neurosynth_dataset.feature_table.ids) self.average_activation = nsar.average_within_regions(neurosynth_dataset, image_file_name) self.get_word_subset_counts(self.word_subset, word_counts_file) def get_word_subset_counts(self, word_subset, word_counts_file): pass
Make more explicit variable name.
Make more explicit variable name.
Python
agpl-3.0
ambimorph/neurotopics-obs
# Neurotopics/code/DataSet.py import neurosynth.analysis.reduce as nsar class DataSet: """ A DataSet takes a NeuroSynth dataset and extracts the DOI's, and word subset of interest. It uses reduce.average_within_regions to get the average activation in the regions of interest (ROIs) of the img. """ def __init__(self, neurosynth_dataset, img, word_counts_file): self.word_subset = neurosynth_dataset.get_feature_names() self.dois = list(neurosynth_dataset.feature_table.ids) self.average_activation = nsar.average_within_regions(neurosynth_dataset, img) self.get_word_subset_counts(self.word_subset, word_counts_file) def get_word_subset_counts(self, word_subset, word_counts_file): pass Make more explicit variable name.
# Neurotopics/code/DataSet.py import neurosynth.analysis.reduce as nsar class DataSet: """ A DataSet takes a NeuroSynth dataset and extracts the DOI's, and word subset of interest. It uses reduce.average_within_regions to get the average activation in the regions of interest (ROIs) of the img. """ def __init__(self, neurosynth_dataset, image_file_name, word_counts_file): self.word_subset = neurosynth_dataset.get_feature_names() self.dois = list(neurosynth_dataset.feature_table.ids) self.average_activation = nsar.average_within_regions(neurosynth_dataset, image_file_name) self.get_word_subset_counts(self.word_subset, word_counts_file) def get_word_subset_counts(self, word_subset, word_counts_file): pass
<commit_before># Neurotopics/code/DataSet.py import neurosynth.analysis.reduce as nsar class DataSet: """ A DataSet takes a NeuroSynth dataset and extracts the DOI's, and word subset of interest. It uses reduce.average_within_regions to get the average activation in the regions of interest (ROIs) of the img. """ def __init__(self, neurosynth_dataset, img, word_counts_file): self.word_subset = neurosynth_dataset.get_feature_names() self.dois = list(neurosynth_dataset.feature_table.ids) self.average_activation = nsar.average_within_regions(neurosynth_dataset, img) self.get_word_subset_counts(self.word_subset, word_counts_file) def get_word_subset_counts(self, word_subset, word_counts_file): pass <commit_msg>Make more explicit variable name.<commit_after>
# Neurotopics/code/DataSet.py import neurosynth.analysis.reduce as nsar class DataSet: """ A DataSet takes a NeuroSynth dataset and extracts the DOI's, and word subset of interest. It uses reduce.average_within_regions to get the average activation in the regions of interest (ROIs) of the img. """ def __init__(self, neurosynth_dataset, image_file_name, word_counts_file): self.word_subset = neurosynth_dataset.get_feature_names() self.dois = list(neurosynth_dataset.feature_table.ids) self.average_activation = nsar.average_within_regions(neurosynth_dataset, image_file_name) self.get_word_subset_counts(self.word_subset, word_counts_file) def get_word_subset_counts(self, word_subset, word_counts_file): pass
# Neurotopics/code/DataSet.py import neurosynth.analysis.reduce as nsar class DataSet: """ A DataSet takes a NeuroSynth dataset and extracts the DOI's, and word subset of interest. It uses reduce.average_within_regions to get the average activation in the regions of interest (ROIs) of the img. """ def __init__(self, neurosynth_dataset, img, word_counts_file): self.word_subset = neurosynth_dataset.get_feature_names() self.dois = list(neurosynth_dataset.feature_table.ids) self.average_activation = nsar.average_within_regions(neurosynth_dataset, img) self.get_word_subset_counts(self.word_subset, word_counts_file) def get_word_subset_counts(self, word_subset, word_counts_file): pass Make more explicit variable name.# Neurotopics/code/DataSet.py import neurosynth.analysis.reduce as nsar class DataSet: """ A DataSet takes a NeuroSynth dataset and extracts the DOI's, and word subset of interest. It uses reduce.average_within_regions to get the average activation in the regions of interest (ROIs) of the img. """ def __init__(self, neurosynth_dataset, image_file_name, word_counts_file): self.word_subset = neurosynth_dataset.get_feature_names() self.dois = list(neurosynth_dataset.feature_table.ids) self.average_activation = nsar.average_within_regions(neurosynth_dataset, image_file_name) self.get_word_subset_counts(self.word_subset, word_counts_file) def get_word_subset_counts(self, word_subset, word_counts_file): pass
<commit_before># Neurotopics/code/DataSet.py import neurosynth.analysis.reduce as nsar class DataSet: """ A DataSet takes a NeuroSynth dataset and extracts the DOI's, and word subset of interest. It uses reduce.average_within_regions to get the average activation in the regions of interest (ROIs) of the img. """ def __init__(self, neurosynth_dataset, img, word_counts_file): self.word_subset = neurosynth_dataset.get_feature_names() self.dois = list(neurosynth_dataset.feature_table.ids) self.average_activation = nsar.average_within_regions(neurosynth_dataset, img) self.get_word_subset_counts(self.word_subset, word_counts_file) def get_word_subset_counts(self, word_subset, word_counts_file): pass <commit_msg>Make more explicit variable name.<commit_after># Neurotopics/code/DataSet.py import neurosynth.analysis.reduce as nsar class DataSet: """ A DataSet takes a NeuroSynth dataset and extracts the DOI's, and word subset of interest. It uses reduce.average_within_regions to get the average activation in the regions of interest (ROIs) of the img. """ def __init__(self, neurosynth_dataset, image_file_name, word_counts_file): self.word_subset = neurosynth_dataset.get_feature_names() self.dois = list(neurosynth_dataset.feature_table.ids) self.average_activation = nsar.average_within_regions(neurosynth_dataset, image_file_name) self.get_word_subset_counts(self.word_subset, word_counts_file) def get_word_subset_counts(self, word_subset, word_counts_file): pass
c47468128ab831133a12f942d32dd73b4198458e
scent.py
scent.py
# -*- coding: utf-8 -*- import os import time import subprocess from sniffer.api import select_runnable, file_validator, runnable try: from pync import Notifier except ImportError: notify = None else: notify = Notifier.notify watch_paths = ['demo/', 'tests/'] @select_runnable('python_tests') @file_validator def py_files(filename): return all((filename.endswith('.py'), not os.path.basename(filename).startswith('.'))) @runnable def python_tests(*args): group = int(time.time()) # unique per run for count, (command, title) in enumerate(( (('make', 'test-unit'), "Unit Tests"), (('make', 'test-int'), "Integration Tests"), (('make', 'test-all'), "Combined Tests"), ), start=1): failure = subprocess.call(command) if failure: if notify: mark = "❌" * count notify(mark + " [FAIL] " + mark, title=title, group=group) return False else: if notify: mark = "✅" * count notify(mark + " [PASS] " + mark, title=title, group=group) return True
# -*- coding: utf-8 -*- import os import time import subprocess from sniffer.api import select_runnable, file_validator, runnable try: from pync import Notifier except ImportError: notify = None else: notify = Notifier.notify watch_paths = ['demo/', 'tests/'] @select_runnable('python_tests') @file_validator def py_files(filename): return all((filename.endswith('.py'), not os.path.basename(filename).startswith('.'))) @runnable def python_tests(*args): group = int(time.time()) # unique per run for count, (command, title) in enumerate(( (('make', 'test-unit'), "Unit Tests"), (('make', 'test-int'), "Integration Tests"), (('make', 'test-all'), "Combined Tests"), (('make', 'check'), "Static Analysis"), (('make', 'doc'), None), ), start=1): failure = subprocess.call(command) if failure: if notify and title: mark = "❌" * count notify(mark + " [FAIL] " + mark, title=title, group=group) return False else: if notify and title: mark = "✅" * count notify(mark + " [PASS] " + mark, title=title, group=group) return True
Deploy Travis CI build 478 to GitHub
Deploy Travis CI build 478 to GitHub
Python
mit
jacebrowning/template-python-demo
# -*- coding: utf-8 -*- import os import time import subprocess from sniffer.api import select_runnable, file_validator, runnable try: from pync import Notifier except ImportError: notify = None else: notify = Notifier.notify watch_paths = ['demo/', 'tests/'] @select_runnable('python_tests') @file_validator def py_files(filename): return all((filename.endswith('.py'), not os.path.basename(filename).startswith('.'))) @runnable def python_tests(*args): group = int(time.time()) # unique per run for count, (command, title) in enumerate(( (('make', 'test-unit'), "Unit Tests"), (('make', 'test-int'), "Integration Tests"), (('make', 'test-all'), "Combined Tests"), ), start=1): failure = subprocess.call(command) if failure: if notify: mark = "❌" * count notify(mark + " [FAIL] " + mark, title=title, group=group) return False else: if notify: mark = "✅" * count notify(mark + " [PASS] " + mark, title=title, group=group) return True Deploy Travis CI build 478 to GitHub
# -*- coding: utf-8 -*- import os import time import subprocess from sniffer.api import select_runnable, file_validator, runnable try: from pync import Notifier except ImportError: notify = None else: notify = Notifier.notify watch_paths = ['demo/', 'tests/'] @select_runnable('python_tests') @file_validator def py_files(filename): return all((filename.endswith('.py'), not os.path.basename(filename).startswith('.'))) @runnable def python_tests(*args): group = int(time.time()) # unique per run for count, (command, title) in enumerate(( (('make', 'test-unit'), "Unit Tests"), (('make', 'test-int'), "Integration Tests"), (('make', 'test-all'), "Combined Tests"), (('make', 'check'), "Static Analysis"), (('make', 'doc'), None), ), start=1): failure = subprocess.call(command) if failure: if notify and title: mark = "❌" * count notify(mark + " [FAIL] " + mark, title=title, group=group) return False else: if notify and title: mark = "✅" * count notify(mark + " [PASS] " + mark, title=title, group=group) return True
<commit_before># -*- coding: utf-8 -*- import os import time import subprocess from sniffer.api import select_runnable, file_validator, runnable try: from pync import Notifier except ImportError: notify = None else: notify = Notifier.notify watch_paths = ['demo/', 'tests/'] @select_runnable('python_tests') @file_validator def py_files(filename): return all((filename.endswith('.py'), not os.path.basename(filename).startswith('.'))) @runnable def python_tests(*args): group = int(time.time()) # unique per run for count, (command, title) in enumerate(( (('make', 'test-unit'), "Unit Tests"), (('make', 'test-int'), "Integration Tests"), (('make', 'test-all'), "Combined Tests"), ), start=1): failure = subprocess.call(command) if failure: if notify: mark = "❌" * count notify(mark + " [FAIL] " + mark, title=title, group=group) return False else: if notify: mark = "✅" * count notify(mark + " [PASS] " + mark, title=title, group=group) return True <commit_msg>Deploy Travis CI build 478 to GitHub<commit_after>
# -*- coding: utf-8 -*- import os import time import subprocess from sniffer.api import select_runnable, file_validator, runnable try: from pync import Notifier except ImportError: notify = None else: notify = Notifier.notify watch_paths = ['demo/', 'tests/'] @select_runnable('python_tests') @file_validator def py_files(filename): return all((filename.endswith('.py'), not os.path.basename(filename).startswith('.'))) @runnable def python_tests(*args): group = int(time.time()) # unique per run for count, (command, title) in enumerate(( (('make', 'test-unit'), "Unit Tests"), (('make', 'test-int'), "Integration Tests"), (('make', 'test-all'), "Combined Tests"), (('make', 'check'), "Static Analysis"), (('make', 'doc'), None), ), start=1): failure = subprocess.call(command) if failure: if notify and title: mark = "❌" * count notify(mark + " [FAIL] " + mark, title=title, group=group) return False else: if notify and title: mark = "✅" * count notify(mark + " [PASS] " + mark, title=title, group=group) return True
# -*- coding: utf-8 -*- import os import time import subprocess from sniffer.api import select_runnable, file_validator, runnable try: from pync import Notifier except ImportError: notify = None else: notify = Notifier.notify watch_paths = ['demo/', 'tests/'] @select_runnable('python_tests') @file_validator def py_files(filename): return all((filename.endswith('.py'), not os.path.basename(filename).startswith('.'))) @runnable def python_tests(*args): group = int(time.time()) # unique per run for count, (command, title) in enumerate(( (('make', 'test-unit'), "Unit Tests"), (('make', 'test-int'), "Integration Tests"), (('make', 'test-all'), "Combined Tests"), ), start=1): failure = subprocess.call(command) if failure: if notify: mark = "❌" * count notify(mark + " [FAIL] " + mark, title=title, group=group) return False else: if notify: mark = "✅" * count notify(mark + " [PASS] " + mark, title=title, group=group) return True Deploy Travis CI build 478 to GitHub# -*- coding: utf-8 -*- import os import time import subprocess from sniffer.api import select_runnable, file_validator, runnable try: from pync import Notifier except ImportError: notify = None else: notify = Notifier.notify watch_paths = ['demo/', 'tests/'] @select_runnable('python_tests') @file_validator def py_files(filename): return all((filename.endswith('.py'), not os.path.basename(filename).startswith('.'))) @runnable def python_tests(*args): group = int(time.time()) # unique per run for count, (command, title) in enumerate(( (('make', 'test-unit'), "Unit Tests"), (('make', 'test-int'), "Integration Tests"), (('make', 'test-all'), "Combined Tests"), (('make', 'check'), "Static Analysis"), (('make', 'doc'), None), ), start=1): failure = subprocess.call(command) if failure: if notify and title: mark = "❌" * count notify(mark + " [FAIL] " + mark, title=title, group=group) return False else: if notify and title: mark = "✅" * count notify(mark + " [PASS] " + mark, title=title, group=group) return True
<commit_before># -*- coding: utf-8 -*- import os import time import subprocess from sniffer.api import select_runnable, file_validator, runnable try: from pync import Notifier except ImportError: notify = None else: notify = Notifier.notify watch_paths = ['demo/', 'tests/'] @select_runnable('python_tests') @file_validator def py_files(filename): return all((filename.endswith('.py'), not os.path.basename(filename).startswith('.'))) @runnable def python_tests(*args): group = int(time.time()) # unique per run for count, (command, title) in enumerate(( (('make', 'test-unit'), "Unit Tests"), (('make', 'test-int'), "Integration Tests"), (('make', 'test-all'), "Combined Tests"), ), start=1): failure = subprocess.call(command) if failure: if notify: mark = "❌" * count notify(mark + " [FAIL] " + mark, title=title, group=group) return False else: if notify: mark = "✅" * count notify(mark + " [PASS] " + mark, title=title, group=group) return True <commit_msg>Deploy Travis CI build 478 to GitHub<commit_after># -*- coding: utf-8 -*- import os import time import subprocess from sniffer.api import select_runnable, file_validator, runnable try: from pync import Notifier except ImportError: notify = None else: notify = Notifier.notify watch_paths = ['demo/', 'tests/'] @select_runnable('python_tests') @file_validator def py_files(filename): return all((filename.endswith('.py'), not os.path.basename(filename).startswith('.'))) @runnable def python_tests(*args): group = int(time.time()) # unique per run for count, (command, title) in enumerate(( (('make', 'test-unit'), "Unit Tests"), (('make', 'test-int'), "Integration Tests"), (('make', 'test-all'), "Combined Tests"), (('make', 'check'), "Static Analysis"), (('make', 'doc'), None), ), start=1): failure = subprocess.call(command) if failure: if notify and title: mark = "❌" * count notify(mark + " [FAIL] " + mark, title=title, group=group) return False else: if notify and title: mark = "✅" * count notify(mark + " [PASS] " + mark, title=title, group=group) return True
cc80f90a4f003c0967c31d5177971061350ee683
pycall/call.py
pycall/call.py
"""A simple wrapper for Asterisk calls.""" class Call(object): """Stores and manipulates Asterisk calls.""" def __init__(self, channel, callerid=None, account=None, wait_time=None, max_retries=None): """Create a new `Call` object. :param str channel: The Asterisk channel to call. Should be in standard Asterisk format. :param str callerid: CallerID to use. :param str account: Account code to associate with this call. :param int wait_time: Amount of time to wait (in seconds) between retry attempts. :param int max_retries: Maximum amount of retry attempts. """ self.channel = channel self.callerid = callerid self.account = account self.wait_time = int(wait_time) self.max_retries = int(max_retries)
"""A simple wrapper for Asterisk calls.""" class Call(object): """Stores and manipulates Asterisk calls.""" def __init__(self, channel, callerid=None, account=None, wait_time=None, max_retries=None): """Create a new `Call` object. :param str channel: The Asterisk channel to call. Should be in standard Asterisk format. :param str callerid: CallerID to use. :param str account: Account code to associate with this call. :param int wait_time: Amount of time to wait (in seconds) between retry attempts. :param int max_retries: Maximum amount of retry attempts. """ self.channel = channel self.callerid = callerid self.account = account self.wait_time = wait_time self.max_retries = max_retries
Revert "Forcing type coersion for int params."
Revert "Forcing type coersion for int params." This is a pointless bit of code. Since we lazy-evaluate them anyhow, it's a duplicate effort. This reverts commit 1ca6b96d492f8f33ac3b3a520937378effb66744.
Python
unlicense
rdegges/pycall
"""A simple wrapper for Asterisk calls.""" class Call(object): """Stores and manipulates Asterisk calls.""" def __init__(self, channel, callerid=None, account=None, wait_time=None, max_retries=None): """Create a new `Call` object. :param str channel: The Asterisk channel to call. Should be in standard Asterisk format. :param str callerid: CallerID to use. :param str account: Account code to associate with this call. :param int wait_time: Amount of time to wait (in seconds) between retry attempts. :param int max_retries: Maximum amount of retry attempts. """ self.channel = channel self.callerid = callerid self.account = account self.wait_time = int(wait_time) self.max_retries = int(max_retries) Revert "Forcing type coersion for int params." This is a pointless bit of code. Since we lazy-evaluate them anyhow, it's a duplicate effort. This reverts commit 1ca6b96d492f8f33ac3b3a520937378effb66744.
"""A simple wrapper for Asterisk calls.""" class Call(object): """Stores and manipulates Asterisk calls.""" def __init__(self, channel, callerid=None, account=None, wait_time=None, max_retries=None): """Create a new `Call` object. :param str channel: The Asterisk channel to call. Should be in standard Asterisk format. :param str callerid: CallerID to use. :param str account: Account code to associate with this call. :param int wait_time: Amount of time to wait (in seconds) between retry attempts. :param int max_retries: Maximum amount of retry attempts. """ self.channel = channel self.callerid = callerid self.account = account self.wait_time = wait_time self.max_retries = max_retries
<commit_before>"""A simple wrapper for Asterisk calls.""" class Call(object): """Stores and manipulates Asterisk calls.""" def __init__(self, channel, callerid=None, account=None, wait_time=None, max_retries=None): """Create a new `Call` object. :param str channel: The Asterisk channel to call. Should be in standard Asterisk format. :param str callerid: CallerID to use. :param str account: Account code to associate with this call. :param int wait_time: Amount of time to wait (in seconds) between retry attempts. :param int max_retries: Maximum amount of retry attempts. """ self.channel = channel self.callerid = callerid self.account = account self.wait_time = int(wait_time) self.max_retries = int(max_retries) <commit_msg>Revert "Forcing type coersion for int params." This is a pointless bit of code. Since we lazy-evaluate them anyhow, it's a duplicate effort. This reverts commit 1ca6b96d492f8f33ac3b3a520937378effb66744.<commit_after>
"""A simple wrapper for Asterisk calls.""" class Call(object): """Stores and manipulates Asterisk calls.""" def __init__(self, channel, callerid=None, account=None, wait_time=None, max_retries=None): """Create a new `Call` object. :param str channel: The Asterisk channel to call. Should be in standard Asterisk format. :param str callerid: CallerID to use. :param str account: Account code to associate with this call. :param int wait_time: Amount of time to wait (in seconds) between retry attempts. :param int max_retries: Maximum amount of retry attempts. """ self.channel = channel self.callerid = callerid self.account = account self.wait_time = wait_time self.max_retries = max_retries
"""A simple wrapper for Asterisk calls.""" class Call(object): """Stores and manipulates Asterisk calls.""" def __init__(self, channel, callerid=None, account=None, wait_time=None, max_retries=None): """Create a new `Call` object. :param str channel: The Asterisk channel to call. Should be in standard Asterisk format. :param str callerid: CallerID to use. :param str account: Account code to associate with this call. :param int wait_time: Amount of time to wait (in seconds) between retry attempts. :param int max_retries: Maximum amount of retry attempts. """ self.channel = channel self.callerid = callerid self.account = account self.wait_time = int(wait_time) self.max_retries = int(max_retries) Revert "Forcing type coersion for int params." This is a pointless bit of code. Since we lazy-evaluate them anyhow, it's a duplicate effort. This reverts commit 1ca6b96d492f8f33ac3b3a520937378effb66744."""A simple wrapper for Asterisk calls.""" class Call(object): """Stores and manipulates Asterisk calls.""" def __init__(self, channel, callerid=None, account=None, wait_time=None, max_retries=None): """Create a new `Call` object. :param str channel: The Asterisk channel to call. Should be in standard Asterisk format. :param str callerid: CallerID to use. :param str account: Account code to associate with this call. :param int wait_time: Amount of time to wait (in seconds) between retry attempts. :param int max_retries: Maximum amount of retry attempts. """ self.channel = channel self.callerid = callerid self.account = account self.wait_time = wait_time self.max_retries = max_retries
<commit_before>"""A simple wrapper for Asterisk calls.""" class Call(object): """Stores and manipulates Asterisk calls.""" def __init__(self, channel, callerid=None, account=None, wait_time=None, max_retries=None): """Create a new `Call` object. :param str channel: The Asterisk channel to call. Should be in standard Asterisk format. :param str callerid: CallerID to use. :param str account: Account code to associate with this call. :param int wait_time: Amount of time to wait (in seconds) between retry attempts. :param int max_retries: Maximum amount of retry attempts. """ self.channel = channel self.callerid = callerid self.account = account self.wait_time = int(wait_time) self.max_retries = int(max_retries) <commit_msg>Revert "Forcing type coersion for int params." This is a pointless bit of code. Since we lazy-evaluate them anyhow, it's a duplicate effort. This reverts commit 1ca6b96d492f8f33ac3b3a520937378effb66744.<commit_after>"""A simple wrapper for Asterisk calls.""" class Call(object): """Stores and manipulates Asterisk calls.""" def __init__(self, channel, callerid=None, account=None, wait_time=None, max_retries=None): """Create a new `Call` object. :param str channel: The Asterisk channel to call. Should be in standard Asterisk format. :param str callerid: CallerID to use. :param str account: Account code to associate with this call. :param int wait_time: Amount of time to wait (in seconds) between retry attempts. :param int max_retries: Maximum amount of retry attempts. """ self.channel = channel self.callerid = callerid self.account = account self.wait_time = wait_time self.max_retries = max_retries
d3c585a0ad11a75c308c70954184a526033f0420
setup.py
setup.py
import sys from setuptools import setup, find_packages, Command from distutils import log setup( name='diffenator', version='0.0.2', author="Google Fonts Project Authors", description="Font regression tester for Google Fonts", url="https://github.com/googlefonts/diffenator", license="Apache Software License 2.0", package_dir={"": "Lib"}, packages=find_packages("Lib"), entry_points={ "console_scripts": [ "diffenator = diffenator.__main__:main", ], }, dependency_links=[ "git+https://github.com/googlei18n/nototools/tarball/master#egg=nototools-0.0.1", ], install_requires=[ "fonttools>=3.4.0", ], )
import sys from setuptools import setup, find_packages, Command from distutils import log setup( name='diffenator', version='0.0.2', author="Google Fonts Project Authors", description="Font regression tester for Google Fonts", url="https://github.com/googlefonts/diffenator", license="Apache Software License 2.0", package_dir={"": "Lib"}, packages=find_packages("Lib"), entry_points={ "console_scripts": [ "diffenator = diffenator.__main__:main", ], }, install_requires=[ "fonttools>=3.4.0", ], )
Remove dependency link to nototools
Remove dependency link to nototools
Python
apache-2.0
googlefonts/fontdiffenator,googlefonts/fontdiffenator
import sys from setuptools import setup, find_packages, Command from distutils import log setup( name='diffenator', version='0.0.2', author="Google Fonts Project Authors", description="Font regression tester for Google Fonts", url="https://github.com/googlefonts/diffenator", license="Apache Software License 2.0", package_dir={"": "Lib"}, packages=find_packages("Lib"), entry_points={ "console_scripts": [ "diffenator = diffenator.__main__:main", ], }, dependency_links=[ "git+https://github.com/googlei18n/nototools/tarball/master#egg=nototools-0.0.1", ], install_requires=[ "fonttools>=3.4.0", ], ) Remove dependency link to nototools
import sys from setuptools import setup, find_packages, Command from distutils import log setup( name='diffenator', version='0.0.2', author="Google Fonts Project Authors", description="Font regression tester for Google Fonts", url="https://github.com/googlefonts/diffenator", license="Apache Software License 2.0", package_dir={"": "Lib"}, packages=find_packages("Lib"), entry_points={ "console_scripts": [ "diffenator = diffenator.__main__:main", ], }, install_requires=[ "fonttools>=3.4.0", ], )
<commit_before>import sys from setuptools import setup, find_packages, Command from distutils import log setup( name='diffenator', version='0.0.2', author="Google Fonts Project Authors", description="Font regression tester for Google Fonts", url="https://github.com/googlefonts/diffenator", license="Apache Software License 2.0", package_dir={"": "Lib"}, packages=find_packages("Lib"), entry_points={ "console_scripts": [ "diffenator = diffenator.__main__:main", ], }, dependency_links=[ "git+https://github.com/googlei18n/nototools/tarball/master#egg=nototools-0.0.1", ], install_requires=[ "fonttools>=3.4.0", ], ) <commit_msg>Remove dependency link to nototools<commit_after>
import sys from setuptools import setup, find_packages, Command from distutils import log setup( name='diffenator', version='0.0.2', author="Google Fonts Project Authors", description="Font regression tester for Google Fonts", url="https://github.com/googlefonts/diffenator", license="Apache Software License 2.0", package_dir={"": "Lib"}, packages=find_packages("Lib"), entry_points={ "console_scripts": [ "diffenator = diffenator.__main__:main", ], }, install_requires=[ "fonttools>=3.4.0", ], )
import sys from setuptools import setup, find_packages, Command from distutils import log setup( name='diffenator', version='0.0.2', author="Google Fonts Project Authors", description="Font regression tester for Google Fonts", url="https://github.com/googlefonts/diffenator", license="Apache Software License 2.0", package_dir={"": "Lib"}, packages=find_packages("Lib"), entry_points={ "console_scripts": [ "diffenator = diffenator.__main__:main", ], }, dependency_links=[ "git+https://github.com/googlei18n/nototools/tarball/master#egg=nototools-0.0.1", ], install_requires=[ "fonttools>=3.4.0", ], ) Remove dependency link to nototoolsimport sys from setuptools import setup, find_packages, Command from distutils import log setup( name='diffenator', version='0.0.2', author="Google Fonts Project Authors", description="Font regression tester for Google Fonts", url="https://github.com/googlefonts/diffenator", license="Apache Software License 2.0", package_dir={"": "Lib"}, packages=find_packages("Lib"), entry_points={ "console_scripts": [ "diffenator = diffenator.__main__:main", ], }, install_requires=[ "fonttools>=3.4.0", ], )
<commit_before>import sys from setuptools import setup, find_packages, Command from distutils import log setup( name='diffenator', version='0.0.2', author="Google Fonts Project Authors", description="Font regression tester for Google Fonts", url="https://github.com/googlefonts/diffenator", license="Apache Software License 2.0", package_dir={"": "Lib"}, packages=find_packages("Lib"), entry_points={ "console_scripts": [ "diffenator = diffenator.__main__:main", ], }, dependency_links=[ "git+https://github.com/googlei18n/nototools/tarball/master#egg=nototools-0.0.1", ], install_requires=[ "fonttools>=3.4.0", ], ) <commit_msg>Remove dependency link to nototools<commit_after>import sys from setuptools import setup, find_packages, Command from distutils import log setup( name='diffenator', version='0.0.2', author="Google Fonts Project Authors", description="Font regression tester for Google Fonts", url="https://github.com/googlefonts/diffenator", license="Apache Software License 2.0", package_dir={"": "Lib"}, packages=find_packages("Lib"), entry_points={ "console_scripts": [ "diffenator = diffenator.__main__:main", ], }, install_requires=[ "fonttools>=3.4.0", ], )
5250a31587f414f6673c13e42095dbb859bf8cb4
setup.py
setup.py
# Copyright 2013-2015 Massachusetts Open Cloud Contributors # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the # License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an "AS # IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either # express or implied. See the License for the specific language # governing permissions and limitations under the License. from setuptools import setup, find_packages setup(name='haas', version='0.2rc2', url='https://github.com/CCI-MOC/haas', packages=find_packages(), scripts=['scripts/haas', 'scripts/create_bridges'], install_requires=['SQLAlchemy==0.9.7', 'Werkzeug>=0.9.4,<0.10', 'Flask>=0.10.1,<0.11', 'schema==0.3.1', 'importlib==1.0.3', 'passlib==1.6.2', 'pexpect==3.3', 'requests==2.4.1', 'pytest>=2.6.2,<3.0', 'pytest-cov==1.8.0', 'pytest-xdist', ])
# Copyright 2013-2015 Massachusetts Open Cloud Contributors # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the # License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an "AS # IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either # express or implied. See the License for the specific language # governing permissions and limitations under the License. from setuptools import setup, find_packages setup(name='haas', version='0.2rc2', url='https://github.com/CCI-MOC/haas', packages=find_packages(), scripts=['scripts/haas', 'scripts/create_bridges'], install_requires=['Flask-SQLAlchemy', 'Werkzeug>=0.9.4,<0.10', 'Flask>=0.10.1,<0.11', 'schema==0.3.1', 'importlib==1.0.3', 'passlib==1.6.2', 'pexpect==3.3', 'requests==2.4.1', 'pytest>=2.6.2,<3.0', 'pytest-cov==1.8.0', 'pytest-xdist', ])
Replace dependency on SQLAlchemy with Flask-SQLALchemy
Replace dependency on SQLAlchemy with Flask-SQLALchemy
Python
apache-2.0
CCI-MOC/haas,kylehogan/haas,kylehogan/hil,henn/hil_sahil,henn/hil,henn/hil,kylehogan/hil,SahilTikale/haas,meng-sun/hil,henn/hil_sahil,meng-sun/hil,henn/haas
# Copyright 2013-2015 Massachusetts Open Cloud Contributors # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the # License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an "AS # IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either # express or implied. See the License for the specific language # governing permissions and limitations under the License. from setuptools import setup, find_packages setup(name='haas', version='0.2rc2', url='https://github.com/CCI-MOC/haas', packages=find_packages(), scripts=['scripts/haas', 'scripts/create_bridges'], install_requires=['SQLAlchemy==0.9.7', 'Werkzeug>=0.9.4,<0.10', 'Flask>=0.10.1,<0.11', 'schema==0.3.1', 'importlib==1.0.3', 'passlib==1.6.2', 'pexpect==3.3', 'requests==2.4.1', 'pytest>=2.6.2,<3.0', 'pytest-cov==1.8.0', 'pytest-xdist', ]) Replace dependency on SQLAlchemy with Flask-SQLALchemy
# Copyright 2013-2015 Massachusetts Open Cloud Contributors # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the # License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an "AS # IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either # express or implied. See the License for the specific language # governing permissions and limitations under the License. from setuptools import setup, find_packages setup(name='haas', version='0.2rc2', url='https://github.com/CCI-MOC/haas', packages=find_packages(), scripts=['scripts/haas', 'scripts/create_bridges'], install_requires=['Flask-SQLAlchemy', 'Werkzeug>=0.9.4,<0.10', 'Flask>=0.10.1,<0.11', 'schema==0.3.1', 'importlib==1.0.3', 'passlib==1.6.2', 'pexpect==3.3', 'requests==2.4.1', 'pytest>=2.6.2,<3.0', 'pytest-cov==1.8.0', 'pytest-xdist', ])
<commit_before># Copyright 2013-2015 Massachusetts Open Cloud Contributors # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the # License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an "AS # IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either # express or implied. See the License for the specific language # governing permissions and limitations under the License. from setuptools import setup, find_packages setup(name='haas', version='0.2rc2', url='https://github.com/CCI-MOC/haas', packages=find_packages(), scripts=['scripts/haas', 'scripts/create_bridges'], install_requires=['SQLAlchemy==0.9.7', 'Werkzeug>=0.9.4,<0.10', 'Flask>=0.10.1,<0.11', 'schema==0.3.1', 'importlib==1.0.3', 'passlib==1.6.2', 'pexpect==3.3', 'requests==2.4.1', 'pytest>=2.6.2,<3.0', 'pytest-cov==1.8.0', 'pytest-xdist', ]) <commit_msg>Replace dependency on SQLAlchemy with Flask-SQLALchemy<commit_after>
# Copyright 2013-2015 Massachusetts Open Cloud Contributors # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the # License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an "AS # IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either # express or implied. See the License for the specific language # governing permissions and limitations under the License. from setuptools import setup, find_packages setup(name='haas', version='0.2rc2', url='https://github.com/CCI-MOC/haas', packages=find_packages(), scripts=['scripts/haas', 'scripts/create_bridges'], install_requires=['Flask-SQLAlchemy', 'Werkzeug>=0.9.4,<0.10', 'Flask>=0.10.1,<0.11', 'schema==0.3.1', 'importlib==1.0.3', 'passlib==1.6.2', 'pexpect==3.3', 'requests==2.4.1', 'pytest>=2.6.2,<3.0', 'pytest-cov==1.8.0', 'pytest-xdist', ])
# Copyright 2013-2015 Massachusetts Open Cloud Contributors # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the # License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an "AS # IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either # express or implied. See the License for the specific language # governing permissions and limitations under the License. from setuptools import setup, find_packages setup(name='haas', version='0.2rc2', url='https://github.com/CCI-MOC/haas', packages=find_packages(), scripts=['scripts/haas', 'scripts/create_bridges'], install_requires=['SQLAlchemy==0.9.7', 'Werkzeug>=0.9.4,<0.10', 'Flask>=0.10.1,<0.11', 'schema==0.3.1', 'importlib==1.0.3', 'passlib==1.6.2', 'pexpect==3.3', 'requests==2.4.1', 'pytest>=2.6.2,<3.0', 'pytest-cov==1.8.0', 'pytest-xdist', ]) Replace dependency on SQLAlchemy with Flask-SQLALchemy# Copyright 2013-2015 Massachusetts Open Cloud Contributors # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the # License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an "AS # IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either # express or implied. See the License for the specific language # governing permissions and limitations under the License. from setuptools import setup, find_packages setup(name='haas', version='0.2rc2', url='https://github.com/CCI-MOC/haas', packages=find_packages(), scripts=['scripts/haas', 'scripts/create_bridges'], install_requires=['Flask-SQLAlchemy', 'Werkzeug>=0.9.4,<0.10', 'Flask>=0.10.1,<0.11', 'schema==0.3.1', 'importlib==1.0.3', 'passlib==1.6.2', 'pexpect==3.3', 'requests==2.4.1', 'pytest>=2.6.2,<3.0', 'pytest-cov==1.8.0', 'pytest-xdist', ])
<commit_before># Copyright 2013-2015 Massachusetts Open Cloud Contributors # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the # License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an "AS # IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either # express or implied. See the License for the specific language # governing permissions and limitations under the License. from setuptools import setup, find_packages setup(name='haas', version='0.2rc2', url='https://github.com/CCI-MOC/haas', packages=find_packages(), scripts=['scripts/haas', 'scripts/create_bridges'], install_requires=['SQLAlchemy==0.9.7', 'Werkzeug>=0.9.4,<0.10', 'Flask>=0.10.1,<0.11', 'schema==0.3.1', 'importlib==1.0.3', 'passlib==1.6.2', 'pexpect==3.3', 'requests==2.4.1', 'pytest>=2.6.2,<3.0', 'pytest-cov==1.8.0', 'pytest-xdist', ]) <commit_msg>Replace dependency on SQLAlchemy with Flask-SQLALchemy<commit_after># Copyright 2013-2015 Massachusetts Open Cloud Contributors # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the # License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an "AS # IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either # express or implied. See the License for the specific language # governing permissions and limitations under the License. from setuptools import setup, find_packages setup(name='haas', version='0.2rc2', url='https://github.com/CCI-MOC/haas', packages=find_packages(), scripts=['scripts/haas', 'scripts/create_bridges'], install_requires=['Flask-SQLAlchemy', 'Werkzeug>=0.9.4,<0.10', 'Flask>=0.10.1,<0.11', 'schema==0.3.1', 'importlib==1.0.3', 'passlib==1.6.2', 'pexpect==3.3', 'requests==2.4.1', 'pytest>=2.6.2,<3.0', 'pytest-cov==1.8.0', 'pytest-xdist', ])
475b89c2d4ce94bc208b610808916246e4e3a575
setup.py
setup.py
from setuptools import setup, Extension module1 = Extension('earl', sources = ['earl.cpp']) setup ( name = "Earl", version = "1.6", description = "Earl, the fanciest External Term Format packer and unpacker available for Python.", ext_modules = [module1], url="https://github.com/ccubed/Earl", author="Charles Click", author_email="CharlesClick@vertinext.com", license="MIT", keywords="Erlang ETF External Term Format" )
from setuptools import setup, Extension module1 = Extension('earl', sources = ['earl.cpp']) setup ( name = "Earl", version = "1.6.1", description = "Earl, the fanciest External Term Format packer and unpacker available for Python.", ext_modules = [module1], url="https://github.com/ccubed/Earl", author="Charles Click", author_email="CharlesClick@vertinext.com", license="MIT", keywords="Erlang ETF External Term Format" )
Update to 1.6.1. Small bug fix.
Update to 1.6.1. Small bug fix.
Python
mit
ccubed/Earl,ccubed/Earl
from setuptools import setup, Extension module1 = Extension('earl', sources = ['earl.cpp']) setup ( name = "Earl", version = "1.6", description = "Earl, the fanciest External Term Format packer and unpacker available for Python.", ext_modules = [module1], url="https://github.com/ccubed/Earl", author="Charles Click", author_email="CharlesClick@vertinext.com", license="MIT", keywords="Erlang ETF External Term Format" ) Update to 1.6.1. Small bug fix.
from setuptools import setup, Extension module1 = Extension('earl', sources = ['earl.cpp']) setup ( name = "Earl", version = "1.6.1", description = "Earl, the fanciest External Term Format packer and unpacker available for Python.", ext_modules = [module1], url="https://github.com/ccubed/Earl", author="Charles Click", author_email="CharlesClick@vertinext.com", license="MIT", keywords="Erlang ETF External Term Format" )
<commit_before>from setuptools import setup, Extension module1 = Extension('earl', sources = ['earl.cpp']) setup ( name = "Earl", version = "1.6", description = "Earl, the fanciest External Term Format packer and unpacker available for Python.", ext_modules = [module1], url="https://github.com/ccubed/Earl", author="Charles Click", author_email="CharlesClick@vertinext.com", license="MIT", keywords="Erlang ETF External Term Format" ) <commit_msg>Update to 1.6.1. Small bug fix.<commit_after>
from setuptools import setup, Extension module1 = Extension('earl', sources = ['earl.cpp']) setup ( name = "Earl", version = "1.6.1", description = "Earl, the fanciest External Term Format packer and unpacker available for Python.", ext_modules = [module1], url="https://github.com/ccubed/Earl", author="Charles Click", author_email="CharlesClick@vertinext.com", license="MIT", keywords="Erlang ETF External Term Format" )
from setuptools import setup, Extension module1 = Extension('earl', sources = ['earl.cpp']) setup ( name = "Earl", version = "1.6", description = "Earl, the fanciest External Term Format packer and unpacker available for Python.", ext_modules = [module1], url="https://github.com/ccubed/Earl", author="Charles Click", author_email="CharlesClick@vertinext.com", license="MIT", keywords="Erlang ETF External Term Format" ) Update to 1.6.1. Small bug fix.from setuptools import setup, Extension module1 = Extension('earl', sources = ['earl.cpp']) setup ( name = "Earl", version = "1.6.1", description = "Earl, the fanciest External Term Format packer and unpacker available for Python.", ext_modules = [module1], url="https://github.com/ccubed/Earl", author="Charles Click", author_email="CharlesClick@vertinext.com", license="MIT", keywords="Erlang ETF External Term Format" )
<commit_before>from setuptools import setup, Extension module1 = Extension('earl', sources = ['earl.cpp']) setup ( name = "Earl", version = "1.6", description = "Earl, the fanciest External Term Format packer and unpacker available for Python.", ext_modules = [module1], url="https://github.com/ccubed/Earl", author="Charles Click", author_email="CharlesClick@vertinext.com", license="MIT", keywords="Erlang ETF External Term Format" ) <commit_msg>Update to 1.6.1. Small bug fix.<commit_after>from setuptools import setup, Extension module1 = Extension('earl', sources = ['earl.cpp']) setup ( name = "Earl", version = "1.6.1", description = "Earl, the fanciest External Term Format packer and unpacker available for Python.", ext_modules = [module1], url="https://github.com/ccubed/Earl", author="Charles Click", author_email="CharlesClick@vertinext.com", license="MIT", keywords="Erlang ETF External Term Format" )
3daaf21879c6b2fc2099708e967cd4d67da1cd4f
setup.py
setup.py
#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup, find_packages __name__ == '__main__' and setup(name='aiohttp-json-rpc', version='0.10.1', author='Florian Scherf', url='https://github.com/pengutronix/aiohttp-json-rpc/', author_email='f.scherf@pengutronix.de', license='Apache 2.0', install_requires=['aiohttp>=3,<3.1'], python_requires='>=3.5', packages=find_packages(), zip_safe=False, entry_points={ 'pytest11': [ 'aiohttp-json-rpc = aiohttp_json_rpc.pytest', ] })
#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup, find_packages __name__ == '__main__' and setup(name='aiohttp-json-rpc', version='0.10.1', author='Florian Scherf', url='https://github.com/pengutronix/aiohttp-json-rpc/', author_email='f.scherf@pengutronix.de', license='Apache 2.0', install_requires=['aiohttp>=3,<3.4'], python_requires='>=3.5', packages=find_packages(), zip_safe=False, entry_points={ 'pytest11': [ 'aiohttp-json-rpc = aiohttp_json_rpc.pytest', ] })
Bump aiohttp version constraint to <3.4
Bump aiohttp version constraint to <3.4
Python
apache-2.0
pengutronix/aiohttp-json-rpc,pengutronix/aiohttp-json-rpc,pengutronix/aiohttp-json-rpc
#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup, find_packages __name__ == '__main__' and setup(name='aiohttp-json-rpc', version='0.10.1', author='Florian Scherf', url='https://github.com/pengutronix/aiohttp-json-rpc/', author_email='f.scherf@pengutronix.de', license='Apache 2.0', install_requires=['aiohttp>=3,<3.1'], python_requires='>=3.5', packages=find_packages(), zip_safe=False, entry_points={ 'pytest11': [ 'aiohttp-json-rpc = aiohttp_json_rpc.pytest', ] }) Bump aiohttp version constraint to <3.4
#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup, find_packages __name__ == '__main__' and setup(name='aiohttp-json-rpc', version='0.10.1', author='Florian Scherf', url='https://github.com/pengutronix/aiohttp-json-rpc/', author_email='f.scherf@pengutronix.de', license='Apache 2.0', install_requires=['aiohttp>=3,<3.4'], python_requires='>=3.5', packages=find_packages(), zip_safe=False, entry_points={ 'pytest11': [ 'aiohttp-json-rpc = aiohttp_json_rpc.pytest', ] })
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup, find_packages __name__ == '__main__' and setup(name='aiohttp-json-rpc', version='0.10.1', author='Florian Scherf', url='https://github.com/pengutronix/aiohttp-json-rpc/', author_email='f.scherf@pengutronix.de', license='Apache 2.0', install_requires=['aiohttp>=3,<3.1'], python_requires='>=3.5', packages=find_packages(), zip_safe=False, entry_points={ 'pytest11': [ 'aiohttp-json-rpc = aiohttp_json_rpc.pytest', ] }) <commit_msg>Bump aiohttp version constraint to <3.4<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup, find_packages __name__ == '__main__' and setup(name='aiohttp-json-rpc', version='0.10.1', author='Florian Scherf', url='https://github.com/pengutronix/aiohttp-json-rpc/', author_email='f.scherf@pengutronix.de', license='Apache 2.0', install_requires=['aiohttp>=3,<3.4'], python_requires='>=3.5', packages=find_packages(), zip_safe=False, entry_points={ 'pytest11': [ 'aiohttp-json-rpc = aiohttp_json_rpc.pytest', ] })
#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup, find_packages __name__ == '__main__' and setup(name='aiohttp-json-rpc', version='0.10.1', author='Florian Scherf', url='https://github.com/pengutronix/aiohttp-json-rpc/', author_email='f.scherf@pengutronix.de', license='Apache 2.0', install_requires=['aiohttp>=3,<3.1'], python_requires='>=3.5', packages=find_packages(), zip_safe=False, entry_points={ 'pytest11': [ 'aiohttp-json-rpc = aiohttp_json_rpc.pytest', ] }) Bump aiohttp version constraint to <3.4#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup, find_packages __name__ == '__main__' and setup(name='aiohttp-json-rpc', version='0.10.1', author='Florian Scherf', url='https://github.com/pengutronix/aiohttp-json-rpc/', author_email='f.scherf@pengutronix.de', license='Apache 2.0', install_requires=['aiohttp>=3,<3.4'], python_requires='>=3.5', packages=find_packages(), zip_safe=False, entry_points={ 'pytest11': [ 'aiohttp-json-rpc = aiohttp_json_rpc.pytest', ] })
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup, find_packages __name__ == '__main__' and setup(name='aiohttp-json-rpc', version='0.10.1', author='Florian Scherf', url='https://github.com/pengutronix/aiohttp-json-rpc/', author_email='f.scherf@pengutronix.de', license='Apache 2.0', install_requires=['aiohttp>=3,<3.1'], python_requires='>=3.5', packages=find_packages(), zip_safe=False, entry_points={ 'pytest11': [ 'aiohttp-json-rpc = aiohttp_json_rpc.pytest', ] }) <commit_msg>Bump aiohttp version constraint to <3.4<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- from setuptools import setup, find_packages __name__ == '__main__' and setup(name='aiohttp-json-rpc', version='0.10.1', author='Florian Scherf', url='https://github.com/pengutronix/aiohttp-json-rpc/', author_email='f.scherf@pengutronix.de', license='Apache 2.0', install_requires=['aiohttp>=3,<3.4'], python_requires='>=3.5', packages=find_packages(), zip_safe=False, entry_points={ 'pytest11': [ 'aiohttp-json-rpc = aiohttp_json_rpc.pytest', ] })
d7201be1fbd5659f60887572f18d103a3f71deaf
setup.py
setup.py
"""Setup for pyexperiment """ from __future__ import print_function # from __future__ import unicode_literals from __future__ import division from __future__ import absolute_import import os from setuptools import setup try: from pypandoc import convert read_md = lambda fname: convert(fname, 'rst') except ImportError: print("Warning: pypandoc module not found") read_md = lambda fname: open( os.path.join(os.path.dirname(__file__), fname), 'r').read() LONG_DESCRIPTION = 'Framework for easy and clean experiments with python.' if os.path.exists('README.md'): LONG_DESCRIPTION = read_md('README.md') setup( name="pyexperiment", version="0.1.15", author="Peter Duerr", author_email="duerrp@gmail.com", description="Run experiments with Python - quick and clean.", license="MIT", keywords="science experiment", url="https://github.com/duerrp/pyexperiment", download_url="https://github.com/duerrp/pyexperiment/tarball/0.1.15", packages=['pyexperiment', 'pyexperiment.conf', 'pyexperiment.state', 'pyexperiment.utils', 'pyexperiment.log', ], long_description=LONG_DESCRIPTION, classifiers=[ "Development Status :: 3 - Alpha", "Topic :: Utilities", "License :: OSI Approved :: MIT License", ], )
"""Setup for pyexperiment """ from __future__ import print_function # from __future__ import unicode_literals from __future__ import division from __future__ import absolute_import import os from setuptools import setup try: from pypandoc import convert read_md = lambda fname: convert(fname, 'rst') except ImportError: print("Warning: pypandoc module not found") read_md = lambda fname: open( os.path.join(os.path.dirname(__file__), fname), 'r').read() LONG_DESCRIPTION = 'Framework for easy and clean experiments with python.' if os.path.exists('README.md'): LONG_DESCRIPTION = read_md('README.md') setup( name="pyexperiment", version="0.1.15-test", author="Peter Duerr", author_email="duerrp@gmail.com", description="Run experiments with Python - quick and clean.", license="MIT", keywords="science experiment", url="https://github.com/duerrp/pyexperiment", # download_url="https://github.com/duerrp/pyexperiment/tarball/0.1.15", packages=['pyexperiment', 'pyexperiment.conf', 'pyexperiment.state', 'pyexperiment.utils', 'pyexperiment.log', ], long_description=LONG_DESCRIPTION, classifiers=[ "Development Status :: 3 - Alpha", "Topic :: Utilities", "License :: OSI Approved :: MIT License", ], )
Move to test version for pypi-test
Move to test version for pypi-test
Python
mit
DeercoderResearch/pyexperiment,shaunstanislaus/pyexperiment,duerrp/pyexperiment,kinverarity1/pyexperiment,duerrp/pyexperiment,shaunstanislaus/pyexperiment,shaunstanislaus/pyexperiment,DeercoderResearch/pyexperiment,DeercoderResearch/pyexperiment,kinverarity1/pyexperiment,shaunstanislaus/pyexperiment,kinverarity1/pyexperiment,kinverarity1/pyexperiment,duerrp/pyexperiment,DeercoderResearch/pyexperiment
"""Setup for pyexperiment """ from __future__ import print_function # from __future__ import unicode_literals from __future__ import division from __future__ import absolute_import import os from setuptools import setup try: from pypandoc import convert read_md = lambda fname: convert(fname, 'rst') except ImportError: print("Warning: pypandoc module not found") read_md = lambda fname: open( os.path.join(os.path.dirname(__file__), fname), 'r').read() LONG_DESCRIPTION = 'Framework for easy and clean experiments with python.' if os.path.exists('README.md'): LONG_DESCRIPTION = read_md('README.md') setup( name="pyexperiment", version="0.1.15", author="Peter Duerr", author_email="duerrp@gmail.com", description="Run experiments with Python - quick and clean.", license="MIT", keywords="science experiment", url="https://github.com/duerrp/pyexperiment", download_url="https://github.com/duerrp/pyexperiment/tarball/0.1.15", packages=['pyexperiment', 'pyexperiment.conf', 'pyexperiment.state', 'pyexperiment.utils', 'pyexperiment.log', ], long_description=LONG_DESCRIPTION, classifiers=[ "Development Status :: 3 - Alpha", "Topic :: Utilities", "License :: OSI Approved :: MIT License", ], ) Move to test version for pypi-test
"""Setup for pyexperiment """ from __future__ import print_function # from __future__ import unicode_literals from __future__ import division from __future__ import absolute_import import os from setuptools import setup try: from pypandoc import convert read_md = lambda fname: convert(fname, 'rst') except ImportError: print("Warning: pypandoc module not found") read_md = lambda fname: open( os.path.join(os.path.dirname(__file__), fname), 'r').read() LONG_DESCRIPTION = 'Framework for easy and clean experiments with python.' if os.path.exists('README.md'): LONG_DESCRIPTION = read_md('README.md') setup( name="pyexperiment", version="0.1.15-test", author="Peter Duerr", author_email="duerrp@gmail.com", description="Run experiments with Python - quick and clean.", license="MIT", keywords="science experiment", url="https://github.com/duerrp/pyexperiment", # download_url="https://github.com/duerrp/pyexperiment/tarball/0.1.15", packages=['pyexperiment', 'pyexperiment.conf', 'pyexperiment.state', 'pyexperiment.utils', 'pyexperiment.log', ], long_description=LONG_DESCRIPTION, classifiers=[ "Development Status :: 3 - Alpha", "Topic :: Utilities", "License :: OSI Approved :: MIT License", ], )
<commit_before>"""Setup for pyexperiment """ from __future__ import print_function # from __future__ import unicode_literals from __future__ import division from __future__ import absolute_import import os from setuptools import setup try: from pypandoc import convert read_md = lambda fname: convert(fname, 'rst') except ImportError: print("Warning: pypandoc module not found") read_md = lambda fname: open( os.path.join(os.path.dirname(__file__), fname), 'r').read() LONG_DESCRIPTION = 'Framework for easy and clean experiments with python.' if os.path.exists('README.md'): LONG_DESCRIPTION = read_md('README.md') setup( name="pyexperiment", version="0.1.15", author="Peter Duerr", author_email="duerrp@gmail.com", description="Run experiments with Python - quick and clean.", license="MIT", keywords="science experiment", url="https://github.com/duerrp/pyexperiment", download_url="https://github.com/duerrp/pyexperiment/tarball/0.1.15", packages=['pyexperiment', 'pyexperiment.conf', 'pyexperiment.state', 'pyexperiment.utils', 'pyexperiment.log', ], long_description=LONG_DESCRIPTION, classifiers=[ "Development Status :: 3 - Alpha", "Topic :: Utilities", "License :: OSI Approved :: MIT License", ], ) <commit_msg>Move to test version for pypi-test<commit_after>
"""Setup for pyexperiment """ from __future__ import print_function # from __future__ import unicode_literals from __future__ import division from __future__ import absolute_import import os from setuptools import setup try: from pypandoc import convert read_md = lambda fname: convert(fname, 'rst') except ImportError: print("Warning: pypandoc module not found") read_md = lambda fname: open( os.path.join(os.path.dirname(__file__), fname), 'r').read() LONG_DESCRIPTION = 'Framework for easy and clean experiments with python.' if os.path.exists('README.md'): LONG_DESCRIPTION = read_md('README.md') setup( name="pyexperiment", version="0.1.15-test", author="Peter Duerr", author_email="duerrp@gmail.com", description="Run experiments with Python - quick and clean.", license="MIT", keywords="science experiment", url="https://github.com/duerrp/pyexperiment", # download_url="https://github.com/duerrp/pyexperiment/tarball/0.1.15", packages=['pyexperiment', 'pyexperiment.conf', 'pyexperiment.state', 'pyexperiment.utils', 'pyexperiment.log', ], long_description=LONG_DESCRIPTION, classifiers=[ "Development Status :: 3 - Alpha", "Topic :: Utilities", "License :: OSI Approved :: MIT License", ], )
"""Setup for pyexperiment """ from __future__ import print_function # from __future__ import unicode_literals from __future__ import division from __future__ import absolute_import import os from setuptools import setup try: from pypandoc import convert read_md = lambda fname: convert(fname, 'rst') except ImportError: print("Warning: pypandoc module not found") read_md = lambda fname: open( os.path.join(os.path.dirname(__file__), fname), 'r').read() LONG_DESCRIPTION = 'Framework for easy and clean experiments with python.' if os.path.exists('README.md'): LONG_DESCRIPTION = read_md('README.md') setup( name="pyexperiment", version="0.1.15", author="Peter Duerr", author_email="duerrp@gmail.com", description="Run experiments with Python - quick and clean.", license="MIT", keywords="science experiment", url="https://github.com/duerrp/pyexperiment", download_url="https://github.com/duerrp/pyexperiment/tarball/0.1.15", packages=['pyexperiment', 'pyexperiment.conf', 'pyexperiment.state', 'pyexperiment.utils', 'pyexperiment.log', ], long_description=LONG_DESCRIPTION, classifiers=[ "Development Status :: 3 - Alpha", "Topic :: Utilities", "License :: OSI Approved :: MIT License", ], ) Move to test version for pypi-test"""Setup for pyexperiment """ from __future__ import print_function # from __future__ import unicode_literals from __future__ import division from __future__ import absolute_import import os from setuptools import setup try: from pypandoc import convert read_md = lambda fname: convert(fname, 'rst') except ImportError: print("Warning: pypandoc module not found") read_md = lambda fname: open( os.path.join(os.path.dirname(__file__), fname), 'r').read() LONG_DESCRIPTION = 'Framework for easy and clean experiments with python.' if os.path.exists('README.md'): LONG_DESCRIPTION = read_md('README.md') setup( name="pyexperiment", version="0.1.15-test", author="Peter Duerr", author_email="duerrp@gmail.com", description="Run experiments with Python - quick and clean.", license="MIT", keywords="science experiment", url="https://github.com/duerrp/pyexperiment", # download_url="https://github.com/duerrp/pyexperiment/tarball/0.1.15", packages=['pyexperiment', 'pyexperiment.conf', 'pyexperiment.state', 'pyexperiment.utils', 'pyexperiment.log', ], long_description=LONG_DESCRIPTION, classifiers=[ "Development Status :: 3 - Alpha", "Topic :: Utilities", "License :: OSI Approved :: MIT License", ], )
<commit_before>"""Setup for pyexperiment """ from __future__ import print_function # from __future__ import unicode_literals from __future__ import division from __future__ import absolute_import import os from setuptools import setup try: from pypandoc import convert read_md = lambda fname: convert(fname, 'rst') except ImportError: print("Warning: pypandoc module not found") read_md = lambda fname: open( os.path.join(os.path.dirname(__file__), fname), 'r').read() LONG_DESCRIPTION = 'Framework for easy and clean experiments with python.' if os.path.exists('README.md'): LONG_DESCRIPTION = read_md('README.md') setup( name="pyexperiment", version="0.1.15", author="Peter Duerr", author_email="duerrp@gmail.com", description="Run experiments with Python - quick and clean.", license="MIT", keywords="science experiment", url="https://github.com/duerrp/pyexperiment", download_url="https://github.com/duerrp/pyexperiment/tarball/0.1.15", packages=['pyexperiment', 'pyexperiment.conf', 'pyexperiment.state', 'pyexperiment.utils', 'pyexperiment.log', ], long_description=LONG_DESCRIPTION, classifiers=[ "Development Status :: 3 - Alpha", "Topic :: Utilities", "License :: OSI Approved :: MIT License", ], ) <commit_msg>Move to test version for pypi-test<commit_after>"""Setup for pyexperiment """ from __future__ import print_function # from __future__ import unicode_literals from __future__ import division from __future__ import absolute_import import os from setuptools import setup try: from pypandoc import convert read_md = lambda fname: convert(fname, 'rst') except ImportError: print("Warning: pypandoc module not found") read_md = lambda fname: open( os.path.join(os.path.dirname(__file__), fname), 'r').read() LONG_DESCRIPTION = 'Framework for easy and clean experiments with python.' if os.path.exists('README.md'): LONG_DESCRIPTION = read_md('README.md') setup( name="pyexperiment", version="0.1.15-test", author="Peter Duerr", author_email="duerrp@gmail.com", description="Run experiments with Python - quick and clean.", license="MIT", keywords="science experiment", url="https://github.com/duerrp/pyexperiment", # download_url="https://github.com/duerrp/pyexperiment/tarball/0.1.15", packages=['pyexperiment', 'pyexperiment.conf', 'pyexperiment.state', 'pyexperiment.utils', 'pyexperiment.log', ], long_description=LONG_DESCRIPTION, classifiers=[ "Development Status :: 3 - Alpha", "Topic :: Utilities", "License :: OSI Approved :: MIT License", ], )
70377c4397680c32b9ee6958bd250dce697fcb62
setup.py
setup.py
from setuptools import setup, find_packages install_requires = [ 'dill==0.2.5', 'easydict==1.6', 'h5py==2.6.0', 'jsonpickle==0.9.3', 'Keras==1.2.0', 'nflgame==1.2.20', 'numpy==1.11.2', 'pandas==0.19.1', 'scikit-learn==0.18.1', 'scipy==0.18.1', 'tensorflow==0.12.0rc1', 'Theano==0.8.2', ] with open('README.md', 'r') as f: readme = f.read() setup( name="wincast", version='0.0.1', url='https://github.com/kahnjw/wincast', author_email='thomas.welfley+djproxy@gmail.com', long_description=readme, license='MIT', packages=find_packages(exclude=['tests', 'tests.*']), install_requires=install_requires, data_files=[ ('', ['models/wincast.model.h5', 'models/wincast.scaler.pkl']), ('', ['data/Xy.csv']) ] )
from setuptools import setup, find_packages install_requires = [ 'dill==0.2.5', 'easydict==1.6', 'h5py==2.6.0', 'jsonpickle==0.9.3', 'Keras==1.2.0', 'nflgame==1.2.20', 'numpy==1.11.2', 'pandas==0.19.1', 'scikit-learn==0.18.1', 'scipy==0.18.1', 'tensorflow==0.12.0rc1', 'Theano==0.8.2', ] with open('README.md', 'r') as f: readme = f.read() setup( name="wincast", version='0.0.1', url='https://github.com/kahnjw/wincast', author_email='thomas.welfley+djproxy@gmail.com', long_description=readme, license='MIT', packages=find_packages(exclude=['tests', 'tests.*']), install_requires=install_requires, data_files=[ ('models', ['models/wincast.model.h5', 'models/wincast.scaler.pkl']), ('data', ['data/Xy.csv']) ] )
Add back models and data
Add back models and data
Python
mit
kahnjw/wincast
from setuptools import setup, find_packages install_requires = [ 'dill==0.2.5', 'easydict==1.6', 'h5py==2.6.0', 'jsonpickle==0.9.3', 'Keras==1.2.0', 'nflgame==1.2.20', 'numpy==1.11.2', 'pandas==0.19.1', 'scikit-learn==0.18.1', 'scipy==0.18.1', 'tensorflow==0.12.0rc1', 'Theano==0.8.2', ] with open('README.md', 'r') as f: readme = f.read() setup( name="wincast", version='0.0.1', url='https://github.com/kahnjw/wincast', author_email='thomas.welfley+djproxy@gmail.com', long_description=readme, license='MIT', packages=find_packages(exclude=['tests', 'tests.*']), install_requires=install_requires, data_files=[ ('', ['models/wincast.model.h5', 'models/wincast.scaler.pkl']), ('', ['data/Xy.csv']) ] ) Add back models and data
from setuptools import setup, find_packages install_requires = [ 'dill==0.2.5', 'easydict==1.6', 'h5py==2.6.0', 'jsonpickle==0.9.3', 'Keras==1.2.0', 'nflgame==1.2.20', 'numpy==1.11.2', 'pandas==0.19.1', 'scikit-learn==0.18.1', 'scipy==0.18.1', 'tensorflow==0.12.0rc1', 'Theano==0.8.2', ] with open('README.md', 'r') as f: readme = f.read() setup( name="wincast", version='0.0.1', url='https://github.com/kahnjw/wincast', author_email='thomas.welfley+djproxy@gmail.com', long_description=readme, license='MIT', packages=find_packages(exclude=['tests', 'tests.*']), install_requires=install_requires, data_files=[ ('models', ['models/wincast.model.h5', 'models/wincast.scaler.pkl']), ('data', ['data/Xy.csv']) ] )
<commit_before>from setuptools import setup, find_packages install_requires = [ 'dill==0.2.5', 'easydict==1.6', 'h5py==2.6.0', 'jsonpickle==0.9.3', 'Keras==1.2.0', 'nflgame==1.2.20', 'numpy==1.11.2', 'pandas==0.19.1', 'scikit-learn==0.18.1', 'scipy==0.18.1', 'tensorflow==0.12.0rc1', 'Theano==0.8.2', ] with open('README.md', 'r') as f: readme = f.read() setup( name="wincast", version='0.0.1', url='https://github.com/kahnjw/wincast', author_email='thomas.welfley+djproxy@gmail.com', long_description=readme, license='MIT', packages=find_packages(exclude=['tests', 'tests.*']), install_requires=install_requires, data_files=[ ('', ['models/wincast.model.h5', 'models/wincast.scaler.pkl']), ('', ['data/Xy.csv']) ] ) <commit_msg>Add back models and data<commit_after>
from setuptools import setup, find_packages install_requires = [ 'dill==0.2.5', 'easydict==1.6', 'h5py==2.6.0', 'jsonpickle==0.9.3', 'Keras==1.2.0', 'nflgame==1.2.20', 'numpy==1.11.2', 'pandas==0.19.1', 'scikit-learn==0.18.1', 'scipy==0.18.1', 'tensorflow==0.12.0rc1', 'Theano==0.8.2', ] with open('README.md', 'r') as f: readme = f.read() setup( name="wincast", version='0.0.1', url='https://github.com/kahnjw/wincast', author_email='thomas.welfley+djproxy@gmail.com', long_description=readme, license='MIT', packages=find_packages(exclude=['tests', 'tests.*']), install_requires=install_requires, data_files=[ ('models', ['models/wincast.model.h5', 'models/wincast.scaler.pkl']), ('data', ['data/Xy.csv']) ] )
from setuptools import setup, find_packages install_requires = [ 'dill==0.2.5', 'easydict==1.6', 'h5py==2.6.0', 'jsonpickle==0.9.3', 'Keras==1.2.0', 'nflgame==1.2.20', 'numpy==1.11.2', 'pandas==0.19.1', 'scikit-learn==0.18.1', 'scipy==0.18.1', 'tensorflow==0.12.0rc1', 'Theano==0.8.2', ] with open('README.md', 'r') as f: readme = f.read() setup( name="wincast", version='0.0.1', url='https://github.com/kahnjw/wincast', author_email='thomas.welfley+djproxy@gmail.com', long_description=readme, license='MIT', packages=find_packages(exclude=['tests', 'tests.*']), install_requires=install_requires, data_files=[ ('', ['models/wincast.model.h5', 'models/wincast.scaler.pkl']), ('', ['data/Xy.csv']) ] ) Add back models and datafrom setuptools import setup, find_packages install_requires = [ 'dill==0.2.5', 'easydict==1.6', 'h5py==2.6.0', 'jsonpickle==0.9.3', 'Keras==1.2.0', 'nflgame==1.2.20', 'numpy==1.11.2', 'pandas==0.19.1', 'scikit-learn==0.18.1', 'scipy==0.18.1', 'tensorflow==0.12.0rc1', 'Theano==0.8.2', ] with open('README.md', 'r') as f: readme = f.read() setup( name="wincast", version='0.0.1', url='https://github.com/kahnjw/wincast', author_email='thomas.welfley+djproxy@gmail.com', long_description=readme, license='MIT', packages=find_packages(exclude=['tests', 'tests.*']), install_requires=install_requires, data_files=[ ('models', ['models/wincast.model.h5', 'models/wincast.scaler.pkl']), ('data', ['data/Xy.csv']) ] )
<commit_before>from setuptools import setup, find_packages install_requires = [ 'dill==0.2.5', 'easydict==1.6', 'h5py==2.6.0', 'jsonpickle==0.9.3', 'Keras==1.2.0', 'nflgame==1.2.20', 'numpy==1.11.2', 'pandas==0.19.1', 'scikit-learn==0.18.1', 'scipy==0.18.1', 'tensorflow==0.12.0rc1', 'Theano==0.8.2', ] with open('README.md', 'r') as f: readme = f.read() setup( name="wincast", version='0.0.1', url='https://github.com/kahnjw/wincast', author_email='thomas.welfley+djproxy@gmail.com', long_description=readme, license='MIT', packages=find_packages(exclude=['tests', 'tests.*']), install_requires=install_requires, data_files=[ ('', ['models/wincast.model.h5', 'models/wincast.scaler.pkl']), ('', ['data/Xy.csv']) ] ) <commit_msg>Add back models and data<commit_after>from setuptools import setup, find_packages install_requires = [ 'dill==0.2.5', 'easydict==1.6', 'h5py==2.6.0', 'jsonpickle==0.9.3', 'Keras==1.2.0', 'nflgame==1.2.20', 'numpy==1.11.2', 'pandas==0.19.1', 'scikit-learn==0.18.1', 'scipy==0.18.1', 'tensorflow==0.12.0rc1', 'Theano==0.8.2', ] with open('README.md', 'r') as f: readme = f.read() setup( name="wincast", version='0.0.1', url='https://github.com/kahnjw/wincast', author_email='thomas.welfley+djproxy@gmail.com', long_description=readme, license='MIT', packages=find_packages(exclude=['tests', 'tests.*']), install_requires=install_requires, data_files=[ ('models', ['models/wincast.model.h5', 'models/wincast.scaler.pkl']), ('data', ['data/Xy.csv']) ] )
739860d8da3e3380e49283a1fca2c43750349909
setup.py
setup.py
from setuptools import setup import mikla setup( name='mikla', version=mikla.__version__.strip(), url='http://dirtymonkey.co.uk/mikla', license='MIT', author=mikla.__author__.strip(), author_email='matt@dirtymonkey.co.uk', description=mikla.__doc__.strip().replace('\n', ' '), long_description=open('README.rst').read(), keywords='encryption security gnupg gpg', packages=['mikla'], include_package_data=True, entry_points={ 'console_scripts': [ 'mikla = mikla.main:main', ], }, install_requires=[ 'docopt>=0.6.2', ], classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: End Users/Desktop', 'License :: OSI Approved :: MIT License', 'Natural Language :: English', 'Operating System :: POSIX :: Linux', 'Operating System :: POSIX :: BSD', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Security :: Cryptography', 'Topic :: Communications', 'Topic :: Utilities', ], )
from __future__ import print_function import sys from setuptools import setup if sys.version_info[:2] < (3, 5): print('Mikla only runs on Python 3.5 or later', file=sys.stderr) sys.exit(1) import mikla setup( name='mikla', version=mikla.__version__.strip(), url='http://dirtymonkey.co.uk/mikla', license='MIT', author=mikla.__author__.strip(), author_email='matt@dirtymonkey.co.uk', description=mikla.__doc__.strip().replace('\n', ' '), long_description=open('README.rst').read(), keywords='encryption security gnupg gpg', packages=['mikla'], include_package_data=True, entry_points={ 'console_scripts': [ 'mikla = mikla.main:main', ], }, install_requires=[ 'docopt>=0.6.2', ], classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: End Users/Desktop', 'License :: OSI Approved :: MIT License', 'Natural Language :: English', 'Operating System :: POSIX :: Linux', 'Operating System :: POSIX :: BSD', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Security :: Cryptography', 'Topic :: Communications', 'Topic :: Utilities', ], )
Add helpful message for users trying to install with Python < 3.5
Add helpful message for users trying to install with Python < 3.5
Python
mit
Matt-Deacalion/Mikla
from setuptools import setup import mikla setup( name='mikla', version=mikla.__version__.strip(), url='http://dirtymonkey.co.uk/mikla', license='MIT', author=mikla.__author__.strip(), author_email='matt@dirtymonkey.co.uk', description=mikla.__doc__.strip().replace('\n', ' '), long_description=open('README.rst').read(), keywords='encryption security gnupg gpg', packages=['mikla'], include_package_data=True, entry_points={ 'console_scripts': [ 'mikla = mikla.main:main', ], }, install_requires=[ 'docopt>=0.6.2', ], classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: End Users/Desktop', 'License :: OSI Approved :: MIT License', 'Natural Language :: English', 'Operating System :: POSIX :: Linux', 'Operating System :: POSIX :: BSD', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Security :: Cryptography', 'Topic :: Communications', 'Topic :: Utilities', ], ) Add helpful message for users trying to install with Python < 3.5
from __future__ import print_function import sys from setuptools import setup if sys.version_info[:2] < (3, 5): print('Mikla only runs on Python 3.5 or later', file=sys.stderr) sys.exit(1) import mikla setup( name='mikla', version=mikla.__version__.strip(), url='http://dirtymonkey.co.uk/mikla', license='MIT', author=mikla.__author__.strip(), author_email='matt@dirtymonkey.co.uk', description=mikla.__doc__.strip().replace('\n', ' '), long_description=open('README.rst').read(), keywords='encryption security gnupg gpg', packages=['mikla'], include_package_data=True, entry_points={ 'console_scripts': [ 'mikla = mikla.main:main', ], }, install_requires=[ 'docopt>=0.6.2', ], classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: End Users/Desktop', 'License :: OSI Approved :: MIT License', 'Natural Language :: English', 'Operating System :: POSIX :: Linux', 'Operating System :: POSIX :: BSD', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Security :: Cryptography', 'Topic :: Communications', 'Topic :: Utilities', ], )
<commit_before>from setuptools import setup import mikla setup( name='mikla', version=mikla.__version__.strip(), url='http://dirtymonkey.co.uk/mikla', license='MIT', author=mikla.__author__.strip(), author_email='matt@dirtymonkey.co.uk', description=mikla.__doc__.strip().replace('\n', ' '), long_description=open('README.rst').read(), keywords='encryption security gnupg gpg', packages=['mikla'], include_package_data=True, entry_points={ 'console_scripts': [ 'mikla = mikla.main:main', ], }, install_requires=[ 'docopt>=0.6.2', ], classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: End Users/Desktop', 'License :: OSI Approved :: MIT License', 'Natural Language :: English', 'Operating System :: POSIX :: Linux', 'Operating System :: POSIX :: BSD', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Security :: Cryptography', 'Topic :: Communications', 'Topic :: Utilities', ], ) <commit_msg>Add helpful message for users trying to install with Python < 3.5<commit_after>
from __future__ import print_function import sys from setuptools import setup if sys.version_info[:2] < (3, 5): print('Mikla only runs on Python 3.5 or later', file=sys.stderr) sys.exit(1) import mikla setup( name='mikla', version=mikla.__version__.strip(), url='http://dirtymonkey.co.uk/mikla', license='MIT', author=mikla.__author__.strip(), author_email='matt@dirtymonkey.co.uk', description=mikla.__doc__.strip().replace('\n', ' '), long_description=open('README.rst').read(), keywords='encryption security gnupg gpg', packages=['mikla'], include_package_data=True, entry_points={ 'console_scripts': [ 'mikla = mikla.main:main', ], }, install_requires=[ 'docopt>=0.6.2', ], classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: End Users/Desktop', 'License :: OSI Approved :: MIT License', 'Natural Language :: English', 'Operating System :: POSIX :: Linux', 'Operating System :: POSIX :: BSD', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Security :: Cryptography', 'Topic :: Communications', 'Topic :: Utilities', ], )
from setuptools import setup import mikla setup( name='mikla', version=mikla.__version__.strip(), url='http://dirtymonkey.co.uk/mikla', license='MIT', author=mikla.__author__.strip(), author_email='matt@dirtymonkey.co.uk', description=mikla.__doc__.strip().replace('\n', ' '), long_description=open('README.rst').read(), keywords='encryption security gnupg gpg', packages=['mikla'], include_package_data=True, entry_points={ 'console_scripts': [ 'mikla = mikla.main:main', ], }, install_requires=[ 'docopt>=0.6.2', ], classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: End Users/Desktop', 'License :: OSI Approved :: MIT License', 'Natural Language :: English', 'Operating System :: POSIX :: Linux', 'Operating System :: POSIX :: BSD', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Security :: Cryptography', 'Topic :: Communications', 'Topic :: Utilities', ], ) Add helpful message for users trying to install with Python < 3.5from __future__ import print_function import sys from setuptools import setup if sys.version_info[:2] < (3, 5): print('Mikla only runs on Python 3.5 or later', file=sys.stderr) sys.exit(1) import mikla setup( name='mikla', version=mikla.__version__.strip(), url='http://dirtymonkey.co.uk/mikla', license='MIT', author=mikla.__author__.strip(), author_email='matt@dirtymonkey.co.uk', description=mikla.__doc__.strip().replace('\n', ' '), long_description=open('README.rst').read(), keywords='encryption security gnupg gpg', packages=['mikla'], include_package_data=True, entry_points={ 'console_scripts': [ 'mikla = mikla.main:main', ], }, install_requires=[ 'docopt>=0.6.2', ], classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: End Users/Desktop', 'License :: OSI Approved :: MIT License', 'Natural Language :: English', 'Operating System :: POSIX :: Linux', 'Operating System :: POSIX :: BSD', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Security :: Cryptography', 'Topic :: Communications', 'Topic :: Utilities', ], )
<commit_before>from setuptools import setup import mikla setup( name='mikla', version=mikla.__version__.strip(), url='http://dirtymonkey.co.uk/mikla', license='MIT', author=mikla.__author__.strip(), author_email='matt@dirtymonkey.co.uk', description=mikla.__doc__.strip().replace('\n', ' '), long_description=open('README.rst').read(), keywords='encryption security gnupg gpg', packages=['mikla'], include_package_data=True, entry_points={ 'console_scripts': [ 'mikla = mikla.main:main', ], }, install_requires=[ 'docopt>=0.6.2', ], classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: End Users/Desktop', 'License :: OSI Approved :: MIT License', 'Natural Language :: English', 'Operating System :: POSIX :: Linux', 'Operating System :: POSIX :: BSD', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Security :: Cryptography', 'Topic :: Communications', 'Topic :: Utilities', ], ) <commit_msg>Add helpful message for users trying to install with Python < 3.5<commit_after>from __future__ import print_function import sys from setuptools import setup if sys.version_info[:2] < (3, 5): print('Mikla only runs on Python 3.5 or later', file=sys.stderr) sys.exit(1) import mikla setup( name='mikla', version=mikla.__version__.strip(), url='http://dirtymonkey.co.uk/mikla', license='MIT', author=mikla.__author__.strip(), author_email='matt@dirtymonkey.co.uk', description=mikla.__doc__.strip().replace('\n', ' '), long_description=open('README.rst').read(), keywords='encryption security gnupg gpg', packages=['mikla'], include_package_data=True, entry_points={ 'console_scripts': [ 'mikla = mikla.main:main', ], }, install_requires=[ 'docopt>=0.6.2', ], classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: End Users/Desktop', 'License :: OSI Approved :: MIT License', 'Natural Language :: English', 'Operating System :: POSIX :: Linux', 'Operating System :: POSIX :: BSD', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Security :: Cryptography', 'Topic :: Communications', 'Topic :: Utilities', ], )
bc6d4f060b36418d7dc57ee4068e705b9bd28678
setup.py
setup.py
# -*- coding: utf-8 -*- from distutils.core import setup from setuptools import find_packages with open('README.rst', 'rb') as readme: readme_text = readme.read().decode('utf-8') setup( name='django-bootstrap-pagination', version='1.6.0', keywords="django bootstrap pagination templatetag", author=u'Jason McClellan', author_email='jason@jasonmccllelan.net', packages=find_packages(), url='https://github.com/jmcclell/django-bootstrap-pagination', license='MIT licence, see LICENCE', description='Render Django Page objects as Bootstrap 3.x Pagination compatible HTML', long_description=readme_text, zip_safe=False, include_package_data=True, classifiers=[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Framework :: Django", "Framework :: Django :: 1.4", "Framework :: Django :: 1.5", "Framework :: Django :: 1.6", "Framework :: Django :: 1.7", "Framework :: Django :: 1.8", "Framework :: Django :: 1.9", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", ] )
# -*- coding: utf-8 -*- from distutils.core import setup from setuptools import find_packages with open('README.rst', 'rb') as readme: readme_text = readme.read().decode('utf-8') setup( name='django-bootstrap-pagination', version='1.6.1', keywords="django bootstrap pagination templatetag", author=u'Jason McClellan', author_email='jason@jasonmccllelan.net', packages=find_packages(), url='https://github.com/jmcclell/django-bootstrap-pagination', license='MIT licence, see LICENCE', description='Render Django Page objects as Bootstrap 3.x Pagination compatible HTML', long_description=readme_text, zip_safe=False, include_package_data=True, classifiers=[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Framework :: Django", "Framework :: Django :: 1.4", "Framework :: Django :: 1.5", "Framework :: Django :: 1.6", "Framework :: Django :: 1.7", "Framework :: Django :: 1.8", "Framework :: Django :: 1.9", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", ] )
Prepare for 1.6.1 on pypi
Prepare for 1.6.1 on pypi
Python
mit
jmcclell/django-bootstrap-pagination,jmcclell/django-bootstrap-pagination
# -*- coding: utf-8 -*- from distutils.core import setup from setuptools import find_packages with open('README.rst', 'rb') as readme: readme_text = readme.read().decode('utf-8') setup( name='django-bootstrap-pagination', version='1.6.0', keywords="django bootstrap pagination templatetag", author=u'Jason McClellan', author_email='jason@jasonmccllelan.net', packages=find_packages(), url='https://github.com/jmcclell/django-bootstrap-pagination', license='MIT licence, see LICENCE', description='Render Django Page objects as Bootstrap 3.x Pagination compatible HTML', long_description=readme_text, zip_safe=False, include_package_data=True, classifiers=[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Framework :: Django", "Framework :: Django :: 1.4", "Framework :: Django :: 1.5", "Framework :: Django :: 1.6", "Framework :: Django :: 1.7", "Framework :: Django :: 1.8", "Framework :: Django :: 1.9", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", ] ) Prepare for 1.6.1 on pypi
# -*- coding: utf-8 -*- from distutils.core import setup from setuptools import find_packages with open('README.rst', 'rb') as readme: readme_text = readme.read().decode('utf-8') setup( name='django-bootstrap-pagination', version='1.6.1', keywords="django bootstrap pagination templatetag", author=u'Jason McClellan', author_email='jason@jasonmccllelan.net', packages=find_packages(), url='https://github.com/jmcclell/django-bootstrap-pagination', license='MIT licence, see LICENCE', description='Render Django Page objects as Bootstrap 3.x Pagination compatible HTML', long_description=readme_text, zip_safe=False, include_package_data=True, classifiers=[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Framework :: Django", "Framework :: Django :: 1.4", "Framework :: Django :: 1.5", "Framework :: Django :: 1.6", "Framework :: Django :: 1.7", "Framework :: Django :: 1.8", "Framework :: Django :: 1.9", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", ] )
<commit_before># -*- coding: utf-8 -*- from distutils.core import setup from setuptools import find_packages with open('README.rst', 'rb') as readme: readme_text = readme.read().decode('utf-8') setup( name='django-bootstrap-pagination', version='1.6.0', keywords="django bootstrap pagination templatetag", author=u'Jason McClellan', author_email='jason@jasonmccllelan.net', packages=find_packages(), url='https://github.com/jmcclell/django-bootstrap-pagination', license='MIT licence, see LICENCE', description='Render Django Page objects as Bootstrap 3.x Pagination compatible HTML', long_description=readme_text, zip_safe=False, include_package_data=True, classifiers=[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Framework :: Django", "Framework :: Django :: 1.4", "Framework :: Django :: 1.5", "Framework :: Django :: 1.6", "Framework :: Django :: 1.7", "Framework :: Django :: 1.8", "Framework :: Django :: 1.9", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", ] ) <commit_msg>Prepare for 1.6.1 on pypi<commit_after>
# -*- coding: utf-8 -*- from distutils.core import setup from setuptools import find_packages with open('README.rst', 'rb') as readme: readme_text = readme.read().decode('utf-8') setup( name='django-bootstrap-pagination', version='1.6.1', keywords="django bootstrap pagination templatetag", author=u'Jason McClellan', author_email='jason@jasonmccllelan.net', packages=find_packages(), url='https://github.com/jmcclell/django-bootstrap-pagination', license='MIT licence, see LICENCE', description='Render Django Page objects as Bootstrap 3.x Pagination compatible HTML', long_description=readme_text, zip_safe=False, include_package_data=True, classifiers=[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Framework :: Django", "Framework :: Django :: 1.4", "Framework :: Django :: 1.5", "Framework :: Django :: 1.6", "Framework :: Django :: 1.7", "Framework :: Django :: 1.8", "Framework :: Django :: 1.9", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", ] )
# -*- coding: utf-8 -*- from distutils.core import setup from setuptools import find_packages with open('README.rst', 'rb') as readme: readme_text = readme.read().decode('utf-8') setup( name='django-bootstrap-pagination', version='1.6.0', keywords="django bootstrap pagination templatetag", author=u'Jason McClellan', author_email='jason@jasonmccllelan.net', packages=find_packages(), url='https://github.com/jmcclell/django-bootstrap-pagination', license='MIT licence, see LICENCE', description='Render Django Page objects as Bootstrap 3.x Pagination compatible HTML', long_description=readme_text, zip_safe=False, include_package_data=True, classifiers=[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Framework :: Django", "Framework :: Django :: 1.4", "Framework :: Django :: 1.5", "Framework :: Django :: 1.6", "Framework :: Django :: 1.7", "Framework :: Django :: 1.8", "Framework :: Django :: 1.9", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", ] ) Prepare for 1.6.1 on pypi# -*- coding: utf-8 -*- from distutils.core import setup from setuptools import find_packages with open('README.rst', 'rb') as readme: readme_text = readme.read().decode('utf-8') setup( name='django-bootstrap-pagination', version='1.6.1', keywords="django bootstrap pagination templatetag", author=u'Jason McClellan', author_email='jason@jasonmccllelan.net', packages=find_packages(), url='https://github.com/jmcclell/django-bootstrap-pagination', license='MIT licence, see LICENCE', description='Render Django Page objects as Bootstrap 3.x Pagination compatible HTML', long_description=readme_text, zip_safe=False, include_package_data=True, classifiers=[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Framework :: Django", "Framework :: Django :: 1.4", "Framework :: Django :: 1.5", "Framework :: Django :: 1.6", "Framework :: Django :: 1.7", "Framework :: Django :: 1.8", "Framework :: Django :: 1.9", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", ] )
<commit_before># -*- coding: utf-8 -*- from distutils.core import setup from setuptools import find_packages with open('README.rst', 'rb') as readme: readme_text = readme.read().decode('utf-8') setup( name='django-bootstrap-pagination', version='1.6.0', keywords="django bootstrap pagination templatetag", author=u'Jason McClellan', author_email='jason@jasonmccllelan.net', packages=find_packages(), url='https://github.com/jmcclell/django-bootstrap-pagination', license='MIT licence, see LICENCE', description='Render Django Page objects as Bootstrap 3.x Pagination compatible HTML', long_description=readme_text, zip_safe=False, include_package_data=True, classifiers=[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Framework :: Django", "Framework :: Django :: 1.4", "Framework :: Django :: 1.5", "Framework :: Django :: 1.6", "Framework :: Django :: 1.7", "Framework :: Django :: 1.8", "Framework :: Django :: 1.9", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", ] ) <commit_msg>Prepare for 1.6.1 on pypi<commit_after># -*- coding: utf-8 -*- from distutils.core import setup from setuptools import find_packages with open('README.rst', 'rb') as readme: readme_text = readme.read().decode('utf-8') setup( name='django-bootstrap-pagination', version='1.6.1', keywords="django bootstrap pagination templatetag", author=u'Jason McClellan', author_email='jason@jasonmccllelan.net', packages=find_packages(), url='https://github.com/jmcclell/django-bootstrap-pagination', license='MIT licence, see LICENCE', description='Render Django Page objects as Bootstrap 3.x Pagination compatible HTML', long_description=readme_text, zip_safe=False, include_package_data=True, classifiers=[ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Framework :: Django", "Framework :: Django :: 1.4", "Framework :: Django :: 1.5", "Framework :: Django :: 1.6", "Framework :: Django :: 1.7", "Framework :: Django :: 1.8", "Framework :: Django :: 1.9", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", ] )
413d9304b5ff0a45e70512dadc4527843eee7b68
setup.py
setup.py
import sys from setuptools import setup if sys.version_info[0] == 2: base_dir = 'python2' elif sys.version_info[0] == 3: base_dir = 'python3' setup( name='raygun4py', version='2.0.0', packages=['raygun4py'], package_dir= { "raygun4py": base_dir + "/raygun4py" }, license='LICENSE', url='http://raygun.io', author='Mindscape', author_email='contact@mindscape.co.nz', long_description=open('README.rst').read(), install_requires=[ "jsonpickle == 0.7.0" ], )
import sys from setuptools import setup if sys.version_info[0] == 2: base_dir = 'python2' elif sys.version_info[0] == 3: base_dir = 'python3' setup( name='raygun4py', version='2.0.0', packages=['raygun4py'], package_dir= { "raygun4py": base_dir + "/raygun4py" }, license='LICENSE', url='http://raygun.io', author='Mindscape', author_email='contact@mindscape.co.nz', long_description=open('README.rst').read(), install_requires=[ "jsonpickle == 0.7.0" ],classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.1', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Communications', ], )
Add classifiers with py ver info
Add classifiers with py ver info
Python
mit
ferringb/raygun4py,MindscapeHQ/raygun4py,Osmose/raygun4py
import sys from setuptools import setup if sys.version_info[0] == 2: base_dir = 'python2' elif sys.version_info[0] == 3: base_dir = 'python3' setup( name='raygun4py', version='2.0.0', packages=['raygun4py'], package_dir= { "raygun4py": base_dir + "/raygun4py" }, license='LICENSE', url='http://raygun.io', author='Mindscape', author_email='contact@mindscape.co.nz', long_description=open('README.rst').read(), install_requires=[ "jsonpickle == 0.7.0" ], ) Add classifiers with py ver info
import sys from setuptools import setup if sys.version_info[0] == 2: base_dir = 'python2' elif sys.version_info[0] == 3: base_dir = 'python3' setup( name='raygun4py', version='2.0.0', packages=['raygun4py'], package_dir= { "raygun4py": base_dir + "/raygun4py" }, license='LICENSE', url='http://raygun.io', author='Mindscape', author_email='contact@mindscape.co.nz', long_description=open('README.rst').read(), install_requires=[ "jsonpickle == 0.7.0" ],classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.1', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Communications', ], )
<commit_before>import sys from setuptools import setup if sys.version_info[0] == 2: base_dir = 'python2' elif sys.version_info[0] == 3: base_dir = 'python3' setup( name='raygun4py', version='2.0.0', packages=['raygun4py'], package_dir= { "raygun4py": base_dir + "/raygun4py" }, license='LICENSE', url='http://raygun.io', author='Mindscape', author_email='contact@mindscape.co.nz', long_description=open('README.rst').read(), install_requires=[ "jsonpickle == 0.7.0" ], ) <commit_msg>Add classifiers with py ver info<commit_after>
import sys from setuptools import setup if sys.version_info[0] == 2: base_dir = 'python2' elif sys.version_info[0] == 3: base_dir = 'python3' setup( name='raygun4py', version='2.0.0', packages=['raygun4py'], package_dir= { "raygun4py": base_dir + "/raygun4py" }, license='LICENSE', url='http://raygun.io', author='Mindscape', author_email='contact@mindscape.co.nz', long_description=open('README.rst').read(), install_requires=[ "jsonpickle == 0.7.0" ],classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.1', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Communications', ], )
import sys from setuptools import setup if sys.version_info[0] == 2: base_dir = 'python2' elif sys.version_info[0] == 3: base_dir = 'python3' setup( name='raygun4py', version='2.0.0', packages=['raygun4py'], package_dir= { "raygun4py": base_dir + "/raygun4py" }, license='LICENSE', url='http://raygun.io', author='Mindscape', author_email='contact@mindscape.co.nz', long_description=open('README.rst').read(), install_requires=[ "jsonpickle == 0.7.0" ], ) Add classifiers with py ver infoimport sys from setuptools import setup if sys.version_info[0] == 2: base_dir = 'python2' elif sys.version_info[0] == 3: base_dir = 'python3' setup( name='raygun4py', version='2.0.0', packages=['raygun4py'], package_dir= { "raygun4py": base_dir + "/raygun4py" }, license='LICENSE', url='http://raygun.io', author='Mindscape', author_email='contact@mindscape.co.nz', long_description=open('README.rst').read(), install_requires=[ "jsonpickle == 0.7.0" ],classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.1', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Communications', ], )
<commit_before>import sys from setuptools import setup if sys.version_info[0] == 2: base_dir = 'python2' elif sys.version_info[0] == 3: base_dir = 'python3' setup( name='raygun4py', version='2.0.0', packages=['raygun4py'], package_dir= { "raygun4py": base_dir + "/raygun4py" }, license='LICENSE', url='http://raygun.io', author='Mindscape', author_email='contact@mindscape.co.nz', long_description=open('README.rst').read(), install_requires=[ "jsonpickle == 0.7.0" ], ) <commit_msg>Add classifiers with py ver info<commit_after>import sys from setuptools import setup if sys.version_info[0] == 2: base_dir = 'python2' elif sys.version_info[0] == 3: base_dir = 'python3' setup( name='raygun4py', version='2.0.0', packages=['raygun4py'], package_dir= { "raygun4py": base_dir + "/raygun4py" }, license='LICENSE', url='http://raygun.io', author='Mindscape', author_email='contact@mindscape.co.nz', long_description=open('README.rst').read(), install_requires=[ "jsonpickle == 0.7.0" ],classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.1', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Communications', ], )
b1bfb600240a006eed0cfce19d3fc87b3c72669f
setup.py
setup.py
#!/usr/bin/env python from distutils.core import setup setup( name = 'jargparse', packages = ['jargparse'], # this must be the same as the name above version = '0.0.3', description = 'A tiny super-dumb python module just because I like to see the usage info on stdout on an error', author = 'Justin Clark-Casey', author_email = 'justincc@justincc.org', url = 'https://github.com/justinccdev/jargparse', # use the URL to the github repo keywords = ['logging'], # arbitrary keywords )
#!/usr/bin/env python from distutils.core import setup setup( name = 'jargparse', packages = ['jargparse'], # this must be the same as the name above version = '0.0.4', description = 'A tiny super-dumb module just because I like to see the usage info on stdout on an error. jargparse.ArgParser just wraps argparse.ArgParser', author = 'Justin Clark-Casey', author_email = 'justincc@justincc.org', url = 'https://github.com/justinccdev/jargparse', # use the URL to the github repo keywords = ['logging'], # arbitrary keywords )
Add a bit more to description
Add a bit more to description
Python
apache-2.0
justinccdev/jargparse
#!/usr/bin/env python from distutils.core import setup setup( name = 'jargparse', packages = ['jargparse'], # this must be the same as the name above version = '0.0.3', description = 'A tiny super-dumb python module just because I like to see the usage info on stdout on an error', author = 'Justin Clark-Casey', author_email = 'justincc@justincc.org', url = 'https://github.com/justinccdev/jargparse', # use the URL to the github repo keywords = ['logging'], # arbitrary keywords ) Add a bit more to description
#!/usr/bin/env python from distutils.core import setup setup( name = 'jargparse', packages = ['jargparse'], # this must be the same as the name above version = '0.0.4', description = 'A tiny super-dumb module just because I like to see the usage info on stdout on an error. jargparse.ArgParser just wraps argparse.ArgParser', author = 'Justin Clark-Casey', author_email = 'justincc@justincc.org', url = 'https://github.com/justinccdev/jargparse', # use the URL to the github repo keywords = ['logging'], # arbitrary keywords )
<commit_before>#!/usr/bin/env python from distutils.core import setup setup( name = 'jargparse', packages = ['jargparse'], # this must be the same as the name above version = '0.0.3', description = 'A tiny super-dumb python module just because I like to see the usage info on stdout on an error', author = 'Justin Clark-Casey', author_email = 'justincc@justincc.org', url = 'https://github.com/justinccdev/jargparse', # use the URL to the github repo keywords = ['logging'], # arbitrary keywords ) <commit_msg>Add a bit more to description<commit_after>
#!/usr/bin/env python from distutils.core import setup setup( name = 'jargparse', packages = ['jargparse'], # this must be the same as the name above version = '0.0.4', description = 'A tiny super-dumb module just because I like to see the usage info on stdout on an error. jargparse.ArgParser just wraps argparse.ArgParser', author = 'Justin Clark-Casey', author_email = 'justincc@justincc.org', url = 'https://github.com/justinccdev/jargparse', # use the URL to the github repo keywords = ['logging'], # arbitrary keywords )
#!/usr/bin/env python from distutils.core import setup setup( name = 'jargparse', packages = ['jargparse'], # this must be the same as the name above version = '0.0.3', description = 'A tiny super-dumb python module just because I like to see the usage info on stdout on an error', author = 'Justin Clark-Casey', author_email = 'justincc@justincc.org', url = 'https://github.com/justinccdev/jargparse', # use the URL to the github repo keywords = ['logging'], # arbitrary keywords ) Add a bit more to description#!/usr/bin/env python from distutils.core import setup setup( name = 'jargparse', packages = ['jargparse'], # this must be the same as the name above version = '0.0.4', description = 'A tiny super-dumb module just because I like to see the usage info on stdout on an error. jargparse.ArgParser just wraps argparse.ArgParser', author = 'Justin Clark-Casey', author_email = 'justincc@justincc.org', url = 'https://github.com/justinccdev/jargparse', # use the URL to the github repo keywords = ['logging'], # arbitrary keywords )
<commit_before>#!/usr/bin/env python from distutils.core import setup setup( name = 'jargparse', packages = ['jargparse'], # this must be the same as the name above version = '0.0.3', description = 'A tiny super-dumb python module just because I like to see the usage info on stdout on an error', author = 'Justin Clark-Casey', author_email = 'justincc@justincc.org', url = 'https://github.com/justinccdev/jargparse', # use the URL to the github repo keywords = ['logging'], # arbitrary keywords ) <commit_msg>Add a bit more to description<commit_after>#!/usr/bin/env python from distutils.core import setup setup( name = 'jargparse', packages = ['jargparse'], # this must be the same as the name above version = '0.0.4', description = 'A tiny super-dumb module just because I like to see the usage info on stdout on an error. jargparse.ArgParser just wraps argparse.ArgParser', author = 'Justin Clark-Casey', author_email = 'justincc@justincc.org', url = 'https://github.com/justinccdev/jargparse', # use the URL to the github repo keywords = ['logging'], # arbitrary keywords )
1fea3970bd6979834f17e1a01bd204f60e4361ba
setup.py
setup.py
#!/usr/bin/env python import setuptools import os setuptools.setup( name='endpoints-proto-datastore', version='0.9.0', description='Google Cloud Endpoints Proto Datastore Library', long_description=open(os.path.join(os.path.dirname(__file__), 'README.md')).read(), url='https://github.com/GoogleCloudPlatform/endpoints-proto-datastore', license='Apache', author='Danny Hermes', author_email='daniel.j.hermes@gmail.com', classifiers=[ 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache License', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python' ], packages=setuptools.find_packages(exclude=['examples*', 'docs*']), )
#!/usr/bin/env python import setuptools import os setuptools.setup( name='endpoints-proto-datastore', version='0.9.0', description='Google Cloud Endpoints Proto Datastore Library', long_description=open(os.path.join(os.path.dirname(__file__), 'README.md')).read(), url='https://github.com/GoogleCloudPlatform/endpoints-proto-datastore', license='Apache', author='Danny Hermes', author_email='daniel.j.hermes@gmail.com', classifiers=[ 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache License', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python' ], packages=setuptools.find_packages(exclude=['examples*', 'docs*']), )
Add trailing newline since PyCharm stripped it
Add trailing newline since PyCharm stripped it
Python
apache-2.0
jbergant/endpoints-proto-datastore,dhermes/endpoints-proto-datastore,GoogleCloudPlatform/endpoints-proto-datastore,mnieper/endpoints-proto-datastore,maxandron/endpoints-proto-datastore
#!/usr/bin/env python import setuptools import os setuptools.setup( name='endpoints-proto-datastore', version='0.9.0', description='Google Cloud Endpoints Proto Datastore Library', long_description=open(os.path.join(os.path.dirname(__file__), 'README.md')).read(), url='https://github.com/GoogleCloudPlatform/endpoints-proto-datastore', license='Apache', author='Danny Hermes', author_email='daniel.j.hermes@gmail.com', classifiers=[ 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache License', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python' ], packages=setuptools.find_packages(exclude=['examples*', 'docs*']), )Add trailing newline since PyCharm stripped it
#!/usr/bin/env python import setuptools import os setuptools.setup( name='endpoints-proto-datastore', version='0.9.0', description='Google Cloud Endpoints Proto Datastore Library', long_description=open(os.path.join(os.path.dirname(__file__), 'README.md')).read(), url='https://github.com/GoogleCloudPlatform/endpoints-proto-datastore', license='Apache', author='Danny Hermes', author_email='daniel.j.hermes@gmail.com', classifiers=[ 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache License', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python' ], packages=setuptools.find_packages(exclude=['examples*', 'docs*']), )
<commit_before>#!/usr/bin/env python import setuptools import os setuptools.setup( name='endpoints-proto-datastore', version='0.9.0', description='Google Cloud Endpoints Proto Datastore Library', long_description=open(os.path.join(os.path.dirname(__file__), 'README.md')).read(), url='https://github.com/GoogleCloudPlatform/endpoints-proto-datastore', license='Apache', author='Danny Hermes', author_email='daniel.j.hermes@gmail.com', classifiers=[ 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache License', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python' ], packages=setuptools.find_packages(exclude=['examples*', 'docs*']), )<commit_msg>Add trailing newline since PyCharm stripped it<commit_after>
#!/usr/bin/env python import setuptools import os setuptools.setup( name='endpoints-proto-datastore', version='0.9.0', description='Google Cloud Endpoints Proto Datastore Library', long_description=open(os.path.join(os.path.dirname(__file__), 'README.md')).read(), url='https://github.com/GoogleCloudPlatform/endpoints-proto-datastore', license='Apache', author='Danny Hermes', author_email='daniel.j.hermes@gmail.com', classifiers=[ 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache License', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python' ], packages=setuptools.find_packages(exclude=['examples*', 'docs*']), )
#!/usr/bin/env python import setuptools import os setuptools.setup( name='endpoints-proto-datastore', version='0.9.0', description='Google Cloud Endpoints Proto Datastore Library', long_description=open(os.path.join(os.path.dirname(__file__), 'README.md')).read(), url='https://github.com/GoogleCloudPlatform/endpoints-proto-datastore', license='Apache', author='Danny Hermes', author_email='daniel.j.hermes@gmail.com', classifiers=[ 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache License', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python' ], packages=setuptools.find_packages(exclude=['examples*', 'docs*']), )Add trailing newline since PyCharm stripped it#!/usr/bin/env python import setuptools import os setuptools.setup( name='endpoints-proto-datastore', version='0.9.0', description='Google Cloud Endpoints Proto Datastore Library', long_description=open(os.path.join(os.path.dirname(__file__), 'README.md')).read(), url='https://github.com/GoogleCloudPlatform/endpoints-proto-datastore', license='Apache', author='Danny Hermes', author_email='daniel.j.hermes@gmail.com', classifiers=[ 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache License', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python' ], packages=setuptools.find_packages(exclude=['examples*', 'docs*']), )
<commit_before>#!/usr/bin/env python import setuptools import os setuptools.setup( name='endpoints-proto-datastore', version='0.9.0', description='Google Cloud Endpoints Proto Datastore Library', long_description=open(os.path.join(os.path.dirname(__file__), 'README.md')).read(), url='https://github.com/GoogleCloudPlatform/endpoints-proto-datastore', license='Apache', author='Danny Hermes', author_email='daniel.j.hermes@gmail.com', classifiers=[ 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache License', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python' ], packages=setuptools.find_packages(exclude=['examples*', 'docs*']), )<commit_msg>Add trailing newline since PyCharm stripped it<commit_after>#!/usr/bin/env python import setuptools import os setuptools.setup( name='endpoints-proto-datastore', version='0.9.0', description='Google Cloud Endpoints Proto Datastore Library', long_description=open(os.path.join(os.path.dirname(__file__), 'README.md')).read(), url='https://github.com/GoogleCloudPlatform/endpoints-proto-datastore', license='Apache', author='Danny Hermes', author_email='daniel.j.hermes@gmail.com', classifiers=[ 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache License', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python' ], packages=setuptools.find_packages(exclude=['examples*', 'docs*']), )
6d702526d04fab65ee62bcec7db668d65f9de97f
setup.py
setup.py
from setuptools import setup import jasinja, sys requires = ['Jinja2'] if sys.version_info < (2, 6): requires += ['simplejson'] setup( name='jasinja', version=jasinja.__version__, url='http://bitbucket.org/djc/jasinja', license='BSD', author='Dirkjan Ochtman', author_email='dirkjan@ochtman.nl', description='A JavaScript code generator for Jinja templates', classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules' ], packages=['jasinja', 'jasinja.tests'], package_data={ 'jasinja': ['*.js'] }, install_requires=requires, test_suite='jasinja.tests.run.suite', test_requires=['python-spidermonkey'], entry_points={ 'console_scripts': ['jasinja-compile = jasinja.compile:main'], }, )
from setuptools import setup import jasinja, sys requires = ['Jinja2'] if sys.version_info < (2, 6): requires += ['simplejson'] setup( name='jasinja', version=jasinja.__version__, url='http://bitbucket.org/djc/jasinja', license='BSD', author='Dirkjan Ochtman', author_email='dirkjan@ochtman.nl', description='A JavaScript code generator for Jinja templates', long_description=open('README.txt').read(), classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules' ], packages=['jasinja', 'jasinja.tests'], package_data={ 'jasinja': ['*.js'] }, install_requires=requires, test_suite='jasinja.tests.run.suite', test_requires=['python-spidermonkey'], entry_points={ 'console_scripts': ['jasinja-compile = jasinja.compile:main'], }, )
Add long description (from README).
Add long description (from README).
Python
bsd-3-clause
djc/jasinja,djc/jasinja
from setuptools import setup import jasinja, sys requires = ['Jinja2'] if sys.version_info < (2, 6): requires += ['simplejson'] setup( name='jasinja', version=jasinja.__version__, url='http://bitbucket.org/djc/jasinja', license='BSD', author='Dirkjan Ochtman', author_email='dirkjan@ochtman.nl', description='A JavaScript code generator for Jinja templates', classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules' ], packages=['jasinja', 'jasinja.tests'], package_data={ 'jasinja': ['*.js'] }, install_requires=requires, test_suite='jasinja.tests.run.suite', test_requires=['python-spidermonkey'], entry_points={ 'console_scripts': ['jasinja-compile = jasinja.compile:main'], }, ) Add long description (from README).
from setuptools import setup import jasinja, sys requires = ['Jinja2'] if sys.version_info < (2, 6): requires += ['simplejson'] setup( name='jasinja', version=jasinja.__version__, url='http://bitbucket.org/djc/jasinja', license='BSD', author='Dirkjan Ochtman', author_email='dirkjan@ochtman.nl', description='A JavaScript code generator for Jinja templates', long_description=open('README.txt').read(), classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules' ], packages=['jasinja', 'jasinja.tests'], package_data={ 'jasinja': ['*.js'] }, install_requires=requires, test_suite='jasinja.tests.run.suite', test_requires=['python-spidermonkey'], entry_points={ 'console_scripts': ['jasinja-compile = jasinja.compile:main'], }, )
<commit_before>from setuptools import setup import jasinja, sys requires = ['Jinja2'] if sys.version_info < (2, 6): requires += ['simplejson'] setup( name='jasinja', version=jasinja.__version__, url='http://bitbucket.org/djc/jasinja', license='BSD', author='Dirkjan Ochtman', author_email='dirkjan@ochtman.nl', description='A JavaScript code generator for Jinja templates', classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules' ], packages=['jasinja', 'jasinja.tests'], package_data={ 'jasinja': ['*.js'] }, install_requires=requires, test_suite='jasinja.tests.run.suite', test_requires=['python-spidermonkey'], entry_points={ 'console_scripts': ['jasinja-compile = jasinja.compile:main'], }, ) <commit_msg>Add long description (from README).<commit_after>
from setuptools import setup import jasinja, sys requires = ['Jinja2'] if sys.version_info < (2, 6): requires += ['simplejson'] setup( name='jasinja', version=jasinja.__version__, url='http://bitbucket.org/djc/jasinja', license='BSD', author='Dirkjan Ochtman', author_email='dirkjan@ochtman.nl', description='A JavaScript code generator for Jinja templates', long_description=open('README.txt').read(), classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules' ], packages=['jasinja', 'jasinja.tests'], package_data={ 'jasinja': ['*.js'] }, install_requires=requires, test_suite='jasinja.tests.run.suite', test_requires=['python-spidermonkey'], entry_points={ 'console_scripts': ['jasinja-compile = jasinja.compile:main'], }, )
from setuptools import setup import jasinja, sys requires = ['Jinja2'] if sys.version_info < (2, 6): requires += ['simplejson'] setup( name='jasinja', version=jasinja.__version__, url='http://bitbucket.org/djc/jasinja', license='BSD', author='Dirkjan Ochtman', author_email='dirkjan@ochtman.nl', description='A JavaScript code generator for Jinja templates', classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules' ], packages=['jasinja', 'jasinja.tests'], package_data={ 'jasinja': ['*.js'] }, install_requires=requires, test_suite='jasinja.tests.run.suite', test_requires=['python-spidermonkey'], entry_points={ 'console_scripts': ['jasinja-compile = jasinja.compile:main'], }, ) Add long description (from README).from setuptools import setup import jasinja, sys requires = ['Jinja2'] if sys.version_info < (2, 6): requires += ['simplejson'] setup( name='jasinja', version=jasinja.__version__, url='http://bitbucket.org/djc/jasinja', license='BSD', author='Dirkjan Ochtman', author_email='dirkjan@ochtman.nl', description='A JavaScript code generator for Jinja templates', long_description=open('README.txt').read(), classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules' ], packages=['jasinja', 'jasinja.tests'], package_data={ 'jasinja': ['*.js'] }, install_requires=requires, test_suite='jasinja.tests.run.suite', test_requires=['python-spidermonkey'], entry_points={ 'console_scripts': ['jasinja-compile = jasinja.compile:main'], }, )
<commit_before>from setuptools import setup import jasinja, sys requires = ['Jinja2'] if sys.version_info < (2, 6): requires += ['simplejson'] setup( name='jasinja', version=jasinja.__version__, url='http://bitbucket.org/djc/jasinja', license='BSD', author='Dirkjan Ochtman', author_email='dirkjan@ochtman.nl', description='A JavaScript code generator for Jinja templates', classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules' ], packages=['jasinja', 'jasinja.tests'], package_data={ 'jasinja': ['*.js'] }, install_requires=requires, test_suite='jasinja.tests.run.suite', test_requires=['python-spidermonkey'], entry_points={ 'console_scripts': ['jasinja-compile = jasinja.compile:main'], }, ) <commit_msg>Add long description (from README).<commit_after>from setuptools import setup import jasinja, sys requires = ['Jinja2'] if sys.version_info < (2, 6): requires += ['simplejson'] setup( name='jasinja', version=jasinja.__version__, url='http://bitbucket.org/djc/jasinja', license='BSD', author='Dirkjan Ochtman', author_email='dirkjan@ochtman.nl', description='A JavaScript code generator for Jinja templates', long_description=open('README.txt').read(), classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules' ], packages=['jasinja', 'jasinja.tests'], package_data={ 'jasinja': ['*.js'] }, install_requires=requires, test_suite='jasinja.tests.run.suite', test_requires=['python-spidermonkey'], entry_points={ 'console_scripts': ['jasinja-compile = jasinja.compile:main'], }, )
1fe7b9c3c9a3764a1e209b2699ef51b84c87e897
setup.py
setup.py
from distutils.core import setup import os setup( name='python-jambel', version='0.1', py_module=['jambel'], url='http://github.com/jambit/python-jambel', license='UNKNOWN', author='Sebastian Rahlf', author_email='sebastian.rahlf@jambit.com', description="Interface to jambit's project traffic lights.", long_description=open(os.path.join(os.path.dirname(__file__), 'README.txt')).read(), classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'License :: Other/Proprietary License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.1', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ] )
from distutils.core import setup import os setup( name='python-jambel', version='0.1', py_module=['jambel'], url='http://github.com/jambit/python-jambel', license='UNKNOWN', author='Sebastian Rahlf', author_email='sebastian.rahlf@jambit.com', description="Interface to jambit's project traffic lights.", long_description=open(os.path.join(os.path.dirname(__file__), 'README.txt')).read(), test_requires=['pytest'], entry_points={ 'console_scripts': [ 'jambel = jambel:main', ] }, classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'License :: Other/Proprietary License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.1', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ] )
Add console script and test requirements.
Add console script and test requirements.
Python
mit
redtoad/python-jambel,jambit/python-jambel
from distutils.core import setup import os setup( name='python-jambel', version='0.1', py_module=['jambel'], url='http://github.com/jambit/python-jambel', license='UNKNOWN', author='Sebastian Rahlf', author_email='sebastian.rahlf@jambit.com', description="Interface to jambit's project traffic lights.", long_description=open(os.path.join(os.path.dirname(__file__), 'README.txt')).read(), classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'License :: Other/Proprietary License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.1', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ] ) Add console script and test requirements.
from distutils.core import setup import os setup( name='python-jambel', version='0.1', py_module=['jambel'], url='http://github.com/jambit/python-jambel', license='UNKNOWN', author='Sebastian Rahlf', author_email='sebastian.rahlf@jambit.com', description="Interface to jambit's project traffic lights.", long_description=open(os.path.join(os.path.dirname(__file__), 'README.txt')).read(), test_requires=['pytest'], entry_points={ 'console_scripts': [ 'jambel = jambel:main', ] }, classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'License :: Other/Proprietary License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.1', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ] )
<commit_before>from distutils.core import setup import os setup( name='python-jambel', version='0.1', py_module=['jambel'], url='http://github.com/jambit/python-jambel', license='UNKNOWN', author='Sebastian Rahlf', author_email='sebastian.rahlf@jambit.com', description="Interface to jambit's project traffic lights.", long_description=open(os.path.join(os.path.dirname(__file__), 'README.txt')).read(), classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'License :: Other/Proprietary License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.1', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ] ) <commit_msg>Add console script and test requirements.<commit_after>
from distutils.core import setup import os setup( name='python-jambel', version='0.1', py_module=['jambel'], url='http://github.com/jambit/python-jambel', license='UNKNOWN', author='Sebastian Rahlf', author_email='sebastian.rahlf@jambit.com', description="Interface to jambit's project traffic lights.", long_description=open(os.path.join(os.path.dirname(__file__), 'README.txt')).read(), test_requires=['pytest'], entry_points={ 'console_scripts': [ 'jambel = jambel:main', ] }, classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'License :: Other/Proprietary License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.1', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ] )
from distutils.core import setup import os setup( name='python-jambel', version='0.1', py_module=['jambel'], url='http://github.com/jambit/python-jambel', license='UNKNOWN', author='Sebastian Rahlf', author_email='sebastian.rahlf@jambit.com', description="Interface to jambit's project traffic lights.", long_description=open(os.path.join(os.path.dirname(__file__), 'README.txt')).read(), classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'License :: Other/Proprietary License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.1', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ] ) Add console script and test requirements.from distutils.core import setup import os setup( name='python-jambel', version='0.1', py_module=['jambel'], url='http://github.com/jambit/python-jambel', license='UNKNOWN', author='Sebastian Rahlf', author_email='sebastian.rahlf@jambit.com', description="Interface to jambit's project traffic lights.", long_description=open(os.path.join(os.path.dirname(__file__), 'README.txt')).read(), test_requires=['pytest'], entry_points={ 'console_scripts': [ 'jambel = jambel:main', ] }, classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'License :: Other/Proprietary License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.1', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ] )
<commit_before>from distutils.core import setup import os setup( name='python-jambel', version='0.1', py_module=['jambel'], url='http://github.com/jambit/python-jambel', license='UNKNOWN', author='Sebastian Rahlf', author_email='sebastian.rahlf@jambit.com', description="Interface to jambit's project traffic lights.", long_description=open(os.path.join(os.path.dirname(__file__), 'README.txt')).read(), classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'License :: Other/Proprietary License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.1', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ] ) <commit_msg>Add console script and test requirements.<commit_after>from distutils.core import setup import os setup( name='python-jambel', version='0.1', py_module=['jambel'], url='http://github.com/jambit/python-jambel', license='UNKNOWN', author='Sebastian Rahlf', author_email='sebastian.rahlf@jambit.com', description="Interface to jambit's project traffic lights.", long_description=open(os.path.join(os.path.dirname(__file__), 'README.txt')).read(), test_requires=['pytest'], entry_points={ 'console_scripts': [ 'jambel = jambel:main', ] }, classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'License :: Other/Proprietary License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.1', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ] )
6ca66af32e522e897495a1fbe1748caec37da9bf
setup.py
setup.py
#!/usr/bin/env python # -*- coding: utf-8 -*- # # Copyright (C) 2008 Jason Davies # All rights reserved. # # This software is licensed as described in the file COPYING, which # you should have received as part of this distribution. try: from setuptools import setup except ImportError: from distutils.core import setup setup( name = 'CouchDB-FUSE', version = '0.1', description = 'CouchDB FUSE module', long_description = \ """This is a Python FUSE module for CouchDB. It allows CouchDB document attachments to be mounted on a virtual filesystem and edited directly.""", author = 'Jason Davies', author_email = 'jason@jasondavies.com', license = 'BSD', url = 'http://code.google.com/p/couchdb-fuse/', zip_safe = True, classifiers = [ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Database :: Front-Ends', ], packages = ['couchdbfuse'], entry_points = { 'console_scripts': [ 'couchmount = couchdbfuse:main', ], }, install_requires = ['CouchDB>=0.5', 'fuse-python>=0.2'], )
#!/usr/bin/env python # -*- coding: utf-8 -*- # # Copyright (C) 2008 Jason Davies # All rights reserved. # # This software is licensed as described in the file COPYING, which # you should have received as part of this distribution. try: from setuptools import setup except ImportError: from distutils.core import setup setup( name = 'CouchDB-FUSE', version = '0.1', description = 'CouchDB FUSE module', long_description = \ """This is a Python FUSE module for CouchDB. It allows CouchDB document attachments to be mounted on a virtual filesystem and edited directly.""", author = 'Jason Davies', author_email = 'jason@jasondavies.com', license = 'BSD', url = 'http://code.google.com/p/couchdb-fuse/', zip_safe = True, classifiers = [ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Database :: Front-Ends', ], packages = ['couchdbfuse'], entry_points = { 'console_scripts': [ 'couchmount = couchdbfuse:main', ], }, install_requires = ['CouchDB>=0.5'], )
Revert previous commit as fuse-python doesn't seem to play nicely with easy_install (at least on Ubuntu).
Revert previous commit as fuse-python doesn't seem to play nicely with easy_install (at least on Ubuntu).
Python
bsd-3-clause
johnko/couchdb-fuse
#!/usr/bin/env python # -*- coding: utf-8 -*- # # Copyright (C) 2008 Jason Davies # All rights reserved. # # This software is licensed as described in the file COPYING, which # you should have received as part of this distribution. try: from setuptools import setup except ImportError: from distutils.core import setup setup( name = 'CouchDB-FUSE', version = '0.1', description = 'CouchDB FUSE module', long_description = \ """This is a Python FUSE module for CouchDB. It allows CouchDB document attachments to be mounted on a virtual filesystem and edited directly.""", author = 'Jason Davies', author_email = 'jason@jasondavies.com', license = 'BSD', url = 'http://code.google.com/p/couchdb-fuse/', zip_safe = True, classifiers = [ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Database :: Front-Ends', ], packages = ['couchdbfuse'], entry_points = { 'console_scripts': [ 'couchmount = couchdbfuse:main', ], }, install_requires = ['CouchDB>=0.5', 'fuse-python>=0.2'], ) Revert previous commit as fuse-python doesn't seem to play nicely with easy_install (at least on Ubuntu).
#!/usr/bin/env python # -*- coding: utf-8 -*- # # Copyright (C) 2008 Jason Davies # All rights reserved. # # This software is licensed as described in the file COPYING, which # you should have received as part of this distribution. try: from setuptools import setup except ImportError: from distutils.core import setup setup( name = 'CouchDB-FUSE', version = '0.1', description = 'CouchDB FUSE module', long_description = \ """This is a Python FUSE module for CouchDB. It allows CouchDB document attachments to be mounted on a virtual filesystem and edited directly.""", author = 'Jason Davies', author_email = 'jason@jasondavies.com', license = 'BSD', url = 'http://code.google.com/p/couchdb-fuse/', zip_safe = True, classifiers = [ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Database :: Front-Ends', ], packages = ['couchdbfuse'], entry_points = { 'console_scripts': [ 'couchmount = couchdbfuse:main', ], }, install_requires = ['CouchDB>=0.5'], )
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- # # Copyright (C) 2008 Jason Davies # All rights reserved. # # This software is licensed as described in the file COPYING, which # you should have received as part of this distribution. try: from setuptools import setup except ImportError: from distutils.core import setup setup( name = 'CouchDB-FUSE', version = '0.1', description = 'CouchDB FUSE module', long_description = \ """This is a Python FUSE module for CouchDB. It allows CouchDB document attachments to be mounted on a virtual filesystem and edited directly.""", author = 'Jason Davies', author_email = 'jason@jasondavies.com', license = 'BSD', url = 'http://code.google.com/p/couchdb-fuse/', zip_safe = True, classifiers = [ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Database :: Front-Ends', ], packages = ['couchdbfuse'], entry_points = { 'console_scripts': [ 'couchmount = couchdbfuse:main', ], }, install_requires = ['CouchDB>=0.5', 'fuse-python>=0.2'], ) <commit_msg>Revert previous commit as fuse-python doesn't seem to play nicely with easy_install (at least on Ubuntu).<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- # # Copyright (C) 2008 Jason Davies # All rights reserved. # # This software is licensed as described in the file COPYING, which # you should have received as part of this distribution. try: from setuptools import setup except ImportError: from distutils.core import setup setup( name = 'CouchDB-FUSE', version = '0.1', description = 'CouchDB FUSE module', long_description = \ """This is a Python FUSE module for CouchDB. It allows CouchDB document attachments to be mounted on a virtual filesystem and edited directly.""", author = 'Jason Davies', author_email = 'jason@jasondavies.com', license = 'BSD', url = 'http://code.google.com/p/couchdb-fuse/', zip_safe = True, classifiers = [ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Database :: Front-Ends', ], packages = ['couchdbfuse'], entry_points = { 'console_scripts': [ 'couchmount = couchdbfuse:main', ], }, install_requires = ['CouchDB>=0.5'], )
#!/usr/bin/env python # -*- coding: utf-8 -*- # # Copyright (C) 2008 Jason Davies # All rights reserved. # # This software is licensed as described in the file COPYING, which # you should have received as part of this distribution. try: from setuptools import setup except ImportError: from distutils.core import setup setup( name = 'CouchDB-FUSE', version = '0.1', description = 'CouchDB FUSE module', long_description = \ """This is a Python FUSE module for CouchDB. It allows CouchDB document attachments to be mounted on a virtual filesystem and edited directly.""", author = 'Jason Davies', author_email = 'jason@jasondavies.com', license = 'BSD', url = 'http://code.google.com/p/couchdb-fuse/', zip_safe = True, classifiers = [ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Database :: Front-Ends', ], packages = ['couchdbfuse'], entry_points = { 'console_scripts': [ 'couchmount = couchdbfuse:main', ], }, install_requires = ['CouchDB>=0.5', 'fuse-python>=0.2'], ) Revert previous commit as fuse-python doesn't seem to play nicely with easy_install (at least on Ubuntu).#!/usr/bin/env python # -*- coding: utf-8 -*- # # Copyright (C) 2008 Jason Davies # All rights reserved. # # This software is licensed as described in the file COPYING, which # you should have received as part of this distribution. try: from setuptools import setup except ImportError: from distutils.core import setup setup( name = 'CouchDB-FUSE', version = '0.1', description = 'CouchDB FUSE module', long_description = \ """This is a Python FUSE module for CouchDB. It allows CouchDB document attachments to be mounted on a virtual filesystem and edited directly.""", author = 'Jason Davies', author_email = 'jason@jasondavies.com', license = 'BSD', url = 'http://code.google.com/p/couchdb-fuse/', zip_safe = True, classifiers = [ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Database :: Front-Ends', ], packages = ['couchdbfuse'], entry_points = { 'console_scripts': [ 'couchmount = couchdbfuse:main', ], }, install_requires = ['CouchDB>=0.5'], )
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- # # Copyright (C) 2008 Jason Davies # All rights reserved. # # This software is licensed as described in the file COPYING, which # you should have received as part of this distribution. try: from setuptools import setup except ImportError: from distutils.core import setup setup( name = 'CouchDB-FUSE', version = '0.1', description = 'CouchDB FUSE module', long_description = \ """This is a Python FUSE module for CouchDB. It allows CouchDB document attachments to be mounted on a virtual filesystem and edited directly.""", author = 'Jason Davies', author_email = 'jason@jasondavies.com', license = 'BSD', url = 'http://code.google.com/p/couchdb-fuse/', zip_safe = True, classifiers = [ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Database :: Front-Ends', ], packages = ['couchdbfuse'], entry_points = { 'console_scripts': [ 'couchmount = couchdbfuse:main', ], }, install_requires = ['CouchDB>=0.5', 'fuse-python>=0.2'], ) <commit_msg>Revert previous commit as fuse-python doesn't seem to play nicely with easy_install (at least on Ubuntu).<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- # # Copyright (C) 2008 Jason Davies # All rights reserved. # # This software is licensed as described in the file COPYING, which # you should have received as part of this distribution. try: from setuptools import setup except ImportError: from distutils.core import setup setup( name = 'CouchDB-FUSE', version = '0.1', description = 'CouchDB FUSE module', long_description = \ """This is a Python FUSE module for CouchDB. It allows CouchDB document attachments to be mounted on a virtual filesystem and edited directly.""", author = 'Jason Davies', author_email = 'jason@jasondavies.com', license = 'BSD', url = 'http://code.google.com/p/couchdb-fuse/', zip_safe = True, classifiers = [ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Database :: Front-Ends', ], packages = ['couchdbfuse'], entry_points = { 'console_scripts': [ 'couchmount = couchdbfuse:main', ], }, install_requires = ['CouchDB>=0.5'], )
8b874d83e3840e1e67e2cd81c18fe0b415130619
setup.py
setup.py
#!/usr/bin/env python try: from setuptools import setup except ImportError: from distutils.core import setup setup(name='chevron', version='0.7.1', license='MIT', description='Mustache templating language renderer', author='noah morrison', author_email='noah@morrison.ph', url='https://github.com/noahmorrison/chevron', packages=['chevron'], entry_points={ 'console_scripts': ['chevron=chevron:cli_main'] }, classifiers=[ 'Development Status :: 4 - Beta', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Text Processing :: Markup' ] )
#!/usr/bin/env python try: from setuptools import setup except ImportError: from distutils.core import setup import pypandoc readme = pypandoc.convert('README.md', 'rst') setup(name='chevron', version='0.7.1', license='MIT', description='Mustache templating language renderer', long_description=readme, author='noah morrison', author_email='noah@morrison.ph', url='https://github.com/noahmorrison/chevron', packages=['chevron'], entry_points={ 'console_scripts': ['chevron=chevron:cli_main'] }, classifiers=[ 'Development Status :: 4 - Beta', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Text Processing :: Markup' ] )
Convert README.md to rst for pypi using pandoc.
Convert README.md to rst for pypi using pandoc. Closes #11 Thanks for the tips/help Marc :)
Python
mit
noahmorrison/chevron,noahmorrison/chevron
#!/usr/bin/env python try: from setuptools import setup except ImportError: from distutils.core import setup setup(name='chevron', version='0.7.1', license='MIT', description='Mustache templating language renderer', author='noah morrison', author_email='noah@morrison.ph', url='https://github.com/noahmorrison/chevron', packages=['chevron'], entry_points={ 'console_scripts': ['chevron=chevron:cli_main'] }, classifiers=[ 'Development Status :: 4 - Beta', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Text Processing :: Markup' ] ) Convert README.md to rst for pypi using pandoc. Closes #11 Thanks for the tips/help Marc :)
#!/usr/bin/env python try: from setuptools import setup except ImportError: from distutils.core import setup import pypandoc readme = pypandoc.convert('README.md', 'rst') setup(name='chevron', version='0.7.1', license='MIT', description='Mustache templating language renderer', long_description=readme, author='noah morrison', author_email='noah@morrison.ph', url='https://github.com/noahmorrison/chevron', packages=['chevron'], entry_points={ 'console_scripts': ['chevron=chevron:cli_main'] }, classifiers=[ 'Development Status :: 4 - Beta', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Text Processing :: Markup' ] )
<commit_before>#!/usr/bin/env python try: from setuptools import setup except ImportError: from distutils.core import setup setup(name='chevron', version='0.7.1', license='MIT', description='Mustache templating language renderer', author='noah morrison', author_email='noah@morrison.ph', url='https://github.com/noahmorrison/chevron', packages=['chevron'], entry_points={ 'console_scripts': ['chevron=chevron:cli_main'] }, classifiers=[ 'Development Status :: 4 - Beta', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Text Processing :: Markup' ] ) <commit_msg>Convert README.md to rst for pypi using pandoc. Closes #11 Thanks for the tips/help Marc :)<commit_after>
#!/usr/bin/env python try: from setuptools import setup except ImportError: from distutils.core import setup import pypandoc readme = pypandoc.convert('README.md', 'rst') setup(name='chevron', version='0.7.1', license='MIT', description='Mustache templating language renderer', long_description=readme, author='noah morrison', author_email='noah@morrison.ph', url='https://github.com/noahmorrison/chevron', packages=['chevron'], entry_points={ 'console_scripts': ['chevron=chevron:cli_main'] }, classifiers=[ 'Development Status :: 4 - Beta', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Text Processing :: Markup' ] )
#!/usr/bin/env python try: from setuptools import setup except ImportError: from distutils.core import setup setup(name='chevron', version='0.7.1', license='MIT', description='Mustache templating language renderer', author='noah morrison', author_email='noah@morrison.ph', url='https://github.com/noahmorrison/chevron', packages=['chevron'], entry_points={ 'console_scripts': ['chevron=chevron:cli_main'] }, classifiers=[ 'Development Status :: 4 - Beta', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Text Processing :: Markup' ] ) Convert README.md to rst for pypi using pandoc. Closes #11 Thanks for the tips/help Marc :)#!/usr/bin/env python try: from setuptools import setup except ImportError: from distutils.core import setup import pypandoc readme = pypandoc.convert('README.md', 'rst') setup(name='chevron', version='0.7.1', license='MIT', description='Mustache templating language renderer', long_description=readme, author='noah morrison', author_email='noah@morrison.ph', url='https://github.com/noahmorrison/chevron', packages=['chevron'], entry_points={ 'console_scripts': ['chevron=chevron:cli_main'] }, classifiers=[ 'Development Status :: 4 - Beta', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Text Processing :: Markup' ] )
<commit_before>#!/usr/bin/env python try: from setuptools import setup except ImportError: from distutils.core import setup setup(name='chevron', version='0.7.1', license='MIT', description='Mustache templating language renderer', author='noah morrison', author_email='noah@morrison.ph', url='https://github.com/noahmorrison/chevron', packages=['chevron'], entry_points={ 'console_scripts': ['chevron=chevron:cli_main'] }, classifiers=[ 'Development Status :: 4 - Beta', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Text Processing :: Markup' ] ) <commit_msg>Convert README.md to rst for pypi using pandoc. Closes #11 Thanks for the tips/help Marc :)<commit_after>#!/usr/bin/env python try: from setuptools import setup except ImportError: from distutils.core import setup import pypandoc readme = pypandoc.convert('README.md', 'rst') setup(name='chevron', version='0.7.1', license='MIT', description='Mustache templating language renderer', long_description=readme, author='noah morrison', author_email='noah@morrison.ph', url='https://github.com/noahmorrison/chevron', packages=['chevron'], entry_points={ 'console_scripts': ['chevron=chevron:cli_main'] }, classifiers=[ 'Development Status :: 4 - Beta', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Topic :: Text Processing :: Markup' ] )
c65d11f82b6d33d4940cdfd7b4d6b81e083c6e34
setup.py
setup.py
from distutils.core import setup setup( name='django_autologin', version='0.1', packages=['django_autologin', 'django_autologin.templatetags'], install_requires=['django>=1.0'], description='Token generator and processor to provide automatic login links for users' )
from distutils.core import setup setup( name='django_autologin', version='0.1', packages=['django_autologin', 'django_autologin.templatetags'], install_requires=['django>=1.0'], description='Token generator and processor to provide automatic login links for users' )
Use 4 spaces for indentation.
Use 4 spaces for indentation.
Python
bsd-3-clause
playfire/django-autologin
from distutils.core import setup setup( name='django_autologin', version='0.1', packages=['django_autologin', 'django_autologin.templatetags'], install_requires=['django>=1.0'], description='Token generator and processor to provide automatic login links for users' ) Use 4 spaces for indentation.
from distutils.core import setup setup( name='django_autologin', version='0.1', packages=['django_autologin', 'django_autologin.templatetags'], install_requires=['django>=1.0'], description='Token generator and processor to provide automatic login links for users' )
<commit_before>from distutils.core import setup setup( name='django_autologin', version='0.1', packages=['django_autologin', 'django_autologin.templatetags'], install_requires=['django>=1.0'], description='Token generator and processor to provide automatic login links for users' ) <commit_msg>Use 4 spaces for indentation.<commit_after>
from distutils.core import setup setup( name='django_autologin', version='0.1', packages=['django_autologin', 'django_autologin.templatetags'], install_requires=['django>=1.0'], description='Token generator and processor to provide automatic login links for users' )
from distutils.core import setup setup( name='django_autologin', version='0.1', packages=['django_autologin', 'django_autologin.templatetags'], install_requires=['django>=1.0'], description='Token generator and processor to provide automatic login links for users' ) Use 4 spaces for indentation.from distutils.core import setup setup( name='django_autologin', version='0.1', packages=['django_autologin', 'django_autologin.templatetags'], install_requires=['django>=1.0'], description='Token generator and processor to provide automatic login links for users' )
<commit_before>from distutils.core import setup setup( name='django_autologin', version='0.1', packages=['django_autologin', 'django_autologin.templatetags'], install_requires=['django>=1.0'], description='Token generator and processor to provide automatic login links for users' ) <commit_msg>Use 4 spaces for indentation.<commit_after>from distutils.core import setup setup( name='django_autologin', version='0.1', packages=['django_autologin', 'django_autologin.templatetags'], install_requires=['django>=1.0'], description='Token generator and processor to provide automatic login links for users' )
2cbc3a5197694905606ce5251516c825b28927d7
setup.py
setup.py
import os from setuptools import setup longDesc = "" if os.path.exists("README.rst"): longDesc = open("README.rst").read().strip() setup( name = "pytesseract", version = "0.1.7", author = "Samuel Hoffstaetter", author_email="pytesseract@madmaze.net", maintainer = "Matthias Lee", maintainer_email = "pytesseract@madmaze.net", description = ("Python-tesseract is a python wrapper for google's Tesseract-OCR"), long_description = longDesc, license = "GPLv3", keywords = "python-tesseract OCR Python", url = "https://github.com/madmaze/python-tesseract", packages=['pytesseract'], package_dir={'pytesseract': 'src'}, package_data = {'pytesseract': ['*.png','*.jpg']}, entry_points = {'console_scripts': ['pytesseract = pytesseract.pytesseract:main']}, classifiers = [ 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3', ] )
import os from setuptools import setup README_PATH = 'README.rst' longDesc = "" if os.path.exists(README_PATH): with open(README_PATH) as readme: longDesc = readme.read() setup( name = "pytesseract", version = "0.1.7", author = "Samuel Hoffstaetter", author_email="pytesseract@madmaze.net", maintainer = "Matthias Lee", maintainer_email = "pytesseract@madmaze.net", description = ("Python-tesseract is a python wrapper for google's Tesseract-OCR"), long_description = longDesc, license = "GPLv3", keywords = "python-tesseract OCR Python", url = "https://github.com/madmaze/python-tesseract", packages=['pytesseract'], package_dir={'pytesseract': 'src'}, package_data = {'pytesseract': ['*.png','*.jpg']}, entry_points = {'console_scripts': ['pytesseract = pytesseract.pytesseract:main']}, classifiers = [ 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3', ] )
Fix long_description loading for PyPI
Fix long_description loading for PyPI
Python
apache-2.0
madmaze/pytesseract
import os from setuptools import setup longDesc = "" if os.path.exists("README.rst"): longDesc = open("README.rst").read().strip() setup( name = "pytesseract", version = "0.1.7", author = "Samuel Hoffstaetter", author_email="pytesseract@madmaze.net", maintainer = "Matthias Lee", maintainer_email = "pytesseract@madmaze.net", description = ("Python-tesseract is a python wrapper for google's Tesseract-OCR"), long_description = longDesc, license = "GPLv3", keywords = "python-tesseract OCR Python", url = "https://github.com/madmaze/python-tesseract", packages=['pytesseract'], package_dir={'pytesseract': 'src'}, package_data = {'pytesseract': ['*.png','*.jpg']}, entry_points = {'console_scripts': ['pytesseract = pytesseract.pytesseract:main']}, classifiers = [ 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3', ] ) Fix long_description loading for PyPI
import os from setuptools import setup README_PATH = 'README.rst' longDesc = "" if os.path.exists(README_PATH): with open(README_PATH) as readme: longDesc = readme.read() setup( name = "pytesseract", version = "0.1.7", author = "Samuel Hoffstaetter", author_email="pytesseract@madmaze.net", maintainer = "Matthias Lee", maintainer_email = "pytesseract@madmaze.net", description = ("Python-tesseract is a python wrapper for google's Tesseract-OCR"), long_description = longDesc, license = "GPLv3", keywords = "python-tesseract OCR Python", url = "https://github.com/madmaze/python-tesseract", packages=['pytesseract'], package_dir={'pytesseract': 'src'}, package_data = {'pytesseract': ['*.png','*.jpg']}, entry_points = {'console_scripts': ['pytesseract = pytesseract.pytesseract:main']}, classifiers = [ 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3', ] )
<commit_before>import os from setuptools import setup longDesc = "" if os.path.exists("README.rst"): longDesc = open("README.rst").read().strip() setup( name = "pytesseract", version = "0.1.7", author = "Samuel Hoffstaetter", author_email="pytesseract@madmaze.net", maintainer = "Matthias Lee", maintainer_email = "pytesseract@madmaze.net", description = ("Python-tesseract is a python wrapper for google's Tesseract-OCR"), long_description = longDesc, license = "GPLv3", keywords = "python-tesseract OCR Python", url = "https://github.com/madmaze/python-tesseract", packages=['pytesseract'], package_dir={'pytesseract': 'src'}, package_data = {'pytesseract': ['*.png','*.jpg']}, entry_points = {'console_scripts': ['pytesseract = pytesseract.pytesseract:main']}, classifiers = [ 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3', ] ) <commit_msg>Fix long_description loading for PyPI<commit_after>
import os from setuptools import setup README_PATH = 'README.rst' longDesc = "" if os.path.exists(README_PATH): with open(README_PATH) as readme: longDesc = readme.read() setup( name = "pytesseract", version = "0.1.7", author = "Samuel Hoffstaetter", author_email="pytesseract@madmaze.net", maintainer = "Matthias Lee", maintainer_email = "pytesseract@madmaze.net", description = ("Python-tesseract is a python wrapper for google's Tesseract-OCR"), long_description = longDesc, license = "GPLv3", keywords = "python-tesseract OCR Python", url = "https://github.com/madmaze/python-tesseract", packages=['pytesseract'], package_dir={'pytesseract': 'src'}, package_data = {'pytesseract': ['*.png','*.jpg']}, entry_points = {'console_scripts': ['pytesseract = pytesseract.pytesseract:main']}, classifiers = [ 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3', ] )
import os from setuptools import setup longDesc = "" if os.path.exists("README.rst"): longDesc = open("README.rst").read().strip() setup( name = "pytesseract", version = "0.1.7", author = "Samuel Hoffstaetter", author_email="pytesseract@madmaze.net", maintainer = "Matthias Lee", maintainer_email = "pytesseract@madmaze.net", description = ("Python-tesseract is a python wrapper for google's Tesseract-OCR"), long_description = longDesc, license = "GPLv3", keywords = "python-tesseract OCR Python", url = "https://github.com/madmaze/python-tesseract", packages=['pytesseract'], package_dir={'pytesseract': 'src'}, package_data = {'pytesseract': ['*.png','*.jpg']}, entry_points = {'console_scripts': ['pytesseract = pytesseract.pytesseract:main']}, classifiers = [ 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3', ] ) Fix long_description loading for PyPIimport os from setuptools import setup README_PATH = 'README.rst' longDesc = "" if os.path.exists(README_PATH): with open(README_PATH) as readme: longDesc = readme.read() setup( name = "pytesseract", version = "0.1.7", author = "Samuel Hoffstaetter", author_email="pytesseract@madmaze.net", maintainer = "Matthias Lee", maintainer_email = "pytesseract@madmaze.net", description = ("Python-tesseract is a python wrapper for google's Tesseract-OCR"), long_description = longDesc, license = "GPLv3", keywords = "python-tesseract OCR Python", url = "https://github.com/madmaze/python-tesseract", packages=['pytesseract'], package_dir={'pytesseract': 'src'}, package_data = {'pytesseract': ['*.png','*.jpg']}, entry_points = {'console_scripts': ['pytesseract = pytesseract.pytesseract:main']}, classifiers = [ 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3', ] )
<commit_before>import os from setuptools import setup longDesc = "" if os.path.exists("README.rst"): longDesc = open("README.rst").read().strip() setup( name = "pytesseract", version = "0.1.7", author = "Samuel Hoffstaetter", author_email="pytesseract@madmaze.net", maintainer = "Matthias Lee", maintainer_email = "pytesseract@madmaze.net", description = ("Python-tesseract is a python wrapper for google's Tesseract-OCR"), long_description = longDesc, license = "GPLv3", keywords = "python-tesseract OCR Python", url = "https://github.com/madmaze/python-tesseract", packages=['pytesseract'], package_dir={'pytesseract': 'src'}, package_data = {'pytesseract': ['*.png','*.jpg']}, entry_points = {'console_scripts': ['pytesseract = pytesseract.pytesseract:main']}, classifiers = [ 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3', ] ) <commit_msg>Fix long_description loading for PyPI<commit_after>import os from setuptools import setup README_PATH = 'README.rst' longDesc = "" if os.path.exists(README_PATH): with open(README_PATH) as readme: longDesc = readme.read() setup( name = "pytesseract", version = "0.1.7", author = "Samuel Hoffstaetter", author_email="pytesseract@madmaze.net", maintainer = "Matthias Lee", maintainer_email = "pytesseract@madmaze.net", description = ("Python-tesseract is a python wrapper for google's Tesseract-OCR"), long_description = longDesc, license = "GPLv3", keywords = "python-tesseract OCR Python", url = "https://github.com/madmaze/python-tesseract", packages=['pytesseract'], package_dir={'pytesseract': 'src'}, package_data = {'pytesseract': ['*.png','*.jpg']}, entry_points = {'console_scripts': ['pytesseract = pytesseract.pytesseract:main']}, classifiers = [ 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3', ] )
13494098e2941ff87d80710ecc00c35142851175
setup.py
setup.py
#!/usr/bin/env python # Copyright (c) 2013 Hewlett-Packard Development Company, L.P. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. # THIS FILE IS MANAGED BY THE GLOBAL REQUIREMENTS REPO - DO NOT EDIT import setuptools setuptools.setup( setup_requires=['pbr'], pbr=True)
#!/usr/bin/env python # Copyright (c) 2013 Hewlett-Packard Development Company, L.P. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. # THIS FILE IS MANAGED BY THE GLOBAL REQUIREMENTS REPO - DO NOT EDIT import setuptools, sys if sys.version_info < (2, 7): sys.exit('Sorry, Python < 2.7 is not supported for' ' python-swiftclient>=3.0') setuptools.setup( setup_requires=['pbr'], pbr=True)
Add python version constraint python>=2.7
Add python version constraint python>=2.7 The 'swift' command from v3.0.0 does not work in Python 2.6, bacause some code is incompatible with Python 2.6 This patch is to add a constraint of python version Change-Id: I5197cba0c2cd3135d08498df827a52f8bba98d4d Closes-bug: #1590334
Python
apache-2.0
openstack/python-swiftclient,openstack/python-swiftclient,sohonetlabs/python-swiftclient,sohonetlabs/python-swiftclient
#!/usr/bin/env python # Copyright (c) 2013 Hewlett-Packard Development Company, L.P. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. # THIS FILE IS MANAGED BY THE GLOBAL REQUIREMENTS REPO - DO NOT EDIT import setuptools setuptools.setup( setup_requires=['pbr'], pbr=True) Add python version constraint python>=2.7 The 'swift' command from v3.0.0 does not work in Python 2.6, bacause some code is incompatible with Python 2.6 This patch is to add a constraint of python version Change-Id: I5197cba0c2cd3135d08498df827a52f8bba98d4d Closes-bug: #1590334
#!/usr/bin/env python # Copyright (c) 2013 Hewlett-Packard Development Company, L.P. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. # THIS FILE IS MANAGED BY THE GLOBAL REQUIREMENTS REPO - DO NOT EDIT import setuptools, sys if sys.version_info < (2, 7): sys.exit('Sorry, Python < 2.7 is not supported for' ' python-swiftclient>=3.0') setuptools.setup( setup_requires=['pbr'], pbr=True)
<commit_before>#!/usr/bin/env python # Copyright (c) 2013 Hewlett-Packard Development Company, L.P. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. # THIS FILE IS MANAGED BY THE GLOBAL REQUIREMENTS REPO - DO NOT EDIT import setuptools setuptools.setup( setup_requires=['pbr'], pbr=True) <commit_msg>Add python version constraint python>=2.7 The 'swift' command from v3.0.0 does not work in Python 2.6, bacause some code is incompatible with Python 2.6 This patch is to add a constraint of python version Change-Id: I5197cba0c2cd3135d08498df827a52f8bba98d4d Closes-bug: #1590334<commit_after>
#!/usr/bin/env python # Copyright (c) 2013 Hewlett-Packard Development Company, L.P. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. # THIS FILE IS MANAGED BY THE GLOBAL REQUIREMENTS REPO - DO NOT EDIT import setuptools, sys if sys.version_info < (2, 7): sys.exit('Sorry, Python < 2.7 is not supported for' ' python-swiftclient>=3.0') setuptools.setup( setup_requires=['pbr'], pbr=True)
#!/usr/bin/env python # Copyright (c) 2013 Hewlett-Packard Development Company, L.P. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. # THIS FILE IS MANAGED BY THE GLOBAL REQUIREMENTS REPO - DO NOT EDIT import setuptools setuptools.setup( setup_requires=['pbr'], pbr=True) Add python version constraint python>=2.7 The 'swift' command from v3.0.0 does not work in Python 2.6, bacause some code is incompatible with Python 2.6 This patch is to add a constraint of python version Change-Id: I5197cba0c2cd3135d08498df827a52f8bba98d4d Closes-bug: #1590334#!/usr/bin/env python # Copyright (c) 2013 Hewlett-Packard Development Company, L.P. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. # THIS FILE IS MANAGED BY THE GLOBAL REQUIREMENTS REPO - DO NOT EDIT import setuptools, sys if sys.version_info < (2, 7): sys.exit('Sorry, Python < 2.7 is not supported for' ' python-swiftclient>=3.0') setuptools.setup( setup_requires=['pbr'], pbr=True)
<commit_before>#!/usr/bin/env python # Copyright (c) 2013 Hewlett-Packard Development Company, L.P. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. # THIS FILE IS MANAGED BY THE GLOBAL REQUIREMENTS REPO - DO NOT EDIT import setuptools setuptools.setup( setup_requires=['pbr'], pbr=True) <commit_msg>Add python version constraint python>=2.7 The 'swift' command from v3.0.0 does not work in Python 2.6, bacause some code is incompatible with Python 2.6 This patch is to add a constraint of python version Change-Id: I5197cba0c2cd3135d08498df827a52f8bba98d4d Closes-bug: #1590334<commit_after>#!/usr/bin/env python # Copyright (c) 2013 Hewlett-Packard Development Company, L.P. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. # THIS FILE IS MANAGED BY THE GLOBAL REQUIREMENTS REPO - DO NOT EDIT import setuptools, sys if sys.version_info < (2, 7): sys.exit('Sorry, Python < 2.7 is not supported for' ' python-swiftclient>=3.0') setuptools.setup( setup_requires=['pbr'], pbr=True)
23d7418aa18f6e4ce1b97938144a8968e2b0cb9b
setup.py
setup.py
#!/usr/bin/env python """Setup script for the pyparsing module distribution.""" # Setuptools depends on pyparsing (via packaging) as of version 34, so allow # installing without it to avoid bootstrap problems. try: from setuptools import setup except ImportError: from distutils.core import setup import sys import os from pyparsing import __version__ as pyparsing_version modules = ["pyparsing",] setup(# Distribution meta-data name = "pyparsing", version = pyparsing_version, description = "Python parsing module", author = "Paul McGuire", author_email = "ptmcg@users.sourceforge.net", url = "https://github.com/pyparsing/pyparsing/", download_url = "https://pypi.org/project/pyparsing/", license = "MIT License", py_modules = modules, classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', ] )
#!/usr/bin/env python """Setup script for the pyparsing module distribution.""" # Setuptools depends on pyparsing (via packaging) as of version 34, so allow # installing without it to avoid bootstrap problems. try: from setuptools import setup except ImportError: from distutils.core import setup import sys import os from pyparsing import __version__ as pyparsing_version modules = ["pyparsing",] setup(# Distribution meta-data name = "pyparsing", version = pyparsing_version, description = "Python parsing module", author = "Paul McGuire", author_email = "ptmcg@users.sourceforge.net", url = "https://github.com/pyparsing/pyparsing/", download_url = "https://pypi.org/project/pyparsing/", license = "MIT License", py_modules = modules, python_requires='>=2.6, !=3.0.*, !=3.1.*, !=3.2.*', classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', ] )
Add python_requires to help pip
Add python_requires to help pip
Python
mit
pyparsing/pyparsing,pyparsing/pyparsing
#!/usr/bin/env python """Setup script for the pyparsing module distribution.""" # Setuptools depends on pyparsing (via packaging) as of version 34, so allow # installing without it to avoid bootstrap problems. try: from setuptools import setup except ImportError: from distutils.core import setup import sys import os from pyparsing import __version__ as pyparsing_version modules = ["pyparsing",] setup(# Distribution meta-data name = "pyparsing", version = pyparsing_version, description = "Python parsing module", author = "Paul McGuire", author_email = "ptmcg@users.sourceforge.net", url = "https://github.com/pyparsing/pyparsing/", download_url = "https://pypi.org/project/pyparsing/", license = "MIT License", py_modules = modules, classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', ] ) Add python_requires to help pip
#!/usr/bin/env python """Setup script for the pyparsing module distribution.""" # Setuptools depends on pyparsing (via packaging) as of version 34, so allow # installing without it to avoid bootstrap problems. try: from setuptools import setup except ImportError: from distutils.core import setup import sys import os from pyparsing import __version__ as pyparsing_version modules = ["pyparsing",] setup(# Distribution meta-data name = "pyparsing", version = pyparsing_version, description = "Python parsing module", author = "Paul McGuire", author_email = "ptmcg@users.sourceforge.net", url = "https://github.com/pyparsing/pyparsing/", download_url = "https://pypi.org/project/pyparsing/", license = "MIT License", py_modules = modules, python_requires='>=2.6, !=3.0.*, !=3.1.*, !=3.2.*', classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', ] )
<commit_before>#!/usr/bin/env python """Setup script for the pyparsing module distribution.""" # Setuptools depends on pyparsing (via packaging) as of version 34, so allow # installing without it to avoid bootstrap problems. try: from setuptools import setup except ImportError: from distutils.core import setup import sys import os from pyparsing import __version__ as pyparsing_version modules = ["pyparsing",] setup(# Distribution meta-data name = "pyparsing", version = pyparsing_version, description = "Python parsing module", author = "Paul McGuire", author_email = "ptmcg@users.sourceforge.net", url = "https://github.com/pyparsing/pyparsing/", download_url = "https://pypi.org/project/pyparsing/", license = "MIT License", py_modules = modules, classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', ] ) <commit_msg>Add python_requires to help pip<commit_after>
#!/usr/bin/env python """Setup script for the pyparsing module distribution.""" # Setuptools depends on pyparsing (via packaging) as of version 34, so allow # installing without it to avoid bootstrap problems. try: from setuptools import setup except ImportError: from distutils.core import setup import sys import os from pyparsing import __version__ as pyparsing_version modules = ["pyparsing",] setup(# Distribution meta-data name = "pyparsing", version = pyparsing_version, description = "Python parsing module", author = "Paul McGuire", author_email = "ptmcg@users.sourceforge.net", url = "https://github.com/pyparsing/pyparsing/", download_url = "https://pypi.org/project/pyparsing/", license = "MIT License", py_modules = modules, python_requires='>=2.6, !=3.0.*, !=3.1.*, !=3.2.*', classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', ] )
#!/usr/bin/env python """Setup script for the pyparsing module distribution.""" # Setuptools depends on pyparsing (via packaging) as of version 34, so allow # installing without it to avoid bootstrap problems. try: from setuptools import setup except ImportError: from distutils.core import setup import sys import os from pyparsing import __version__ as pyparsing_version modules = ["pyparsing",] setup(# Distribution meta-data name = "pyparsing", version = pyparsing_version, description = "Python parsing module", author = "Paul McGuire", author_email = "ptmcg@users.sourceforge.net", url = "https://github.com/pyparsing/pyparsing/", download_url = "https://pypi.org/project/pyparsing/", license = "MIT License", py_modules = modules, classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', ] ) Add python_requires to help pip#!/usr/bin/env python """Setup script for the pyparsing module distribution.""" # Setuptools depends on pyparsing (via packaging) as of version 34, so allow # installing without it to avoid bootstrap problems. try: from setuptools import setup except ImportError: from distutils.core import setup import sys import os from pyparsing import __version__ as pyparsing_version modules = ["pyparsing",] setup(# Distribution meta-data name = "pyparsing", version = pyparsing_version, description = "Python parsing module", author = "Paul McGuire", author_email = "ptmcg@users.sourceforge.net", url = "https://github.com/pyparsing/pyparsing/", download_url = "https://pypi.org/project/pyparsing/", license = "MIT License", py_modules = modules, python_requires='>=2.6, !=3.0.*, !=3.1.*, !=3.2.*', classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', ] )
<commit_before>#!/usr/bin/env python """Setup script for the pyparsing module distribution.""" # Setuptools depends on pyparsing (via packaging) as of version 34, so allow # installing without it to avoid bootstrap problems. try: from setuptools import setup except ImportError: from distutils.core import setup import sys import os from pyparsing import __version__ as pyparsing_version modules = ["pyparsing",] setup(# Distribution meta-data name = "pyparsing", version = pyparsing_version, description = "Python parsing module", author = "Paul McGuire", author_email = "ptmcg@users.sourceforge.net", url = "https://github.com/pyparsing/pyparsing/", download_url = "https://pypi.org/project/pyparsing/", license = "MIT License", py_modules = modules, classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', ] ) <commit_msg>Add python_requires to help pip<commit_after>#!/usr/bin/env python """Setup script for the pyparsing module distribution.""" # Setuptools depends on pyparsing (via packaging) as of version 34, so allow # installing without it to avoid bootstrap problems. try: from setuptools import setup except ImportError: from distutils.core import setup import sys import os from pyparsing import __version__ as pyparsing_version modules = ["pyparsing",] setup(# Distribution meta-data name = "pyparsing", version = pyparsing_version, description = "Python parsing module", author = "Paul McGuire", author_email = "ptmcg@users.sourceforge.net", url = "https://github.com/pyparsing/pyparsing/", download_url = "https://pypi.org/project/pyparsing/", license = "MIT License", py_modules = modules, python_requires='>=2.6, !=3.0.*, !=3.1.*, !=3.2.*', classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', ] )
a71c9ef43b1801343bccbaf4e9f3a4a2eaa570f0
setup.py
setup.py
import os from setuptools import setup def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() setup(name='windpowerlib', version='0.2.0dev', description='Creating time series of wind power plants.', url='http://github.com/wind-python/windpowerlib', author='oemof developer group', author_email='windpowerlib@rl-institut.de', license=None, packages=['windpowerlib'], package_data={ 'windpowerlib': [os.path.join('data', '*.csv'), os.path.join('oedb', '*.csv')]}, long_description=read('README.rst'), zip_safe=False, install_requires=['pandas >= 0.19.1, < 0.25', 'requests < 3.0'], extras_require={ 'dev': ['pytest', 'jupyter', 'sphinx_rtd_theme', 'nbformat', 'numpy']})
import os from setuptools import setup def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() setup(name='windpowerlib', version='0.2.0dev', description='Creating time series of wind power plants.', url='http://github.com/wind-python/windpowerlib', author='oemof developer group', author_email='windpowerlib@rl-institut.de', license=None, packages=['windpowerlib'], package_data={ 'windpowerlib': [os.path.join('data', '*.csv'), os.path.join('oedb', '*.csv')]}, long_description=read('README.rst'), long_description_content_type='text/x-rst', zip_safe=False, install_requires=['pandas >= 0.19.1, < 0.25', 'requests < 3.0'], extras_require={ 'dev': ['pytest', 'jupyter', 'sphinx_rtd_theme', 'nbformat', 'numpy']})
Add content type for long description
Add content type for long description
Python
mit
wind-python/windpowerlib
import os from setuptools import setup def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() setup(name='windpowerlib', version='0.2.0dev', description='Creating time series of wind power plants.', url='http://github.com/wind-python/windpowerlib', author='oemof developer group', author_email='windpowerlib@rl-institut.de', license=None, packages=['windpowerlib'], package_data={ 'windpowerlib': [os.path.join('data', '*.csv'), os.path.join('oedb', '*.csv')]}, long_description=read('README.rst'), zip_safe=False, install_requires=['pandas >= 0.19.1, < 0.25', 'requests < 3.0'], extras_require={ 'dev': ['pytest', 'jupyter', 'sphinx_rtd_theme', 'nbformat', 'numpy']}) Add content type for long description
import os from setuptools import setup def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() setup(name='windpowerlib', version='0.2.0dev', description='Creating time series of wind power plants.', url='http://github.com/wind-python/windpowerlib', author='oemof developer group', author_email='windpowerlib@rl-institut.de', license=None, packages=['windpowerlib'], package_data={ 'windpowerlib': [os.path.join('data', '*.csv'), os.path.join('oedb', '*.csv')]}, long_description=read('README.rst'), long_description_content_type='text/x-rst', zip_safe=False, install_requires=['pandas >= 0.19.1, < 0.25', 'requests < 3.0'], extras_require={ 'dev': ['pytest', 'jupyter', 'sphinx_rtd_theme', 'nbformat', 'numpy']})
<commit_before>import os from setuptools import setup def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() setup(name='windpowerlib', version='0.2.0dev', description='Creating time series of wind power plants.', url='http://github.com/wind-python/windpowerlib', author='oemof developer group', author_email='windpowerlib@rl-institut.de', license=None, packages=['windpowerlib'], package_data={ 'windpowerlib': [os.path.join('data', '*.csv'), os.path.join('oedb', '*.csv')]}, long_description=read('README.rst'), zip_safe=False, install_requires=['pandas >= 0.19.1, < 0.25', 'requests < 3.0'], extras_require={ 'dev': ['pytest', 'jupyter', 'sphinx_rtd_theme', 'nbformat', 'numpy']}) <commit_msg>Add content type for long description<commit_after>
import os from setuptools import setup def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() setup(name='windpowerlib', version='0.2.0dev', description='Creating time series of wind power plants.', url='http://github.com/wind-python/windpowerlib', author='oemof developer group', author_email='windpowerlib@rl-institut.de', license=None, packages=['windpowerlib'], package_data={ 'windpowerlib': [os.path.join('data', '*.csv'), os.path.join('oedb', '*.csv')]}, long_description=read('README.rst'), long_description_content_type='text/x-rst', zip_safe=False, install_requires=['pandas >= 0.19.1, < 0.25', 'requests < 3.0'], extras_require={ 'dev': ['pytest', 'jupyter', 'sphinx_rtd_theme', 'nbformat', 'numpy']})
import os from setuptools import setup def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() setup(name='windpowerlib', version='0.2.0dev', description='Creating time series of wind power plants.', url='http://github.com/wind-python/windpowerlib', author='oemof developer group', author_email='windpowerlib@rl-institut.de', license=None, packages=['windpowerlib'], package_data={ 'windpowerlib': [os.path.join('data', '*.csv'), os.path.join('oedb', '*.csv')]}, long_description=read('README.rst'), zip_safe=False, install_requires=['pandas >= 0.19.1, < 0.25', 'requests < 3.0'], extras_require={ 'dev': ['pytest', 'jupyter', 'sphinx_rtd_theme', 'nbformat', 'numpy']}) Add content type for long descriptionimport os from setuptools import setup def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() setup(name='windpowerlib', version='0.2.0dev', description='Creating time series of wind power plants.', url='http://github.com/wind-python/windpowerlib', author='oemof developer group', author_email='windpowerlib@rl-institut.de', license=None, packages=['windpowerlib'], package_data={ 'windpowerlib': [os.path.join('data', '*.csv'), os.path.join('oedb', '*.csv')]}, long_description=read('README.rst'), long_description_content_type='text/x-rst', zip_safe=False, install_requires=['pandas >= 0.19.1, < 0.25', 'requests < 3.0'], extras_require={ 'dev': ['pytest', 'jupyter', 'sphinx_rtd_theme', 'nbformat', 'numpy']})
<commit_before>import os from setuptools import setup def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() setup(name='windpowerlib', version='0.2.0dev', description='Creating time series of wind power plants.', url='http://github.com/wind-python/windpowerlib', author='oemof developer group', author_email='windpowerlib@rl-institut.de', license=None, packages=['windpowerlib'], package_data={ 'windpowerlib': [os.path.join('data', '*.csv'), os.path.join('oedb', '*.csv')]}, long_description=read('README.rst'), zip_safe=False, install_requires=['pandas >= 0.19.1, < 0.25', 'requests < 3.0'], extras_require={ 'dev': ['pytest', 'jupyter', 'sphinx_rtd_theme', 'nbformat', 'numpy']}) <commit_msg>Add content type for long description<commit_after>import os from setuptools import setup def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() setup(name='windpowerlib', version='0.2.0dev', description='Creating time series of wind power plants.', url='http://github.com/wind-python/windpowerlib', author='oemof developer group', author_email='windpowerlib@rl-institut.de', license=None, packages=['windpowerlib'], package_data={ 'windpowerlib': [os.path.join('data', '*.csv'), os.path.join('oedb', '*.csv')]}, long_description=read('README.rst'), long_description_content_type='text/x-rst', zip_safe=False, install_requires=['pandas >= 0.19.1, < 0.25', 'requests < 3.0'], extras_require={ 'dev': ['pytest', 'jupyter', 'sphinx_rtd_theme', 'nbformat', 'numpy']})
269b3951049255fd3b459ce254afe3b8d6ffea1b
setup.py
setup.py
"""Setup for drewtils project.""" import os try: from setuptools import setup setupTools = True except ImportError: from distutils.core import setup setupTools = False _classifiers = [ 'License :: OSI Approved :: MIT License', ] if os.path.exists('README.rst'): with open('README.rst') as readme: long_description = readme.read() else: long_description = '' setupArgs = { 'name': 'drewtils', 'version': "0.1.9", 'packages': ['drewtils'], 'author': 'Andrew Johnson', 'author_email': 'drewej@protonmail.com', 'description': 'Simple tools to make testing and file parsing easier', 'long_description': long_description, 'license': 'MIT', 'keywords': 'parsing files', 'url': 'https://github.com/drewejohnson/drewtils', 'classifiers': _classifiers, } if setupTools: setupArgs.update(**{ 'test_suite': 'drewtils.tests', 'python_requires': '>=2.7,!=3.1,!=3.2,!=3.3,!=3.4' }) setup(**setupArgs)
"""Setup for drewtils project.""" try: from setuptools import setup setupTools = True except ImportError: from distutils.core import setup setupTools = False _classifiers = [ 'License :: OSI Approved :: MIT License', ] with open('README.rst') as readme: long_description = readme.read() setupArgs = { 'name': 'drewtils', 'version': "0.1.9", 'packages': ['drewtils'], 'author': 'Andrew Johnson', 'author_email': 'drewej@protonmail.com', 'description': 'Simple tools to make testing and file parsing easier', 'long_description': long_description, 'license': 'MIT', 'keywords': 'parsing files', 'url': 'https://github.com/drewejohnson/drewtils', 'classifiers': _classifiers, } if setupTools: setupArgs.update(**{ 'test_suite': 'drewtils.tests', 'python_requires': '>=2.7,!=3.1,!=3.2,!=3.3,!=3.4' }) setup(**setupArgs)
Read long description from readme always
Read long description from readme always
Python
mit
drewejohnson/drewtils
"""Setup for drewtils project.""" import os try: from setuptools import setup setupTools = True except ImportError: from distutils.core import setup setupTools = False _classifiers = [ 'License :: OSI Approved :: MIT License', ] if os.path.exists('README.rst'): with open('README.rst') as readme: long_description = readme.read() else: long_description = '' setupArgs = { 'name': 'drewtils', 'version': "0.1.9", 'packages': ['drewtils'], 'author': 'Andrew Johnson', 'author_email': 'drewej@protonmail.com', 'description': 'Simple tools to make testing and file parsing easier', 'long_description': long_description, 'license': 'MIT', 'keywords': 'parsing files', 'url': 'https://github.com/drewejohnson/drewtils', 'classifiers': _classifiers, } if setupTools: setupArgs.update(**{ 'test_suite': 'drewtils.tests', 'python_requires': '>=2.7,!=3.1,!=3.2,!=3.3,!=3.4' }) setup(**setupArgs) Read long description from readme always
"""Setup for drewtils project.""" try: from setuptools import setup setupTools = True except ImportError: from distutils.core import setup setupTools = False _classifiers = [ 'License :: OSI Approved :: MIT License', ] with open('README.rst') as readme: long_description = readme.read() setupArgs = { 'name': 'drewtils', 'version': "0.1.9", 'packages': ['drewtils'], 'author': 'Andrew Johnson', 'author_email': 'drewej@protonmail.com', 'description': 'Simple tools to make testing and file parsing easier', 'long_description': long_description, 'license': 'MIT', 'keywords': 'parsing files', 'url': 'https://github.com/drewejohnson/drewtils', 'classifiers': _classifiers, } if setupTools: setupArgs.update(**{ 'test_suite': 'drewtils.tests', 'python_requires': '>=2.7,!=3.1,!=3.2,!=3.3,!=3.4' }) setup(**setupArgs)
<commit_before>"""Setup for drewtils project.""" import os try: from setuptools import setup setupTools = True except ImportError: from distutils.core import setup setupTools = False _classifiers = [ 'License :: OSI Approved :: MIT License', ] if os.path.exists('README.rst'): with open('README.rst') as readme: long_description = readme.read() else: long_description = '' setupArgs = { 'name': 'drewtils', 'version': "0.1.9", 'packages': ['drewtils'], 'author': 'Andrew Johnson', 'author_email': 'drewej@protonmail.com', 'description': 'Simple tools to make testing and file parsing easier', 'long_description': long_description, 'license': 'MIT', 'keywords': 'parsing files', 'url': 'https://github.com/drewejohnson/drewtils', 'classifiers': _classifiers, } if setupTools: setupArgs.update(**{ 'test_suite': 'drewtils.tests', 'python_requires': '>=2.7,!=3.1,!=3.2,!=3.3,!=3.4' }) setup(**setupArgs) <commit_msg>Read long description from readme always<commit_after>
"""Setup for drewtils project.""" try: from setuptools import setup setupTools = True except ImportError: from distutils.core import setup setupTools = False _classifiers = [ 'License :: OSI Approved :: MIT License', ] with open('README.rst') as readme: long_description = readme.read() setupArgs = { 'name': 'drewtils', 'version': "0.1.9", 'packages': ['drewtils'], 'author': 'Andrew Johnson', 'author_email': 'drewej@protonmail.com', 'description': 'Simple tools to make testing and file parsing easier', 'long_description': long_description, 'license': 'MIT', 'keywords': 'parsing files', 'url': 'https://github.com/drewejohnson/drewtils', 'classifiers': _classifiers, } if setupTools: setupArgs.update(**{ 'test_suite': 'drewtils.tests', 'python_requires': '>=2.7,!=3.1,!=3.2,!=3.3,!=3.4' }) setup(**setupArgs)
"""Setup for drewtils project.""" import os try: from setuptools import setup setupTools = True except ImportError: from distutils.core import setup setupTools = False _classifiers = [ 'License :: OSI Approved :: MIT License', ] if os.path.exists('README.rst'): with open('README.rst') as readme: long_description = readme.read() else: long_description = '' setupArgs = { 'name': 'drewtils', 'version': "0.1.9", 'packages': ['drewtils'], 'author': 'Andrew Johnson', 'author_email': 'drewej@protonmail.com', 'description': 'Simple tools to make testing and file parsing easier', 'long_description': long_description, 'license': 'MIT', 'keywords': 'parsing files', 'url': 'https://github.com/drewejohnson/drewtils', 'classifiers': _classifiers, } if setupTools: setupArgs.update(**{ 'test_suite': 'drewtils.tests', 'python_requires': '>=2.7,!=3.1,!=3.2,!=3.3,!=3.4' }) setup(**setupArgs) Read long description from readme always"""Setup for drewtils project.""" try: from setuptools import setup setupTools = True except ImportError: from distutils.core import setup setupTools = False _classifiers = [ 'License :: OSI Approved :: MIT License', ] with open('README.rst') as readme: long_description = readme.read() setupArgs = { 'name': 'drewtils', 'version': "0.1.9", 'packages': ['drewtils'], 'author': 'Andrew Johnson', 'author_email': 'drewej@protonmail.com', 'description': 'Simple tools to make testing and file parsing easier', 'long_description': long_description, 'license': 'MIT', 'keywords': 'parsing files', 'url': 'https://github.com/drewejohnson/drewtils', 'classifiers': _classifiers, } if setupTools: setupArgs.update(**{ 'test_suite': 'drewtils.tests', 'python_requires': '>=2.7,!=3.1,!=3.2,!=3.3,!=3.4' }) setup(**setupArgs)
<commit_before>"""Setup for drewtils project.""" import os try: from setuptools import setup setupTools = True except ImportError: from distutils.core import setup setupTools = False _classifiers = [ 'License :: OSI Approved :: MIT License', ] if os.path.exists('README.rst'): with open('README.rst') as readme: long_description = readme.read() else: long_description = '' setupArgs = { 'name': 'drewtils', 'version': "0.1.9", 'packages': ['drewtils'], 'author': 'Andrew Johnson', 'author_email': 'drewej@protonmail.com', 'description': 'Simple tools to make testing and file parsing easier', 'long_description': long_description, 'license': 'MIT', 'keywords': 'parsing files', 'url': 'https://github.com/drewejohnson/drewtils', 'classifiers': _classifiers, } if setupTools: setupArgs.update(**{ 'test_suite': 'drewtils.tests', 'python_requires': '>=2.7,!=3.1,!=3.2,!=3.3,!=3.4' }) setup(**setupArgs) <commit_msg>Read long description from readme always<commit_after>"""Setup for drewtils project.""" try: from setuptools import setup setupTools = True except ImportError: from distutils.core import setup setupTools = False _classifiers = [ 'License :: OSI Approved :: MIT License', ] with open('README.rst') as readme: long_description = readme.read() setupArgs = { 'name': 'drewtils', 'version': "0.1.9", 'packages': ['drewtils'], 'author': 'Andrew Johnson', 'author_email': 'drewej@protonmail.com', 'description': 'Simple tools to make testing and file parsing easier', 'long_description': long_description, 'license': 'MIT', 'keywords': 'parsing files', 'url': 'https://github.com/drewejohnson/drewtils', 'classifiers': _classifiers, } if setupTools: setupArgs.update(**{ 'test_suite': 'drewtils.tests', 'python_requires': '>=2.7,!=3.1,!=3.2,!=3.3,!=3.4' }) setup(**setupArgs)
0578b55f4e1e62b7c5c9c6f62721576970f43fdd
setup.py
setup.py
#!/usr/bin/env python # # Setup script for Django Evolution from ez_setup import use_setuptools use_setuptools() from setuptools import setup, find_packages from setuptools.command.test import test from django_evolution import get_package_version, VERSION def run_tests(*args): import os os.system('tests/runtests.py') test.run_tests = run_tests PACKAGE_NAME = 'django_evolution' download_url = ( 'http://downloads.reviewboard.org/releases/django-evolution/%s.%s/' % (VERSION[0], VERSION[1])) # Build the package setup( name=PACKAGE_NAME, version=get_package_version(), description='A database schema evolution tool for the Django web framework.', url='http://code.google.com/p/django-evolution/', author='Ben Khoo', author_email='khoobks@westnet.com.au', maintainer='Christian Hammond', maintainer_email='christian@beanbaginc.com', download_url=download_url, packages=find_packages(exclude=['tests']), install_requires=[ 'Django>=1.4.10', ], include_package_data=True, classifiers=[ 'Development Status :: 4 - Beta', 'Framework :: Django', 'Intended Audience :: Developers', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Software Development', 'Topic :: Software Development :: Libraries :: Python Modules', ] )
#!/usr/bin/env python # # Setup script for Django Evolution from ez_setup import use_setuptools use_setuptools() from setuptools import setup, find_packages from setuptools.command.test import test from django_evolution import get_package_version, VERSION def run_tests(*args): import os os.system('tests/runtests.py') test.run_tests = run_tests PACKAGE_NAME = 'django_evolution' download_url = ( 'http://downloads.reviewboard.org/releases/django-evolution/%s.%s/' % (VERSION[0], VERSION[1])) # Build the package setup( name=PACKAGE_NAME, version=get_package_version(), description='A database schema evolution tool for the Django web framework.', url='http://code.google.com/p/django-evolution/', author='Ben Khoo', author_email='khoobks@westnet.com.au', maintainer='Christian Hammond', maintainer_email='christian@beanbaginc.com', download_url=download_url, packages=find_packages(exclude=['tests']), install_requires=[ 'Django>=1.4.10,<1.7.0', ], include_package_data=True, classifiers=[ 'Development Status :: 4 - Beta', 'Framework :: Django', 'Intended Audience :: Developers', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Software Development', 'Topic :: Software Development :: Libraries :: Python Modules', ] )
Revert "Allow Django Evolution to install along with Django >= 1.7."
Revert "Allow Django Evolution to install along with Django >= 1.7." This reverts commit 28b280bb04f806f614f6f2cd25ce779b551fef9e. This was committed to the wrong branch.
Python
bsd-3-clause
beanbaginc/django-evolution
#!/usr/bin/env python # # Setup script for Django Evolution from ez_setup import use_setuptools use_setuptools() from setuptools import setup, find_packages from setuptools.command.test import test from django_evolution import get_package_version, VERSION def run_tests(*args): import os os.system('tests/runtests.py') test.run_tests = run_tests PACKAGE_NAME = 'django_evolution' download_url = ( 'http://downloads.reviewboard.org/releases/django-evolution/%s.%s/' % (VERSION[0], VERSION[1])) # Build the package setup( name=PACKAGE_NAME, version=get_package_version(), description='A database schema evolution tool for the Django web framework.', url='http://code.google.com/p/django-evolution/', author='Ben Khoo', author_email='khoobks@westnet.com.au', maintainer='Christian Hammond', maintainer_email='christian@beanbaginc.com', download_url=download_url, packages=find_packages(exclude=['tests']), install_requires=[ 'Django>=1.4.10', ], include_package_data=True, classifiers=[ 'Development Status :: 4 - Beta', 'Framework :: Django', 'Intended Audience :: Developers', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Software Development', 'Topic :: Software Development :: Libraries :: Python Modules', ] ) Revert "Allow Django Evolution to install along with Django >= 1.7." This reverts commit 28b280bb04f806f614f6f2cd25ce779b551fef9e. This was committed to the wrong branch.
#!/usr/bin/env python # # Setup script for Django Evolution from ez_setup import use_setuptools use_setuptools() from setuptools import setup, find_packages from setuptools.command.test import test from django_evolution import get_package_version, VERSION def run_tests(*args): import os os.system('tests/runtests.py') test.run_tests = run_tests PACKAGE_NAME = 'django_evolution' download_url = ( 'http://downloads.reviewboard.org/releases/django-evolution/%s.%s/' % (VERSION[0], VERSION[1])) # Build the package setup( name=PACKAGE_NAME, version=get_package_version(), description='A database schema evolution tool for the Django web framework.', url='http://code.google.com/p/django-evolution/', author='Ben Khoo', author_email='khoobks@westnet.com.au', maintainer='Christian Hammond', maintainer_email='christian@beanbaginc.com', download_url=download_url, packages=find_packages(exclude=['tests']), install_requires=[ 'Django>=1.4.10,<1.7.0', ], include_package_data=True, classifiers=[ 'Development Status :: 4 - Beta', 'Framework :: Django', 'Intended Audience :: Developers', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Software Development', 'Topic :: Software Development :: Libraries :: Python Modules', ] )
<commit_before>#!/usr/bin/env python # # Setup script for Django Evolution from ez_setup import use_setuptools use_setuptools() from setuptools import setup, find_packages from setuptools.command.test import test from django_evolution import get_package_version, VERSION def run_tests(*args): import os os.system('tests/runtests.py') test.run_tests = run_tests PACKAGE_NAME = 'django_evolution' download_url = ( 'http://downloads.reviewboard.org/releases/django-evolution/%s.%s/' % (VERSION[0], VERSION[1])) # Build the package setup( name=PACKAGE_NAME, version=get_package_version(), description='A database schema evolution tool for the Django web framework.', url='http://code.google.com/p/django-evolution/', author='Ben Khoo', author_email='khoobks@westnet.com.au', maintainer='Christian Hammond', maintainer_email='christian@beanbaginc.com', download_url=download_url, packages=find_packages(exclude=['tests']), install_requires=[ 'Django>=1.4.10', ], include_package_data=True, classifiers=[ 'Development Status :: 4 - Beta', 'Framework :: Django', 'Intended Audience :: Developers', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Software Development', 'Topic :: Software Development :: Libraries :: Python Modules', ] ) <commit_msg>Revert "Allow Django Evolution to install along with Django >= 1.7." This reverts commit 28b280bb04f806f614f6f2cd25ce779b551fef9e. This was committed to the wrong branch.<commit_after>
#!/usr/bin/env python # # Setup script for Django Evolution from ez_setup import use_setuptools use_setuptools() from setuptools import setup, find_packages from setuptools.command.test import test from django_evolution import get_package_version, VERSION def run_tests(*args): import os os.system('tests/runtests.py') test.run_tests = run_tests PACKAGE_NAME = 'django_evolution' download_url = ( 'http://downloads.reviewboard.org/releases/django-evolution/%s.%s/' % (VERSION[0], VERSION[1])) # Build the package setup( name=PACKAGE_NAME, version=get_package_version(), description='A database schema evolution tool for the Django web framework.', url='http://code.google.com/p/django-evolution/', author='Ben Khoo', author_email='khoobks@westnet.com.au', maintainer='Christian Hammond', maintainer_email='christian@beanbaginc.com', download_url=download_url, packages=find_packages(exclude=['tests']), install_requires=[ 'Django>=1.4.10,<1.7.0', ], include_package_data=True, classifiers=[ 'Development Status :: 4 - Beta', 'Framework :: Django', 'Intended Audience :: Developers', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Software Development', 'Topic :: Software Development :: Libraries :: Python Modules', ] )
#!/usr/bin/env python # # Setup script for Django Evolution from ez_setup import use_setuptools use_setuptools() from setuptools import setup, find_packages from setuptools.command.test import test from django_evolution import get_package_version, VERSION def run_tests(*args): import os os.system('tests/runtests.py') test.run_tests = run_tests PACKAGE_NAME = 'django_evolution' download_url = ( 'http://downloads.reviewboard.org/releases/django-evolution/%s.%s/' % (VERSION[0], VERSION[1])) # Build the package setup( name=PACKAGE_NAME, version=get_package_version(), description='A database schema evolution tool for the Django web framework.', url='http://code.google.com/p/django-evolution/', author='Ben Khoo', author_email='khoobks@westnet.com.au', maintainer='Christian Hammond', maintainer_email='christian@beanbaginc.com', download_url=download_url, packages=find_packages(exclude=['tests']), install_requires=[ 'Django>=1.4.10', ], include_package_data=True, classifiers=[ 'Development Status :: 4 - Beta', 'Framework :: Django', 'Intended Audience :: Developers', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Software Development', 'Topic :: Software Development :: Libraries :: Python Modules', ] ) Revert "Allow Django Evolution to install along with Django >= 1.7." This reverts commit 28b280bb04f806f614f6f2cd25ce779b551fef9e. This was committed to the wrong branch.#!/usr/bin/env python # # Setup script for Django Evolution from ez_setup import use_setuptools use_setuptools() from setuptools import setup, find_packages from setuptools.command.test import test from django_evolution import get_package_version, VERSION def run_tests(*args): import os os.system('tests/runtests.py') test.run_tests = run_tests PACKAGE_NAME = 'django_evolution' download_url = ( 'http://downloads.reviewboard.org/releases/django-evolution/%s.%s/' % (VERSION[0], VERSION[1])) # Build the package setup( name=PACKAGE_NAME, version=get_package_version(), description='A database schema evolution tool for the Django web framework.', url='http://code.google.com/p/django-evolution/', author='Ben Khoo', author_email='khoobks@westnet.com.au', maintainer='Christian Hammond', maintainer_email='christian@beanbaginc.com', download_url=download_url, packages=find_packages(exclude=['tests']), install_requires=[ 'Django>=1.4.10,<1.7.0', ], include_package_data=True, classifiers=[ 'Development Status :: 4 - Beta', 'Framework :: Django', 'Intended Audience :: Developers', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Software Development', 'Topic :: Software Development :: Libraries :: Python Modules', ] )
<commit_before>#!/usr/bin/env python # # Setup script for Django Evolution from ez_setup import use_setuptools use_setuptools() from setuptools import setup, find_packages from setuptools.command.test import test from django_evolution import get_package_version, VERSION def run_tests(*args): import os os.system('tests/runtests.py') test.run_tests = run_tests PACKAGE_NAME = 'django_evolution' download_url = ( 'http://downloads.reviewboard.org/releases/django-evolution/%s.%s/' % (VERSION[0], VERSION[1])) # Build the package setup( name=PACKAGE_NAME, version=get_package_version(), description='A database schema evolution tool for the Django web framework.', url='http://code.google.com/p/django-evolution/', author='Ben Khoo', author_email='khoobks@westnet.com.au', maintainer='Christian Hammond', maintainer_email='christian@beanbaginc.com', download_url=download_url, packages=find_packages(exclude=['tests']), install_requires=[ 'Django>=1.4.10', ], include_package_data=True, classifiers=[ 'Development Status :: 4 - Beta', 'Framework :: Django', 'Intended Audience :: Developers', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Software Development', 'Topic :: Software Development :: Libraries :: Python Modules', ] ) <commit_msg>Revert "Allow Django Evolution to install along with Django >= 1.7." This reverts commit 28b280bb04f806f614f6f2cd25ce779b551fef9e. This was committed to the wrong branch.<commit_after>#!/usr/bin/env python # # Setup script for Django Evolution from ez_setup import use_setuptools use_setuptools() from setuptools import setup, find_packages from setuptools.command.test import test from django_evolution import get_package_version, VERSION def run_tests(*args): import os os.system('tests/runtests.py') test.run_tests = run_tests PACKAGE_NAME = 'django_evolution' download_url = ( 'http://downloads.reviewboard.org/releases/django-evolution/%s.%s/' % (VERSION[0], VERSION[1])) # Build the package setup( name=PACKAGE_NAME, version=get_package_version(), description='A database schema evolution tool for the Django web framework.', url='http://code.google.com/p/django-evolution/', author='Ben Khoo', author_email='khoobks@westnet.com.au', maintainer='Christian Hammond', maintainer_email='christian@beanbaginc.com', download_url=download_url, packages=find_packages(exclude=['tests']), install_requires=[ 'Django>=1.4.10,<1.7.0', ], include_package_data=True, classifiers=[ 'Development Status :: 4 - Beta', 'Framework :: Django', 'Intended Audience :: Developers', 'Natural Language :: English', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Software Development', 'Topic :: Software Development :: Libraries :: Python Modules', ] )
c55b69aec422a28429f759893bfc1e1bd7352a59
setup.py
setup.py
#!/usr/bin/env python import sys from setuptools import setup, find_packages import versioneer versioneer.versionfile_source = 'obfsproxy/_version.py' versioneer.versionfile_build = 'obfsproxy/_version.py' versioneer.tag_prefix = 'obfsproxy-' # tags are like 1.2.0 versioneer.parentdir_prefix = 'obfsproxy-' # dirname like 'myproject-1.2.0' setup( name = "obfsproxy", author = "asn", author_email = "asn@torproject.org", description = ("A pluggable transport proxy written in Python"), license = "BSD", keywords = ['tor', 'obfuscation', 'twisted'], version=versioneer.get_version(), cmdclass=versioneer.get_cmdclass(), packages = find_packages(), entry_points = { 'console_scripts': [ 'obfsproxy = obfsproxy.pyobfsproxy:run' ] }, install_requires = [ 'setuptools', 'PyCrypto', 'Twisted', 'argparse', 'pyptlib >= 0.0.5', 'gmpy' ], )
#!/usr/bin/env python import sys from setuptools import setup, find_packages import versioneer versioneer.versionfile_source = 'obfsproxy/_version.py' versioneer.versionfile_build = 'obfsproxy/_version.py' versioneer.tag_prefix = 'obfsproxy-' # tags are like 1.2.0 versioneer.parentdir_prefix = 'obfsproxy-' # dirname like 'myproject-1.2.0' setup( name = "obfsproxy", author = "asn", author_email = "asn@torproject.org", description = ("A pluggable transport proxy written in Python"), license = "BSD", keywords = ['tor', 'obfuscation', 'twisted'], version=versioneer.get_version(), cmdclass=versioneer.get_cmdclass(), packages = find_packages(), entry_points = { 'console_scripts': [ 'obfsproxy = obfsproxy.pyobfsproxy:run' ] }, install_requires = [ 'setuptools', 'PyCrypto', 'Twisted', 'argparse', 'pyptlib >= 0.0.5', 'gmpy', 'pyyaml' ], )
Add "pyyaml" because it is used by ScrambleSuit.
Add "pyyaml" because it is used by ScrambleSuit.
Python
bsd-3-clause
isislovecruft/obfsproxy,masterkorp/obfsproxy,qdzheng/obfsproxy,infinity0/obfsproxy,catinred2/obfsproxy,Yawning/obfsproxy,david415/obfsproxy,Yawning/obfsproxy-wfpadtools,NullHypothesis/obfsproxy
#!/usr/bin/env python import sys from setuptools import setup, find_packages import versioneer versioneer.versionfile_source = 'obfsproxy/_version.py' versioneer.versionfile_build = 'obfsproxy/_version.py' versioneer.tag_prefix = 'obfsproxy-' # tags are like 1.2.0 versioneer.parentdir_prefix = 'obfsproxy-' # dirname like 'myproject-1.2.0' setup( name = "obfsproxy", author = "asn", author_email = "asn@torproject.org", description = ("A pluggable transport proxy written in Python"), license = "BSD", keywords = ['tor', 'obfuscation', 'twisted'], version=versioneer.get_version(), cmdclass=versioneer.get_cmdclass(), packages = find_packages(), entry_points = { 'console_scripts': [ 'obfsproxy = obfsproxy.pyobfsproxy:run' ] }, install_requires = [ 'setuptools', 'PyCrypto', 'Twisted', 'argparse', 'pyptlib >= 0.0.5', 'gmpy' ], ) Add "pyyaml" because it is used by ScrambleSuit.
#!/usr/bin/env python import sys from setuptools import setup, find_packages import versioneer versioneer.versionfile_source = 'obfsproxy/_version.py' versioneer.versionfile_build = 'obfsproxy/_version.py' versioneer.tag_prefix = 'obfsproxy-' # tags are like 1.2.0 versioneer.parentdir_prefix = 'obfsproxy-' # dirname like 'myproject-1.2.0' setup( name = "obfsproxy", author = "asn", author_email = "asn@torproject.org", description = ("A pluggable transport proxy written in Python"), license = "BSD", keywords = ['tor', 'obfuscation', 'twisted'], version=versioneer.get_version(), cmdclass=versioneer.get_cmdclass(), packages = find_packages(), entry_points = { 'console_scripts': [ 'obfsproxy = obfsproxy.pyobfsproxy:run' ] }, install_requires = [ 'setuptools', 'PyCrypto', 'Twisted', 'argparse', 'pyptlib >= 0.0.5', 'gmpy', 'pyyaml' ], )
<commit_before>#!/usr/bin/env python import sys from setuptools import setup, find_packages import versioneer versioneer.versionfile_source = 'obfsproxy/_version.py' versioneer.versionfile_build = 'obfsproxy/_version.py' versioneer.tag_prefix = 'obfsproxy-' # tags are like 1.2.0 versioneer.parentdir_prefix = 'obfsproxy-' # dirname like 'myproject-1.2.0' setup( name = "obfsproxy", author = "asn", author_email = "asn@torproject.org", description = ("A pluggable transport proxy written in Python"), license = "BSD", keywords = ['tor', 'obfuscation', 'twisted'], version=versioneer.get_version(), cmdclass=versioneer.get_cmdclass(), packages = find_packages(), entry_points = { 'console_scripts': [ 'obfsproxy = obfsproxy.pyobfsproxy:run' ] }, install_requires = [ 'setuptools', 'PyCrypto', 'Twisted', 'argparse', 'pyptlib >= 0.0.5', 'gmpy' ], ) <commit_msg>Add "pyyaml" because it is used by ScrambleSuit.<commit_after>
#!/usr/bin/env python import sys from setuptools import setup, find_packages import versioneer versioneer.versionfile_source = 'obfsproxy/_version.py' versioneer.versionfile_build = 'obfsproxy/_version.py' versioneer.tag_prefix = 'obfsproxy-' # tags are like 1.2.0 versioneer.parentdir_prefix = 'obfsproxy-' # dirname like 'myproject-1.2.0' setup( name = "obfsproxy", author = "asn", author_email = "asn@torproject.org", description = ("A pluggable transport proxy written in Python"), license = "BSD", keywords = ['tor', 'obfuscation', 'twisted'], version=versioneer.get_version(), cmdclass=versioneer.get_cmdclass(), packages = find_packages(), entry_points = { 'console_scripts': [ 'obfsproxy = obfsproxy.pyobfsproxy:run' ] }, install_requires = [ 'setuptools', 'PyCrypto', 'Twisted', 'argparse', 'pyptlib >= 0.0.5', 'gmpy', 'pyyaml' ], )
#!/usr/bin/env python import sys from setuptools import setup, find_packages import versioneer versioneer.versionfile_source = 'obfsproxy/_version.py' versioneer.versionfile_build = 'obfsproxy/_version.py' versioneer.tag_prefix = 'obfsproxy-' # tags are like 1.2.0 versioneer.parentdir_prefix = 'obfsproxy-' # dirname like 'myproject-1.2.0' setup( name = "obfsproxy", author = "asn", author_email = "asn@torproject.org", description = ("A pluggable transport proxy written in Python"), license = "BSD", keywords = ['tor', 'obfuscation', 'twisted'], version=versioneer.get_version(), cmdclass=versioneer.get_cmdclass(), packages = find_packages(), entry_points = { 'console_scripts': [ 'obfsproxy = obfsproxy.pyobfsproxy:run' ] }, install_requires = [ 'setuptools', 'PyCrypto', 'Twisted', 'argparse', 'pyptlib >= 0.0.5', 'gmpy' ], ) Add "pyyaml" because it is used by ScrambleSuit.#!/usr/bin/env python import sys from setuptools import setup, find_packages import versioneer versioneer.versionfile_source = 'obfsproxy/_version.py' versioneer.versionfile_build = 'obfsproxy/_version.py' versioneer.tag_prefix = 'obfsproxy-' # tags are like 1.2.0 versioneer.parentdir_prefix = 'obfsproxy-' # dirname like 'myproject-1.2.0' setup( name = "obfsproxy", author = "asn", author_email = "asn@torproject.org", description = ("A pluggable transport proxy written in Python"), license = "BSD", keywords = ['tor', 'obfuscation', 'twisted'], version=versioneer.get_version(), cmdclass=versioneer.get_cmdclass(), packages = find_packages(), entry_points = { 'console_scripts': [ 'obfsproxy = obfsproxy.pyobfsproxy:run' ] }, install_requires = [ 'setuptools', 'PyCrypto', 'Twisted', 'argparse', 'pyptlib >= 0.0.5', 'gmpy', 'pyyaml' ], )
<commit_before>#!/usr/bin/env python import sys from setuptools import setup, find_packages import versioneer versioneer.versionfile_source = 'obfsproxy/_version.py' versioneer.versionfile_build = 'obfsproxy/_version.py' versioneer.tag_prefix = 'obfsproxy-' # tags are like 1.2.0 versioneer.parentdir_prefix = 'obfsproxy-' # dirname like 'myproject-1.2.0' setup( name = "obfsproxy", author = "asn", author_email = "asn@torproject.org", description = ("A pluggable transport proxy written in Python"), license = "BSD", keywords = ['tor', 'obfuscation', 'twisted'], version=versioneer.get_version(), cmdclass=versioneer.get_cmdclass(), packages = find_packages(), entry_points = { 'console_scripts': [ 'obfsproxy = obfsproxy.pyobfsproxy:run' ] }, install_requires = [ 'setuptools', 'PyCrypto', 'Twisted', 'argparse', 'pyptlib >= 0.0.5', 'gmpy' ], ) <commit_msg>Add "pyyaml" because it is used by ScrambleSuit.<commit_after>#!/usr/bin/env python import sys from setuptools import setup, find_packages import versioneer versioneer.versionfile_source = 'obfsproxy/_version.py' versioneer.versionfile_build = 'obfsproxy/_version.py' versioneer.tag_prefix = 'obfsproxy-' # tags are like 1.2.0 versioneer.parentdir_prefix = 'obfsproxy-' # dirname like 'myproject-1.2.0' setup( name = "obfsproxy", author = "asn", author_email = "asn@torproject.org", description = ("A pluggable transport proxy written in Python"), license = "BSD", keywords = ['tor', 'obfuscation', 'twisted'], version=versioneer.get_version(), cmdclass=versioneer.get_cmdclass(), packages = find_packages(), entry_points = { 'console_scripts': [ 'obfsproxy = obfsproxy.pyobfsproxy:run' ] }, install_requires = [ 'setuptools', 'PyCrypto', 'Twisted', 'argparse', 'pyptlib >= 0.0.5', 'gmpy', 'pyyaml' ], )
5b7685131030249cd1a6213eca8134e782e8d12d
setup.py
setup.py
import sys try: from setuptools import setup except ImportError: from distutils.core import setup requirements = [req.strip() for req in open('requirements.pip')] setup( name = 'leaderboard', version = "2.0.0", author = 'David Czarnecki', author_email = "dczarnecki@agoragames.com", packages = ['leaderboard'], install_requires = requirements, url = 'https://github.com/agoragames/leaderboard-python', license = "LICENSE.txt", description = 'Leaderboards backed by Redis in Python', long_description = open('README.md').read(), keywords = ['python', 'redis', 'leaderboard'], classifiers = [ 'Development Status :: 1 - Production', 'License :: OSI Approved :: MIT License', "Intended Audience :: Developers", "Operating System :: POSIX", "Topic :: Communications", "Topic :: System :: Distributed Computing", "Topic :: Software Development :: Libraries :: Python Modules", 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries' ] )
import sys try: from setuptools import setup except ImportError: from distutils.core import setup requirements = [req.strip() for req in open('requirements.pip')] setup( name = 'leaderboard', version = "2.0.0", author = 'David Czarnecki', author_email = "dczarnecki@agoragames.com", packages = ['leaderboard'], install_requires = requirements, url = 'https://github.com/agoragames/leaderboard-python', license = "LICENSE.txt", description = 'Leaderboards backed by Redis in Python', long_description = open('README.md').read(), keywords = ['python', 'redis', 'leaderboard'], classifiers = [ 'Development Status :: 5 - Production/Stable', 'License :: OSI Approved :: MIT License', "Intended Audience :: Developers", "Operating System :: POSIX", "Topic :: Communications", "Topic :: System :: Distributed Computing", "Topic :: Software Development :: Libraries :: Python Modules", 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries' ] )
Use the correct classifier for production
Use the correct classifier for production
Python
mit
agoragames/leaderboard-python,SixMinute/leaderboard-python
import sys try: from setuptools import setup except ImportError: from distutils.core import setup requirements = [req.strip() for req in open('requirements.pip')] setup( name = 'leaderboard', version = "2.0.0", author = 'David Czarnecki', author_email = "dczarnecki@agoragames.com", packages = ['leaderboard'], install_requires = requirements, url = 'https://github.com/agoragames/leaderboard-python', license = "LICENSE.txt", description = 'Leaderboards backed by Redis in Python', long_description = open('README.md').read(), keywords = ['python', 'redis', 'leaderboard'], classifiers = [ 'Development Status :: 1 - Production', 'License :: OSI Approved :: MIT License', "Intended Audience :: Developers", "Operating System :: POSIX", "Topic :: Communications", "Topic :: System :: Distributed Computing", "Topic :: Software Development :: Libraries :: Python Modules", 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries' ] ) Use the correct classifier for production
import sys try: from setuptools import setup except ImportError: from distutils.core import setup requirements = [req.strip() for req in open('requirements.pip')] setup( name = 'leaderboard', version = "2.0.0", author = 'David Czarnecki', author_email = "dczarnecki@agoragames.com", packages = ['leaderboard'], install_requires = requirements, url = 'https://github.com/agoragames/leaderboard-python', license = "LICENSE.txt", description = 'Leaderboards backed by Redis in Python', long_description = open('README.md').read(), keywords = ['python', 'redis', 'leaderboard'], classifiers = [ 'Development Status :: 5 - Production/Stable', 'License :: OSI Approved :: MIT License', "Intended Audience :: Developers", "Operating System :: POSIX", "Topic :: Communications", "Topic :: System :: Distributed Computing", "Topic :: Software Development :: Libraries :: Python Modules", 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries' ] )
<commit_before>import sys try: from setuptools import setup except ImportError: from distutils.core import setup requirements = [req.strip() for req in open('requirements.pip')] setup( name = 'leaderboard', version = "2.0.0", author = 'David Czarnecki', author_email = "dczarnecki@agoragames.com", packages = ['leaderboard'], install_requires = requirements, url = 'https://github.com/agoragames/leaderboard-python', license = "LICENSE.txt", description = 'Leaderboards backed by Redis in Python', long_description = open('README.md').read(), keywords = ['python', 'redis', 'leaderboard'], classifiers = [ 'Development Status :: 1 - Production', 'License :: OSI Approved :: MIT License', "Intended Audience :: Developers", "Operating System :: POSIX", "Topic :: Communications", "Topic :: System :: Distributed Computing", "Topic :: Software Development :: Libraries :: Python Modules", 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries' ] ) <commit_msg>Use the correct classifier for production<commit_after>
import sys try: from setuptools import setup except ImportError: from distutils.core import setup requirements = [req.strip() for req in open('requirements.pip')] setup( name = 'leaderboard', version = "2.0.0", author = 'David Czarnecki', author_email = "dczarnecki@agoragames.com", packages = ['leaderboard'], install_requires = requirements, url = 'https://github.com/agoragames/leaderboard-python', license = "LICENSE.txt", description = 'Leaderboards backed by Redis in Python', long_description = open('README.md').read(), keywords = ['python', 'redis', 'leaderboard'], classifiers = [ 'Development Status :: 5 - Production/Stable', 'License :: OSI Approved :: MIT License', "Intended Audience :: Developers", "Operating System :: POSIX", "Topic :: Communications", "Topic :: System :: Distributed Computing", "Topic :: Software Development :: Libraries :: Python Modules", 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries' ] )
import sys try: from setuptools import setup except ImportError: from distutils.core import setup requirements = [req.strip() for req in open('requirements.pip')] setup( name = 'leaderboard', version = "2.0.0", author = 'David Czarnecki', author_email = "dczarnecki@agoragames.com", packages = ['leaderboard'], install_requires = requirements, url = 'https://github.com/agoragames/leaderboard-python', license = "LICENSE.txt", description = 'Leaderboards backed by Redis in Python', long_description = open('README.md').read(), keywords = ['python', 'redis', 'leaderboard'], classifiers = [ 'Development Status :: 1 - Production', 'License :: OSI Approved :: MIT License', "Intended Audience :: Developers", "Operating System :: POSIX", "Topic :: Communications", "Topic :: System :: Distributed Computing", "Topic :: Software Development :: Libraries :: Python Modules", 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries' ] ) Use the correct classifier for productionimport sys try: from setuptools import setup except ImportError: from distutils.core import setup requirements = [req.strip() for req in open('requirements.pip')] setup( name = 'leaderboard', version = "2.0.0", author = 'David Czarnecki', author_email = "dczarnecki@agoragames.com", packages = ['leaderboard'], install_requires = requirements, url = 'https://github.com/agoragames/leaderboard-python', license = "LICENSE.txt", description = 'Leaderboards backed by Redis in Python', long_description = open('README.md').read(), keywords = ['python', 'redis', 'leaderboard'], classifiers = [ 'Development Status :: 5 - Production/Stable', 'License :: OSI Approved :: MIT License', "Intended Audience :: Developers", "Operating System :: POSIX", "Topic :: Communications", "Topic :: System :: Distributed Computing", "Topic :: Software Development :: Libraries :: Python Modules", 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries' ] )
<commit_before>import sys try: from setuptools import setup except ImportError: from distutils.core import setup requirements = [req.strip() for req in open('requirements.pip')] setup( name = 'leaderboard', version = "2.0.0", author = 'David Czarnecki', author_email = "dczarnecki@agoragames.com", packages = ['leaderboard'], install_requires = requirements, url = 'https://github.com/agoragames/leaderboard-python', license = "LICENSE.txt", description = 'Leaderboards backed by Redis in Python', long_description = open('README.md').read(), keywords = ['python', 'redis', 'leaderboard'], classifiers = [ 'Development Status :: 1 - Production', 'License :: OSI Approved :: MIT License', "Intended Audience :: Developers", "Operating System :: POSIX", "Topic :: Communications", "Topic :: System :: Distributed Computing", "Topic :: Software Development :: Libraries :: Python Modules", 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries' ] ) <commit_msg>Use the correct classifier for production<commit_after>import sys try: from setuptools import setup except ImportError: from distutils.core import setup requirements = [req.strip() for req in open('requirements.pip')] setup( name = 'leaderboard', version = "2.0.0", author = 'David Czarnecki', author_email = "dczarnecki@agoragames.com", packages = ['leaderboard'], install_requires = requirements, url = 'https://github.com/agoragames/leaderboard-python', license = "LICENSE.txt", description = 'Leaderboards backed by Redis in Python', long_description = open('README.md').read(), keywords = ['python', 'redis', 'leaderboard'], classifiers = [ 'Development Status :: 5 - Production/Stable', 'License :: OSI Approved :: MIT License', "Intended Audience :: Developers", "Operating System :: POSIX", "Topic :: Communications", "Topic :: System :: Distributed Computing", "Topic :: Software Development :: Libraries :: Python Modules", 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries' ] )
e889c0420a0770d2be633af5ad2b20fb27e0d05c
setup.py
setup.py
import codecs from setuptools import find_packages, setup import digestive setup( name='digestive', version=digestive.__version__, url='https://github.com/akaIDIOT/Digestive', packages=find_packages(), description='TODO', author='Mattijs Ugen', author_email=codecs.encode('nxnvqvbg@hfref.abercyl.tvguho.pbz', 'rot_13'), license='ISC', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'License :: OSI Approved :: ISC License (ISCL)', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], tests_require=['pytest'], scripts=['bin/digestive'], # TODO: turn this into an entry_points= )
import codecs from setuptools import find_packages, setup import digestive setup( name='digestive', version=digestive.__version__, url='https://github.com/akaIDIOT/Digestive', packages=find_packages(), description='TODO', author='Mattijs Ugen', author_email=codecs.encode('nxnvqvbg@hfref.abercyl.tvguho.pbz', 'rot_13'), license='ISC', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'License :: OSI Approved :: ISC License (ISCL)', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], tests_require=['pytest'], entry_points={ 'console_scripts': { 'digestive = digestive.main:main' } } )
Use console_scripts entry_point instead of scripts
Use console_scripts entry_point instead of scripts
Python
isc
akaIDIOT/Digestive
import codecs from setuptools import find_packages, setup import digestive setup( name='digestive', version=digestive.__version__, url='https://github.com/akaIDIOT/Digestive', packages=find_packages(), description='TODO', author='Mattijs Ugen', author_email=codecs.encode('nxnvqvbg@hfref.abercyl.tvguho.pbz', 'rot_13'), license='ISC', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'License :: OSI Approved :: ISC License (ISCL)', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], tests_require=['pytest'], scripts=['bin/digestive'], # TODO: turn this into an entry_points= ) Use console_scripts entry_point instead of scripts
import codecs from setuptools import find_packages, setup import digestive setup( name='digestive', version=digestive.__version__, url='https://github.com/akaIDIOT/Digestive', packages=find_packages(), description='TODO', author='Mattijs Ugen', author_email=codecs.encode('nxnvqvbg@hfref.abercyl.tvguho.pbz', 'rot_13'), license='ISC', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'License :: OSI Approved :: ISC License (ISCL)', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], tests_require=['pytest'], entry_points={ 'console_scripts': { 'digestive = digestive.main:main' } } )
<commit_before>import codecs from setuptools import find_packages, setup import digestive setup( name='digestive', version=digestive.__version__, url='https://github.com/akaIDIOT/Digestive', packages=find_packages(), description='TODO', author='Mattijs Ugen', author_email=codecs.encode('nxnvqvbg@hfref.abercyl.tvguho.pbz', 'rot_13'), license='ISC', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'License :: OSI Approved :: ISC License (ISCL)', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], tests_require=['pytest'], scripts=['bin/digestive'], # TODO: turn this into an entry_points= ) <commit_msg>Use console_scripts entry_point instead of scripts<commit_after>
import codecs from setuptools import find_packages, setup import digestive setup( name='digestive', version=digestive.__version__, url='https://github.com/akaIDIOT/Digestive', packages=find_packages(), description='TODO', author='Mattijs Ugen', author_email=codecs.encode('nxnvqvbg@hfref.abercyl.tvguho.pbz', 'rot_13'), license='ISC', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'License :: OSI Approved :: ISC License (ISCL)', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], tests_require=['pytest'], entry_points={ 'console_scripts': { 'digestive = digestive.main:main' } } )
import codecs from setuptools import find_packages, setup import digestive setup( name='digestive', version=digestive.__version__, url='https://github.com/akaIDIOT/Digestive', packages=find_packages(), description='TODO', author='Mattijs Ugen', author_email=codecs.encode('nxnvqvbg@hfref.abercyl.tvguho.pbz', 'rot_13'), license='ISC', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'License :: OSI Approved :: ISC License (ISCL)', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], tests_require=['pytest'], scripts=['bin/digestive'], # TODO: turn this into an entry_points= ) Use console_scripts entry_point instead of scriptsimport codecs from setuptools import find_packages, setup import digestive setup( name='digestive', version=digestive.__version__, url='https://github.com/akaIDIOT/Digestive', packages=find_packages(), description='TODO', author='Mattijs Ugen', author_email=codecs.encode('nxnvqvbg@hfref.abercyl.tvguho.pbz', 'rot_13'), license='ISC', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'License :: OSI Approved :: ISC License (ISCL)', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], tests_require=['pytest'], entry_points={ 'console_scripts': { 'digestive = digestive.main:main' } } )
<commit_before>import codecs from setuptools import find_packages, setup import digestive setup( name='digestive', version=digestive.__version__, url='https://github.com/akaIDIOT/Digestive', packages=find_packages(), description='TODO', author='Mattijs Ugen', author_email=codecs.encode('nxnvqvbg@hfref.abercyl.tvguho.pbz', 'rot_13'), license='ISC', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'License :: OSI Approved :: ISC License (ISCL)', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], tests_require=['pytest'], scripts=['bin/digestive'], # TODO: turn this into an entry_points= ) <commit_msg>Use console_scripts entry_point instead of scripts<commit_after>import codecs from setuptools import find_packages, setup import digestive setup( name='digestive', version=digestive.__version__, url='https://github.com/akaIDIOT/Digestive', packages=find_packages(), description='TODO', author='Mattijs Ugen', author_email=codecs.encode('nxnvqvbg@hfref.abercyl.tvguho.pbz', 'rot_13'), license='ISC', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'License :: OSI Approved :: ISC License (ISCL)', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], tests_require=['pytest'], entry_points={ 'console_scripts': { 'digestive = digestive.main:main' } } )