commit stringlengths 40 40 | old_file stringlengths 4 150 | new_file stringlengths 4 150 | old_contents stringlengths 0 3.26k | new_contents stringlengths 1 4.43k | subject stringlengths 15 501 | message stringlengths 15 4.06k | lang stringclasses 4 values | license stringclasses 13 values | repos stringlengths 5 91.5k | diff stringlengths 0 4.35k |
|---|---|---|---|---|---|---|---|---|---|---|
a557bae533debc43053294fdb60f774ecb4f2a4e | corehq/blobs/migrations/0003_auto_20161012_1358.py | corehq/blobs/migrations/0003_auto_20161012_1358.py | # -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
from corehq.blobs.migrate import assert_migration_complete
from corehq.sql_db.operations import HqRunPython
class Migration(migrations.Migration):
dependencies = [
('blobs', '0002_auto_20151221_1623'),
]
operations = [
HqRunPython(*assert_migration_complete("applications"))
HqRunPython(*assert_migration_complete("multimedia"))
HqRunPython(*assert_migration_complete("xforms"))
]
| # -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
from corehq.blobs.migrate import assert_migration_complete
from corehq.sql_db.operations import HqRunPython
class Migration(migrations.Migration):
dependencies = [
('blobs', '0002_auto_20151221_1623'),
]
operations = [
HqRunPython(*assert_migration_complete("applications")),
HqRunPython(*assert_migration_complete("multimedia")),
HqRunPython(*assert_migration_complete("xforms")),
]
| Fix bad syntax in migration | Fix bad syntax in migration | Python | bsd-3-clause | qedsoftware/commcare-hq,dimagi/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq | ---
+++
@@ -14,7 +14,7 @@
]
operations = [
- HqRunPython(*assert_migration_complete("applications"))
- HqRunPython(*assert_migration_complete("multimedia"))
- HqRunPython(*assert_migration_complete("xforms"))
+ HqRunPython(*assert_migration_complete("applications")),
+ HqRunPython(*assert_migration_complete("multimedia")),
+ HqRunPython(*assert_migration_complete("xforms")),
] |
f83011c7b5fdb7c1865d04b03e44660851fc8dcd | test/handler_fixture.py | test/handler_fixture.py | import unittest
import tempfile
import shutil
import uuid
import groundstation.node
import groundstation.transfer.response
import groundstation.transfer.request
from groundstation.station import Station
class MockStream(list):
def enqueue(self, *args, **kwargs):
self.append(*args, **kwargs)
def MockTERMINATE():
pass
class MockStation(object):
def __init__(self, **kwargs):
self.tmpdir = tempfile.mkdtemp()
self.node = groundstation.node.Node()
self.station = Station(self.tmpdir, self.node)
self.stream = MockStream()
self.TERMINATE = MockTERMINATE
self.id = "test_station"
if 'origin' in kwargs:
self.origin = kwargs['origin']
else:
self.origin = uuid.uuid1()
def _Response(self, *args, **kwargs):
kwargs['station'] = self.station
return groundstation.transfer.response.Response(*args, **kwargs)
def _Request(self, *args, **kwargs):
kwargs['station'] = self.station
return groundstation.transfer.request.Request(*args, **kwargs)
def __del__(self):
shutil.rmtree(self.tmpdir)
class StationHandlerTestCase(unittest.TestCase):
def setUp(self):
self.station = MockStation()
| import unittest
import tempfile
import shutil
import uuid
import groundstation.node
import groundstation.transfer.response
import groundstation.transfer.request
from groundstation.station import Station
class MockStream(list):
def enqueue(self, *args, **kwargs):
self.append(*args, **kwargs)
def MockTERMINATE():
pass
class MockStation(object):
def __init__(self, **kwargs):
self.tmpdir = tempfile.mkdtemp()
self.node = groundstation.node.Node()
self.station = Station(self.tmpdir, self.node)
self.stream = MockStream()
self.TERMINATE = MockTERMINATE
self.id = "test_station"
if 'origin' in kwargs:
self.origin = kwargs['origin']
else:
self.origin = uuid.uuid1()
def _Response(self, *args, **kwargs):
kwargs['station'] = self.station
return groundstation.transfer.response.Response(*args, **kwargs)
def _Request(self, *args, **kwargs):
kwargs['station'] = self.station
return groundstation.transfer.request.Request(*args, **kwargs)
def __del__(self):
shutil.rmtree(self.tmpdir)
def teardown(self):
pass
class StationHandlerTestCase(unittest.TestCase):
def setUp(self):
self.station = MockStation()
| Implement teardown on Station mock | Implement teardown on Station mock
| Python | mit | richo/groundstation,richo/groundstation,richo/groundstation,richo/groundstation,richo/groundstation | ---
+++
@@ -43,6 +43,9 @@
def __del__(self):
shutil.rmtree(self.tmpdir)
+ def teardown(self):
+ pass
+
class StationHandlerTestCase(unittest.TestCase):
def setUp(self): |
d8c0f7cd89ec52d9ad7553a27421c899b3a94417 | dbaas/workflow/workflow.py | dbaas/workflow/workflow.py | # -*- coding: utf-8 -*-
from django.utils.module_loading import import_by_path
from notification.models import TaskHistory
import logging
LOG = logging.getLogger(__name__)
def start_workflow(workflow_dict, task=None):
try :
if not 'steps' in workflow_dict:
return False
workflow_dict['step_counter'] = 0
for step in workflow_dict['steps']:
workflow_dict['step_counter']+=1
LOG.info("Step %s number %i" % (step, workflow_dict['step_counter']))
my_class = import_by_path(step)
my_instance = my_class()
if task:
task.update_details(persist=True, details=str(my_instance))
if my_instance.do(workflow_dict)!=True:
raise Exception
except Exception, e:
print e
workflow_dict['steps'] = workflow_dict['steps'][:workflow_dict['step_counter']]
stop_workflow(workflow_dict)
def stop_workflow(workflow_dict):
for step in workflow_dict['steps'][::-1]:
my_class = import_by_path(step)
my_instance = my_class()
my_instance.undo(workflow_dict)
| # -*- coding: utf-8 -*-
from django.utils.module_loading import import_by_path
import logging
LOG = logging.getLogger(__name__)
def start_workflow(workflow_dict, task=None):
try:
if not 'steps' in workflow_dict:
return False
workflow_dict['step_counter'] = 0
for step in workflow_dict['steps']:
workflow_dict['step_counter'] += 1
my_class = import_by_path(step)
my_instance = my_class()
LOG.info("Step %i %s " %
(workflow_dict['step_counter'], str(my_instance)))
if task:
task.update_details(persist=True, details=str(my_instance))
if my_instance.do(workflow_dict) != True:
raise Exception
except Exception, e:
print e
workflow_dict['steps'] = workflow_dict[
'steps'][:workflow_dict['step_counter']]
stop_workflow(workflow_dict)
def stop_workflow(workflow_dict):
LOG.info("Running undo...")
for step in workflow_dict['steps'][::-1]:
workflow_dict['step_counter'] -= 1
my_class = import_by_path(step)
my_instance = my_class()
LOG.info("Step %i %s " %
(workflow_dict['step_counter'], str(my_instance)))
my_instance.undo(workflow_dict)
| Improve logs and refactor to pep8 | Improve logs and refactor to pep8
| Python | bsd-3-clause | globocom/database-as-a-service,globocom/database-as-a-service,globocom/database-as-a-service,globocom/database-as-a-service | ---
+++
@@ -1,39 +1,46 @@
# -*- coding: utf-8 -*-
from django.utils.module_loading import import_by_path
-from notification.models import TaskHistory
import logging
LOG = logging.getLogger(__name__)
+
def start_workflow(workflow_dict, task=None):
- try :
- if not 'steps' in workflow_dict:
- return False
- workflow_dict['step_counter'] = 0
+ try:
+ if not 'steps' in workflow_dict:
+ return False
+ workflow_dict['step_counter'] = 0
- for step in workflow_dict['steps']:
- workflow_dict['step_counter']+=1
+ for step in workflow_dict['steps']:
+ workflow_dict['step_counter'] += 1
- LOG.info("Step %s number %i" % (step, workflow_dict['step_counter']))
+ my_class = import_by_path(step)
+ my_instance = my_class()
- my_class = import_by_path(step)
- my_instance = my_class()
+ LOG.info("Step %i %s " %
+ (workflow_dict['step_counter'], str(my_instance)))
- if task:
- task.update_details(persist=True, details=str(my_instance))
+ if task:
+ task.update_details(persist=True, details=str(my_instance))
- if my_instance.do(workflow_dict)!=True:
- raise Exception
+ if my_instance.do(workflow_dict) != True:
+ raise Exception
- except Exception, e:
- print e
- workflow_dict['steps'] = workflow_dict['steps'][:workflow_dict['step_counter']]
- stop_workflow(workflow_dict)
+ except Exception, e:
+ print e
+ workflow_dict['steps'] = workflow_dict[
+ 'steps'][:workflow_dict['step_counter']]
+ stop_workflow(workflow_dict)
def stop_workflow(workflow_dict):
+ LOG.info("Running undo...")
- for step in workflow_dict['steps'][::-1]:
- my_class = import_by_path(step)
- my_instance = my_class()
- my_instance.undo(workflow_dict)
+ for step in workflow_dict['steps'][::-1]:
+ workflow_dict['step_counter'] -= 1
+ my_class = import_by_path(step)
+ my_instance = my_class()
+
+ LOG.info("Step %i %s " %
+ (workflow_dict['step_counter'], str(my_instance)))
+ my_instance.undo(workflow_dict) |
adc78d6c23b15bc06d16c7365b1875ebe39dc328 | jukebox/scanner.py | jukebox/scanner.py | import os
import zope.interface
import mutagen
import jukebox.song
class IScanner(zope.interface.Interface):
def scan():
"""
Start the scanning process. It is expected to be called at startup and
can block.
"""
def watch():
"""
Starts an async watcher that can add files to the IStorage
"""
def stop():
"""
Stops the async watcher
"""
class DirScanner(object):
zope.interface.implements(IScanner)
def __init__(self, storage, path):
self.storage = storage
self.path = path
def scan(self):
for root, dirs, files in os.walk(self.path):
for name in files:
path = os.path.join(root, name)
try:
music_file = mutagen.File(path, easy=True)
except:
continue
if not music_file:
continue
self.storage.add_song(jukebox.song.Song(
title=(music_file['title'] + [None])[0],
album=(music_file['album'] + [None])[0],
artist=(music_file['artist'] + [None])[0],
path=path,
))
def watch(self):
pass
def stop(self):
pass
| import os
import zope.interface
import mutagen
import jukebox.song
class IScanner(zope.interface.Interface):
def scan():
"""
Start the scanning process. It is expected to be called at startup and
can block.
"""
def watch():
"""
Starts an async watcher that can add files to the IStorage
"""
def stop():
"""
Stops the async watcher
"""
class DirScanner(object):
zope.interface.implements(IScanner)
def __init__(self, storage, *paths):
self.storage = storage
self.paths = paths
def _scan_path(self, path):
print path
for root, dirs, files in os.walk(path):
for name in files:
file_path = os.path.join(root, name)
try:
music_file = mutagen.File(file_path, easy=True)
except:
continue
if not music_file:
continue
try:
title = music_file['title'][0]
except KeyError:
title = None
try:
album = music_file['album'][0]
except KeyError:
album = None
try:
artist = music_file['artist'][0]
except KeyError:
artist = None
self.storage.add_song(jukebox.song.Song(
title=title,
album=album,
artist=artist,
path=file_path,
))
def scan(self):
for path in self.paths:
self._scan_path(path)
def watch(self):
pass
def stop(self):
pass
| Support more than on path when building a DirScanner | Support more than on path when building a DirScanner
| Python | mit | armooo/jukebox,armooo/jukebox | ---
+++
@@ -26,26 +26,45 @@
class DirScanner(object):
zope.interface.implements(IScanner)
- def __init__(self, storage, path):
+ def __init__(self, storage, *paths):
self.storage = storage
- self.path = path
+ self.paths = paths
- def scan(self):
- for root, dirs, files in os.walk(self.path):
+ def _scan_path(self, path):
+ print path
+ for root, dirs, files in os.walk(path):
for name in files:
- path = os.path.join(root, name)
+ file_path = os.path.join(root, name)
try:
- music_file = mutagen.File(path, easy=True)
+ music_file = mutagen.File(file_path, easy=True)
except:
continue
if not music_file:
continue
+
+ try:
+ title = music_file['title'][0]
+ except KeyError:
+ title = None
+ try:
+ album = music_file['album'][0]
+ except KeyError:
+ album = None
+ try:
+ artist = music_file['artist'][0]
+ except KeyError:
+ artist = None
+
self.storage.add_song(jukebox.song.Song(
- title=(music_file['title'] + [None])[0],
- album=(music_file['album'] + [None])[0],
- artist=(music_file['artist'] + [None])[0],
- path=path,
+ title=title,
+ album=album,
+ artist=artist,
+ path=file_path,
))
+
+ def scan(self):
+ for path in self.paths:
+ self._scan_path(path)
def watch(self):
pass |
949f629b349707c2f4caf0a288969b5a6143a730 | kyokai/response.py | kyokai/response.py | """
Module for a Response object.
A Response is returned by Routes when the underlying coroutine is done.
"""
from http_parser.util import IOrderedDict
from .util import HTTP_CODES
class Response(object):
"""
A response is responsible (no pun intended) for delivering data to the client, again.
The method `to_bytes()` transforms this into a bytes response.
"""
def __init__(self, code: int, body: str, headers: dict):
"""
Create a new response.
"""
self.code = code
self.body = str(body)
self.headers = IOrderedDict(headers)
def _recalculate_headers(self):
"""
Override certain headers, like Content-Size.
"""
self.headers["Content-Length"] = len(self.body)
def to_bytes(self):
"""
Return the correct response.
"""
self._recalculate_headers()
fmt = "HTTP/1.1 {code} {msg}\r\n{headers}\r\n\r\n{body}\r\n"
headers_fmt = ""
# Calculate headers
for name, val in self.headers.items():
headers_fmt += "{}: {}".format(name, val)
built = fmt.format(code=self.code, msg=HTTP_CODES.get(self.code, "Unknown"), headers=headers_fmt,
body=self.body)
return built.encode()
| """
Module for a Response object.
A Response is returned by Routes when the underlying coroutine is done.
"""
from http_parser.util import IOrderedDict
from .util import HTTP_CODES
class Response(object):
"""
A response is responsible (no pun intended) for delivering data to the client, again.
The method `to_bytes()` transforms this into a bytes response.
"""
def __init__(self, code: int, body: str, headers: dict):
"""
Create a new response.
"""
self.code = code
self.body = str(body)
self.headers = IOrderedDict(headers)
def _recalculate_headers(self):
"""
Override certain headers, like Content-Size.
"""
self.headers["Content-Length"] = len(self.body)
if 'Content-Type' not in self.headers:
self.headers["Content-Type"] = "text/plain"
def to_bytes(self):
"""
Return the correct response.
"""
self._recalculate_headers()
fmt = "HTTP/1.1 {code} {msg}\r\n{headers}\r\n{body}\r\n"
headers_fmt = ""
# Calculate headers
for name, val in self.headers.items():
headers_fmt += "{}: {}\r\n".format(name, val)
built = fmt.format(code=self.code, msg=HTTP_CODES.get(self.code, "Unknown"), headers=headers_fmt,
body=self.body)
return built.encode()
| Add \r\n to the end of each header too | Add \r\n to the end of each header too
| Python | mit | SunDwarf/Kyoukai | ---
+++
@@ -28,17 +28,19 @@
Override certain headers, like Content-Size.
"""
self.headers["Content-Length"] = len(self.body)
+ if 'Content-Type' not in self.headers:
+ self.headers["Content-Type"] = "text/plain"
def to_bytes(self):
"""
Return the correct response.
"""
self._recalculate_headers()
- fmt = "HTTP/1.1 {code} {msg}\r\n{headers}\r\n\r\n{body}\r\n"
+ fmt = "HTTP/1.1 {code} {msg}\r\n{headers}\r\n{body}\r\n"
headers_fmt = ""
# Calculate headers
for name, val in self.headers.items():
- headers_fmt += "{}: {}".format(name, val)
+ headers_fmt += "{}: {}\r\n".format(name, val)
built = fmt.format(code=self.code, msg=HTTP_CODES.get(self.code, "Unknown"), headers=headers_fmt,
body=self.body)
|
e08a382e215569b2ad147ea82d7ede1319722724 | configurator/__init__.py | configurator/__init__.py | """Adaptive configuration dialogs.
Attributes:
__version__: The current version string.
"""
import os
import subprocess
def _get_version(version=None): # overwritten by setup.py
if version is None:
pkg_dir = os.path.dirname(__file__)
src_dir = os.path.abspath(os.path.join(pkg_dir, os.pardir))
git_dir = os.path.join(src_dir, ".git")
git_args = ("git", "--work-tree", src_dir, "--git-dir",
git_dir, "describe", "--tags")
output = subprocess.check_output(git_args)
version = output.decode("utf-8").strip()
if version.rfind("-") >= 0:
version = version[:version.rfind("-")] # strip SHA1 hash
version = version.replace("-", ".post") # PEP 440 compatible
return version
__version__ = _get_version()
| """Adaptive configuration dialogs.
Attributes:
__version__: The current version string.
"""
import os
import subprocess
def _get_version(version=None): # overwritten by setup.py
if version is None:
pkg_dir = os.path.dirname(__file__)
src_dir = os.path.abspath(os.path.join(pkg_dir, os.pardir))
git_dir = os.path.join(src_dir, ".git")
git_args = ("git", "--work-tree", src_dir, "--git-dir",
git_dir, "describe", "--tags", "--always")
output = subprocess.check_output(git_args)
version = output.decode("utf-8").strip()
if version.rfind("-") >= 0:
version = version[:version.rfind("-")] # strip SHA1 hash
version = version.replace("-", ".post") # PEP 440 compatible
return version
__version__ = _get_version()
| Use git describe --always in _get_version() | Use git describe --always in _get_version()
Travis CI does a truncated clone and causes 'git describe' to fail if
the version tag is not part of the truncated history.
| Python | apache-2.0 | yasserglez/configurator,yasserglez/configurator | ---
+++
@@ -14,7 +14,7 @@
src_dir = os.path.abspath(os.path.join(pkg_dir, os.pardir))
git_dir = os.path.join(src_dir, ".git")
git_args = ("git", "--work-tree", src_dir, "--git-dir",
- git_dir, "describe", "--tags")
+ git_dir, "describe", "--tags", "--always")
output = subprocess.check_output(git_args)
version = output.decode("utf-8").strip()
if version.rfind("-") >= 0: |
749181ac27583900af94feca848f3039d7c69bcc | app/backend/gwells/views/api.py | app/backend/gwells/views/api.py | from rest_framework.response import Response
from rest_framework.views import APIView
from gwells.settings.base import get_env_variable
class KeycloakConfig(APIView):
""" serves keycloak config """
def get(self, request):
config = {
"realm": get_env_variable("SSO_REALM"),
"auth-server-url": get_env_variable("SSO_AUTH_HOST"),
"ssl-required": "external",
"resource": get_env_variable("SSO_CLIENT"),
"public-client": True,
"confidential-port": int(get_env_variable("SSO_PORT", "0"))
}
return Response(config)
class GeneralConfig(APIView):
""" serves general configuration """
def get(self, request):
config = {
"enable_data_entry": get_env_variable("ENABLE_DATA_ENTRY") == "True",
"enable_google_analytics": get_env_variable("ENABLE_GOOGLE_ANALYTICS") == "True",
"sso_idp_hint": get_env_variable("SSO_IDP_HINT")
}
return Response(config)
| from rest_framework.response import Response
from rest_framework.views import APIView
from gwells.settings.base import get_env_variable
class KeycloakConfig(APIView):
""" serves keycloak config """
def get(self, request):
config = {
"realm": get_env_variable("SSO_REALM"),
"auth-server-url": get_env_variable("SSO_AUTH_HOST"),
"ssl-required": "external",
"resource": get_env_variable("SSO_CLIENT"),
"public-client": True,
"confidential-port": int(get_env_variable("SSO_PORT", "0"))
}
return Response(config)
class GeneralConfig(APIView):
""" serves general configuration """
def get(self, request):
config = {
"enable_data_entry": get_env_variable("ENABLE_DATA_ENTRY") == "True",
"enable_google_analytics": get_env_variable("ENABLE_GOOGLE_ANALYTICS") == "True",
"sso_idp_hint": get_env_variable("SSO_IDP_HINT") or "idir"
}
return Response(config)
| Add "idir" as default value. | Add "idir" as default value.
| Python | apache-2.0 | bcgov/gwells,bcgov/gwells,bcgov/gwells,bcgov/gwells | ---
+++
@@ -26,6 +26,6 @@
config = {
"enable_data_entry": get_env_variable("ENABLE_DATA_ENTRY") == "True",
"enable_google_analytics": get_env_variable("ENABLE_GOOGLE_ANALYTICS") == "True",
- "sso_idp_hint": get_env_variable("SSO_IDP_HINT")
+ "sso_idp_hint": get_env_variable("SSO_IDP_HINT") or "idir"
}
return Response(config) |
e4564e790dca3079e76540f452836353b902f5b6 | flexbe_core/src/flexbe_core/core/loopback_state.py | flexbe_core/src/flexbe_core/core/loopback_state.py | #!/usr/bin/env python
import rospy
from flexbe_core.core.lockable_state import LockableState
class LoopbackState(LockableState):
"""
A state that can refer back to itself.
It periodically transitions to itself while no other outcome is fulfilled.
"""
_loopback_name = 'loopback'
def __init__(self, *args, **kwargs):
self._rate = rospy.Rate(10)
# add loopback outcome
if len(args) > 0 and type(args[0]) is list:
# need this ugly check for list type, because first argument in CBState is the callback
args[0].append(self._loopback_name)
else:
outcomes = kwargs.get('outcomes', [])
outcomes.append(self._loopback_name)
kwargs['outcomes'] = outcomes
super(LoopbackState, self).__init__(*args, **kwargs)
self.__execute = self.execute
self.execute = self._loopback_execute
def _loopback_execute(self, *args, **kwargs):
result = self.__execute(*args, **kwargs)
if result is None or result == 'None':
result = self._loopback_name
return result
| #!/usr/bin/env python
import rospy
from flexbe_core.core.lockable_state import LockableState
class LoopbackState(LockableState):
"""
A state that can refer back to itself.
It periodically transitions to itself while no other outcome is fulfilled.
"""
_loopback_name = 'loopback'
def __init__(self, *args, **kwargs):
self._rate = rospy.Rate(10)
# add loopback outcome
if len(args) > 0 and type(args[0]) is list:
# need this ugly check for list type, because first argument in CBState is the callback
args[0].append(self._loopback_name)
else:
outcomes = kwargs.get('outcomes', [])
outcomes.append(self._loopback_name)
kwargs['outcomes'] = outcomes
super(LoopbackState, self).__init__(*args, **kwargs)
self.__execute = self.execute
self.execute = self._loopback_execute
def _loopback_execute(self, *args, **kwargs):
result = self.__execute(*args, **kwargs)
if result is None or result == 'None':
result = self._loopback_name
return result
def set_rate(self, rate):
"""
Set the execution rate of this state,
i.e., the rate with which the execute method is being called.
Note: The rate is best-effort,
a rospy.Rate does not guarantee real-time properties.
@type label: float
@param label: The desired rate in Hz.
"""
self._rate = rospy.Rate(rate)
| Add method to set a custom execute rate for states | [flexbe_core] Add method to set a custom execute rate for states
| Python | bsd-3-clause | team-vigir/flexbe_behavior_engine,team-vigir/flexbe_behavior_engine | ---
+++
@@ -31,3 +31,16 @@
if result is None or result == 'None':
result = self._loopback_name
return result
+
+ def set_rate(self, rate):
+ """
+ Set the execution rate of this state,
+ i.e., the rate with which the execute method is being called.
+
+ Note: The rate is best-effort,
+ a rospy.Rate does not guarantee real-time properties.
+
+ @type label: float
+ @param label: The desired rate in Hz.
+ """
+ self._rate = rospy.Rate(rate) |
f1fedff9247b78120df7335b64cdf46c8f60ef03 | test/test_fixtures.py | test/test_fixtures.py | import pytest
from tornado import gen
_used_fixture = False
@gen.coroutine
def dummy(io_loop):
yield gen.Task(io_loop.add_callback)
raise gen.Return(True)
@pytest.fixture(scope='module')
def preparations():
global _used_fixture
_used_fixture = True
pytestmark = pytest.mark.usefixtures('preparations')
@pytest.mark.xfail(pytest.__version__ < '2.7.0',
reason='py.test 2.7 adds hookwrapper, fixes collection')
@pytest.mark.gen_test
def test_uses_pytestmark_fixtures(io_loop):
assert (yield dummy(io_loop))
assert _used_fixture
| import pytest
from tornado import gen
_used_fixture = False
@gen.coroutine
def dummy(io_loop):
yield gen.Task(io_loop.add_callback)
raise gen.Return(True)
@pytest.fixture(scope='module')
def preparations():
global _used_fixture
_used_fixture = True
pytestmark = pytest.mark.usefixtures('preparations')
@pytest.mark.xfail(pytest.__version__ < '2.7.0',
reason='py.test 2.7 adds hookwrapper, fixes collection')
@pytest.mark.gen_test
def test_uses_pytestmark_fixtures(io_loop):
assert (yield dummy(io_loop))
assert _used_fixture
class TestClass:
@pytest.mark.gen_test
def test_uses_pytestmark_fixtures(self, io_loop):
assert (yield dummy(io_loop))
assert _used_fixture
| Add some test for method signature inspection | Add some test for method signature inspection
| Python | apache-2.0 | eugeniy/pytest-tornado | ---
+++
@@ -25,3 +25,9 @@
def test_uses_pytestmark_fixtures(io_loop):
assert (yield dummy(io_loop))
assert _used_fixture
+
+class TestClass:
+ @pytest.mark.gen_test
+ def test_uses_pytestmark_fixtures(self, io_loop):
+ assert (yield dummy(io_loop))
+ assert _used_fixture |
03e864845d0b06cfa4b6cec1adf9b08152a49f0b | forms_builder/example_project/manage.py | forms_builder/example_project/manage.py | #!/usr/bin/env python
from django.core.management import execute_manager
import imp
try:
imp.find_module('settings') # Assumed to be in the same directory.
except ImportError:
import sys
sys.stderr.write("Error: Can't find the file 'settings.py' in the directory containing %r. It appears you've customized things.\nYou'll have to run django-admin.py, passing it your settings module.\n" % __file__)
sys.exit(1)
import settings
if __name__ == "__main__":
execute_manager(settings)
| #!/usr/bin/env python
import sys
import os
sys.path.insert(0, os.path.join(os.path.dirname(__file__), '../..'))
from django.core.management import execute_manager
import imp
try:
imp.find_module('settings') # Assumed to be in the same directory.
except ImportError:
import sys
sys.stderr.write("Error: Can't find the file 'settings.py' in the directory containing %r. It appears you've customized things.\nYou'll have to run django-admin.py, passing it your settings module.\n" % __file__)
sys.exit(1)
import settings
if __name__ == "__main__":
execute_manager(settings)
| Allow tests to be run from example_project | Allow tests to be run from example_project
| Python | bsd-2-clause | iddqd1/django-forms-builder,ixc/django-forms-builder,frontendr/django-forms-builder,vinnyrose/django-forms-builder,GetHappie/django-forms-builder,Afnarel/django-forms-builder,JostCrow/django-forms-builder,GetHappie/django-forms-builder,nimbis/django-forms-builder,Afnarel/django-forms-builder,simas/django-forms-builder,ixc/django-forms-builder,stephenmcd/django-forms-builder,iddqd1/django-forms-builder,bichocj/django-forms-builder,stephenmcd/django-forms-builder,vinnyrose/django-forms-builder,JostCrow/django-forms-builder,frontendr/django-forms-builder,simas/django-forms-builder,bichocj/django-forms-builder,nimbis/django-forms-builder | ---
+++
@@ -1,4 +1,9 @@
#!/usr/bin/env python
+import sys
+import os
+
+sys.path.insert(0, os.path.join(os.path.dirname(__file__), '../..'))
+
from django.core.management import execute_manager
import imp
try: |
3e9a90890f122090be027a3af3d6cbd8a713963c | test/test_issue655.py | test/test_issue655.py | from rdflib import Graph, Namespace, URIRef, Literal
from rdflib.compare import to_isomorphic
import unittest
class TestIssue655(unittest.TestCase):
def test_issue655(self):
PROV = Namespace('http://www.w3.org/ns/prov#')
bob = URIRef("http://example.org/object/Bob")
value = Literal(float("inf"))
# g1 is a simple graph with one attribute having an infinite value
g1 = Graph()
g1.add((bob, PROV.value, value))
# Build g2 out of the deserialisation of g1 serialisation
g2 = Graph()
g2.parse(data=g1.serialize(format='turtle'), format='turtle')
self.assertTrue(g1.serialize(
format='turtle') == g2.serialize(format='turtle'))
self.assertTrue(to_isomorphic(g1) == to_isomorphic(g2))
if __name__ == "__main__":
unittest.main()
| from rdflib import Graph, Namespace, URIRef, Literal
from rdflib.compare import to_isomorphic
import unittest
class TestIssue655(unittest.TestCase):
def test_issue655(self):
PROV = Namespace('http://www.w3.org/ns/prov#')
bob = URIRef("http://example.org/object/Bob")
value = Literal(float("inf"))
# g1 is a simple graph with one attribute having an infinite value
g1 = Graph()
g1.add((bob, PROV.value, value))
# Build g2 out of the deserialisation of g1 serialisation
g2 = Graph()
g2.parse(data=g1.serialize(format='turtle'), format='turtle')
self.assertTrue(g1.serialize(
format='turtle') == g2.serialize(format='turtle'))
self.assertTrue(to_isomorphic(g1) == to_isomorphic(g2))
self.assertTrue(Literal(float("inf")).n3().split("^")[0] == '"INF"')
self.assertTrue(Literal(float("-inf")).n3().split("^")[0] == '"-INF"')
if __name__ == "__main__":
unittest.main()
| Add tests requested by @joernhees | Add tests requested by @joernhees
| Python | bsd-3-clause | RDFLib/rdflib,RDFLib/rdflib,RDFLib/rdflib,RDFLib/rdflib | ---
+++
@@ -23,5 +23,7 @@
format='turtle') == g2.serialize(format='turtle'))
self.assertTrue(to_isomorphic(g1) == to_isomorphic(g2))
+ self.assertTrue(Literal(float("inf")).n3().split("^")[0] == '"INF"')
+ self.assertTrue(Literal(float("-inf")).n3().split("^")[0] == '"-INF"')
if __name__ == "__main__":
unittest.main() |
a66f0227946984702fe6247f64b65152774c5d06 | test/test_packages.py | test/test_packages.py | import pytest
@pytest.mark.parametrize("name", [
("bash-completion"),
("bind-utils"),
("bridge-utils"),
("docker"),
("epel-release"),
("git"),
("iptables-services"),
("libnfsidmap"),
("net-tools"),
("nfs-utils"),
("pyOpenSSL"),
("screen"),
("strace"),
("tcpdump"),
("wget"),
])
def test_packages(Package, name):
assert Package(name).is_installed
| import pytest
@pytest.mark.parametrize("name", [
("bash-completion"),
("bind-utils"),
("bridge-utils"),
("docker"),
("epel-release"),
("git"),
("iptables-services"),
("libnfsidmap"),
("net-tools"),
("nfs-utils"),
("pyOpenSSL"),
("screen"),
("strace"),
("tcpdump"),
("wget"),
])
def test_packages(host, name):
pkg = host.package(name)
assert pkg.is_installed
| Rework based on examples in docs | Rework based on examples in docs
| Python | mit | wicksy/vagrant-openshift,wicksy/vagrant-openshift,wicksy/vagrant-openshift | ---
+++
@@ -18,5 +18,6 @@
("wget"),
])
-def test_packages(Package, name):
- assert Package(name).is_installed
+def test_packages(host, name):
+ pkg = host.package(name)
+ assert pkg.is_installed |
e6d483c867687188ad89dae9ea00b0d651598605 | tw_begins.py | tw_begins.py | #!/usr/bin/env python
import begin
import twitterlib
@begin.subcommand
def timeline():
"Display recent tweets from users timeline"
for status in begin.context.api.timeline:
print u"%s: %s" % (status.user.screen_name, status.text)
@begin.subcommand
def mentions():
"Display recent tweets mentioning user"
for status in begin.context.api.mentions:
print u"%s: %s" % (status.user.screen_name, status.text)
@begin.subcommand
def retweets():
"Display recent retweets from user's timeline"
for status in begin.context.api.retweets:
print u"%s: %s" % (status.user.screen_name, status.text)
@begin.start(env_prefix='')
def main(api_key='', api_secret='', access_token='', access_secret=''):
"""Minimal Twitter client
Demonstrate the use of the begins command line application framework by
implementing a simple Twitter command line client.
"""
api = twitterlib.API(api_key, api_secret, access_token, access_secret)
begin.context.api = api
| #!/usr/bin/env python
import begin
import twitterlib
@begin.subcommand
def timeline():
"Display recent tweets from users timeline"
for status in begin.context.api.timeline:
print u"%s: %s" % (status.user.screen_name, status.text)
@begin.subcommand
def mentions():
"Display recent tweets mentioning user"
for status in begin.context.api.mentions:
print u"%s: %s" % (status.user.screen_name, status.text)
@begin.subcommand
def retweets():
"Display recent retweets from user's timeline"
for status in begin.context.api.retweets:
print u"%s: %s" % (status.user.screen_name, status.text)
@begin.start(env_prefix='', short_args=False)
def main(api_key='', api_secret='', access_token='', access_secret=''):
"""Minimal Twitter client
Demonstrate the use of the begins command line application framework by
implementing a simple Twitter command line client.
"""
api = twitterlib.API(api_key, api_secret, access_token, access_secret)
begin.context.api = api
| Use only long form arguments | Use only long form arguments
Do not use any short (single character) arguments for command line
options.
| Python | mit | aliles/cmdline_examples | ---
+++
@@ -22,7 +22,7 @@
for status in begin.context.api.retweets:
print u"%s: %s" % (status.user.screen_name, status.text)
-@begin.start(env_prefix='')
+@begin.start(env_prefix='', short_args=False)
def main(api_key='', api_secret='', access_token='', access_secret=''):
"""Minimal Twitter client
|
21beae1231d0d721565015bb9eba6925ad5d9fd4 | linter.py | linter.py | from SublimeLinter.lint import PythonLinter
import re
class Pyflakes(PythonLinter):
cmd = 'pyflakes'
regex = r'''(?x)
^(?P<filename>[^:\n]+):(?P<line>\d+):((?P<col>\d+):)?\s
# The rest of the line is the error message.
# Within that, capture anything within single quotes as `near`.
(?P<message>[^\'\n\r]*(?P<near>\'.+?\')?.*)
'''
multiline = True
# stderr has all syntax errors, parse it via our regex
on_stderr = None
defaults = {
'selector': 'source.python'
}
def reposition_match(self, line, col, match, vv):
if 'imported but unused' in match.message:
# Consider:
# from foo import bar
# import foo.bar
# In both cases `pyflakes` reports `'foo.bar' ... unused`.
import_id = re.escape(match.near[1:-1]) # unquote
last_part = import_id.split('.')[-1]
# So we match either `bar` or `foo.bar` against the line content
text = vv.select_line(line)
pattern = r"\s({}|{})".format(last_part, import_id)
re_match = re.search(pattern, text)
if re_match:
return line, re_match.start(1), re_match.end(1)
return super().reposition_match(line, col, match, vv)
| from SublimeLinter.lint import PythonLinter
import re
class Pyflakes(PythonLinter):
cmd = 'pyflakes'
regex = r'''(?x)
^(?P<filename>.+):(?P<line>\d+):((?P<col>\d+):?)?\s
# The rest of the line is the error message.
# Within that, capture anything within single quotes as `near`.
(?P<message>[^\'\n\r]*(?P<near>\'.+?\')?.*)
'''
multiline = True
# stderr has all syntax errors, parse it via our regex
on_stderr = None
defaults = {
'selector': 'source.python'
}
def reposition_match(self, line, col, match, vv):
if 'imported but unused' in match.message:
# Consider:
# from foo import bar
# import foo.bar
# In both cases `pyflakes` reports `'foo.bar' ... unused`.
import_id = re.escape(match.near[1:-1]) # unquote
last_part = import_id.split('.')[-1]
# So we match either `bar` or `foo.bar` against the line content
text = vv.select_line(line)
pattern = r"\s({}|{})".format(last_part, import_id)
re_match = re.search(pattern, text)
if re_match:
return line, re_match.start(1), re_match.end(1)
return super().reposition_match(line, col, match, vv)
| Fix regex for pyflakes output format since `2.2.0` | Fix regex for pyflakes output format since `2.2.0`
| Python | mit | SublimeLinter/SublimeLinter-pyflakes | ---
+++
@@ -5,7 +5,7 @@
class Pyflakes(PythonLinter):
cmd = 'pyflakes'
regex = r'''(?x)
- ^(?P<filename>[^:\n]+):(?P<line>\d+):((?P<col>\d+):)?\s
+ ^(?P<filename>.+):(?P<line>\d+):((?P<col>\d+):?)?\s
# The rest of the line is the error message.
# Within that, capture anything within single quotes as `near`. |
2ac4bca0db8609bc92c9de8b1c272b2a607f6c15 | tests/resource_tests.py | tests/resource_tests.py | #
# Project: retdec-python
# Copyright: (c) 2015 by Petr Zemek <s3rvac@gmail.com> and contributors
# License: MIT, see the LICENSE file for more details
#
"""Tests for the :mod:`retdec.resource` module."""
import unittest
from unittest import mock
from retdec.conn import APIConnection
from retdec.resource import Resource
class ResourceTests(unittest.TestCase):
"""Tests for :class:`retdec.resource.Resource`."""
def test_id_returns_passed_id(self):
r = Resource('ID', mock.Mock(spec_set=APIConnection))
self.assertEqual(r.id, 'ID')
def test_wait_until_finished_returns_when_resource_is_finished(self):
conn_mock = mock.Mock(spec_set=APIConnection)
conn_mock.send_get_request.return_value = {'finished': True}
r = Resource('ID', conn_mock)
r.wait_until_finished()
conn_mock.send_get_request.assert_called_once_with('/ID/status')
| #
# Project: retdec-python
# Copyright: (c) 2015 by Petr Zemek <s3rvac@gmail.com> and contributors
# License: MIT, see the LICENSE file for more details
#
"""Tests for the :mod:`retdec.resource` module."""
import unittest
from unittest import mock
from retdec.conn import APIConnection
from retdec.resource import Resource
class ResourceTests(unittest.TestCase):
"""Tests for :class:`retdec.resource.Resource`."""
def test_id_returns_passed_id(self):
r = Resource('ID', mock.Mock(spec_set=APIConnection))
self.assertEqual(r.id, 'ID')
class ResourceWaitUntilFinishedTests(unittest.TestCase):
"""Tests for :func:`retdec.resource.Resource.wait_until_finished()`."""
def test_returns_when_resource_is_finished(self):
conn_mock = mock.Mock(spec_set=APIConnection)
conn_mock.send_get_request.return_value = {'finished': True}
r = Resource('ID', conn_mock)
r.wait_until_finished()
conn_mock.send_get_request.assert_called_once_with('/ID/status')
| Move tests for Resource.wait_until_finished() into a separate class. | Move tests for Resource.wait_until_finished() into a separate class.
| Python | mit | s3rvac/retdec-python | ---
+++
@@ -20,7 +20,11 @@
r = Resource('ID', mock.Mock(spec_set=APIConnection))
self.assertEqual(r.id, 'ID')
- def test_wait_until_finished_returns_when_resource_is_finished(self):
+
+class ResourceWaitUntilFinishedTests(unittest.TestCase):
+ """Tests for :func:`retdec.resource.Resource.wait_until_finished()`."""
+
+ def test_returns_when_resource_is_finished(self):
conn_mock = mock.Mock(spec_set=APIConnection)
conn_mock.send_get_request.return_value = {'finished': True}
|
95eb2c11a4f35e594eda25c10bdf85a25b2f4392 | src/ConfigLoader.py | src/ConfigLoader.py | import json
import sys
def load_config_file(out=sys.stdout):
default_filepath = "../resources/config/default-config.json"
user_filepath = "../resources/config/user-config.json"
try:
default_json = read_json(default_filepath)
user_json = read_json(user_filepath)
for property in user_json:
default_json[property] = user_json[property]
except FileNotFoundError as e:
out.write("Cannot find file: " + e.filename)
else:
out.write("Read styling config JSON correctly.")
return default_json
def read_json(filepath):
config_string = ''
with open(filepath) as f:
for line in f:
line = line.lstrip()
if not line.startswith("//"):
config_string += line
config_json = json.loads(config_string)
return config_json
if __name__ == "__main__":
load_config_file() | import json
import sys
def load_config_file(out=sys.stdout):
if sys.argv[0].endswith('nosetests'):
default_filepath = "./resources/config/default-config.json"
user_filepath = "./resources/config/user-config.json"
else:
default_filepath = "../resources/config/default-config.json"
user_filepath = "../resources/config/user-config.json"
try:
default_json = read_json(default_filepath)
user_json = read_json(user_filepath)
for property in user_json:
default_json[property] = user_json[property]
except FileNotFoundError as e:
out.write("Cannot find file: " + e.filename)
else:
out.write("Read styling config JSON correctly.")
return default_json
def read_json(filepath):
config_string = ''
with open(filepath) as f:
for line in f:
line = line.lstrip()
if not line.startswith("//"):
config_string += line
config_json = json.loads(config_string)
return config_json
if __name__ == "__main__":
load_config_file() | Fix nosetests for config file loading | Fix nosetests for config file loading
| Python | bsd-3-clause | sky-uk/bslint | ---
+++
@@ -3,8 +3,12 @@
def load_config_file(out=sys.stdout):
- default_filepath = "../resources/config/default-config.json"
- user_filepath = "../resources/config/user-config.json"
+ if sys.argv[0].endswith('nosetests'):
+ default_filepath = "./resources/config/default-config.json"
+ user_filepath = "./resources/config/user-config.json"
+ else:
+ default_filepath = "../resources/config/default-config.json"
+ user_filepath = "../resources/config/user-config.json"
try:
default_json = read_json(default_filepath) |
7d8283b2d233a8fbee97de122f0b4ba293cf788d | app/emails.py | app/emails.py | # -*- coding: utf-8 -*-
from flask import render_template,g
from flask.ext.mail import Message
from app import mail, db
from .models import User
from config import MAIL_SENDER
# Wrapper function for sending mails using flask-mail plugin
def send_email(subject, sender, recipients, text_body):
msg = Message(subject, sender=sender, recipients=recipients)
msg.body = text_body
mail.send(msg)
def add_movie_notification(movie):
users = User.query.all()
you_user = False
for cur_user in users:
# Check if the cur_user is the logged user who added the movie
# in order to change the mail text
if cur_user.id==g.user.id:
you_user=True
send_email('[Cineapp] - Ajout d\'un film' , MAIL_SENDER,[ cur_user.email ] ,
render_template('add_movie_notification.txt', dest_user=cur_user, add_user=g.user,movie=movie,you_user=you_user))
| # -*- coding: utf-8 -*-
from flask import render_template,g
from flask.ext.mail import Message
from app import mail, db
from .models import User
from config import MAIL_SENDER
from threading import Thread
from app import app
# Send mail into a dedicated thread in order to avoir the web app to wait
def send_async_email(app, msg):
with app.app_context():
mail.send(msg)
# Wrapper function for sending mails using flask-mail plugin
def send_email(subject, sender, recipients, text_body):
msg = Message(subject, sender=sender, recipients=recipients)
msg.body = text_body
thr = Thread(target=send_async_email, args=[app, msg])
thr.start()
# Function which sends notifications to users when a movie is added
def add_movie_notification(movie):
users = User.query.all()
you_user = False
for cur_user in users:
# Check if the cur_user is the logged user who added the movie
# in order to change the mail text
if cur_user.id==g.user.id:
you_user=True
send_email('[Cineapp] - Ajout d\'un film' , MAIL_SENDER,[ cur_user.email ] ,
render_template('add_movie_notification.txt', dest_user=cur_user, add_user=g.user,movie=movie,you_user=you_user))
| Send email notifications in asynchronous mode | Send email notifications in asynchronous mode
Each mail notification is done on a thread in order to not block the main thread
of the web app.
| Python | mit | ptitoliv/cineapp,ptitoliv/cineapp,ptitoliv/cineapp | ---
+++
@@ -5,13 +5,22 @@
from app import mail, db
from .models import User
from config import MAIL_SENDER
+from threading import Thread
+from app import app
+
+# Send mail into a dedicated thread in order to avoir the web app to wait
+def send_async_email(app, msg):
+ with app.app_context():
+ mail.send(msg)
# Wrapper function for sending mails using flask-mail plugin
def send_email(subject, sender, recipients, text_body):
msg = Message(subject, sender=sender, recipients=recipients)
msg.body = text_body
- mail.send(msg)
+ thr = Thread(target=send_async_email, args=[app, msg])
+ thr.start()
+# Function which sends notifications to users when a movie is added
def add_movie_notification(movie):
users = User.query.all()
you_user = False |
57db4140237b79cc8b6958ac503bc087f57ddad2 | ibmcnx/doc/DataSources.py | ibmcnx/doc/DataSources.py | ######
# Check ExId (GUID) by Email through JDBC
#
# Author: Christoph Stoettner
# Mail: christoph.stoettner@stoeps.de
# Documentation: http://scripting101.stoeps.de
#
# Version: 2.0
# Date: 2014-06-04
#
# License: Apache 2.0
#
# Check ExId of a User in all Connections Applications
import ibmcnx.functions
cell = '/' + AdminControl.getCell() + '/'
dbs = AdminConfig.list( 'DataSource', cell )
for db in dbs:
t1 = ibmcnx.functions.getDSId( db )
AdminConfig.list( t1 ) | ######
# Check ExId (GUID) by Email through JDBC
#
# Author: Christoph Stoettner
# Mail: christoph.stoettner@stoeps.de
# Documentation: http://scripting101.stoeps.de
#
# Version: 2.0
# Date: 2014-06-04
#
# License: Apache 2.0
#
# Check ExId of a User in all Connections Applications
import ibmcnx.functions
cell = AdminConfig.getid( '/Cell:' + AdminControl.getCell() + '/' )
dbs = AdminConfig.list( 'DataSource', cell )
for db in dbs:
t1 = ibmcnx.functions.getDSId( db )
AdminConfig.list( t1 ) | Create script to save documentation to a file | 4: Create script to save documentation to a file
Task-Url: http://github.com/stoeps13/ibmcnx2/issues/issue/4 | Python | apache-2.0 | stoeps13/ibmcnx2,stoeps13/ibmcnx2 | ---
+++
@@ -14,7 +14,7 @@
import ibmcnx.functions
-cell = '/' + AdminControl.getCell() + '/'
+cell = AdminConfig.getid( '/Cell:' + AdminControl.getCell() + '/' )
dbs = AdminConfig.list( 'DataSource', cell )
for db in dbs: |
30e1e82573437cfaf8b75ec9f42c520a5f4f60d5 | ibmcnx/doc/DataSources.py | ibmcnx/doc/DataSources.py | ######
# Check ExId (GUID) by Email through JDBC
#
# Author: Christoph Stoettner
# Mail: christoph.stoettner@stoeps.de
# Documentation: http://scripting101.stoeps.de
#
# Version: 2.0
# Date: 2014-06-04
#
# License: Apache 2.0
#
# Check ExId of a User in all Connections Applications
import ibmcnx.functions
cell = AdminControl.getCell()
cellname = "/Cell:" + cell + "/"
# Get a list of all databases except DefaultEJBTimerDataSource and OTiSDataSource
dbs = AdminConfig.list('DataSource',AdminConfig.getid(cellname)).splitlines()
dblist = []
for db in dbs:
db = db.split('(')
n = 0
for i in db:
if n == 0 and i != "DefaultEJBTimerDataSource" and i != 'OTiSDataSource':
dblist.append(str(i).replace('"',''))
n += 1
dblist.sort()
for db in dblist:
t1 = ibmcnx.functions.getDSId( db )
print t1
# AdminConfig.show( t1 )
# print '\n\n'
# AdminConfig.showall( t1 )
# AdminConfig.showAttribute(t1,'statementCacheSize' )
# AdminConfig.showAttribute(t1,'[statementCacheSize]' ) | ######
# Check ExId (GUID) by Email through JDBC
#
# Author: Christoph Stoettner
# Mail: christoph.stoettner@stoeps.de
# Documentation: http://scripting101.stoeps.de
#
# Version: 2.0
# Date: 2014-06-04
#
# License: Apache 2.0
#
# Check ExId of a User in all Connections Applications
import ibmcnx.functions
cell = AdminControl.getCell()
cellname = "/Cell:" + cell + "/"
# Get a list of all databases except DefaultEJBTimerDataSource and OTiSDataSource
dbs = AdminConfig.list('DataSource',AdminConfig.getid(cellname)).splitlines()
dblist = []
for db in dbs:
dbname = db.split('(')
n = 0
for i in dbname:
if n == 0 and i != "DefaultEJBTimerDataSource" and i != 'OTiSDataSource':
dblist.append(str(db).replace('"',''))
n += 1
dblist.sort()
for db in dblist:
#t1 = ibmcnx.functions.getDSId( db )
print db
# AdminConfig.show( t1 )
# print '\n\n'
# AdminConfig.showall( t1 )
# AdminConfig.showAttribute(t1,'statementCacheSize' )
# AdminConfig.showAttribute(t1,'[statementCacheSize]' ) | Create documentation of DataSource Settings | 8: Create documentation of DataSource Settings
Task-Url: http://github.com/stoeps13/ibmcnx2/issues/issue/8 | Python | apache-2.0 | stoeps13/ibmcnx2,stoeps13/ibmcnx2 | ---
+++
@@ -21,18 +21,18 @@
dbs = AdminConfig.list('DataSource',AdminConfig.getid(cellname)).splitlines()
dblist = []
for db in dbs:
- db = db.split('(')
+ dbname = db.split('(')
n = 0
- for i in db:
+ for i in dbname:
if n == 0 and i != "DefaultEJBTimerDataSource" and i != 'OTiSDataSource':
- dblist.append(str(i).replace('"',''))
+ dblist.append(str(db).replace('"',''))
n += 1
dblist.sort()
for db in dblist:
- t1 = ibmcnx.functions.getDSId( db )
- print t1
+ #t1 = ibmcnx.functions.getDSId( db )
+ print db
# AdminConfig.show( t1 )
# print '\n\n'
# AdminConfig.showall( t1 ) |
4d2a2b3e5b9f734e688c104845d33b90dda2c159 | commands/say.py | commands/say.py | from CommandTemplate import CommandTemplate
class Command(CommandTemplate):
triggers = ['say']
helptext = "Makes the bot say the provided text in the provided channel (format 'say [channel/user] text')"
adminOnly = True
showInCommandList = False
def execute(self, bot, user, target, triggerInMsg, msg, msgWithoutFirstWord, msgParts, msgPartsLength):
if msgPartsLength < 3:
bot.say(target, "Please provide both a channel or user name to say something to, and the text to say")
return
bot.say(msgParts[1], " ".join(msgParts[2:]))
| from CommandTemplate import CommandTemplate
class Command(CommandTemplate):
triggers = ['say']
helptext = "Makes the bot say the provided text in the provided channel (format 'say [channel/user] text')"
adminOnly = True
showInCommandList = False
def execute(self, bot, user, target, triggerInMsg, msg, msgWithoutFirstWord, msgParts, msgPartsLength):
if msgPartsLength < 3:
bot.say(target, "Please provide both a channel or user name to say something to, and the text to say")
#Check if we're in the channel we have to say something to
elif msgParts[1].startswith('#') and msgParts[1] not in bot.channelsUserList:
bot.say(target, "I'm not in that channel, so I can't say anything in there, sorry.")
#Nothing's stopping us now! Say it!
else:
bot.say(msgParts[1], " ".join(msgParts[2:]))
| Add a check to see if the bot's in the channel the message should go to | Add a check to see if the bot's in the channel the message should go to
| Python | mit | Didero/DideRobot | ---
+++
@@ -9,6 +9,9 @@
def execute(self, bot, user, target, triggerInMsg, msg, msgWithoutFirstWord, msgParts, msgPartsLength):
if msgPartsLength < 3:
bot.say(target, "Please provide both a channel or user name to say something to, and the text to say")
- return
-
- bot.say(msgParts[1], " ".join(msgParts[2:]))
+ #Check if we're in the channel we have to say something to
+ elif msgParts[1].startswith('#') and msgParts[1] not in bot.channelsUserList:
+ bot.say(target, "I'm not in that channel, so I can't say anything in there, sorry.")
+ #Nothing's stopping us now! Say it!
+ else:
+ bot.say(msgParts[1], " ".join(msgParts[2:])) |
3d29ca66f17bcf5dcc72f9fe12a83c4949066a8b | linter.py | linter.py | from SublimeLinter.lint import PythonLinter
import re
class Pyflakes(PythonLinter):
cmd = 'pyflakes'
regex = r'''(?x)
^(?P<filename>.+):(?P<line>\d+):((?P<col>\d+):?)?\s
# The rest of the line is the error message.
# Within that, capture anything within single quotes as `near`.
(?P<message>[^\'\n\r]*(?P<near>\'.+?\')?.*)
'''
multiline = True
# stderr has all syntax errors, parse it via our regex
on_stderr = None
defaults = {
'selector': 'source.python'
}
def reposition_match(self, line, col, match, vv):
if 'imported but unused' in match.message:
# Consider:
# from foo import bar
# import foo.bar
# In both cases `pyflakes` reports `'foo.bar' ... unused`.
import_id = re.escape(match.near[1:-1]) # unquote
last_part = import_id.split('.')[-1]
# So we match either `bar` or `foo.bar` against the line content
text = vv.select_line(line)
pattern = r"\s({}|{})".format(last_part, import_id)
re_match = re.search(pattern, text)
if re_match:
return line, re_match.start(1), re_match.end(1)
return super().reposition_match(line, col, match, vv)
| from SublimeLinter.lint import PythonLinter
import re
class Pyflakes(PythonLinter):
cmd = 'pyflakes'
regex = r'''(?x)
^(?P<filename>.+?):(?P<line>\d+):((?P<col>\d+):?)?\s
# The rest of the line is the error message.
# Within that, capture anything within single quotes as `near`.
(?P<message>[^\'\n\r]*(?P<near>\'.+?\')?.*)
'''
multiline = True
# stderr has all syntax errors, parse it via our regex
on_stderr = None
defaults = {
'selector': 'source.python'
}
def reposition_match(self, line, col, match, vv):
if 'imported but unused' in match.message:
# Consider:
# from foo import bar
# import foo.bar
# In both cases `pyflakes` reports `'foo.bar' ... unused`.
import_id = re.escape(match.near[1:-1]) # unquote
last_part = import_id.split('.')[-1]
# So we match either `bar` or `foo.bar` against the line content
text = vv.select_line(line)
pattern = r"\s({}|{})".format(last_part, import_id)
re_match = re.search(pattern, text)
if re_match:
return line, re_match.start(1), re_match.end(1)
return super().reposition_match(line, col, match, vv)
| Fix regex as the `:` after the column is optional | Fix regex as the `:` after the column is optional
Fixes SublimeLinter/SublimeLinter#1847
Closes #17
Output from pyflakes can be
```
<stdin>:1044:12 undefined name 'settingss'
<stdin>:34:45: invalid syntax
```
t.i. the `:` after the column is optional. In the latter case
`filename` overmatched ("not lazy") to "<stdin>:34".
| Python | mit | SublimeLinter/SublimeLinter-pyflakes | ---
+++
@@ -5,7 +5,7 @@
class Pyflakes(PythonLinter):
cmd = 'pyflakes'
regex = r'''(?x)
- ^(?P<filename>.+):(?P<line>\d+):((?P<col>\d+):?)?\s
+ ^(?P<filename>.+?):(?P<line>\d+):((?P<col>\d+):?)?\s
# The rest of the line is the error message.
# Within that, capture anything within single quotes as `near`. |
236816340df75e1ef4fc3f6ee4540a02c44fc279 | tests/helpers/helpers.py | tests/helpers/helpers.py | from inspect import getdoc
import pytest
class CodeCollector(object):
def __init__(self):
self.collected = []
def __call__(self, f):
self.collected.append(f)
return f
def __iter__(self):
return iter(self.collected)
def parametrize(self, test_func):
return pytest.mark.parametrize('code', self, ids=getdoc)(test_func)
| from inspect import getdoc
import pytest
class CodeCollector(object):
def __init__(self, name='code'):
self.name = name
self.collected = []
def __call__(self, f):
self.collected.append(f)
return f
def __iter__(self):
return iter(self.collected)
def parametrize(self, test_func):
return pytest.mark.parametrize(self.name, self, ids=getdoc)(test_func)
| Support argument name in the CodeCollector helper. | Support argument name in the CodeCollector helper.
| Python | bsd-2-clause | proofit404/dependencies,proofit404/dependencies,proofit404/dependencies,proofit404/dependencies | ---
+++
@@ -5,8 +5,9 @@
class CodeCollector(object):
- def __init__(self):
+ def __init__(self, name='code'):
+ self.name = name
self.collected = []
def __call__(self, f):
@@ -20,4 +21,4 @@
def parametrize(self, test_func):
- return pytest.mark.parametrize('code', self, ids=getdoc)(test_func)
+ return pytest.mark.parametrize(self.name, self, ids=getdoc)(test_func) |
e822ee6fe844c18b2a459ed188f4fbae26c617e5 | lingcod/bookmarks/forms.py | lingcod/bookmarks/forms.py | from lingcod.features.forms import FeatureForm
from lingcod.bookmarks.models import Bookmark
from django import forms
class BookmarkForm(FeatureForm):
name = forms.CharField(label='Bookmark Name')
latitude = forms.FloatField(widget=forms.HiddenInput())
longitude = forms.FloatField(widget=forms.HiddenInput())
altitude = forms.FloatField(widget=forms.HiddenInput())
heading = forms.FloatField(widget=forms.HiddenInput())
tilt = forms.FloatField(widget=forms.HiddenInput())
roll = forms.FloatField(widget=forms.HiddenInput())
altitudeMode = forms.FloatField(widget=forms.HiddenInput())
publicstate = forms.CharField(widget=forms.HiddenInput())
class Meta(FeatureForm.Meta):
model = Bookmark
| from lingcod.features.forms import FeatureForm
from lingcod.bookmarks.models import Bookmark
from django import forms
class BookmarkForm(FeatureForm):
name = forms.CharField(label='Bookmark Name')
latitude = forms.FloatField(widget=forms.HiddenInput())
longitude = forms.FloatField(widget=forms.HiddenInput())
altitude = forms.FloatField(widget=forms.HiddenInput())
heading = forms.FloatField(widget=forms.HiddenInput())
tilt = forms.FloatField(widget=forms.HiddenInput())
roll = forms.FloatField(widget=forms.HiddenInput())
altitudeMode = forms.FloatField(widget=forms.HiddenInput())
publicstate = forms.CharField(widget=forms.HiddenInput())
ip = forms.CharField(widget=forms.HiddenInput())
class Meta(FeatureForm.Meta):
model = Bookmark
| Hide IP from input form | Hide IP from input form
--HG--
branch : bookmarks
| Python | bsd-3-clause | underbluewaters/marinemap,underbluewaters/marinemap,underbluewaters/marinemap | ---
+++
@@ -12,5 +12,6 @@
roll = forms.FloatField(widget=forms.HiddenInput())
altitudeMode = forms.FloatField(widget=forms.HiddenInput())
publicstate = forms.CharField(widget=forms.HiddenInput())
+ ip = forms.CharField(widget=forms.HiddenInput())
class Meta(FeatureForm.Meta):
model = Bookmark |
184e726d44d113a46ddb9cf3a5762f453ed7b512 | myuw/management/commands/clear_expired_sessions.py | myuw/management/commands/clear_expired_sessions.py | """
The django clearsessions commend internally calls:
cls.get_model_class().objects.filter(
expire_date__lt=timezone.now()).delete()
which could lock the DB table for a long time when
having a large number of records to delete.
To prevent the job running forever, we only delete a limit number of
expired django sessions in a single run
"""
import logging
from datetime import timedelta
from django.core.management.base import BaseCommand, CommandError
from django.contrib.sessions.models import Session
from django.utils import timezone
from myuw.logger.timer import Timer
logger = logging.getLogger(__name__)
begin_delta = 1920
log_format = "Deleted django sessions expired before {}, Time={} seconds"
class Command(BaseCommand):
def handle(self, *args, **options):
now = timezone.now()
for ddelta in range(begin_delta, 0, -1):
timer = Timer()
cut_off_dt = now - timedelta(days=ddelta)
try:
qset = Session.objects.filter(expire_date__lt=cut_off_dt)
if qset.exists():
qset.delete()
logger.info(log_format.format(cut_off_dt.date(),
timer.get_elapsed()))
except Exception as ex:
logger.error(str(ex))
| """
The django clearsessions commend internally calls:
cls.get_model_class().objects.filter(
expire_date__lt=timezone.now()).delete()
which could lock the DB table for a long time when
having a large number of records to delete.
To prevent the job running forever, we only delete a limit number of
expired django sessions in a single run
"""
import logging
from datetime import timedelta
import time
from django.core.management.base import BaseCommand, CommandError
from django.contrib.sessions.models import Session
from django.utils import timezone
from myuw.logger.timer import Timer
logger = logging.getLogger(__name__)
begin_delta = 1920
log_format = "Deleted django sessions expired before {}, Time={} seconds"
class Command(BaseCommand):
def handle(self, *args, **options):
now = timezone.now()
for ddelta in range(begin_delta, 0, -1):
timer = Timer()
cut_off_dt = now - timedelta(days=ddelta)
try:
qset = Session.objects.filter(expire_date__lt=cut_off_dt)
if qset.exists():
qset.delete()
logger.info(log_format.format(cut_off_dt.date(),
timer.get_elapsed()))
time.sleep(5)
except Exception as ex:
logger.error(str(ex))
| Add a 5 second pause | Add a 5 second pause
| Python | apache-2.0 | uw-it-aca/myuw,uw-it-aca/myuw,uw-it-aca/myuw,uw-it-aca/myuw | ---
+++
@@ -11,6 +11,7 @@
import logging
from datetime import timedelta
+import time
from django.core.management.base import BaseCommand, CommandError
from django.contrib.sessions.models import Session
from django.utils import timezone
@@ -34,6 +35,6 @@
qset.delete()
logger.info(log_format.format(cut_off_dt.date(),
timer.get_elapsed()))
-
+ time.sleep(5)
except Exception as ex:
logger.error(str(ex)) |
f00c7f3a976ba4790963a5701c5ce13f6dcd84fa | tests/test_funcmakers.py | tests/test_funcmakers.py | import inspect
from collections import defaultdict
import pytest
from funcy.funcmakers import *
def test_callable():
assert make_func(lambda x: x + 42)(0) == 42
def test_int():
assert make_func(0)('abc') == 'a'
assert make_func(2)([1,2,3]) == 3
assert make_func(1)({1: 'a'}) == 'a'
with pytest.raises(IndexError): make_func(1)('a')
with pytest.raises(TypeError): make_func(1)(42)
def test_slice():
assert make_func(slice(1, None))('abc') == 'bc'
def test_str():
assert make_func('\d+')('ab42c') == '42'
assert make_func('\d+')('abc') is None
assert make_pred('\d+')('ab42c') is True
assert make_pred('\d+')('abc') is False
def test_dict():
assert make_func({1: 'a'})(1) == 'a'
with pytest.raises(KeyError): make_func({1: 'a'})(2)
d = defaultdict(int, a=42)
assert make_func(d)('a') == 42
assert make_func(d)('b') == 0
def test_set():
s = set([1,2,3])
assert make_func(s)(1) is True
assert make_func(s)(4) is False
| from collections import defaultdict
import pytest
from funcy.funcmakers import *
def test_callable():
assert make_func(lambda x: x + 42)(0) == 42
def test_int():
assert make_func(0)('abc') == 'a'
assert make_func(2)([1,2,3]) == 3
assert make_func(1)({1: 'a'}) == 'a'
with pytest.raises(IndexError): make_func(1)('a')
with pytest.raises(TypeError): make_func(1)(42)
def test_slice():
assert make_func(slice(1, None))('abc') == 'bc'
def test_str():
assert make_func('\d+')('ab42c') == '42'
assert make_func('\d+')('abc') is None
assert make_pred('\d+')('ab42c') is True
assert make_pred('\d+')('abc') is False
def test_dict():
assert make_func({1: 'a'})(1) == 'a'
with pytest.raises(KeyError): make_func({1: 'a'})(2)
d = defaultdict(int, a=42)
assert make_func(d)('a') == 42
assert make_func(d)('b') == 0
def test_set():
s = set([1,2,3])
assert make_func(s)(1) is True
assert make_func(s)(4) is False
| Remove unused import from tests | Remove unused import from tests
| Python | bsd-3-clause | Suor/funcy | ---
+++
@@ -1,4 +1,3 @@
-import inspect
from collections import defaultdict
import pytest |
b602e732ae091b78c1a0e7258bec878a12cf063f | tests/test_account.py | tests/test_account.py | # coding: utf-8
import unittest
from lastpass.account import Account
class AccountTestCase(unittest.TestCase):
def setUp(self):
self.id = 'id'
self.name = 'name'
self.username = 'username'
self.password = 'password'
self.url = 'url'
self.group = 'group'
self.account = Account(self.id, self.name, self.username, self.password, self.url, self.group)
def test_id_returns_the_correct_value(self):
self.assertEqual(self.account.id, self.id)
def test_name_returns_the_correct_value(self):
self.assertEqual(self.account.name, self.name)
def test_username_returns_the_correct_value(self):
self.assertEqual(self.account.username, self.username)
def test_password_returns_the_correct_value(self):
self.assertEqual(self.account.password, self.password)
def test_url_returns_the_correct_value(self):
self.assertEqual(self.account.url, self.url)
def test_group_returns_the_correct_value(self):
self.assertEqual(self.account.group, self.group)
| # coding: utf-8
import unittest
from lastpass.account import Account
class AccountTestCase(unittest.TestCase):
def setUp(self):
self.id = 'id'
self.name = 'name'
self.username = 'username'
self.password = 'password'
self.url = 'url'
self.group = 'group'
self.notes = 'notes'
self.account = Account(self.id, self.name, self.username, self.password, self.url, self.group, self.notes)
def test_id_returns_the_correct_value(self):
self.assertEqual(self.account.id, self.id)
def test_name_returns_the_correct_value(self):
self.assertEqual(self.account.name, self.name)
def test_username_returns_the_correct_value(self):
self.assertEqual(self.account.username, self.username)
def test_password_returns_the_correct_value(self):
self.assertEqual(self.account.password, self.password)
def test_url_returns_the_correct_value(self):
self.assertEqual(self.account.url, self.url)
def test_group_returns_the_correct_value(self):
self.assertEqual(self.account.group, self.group)
def test_notes_returns_the_correct_value(self):
self.assertEqual(self.account.notes, self.notes)
| Add a test for `notes` attribute | Add a test for `notes` attribute
| Python | mit | konomae/lastpass-python | ---
+++
@@ -11,7 +11,8 @@
self.password = 'password'
self.url = 'url'
self.group = 'group'
- self.account = Account(self.id, self.name, self.username, self.password, self.url, self.group)
+ self.notes = 'notes'
+ self.account = Account(self.id, self.name, self.username, self.password, self.url, self.group, self.notes)
def test_id_returns_the_correct_value(self):
self.assertEqual(self.account.id, self.id)
@@ -30,3 +31,6 @@
def test_group_returns_the_correct_value(self):
self.assertEqual(self.account.group, self.group)
+
+ def test_notes_returns_the_correct_value(self):
+ self.assertEqual(self.account.notes, self.notes) |
55d4b7b939fc218d47a920761350aee7bee91eb9 | opps/article/views.py | opps/article/views.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from django.views.generic.detail import DetailView
from django.views.generic.list import ListView
from opps.article.models import Post
class OppsList(ListView):
context_object_name = "context"
@property
def template_name(self):
return 'channel/{0}.html'.format(self.kwargs['channel__long_slug'])
@property
def queryset(self):
if not self.kwargs.get('channel__long_slug'):
return Post.objects.filter(channel__homepage=True).all()
return Post.objects.filter(
channel__long_slug=self.kwargs['channel__long_slug']).all()
class OppsDetail(DetailView):
context_object_name = "context"
@property
def template_name(self):
return 'article/{0}/{1}.html'.format(
self.kwargs['channel__long_slug'], self.kwargs['slug'])
@property
def queryset(self):
return Post.objects.filter(
channel__long_slug=self.kwargs['channel__long_slug'],
slug=self.kwargs['slug']).all()
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
from django.views.generic.detail import DetailView
from django.views.generic.list import ListView
from opps.article.models import Post
class OppsList(ListView):
context_object_name = "context"
@property
def template_name(self):
long_slug = self.kwargs.get('channel__long_slug', 'home')
return 'channel/{0}.html'.format(long_slug)
@property
def queryset(self):
if not self.kwargs.get('channel__long_slug'):
return Post.objects.filter(channel__homepage=True).all()
return Post.objects.filter(
channel__long_slug=self.kwargs['channel__long_slug']).all()
class OppsDetail(DetailView):
context_object_name = "context"
@property
def template_name(self):
return 'article/{0}/{1}.html'.format(
self.kwargs['channel__long_slug'], self.kwargs['slug'])
@property
def queryset(self):
return Post.objects.filter(
channel__long_slug=self.kwargs['channel__long_slug'],
slug=self.kwargs['slug']).all()
| Fix template name on entry home page (/) on list page | Fix template name on entry home page (/) on list page
| Python | mit | opps/opps,williamroot/opps,williamroot/opps,opps/opps,opps/opps,jeanmask/opps,jeanmask/opps,williamroot/opps,jeanmask/opps,YACOWS/opps,opps/opps,jeanmask/opps,YACOWS/opps,YACOWS/opps,YACOWS/opps,williamroot/opps | ---
+++
@@ -13,7 +13,8 @@
@property
def template_name(self):
- return 'channel/{0}.html'.format(self.kwargs['channel__long_slug'])
+ long_slug = self.kwargs.get('channel__long_slug', 'home')
+ return 'channel/{0}.html'.format(long_slug)
@property
def queryset(self): |
1f5403108a25257dfa63d235c8faf4e02f69c6bf | opps/sitemaps/urls.py | opps/sitemaps/urls.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from django.conf.urls import patterns, url
from django.contrib.sitemaps import views as sitemap_views
from opps.sitemaps.sitemaps import GenericSitemap, InfoDisct
sitemaps = {
'articles': GenericSitemap(InfoDisct(), priority=0.6),
}
sitemaps_googlenews = {
'articles': GenericSitemap(InfoDisct(True), priority=0.6),
}
urlpatterns = patterns(
'',
url(r'^\.xml$', sitemap_views.index,
{'sitemaps': sitemaps}),
url(r'^-googlenews\.xml$', sitemap_views.sitemap,
{'sitemaps': sitemaps_googlenews,
'template_name': 'sitemap_googlenews.xml'}),
url(r'^-(?P<section>.+)\.xml$', sitemap_views.sitemap,
{'sitemaps': sitemaps}),
)
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
from django.conf.urls import patterns, url
from django.views.decorators.cache import cache_page
from django.contrib.sitemaps import views as sitemap_views
from opps.sitemaps.sitemaps import GenericSitemap, InfoDisct
sitemaps = {
'articles': GenericSitemap(InfoDisct(), priority=0.6),
}
sitemaps_googlenews = {
'articles': GenericSitemap(InfoDisct(True), priority=0.6),
}
urlpatterns = patterns(
'',
url(r'^\.xml$', cache_page(86400)(sitemap_views.index),
{'sitemaps': sitemaps}),
url(r'^-googlenews\.xml$', cache_page(86400)(sitemap_views.sitemap),
{'sitemaps': sitemaps_googlenews,
'template_name': 'sitemap_googlenews.xml'}),
url(r'^-(?P<section>.+)\.xml$', sitemap_views.sitemap,
{'sitemaps': sitemaps}),
)
| Add cache page on sitemap fixed | Add cache page on sitemap fixed
| Python | mit | jeanmask/opps,williamroot/opps,YACOWS/opps,opps/opps,YACOWS/opps,opps/opps,opps/opps,jeanmask/opps,opps/opps,jeanmask/opps,williamroot/opps,YACOWS/opps,jeanmask/opps,williamroot/opps,williamroot/opps,YACOWS/opps | ---
+++
@@ -1,6 +1,7 @@
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from django.conf.urls import patterns, url
+from django.views.decorators.cache import cache_page
from django.contrib.sitemaps import views as sitemap_views
from opps.sitemaps.sitemaps import GenericSitemap, InfoDisct
@@ -16,9 +17,9 @@
urlpatterns = patterns(
'',
- url(r'^\.xml$', sitemap_views.index,
+ url(r'^\.xml$', cache_page(86400)(sitemap_views.index),
{'sitemaps': sitemaps}),
- url(r'^-googlenews\.xml$', sitemap_views.sitemap,
+ url(r'^-googlenews\.xml$', cache_page(86400)(sitemap_views.sitemap),
{'sitemaps': sitemaps_googlenews,
'template_name': 'sitemap_googlenews.xml'}),
url(r'^-(?P<section>.+)\.xml$', sitemap_views.sitemap, |
18946d7f464e1f71ad91229261bd8d191a16dea3 | demo/setup.py | demo/setup.py | import os
from setuptools import setup
def read_relative_file(filename):
"""Returns contents of the given file, which path is supposed relative
to this module."""
with open(os.path.join(os.path.dirname(__file__), filename)) as f:
return f.read()
NAME = 'django-genericfilters-demo'
README = read_relative_file('README')
VERSION = '0.1'
PACKAGES = ['demoproject']
REQUIRES = [
'django-generic-filters',
'mock',
'Django>=1.8,<1.12',
'coverage', ]
setup(name=NAME,
version=VERSION,
description='Demo project for django-generic-filters.',
long_description=README,
classifiers=['Development Status :: 1 - Planning',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 3',
'Framework :: Django',
],
keywords='class-based view, generic view, filters',
author='Novapost',
author_email='peopleask@novapost.fr',
url='https://github.com/novapost/django-generic-filters',
license='BSD',
packages=PACKAGES,
include_package_data=True,
zip_safe=False,
install_requires=REQUIRES,
entry_points={
'console_scripts': [
'demo = demoproject.manage:main',
]
},
)
| import os
from setuptools import setup, find_packages
def read_relative_file(filename):
"""Returns contents of the given file, which path is supposed relative
to this module."""
with open(os.path.join(os.path.dirname(__file__), filename)) as f:
return f.read()
NAME = 'django-genericfilters-demo'
README = read_relative_file('README')
VERSION = '0.1'
PACKAGES = find_packages()
REQUIRES = [
'django-generic-filters',
'mock',
'Django',
'coverage', ]
setup(name=NAME,
version=VERSION,
description='Demo project for django-generic-filters.',
long_description=README,
classifiers=['Development Status :: 1 - Planning',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 3',
'Framework :: Django',
],
keywords='class-based view, generic view, filters',
author='Novapost',
author_email='peopleask@novapost.fr',
url='https://github.com/novapost/django-generic-filters',
license='BSD',
packages=PACKAGES,
include_package_data=True,
zip_safe=False,
install_requires=REQUIRES,
entry_points={
'console_scripts': [
'demo = demoproject.manage:main',
]
},
)
| Fix demo. And *actually* test new Django versions | Fix demo. And *actually* test new Django versions
| Python | bsd-3-clause | novapost/django-generic-filters,novapost/django-generic-filters,novapost/django-generic-filters | ---
+++
@@ -1,5 +1,5 @@
import os
-from setuptools import setup
+from setuptools import setup, find_packages
def read_relative_file(filename):
@@ -12,11 +12,11 @@
NAME = 'django-genericfilters-demo'
README = read_relative_file('README')
VERSION = '0.1'
-PACKAGES = ['demoproject']
+PACKAGES = find_packages()
REQUIRES = [
'django-generic-filters',
'mock',
- 'Django>=1.8,<1.12',
+ 'Django',
'coverage', ]
|
99b1e11542c7e102e9fe739ea355b84d22b913d3 | src/experiments/forms.py | src/experiments/forms.py | from crispy_forms.bootstrap import FormActions
from crispy_forms.helper import FormHelper
from crispy_forms.layout import Fieldset, Layout, Submit
from django import forms
from django.utils.functional import cached_property
from django.utils.translation import ugettext_lazy as _
from core.widgets import SimpleMDEWidget
from .models import Experiment, Condition
class ExperimentCreateForm(forms.ModelForm):
class Meta:
model = Experiment
fields = ['name', 'description']
widgets = {
'description': SimpleMDEWidget(),
}
@cached_property
def helper(self):
helper = FormHelper()
helper.layout = Layout(
Fieldset(
'',
'name',
'description',
),
FormActions(
Submit(
'save', _('Create New Experiment'), css_class='btn-lg',
)
)
)
return helper
| from crispy_forms.bootstrap import FormActions
from crispy_forms.helper import FormHelper
from crispy_forms.layout import Fieldset, Layout, Submit, Div
from django import forms
from django.utils.functional import cached_property
from django.utils.translation import ugettext_lazy as _
from core.widgets import SimpleMDEWidget
from .models import Experiment, Condition
class ExperimentCreateForm(forms.ModelForm):
class Meta:
model = Experiment
fields = ['name', 'description']
widgets = {
'description': SimpleMDEWidget(),
}
@cached_property
def helper(self):
helper = FormHelper()
helper.layout = Layout(
Fieldset(
'',
'name',
'description',
),
)
return helper
| Disable form submit for now | Disable form submit for now
| Python | mit | ccwang002/biocloud-server-kai,ccwang002/biocloud-server-kai,ccwang002/biocloud-server-kai | ---
+++
@@ -1,6 +1,6 @@
from crispy_forms.bootstrap import FormActions
from crispy_forms.helper import FormHelper
-from crispy_forms.layout import Fieldset, Layout, Submit
+from crispy_forms.layout import Fieldset, Layout, Submit, Div
from django import forms
from django.utils.functional import cached_property
from django.utils.translation import ugettext_lazy as _
@@ -27,10 +27,5 @@
'name',
'description',
),
- FormActions(
- Submit(
- 'save', _('Create New Experiment'), css_class='btn-lg',
- )
- )
)
return helper |
0b632368b4991621696a7f7a396afecf61e6ccc3 | tools/examples/geturl.py | tools/examples/geturl.py | #!/usr/bin/env python2
#
# USAGE: geturl.py FILE_OR_DIR1 FILE_OR_DIR2 ...
#
# prints out the URL associated with each item
#
import sys
import svn._wc
import svn.util
def main(pool, files):
for f in files:
entry = svn._wc.svn_wc_entry(f, 0, pool)
print svn._wc.svn_wc_entry_t_url_get(entry)
if __name__ == '__main__':
svn.util.run_app(main, sys.argv[1:])
| #!/usr/bin/env python2
#
# USAGE: geturl.py FILE_OR_DIR1 FILE_OR_DIR2 ...
#
# prints out the URL associated with each item
#
import os
import sys
import svn.wc
import svn.util
def main(pool, files):
for f in files:
dirpath = fullpath = os.path.abspath(f)
if not os.path.isdir(dirpath):
dirpath = os.path.dirname(dirpath)
adm_baton = svn.wc.svn_wc_adm_open(None, dirpath, 1, 1, pool)
try:
entry = svn.wc.svn_wc_entry(fullpath, adm_baton, 0, pool)
print svn.wc.svn_wc_entry_t_url_get(entry)
except:
svn.wc.svn_wc_adm_close(adm_baton)
if __name__ == '__main__':
svn.util.run_app(main, sys.argv[1:])
| Update the example to use the new access baton stuff. | Update the example to use the new access baton stuff.
| Python | apache-2.0 | jmckaskill/subversion,jmckaskill/subversion,jmckaskill/subversion,jmckaskill/subversion,jmckaskill/subversion,jmckaskill/subversion,jmckaskill/subversion,jmckaskill/subversion | ---
+++
@@ -5,14 +5,23 @@
# prints out the URL associated with each item
#
+import os
import sys
-import svn._wc
+
+import svn.wc
import svn.util
def main(pool, files):
for f in files:
- entry = svn._wc.svn_wc_entry(f, 0, pool)
- print svn._wc.svn_wc_entry_t_url_get(entry)
+ dirpath = fullpath = os.path.abspath(f)
+ if not os.path.isdir(dirpath):
+ dirpath = os.path.dirname(dirpath)
+ adm_baton = svn.wc.svn_wc_adm_open(None, dirpath, 1, 1, pool)
+ try:
+ entry = svn.wc.svn_wc_entry(fullpath, adm_baton, 0, pool)
+ print svn.wc.svn_wc_entry_t_url_get(entry)
+ except:
+ svn.wc.svn_wc_adm_close(adm_baton)
if __name__ == '__main__':
svn.util.run_app(main, sys.argv[1:]) |
7148264a374301cb6cf9d35f3b17f3a02652600f | manage.py | manage.py | #!/usr/bin/env python
import os
from app import create_app, db
from app.models import User, Role, Permission
from flask.ext.script import Manager, Shell
from flask.ext.migrate import Migrate, MigrateCommand
app = create_app(os.getenv('FLASK_CONFIG') or 'default')
manager = Manager(app)
migrate = Migrate(app, db)
def make_shell_context():
return dict(app=app, db=db, User=User, Role=Role, Permission=Permission)
manager.add_command("shell", Shell(make_context=make_shell_context))
manager.add_command('db', MigrateCommand)
@manager.command
def test():
"""Run the unit tests."""
import unittest
tests = unittest.TestLoader().discover('tests')
unittest.TextTestRunner(verbosity=2).run(tests)
if __name__ == '__main__':
manager.run()
| #!/usr/bin/env python
import os
from app import create_app, db
from app.models import User, Role, Permission
from flask.ext.script import Manager, Shell
from flask.ext.migrate import Migrate, MigrateCommand
app = create_app(os.getenv('FLASK_CONFIG') or 'default')
manager = Manager(app)
migrate = Migrate(app, db)
def make_shell_context():
return dict(app=app, db=db, User=User, Role=Role, Permission=Permission)
manager.add_command("shell", Shell(make_context=make_shell_context))
manager.add_command('db', MigrateCommand)
@manager.command
def seed():
"""Seed the database."""
Role.insert_roles()
@manager.command
def test():
"""Run the unit tests."""
import unittest
tests = unittest.TestLoader().discover('tests')
unittest.TextTestRunner(verbosity=2).run(tests)
if __name__ == '__main__':
manager.run()
| Add command to seed database | Add command to seed database
| Python | mit | richgieg/flask-now,richgieg/flask-now | ---
+++
@@ -17,6 +17,12 @@
@manager.command
+def seed():
+ """Seed the database."""
+ Role.insert_roles()
+
+
+@manager.command
def test():
"""Run the unit tests."""
import unittest |
057e89c7beb8eb49321fe26513d61656b91ded6f | manage.py | manage.py | #!/usr/bin/env python
import os
import sys
if __name__ == "__main__":
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "tests.settings.dev")
from django.core.management import execute_from_command_line
execute_from_command_line(sys.argv)
| #!/usr/bin/env python
import os
import sys
if __name__ == "__main__":
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "tests.settings")
from django.core.management import execute_from_command_line
execute_from_command_line(sys.argv)
| Update the path to the setting file. | Update the path to the setting file.
| Python | bsd-3-clause | unt-libraries/django-premis-event-service,unt-libraries/django-premis-event-service,unt-libraries/django-premis-event-service | ---
+++
@@ -3,7 +3,7 @@
import sys
if __name__ == "__main__":
- os.environ.setdefault("DJANGO_SETTINGS_MODULE", "tests.settings.dev")
+ os.environ.setdefault("DJANGO_SETTINGS_MODULE", "tests.settings")
from django.core.management import execute_from_command_line
|
c1a8d0e1fdc0d2ebd822a041874ba6de8ed6d9ac | mod/default/mod_default.py | mod/default/mod_default.py | #! /usr/bin/env python
# coding:utf-8
from mod import Mod
import random
import os
class ModDefault(Mod):
def __init__(
self,
filename=None,
logger=None
):
Mod.__init__(self, logger)
text_path = filename or os.path.join(
os.path.abspath(os.path.dirname(__file__)),
"default.txt"
)
self.default_texts = [line.strip() for line in open(text_path)]
def is_fire(self, message, master):
return True
def reses(self, message, master):
return [
(random.random(),
text, "default", dict())
for text in self.default_texts
]
| #! /usr/bin/env python
# coding:utf-8
from mod import Mod
import random
import os
class ModDefault(Mod):
def __init__(
self,
filename=None,
logger=None
):
Mod.__init__(self, logger)
text_path = filename or os.path.join(
os.path.abspath(os.path.dirname(__file__)),
"default.txt"
)
self.default_texts = [line.strip() for line in open(text_path)]
def is_fire(self, message, master):
return True
def reses(self, message, master):
return [
(random.uniform(0, 0.3),
text, "default", dict())
for text in self.default_texts
]
| Use uniform instead of random | Use uniform instead of random
| Python | mit | kenkov/kovot,kenkov/kovot | ---
+++
@@ -25,7 +25,7 @@
def reses(self, message, master):
return [
- (random.random(),
+ (random.uniform(0, 0.3),
text, "default", dict())
for text in self.default_texts
] |
30ade212225344c9ca3e94cc47ee93e3b0b702b5 | website/addons/osfstorage/settings/defaults.py | website/addons/osfstorage/settings/defaults.py | # encoding: utf-8
import hashlib
from website import settings
DOMAIN = settings.DOMAIN
UPLOAD_SERVICE_URLS = ['changeme']
PING_TIMEOUT = 5 * 60
SIGNED_REQUEST_KWARGS = {}
# HMAC options
SIGNATURE_HEADER_KEY = 'X-Signature'
URLS_HMAC_SECRET = 'changeme'
URLS_HMAC_DIGEST = hashlib.sha1
WEBHOOK_HMAC_SECRET = 'changeme'
WEBHOOK_HMAC_DIGEST = hashlib.sha1
REVISIONS_PAGE_SIZE = 10
# IDENTITY = {
# 'provider': 's3',
# 'access_key': '',
# 'secret_key': ''
# }
WATERBUTLER_CREDENTIALS = {
'username': 'changeme',
'token': 'changeme',
'region': 'changeme'
}
WATERBUTLER_SETTINGS = {
'provider': 'buttfiles',
'container': 'changeme',
}
WATERBUTLER_RESOURCE = 'container'
| # encoding: utf-8
import hashlib
from website import settings
DOMAIN = settings.DOMAIN
UPLOAD_SERVICE_URLS = ['changeme']
PING_TIMEOUT = 5 * 60
SIGNED_REQUEST_KWARGS = {}
# HMAC options
SIGNATURE_HEADER_KEY = 'X-Signature'
URLS_HMAC_SECRET = 'changeme'
URLS_HMAC_DIGEST = hashlib.sha1
WEBHOOK_HMAC_SECRET = 'changeme'
WEBHOOK_HMAC_DIGEST = hashlib.sha1
REVISIONS_PAGE_SIZE = 10
# IDENTITY = {
# 'provider': 's3',
# 'access_key': '',
# 'secret_key': ''
# }
WATERBUTLER_CREDENTIALS = {
'username': 'changeme',
'token': 'changeme',
'region': 'changeme'
}
WATERBUTLER_SETTINGS = {
'provider': 'cloudfiles',
'container': 'changeme',
}
WATERBUTLER_RESOURCE = 'container'
| Fix copy paste typo caused by cloud to butt | Fix copy paste typo caused by cloud to butt
| Python | apache-2.0 | RomanZWang/osf.io,zachjanicki/osf.io,billyhunt/osf.io,jolene-esposito/osf.io,ZobairAlijan/osf.io,revanthkolli/osf.io,HalcyonChimera/osf.io,alexschiller/osf.io,kch8qx/osf.io,mluke93/osf.io,SSJohns/osf.io,brandonPurvis/osf.io,aaxelb/osf.io,arpitar/osf.io,samchrisinger/osf.io,arpitar/osf.io,chennan47/osf.io,mluo613/osf.io,jinluyuan/osf.io,lyndsysimon/osf.io,caseyrygt/osf.io,billyhunt/osf.io,lamdnhan/osf.io,amyshi188/osf.io,cosenal/osf.io,TomHeatwole/osf.io,cosenal/osf.io,monikagrabowska/osf.io,baylee-d/osf.io,crcresearch/osf.io,Nesiehr/osf.io,HalcyonChimera/osf.io,SSJohns/osf.io,rdhyee/osf.io,Nesiehr/osf.io,doublebits/osf.io,felliott/osf.io,doublebits/osf.io,HalcyonChimera/osf.io,bdyetton/prettychart,icereval/osf.io,jeffreyliu3230/osf.io,zamattiac/osf.io,chrisseto/osf.io,caseyrygt/osf.io,doublebits/osf.io,jnayak1/osf.io,hmoco/osf.io,dplorimer/osf,alexschiller/osf.io,sbt9uc/osf.io,mluo613/osf.io,kushG/osf.io,aaxelb/osf.io,danielneis/osf.io,sloria/osf.io,dplorimer/osf,lamdnhan/osf.io,CenterForOpenScience/osf.io,erinspace/osf.io,sloria/osf.io,brianjgeiger/osf.io,GageGaskins/osf.io,haoyuchen1992/osf.io,samchrisinger/osf.io,ckc6cz/osf.io,jmcarp/osf.io,binoculars/osf.io,GageGaskins/osf.io,lamdnhan/osf.io,samchrisinger/osf.io,adlius/osf.io,pattisdr/osf.io,Johnetordoff/osf.io,ticklemepierce/osf.io,baylee-d/osf.io,haoyuchen1992/osf.io,cslzchen/osf.io,Nesiehr/osf.io,petermalcolm/osf.io,rdhyee/osf.io,acshi/osf.io,arpitar/osf.io,kch8qx/osf.io,lyndsysimon/osf.io,amyshi188/osf.io,Nesiehr/osf.io,wearpants/osf.io,jnayak1/osf.io,hmoco/osf.io,dplorimer/osf,caneruguz/osf.io,asanfilippo7/osf.io,aaxelb/osf.io,ticklemepierce/osf.io,njantrania/osf.io,zamattiac/osf.io,alexschiller/osf.io,chennan47/osf.io,mfraezz/osf.io,jolene-esposito/osf.io,acshi/osf.io,petermalcolm/osf.io,caseyrollins/osf.io,zkraime/osf.io,saradbowman/osf.io,zamattiac/osf.io,hmoco/osf.io,Ghalko/osf.io,abought/osf.io,samanehsan/osf.io,mluke93/osf.io,ZobairAlijan/osf.io,kushG/osf.io,samchrisinger/osf.io,amyshi188/osf.io,jolene-esposito/osf.io,cldershem/osf.io,sbt9uc/osf.io,samanehsan/osf.io,fabianvf/osf.io,CenterForOpenScience/osf.io,acshi/osf.io,Ghalko/osf.io,reinaH/osf.io,cosenal/osf.io,fabianvf/osf.io,zkraime/osf.io,erinspace/osf.io,MerlinZhang/osf.io,saradbowman/osf.io,zkraime/osf.io,jmcarp/osf.io,aaxelb/osf.io,haoyuchen1992/osf.io,rdhyee/osf.io,jinluyuan/osf.io,kch8qx/osf.io,wearpants/osf.io,fabianvf/osf.io,KAsante95/osf.io,DanielSBrown/osf.io,MerlinZhang/osf.io,haoyuchen1992/osf.io,HalcyonChimera/osf.io,GaryKriebel/osf.io,felliott/osf.io,himanshuo/osf.io,CenterForOpenScience/osf.io,jmcarp/osf.io,CenterForOpenScience/osf.io,billyhunt/osf.io,danielneis/osf.io,pattisdr/osf.io,asanfilippo7/osf.io,rdhyee/osf.io,brianjgeiger/osf.io,jolene-esposito/osf.io,monikagrabowska/osf.io,barbour-em/osf.io,icereval/osf.io,SSJohns/osf.io,Ghalko/osf.io,emetsger/osf.io,asanfilippo7/osf.io,caseyrygt/osf.io,monikagrabowska/osf.io,billyhunt/osf.io,chennan47/osf.io,alexschiller/osf.io,MerlinZhang/osf.io,cwisecarver/osf.io,petermalcolm/osf.io,ticklemepierce/osf.io,RomanZWang/osf.io,RomanZWang/osf.io,billyhunt/osf.io,caseyrollins/osf.io,leb2dg/osf.io,brianjgeiger/osf.io,reinaH/osf.io,crcresearch/osf.io,bdyetton/prettychart,mattclark/osf.io,cwisecarver/osf.io,RomanZWang/osf.io,Johnetordoff/osf.io,abought/osf.io,reinaH/osf.io,zachjanicki/osf.io,reinaH/osf.io,jeffreyliu3230/osf.io,lamdnhan/osf.io,adlius/osf.io,Johnetordoff/osf.io,TomBaxter/osf.io,emetsger/osf.io,baylee-d/osf.io,cldershem/osf.io,crcresearch/osf.io,danielneis/osf.io,felliott/osf.io,kch8qx/osf.io,DanielSBrown/osf.io,jinluyuan/osf.io,MerlinZhang/osf.io,mfraezz/osf.io,laurenrevere/osf.io,njantrania/osf.io,caseyrollins/osf.io,mluke93/osf.io,GageGaskins/osf.io,monikagrabowska/osf.io,mluke93/osf.io,GageGaskins/osf.io,monikagrabowska/osf.io,zkraime/osf.io,barbour-em/osf.io,hmoco/osf.io,doublebits/osf.io,cldershem/osf.io,Johnetordoff/osf.io,petermalcolm/osf.io,cosenal/osf.io,kwierman/osf.io,KAsante95/osf.io,caneruguz/osf.io,GaryKriebel/osf.io,binoculars/osf.io,kushG/osf.io,Ghalko/osf.io,ckc6cz/osf.io,jinluyuan/osf.io,cslzchen/osf.io,brandonPurvis/osf.io,kushG/osf.io,bdyetton/prettychart,caneruguz/osf.io,brianjgeiger/osf.io,sbt9uc/osf.io,HarryRybacki/osf.io,revanthkolli/osf.io,cwisecarver/osf.io,KAsante95/osf.io,ticklemepierce/osf.io,ZobairAlijan/osf.io,kwierman/osf.io,acshi/osf.io,brandonPurvis/osf.io,mluo613/osf.io,mluo613/osf.io,GaryKriebel/osf.io,erinspace/osf.io,KAsante95/osf.io,sbt9uc/osf.io,jmcarp/osf.io,leb2dg/osf.io,wearpants/osf.io,TomHeatwole/osf.io,himanshuo/osf.io,felliott/osf.io,HarryRybacki/osf.io,zamattiac/osf.io,TomHeatwole/osf.io,sloria/osf.io,wearpants/osf.io,mfraezz/osf.io,dplorimer/osf,GaryKriebel/osf.io,TomBaxter/osf.io,himanshuo/osf.io,asanfilippo7/osf.io,doublebits/osf.io,HarryRybacki/osf.io,mattclark/osf.io,brandonPurvis/osf.io,amyshi188/osf.io,lyndsysimon/osf.io,caseyrygt/osf.io,mattclark/osf.io,DanielSBrown/osf.io,cslzchen/osf.io,ckc6cz/osf.io,leb2dg/osf.io,jnayak1/osf.io,danielneis/osf.io,pattisdr/osf.io,jnayak1/osf.io,GageGaskins/osf.io,chrisseto/osf.io,bdyetton/prettychart,binoculars/osf.io,emetsger/osf.io,ZobairAlijan/osf.io,RomanZWang/osf.io,brandonPurvis/osf.io,SSJohns/osf.io,njantrania/osf.io,caneruguz/osf.io,laurenrevere/osf.io,abought/osf.io,barbour-em/osf.io,laurenrevere/osf.io,revanthkolli/osf.io,revanthkolli/osf.io,barbour-em/osf.io,himanshuo/osf.io,chrisseto/osf.io,kwierman/osf.io,adlius/osf.io,zachjanicki/osf.io,KAsante95/osf.io,mfraezz/osf.io,kwierman/osf.io,abought/osf.io,icereval/osf.io,lyndsysimon/osf.io,leb2dg/osf.io,emetsger/osf.io,mluo613/osf.io,cwisecarver/osf.io,alexschiller/osf.io,fabianvf/osf.io,jeffreyliu3230/osf.io,cldershem/osf.io,arpitar/osf.io,zachjanicki/osf.io,kch8qx/osf.io,DanielSBrown/osf.io,samanehsan/osf.io,TomBaxter/osf.io,TomHeatwole/osf.io,samanehsan/osf.io,chrisseto/osf.io,jeffreyliu3230/osf.io,acshi/osf.io,ckc6cz/osf.io,adlius/osf.io,cslzchen/osf.io,njantrania/osf.io,HarryRybacki/osf.io | ---
+++
@@ -33,7 +33,7 @@
}
WATERBUTLER_SETTINGS = {
- 'provider': 'buttfiles',
+ 'provider': 'cloudfiles',
'container': 'changeme',
}
|
a6017e692ee8d602a228a91ad76b344459292bbf | src/hatchit/urls.py | src/hatchit/urls.py | from django.conf.urls import patterns, include, url
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
url(r'^$', 'event_manager.views.home', name='home'),
url(r'^admin/', include(admin.site.urls)),
)
| from django.conf.urls import patterns, include, url
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
url(r'^$', 'event_manager.views.home', name='home'),
url(r'^s/$', 'event_manager.views.my_suggestions', name='suggestions'),
url(r'^e/$', 'event_manager.views.my_events', name='events'),
url(r'^admin/', include(admin.site.urls)),
)
| Add routes for new views | Add routes for new views | Python | agpl-3.0 | DavidJFelix/hatchit,DavidJFelix/hatchit,DavidJFelix/hatchit | ---
+++
@@ -5,5 +5,7 @@
urlpatterns = patterns('',
url(r'^$', 'event_manager.views.home', name='home'),
+ url(r'^s/$', 'event_manager.views.my_suggestions', name='suggestions'),
+ url(r'^e/$', 'event_manager.views.my_events', name='events'),
url(r'^admin/', include(admin.site.urls)),
) |
df84686302697bbc2cce70828dd7cb3ad5ee65d5 | avacon_website/apps/simpleregistration/models.py | avacon_website/apps/simpleregistration/models.py |
from selvbetjening.core.events.models import request_attendee_pks_signal, find_attendee_signal, Attend
def avacon_style_attendee_pks_handler(sender, **kwargs):
attendee = kwargs['attendee']
key = 'Avacon.%s.%s.%s' % (attendee.event.pk,
attendee.user.pk,
attendee.invoice.pk)
return ('Avacon-Style ID', key)
request_attendee_pks_signal.connect(avacon_style_attendee_pks_handler)
def avacon_style_find_attendee_handler(sender, **kwargs):
pk = kwargs['pk']
try:
avacon, event_pk, user_pk, invoice_pk = pk.split('.')
attendee = Attend.objects.get(user__pk=user_pk,
invoice__pk=invoice_pk,
event__pk=event_pk)
return ('Avacon', attendee)
except:
return None
find_attendee_signal.connect(avacon_style_find_attendee_handler) | Add avacon-style payment ids like the ones used at svscon | Add avacon-style payment ids like the ones used at svscon
| Python | mit | sema/selvbetjening-avacon,sema/selvbetjening-avacon | ---
+++
@@ -0,0 +1,31 @@
+
+from selvbetjening.core.events.models import request_attendee_pks_signal, find_attendee_signal, Attend
+
+
+def avacon_style_attendee_pks_handler(sender, **kwargs):
+ attendee = kwargs['attendee']
+ key = 'Avacon.%s.%s.%s' % (attendee.event.pk,
+ attendee.user.pk,
+ attendee.invoice.pk)
+
+ return ('Avacon-Style ID', key)
+
+request_attendee_pks_signal.connect(avacon_style_attendee_pks_handler)
+
+
+def avacon_style_find_attendee_handler(sender, **kwargs):
+ pk = kwargs['pk']
+
+ try:
+ avacon, event_pk, user_pk, invoice_pk = pk.split('.')
+
+ attendee = Attend.objects.get(user__pk=user_pk,
+ invoice__pk=invoice_pk,
+ event__pk=event_pk)
+
+ return ('Avacon', attendee)
+
+ except:
+ return None
+
+find_attendee_signal.connect(avacon_style_find_attendee_handler) | |
3ceb39e4bbc4c5de7cbcce9c1ecfe94daa57266e | zhihudaily/models.py | zhihudaily/models.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from __future__ import absolute_import, unicode_literals
from peewee import Model, IntegerField, CharField
from zhihudaily.configs import Config
class BaseModel(Model):
class Meta:
database = Config.database
class Zhihudaily(BaseModel):
date = IntegerField()
json_news = CharField()
display_date = CharField()
def create_tables():
database.connect()
database.create_tables([Zhihudaily])
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
from __future__ import absolute_import, unicode_literals
from peewee import Model, IntegerField, CharField
from zhihudaily.configs import Config
class BaseModel(Model):
class Meta:
database = Config.database
class Zhihudaily(BaseModel):
date = IntegerField()
json_news = CharField()
display_date = CharField()
def create_tables():
Config.database.connect()
Config.database.create_tables([Zhihudaily])
| Fix bug when create the datebase table | Fix bug when create the datebase table
| Python | mit | lord63/zhihudaily,lord63/zhihudaily,lord63/zhihudaily | ---
+++
@@ -20,5 +20,5 @@
def create_tables():
- database.connect()
- database.create_tables([Zhihudaily])
+ Config.database.connect()
+ Config.database.create_tables([Zhihudaily]) |
ba3009d1d19243c703743070df68ad1ea11d454e | addons/hr_contract/__terp__.py | addons/hr_contract/__terp__.py | {
"name" : "Human Resources Contracts",
"version" : "1.0",
"author" : "Tiny",
"category" : "Generic Modules/Human Resources",
"website" : "http://tinyerp.com/module_hr.html",
"depends" : ["hr"],
"module": "",
"description": """
Add all information on the employee form to manage contracts:
* Martial status,
* Security number,
* Place of birth, birth date, ...
You can assign several contracts per employee.
""",
"init_xml" : ["hr_contract_data.xml"],
"demo_xml" : [],
"update_xml" : ["hr_contract_view.xml"],
"active": False,
"installable": True
}
| {
"name" : "Human Resources Contracts",
"version" : "1.0",
"author" : "Tiny",
"category" : "Generic Modules/Human Resources",
"website" : "http://tinyerp.com/module_hr.html",
"depends" : ["hr"],
"module": "",
"description": """
Add all information on the employee form to manage contracts:
* Martial status,
* Security number,
* Place of birth, birth date, ...
You can assign several contracts per employee.
""",
"init_xml" : ["hr_contract_data.xml"],
"demo_xml" : [],
"update_xml" : [
"hr_contract_view.xml",
"hr_contract_security.xml"
],
"active": False,
"installable": True
}
| Add hr_contract_security.xml file entry in update_xml section | Add hr_contract_security.xml file entry in update_xml section
bzr revid: mga@tinyerp.com-cceb329e8e30908ba1f6594b9502506150d105b2 | Python | agpl-3.0 | ygol/odoo,fossoult/odoo,leoliujie/odoo,hoatle/odoo,vrenaville/ngo-addons-backport,matrixise/odoo,cedk/odoo,chiragjogi/odoo,lsinfo/odoo,Nick-OpusVL/odoo,ihsanudin/odoo,dfang/odoo,gsmartway/odoo,ojengwa/odoo,windedge/odoo,tangyiyong/odoo,agrista/odoo-saas,mlaitinen/odoo,Daniel-CA/odoo,incaser/odoo-odoo,VitalPet/odoo,xujb/odoo,ojengwa/odoo,frouty/odoo_oph,ClearCorp-dev/odoo,sysadminmatmoz/OCB,syci/OCB,klunwebale/odoo,jusdng/odoo,VielSoft/odoo,goliveirab/odoo,Nowheresly/odoo,VielSoft/odoo,gorjuce/odoo,ccomb/OpenUpgrade,ChanduERP/odoo,fuselock/odoo,odoousers2014/odoo,acshan/odoo,patmcb/odoo,PongPi/isl-odoo,hmen89/odoo,Eric-Zhong/odoo,zchking/odoo,guerrerocarlos/odoo,Noviat/odoo,naousse/odoo,bakhtout/odoo-educ,AuyaJackie/odoo,jiangzhixiao/odoo,kittiu/odoo,Kilhog/odoo,xzYue/odoo,bguillot/OpenUpgrade,tangyiyong/odoo,jesramirez/odoo,markeTIC/OCB,shingonoide/odoo,xzYue/odoo,savoirfairelinux/odoo,matrixise/odoo,dariemp/odoo,ClearCorp-dev/odoo,dkubiak789/odoo,naousse/odoo,x111ong/odoo,virgree/odoo,odoousers2014/odoo,Nick-OpusVL/odoo,OSSESAC/odoopubarquiluz,lombritz/odoo,Drooids/odoo,Eric-Zhong/odoo,TRESCLOUD/odoopub,chiragjogi/odoo,prospwro/odoo,lsinfo/odoo,QianBIG/odoo,factorlibre/OCB,Drooids/odoo,florian-dacosta/OpenUpgrade,bakhtout/odoo-educ,guerrerocarlos/odoo,erkrishna9/odoo,syci/OCB,dkubiak789/odoo,FlorianLudwig/odoo,mvaled/OpenUpgrade,JonathanStein/odoo,jiachenning/odoo,Antiun/odoo,ShineFan/odoo,joshuajan/odoo,brijeshkesariya/odoo,Maspear/odoo,rubencabrera/odoo,PongPi/isl-odoo,RafaelTorrealba/odoo,TRESCLOUD/odoopub,blaggacao/OpenUpgrade,ingadhoc/odoo,gvb/odoo,leorochael/odoo,rdeheele/odoo,ovnicraft/odoo,colinnewell/odoo,gsmartway/odoo,syci/OCB,thanhacun/odoo,hubsaysnuaa/odoo,kybriainfotech/iSocioCRM,Elico-Corp/odoo_OCB,CopeX/odoo,codekaki/odoo,bkirui/odoo,bplancher/odoo,JonathanStein/odoo,massot/odoo,florian-dacosta/OpenUpgrade,Kilhog/odoo,BT-ojossen/odoo,slevenhagen/odoo,mmbtba/odoo,BT-ojossen/odoo,SerpentCS/odoo,ojengwa/odoo,ygol/odoo,charbeljc/OCB,ThinkOpen-Solutions/odoo,ihsanudin/odoo,synconics/odoo,stonegithubs/odoo,mkieszek/odoo,optima-ict/odoo,rschnapka/odoo,takis/odoo,Adel-Magebinary/odoo,brijeshkesariya/odoo,florian-dacosta/OpenUpgrade,OpenUpgrade-dev/OpenUpgrade,MarcosCommunity/odoo,Grirrane/odoo,aviciimaxwell/odoo,juanalfonsopr/odoo,CubicERP/odoo,nagyistoce/odoo-dev-odoo,hip-odoo/odoo,podemos-info/odoo,jfpla/odoo,alexcuellar/odoo,optima-ict/odoo,cedk/odoo,ihsanudin/odoo,chiragjogi/odoo,OpenUpgrade/OpenUpgrade,jeasoft/odoo,lombritz/odoo,luiseduardohdbackup/odoo,vnsofthe/odoo,xujb/odoo,apanju/GMIO_Odoo,Ichag/odoo,srsman/odoo,mmbtba/odoo,BT-rmartin/odoo,addition-it-solutions/project-all,hip-odoo/odoo,andreparames/odoo,SerpentCS/odoo,ramadhane/odoo,OpenUpgrade-dev/OpenUpgrade,savoirfairelinux/OpenUpgrade,Nowheresly/odoo,fuhongliang/odoo,mlaitinen/odoo,ingadhoc/odoo,tinkerthaler/odoo,fossoult/odoo,microcom/odoo,apocalypsebg/odoo,bealdav/OpenUpgrade,GauravSahu/odoo,kirca/OpenUpgrade,OpenUpgrade-dev/OpenUpgrade,apanju/odoo,dfang/odoo,oasiswork/odoo,ingadhoc/odoo,OSSESAC/odoopubarquiluz,oliverhr/odoo,fgesora/odoo,gdgellatly/OCB1,hassoon3/odoo,janocat/odoo,mustafat/odoo-1,vnsofthe/odoo,osvalr/odoo,Drooids/odoo,slevenhagen/odoo,credativUK/OCB,tvibliani/odoo,sysadminmatmoz/OCB,CatsAndDogsbvba/odoo,blaggacao/OpenUpgrade,spadae22/odoo,sve-odoo/odoo,NeovaHealth/odoo,mkieszek/odoo,sv-dev1/odoo,feroda/odoo,hassoon3/odoo,apocalypsebg/odoo,SAM-IT-SA/odoo,savoirfairelinux/OpenUpgrade,prospwro/odoo,pedrobaeza/odoo,alexcuellar/odoo,lombritz/odoo,OSSESAC/odoopubarquiluz,Nowheresly/odoo,florentx/OpenUpgrade,x111ong/odoo,sysadminmatmoz/OCB,shaufi10/odoo,tangyiyong/odoo,datenbetrieb/odoo,Grirrane/odoo,dllsf/odootest,kybriainfotech/iSocioCRM,slevenhagen/odoo,shivam1111/odoo,oliverhr/odoo,ojengwa/odoo,rahuldhote/odoo,ovnicraft/odoo,cysnake4713/odoo,fjbatresv/odoo,csrocha/OpenUpgrade,andreparames/odoo,bkirui/odoo,BT-rmartin/odoo,datenbetrieb/odoo,JCA-Developpement/Odoo,elmerdpadilla/iv,abenzbiria/clients_odoo,gsmartway/odoo,dezynetechnologies/odoo,javierTerry/odoo,apanju/odoo,camptocamp/ngo-addons-backport,takis/odoo,patmcb/odoo,minhtuancn/odoo,storm-computers/odoo,fevxie/odoo,CopeX/odoo,erkrishna9/odoo,Drooids/odoo,omprakasha/odoo,stephen144/odoo,shivam1111/odoo,joshuajan/odoo,Grirrane/odoo,kifcaliph/odoo,colinnewell/odoo,mmbtba/odoo,javierTerry/odoo,microcom/odoo,thanhacun/odoo,mvaled/OpenUpgrade,kittiu/odoo,alexteodor/odoo,credativUK/OCB,synconics/odoo,srsman/odoo,apanju/odoo,markeTIC/OCB,poljeff/odoo,CatsAndDogsbvba/odoo,fuhongliang/odoo,steedos/odoo,BT-ojossen/odoo,kittiu/odoo,aviciimaxwell/odoo,fevxie/odoo,laslabs/odoo,diagramsoftware/odoo,cysnake4713/odoo,incaser/odoo-odoo,waytai/odoo,ovnicraft/odoo,agrista/odoo-saas,omprakasha/odoo,omprakasha/odoo,BT-astauder/odoo,jeasoft/odoo,minhtuancn/odoo,ygol/odoo,rubencabrera/odoo,grap/OCB,leorochael/odoo,Ichag/odoo,numerigraphe/odoo,BT-fgarbely/odoo,omprakasha/odoo,jusdng/odoo,diagramsoftware/odoo,sinbazhou/odoo,podemos-info/odoo,rubencabrera/odoo,bakhtout/odoo-educ,diagramsoftware/odoo,salaria/odoo,optima-ict/odoo,Nowheresly/odoo,shaufi/odoo,dalegregory/odoo,oihane/odoo,mvaled/OpenUpgrade,jfpla/odoo,patmcb/odoo,bwrsandman/OpenUpgrade,deKupini/erp,ramadhane/odoo,factorlibre/OCB,Nowheresly/odoo,luiseduardohdbackup/odoo,shaufi10/odoo,CopeX/odoo,laslabs/odoo,windedge/odoo,abdellatifkarroum/odoo,ujjwalwahi/odoo,mvaled/OpenUpgrade,fuhongliang/odoo,grap/OCB,BT-ojossen/odoo,gavin-feng/odoo,diagramsoftware/odoo,markeTIC/OCB,doomsterinc/odoo,oliverhr/odoo,salaria/odoo,nitinitprof/odoo,dezynetechnologies/odoo,numerigraphe/odoo,Elico-Corp/odoo_OCB,OpenPymeMx/OCB,addition-it-solutions/project-all,ihsanudin/odoo,doomsterinc/odoo,incaser/odoo-odoo,shingonoide/odoo,avoinsystems/odoo,salaria/odoo,mmbtba/odoo,Kilhog/odoo,numerigraphe/odoo,Bachaco-ve/odoo,tinkerthaler/odoo,shaufi/odoo,storm-computers/odoo,shingonoide/odoo,CatsAndDogsbvba/odoo,fdvarela/odoo8,tinkhaven-organization/odoo,slevenhagen/odoo-npg,savoirfairelinux/OpenUpgrade,ubic135/odoo-design,Gitlab11/odoo,alqfahad/odoo,Nick-OpusVL/odoo,guerrerocarlos/odoo,OpenUpgrade/OpenUpgrade,alhashash/odoo,vnsofthe/odoo,windedge/odoo,collex100/odoo,storm-computers/odoo,rahuldhote/odoo,massot/odoo,nagyistoce/odoo-dev-odoo,funkring/fdoo,dgzurita/odoo,eino-makitalo/odoo,savoirfairelinux/odoo,hanicker/odoo,sysadminmatmoz/OCB,srsman/odoo,apanju/odoo,hifly/OpenUpgrade,tinkerthaler/odoo,fjbatresv/odoo,bealdav/OpenUpgrade,SerpentCS/odoo,lightcn/odoo,ujjwalwahi/odoo,synconics/odoo,hopeall/odoo,jiangzhixiao/odoo,PongPi/isl-odoo,papouso/odoo,vnsofthe/odoo,bplancher/odoo,pplatek/odoo,stonegithubs/odoo,fevxie/odoo,Eric-Zhong/odoo,damdam-s/OpenUpgrade,nagyistoce/odoo-dev-odoo,RafaelTorrealba/odoo,brijeshkesariya/odoo,papouso/odoo,ovnicraft/odoo,BT-fgarbely/odoo,datenbetrieb/odoo,abdellatifkarroum/odoo,Nick-OpusVL/odoo,lgscofield/odoo,csrocha/OpenUpgrade,bplancher/odoo,ShineFan/odoo,rowemoore/odoo,chiragjogi/odoo,sv-dev1/odoo,datenbetrieb/odoo,ramadhane/odoo,dariemp/odoo,ThinkOpen-Solutions/odoo,sebalix/OpenUpgrade,doomsterinc/odoo,sinbazhou/odoo,funkring/fdoo,OpenPymeMx/OCB,ygol/odoo,Danisan/odoo-1,stonegithubs/odoo,elmerdpadilla/iv,dsfsdgsbngfggb/odoo,odoo-turkiye/odoo,aviciimaxwell/odoo,VitalPet/odoo,leoliujie/odoo,grap/OpenUpgrade,kybriainfotech/iSocioCRM,xujb/odoo,ubic135/odoo-design,fevxie/odoo,apanju/odoo,stonegithubs/odoo,sinbazhou/odoo,glovebx/odoo,avoinsystems/odoo,ChanduERP/odoo,gvb/odoo,kifcaliph/odoo,Adel-Magebinary/odoo,cedk/odoo,BT-rmartin/odoo,odooindia/odoo,oasiswork/odoo,arthru/OpenUpgrade,nitinitprof/odoo,grap/OCB,SAM-IT-SA/odoo,joariasl/odoo,Nick-OpusVL/odoo,christophlsa/odoo,shivam1111/odoo,rowemoore/odoo,takis/odoo,tvtsoft/odoo8,JGarcia-Panach/odoo,cedk/odoo,waytai/odoo,oasiswork/odoo,oihane/odoo,tangyiyong/odoo,gavin-feng/odoo,hopeall/odoo,JGarcia-Panach/odoo,slevenhagen/odoo-npg,bkirui/odoo,Endika/OpenUpgrade,abenzbiria/clients_odoo,sergio-incaser/odoo,glovebx/odoo,ramitalat/odoo,GauravSahu/odoo,Bachaco-ve/odoo,nhomar/odoo,janocat/odoo,odootr/odoo,ApuliaSoftware/odoo,makinacorpus/odoo,NeovaHealth/odoo,PongPi/isl-odoo,fevxie/odoo,CubicERP/odoo,Endika/OpenUpgrade,arthru/OpenUpgrade,ApuliaSoftware/odoo,shingonoide/odoo,hoatle/odoo,brijeshkesariya/odoo,matrixise/odoo,joariasl/odoo,rahuldhote/odoo,frouty/odoogoeen,wangjun/odoo,patmcb/odoo,tarzan0820/odoo,PongPi/isl-odoo,odoo-turkiye/odoo,diagramsoftware/odoo,ApuliaSoftware/odoo,damdam-s/OpenUpgrade,omprakasha/odoo,rgeleta/odoo,guerrerocarlos/odoo,alqfahad/odoo,nhomar/odoo,RafaelTorrealba/odoo,lombritz/odoo,realsaiko/odoo,AuyaJackie/odoo,ingadhoc/odoo,dalegregory/odoo,cpyou/odoo,colinnewell/odoo,numerigraphe/odoo,funkring/fdoo,cedk/odoo,csrocha/OpenUpgrade,charbeljc/OCB,tinkerthaler/odoo,salaria/odoo,jpshort/odoo,provaleks/o8,sebalix/OpenUpgrade,jfpla/odoo,RafaelTorrealba/odoo,dgzurita/odoo,vnsofthe/odoo,nexiles/odoo,xzYue/odoo,nexiles/odoo,kirca/OpenUpgrade,hbrunn/OpenUpgrade,klunwebale/odoo,christophlsa/odoo,JonathanStein/odoo,sve-odoo/odoo,janocat/odoo,srimai/odoo,makinacorpus/odoo,javierTerry/odoo,luistorresm/odoo,inspyration/odoo,jusdng/odoo,Adel-Magebinary/odoo,realsaiko/odoo,QianBIG/odoo,hip-odoo/odoo,ramadhane/odoo,jolevq/odoopub,GauravSahu/odoo,frouty/odoo_oph,alhashash/odoo,QianBIG/odoo,arthru/OpenUpgrade,andreparames/odoo,Danisan/odoo-1,cpyou/odoo,dllsf/odootest,Endika/odoo,draugiskisprendimai/odoo,dezynetechnologies/odoo,markeTIC/OCB,mustafat/odoo-1,JGarcia-Panach/odoo,OpusVL/odoo,tinkhaven-organization/odoo,OSSESAC/odoopubarquiluz,RafaelTorrealba/odoo,fuselock/odoo,OpenPymeMx/OCB,tarzan0820/odoo,rahuldhote/odoo,credativUK/OCB,hbrunn/OpenUpgrade,ujjwalwahi/odoo,frouty/odoogoeen,ShineFan/odoo,goliveirab/odoo,sv-dev1/odoo,jeasoft/odoo,steedos/odoo,Codefans-fan/odoo,TRESCLOUD/odoopub,hifly/OpenUpgrade,ecosoft-odoo/odoo,acshan/odoo,dsfsdgsbngfggb/odoo,nhomar/odoo-mirror,thanhacun/odoo,brijeshkesariya/odoo,lsinfo/odoo,jaxkodex/odoo,ihsanudin/odoo,nhomar/odoo-mirror,dfang/odoo,Daniel-CA/odoo,havt/odoo,slevenhagen/odoo,hbrunn/OpenUpgrade,Daniel-CA/odoo,andreparames/odoo,fuselock/odoo,VitalPet/odoo,rgeleta/odoo,credativUK/OCB,erkrishna9/odoo,odoo-turkiye/odoo,Endika/odoo,florentx/OpenUpgrade,nuuuboo/odoo,OpenUpgrade/OpenUpgrade,guewen/OpenUpgrade,sergio-incaser/odoo,x111ong/odoo,VielSoft/odoo,abdellatifkarroum/odoo,abenzbiria/clients_odoo,camptocamp/ngo-addons-backport,tvtsoft/odoo8,BT-astauder/odoo,massot/odoo,sebalix/OpenUpgrade,joariasl/odoo,acshan/odoo,tarzan0820/odoo,lsinfo/odoo,JGarcia-Panach/odoo,dgzurita/odoo,chiragjogi/odoo,BT-rmartin/odoo,zchking/odoo,funkring/fdoo,bwrsandman/OpenUpgrade,bwrsandman/OpenUpgrade,rdeheele/odoo,bealdav/OpenUpgrade,0k/odoo,synconics/odoo,massot/odoo,blaggacao/OpenUpgrade,vrenaville/ngo-addons-backport,NeovaHealth/odoo,OpenPymeMx/OCB,bguillot/OpenUpgrade,stephen144/odoo,Adel-Magebinary/odoo,mszewczy/odoo,QianBIG/odoo,fossoult/odoo,odooindia/odoo,sadleader/odoo,abdellatifkarroum/odoo,Ernesto99/odoo,alexteodor/odoo,vrenaville/ngo-addons-backport,brijeshkesariya/odoo,hifly/OpenUpgrade,hubsaysnuaa/odoo,Endika/odoo,Ichag/odoo,luistorresm/odoo,kittiu/odoo,leorochael/odoo,vnsofthe/odoo,OpenPymeMx/OCB,AuyaJackie/odoo,ramadhane/odoo,patmcb/odoo,VielSoft/odoo,waytai/odoo,OpusVL/odoo,jiachenning/odoo,CatsAndDogsbvba/odoo,bplancher/odoo,credativUK/OCB,avoinsystems/odoo,mlaitinen/odoo,nexiles/odoo,BT-fgarbely/odoo,nitinitprof/odoo,tinkerthaler/odoo,makinacorpus/odoo,jaxkodex/odoo,camptocamp/ngo-addons-backport,luiseduardohdbackup/odoo,tvibliani/odoo,bobisme/odoo,oliverhr/odoo,ChanduERP/odoo,FlorianLudwig/odoo,christophlsa/odoo,waytai/odoo,pedrobaeza/OpenUpgrade,eino-makitalo/odoo,mmbtba/odoo,papouso/odoo,gvb/odoo,hubsaysnuaa/odoo,sinbazhou/odoo,NL66278/OCB,poljeff/odoo,KontorConsulting/odoo,savoirfairelinux/odoo,abstract-open-solutions/OCB,hmen89/odoo,rowemoore/odoo,shivam1111/odoo,numerigraphe/odoo,BT-rmartin/odoo,prospwro/odoo,FlorianLudwig/odoo,n0m4dz/odoo,fevxie/odoo,spadae22/odoo,kifcaliph/odoo,bplancher/odoo,pedrobaeza/odoo,tvtsoft/odoo8,ingadhoc/odoo,massot/odoo,datenbetrieb/odoo,slevenhagen/odoo-npg,alqfahad/odoo,odootr/odoo,havt/odoo,BT-rmartin/odoo,0k/OpenUpgrade,bwrsandman/OpenUpgrade,draugiskisprendimai/odoo,prospwro/odoo,mustafat/odoo-1,sinbazhou/odoo,jusdng/odoo,sergio-incaser/odoo,papouso/odoo,cloud9UG/odoo,rdeheele/odoo,nuuuboo/odoo,ClearCorp-dev/odoo,luistorresm/odoo,srimai/odoo,syci/OCB,simongoffin/website_version,ramitalat/odoo,steedos/odoo,ccomb/OpenUpgrade,agrista/odoo-saas,colinnewell/odoo,hopeall/odoo,addition-it-solutions/project-all,oliverhr/odoo,waytai/odoo,Nick-OpusVL/odoo,abdellatifkarroum/odoo,klunwebale/odoo,abdellatifkarroum/odoo,cloud9UG/odoo,juanalfonsopr/odoo,shaufi/odoo,rschnapka/odoo,oihane/odoo,Codefans-fan/odoo,lombritz/odoo,oasiswork/odoo,OpenUpgrade/OpenUpgrade,mustafat/odoo-1,jusdng/odoo,virgree/odoo,tarzan0820/odoo,fuselock/odoo,tvibliani/odoo,ramadhane/odoo,joshuajan/odoo,matrixise/odoo,jesramirez/odoo,shivam1111/odoo,credativUK/OCB,VitalPet/odoo,MarcosCommunity/odoo,jeasoft/odoo,windedge/odoo,dllsf/odootest,NL66278/OCB,hoatle/odoo,lgscofield/odoo,kybriainfotech/iSocioCRM,arthru/OpenUpgrade,optima-ict/odoo,hanicker/odoo,blaggacao/OpenUpgrade,goliveirab/odoo,provaleks/o8,hassoon3/odoo,charbeljc/OCB,frouty/odoogoeen,Grirrane/odoo,dsfsdgsbngfggb/odoo,OpenPymeMx/OCB,Antiun/odoo,janocat/odoo,x111ong/odoo,sinbazhou/odoo,gavin-feng/odoo,Kilhog/odoo,alexcuellar/odoo,waytai/odoo,sadleader/odoo,srimai/odoo,sergio-incaser/odoo,Elico-Corp/odoo_OCB,nhomar/odoo,bakhtout/odoo-educ,storm-computers/odoo,abenzbiria/clients_odoo,jfpla/odoo,joshuajan/odoo,odootr/odoo,apanju/GMIO_Odoo,srimai/odoo,OpenUpgrade-dev/OpenUpgrade,tarzan0820/odoo,tvibliani/odoo,abdellatifkarroum/odoo,0k/odoo,zchking/odoo,tinkerthaler/odoo,xujb/odoo,hopeall/odoo,hassoon3/odoo,Ichag/odoo,aviciimaxwell/odoo,inspyration/odoo,srsman/odoo,fgesora/odoo,TRESCLOUD/odoopub,bobisme/odoo,bakhtout/odoo-educ,virgree/odoo,bkirui/odoo,savoirfairelinux/OpenUpgrade,rowemoore/odoo,grap/OpenUpgrade,gdgellatly/OCB1,grap/OCB,hmen89/odoo,havt/odoo,KontorConsulting/odoo,NeovaHealth/odoo,nitinitprof/odoo,salaria/odoo,SerpentCS/odoo,pplatek/odoo,realsaiko/odoo,tarzan0820/odoo,KontorConsulting/odoo,xujb/odoo,codekaki/odoo,stephen144/odoo,x111ong/odoo,funkring/fdoo,idncom/odoo,cysnake4713/odoo,ramitalat/odoo,jiangzhixiao/odoo,hoatle/odoo,kittiu/odoo,hip-odoo/odoo,Drooids/odoo,ApuliaSoftware/odoo,jeasoft/odoo,oliverhr/odoo,omprakasha/odoo,aviciimaxwell/odoo,realsaiko/odoo,jiachenning/odoo,stephen144/odoo,rgeleta/odoo,dllsf/odootest,odooindia/odoo,0k/OpenUpgrade,sebalix/OpenUpgrade,NeovaHealth/odoo,JGarcia-Panach/odoo,OpenUpgrade/OpenUpgrade,Ernesto99/odoo,factorlibre/OCB,csrocha/OpenUpgrade,mszewczy/odoo,pedrobaeza/odoo,lightcn/odoo,hopeall/odoo,fuhongliang/odoo,elmerdpadilla/iv,Bachaco-ve/odoo,dkubiak789/odoo,xzYue/odoo,doomsterinc/odoo,Antiun/odoo,odoousers2014/odoo,xujb/odoo,OSSESAC/odoopubarquiluz,idncom/odoo,Elico-Corp/odoo_OCB,Adel-Magebinary/odoo,cdrooom/odoo,Gitlab11/odoo,PongPi/isl-odoo,pedrobaeza/OpenUpgrade,nitinitprof/odoo,synconics/odoo,apanju/GMIO_Odoo,vnsofthe/odoo,fossoult/odoo,Maspear/odoo,makinacorpus/odoo,nuncjo/odoo,FlorianLudwig/odoo,dfang/odoo,zchking/odoo,minhtuancn/odoo,csrocha/OpenUpgrade,gorjuce/odoo,sinbazhou/odoo,grap/OpenUpgrade,leoliujie/odoo,pedrobaeza/OpenUpgrade,feroda/odoo,acshan/odoo,ApuliaSoftware/odoo,vrenaville/ngo-addons-backport,sebalix/OpenUpgrade,nuuuboo/odoo,gorjuce/odoo,fgesora/odoo,camptocamp/ngo-addons-backport,bkirui/odoo,rschnapka/odoo,ygol/odoo,dkubiak789/odoo,odoousers2014/odoo,cloud9UG/odoo,hmen89/odoo,pplatek/odoo,bobisme/odoo,luistorresm/odoo,aviciimaxwell/odoo,n0m4dz/odoo,lgscofield/odoo,cysnake4713/odoo,CopeX/odoo,thanhacun/odoo,JCA-Developpement/Odoo,apanju/GMIO_Odoo,shingonoide/odoo,Antiun/odoo,hifly/OpenUpgrade,stephen144/odoo,highco-groupe/odoo,nagyistoce/odoo-dev-odoo,oliverhr/odoo,podemos-info/odoo,SAM-IT-SA/odoo,jolevq/odoopub,colinnewell/odoo,colinnewell/odoo,OpenUpgrade-dev/OpenUpgrade,markeTIC/OCB,sadleader/odoo,apocalypsebg/odoo,Gitlab11/odoo,fuhongliang/odoo,cysnake4713/odoo,x111ong/odoo,bwrsandman/OpenUpgrade,realsaiko/odoo,datenbetrieb/odoo,alexcuellar/odoo,fdvarela/odoo8,Adel-Magebinary/odoo,klunwebale/odoo,pedrobaeza/odoo,codekaki/odoo,xzYue/odoo,odoo-turkiye/odoo,Noviat/odoo,ChanduERP/odoo,AuyaJackie/odoo,gvb/odoo,jfpla/odoo,sysadminmatmoz/OCB,papouso/odoo,rubencabrera/odoo,jiachenning/odoo,eino-makitalo/odoo,camptocamp/ngo-addons-backport,sve-odoo/odoo,osvalr/odoo,janocat/odoo,virgree/odoo,Antiun/odoo,SAM-IT-SA/odoo,tvtsoft/odoo8,nitinitprof/odoo,JGarcia-Panach/odoo,spadae22/odoo,Codefans-fan/odoo,christophlsa/odoo,collex100/odoo,gorjuce/odoo,spadae22/odoo,guewen/OpenUpgrade,goliveirab/odoo,shaufi10/odoo,incaser/odoo-odoo,bealdav/OpenUpgrade,idncom/odoo,microcom/odoo,osvalr/odoo,ojengwa/odoo,alexcuellar/odoo,klunwebale/odoo,funkring/fdoo,highco-groupe/odoo,ccomb/OpenUpgrade,brijeshkesariya/odoo,ramitalat/odoo,idncom/odoo,makinacorpus/odoo,savoirfairelinux/OpenUpgrade,hopeall/odoo,gdgellatly/OCB1,hifly/OpenUpgrade,nexiles/odoo,doomsterinc/odoo,Kilhog/odoo,jesramirez/odoo,jolevq/odoopub,luiseduardohdbackup/odoo,Codefans-fan/odoo,steedos/odoo,fgesora/odoo,charbeljc/OCB,prospwro/odoo,ecosoft-odoo/odoo,BT-ojossen/odoo,storm-computers/odoo,jfpla/odoo,CopeX/odoo,jiangzhixiao/odoo,VitalPet/odoo,n0m4dz/odoo,fuhongliang/odoo,laslabs/odoo,grap/OpenUpgrade,tinkhaven-organization/odoo,cloud9UG/odoo,ecosoft-odoo/odoo,provaleks/o8,mszewczy/odoo,OpenUpgrade/OpenUpgrade,wangjun/odoo,ihsanudin/odoo,slevenhagen/odoo-npg,zchking/odoo,JCA-Developpement/Odoo,janocat/odoo,thanhacun/odoo,dfang/odoo,CatsAndDogsbvba/odoo,fjbatresv/odoo,oihane/odoo,eino-makitalo/odoo,sv-dev1/odoo,spadae22/odoo,joariasl/odoo,BT-rmartin/odoo,bobisme/odoo,eino-makitalo/odoo,hbrunn/OpenUpgrade,doomsterinc/odoo,frouty/odoogoeen,minhtuancn/odoo,abstract-open-solutions/OCB,gavin-feng/odoo,grap/OCB,dezynetechnologies/odoo,jolevq/odoopub,GauravSahu/odoo,tangyiyong/odoo,gorjuce/odoo,feroda/odoo,optima-ict/odoo,Bachaco-ve/odoo,mkieszek/odoo,ehirt/odoo,salaria/odoo,odoo-turkiye/odoo,Bachaco-ve/odoo,csrocha/OpenUpgrade,luiseduardohdbackup/odoo,frouty/odoo_oph,ccomb/OpenUpgrade,BT-astauder/odoo,steedos/odoo,jpshort/odoo,chiragjogi/odoo,Elico-Corp/odoo_OCB,leoliujie/odoo,janocat/odoo,acshan/odoo,demon-ru/iml-crm,JonathanStein/odoo,oasiswork/odoo,syci/OCB,gvb/odoo,Ichag/odoo,srsman/odoo,n0m4dz/odoo,dariemp/odoo,nuncjo/odoo,mustafat/odoo-1,shingonoide/odoo,odootr/odoo,jaxkodex/odoo,stephen144/odoo,dariemp/odoo,csrocha/OpenUpgrade,papouso/odoo,rgeleta/odoo,Maspear/odoo,GauravSahu/odoo,xzYue/odoo,OpusVL/odoo,Danisan/odoo-1,gsmartway/odoo,fuselock/odoo,srsman/odoo,odoo-turkiye/odoo,mustafat/odoo-1,alhashash/odoo,dariemp/odoo,hmen89/odoo,lgscofield/odoo,juanalfonsopr/odoo,grap/OpenUpgrade,pplatek/odoo,slevenhagen/odoo,wangjun/odoo,Danisan/odoo-1,frouty/odoogoeen,Ichag/odoo,CubicERP/odoo,damdam-s/OpenUpgrade,damdam-s/OpenUpgrade,nhomar/odoo-mirror,kirca/OpenUpgrade,tangyiyong/odoo,arthru/OpenUpgrade,joariasl/odoo,ehirt/odoo,pedrobaeza/OpenUpgrade,VielSoft/odoo,odootr/odoo,florentx/OpenUpgrade,florian-dacosta/OpenUpgrade,ojengwa/odoo,zchking/odoo,glovebx/odoo,gavin-feng/odoo,Endika/OpenUpgrade,RafaelTorrealba/odoo,mvaled/OpenUpgrade,nhomar/odoo,Noviat/odoo,SAM-IT-SA/odoo,jaxkodex/odoo,nhomar/odoo,codekaki/odoo,podemos-info/odoo,guewen/OpenUpgrade,factorlibre/OCB,colinnewell/odoo,shaufi10/odoo,charbeljc/OCB,odootr/odoo,hoatle/odoo,bplancher/odoo,ubic135/odoo-design,ApuliaSoftware/odoo,draugiskisprendimai/odoo,fjbatresv/odoo,osvalr/odoo,joshuajan/odoo,leorochael/odoo,eino-makitalo/odoo,rgeleta/odoo,arthru/OpenUpgrade,srimai/odoo,fgesora/odoo,gorjuce/odoo,Gitlab11/odoo,guewen/OpenUpgrade,joariasl/odoo,windedge/odoo,poljeff/odoo,apocalypsebg/odoo,nitinitprof/odoo,hoatle/odoo,Endika/OpenUpgrade,jesramirez/odoo,NL66278/OCB,lightcn/odoo,christophlsa/odoo,Codefans-fan/odoo,ThinkOpen-Solutions/odoo,ramitalat/odoo,hanicker/odoo,JonathanStein/odoo,gvb/odoo,poljeff/odoo,dllsf/odootest,mlaitinen/odoo,matrixise/odoo,sergio-incaser/odoo,bobisme/odoo,Danisan/odoo-1,kifcaliph/odoo,lightcn/odoo,shingonoide/odoo,hassoon3/odoo,guewen/OpenUpgrade,camptocamp/ngo-addons-backport,abstract-open-solutions/OCB,codekaki/odoo,Maspear/odoo,andreparames/odoo,laslabs/odoo,codekaki/odoo,nagyistoce/odoo-dev-odoo,juanalfonsopr/odoo,shaufi/odoo,cloud9UG/odoo,sebalix/OpenUpgrade,hifly/OpenUpgrade,odooindia/odoo,BT-ojossen/odoo,bkirui/odoo,ingadhoc/odoo,ovnicraft/odoo,jfpla/odoo,Grirrane/odoo,Daniel-CA/odoo,jiangzhixiao/odoo,javierTerry/odoo,kittiu/odoo,nuuuboo/odoo,shaufi/odoo,luistorresm/odoo,srsman/odoo,mvaled/OpenUpgrade,naousse/odoo,simongoffin/website_version,tinkhaven-organization/odoo,bguillot/OpenUpgrade,apanju/GMIO_Odoo,FlorianLudwig/odoo,nhomar/odoo-mirror,AuyaJackie/odoo,dkubiak789/odoo,mszewczy/odoo,draugiskisprendimai/odoo,highco-groupe/odoo,markeTIC/OCB,doomsterinc/odoo,ecosoft-odoo/odoo,eino-makitalo/odoo,ThinkOpen-Solutions/odoo,OpenPymeMx/OCB,alhashash/odoo,Gitlab11/odoo,podemos-info/odoo,kybriainfotech/iSocioCRM,frouty/odoogoeen,OSSESAC/odoopubarquiluz,gdgellatly/OCB1,erkrishna9/odoo,ovnicraft/odoo,mmbtba/odoo,OpenUpgrade-dev/OpenUpgrade,jusdng/odoo,avoinsystems/odoo,naousse/odoo,elmerdpadilla/iv,dsfsdgsbngfggb/odoo,fuselock/odoo,pedrobaeza/OpenUpgrade,luiseduardohdbackup/odoo,JonathanStein/odoo,hanicker/odoo,ujjwalwahi/odoo,nuuuboo/odoo,Noviat/odoo,luiseduardohdbackup/odoo,Elico-Corp/odoo_OCB,jiangzhixiao/odoo,windedge/odoo,Antiun/odoo,xujb/odoo,minhtuancn/odoo,erkrishna9/odoo,Nowheresly/odoo,mszewczy/odoo,klunwebale/odoo,rahuldhote/odoo,cdrooom/odoo,andreparames/odoo,idncom/odoo,VitalPet/odoo,Daniel-CA/odoo,demon-ru/iml-crm,guerrerocarlos/odoo,guewen/OpenUpgrade,luistorresm/odoo,Endika/odoo,klunwebale/odoo,gavin-feng/odoo,dariemp/odoo,shaufi10/odoo,highco-groupe/odoo,MarcosCommunity/odoo,nexiles/odoo,0k/OpenUpgrade,slevenhagen/odoo-npg,poljeff/odoo,makinacorpus/odoo,CubicERP/odoo,hip-odoo/odoo,ShineFan/odoo,ChanduERP/odoo,KontorConsulting/odoo,grap/OpenUpgrade,oihane/odoo,gsmartway/odoo,juanalfonsopr/odoo,Endika/OpenUpgrade,dsfsdgsbngfggb/odoo,abstract-open-solutions/OCB,jiachenning/odoo,funkring/fdoo,kirca/OpenUpgrade,Maspear/odoo,hubsaysnuaa/odoo,havt/odoo,gorjuce/odoo,sv-dev1/odoo,dgzurita/odoo,draugiskisprendimai/odoo,javierTerry/odoo,rowemoore/odoo,abenzbiria/clients_odoo,stonegithubs/odoo,Bachaco-ve/odoo,rdeheele/odoo,zchking/odoo,sergio-incaser/odoo,ccomb/OpenUpgrade,alexcuellar/odoo,ChanduERP/odoo,glovebx/odoo,avoinsystems/odoo,demon-ru/iml-crm,slevenhagen/odoo-npg,shaufi/odoo,0k/OpenUpgrade,nexiles/odoo,mkieszek/odoo,lightcn/odoo,alhashash/odoo,factorlibre/OCB,optima-ict/odoo,avoinsystems/odoo,rschnapka/odoo,dsfsdgsbngfggb/odoo,fuhongliang/odoo,cedk/odoo,fjbatresv/odoo,shaufi/odoo,bguillot/OpenUpgrade,ecosoft-odoo/odoo,Maspear/odoo,dkubiak789/odoo,kirca/OpenUpgrade,frouty/odoogoeen,steedos/odoo,takis/odoo,Endika/odoo,fossoult/odoo,MarcosCommunity/odoo,poljeff/odoo,bobisme/odoo,nuncjo/odoo,lgscofield/odoo,sysadminmatmoz/OCB,ChanduERP/odoo,apanju/odoo,sv-dev1/odoo,dezynetechnologies/odoo,dkubiak789/odoo,SAM-IT-SA/odoo,collex100/odoo,0k/odoo,tvibliani/odoo,Nick-OpusVL/odoo,gvb/odoo,blaggacao/OpenUpgrade,FlorianLudwig/odoo,gdgellatly/OCB1,virgree/odoo,Eric-Zhong/odoo,cpyou/odoo,alqfahad/odoo,Maspear/odoo,minhtuancn/odoo,mustafat/odoo-1,acshan/odoo,mkieszek/odoo,MarcosCommunity/odoo,Ernesto99/odoo,Ichag/odoo,virgree/odoo,feroda/odoo,hassoon3/odoo,sebalix/OpenUpgrade,spadae22/odoo,javierTerry/odoo,lightcn/odoo,apanju/GMIO_Odoo,Nowheresly/odoo,credativUK/OCB,ramadhane/odoo,kirca/OpenUpgrade,prospwro/odoo,provaleks/o8,dalegregory/odoo,n0m4dz/odoo,hifly/OpenUpgrade,0k/odoo,Ernesto99/odoo,ehirt/odoo,guerrerocarlos/odoo,SAM-IT-SA/odoo,leoliujie/odoo,Danisan/odoo-1,dalegregory/odoo,cpyou/odoo,havt/odoo,Adel-Magebinary/odoo,ujjwalwahi/odoo,Ernesto99/odoo,n0m4dz/odoo,KontorConsulting/odoo,Endika/odoo,rschnapka/odoo,apanju/GMIO_Odoo,apocalypsebg/odoo,FlorianLudwig/odoo,vrenaville/ngo-addons-backport,bakhtout/odoo-educ,florian-dacosta/OpenUpgrade,blaggacao/OpenUpgrade,lightcn/odoo,rschnapka/odoo,tinkerthaler/odoo,gavin-feng/odoo,dalegregory/odoo,hbrunn/OpenUpgrade,alexcuellar/odoo,BT-fgarbely/odoo,jeasoft/odoo,slevenhagen/odoo,damdam-s/OpenUpgrade,camptocamp/ngo-addons-backport,VielSoft/odoo,camptocamp/ngo-addons-backport,codekaki/odoo,juanalfonsopr/odoo,ehirt/odoo,apocalypsebg/odoo,tangyiyong/odoo,mmbtba/odoo,fevxie/odoo,ingadhoc/odoo,alqfahad/odoo,jpshort/odoo,ShineFan/odoo,cloud9UG/odoo,pedrobaeza/OpenUpgrade,synconics/odoo,tinkhaven-organization/odoo,syci/OCB,slevenhagen/odoo,ujjwalwahi/odoo,Codefans-fan/odoo,ehirt/odoo,jpshort/odoo,Ernesto99/odoo,factorlibre/OCB,christophlsa/odoo,NL66278/OCB,ShineFan/odoo,fdvarela/odoo8,bakhtout/odoo-educ,jolevq/odoopub,incaser/odoo-odoo,Endika/OpenUpgrade,vrenaville/ngo-addons-backport,deKupini/erp,waytai/odoo,SerpentCS/odoo,pplatek/odoo,tvibliani/odoo,florentx/OpenUpgrade,alqfahad/odoo,srimai/odoo,gdgellatly/OCB1,feroda/odoo,shaufi10/odoo,addition-it-solutions/project-all,sve-odoo/odoo,thanhacun/odoo,tarzan0820/odoo,sve-odoo/odoo,chiragjogi/odoo,damdam-s/OpenUpgrade,Noviat/odoo,mlaitinen/odoo,Bachaco-ve/odoo,ThinkOpen-Solutions/odoo,CopeX/odoo,papouso/odoo,CopeX/odoo,rgeleta/odoo,frouty/odoo_oph,oihane/odoo,SerpentCS/odoo,kittiu/odoo,microcom/odoo,QianBIG/odoo,addition-it-solutions/project-all,Drooids/odoo,KontorConsulting/odoo,pplatek/odoo,joshuajan/odoo,naousse/odoo,0k/OpenUpgrade,ehirt/odoo,SerpentCS/odoo,nuuuboo/odoo,storm-computers/odoo,pedrobaeza/odoo,fgesora/odoo,savoirfairelinux/odoo,bealdav/OpenUpgrade,QianBIG/odoo,savoirfairelinux/OpenUpgrade,bguillot/OpenUpgrade,GauravSahu/odoo,hanicker/odoo,lgscofield/odoo,JGarcia-Panach/odoo,alexteodor/odoo,poljeff/odoo,jusdng/odoo,dezynetechnologies/odoo,hubsaysnuaa/odoo,BT-fgarbely/odoo,lsinfo/odoo,inspyration/odoo,leorochael/odoo,BT-astauder/odoo,ubic135/odoo-design,MarcosCommunity/odoo,abstract-open-solutions/OCB,lsinfo/odoo,naousse/odoo,ojengwa/odoo,fjbatresv/odoo,avoinsystems/odoo,abstract-open-solutions/OCB,christophlsa/odoo,virgree/odoo,elmerdpadilla/iv,rubencabrera/odoo,BT-fgarbely/odoo,highco-groupe/odoo,Drooids/odoo,minhtuancn/odoo,mkieszek/odoo,odootr/odoo,deKupini/erp,feroda/odoo,windedge/odoo,bwrsandman/OpenUpgrade,CatsAndDogsbvba/odoo,fgesora/odoo,oasiswork/odoo,havt/odoo,simongoffin/website_version,acshan/odoo,Gitlab11/odoo,leorochael/odoo,apanju/odoo,diagramsoftware/odoo,jpshort/odoo,Gitlab11/odoo,podemos-info/odoo,ThinkOpen-Solutions/odoo,steedos/odoo,mszewczy/odoo,nuncjo/odoo,havt/odoo,shivam1111/odoo,jiangzhixiao/odoo,kybriainfotech/iSocioCRM,takis/odoo,demon-ru/iml-crm,goliveirab/odoo,collex100/odoo,dgzurita/odoo,ClearCorp-dev/odoo,jaxkodex/odoo,hanicker/odoo,nhomar/odoo-mirror,rdeheele/odoo,cpyou/odoo,xzYue/odoo,bealdav/OpenUpgrade,alexteodor/odoo,mszewczy/odoo,nuuuboo/odoo,slevenhagen/odoo-npg,laslabs/odoo,podemos-info/odoo,juanalfonsopr/odoo,patmcb/odoo,frouty/odoogoeen,jiachenning/odoo,javierTerry/odoo,tinkhaven-organization/odoo,0k/OpenUpgrade,collex100/odoo,MarcosCommunity/odoo,frouty/odoo_oph,simongoffin/website_version,fossoult/odoo,goliveirab/odoo,florentx/OpenUpgrade,agrista/odoo-saas,PongPi/isl-odoo,jpshort/odoo,alexteodor/odoo,jeasoft/odoo,synconics/odoo,GauravSahu/odoo,kirca/OpenUpgrade,rahuldhote/odoo,dariemp/odoo,Eric-Zhong/odoo,grap/OCB,Endika/OpenUpgrade,fdvarela/odoo8,stonegithubs/odoo,ihsanudin/odoo,osvalr/odoo,microcom/odoo,OpenUpgrade/OpenUpgrade,collex100/odoo,microcom/odoo,pedrobaeza/OpenUpgrade,0k/odoo,cdrooom/odoo,VielSoft/odoo,numerigraphe/odoo,ShineFan/odoo,wangjun/odoo,n0m4dz/odoo,sysadminmatmoz/OCB,alhashash/odoo,florian-dacosta/OpenUpgrade,x111ong/odoo,salaria/odoo,NeovaHealth/odoo,cdrooom/odoo,OpusVL/odoo,grap/OCB,rubencabrera/odoo,Noviat/odoo,stonegithubs/odoo,ThinkOpen-Solutions/odoo,diagramsoftware/odoo,ccomb/OpenUpgrade,JCA-Developpement/Odoo,NeovaHealth/odoo,tvibliani/odoo,JCA-Developpement/Odoo,blaggacao/OpenUpgrade,grap/OCB,oasiswork/odoo,Noviat/odoo,credativUK/OCB,fossoult/odoo,hanicker/odoo,odooindia/odoo,shivam1111/odoo,MarcosCommunity/odoo,nhomar/odoo,tvtsoft/odoo8,mlaitinen/odoo,glovebx/odoo,vrenaville/ngo-addons-backport,glovebx/odoo,naousse/odoo,sadleader/odoo,hubsaysnuaa/odoo,leoliujie/odoo,tinkhaven-organization/odoo,wangjun/odoo,Endika/odoo,ramitalat/odoo,jesramirez/odoo,odoo-turkiye/odoo,kybriainfotech/iSocioCRM,srimai/odoo,CubicERP/odoo,pplatek/odoo,bobisme/odoo,rowemoore/odoo,incaser/odoo-odoo,guewen/OpenUpgrade,vrenaville/ngo-addons-backport,simongoffin/website_version,OpenPymeMx/OCB,osvalr/odoo,feroda/odoo,ujjwalwahi/odoo,glovebx/odoo,nexiles/odoo,Eric-Zhong/odoo,prospwro/odoo,nagyistoce/odoo-dev-odoo,bwrsandman/OpenUpgrade,ecosoft-odoo/odoo,patmcb/odoo,Kilhog/odoo,charbeljc/OCB,CatsAndDogsbvba/odoo,dalegregory/odoo,draugiskisprendimai/odoo,charbeljc/OCB,florentx/OpenUpgrade,VitalPet/odoo,AuyaJackie/odoo,draugiskisprendimai/odoo,ecosoft-odoo/odoo,abstract-open-solutions/OCB,fdvarela/odoo8,KontorConsulting/odoo,wangjun/odoo,addition-it-solutions/project-all,dgzurita/odoo,ApuliaSoftware/odoo,damdam-s/OpenUpgrade,Kilhog/odoo,lombritz/odoo,provaleks/o8,rahuldhote/odoo,Codefans-fan/odoo,BT-ojossen/odoo,bkirui/odoo,gsmartway/odoo,Eric-Zhong/odoo,collex100/odoo,jaxkodex/odoo,frouty/odoo_oph,NL66278/OCB,ovnicraft/odoo,savoirfairelinux/odoo,sadleader/odoo,Danisan/odoo-1,dalegregory/odoo,Antiun/odoo,dsfsdgsbngfggb/odoo,lsinfo/odoo,jeasoft/odoo,cedk/odoo,hopeall/odoo,TRESCLOUD/odoopub,kifcaliph/odoo,dfang/odoo,factorlibre/OCB,hoatle/odoo,bguillot/OpenUpgrade,codekaki/odoo,gsmartway/odoo,numerigraphe/odoo,ehirt/odoo,nuncjo/odoo,laslabs/odoo,makinacorpus/odoo,ygol/odoo,Daniel-CA/odoo,incaser/odoo-odoo,thanhacun/odoo,oihane/odoo,gdgellatly/OCB1,leoliujie/odoo,Ernesto99/odoo,apocalypsebg/odoo,fuselock/odoo,dezynetechnologies/odoo,CubicERP/odoo,alqfahad/odoo,wangjun/odoo,omprakasha/odoo,markeTIC/OCB,andreparames/odoo,RafaelTorrealba/odoo,idncom/odoo,shaufi10/odoo,takis/odoo,agrista/odoo-saas,ccomb/OpenUpgrade,jpshort/odoo,lombritz/odoo,odoousers2014/odoo,spadae22/odoo,rgeleta/odoo,CubicERP/odoo,nagyistoce/odoo-dev-odoo,VitalPet/odoo,savoirfairelinux/odoo,JonathanStein/odoo,deKupini/erp,hip-odoo/odoo,rschnapka/odoo,BT-fgarbely/odoo,leorochael/odoo,provaleks/o8,takis/odoo,fjbatresv/odoo,cloud9UG/odoo,mvaled/OpenUpgrade,jaxkodex/odoo,idncom/odoo,datenbetrieb/odoo,hubsaysnuaa/odoo,mlaitinen/odoo,rschnapka/odoo,sv-dev1/odoo,AuyaJackie/odoo,rubencabrera/odoo,dgzurita/odoo,grap/OpenUpgrade,aviciimaxwell/odoo,provaleks/o8,gdgellatly/OCB1,odoousers2014/odoo,Grirrane/odoo,Daniel-CA/odoo,joariasl/odoo,luistorresm/odoo,deKupini/erp,osvalr/odoo,BT-astauder/odoo,tvtsoft/odoo8,ygol/odoo,hbrunn/OpenUpgrade,bguillot/OpenUpgrade,ClearCorp-dev/odoo,nuncjo/odoo,lgscofield/odoo,ubic135/odoo-design,inspyration/odoo,guerrerocarlos/odoo,pedrobaeza/odoo,nuncjo/odoo,rowemoore/odoo,demon-ru/iml-crm,goliveirab/odoo | ---
+++
@@ -16,7 +16,10 @@
""",
"init_xml" : ["hr_contract_data.xml"],
"demo_xml" : [],
- "update_xml" : ["hr_contract_view.xml"],
+ "update_xml" : [
+ "hr_contract_view.xml",
+ "hr_contract_security.xml"
+ ],
"active": False,
"installable": True
} |
7461666cde3c0206058d10f2341e0a57bf33e504 | src/renderers/status.py | src/renderers/status.py | from flask import Blueprint
from models import db, Status
import json
status_renderer = Blueprint('status', __name__)
@status_renderer.route('/status/<int:user_id>')
def get_tweet(user_id):
status = db.session.query(Status).order_by(Status.id.desc()).one()
return json.dumps({'type' : 'text',
'status_text' : status.status_text,
'posted_by' : status.posted_by,
'image_url' : status.pic_url,
'profile_pic': status.profile_pic
})
| from flask import Blueprint
from models import db, Status
import json
status_renderer = Blueprint('status', __name__)
@status_renderer.route('/status')
def get_status():
status = db.session.query(Status).order_by(Status.id.desc()).one()
return json.dumps({'type' : 'text',
'status_text' : status.status_text,
'posted_by' : status.posted_by,
'image_url' : status.pic_url,
'profile_pic': status.profile_pic
})
| Remove user_id requirement for FB endpoint | Remove user_id requirement for FB endpoint
| Python | mit | ndm25/notifyable | ---
+++
@@ -4,8 +4,8 @@
status_renderer = Blueprint('status', __name__)
-@status_renderer.route('/status/<int:user_id>')
-def get_tweet(user_id):
+@status_renderer.route('/status')
+def get_status():
status = db.session.query(Status).order_by(Status.id.desc()).one()
return json.dumps({'type' : 'text',
'status_text' : status.status_text, |
e877f379bdb864b4f4297796b5c4ed1535354772 | mcrouter/test/test_mcrouter_to_mcrouter_tko.py | mcrouter/test/test_mcrouter_to_mcrouter_tko.py | # Copyright (c) 2015, Facebook, Inc.
# All rights reserved.
#
# This source code is licensed under the BSD-style license found in the
# LICENSE file in the root directory of this source tree. An additional grant
# of patent rights can be found in the PATENTS file in the same directory.
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import re
from mcrouter.test.McrouterTestCase import McrouterTestCase
class TestMcrouterToMcrouterTko(McrouterTestCase):
config = './mcrouter/test/test_mcrouter_to_mcrouter_tko.json'
extra_args = ['--timeouts-until-tko', '1']
def setUp(self):
self.add_mcrouter(self.config)
def get_mcrouter(self):
return self.add_mcrouter(self.config, extra_args=self.extra_args)
def test_underlying_tko(self):
mcr = self.get_mcrouter()
self.assertFalse(mcr.delete("key"))
stats = mcr.stats("suspect_servers")
self.assertEqual(1, len(stats))
self.assertTrue(re.match("status:[tko|down]", stats.values()[0]))
| # Copyright (c) 2015, Facebook, Inc.
# All rights reserved.
#
# This source code is licensed under the BSD-style license found in the
# LICENSE file in the root directory of this source tree. An additional grant
# of patent rights can be found in the PATENTS file in the same directory.
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import re
from mcrouter.test.McrouterTestCase import McrouterTestCase
class TestMcrouterToMcrouterTko(McrouterTestCase):
config = './mcrouter/test/test_mcrouter_to_mcrouter_tko.json'
extra_args = ['--timeouts-until-tko', '1']
def setUp(self):
self.add_mcrouter(self.config)
def get_mcrouter(self):
return self.add_mcrouter(self.config, extra_args=self.extra_args)
def test_underlying_tko(self):
mcr = self.get_mcrouter()
self.assertFalse(mcr.delete("key"))
stats = mcr.stats("suspect_servers")
self.assertEqual(1, len(stats))
self.assertTrue(re.match("status:(tko|down)", stats.values()[0]))
| Correct regex in unit test | Correct regex in unit test
Summary: Correct regex used in unit tests.
Test Plan: Unit test
Reviewed By: @alikhtarov
Differential Revision: D1804582 | Python | mit | easyfmxu/mcrouter,facebook/mcrouter,leitao/mcrouter,reddit/mcrouter,synecdoche/mcrouter,facebook/mcrouter,glensc/mcrouter,evertrue/mcrouter,leitao/mcrouter,leitao/mcrouter,seem-sky/mcrouter,apinski-cavium/mcrouter,apinski-cavium/mcrouter,facebook/mcrouter,reddit/mcrouter,evertrue/mcrouter,yqzhang/mcrouter,reddit/mcrouter,zhlong73/mcrouter,reddit/mcrouter,nvaller/mcrouter,synecdoche/mcrouter,leitao/mcrouter,facebook/mcrouter,zhlong73/mcrouter,seem-sky/mcrouter,synecdoche/mcrouter,tempbottle/mcrouter,evertrue/mcrouter,glensc/mcrouter,seem-sky/mcrouter,tempbottle/mcrouter,yqzhang/mcrouter,easyfmxu/mcrouter,apinski-cavium/mcrouter,tempbottle/mcrouter,nvaller/mcrouter,apinski-cavium/mcrouter,nvaller/mcrouter,is00hcw/mcrouter,zhlong73/mcrouter,synecdoche/mcrouter,zhlong73/mcrouter,glensc/mcrouter,seem-sky/mcrouter,yqzhang/mcrouter,easyfmxu/mcrouter,is00hcw/mcrouter,nvaller/mcrouter,tempbottle/mcrouter,easyfmxu/mcrouter,is00hcw/mcrouter,yqzhang/mcrouter,evertrue/mcrouter,is00hcw/mcrouter,glensc/mcrouter | ---
+++
@@ -31,4 +31,4 @@
stats = mcr.stats("suspect_servers")
self.assertEqual(1, len(stats))
- self.assertTrue(re.match("status:[tko|down]", stats.values()[0]))
+ self.assertTrue(re.match("status:(tko|down)", stats.values()[0])) |
8861ff87f20de238ea3ddccd7958973dcf992e4a | us_ignite/people/urls.py | us_ignite/people/urls.py | from django.conf.urls import patterns, url
urlpatterns = patterns(
'us_ignite.people.views',
url(r'^$', 'profile_list', name='profile_list'),
url(r'^(?P<slug>[-\w]{1,32})/$', 'profile_detail', name='profile_detail'),
)
| from django.conf.urls import patterns, url
urlpatterns = patterns(
'us_ignite.people.views',
url(r'^$', 'profile_list', name='profile_list'),
url(r'^(?P<slug>[-_\w]{1,32})/$', 'profile_detail', name='profile_detail'),
)
| Allow underscores as the user profile slug. | Allow underscores as the user profile slug.
https://github.com/madewithbytes/us_ignite/issues/316
| Python | bsd-3-clause | us-ignite/us_ignite,us-ignite/us_ignite,us-ignite/us_ignite,us-ignite/us_ignite,us-ignite/us_ignite | ---
+++
@@ -3,5 +3,5 @@
urlpatterns = patterns(
'us_ignite.people.views',
url(r'^$', 'profile_list', name='profile_list'),
- url(r'^(?P<slug>[-\w]{1,32})/$', 'profile_detail', name='profile_detail'),
+ url(r'^(?P<slug>[-_\w]{1,32})/$', 'profile_detail', name='profile_detail'),
) |
7d9c7133de36d2fd7587d7be361cd0ff964d4e94 | deflect/urls.py | deflect/urls.py | from django.conf.urls import patterns
from django.conf.urls import url
from .views import redirect
urlpatterns = patterns('',
url(r'^(?P<key>[a-zA-Z0-9]+)$', redirect, name='deflect-redirect'),
)
| from django.conf import settings
from django.conf.urls import patterns
from django.conf.urls import url
from .views import alias
from .views import redirect
urlpatterns = patterns('',
url(r'^(?P<key>[a-zA-Z0-9]+)$', redirect, name='deflect-redirect'),
)
alias_prefix = getattr(settings, 'DEFLECT_ALIAS_PREFIX', '')
if alias_prefix:
urlpatterns += patterns('',
url(r'^%s(?P<key>[a-zA-Z0-9]+)$' % alias_prefix, alias,
name='deflect-alias'),
)
| Add custom URL alias paths to URLconf | Add custom URL alias paths to URLconf
| Python | bsd-3-clause | jbittel/django-deflect | ---
+++
@@ -1,9 +1,18 @@
+from django.conf import settings
from django.conf.urls import patterns
from django.conf.urls import url
+from .views import alias
from .views import redirect
urlpatterns = patterns('',
url(r'^(?P<key>[a-zA-Z0-9]+)$', redirect, name='deflect-redirect'),
)
+
+alias_prefix = getattr(settings, 'DEFLECT_ALIAS_PREFIX', '')
+if alias_prefix:
+ urlpatterns += patterns('',
+ url(r'^%s(?P<key>[a-zA-Z0-9]+)$' % alias_prefix, alias,
+ name='deflect-alias'),
+ ) |
403ad86bde44c1a015d8d35ac2826221ef98f9da | drftest/shop/api/views.py | drftest/shop/api/views.py | from django.shortcuts import render
from rest_framework.views import APIView
from rest_framework.response import Response
from rest_framework import status
from rest_framework.permissions import IsAuthenticated
class ShopAPIView(APIView):
permission_classes = (IsAuthenticated,)
class OrdersView(ShopAPIView):
"""
Place an order. Create the "order" record with general informations, create the "order_details" records with
the details of the order. During this transaction any stamp earned by the user is added to the database and
at the end voucher(s) are created if there are enough stamps available for that user.
"""
def post(self, request, format=None):
return Response({'success': True})
| from django.shortcuts import render
from rest_framework.views import APIView
from rest_framework.response import Response
from rest_framework import status
from rest_framework.permissions import IsAuthenticated
from django.db import transaction
class ShopAPIView(APIView):
permission_classes = (IsAuthenticated,)
class OrdersView(ShopAPIView):
"""
Place an order. Create the "order" record with general informations, create the "order_details" records with
the details of the order. During this transaction any stamp earned by the user is added to the database and
at the end voucher(s) are created if there are enough stamps available for that user.
"""
def post(self, request, format=None):
with transaction.atomic():
return Response({'success': True})
| Add transaction support to the orders view | Add transaction support to the orders view
| Python | mit | andreagrandi/drf3-test,andreagrandi/drf3-test,andreagrandi/drf3-test | ---
+++
@@ -3,10 +3,10 @@
from rest_framework.response import Response
from rest_framework import status
from rest_framework.permissions import IsAuthenticated
+from django.db import transaction
class ShopAPIView(APIView):
permission_classes = (IsAuthenticated,)
-
class OrdersView(ShopAPIView):
"""
@@ -15,4 +15,5 @@
at the end voucher(s) are created if there are enough stamps available for that user.
"""
def post(self, request, format=None):
- return Response({'success': True})
+ with transaction.atomic():
+ return Response({'success': True}) |
fbbe6b46b93274567e031d2ba7874fe3231b1557 | openacademy/model/openacademy_course.py | openacademy/model/openacademy_course.py | # -*- coding: utf-8 -*-
from openerp import models, fields, api
'''
This module create model of Course
'''
class Course(models.Model):
'''
This class create model of Course
'''
_name = 'openacademy.course' # model odoo name
name = fields.Char(string="Title", required=True)
description = fields.Text()
responsible_id = fields.Many2one('res.users',
ondelete='set null', string="Responsible", index=True)
session_ids = fields.One2many('openacademy.session',
'course_id', string="Sessions")
_sql_constraints = [
('name_description_check',
'CHECK(name != description)',
"The title of the course should not be the description"),
('name_unique',
'UNIQUE(name)',
"The course title must be unique"),
]
| # -*- coding: utf-8 -*-
from openerp import models, fields, api
'''
This module create model of Course
'''
class Course(models.Model):
'''
This class create model of Course
'''
_name = 'openacademy.course' # model odoo name
name = fields.Char(string="Title", required=True)
description = fields.Text()
responsible_id = fields.Many2one('res.users',
ondelete='set null', string="Responsible", index=True)
session_ids = fields.One2many('openacademy.session',
'course_id', string="Sessions")
@api.multi # send defaults params: cr, uid, id , context
def copy(self, default=None):
default = dict(default or {})
copied_count = self.search_count(
[('name', '=like', u"Copy of {}%".format(self.name))])
if not copied_count:
new_name = u"Copy of {}".format(self.name)
else:
new_name = u"Copy of {} ({})".format(self.name, copied_count)
default['name'] = new_name
return super(Course, self).copy(default)
_sql_constraints = [
('name_description_check',
'CHECK(name != description)',
"The title of the course should not be the description"),
('name_unique',
'UNIQUE(name)',
"The course title must be unique"),
]
| Modify copy method into inherit | [REF] openacademy: Modify copy method into inherit
| Python | apache-2.0 | lescobarvx/curso-tecnico-odoo | ---
+++
@@ -19,6 +19,21 @@
session_ids = fields.One2many('openacademy.session',
'course_id', string="Sessions")
+ @api.multi # send defaults params: cr, uid, id , context
+ def copy(self, default=None):
+ default = dict(default or {})
+
+ copied_count = self.search_count(
+ [('name', '=like', u"Copy of {}%".format(self.name))])
+ if not copied_count:
+ new_name = u"Copy of {}".format(self.name)
+ else:
+ new_name = u"Copy of {} ({})".format(self.name, copied_count)
+
+ default['name'] = new_name
+ return super(Course, self).copy(default)
+
+
_sql_constraints = [
('name_description_check',
'CHECK(name != description)', |
979d1b8ce4567432a6816b55fe5c29ef7c190459 | aospy_user/calcs/tendencies.py | aospy_user/calcs/tendencies.py | """Calculations involved in mass and energy budgets."""
import numpy as np
from aospy.utils import coord_to_new_dataarray
from .. import TIME_STR
def first_to_last_vals_dur(arr, freq='1M'):
"""Time elapsed between 1st and last values in each given time period."""
time = coord_to_new_dataarray(arr, TIME_STR)
first = time.resample(freq, TIME_STR, how='first')
last = time.resample(freq, TIME_STR, how='last')
delta_time = last - first
# Divide by a 1 sec timedelta to convert to seconds.
delta_time.values = delta_time.values / np.timedelta64(1, 's')
return delta_time
def time_tendency(arr, freq='1M'):
"""Monthly time tendency of the given field."""
first = arr.resample(freq, TIME_STR, how='first').dropna(TIME_STR)
last = arr.resample(freq, TIME_STR, how='last').dropna(TIME_STR)
return (last - first) / first_to_last_vals_dur(arr, freq)
| """Calculations involved in mass and energy budgets."""
import numpy as np
from aospy.utils import coord_to_new_dataarray
from .. import TIME_STR
def first_to_last_vals_dur(arr, freq='1M'):
"""Time elapsed between 1st and last values in each given time period."""
time = coord_to_new_dataarray(arr, TIME_STR)
first = time.resample(freq, TIME_STR, how='first')
last = time.resample(freq, TIME_STR, how='last')
delta_time = last - first
# Divide by a 1 sec timedelta to convert to seconds.
delta_time.values = delta_time.values / np.timedelta64(1, 's')
return delta_time
def time_tendency(arr, freq='1M'):
"""Monthly time tendency of the given field."""
first = arr.resample(freq, TIME_STR, how='first').dropna(TIME_STR)
last = arr.resample(freq, TIME_STR, how='last').dropna(TIME_STR)
return (last - first) / first_to_last_vals_dur(arr, freq)
def time_tendency_each_timestep(arr):
"""Time tendency of the given field between each timestep.
The last or first timestep will not have a tendency, due to the finite
differencing. So the output array will have length one less.
"""
return (arr.diff(dim=TIME_STR, label='upper') /
arr[TIME_STR].diff(dim=TIME_STR, label='upper'))
| Add experimental every-timestep tendency function | Add experimental every-timestep tendency function
| Python | apache-2.0 | spencerahill/aospy-obj-lib | ---
+++
@@ -21,3 +21,13 @@
first = arr.resample(freq, TIME_STR, how='first').dropna(TIME_STR)
last = arr.resample(freq, TIME_STR, how='last').dropna(TIME_STR)
return (last - first) / first_to_last_vals_dur(arr, freq)
+
+
+def time_tendency_each_timestep(arr):
+ """Time tendency of the given field between each timestep.
+
+ The last or first timestep will not have a tendency, due to the finite
+ differencing. So the output array will have length one less.
+ """
+ return (arr.diff(dim=TIME_STR, label='upper') /
+ arr[TIME_STR].diff(dim=TIME_STR, label='upper')) |
7d76b44c371d74cb8c7b272fd9bf8021db6c6702 | qa/rpc-tests/test_framework/cashlib/__init__.py | qa/rpc-tests/test_framework/cashlib/__init__.py | # Copyright (c) 2018 The Bitcoin Unlimited developers
# Distributed under the MIT software license, see the accompanying
# file COPYING or http://www.opensource.org/licenses/mit-license.php.
from test_framework.cashlib.cashlib import init, bin2hex, signTxInput, randombytes, pubkey, spendscript, addrbin, txid, SIGHASH_ALL, SIGHASH_NONE, SIGHASH_SINGLE, SIGHASH_FORKID, SIGHASH_ANYONECANPAY
| # Copyright (c) 2018 The Bitcoin Unlimited developers
# Distributed under the MIT software license, see the accompanying
# file COPYING or http://www.opensource.org/licenses/mit-license.php.
from .cashlib import init, bin2hex, signTxInput, randombytes, pubkey, spendscript, addrbin, txid, SIGHASH_ALL, SIGHASH_NONE, SIGHASH_SINGLE, SIGHASH_FORKID, SIGHASH_ANYONECANPAY
| Use relative import for cashlib | Use relative import for cashlib
| Python | mit | BitcoinUnlimited/BitcoinUnlimited,Justaphf/BitcoinUnlimited,BitcoinUnlimited/BitcoinUnlimited,BitcoinUnlimited/BitcoinUnlimited,BitcoinUnlimited/BitcoinUnlimited,Justaphf/BitcoinUnlimited,Justaphf/BitcoinUnlimited,Justaphf/BitcoinUnlimited,BitcoinUnlimited/BitcoinUnlimited,BitcoinUnlimited/BitcoinUnlimited,Justaphf/BitcoinUnlimited,Justaphf/BitcoinUnlimited | ---
+++
@@ -2,4 +2,4 @@
# Distributed under the MIT software license, see the accompanying
# file COPYING or http://www.opensource.org/licenses/mit-license.php.
-from test_framework.cashlib.cashlib import init, bin2hex, signTxInput, randombytes, pubkey, spendscript, addrbin, txid, SIGHASH_ALL, SIGHASH_NONE, SIGHASH_SINGLE, SIGHASH_FORKID, SIGHASH_ANYONECANPAY
+from .cashlib import init, bin2hex, signTxInput, randombytes, pubkey, spendscript, addrbin, txid, SIGHASH_ALL, SIGHASH_NONE, SIGHASH_SINGLE, SIGHASH_FORKID, SIGHASH_ANYONECANPAY |
ad5da54795ac1329c683e069d3148120470b6451 | brainhack/covariance/pearson.py | brainhack/covariance/pearson.py | import numpy as np
from scipy.sparse import coo_matrix
from sklearn.base import BaseEstimator
class PearsonCorrelation(BaseEstimator):
"""Pearson correlation estimator
"""
def __init__(self, assume_centered=False):
self.assume_centered = assume_centered
def fit(self, X, y=None, connectivity=None):
""" Compute Pearson correlation coefficient
Parameters
----------
X : array-like, shape = [n_samples, n_features]
Training data, where n_samples is the number of samples and
n_features is the number of features.
y : not used, present for API consistence purpose.
Returns
-------
self : object
Returns self.
"""
if connectivity is None:
self.covariance_ = np.corrcoef(X, rowvar=0)
else:
# We suppose connectivity as coo but most of this code would work
# with dense matrix
rows, cols = connectivity.nonzero()
values = np.zeros(rows.shape)
for i, (r, c) in enumerate(zip(rows, cols)):
corr = np.corrcoef(X[r], X[c])
if not np.isnan(corr):
values[i] = corr
self.covariance_ = coo_matrix((values, (rows, cols)))
return self
| import numpy as np
from scipy.sparse import coo_matrix
from sklearn.base import BaseEstimator
import scipy as sp
class PearsonCorrelation(BaseEstimator):
"""Pearson correlation estimator
"""
def __init__(self, assume_centered=False, spatial=False):
self.assume_centered = assume_centered
self.spatial = spatial
def fit(self, X, y=None, connectivity=None):
""" Compute Pearson correlation coefficient
Parameters
----------
X : array-like, shape = [n_samples, n_features]
Training data, where n_samples is the number of samples and
n_features is the number of features.
y : not used, present for API consistence purpose.
Returns
-------
self : object
Returns self.
"""
if connectivity is None:
self.covariance_ = np.corrcoef(X, rowvar=0)
else:
# We suppose connectivity as coo but most of this code would work
# with dense matrix
rows, cols = connectivity.nonzero()
values = np.zeros(rows.shape)
for i, (r, c) in enumerate(zip(rows, cols)):
if self.spatial:
corr = sp.stats.pearsonr(np.dot(X[:, r], X),
np.dot(X[:, c], X))[0]
else:
corr = sp.stats.pearsonr(X[:, r], X[:, c])[0]
if not np.isnan(corr):
values[i] = corr
self.covariance_ = coo_matrix((values, (rows, cols)))
return self
| Add option for spatial correlation | Add option for spatial correlation
| Python | bsd-3-clause | AlexandreAbraham/brainhack2013 | ---
+++
@@ -1,14 +1,16 @@
import numpy as np
from scipy.sparse import coo_matrix
from sklearn.base import BaseEstimator
+import scipy as sp
class PearsonCorrelation(BaseEstimator):
"""Pearson correlation estimator
"""
- def __init__(self, assume_centered=False):
+ def __init__(self, assume_centered=False, spatial=False):
self.assume_centered = assume_centered
+ self.spatial = spatial
def fit(self, X, y=None, connectivity=None):
""" Compute Pearson correlation coefficient
@@ -34,7 +36,11 @@
rows, cols = connectivity.nonzero()
values = np.zeros(rows.shape)
for i, (r, c) in enumerate(zip(rows, cols)):
- corr = np.corrcoef(X[r], X[c])
+ if self.spatial:
+ corr = sp.stats.pearsonr(np.dot(X[:, r], X),
+ np.dot(X[:, c], X))[0]
+ else:
+ corr = sp.stats.pearsonr(X[:, r], X[:, c])[0]
if not np.isnan(corr):
values[i] = corr
self.covariance_ = coo_matrix((values, (rows, cols))) |
6d44672ab0689f6d000001749e0e81b4a9b375f7 | reporting_scripts/user_info.py | reporting_scripts/user_info.py | '''
This module will retrieve info about students registered in the course
Usage:
python user_info.py
'''
from collections import defaultdict
from base_edx import EdXConnection
from generate_csv_report import CSV
connection = EdXConnection('certificates_generatedcertificate', 'auth_userprofile')
collection = connection.get_access_to_collection()
documents = collection['auth_userprofile'].find()
result = []
for document in documents:
user_id = document['user_id']
try:
final_grade = collection['certificates_generatedcertificate'].find_one({'user_id' : user_id})['grade']
result.append([user_id, document['name'], final_grade, document['gender'], document['year_of_birth'], document['level_of_education'], document['country'], document['city']])
except:
# Handle users with no grades
pass
output = CSV(result, ['User ID','Username', 'Final Grade', 'Gender', 'Year of Birth', 'Level of Education', 'Country', 'City'], output_file='atoc185x_user_info.csv')
output.generate_csv()
| '''
This module will retrieve info about students registered in the course
Usage:
python user_info.py
'''
from base_edx import EdXConnection
from generate_csv_report import CSV
connection = EdXConnection('certificates_generatedcertificate', 'auth_userprofile')
collection = connection.get_access_to_collection()
documents = collection['auth_userprofile'].find()
result = []
for document in documents:
user_id = document['user_id']
try:
final_grade = collection['certificates_generatedcertificate'].find_one({'user_id' : user_id})['grade']
result.append([user_id, document['name'], final_grade, document['gender'], document['year_of_birth'], document['level_of_education'], document['country'], document['city']])
except:
# Handle users with no grades
pass
output = CSV(result, ['User ID','Username', 'Final Grade', 'Gender', 'Year of Birth', 'Level of Education', 'Country', 'City'], output_file='atoc185x_user_info.csv')
output.generate_csv()
| Update to handle users with no final grade | Update to handle users with no final grade
| Python | mit | McGillX/edx_data_research,andyzsf/edx_data_research,McGillX/edx_data_research,andyzsf/edx_data_research,McGillX/edx_data_research | ---
+++
@@ -7,18 +7,15 @@
'''
-from collections import defaultdict
from base_edx import EdXConnection
from generate_csv_report import CSV
connection = EdXConnection('certificates_generatedcertificate', 'auth_userprofile')
collection = connection.get_access_to_collection()
-
documents = collection['auth_userprofile'].find()
result = []
-
for document in documents:
user_id = document['user_id']
try: |
40a59efec51661d4325e97f2e307963811336b94 | calaccess_processed/__init__.py | calaccess_processed/__init__.py | from __future__ import absolute_import
default_app_config = 'calaccess_processed.apps.CalAccessProcessedConfig'
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
import os
default_app_config = 'calaccess_processed.apps.CalAccessProcessedConfig'
def get_model_list():
"""
Returns a model list of
"""
from django.apps import apps
model_list = apps.get_app_config("calaccess_processed").models.values()
return [
m for m in model_list
if m.__module__.split('.')[-1] != 'tracking'
]
def archive_directory_path(instance, filename):
"""
Returns a path to an archived processed data file or zip
"""
from calaccess_processed.models.tracking import (
ProcessedDataVersion,
ProcessedDataFile,
)
if isinstance(instance, ProcessedDataVersion):
release_datetime = instance.raw_version.release_datetime
f_name, f_ext = filename.split('.')
path = '{fn}_{dt:%Y-%m-%d_%H-%M-%S}.{fx}'.format(
fn=f_name,
dt=release_datetime,
fx=f_ext,
)
elif isinstance(instance, ProcessedDataFile):
release_datetime = instance.version.raw_version.release_datetime
path = '{dt:%Y-%m-%d_%H-%M-%S}/{f}'.format(dt=release_datetime, f=filename)
else:
raise TypeError(
"Must be ProcessedDataVersion or ProcessedDataFile instance."
)
return path
| Add get_model_list and archive_directory_path functions | Add get_model_list and archive_directory_path functions
| Python | mit | california-civic-data-coalition/django-calaccess-processed-data,california-civic-data-coalition/django-calaccess-processed-data | ---
+++
@@ -1,2 +1,42 @@
-from __future__ import absolute_import
+#!/usr/bin/env python
+# -*- coding: utf-8 -*-
+import os
default_app_config = 'calaccess_processed.apps.CalAccessProcessedConfig'
+
+
+def get_model_list():
+ """
+ Returns a model list of
+ """
+ from django.apps import apps
+ model_list = apps.get_app_config("calaccess_processed").models.values()
+ return [
+ m for m in model_list
+ if m.__module__.split('.')[-1] != 'tracking'
+ ]
+
+def archive_directory_path(instance, filename):
+ """
+ Returns a path to an archived processed data file or zip
+ """
+ from calaccess_processed.models.tracking import (
+ ProcessedDataVersion,
+ ProcessedDataFile,
+ )
+
+ if isinstance(instance, ProcessedDataVersion):
+ release_datetime = instance.raw_version.release_datetime
+ f_name, f_ext = filename.split('.')
+ path = '{fn}_{dt:%Y-%m-%d_%H-%M-%S}.{fx}'.format(
+ fn=f_name,
+ dt=release_datetime,
+ fx=f_ext,
+ )
+ elif isinstance(instance, ProcessedDataFile):
+ release_datetime = instance.version.raw_version.release_datetime
+ path = '{dt:%Y-%m-%d_%H-%M-%S}/{f}'.format(dt=release_datetime, f=filename)
+ else:
+ raise TypeError(
+ "Must be ProcessedDataVersion or ProcessedDataFile instance."
+ )
+ return path |
df2bce1dc4542615a61b5346d2f7e89029e80de9 | caniusepython3/test/__init__.py | caniusepython3/test/__init__.py | try:
import unittest2 as unittest
except ImportError:
import unittest
try:
from unittest import mock
except ImportError:
import mock
try:
import xmlrpc.client as xmlrpc_client
except ImportError:
import xmlrpclib as xmlrpc_client
import functools
def skip_pypi_timeouts(method):
@functools.wraps(method)
def closure(*args, **kwargs):
try:
method(*args, **kwargs)
except xmlrpc_client.ProtocolError as exc:
if exc.errcode >= 500:
raise unittest.SkipTest('PyPI had an error (probably timed out)')
else:
raise
except xmlrpc_client.Fault as exc:
if exc.faultCode == 1:
raise unittest.SkipTest('PyPI had an error (probably timed out)')
else:
raise
return closure
| import requests
try:
import unittest2 as unittest
except ImportError:
import unittest
try:
from unittest import mock
except ImportError:
import mock
import functools
def skip_pypi_timeouts(method):
@functools.wraps(method)
def closure(*args, **kwargs):
try:
method(*args, **kwargs)
except requests.ConnectionError as exc:
raise unittest.SkipTest('PyPI had an error:' + str(exc))
return closure
| Update test skipping based on requests | Update test skipping based on requests
| Python | apache-2.0 | brettcannon/caniusepython3 | ---
+++
@@ -1,3 +1,5 @@
+import requests
+
try:
import unittest2 as unittest
except ImportError:
@@ -8,11 +10,6 @@
except ImportError:
import mock
-try:
- import xmlrpc.client as xmlrpc_client
-except ImportError:
- import xmlrpclib as xmlrpc_client
-
import functools
def skip_pypi_timeouts(method):
@@ -20,14 +17,6 @@
def closure(*args, **kwargs):
try:
method(*args, **kwargs)
- except xmlrpc_client.ProtocolError as exc:
- if exc.errcode >= 500:
- raise unittest.SkipTest('PyPI had an error (probably timed out)')
- else:
- raise
- except xmlrpc_client.Fault as exc:
- if exc.faultCode == 1:
- raise unittest.SkipTest('PyPI had an error (probably timed out)')
- else:
- raise
+ except requests.ConnectionError as exc:
+ raise unittest.SkipTest('PyPI had an error:' + str(exc))
return closure |
18d2c4be27b58a142145f0726e6be21c358064cd | src/rnaseq_lib/docker/__init__.py | src/rnaseq_lib/docker/__init__.py | import os
from subprocess import call
def base_docker_call(mount):
return ['docker', 'run', '--rm', '-v', '{}:/data'.format(mount)]
def fix_directory_ownership(output_dir, tool):
"""
Uses a Docker container to change ownership recursively of a directory
:param str output_dir: Directory to change ownership of
:param str tool: Docker tool to use
"""
stat = os.stat(output_dir)
call(['docker', 'run', '--rm', '--entrypoint=chown', '-v', '{}:/data'.format(output_dir),
tool, '-R', '{}:{}'.format(stat.st_uid, stat.st_gid), '/data'])
| import os
from subprocess import call
def base_docker_call(mount):
"""
Returns the boilerplate array used for Docker calls
:param str mount: Directory to mount
:return: Docker run parameters
:rtype: list(str)
"""
return ['docker', 'run', '--rm', '-v', '{}:/data'.format(os.path.abspath(mount))]
def fix_directory_ownership(output_dir, tool):
"""
Uses a Docker container to change ownership recursively of a directory
:param str output_dir: Directory to change ownership of
:param str tool: Docker tool to use
"""
stat = os.stat(output_dir)
call(['docker', 'run', '--rm', '--entrypoint=chown', '-v', '{}:/data'.format(output_dir),
tool, '-R', '{}:{}'.format(stat.st_uid, stat.st_gid), '/data'])
| Use abspath for docker mount | Use abspath for docker mount
| Python | mit | jvivian/rnaseq-lib,jvivian/rnaseq-lib | ---
+++
@@ -3,7 +3,14 @@
def base_docker_call(mount):
- return ['docker', 'run', '--rm', '-v', '{}:/data'.format(mount)]
+ """
+ Returns the boilerplate array used for Docker calls
+
+ :param str mount: Directory to mount
+ :return: Docker run parameters
+ :rtype: list(str)
+ """
+ return ['docker', 'run', '--rm', '-v', '{}:/data'.format(os.path.abspath(mount))]
def fix_directory_ownership(output_dir, tool): |
76f77023095b068402553b878a65f0566d8add81 | runapp.py | runapp.py | import os
from paste.deploy import loadapp
from waitress import serve
if __name__ == "__main__":
port = int(os.environ.get("PORT", 5000))
app = loadapp('config:production.ini', relative_to='.')
serve(app, host='0.0.0.0', port=port)
| import os
from paste.deploy import loadapp
from pyramid.paster import setup_logging
from waitress import serve
if __name__ == "__main__":
port = int(os.environ.get("PORT", 5000))
setup_logging('production.ini')
app = loadapp('config:production.ini', relative_to='.')
serve(app, host='0.0.0.0', port=port)
| Enable logs when run in Heroku | Enable logs when run in Heroku
| Python | agpl-3.0 | Yaco-Sistemas/yith-library-server,lorenzogil/yith-library-server,Yaco-Sistemas/yith-library-server,lorenzogil/yith-library-server,Yaco-Sistemas/yith-library-server,lorenzogil/yith-library-server | ---
+++
@@ -1,10 +1,12 @@
import os
from paste.deploy import loadapp
+from pyramid.paster import setup_logging
from waitress import serve
if __name__ == "__main__":
port = int(os.environ.get("PORT", 5000))
+ setup_logging('production.ini')
app = loadapp('config:production.ini', relative_to='.')
serve(app, host='0.0.0.0', port=port) |
2c965c0a75be129f429e40ade34ef608f8ceea27 | micropress/urls.py | micropress/urls.py | from django.conf.urls.defaults import *
urlpatterns = patterns('micropress.views',
(r'^$', 'article_list'),
(r'^issue/(?P<issue>\d+)/$', 'issue_list'),
(r'^article/(?P<slug>[-\w]+)/$', 'article_detail'),
#(r'^new/$', 'article_create'),
)
| from django.conf.urls.defaults import *
urlpatterns = patterns('micropress.views',
(r'^$', 'article_list'),
url(r'^issue/(?P<issue>\d+)/$', 'article_list', name='issue_list'),
(r'^article/(?P<slug>[-\w]+)/$', 'article_detail'),
#(r'^new/$', 'article_create'),
)
| Fix url spec for article list by issue. | Fix url spec for article list by issue.
| Python | mit | jbradberry/django-micro-press,jbradberry/django-micro-press | ---
+++
@@ -3,7 +3,7 @@
urlpatterns = patterns('micropress.views',
(r'^$', 'article_list'),
- (r'^issue/(?P<issue>\d+)/$', 'issue_list'),
+ url(r'^issue/(?P<issue>\d+)/$', 'article_list', name='issue_list'),
(r'^article/(?P<slug>[-\w]+)/$', 'article_detail'),
#(r'^new/$', 'article_create'),
) |
794f3d7e229f94b71a7cb3aaefd4640b185e2572 | feder/letters/management/commands/reimport_mailbox.py | feder/letters/management/commands/reimport_mailbox.py | from __future__ import unicode_literals
from django.core.management.base import BaseCommand
from django_mailbox.models import Message
from feder.letters.signals import MessageParser
class Command(BaseCommand):
help = "Reimport mailbox archived emails as letter."
def handle(self, *args, **options):
pass
for message in Message.objects.filter(letter=None).all().iterator():
self.stdout.write(message)
try:
MessageParser(message).insert()
except IOError:
message.delete()
import ipdb;ipdb.set_trace()
| from __future__ import unicode_literals
from django.core.management.base import BaseCommand
from django_mailbox.models import Message
from feder.letters.signals import MessageParser
class Command(BaseCommand):
help = "Reimport mailbox archived emails as letter."
def handle(self, *args, **options):
for message in Message.objects.filter(letter=None).all().iterator():
self.stdout.write(message)
try:
MessageParser(message).insert()
except IOError as e:
print("IO error for message", message, e)
# message.delete()
| Disable delete message on IO error | Disable delete message on IO error
| Python | mit | watchdogpolska/feder,watchdogpolska/feder,watchdogpolska/feder,watchdogpolska/feder | ---
+++
@@ -11,11 +11,10 @@
help = "Reimport mailbox archived emails as letter."
def handle(self, *args, **options):
- pass
for message in Message.objects.filter(letter=None).all().iterator():
self.stdout.write(message)
try:
MessageParser(message).insert()
- except IOError:
- message.delete()
- import ipdb;ipdb.set_trace()
+ except IOError as e:
+ print("IO error for message", message, e)
+ # message.delete() |
673d96c8845a69f567ffe4475d5eb9f110b9995c | cloudenvy/commands/envy_list.py | cloudenvy/commands/envy_list.py | from cloudenvy.envy import Envy
class EnvyList(object):
def __init__(self, argparser):
self._build_subparser(argparser)
def _build_subparser(self, subparsers):
help_str = 'List all ENVys in your current project.'
subparser = subparsers.add_parser('list', help=help_str,
description=help_str)
subparser.set_defaults(func=self.run)
subparser.add_argument('-n', '--name', action='store', default='',
help='Specify custom name for an ENVy.')
return subparser
def run(self, config, args):
envy = Envy(config)
for server in envy.list_servers():
if server.name.startswith(envy.name):
print server.name
| from cloudenvy.envy import Envy
class EnvyList(object):
def __init__(self, argparser):
self._build_subparser(argparser)
def _build_subparser(self, subparsers):
help_str = 'List all ENVys in your current project.'
subparser = subparsers.add_parser('list', help=help_str,
description=help_str)
subparser.set_defaults(func=self.run)
return subparser
def run(self, config, args):
envy = Envy(config)
for server in envy.list_servers():
if server.name.startswith(envy.name):
print server.name
| Remove --name from 'envy list' | Remove --name from 'envy list'
| Python | apache-2.0 | cloudenvy/cloudenvy | ---
+++
@@ -11,9 +11,6 @@
subparser = subparsers.add_parser('list', help=help_str,
description=help_str)
subparser.set_defaults(func=self.run)
- subparser.add_argument('-n', '--name', action='store', default='',
- help='Specify custom name for an ENVy.')
-
return subparser
def run(self, config, args): |
a0b4476d08c59da74eb64cbcc92621cad160fbce | scipy_distutils/setup.py | scipy_distutils/setup.py | import sys
sys.path.insert(0,'..')
import os
d = os.path.basename(os.path.dirname(os.path.abspath(__file__)))
if d == 'scipy_distutils':
execfile('setup_scipy_distutils.py')
else:
os.system('cd .. && ln -s %s scipy_distutils' % (d))
execfile('setup_scipy_distutils.py')
os.system('cd .. && rm -f scipy_distutils')
| import sys
sys.path.insert(0,'..')
import os
d = os.path.basename(os.path.dirname(os.path.abspath(__file__)))
if d == 'scipy_distutils':
import scipy_distutils
del sys.path[0]
execfile('setup_scipy_distutils.py')
else:
os.system('cd .. && ln -s %s scipy_distutils' % (d))
import scipy_distutils
del sys.path[0]
execfile('setup_scipy_distutils.py')
os.system('cd .. && rm -f scipy_distutils')
| Clean up sys.path after scipy_distutils has been imported. | Clean up sys.path after scipy_distutils has been imported.
| Python | bsd-3-clause | mattip/numpy,andsor/numpy,naritta/numpy,GaZ3ll3/numpy,jschueller/numpy,felipebetancur/numpy,bertrand-l/numpy,argriffing/numpy,pbrod/numpy,dato-code/numpy,AustereCuriosity/numpy,cjermain/numpy,argriffing/numpy,astrofrog/numpy,seberg/numpy,jankoslavic/numpy,rmcgibbo/numpy,Yusa95/numpy,cjermain/numpy,nguyentu1602/numpy,githubmlai/numpy,behzadnouri/numpy,maniteja123/numpy,cjermain/numpy,WarrenWeckesser/numpy,joferkington/numpy,SunghanKim/numpy,sinhrks/numpy,numpy/numpy,skymanaditya1/numpy,rajathkumarmp/numpy,bringingheavendown/numpy,MSeifert04/numpy,madphysicist/numpy,Eric89GXL/numpy,MaPePeR/numpy,pyparallel/numpy,jankoslavic/numpy,maniteja123/numpy,yiakwy/numpy,KaelChen/numpy,rmcgibbo/numpy,pbrod/numpy,ahaldane/numpy,numpy/numpy,rhythmsosad/numpy,nguyentu1602/numpy,MSeifert04/numpy,larsmans/numpy,musically-ut/numpy,ContinuumIO/numpy,NextThought/pypy-numpy,dato-code/numpy,NextThought/pypy-numpy,kirillzhuravlev/numpy,MichaelAquilina/numpy,madphysicist/numpy,embray/numpy,gfyoung/numpy,MaPePeR/numpy,stuarteberg/numpy,WarrenWeckesser/numpy,ahaldane/numpy,jankoslavic/numpy,Srisai85/numpy,madphysicist/numpy,embray/numpy,anntzer/numpy,Dapid/numpy,ogrisel/numpy,stefanv/numpy,madphysicist/numpy,dwf/numpy,KaelChen/numpy,ChanderG/numpy,utke1/numpy,ChanderG/numpy,rajathkumarmp/numpy,kiwifb/numpy,dwillmer/numpy,SunghanKim/numpy,stefanv/numpy,simongibbons/numpy,stuarteberg/numpy,skwbc/numpy,rherault-insa/numpy,tdsmith/numpy,pizzathief/numpy,endolith/numpy,chatcannon/numpy,sonnyhu/numpy,sinhrks/numpy,leifdenby/numpy,cowlicks/numpy,shoyer/numpy,ajdawson/numpy,BabeNovelty/numpy,chatcannon/numpy,behzadnouri/numpy,GrimDerp/numpy,joferkington/numpy,grlee77/numpy,seberg/numpy,dwf/numpy,dch312/numpy,Anwesh43/numpy,AustereCuriosity/numpy,ChristopherHogan/numpy,grlee77/numpy,ogrisel/numpy,Eric89GXL/numpy,Yusa95/numpy,yiakwy/numpy,numpy/numpy-refactor,anntzer/numpy,jakirkham/numpy,b-carter/numpy,skwbc/numpy,BMJHayward/numpy,jorisvandenbossche/numpy,Anwesh43/numpy,githubmlai/numpy,has2k1/numpy,cjermain/numpy,mortada/numpy,groutr/numpy,dwillmer/numpy,tynn/numpy,SiccarPoint/numpy,rherault-insa/numpy,anntzer/numpy,dch312/numpy,ahaldane/numpy,andsor/numpy,gfyoung/numpy,chiffa/numpy,grlee77/numpy,MaPePeR/numpy,BabeNovelty/numpy,matthew-brett/numpy,utke1/numpy,ViralLeadership/numpy,rudimeier/numpy,SiccarPoint/numpy,ViralLeadership/numpy,abalkin/numpy,bringingheavendown/numpy,ewmoore/numpy,mingwpy/numpy,mathdd/numpy,pelson/numpy,astrofrog/numpy,brandon-rhodes/numpy,KaelChen/numpy,nguyentu1602/numpy,sinhrks/numpy,ESSS/numpy,embray/numpy,bertrand-l/numpy,groutr/numpy,stefanv/numpy,mwiebe/numpy,pyparallel/numpy,mindw/numpy,yiakwy/numpy,jonathanunderwood/numpy,pelson/numpy,AustereCuriosity/numpy,musically-ut/numpy,dwf/numpy,bmorris3/numpy,Yusa95/numpy,gmcastil/numpy,immerrr/numpy,tacaswell/numpy,drasmuss/numpy,nbeaver/numpy,ContinuumIO/numpy,hainm/numpy,Anwesh43/numpy,madphysicist/numpy,mathdd/numpy,numpy/numpy-refactor,naritta/numpy,pyparallel/numpy,simongibbons/numpy,stefanv/numpy,skwbc/numpy,mathdd/numpy,BabeNovelty/numpy,mortada/numpy,stuarteberg/numpy,rhythmsosad/numpy,anntzer/numpy,simongibbons/numpy,skymanaditya1/numpy,utke1/numpy,gfyoung/numpy,dimasad/numpy,ssanderson/numpy,bringingheavendown/numpy,ogrisel/numpy,empeeu/numpy,jakirkham/numpy,andsor/numpy,ogrisel/numpy,chatcannon/numpy,pizzathief/numpy,numpy/numpy-refactor,MSeifert04/numpy,hainm/numpy,BMJHayward/numpy,immerrr/numpy,CMartelLML/numpy,musically-ut/numpy,leifdenby/numpy,skymanaditya1/numpy,charris/numpy,Yusa95/numpy,rudimeier/numpy,joferkington/numpy,larsmans/numpy,charris/numpy,ContinuumIO/numpy,jorisvandenbossche/numpy,bmorris3/numpy,endolith/numpy,charris/numpy,mindw/numpy,SunghanKim/numpy,WillieMaddox/numpy,chiffa/numpy,brandon-rhodes/numpy,trankmichael/numpy,WarrenWeckesser/numpy,stefanv/numpy,dato-code/numpy,CMartelLML/numpy,ChristopherHogan/numpy,Dapid/numpy,pelson/numpy,ekalosak/numpy,cowlicks/numpy,ogrisel/numpy,ddasilva/numpy,trankmichael/numpy,rhythmsosad/numpy,ssanderson/numpy,kirillzhuravlev/numpy,dch312/numpy,numpy/numpy-refactor,nbeaver/numpy,gmcastil/numpy,bmorris3/numpy,simongibbons/numpy,rhythmsosad/numpy,abalkin/numpy,felipebetancur/numpy,tacaswell/numpy,immerrr/numpy,MSeifert04/numpy,drasmuss/numpy,mhvk/numpy,Eric89GXL/numpy,numpy/numpy-refactor,Linkid/numpy,Srisai85/numpy,sonnyhu/numpy,leifdenby/numpy,maniteja123/numpy,rgommers/numpy,empeeu/numpy,rgommers/numpy,mingwpy/numpy,mortada/numpy,mwiebe/numpy,ewmoore/numpy,rgommers/numpy,BMJHayward/numpy,SunghanKim/numpy,pdebuyl/numpy,larsmans/numpy,b-carter/numpy,naritta/numpy,ajdawson/numpy,Linkid/numpy,drasmuss/numpy,MichaelAquilina/numpy,jorisvandenbossche/numpy,KaelChen/numpy,dwf/numpy,larsmans/numpy,groutr/numpy,CMartelLML/numpy,ekalosak/numpy,simongibbons/numpy,hainm/numpy,stuarteberg/numpy,GaZ3ll3/numpy,dato-code/numpy,sonnyhu/numpy,ssanderson/numpy,behzadnouri/numpy,endolith/numpy,mwiebe/numpy,WillieMaddox/numpy,pizzathief/numpy,tacaswell/numpy,SiccarPoint/numpy,dch312/numpy,Srisai85/numpy,brandon-rhodes/numpy,numpy/numpy,Dapid/numpy,jankoslavic/numpy,mindw/numpy,dimasad/numpy,jakirkham/numpy,mortada/numpy,WarrenWeckesser/numpy,pbrod/numpy,mindw/numpy,matthew-brett/numpy,seberg/numpy,ddasilva/numpy,solarjoe/numpy,pdebuyl/numpy,Anwesh43/numpy,ESSS/numpy,sonnyhu/numpy,joferkington/numpy,grlee77/numpy,rherault-insa/numpy,rmcgibbo/numpy,dwillmer/numpy,andsor/numpy,ChanderG/numpy,pizzathief/numpy,grlee77/numpy,tynn/numpy,ddasilva/numpy,BMJHayward/numpy,moreati/numpy,rmcgibbo/numpy,njase/numpy,immerrr/numpy,mattip/numpy,NextThought/pypy-numpy,gmcastil/numpy,Linkid/numpy,moreati/numpy,pizzathief/numpy,musically-ut/numpy,GrimDerp/numpy,cowlicks/numpy,pelson/numpy,jschueller/numpy,mathdd/numpy,dwf/numpy,Linkid/numpy,dimasad/numpy,Srisai85/numpy,has2k1/numpy,nbeaver/numpy,matthew-brett/numpy,sigma-random/numpy,kiwifb/numpy,trankmichael/numpy,jorisvandenbossche/numpy,shoyer/numpy,kirillzhuravlev/numpy,ahaldane/numpy,mhvk/numpy,brandon-rhodes/numpy,rajathkumarmp/numpy,jschueller/numpy,has2k1/numpy,felipebetancur/numpy,cowlicks/numpy,pdebuyl/numpy,mhvk/numpy,rudimeier/numpy,solarjoe/numpy,sigma-random/numpy,jorisvandenbossche/numpy,GaZ3ll3/numpy,ESSS/numpy,shoyer/numpy,felipebetancur/numpy,ewmoore/numpy,has2k1/numpy,pelson/numpy,empeeu/numpy,shoyer/numpy,moreati/numpy,hainm/numpy,SiccarPoint/numpy,rajathkumarmp/numpy,dimasad/numpy,shoyer/numpy,pbrod/numpy,sigma-random/numpy,chiffa/numpy,ajdawson/numpy,MaPePeR/numpy,tdsmith/numpy,empeeu/numpy,skymanaditya1/numpy,nguyentu1602/numpy,GrimDerp/numpy,rgommers/numpy,trankmichael/numpy,jakirkham/numpy,Eric89GXL/numpy,jonathanunderwood/numpy,ChristopherHogan/numpy,GaZ3ll3/numpy,argriffing/numpy,tdsmith/numpy,MichaelAquilina/numpy,kiwifb/numpy,jschueller/numpy,MichaelAquilina/numpy,seberg/numpy,githubmlai/numpy,matthew-brett/numpy,bertrand-l/numpy,sinhrks/numpy,b-carter/numpy,tdsmith/numpy,ewmoore/numpy,mhvk/numpy,bmorris3/numpy,numpy/numpy,ViralLeadership/numpy,GrimDerp/numpy,ekalosak/numpy,njase/numpy,tynn/numpy,MSeifert04/numpy,mattip/numpy,pbrod/numpy,yiakwy/numpy,naritta/numpy,ekalosak/numpy,jonathanunderwood/numpy,githubmlai/numpy,WarrenWeckesser/numpy,njase/numpy,jakirkham/numpy,embray/numpy,ChanderG/numpy,astrofrog/numpy,kirillzhuravlev/numpy,endolith/numpy,mingwpy/numpy,BabeNovelty/numpy,abalkin/numpy,CMartelLML/numpy,astrofrog/numpy,pdebuyl/numpy,mingwpy/numpy,ewmoore/numpy,astrofrog/numpy,ajdawson/numpy,mhvk/numpy,WillieMaddox/numpy,rudimeier/numpy,NextThought/pypy-numpy,dwillmer/numpy,ChristopherHogan/numpy,charris/numpy,solarjoe/numpy,ahaldane/numpy,mattip/numpy,sigma-random/numpy,matthew-brett/numpy,embray/numpy | ---
+++
@@ -3,8 +3,12 @@
import os
d = os.path.basename(os.path.dirname(os.path.abspath(__file__)))
if d == 'scipy_distutils':
+ import scipy_distutils
+ del sys.path[0]
execfile('setup_scipy_distutils.py')
else:
os.system('cd .. && ln -s %s scipy_distutils' % (d))
+ import scipy_distutils
+ del sys.path[0]
execfile('setup_scipy_distutils.py')
os.system('cd .. && rm -f scipy_distutils') |
8d63079647d37b3d479dab13efacd1eafeac9629 | .travis/run_all_tests.py | .travis/run_all_tests.py | #!/usr/bin/python
import sys, os, os.path
from subprocess import call
cur_dir = os.path.dirname(os.path.realpath(__file__))
parent_dir = os.path.dirname(cur_dir)
statuses = [
call(["echo", "Running python unit tests via nose..."]),
call(["/usr/bin/env", "nosetests", parent_dir], env=os.environ.copy()),
call([os.path.join(cur_dir, "prepare_tests.py")], env=os.environ.copy()),
call(["/usr/bin/env", "phantomjs", os.path.join(cur_dir, "jasmine.js")], env=os.environ.copy()),
]
final_status = 0
for status in statuses:
if status != 0:
final_status = status
break
sys.exit(final_status)
| #!/usr/bin/python
import sys, os, os.path
from subprocess import call
cur_dir = os.path.dirname(os.path.realpath(__file__))
parent_dir = os.path.dirname(cur_dir)
statuses = [
call(["echo", "Running python unit tests via nose..."]),
call([os.path.join(parent_dir, "manage.py"), "test", "harmony.apps.lab.tests"], env=os.environ.copy()),
call([os.path.join(cur_dir, "prepare_tests.py")], env=os.environ.copy()),
call(["/usr/bin/env", "phantomjs", os.path.join(cur_dir, "jasmine.js")], env=os.environ.copy()),
]
final_status = 0
for status in statuses:
if status != 0:
final_status = status
break
sys.exit(final_status)
| Update test harness configuration for Travis CI. | Update test harness configuration for Travis CI.
| Python | bsd-3-clause | Harvard-ATG/HarmonyLab,Harvard-ATG/HarmonyLab,Harvard-ATG/HarmonyLab,Harvard-ATG/HarmonyLab | ---
+++
@@ -8,7 +8,7 @@
statuses = [
call(["echo", "Running python unit tests via nose..."]),
- call(["/usr/bin/env", "nosetests", parent_dir], env=os.environ.copy()),
+ call([os.path.join(parent_dir, "manage.py"), "test", "harmony.apps.lab.tests"], env=os.environ.copy()),
call([os.path.join(cur_dir, "prepare_tests.py")], env=os.environ.copy()),
call(["/usr/bin/env", "phantomjs", os.path.join(cur_dir, "jasmine.js")], env=os.environ.copy()),
] |
1b73399006c6dbd972814f680eb1d2e582699ad7 | examples/xor-classfier.py | examples/xor-classfier.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
'''Example using the theanets package for learning the XOR relation.'''
import climate
import numpy as np
import theanets
climate.enable_default_logging()
X = np.array([[0.0, 0.0], [0.0, 1.0], [1.0, 0.0], [1.0, 1.0]])
Y = np.array([0, 1, 1, 0, ])
Xi = np.random.randint(0, 2, size=(256, 2))
train = [
(Xi + 0.1 * np.random.randn(*Xi.shape)).astype('f'),
(Xi[:, 0] ^ Xi[:, 1]).astype('f')[:, None],
]
e = theanets.Experiment(theanets.Regressor,
layers=(2, 2, 1),
learning_rate=0.1,
learning_rate_decay=0,
momentum=0.5,
patience=300,
num_updates=5000)
e.run(train, train)
print "Input:"
print X
print "XOR output"
print Y
print "NN XOR predictions"
print e.network(X.astype('f'))
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
'''Example using the theanets package for learning the XOR relation.'''
import climate
import numpy as np
import theanets
climate.enable_default_logging()
X = np.array([[0.0, 0.0], [0.0, 1.0], [1.0, 0.0], [1.0, 1.0]])
Y = np.array([0, 1, 1, 0, ])
Xi = np.random.randint(0, 2, size=(256, 2))
train = [
(Xi + 0.1 * np.random.randn(*Xi.shape)).astype('f'),
(Xi[:, 0] ^ Xi[:, 1]).astype('f')[:, None],
]
e = theanets.Experiment(theanets.Regressor,
layers=(2, 2, 1),
learning_rate=0.1,
momentum=0.5,
patience=300,
num_updates=500)
e.run(train, train)
print "Input:"
print X
print "XOR output"
print Y
print "NN XOR predictions"
print e.network(X.astype('f'))
| Adjust default parameters for xor classifier. | Adjust default parameters for xor classifier.
| Python | mit | lmjohns3/theanets,chrinide/theanets,devdoer/theanets | ---
+++
@@ -21,10 +21,9 @@
e = theanets.Experiment(theanets.Regressor,
layers=(2, 2, 1),
learning_rate=0.1,
- learning_rate_decay=0,
momentum=0.5,
patience=300,
- num_updates=5000)
+ num_updates=500)
e.run(train, train)
print "Input:" |
72cb23aeffb2ec9ec29ad063fb4d586f00b67b57 | openpassword/pkcs_utils.py | openpassword/pkcs_utils.py | from math import fmod
def byte_pad(input_bytes, length=8):
if length > 256:
raise ValueError("Maximum padding length is 256")
# Modulo input bytes length with padding length to see how many bytes to pad with
bytes_to_pad = int(fmod(len(input_bytes), length))
# Pad input bytes with a sequence of bytes containing the number of padded bytes
input_bytes += bytes([bytes_to_pad] * bytes_to_pad)
return input_bytes
def strip_byte_padding(input_bytes, length=8):
if fmod(len(input_bytes), length) != 0:
raise ValueError("Input byte length is not divisible by %s " % length)
# Get the last {length} bytes of the input bytes, reversed
byte_block = bytes(input_bytes[:length:-1])
# If input bytes is padded, the padding is equal to byte value of the number
# of bytes padded. So we can read the padding value from the last byte..
padding_byte = byte_block[0:1]
for i in range(1, ord(padding_byte)):
if byte_block[i:i+1] != padding_byte:
return input_bytes
return input_bytes[0:-ord(padding_byte)]
| from math import fmod
def byte_pad(input_bytes, length=8):
if length > 256:
raise ValueError("Maximum padding length is 256")
# Modulo input bytes length with padding length to see how many bytes to pad with
bytes_to_pad = int(fmod(len(input_bytes), length))
# Pad input bytes with a sequence of bytes containing the number of padded bytes
input_bytes += bytes([bytes_to_pad] * bytes_to_pad)
return input_bytes
def strip_byte_padding(input_bytes, length=8):
if fmod(len(input_bytes), length) != 0:
raise ValueError("Input byte length is not divisible by %s " % length)
# Get the last {length} bytes of the input bytes, reversed
byte_block = bytes(input_bytes[:length:-1])
# If input bytes is padded, the padding is equal to byte value of the number
# of bytes padded. So we can read the padding value from the last byte..
padding_byte = byte_block[0:1]
for i in range(1, ord(padding_byte.decode())):
if byte_block[i:i+1] != padding_byte:
return input_bytes
return input_bytes[0:-ord(padding_byte.decode())]
| Convert byte to string before using it with ord() | Convert byte to string before using it with ord()
| Python | mit | openpassword/blimey,openpassword/blimey | ---
+++
@@ -25,8 +25,8 @@
# of bytes padded. So we can read the padding value from the last byte..
padding_byte = byte_block[0:1]
- for i in range(1, ord(padding_byte)):
+ for i in range(1, ord(padding_byte.decode())):
if byte_block[i:i+1] != padding_byte:
return input_bytes
- return input_bytes[0:-ord(padding_byte)]
+ return input_bytes[0:-ord(padding_byte.decode())] |
2b7f7705f709505ef89f759614cf76bc691c8b20 | zc_common/settings.py | zc_common/settings.py | from django.conf import settings
from rest_framework.settings import APISettings
import os
DEFAULTS = {
'GATEWAY_ROOT_PATH': getattr(
settings, 'GATEWAY_ROOT_PATH', os.environ.get('GATEWAY_ROOT_PATH', 'http://gateway:3000/'))
}
zc_settings = APISettings(None, DEFAULTS, None)
| from django.conf import settings
from rest_framework.settings import APISettings
import os
DEFAULTS = {
'GATEWAY_ROOT_PATH': getattr(
settings, 'GATEWAY_ROOT_PATH', os.environ.get('GATEWAY_ROOT_PATH', 'http://gateway:4000/'))
}
zc_settings = APISettings(None, DEFAULTS, None)
| Update default gateway port for local development | Update default gateway port for local development
| Python | mit | ZeroCater/zc_common,ZeroCater/zc_common | ---
+++
@@ -4,7 +4,7 @@
DEFAULTS = {
'GATEWAY_ROOT_PATH': getattr(
- settings, 'GATEWAY_ROOT_PATH', os.environ.get('GATEWAY_ROOT_PATH', 'http://gateway:3000/'))
+ settings, 'GATEWAY_ROOT_PATH', os.environ.get('GATEWAY_ROOT_PATH', 'http://gateway:4000/'))
}
zc_settings = APISettings(None, DEFAULTS, None) |
d39361c3545e60467ff305e00cc09e1790f78c8b | src/lib/sd2/gen_hosts.py | src/lib/sd2/gen_hosts.py | #!/usr/bin/env python
#############################################################################
# Copyright (c) 2017 SiteWare Corp. All right reserved
#############################################################################
import os
from . import get_hosts
from .file_rewriter import FileRewriter
g_etc_hosts = os.getenv('SD2_ETC_HOSTS', '/etc/hosts')
def get_our_config():
rr = ''
for host in get_hosts(enabled=False):
if not host.get('containers'):
continue
rr += '{}\t{}\n'.format(host['local-ip'], host['name'] + '-local')
for cont in host['containers']:
rr += '{}\t{}\n'.format(cont['ip'], cont['name'])
return rr
def gen_etc_hosts():
fr = FileRewriter('/etc/hosts')
before, after = fr.read_config()
rr = get_our_config()
fr.write_config(
before,
rr.split('\n'),
after,
sudo=True
) | #!/usr/bin/env python
#############################################################################
# Copyright (c) 2017 SiteWare Corp. All right reserved
#############################################################################
import os
from . import get_hosts
from .file_rewriter import FileRewriter
g_etc_hosts = os.getenv('SD2_ETC_HOSTS', '/etc/hosts')
def get_our_config():
rr = ''
for host in get_hosts(enabled=False):
if not host.get('containers'):
continue
rr += '{}\t{}\n'.format(host['local-ip'], host['name'] + '-local')
for cont in host['containers']:
rr += '{}\t{}\n'.format(cont['ip'], cont['name'])
return rr
def gen_etc_hosts():
fr = FileRewriter(g_etc_hosts)
before, after = fr.read_config()
rr = get_our_config()
fr.write_config(
before,
rr.split('\n'),
after,
sudo=True
) | Fix it to honor the environment variable when set | Fix it to honor the environment variable when set
| Python | apache-2.0 | gae123/sd2,gae123/sd2 | ---
+++
@@ -19,7 +19,7 @@
return rr
def gen_etc_hosts():
- fr = FileRewriter('/etc/hosts')
+ fr = FileRewriter(g_etc_hosts)
before, after = fr.read_config()
rr = get_our_config()
fr.write_config( |
0f6b0e6288a4bf24902fe52e0d157a5976a0aa59 | scripts/web-server/alexa-pi.py | scripts/web-server/alexa-pi.py | from flask import Flask
from flask_ask import Ask, statement, convert_errors
import logging
from rfsend import rf_send
GPIO.setmode(GPIO.BCM)
app = Flask(__name__)
ask = Ask(app, '/')
logging.getLogger("flask_ask").setLevel(logging.DEBUG)
@ask.intent('LocationControlIntent', mapping={'status': 'status', 'location': 'location'})
def location_control(status, location):
render_message(location, status)
return statement('Turning {} {}!'.format(location, status))
def render_message(location, status):
loc = location.lower()
status = status.lower()
if loc == 'lights':
rf_send('4', status)
rf_send('5', status)
rf_send('2', status)
elif loc == 'christmas':
rf_send('2', status)
elif location == 'kettle':
rf_send('1', status)
if __name__ == '__main__':
port = 5500 #the custom port you want
app.run(host='0.0.0.0', port=port)
| from flask import Flask
from flask_ask import Ask, statement, convert_errors
import logging
from rfsend import rf_send
app = Flask(__name__)
ask = Ask(app, '/')
logging.getLogger("flask_ask").setLevel(logging.DEBUG)
@ask.intent('LocationControlIntent', mapping={'status': 'status', 'location': 'location'})
def location_control(status, location):
render_message(location, status)
return statement('Turning {} {}!'.format(location, status))
def render_message(location, status):
loc = location.lower()
status = status.lower()
if loc == 'lights':
rf_send('4', status)
rf_send('5', status)
rf_send('2', status)
elif loc == 'christmas':
rf_send('2', status)
elif location == 'kettle':
rf_send('1', status)
if __name__ == '__main__':
port = 5500 #the custom port you want
app.run(host='0.0.0.0', port=port)
| Update alexa pi server script. | Update alexa pi server script.
| Python | bsd-3-clause | kbsezginel/raspberry-pi,kbsezginel/raspberry-pi,kbsezginel/raspberry-pi,kbsezginel/raspberry-pi | ---
+++
@@ -2,8 +2,6 @@
from flask_ask import Ask, statement, convert_errors
import logging
from rfsend import rf_send
-
-GPIO.setmode(GPIO.BCM)
app = Flask(__name__)
ask = Ask(app, '/')
@@ -18,7 +16,7 @@
def render_message(location, status):
loc = location.lower()
status = status.lower()
- if loc == 'lights':
+ if loc == 'lights':
rf_send('4', status)
rf_send('5', status)
rf_send('2', status) |
7cfe992263730e7b6e75ae8a84281f4e9d2ac644 | test/benchmark/params_examples.py | test/benchmark/params_examples.py | # -*- coding: utf-8 -*-
# Licensed under a 3-clause BSD style license - see LICENSE.rst
from __future__ import (absolute_import, division, print_function,
unicode_literals)
def track_param(n):
return 42
track_param.params = [10, 20]
def mem_param(n, m):
return [[0]*m]*n
mem_param.params = ([10, 20], [2, 3])
mem_param.param_names = ['number', 'depth']
class ParamSuite:
params = ['a', 'b', 'c']
def setup(self):
self.values = {'a': 1, 'b': 2, 'c': 3}
self.count = 0
def setup_param(self, p):
self.value = self.values[p]
def track_value(self, p):
return self.value + self.count
def teardown_param(self, p):
self.count += 1
del self.value
def teardown(self):
del self.values
| # -*- coding: utf-8 -*-
# Licensed under a 3-clause BSD style license - see LICENSE.rst
from __future__ import (absolute_import, division, print_function,
unicode_literals)
def track_param(n):
return 42
track_param.params = [10, 20]
def mem_param(n, m):
return [[0]*m]*n
mem_param.params = ([10, 20], [2, 3])
mem_param.param_names = ['number', 'depth']
class ParamSuite:
params = ['a', 'b', 'c']
def setup(self):
self.values = {'a': 1, 'b': 2, 'c': 3}
self.count = 0
def setup_params(self, p):
self.value = self.values[p]
def track_value(self, p):
return self.value + self.count
def teardown_params(self, p):
self.count += 1
del self.value
def teardown(self):
del self.values
| Fix test vs parameterized setup/teardown function names | Fix test vs parameterized setup/teardown function names
| Python | bsd-3-clause | qwhelan/asv,mdboom/asv,qwhelan/asv,qwhelan/asv,waylonflinn/asv,ericdill/asv,waylonflinn/asv,spacetelescope/asv,edisongustavo/asv,ericdill/asv,pv/asv,airspeed-velocity/asv,pv/asv,edisongustavo/asv,waylonflinn/asv,spacetelescope/asv,airspeed-velocity/asv,edisongustavo/asv,airspeed-velocity/asv,mdboom/asv,mdboom/asv,pv/asv,pv/asv,spacetelescope/asv,mdboom/asv,airspeed-velocity/asv,spacetelescope/asv,giltis/asv,ericdill/asv,giltis/asv,qwhelan/asv,ericdill/asv,giltis/asv | ---
+++
@@ -25,13 +25,13 @@
self.values = {'a': 1, 'b': 2, 'c': 3}
self.count = 0
- def setup_param(self, p):
+ def setup_params(self, p):
self.value = self.values[p]
def track_value(self, p):
return self.value + self.count
- def teardown_param(self, p):
+ def teardown_params(self, p):
self.count += 1
del self.value
|
f3fa16aeee901f7ee1438bbc5b12170f82a184bc | project/api/management/commands/song_titles.py | project/api/management/commands/song_titles.py | # Django
from django.core.management.base import BaseCommand
from django.core.mail import EmailMessage
from openpyxl import Workbook
# First-Party
from api.models import Chart
class Command(BaseCommand):
help = "Command to sync database with BHS ."
def handle(self, *args, **options):
self.stdout.write("Sending song title report...")
wb = Workbook()
ws = wb.active
fieldnames = [
'title',
]
ws.append(fieldnames)
charts = Chart.objects.all().distinct(
'title'
).order_by(
'title'
)
for chart in charts:
title = chart.title.strip()
row = [
title,
]
ws.append(row)
wb.save('song_title_report.xlsx')
message = EmailMessage(
subject='Song Title Report',
body='Song Title Report Attached',
from_email='admin@barberscore.com',
to=['chris.buechler@verizon.net', 'dbinetti@gmail.com']
)
message.attach_file('song_title_report.xlsx')
result = message.send()
if result == 1:
self.stdout.write("Sent.")
else:
self.stdout.write("Error. Not sent.")
| # Django
from django.core.management.base import BaseCommand
from django.core.mail import EmailMessage
from openpyxl import Workbook
# First-Party
from api.models import Chart
class Command(BaseCommand):
help = "Command to sync database with BHS ."
def handle(self, *args, **options):
self.stdout.write("Sending song title report...")
wb = Workbook()
ws = wb.active
fieldnames = [
'title',
]
ws.append(fieldnames)
charts = Chart.objects.all().distinct(
'title'
).order_by(
'title'
)
for chart in charts:
title = chart.title.strip()
row = [
title,
]
ws.append(row)
wb.save('song_title_report.xlsx')
message = EmailMessage(
subject='Song Title Report',
body='Song Title Report Attached',
from_email='admin@barberscore.com',
to=['chris.buechler@verizon.net', ]
)
message.attach_file('song_title_report.xlsx')
result = message.send()
if result == 1:
self.stdout.write("Sent.")
else:
self.stdout.write("Error. Not sent.")
| Remove me from song title report send | Remove me from song title report send
| Python | bsd-2-clause | dbinetti/barberscore-django,dbinetti/barberscore,barberscore/barberscore-api,dbinetti/barberscore,barberscore/barberscore-api,barberscore/barberscore-api,dbinetti/barberscore-django,barberscore/barberscore-api | ---
+++
@@ -35,7 +35,7 @@
subject='Song Title Report',
body='Song Title Report Attached',
from_email='admin@barberscore.com',
- to=['chris.buechler@verizon.net', 'dbinetti@gmail.com']
+ to=['chris.buechler@verizon.net', ]
)
message.attach_file('song_title_report.xlsx')
result = message.send() |
4aa8cd1f2230b5c2b180a097815802ed1bcb4905 | Lib/sandbox/pyem/__init__.py | Lib/sandbox/pyem/__init__.py | #! /usr/bin/env python
# Last Change: Sat Jun 09 10:00 PM 2007 J
from info import __doc__
from gauss_mix import GmParamError, GM
from gmm_em import GmmParamError, GMM, EM
#from online_em import OnGMM as _OnGMM
#import examples as _examples
__all__ = filter(lambda s:not s.startswith('_'), dir())
from numpy.testing import NumpyTest
test = NumpyTest().test
| #! /usr/bin/env python
# Last Change: Sun Jul 22 11:00 AM 2007 J
raise ImportError(
"""pyem has been moved to scikits and renamed to em. Please install
scikits.learn instead, and change your import to the following:
from scickits.learn.machine import em.""")
from info import __doc__
from gauss_mix import GmParamError, GM
from gmm_em import GmmParamError, GMM, EM
#from online_em import OnGMM as _OnGMM
#import examples as _examples
__all__ = filter(lambda s:not s.startswith('_'), dir())
from numpy.testing import NumpyTest
test = NumpyTest().test
| Raise an import error when importing pyem as it has been moved to scikits. | Raise an import error when importing pyem as it has been moved to scikits.
| Python | bsd-3-clause | newemailjdm/scipy,mdhaber/scipy,Newman101/scipy,Shaswat27/scipy,ilayn/scipy,haudren/scipy,nmayorov/scipy,pschella/scipy,zxsted/scipy,njwilson23/scipy,jseabold/scipy,anntzer/scipy,Srisai85/scipy,pyramania/scipy,jor-/scipy,jseabold/scipy,lukauskas/scipy,woodscn/scipy,raoulbq/scipy,grlee77/scipy,sargas/scipy,aman-iitj/scipy,woodscn/scipy,hainm/scipy,vigna/scipy,vigna/scipy,aeklant/scipy,ogrisel/scipy,fredrikw/scipy,mhogg/scipy,haudren/scipy,nmayorov/scipy,jsilter/scipy,e-q/scipy,pnedunuri/scipy,anielsen001/scipy,aeklant/scipy,maciejkula/scipy,Eric89GXL/scipy,zxsted/scipy,richardotis/scipy,Gillu13/scipy,mgaitan/scipy,maniteja123/scipy,ChanderG/scipy,minhlongdo/scipy,newemailjdm/scipy,scipy/scipy,aarchiba/scipy,anielsen001/scipy,raoulbq/scipy,ilayn/scipy,zerothi/scipy,befelix/scipy,juliantaylor/scipy,nvoron23/scipy,perimosocordiae/scipy,petebachant/scipy,vanpact/scipy,FRidh/scipy,gdooper/scipy,grlee77/scipy,efiring/scipy,Gillu13/scipy,matthew-brett/scipy,ilayn/scipy,apbard/scipy,nmayorov/scipy,perimosocordiae/scipy,sauliusl/scipy,minhlongdo/scipy,mikebenfield/scipy,apbard/scipy,pizzathief/scipy,jakevdp/scipy,nonhermitian/scipy,piyush0609/scipy,teoliphant/scipy,kleskjr/scipy,felipebetancur/scipy,zaxliu/scipy,arokem/scipy,petebachant/scipy,matthewalbani/scipy,arokem/scipy,fredrikw/scipy,jjhelmus/scipy,lhilt/scipy,trankmichael/scipy,mgaitan/scipy,anielsen001/scipy,njwilson23/scipy,andim/scipy,Srisai85/scipy,woodscn/scipy,sriki18/scipy,fernand/scipy,witcxc/scipy,pnedunuri/scipy,pbrod/scipy,mdhaber/scipy,niknow/scipy,haudren/scipy,larsmans/scipy,jor-/scipy,andim/scipy,gertingold/scipy,witcxc/scipy,trankmichael/scipy,gef756/scipy,mingwpy/scipy,arokem/scipy,jseabold/scipy,jamestwebber/scipy,pbrod/scipy,anielsen001/scipy,befelix/scipy,person142/scipy,maniteja123/scipy,ales-erjavec/scipy,ndchorley/scipy,mingwpy/scipy,zaxliu/scipy,efiring/scipy,surhudm/scipy,arokem/scipy,endolith/scipy,WillieMaddox/scipy,endolith/scipy,vberaudi/scipy,felipebetancur/scipy,argriffing/scipy,andim/scipy,zxsted/scipy,teoliphant/scipy,rmcgibbo/scipy,kleskjr/scipy,aeklant/scipy,vanpact/scipy,petebachant/scipy,e-q/scipy,ndchorley/scipy,teoliphant/scipy,richardotis/scipy,niknow/scipy,WillieMaddox/scipy,mikebenfield/scipy,FRidh/scipy,petebachant/scipy,juliantaylor/scipy,Newman101/scipy,hainm/scipy,fredrikw/scipy,Shaswat27/scipy,vanpact/scipy,Stefan-Endres/scipy,sargas/scipy,vhaasteren/scipy,apbard/scipy,vigna/scipy,mingwpy/scipy,efiring/scipy,bkendzior/scipy,pbrod/scipy,haudren/scipy,nvoron23/scipy,ChanderG/scipy,cpaulik/scipy,mgaitan/scipy,maciejkula/scipy,mdhaber/scipy,endolith/scipy,pizzathief/scipy,giorgiop/scipy,richardotis/scipy,grlee77/scipy,bkendzior/scipy,scipy/scipy,gertingold/scipy,zaxliu/scipy,sriki18/scipy,nonhermitian/scipy,Gillu13/scipy,anielsen001/scipy,ales-erjavec/scipy,lukauskas/scipy,behzadnouri/scipy,maniteja123/scipy,vanpact/scipy,fernand/scipy,minhlongdo/scipy,josephcslater/scipy,jakevdp/scipy,mtrbean/scipy,larsmans/scipy,felipebetancur/scipy,zaxliu/scipy,fredrikw/scipy,surhudm/scipy,mhogg/scipy,gertingold/scipy,kalvdans/scipy,gef756/scipy,fernand/scipy,rmcgibbo/scipy,gdooper/scipy,piyush0609/scipy,mikebenfield/scipy,ilayn/scipy,mdhaber/scipy,ales-erjavec/scipy,scipy/scipy,jonycgn/scipy,efiring/scipy,lhilt/scipy,aarchiba/scipy,cpaulik/scipy,zerothi/scipy,ogrisel/scipy,tylerjereddy/scipy,bkendzior/scipy,andyfaff/scipy,jjhelmus/scipy,gef756/scipy,mhogg/scipy,haudren/scipy,ChanderG/scipy,futurulus/scipy,giorgiop/scipy,felipebetancur/scipy,kalvdans/scipy,Kamp9/scipy,perimosocordiae/scipy,zerothi/scipy,felipebetancur/scipy,surhudm/scipy,dominicelse/scipy,jamestwebber/scipy,richardotis/scipy,surhudm/scipy,jsilter/scipy,jonycgn/scipy,anntzer/scipy,sauliusl/scipy,nmayorov/scipy,tylerjereddy/scipy,futurulus/scipy,larsmans/scipy,lukauskas/scipy,FRidh/scipy,rmcgibbo/scipy,dominicelse/scipy,sonnyhu/scipy,FRidh/scipy,mgaitan/scipy,tylerjereddy/scipy,Eric89GXL/scipy,apbard/scipy,dominicelse/scipy,kalvdans/scipy,mortonjt/scipy,aeklant/scipy,vberaudi/scipy,futurulus/scipy,pschella/scipy,chatcannon/scipy,Kamp9/scipy,Gillu13/scipy,vberaudi/scipy,apbard/scipy,njwilson23/scipy,raoulbq/scipy,jakevdp/scipy,sonnyhu/scipy,futurulus/scipy,WillieMaddox/scipy,person142/scipy,WarrenWeckesser/scipy,zaxliu/scipy,raoulbq/scipy,pbrod/scipy,rgommers/scipy,maniteja123/scipy,scipy/scipy,pbrod/scipy,minhlongdo/scipy,trankmichael/scipy,dominicelse/scipy,grlee77/scipy,aman-iitj/scipy,nvoron23/scipy,zerothi/scipy,mortada/scipy,kalvdans/scipy,felipebetancur/scipy,sargas/scipy,WarrenWeckesser/scipy,gdooper/scipy,gdooper/scipy,ales-erjavec/scipy,ndchorley/scipy,jor-/scipy,sauliusl/scipy,niknow/scipy,mtrbean/scipy,woodscn/scipy,sonnyhu/scipy,ndchorley/scipy,sargas/scipy,rgommers/scipy,newemailjdm/scipy,haudren/scipy,tylerjereddy/scipy,Stefan-Endres/scipy,nvoron23/scipy,aarchiba/scipy,mtrbean/scipy,Kamp9/scipy,mhogg/scipy,gef756/scipy,vberaudi/scipy,pbrod/scipy,maciejkula/scipy,fredrikw/scipy,larsmans/scipy,lukauskas/scipy,gertingold/scipy,gfyoung/scipy,WarrenWeckesser/scipy,petebachant/scipy,jseabold/scipy,zxsted/scipy,futurulus/scipy,dch312/scipy,Dapid/scipy,dch312/scipy,efiring/scipy,ogrisel/scipy,jonycgn/scipy,behzadnouri/scipy,pnedunuri/scipy,argriffing/scipy,gfyoung/scipy,arokem/scipy,person142/scipy,Kamp9/scipy,mortada/scipy,ogrisel/scipy,jsilter/scipy,Stefan-Endres/scipy,jsilter/scipy,matthew-brett/scipy,behzadnouri/scipy,giorgiop/scipy,Shaswat27/scipy,josephcslater/scipy,ChanderG/scipy,matthewalbani/scipy,ortylp/scipy,jamestwebber/scipy,anielsen001/scipy,Srisai85/scipy,ales-erjavec/scipy,piyush0609/scipy,maciejkula/scipy,sonnyhu/scipy,ales-erjavec/scipy,pnedunuri/scipy,rgommers/scipy,jjhelmus/scipy,dch312/scipy,mhogg/scipy,Dapid/scipy,fernand/scipy,rgommers/scipy,trankmichael/scipy,fredrikw/scipy,mortada/scipy,chatcannon/scipy,nonhermitian/scipy,Shaswat27/scipy,petebachant/scipy,andyfaff/scipy,pizzathief/scipy,zaxliu/scipy,witcxc/scipy,larsmans/scipy,aeklant/scipy,richardotis/scipy,mingwpy/scipy,rmcgibbo/scipy,juliantaylor/scipy,sriki18/scipy,mtrbean/scipy,FRidh/scipy,vhaasteren/scipy,larsmans/scipy,njwilson23/scipy,chatcannon/scipy,mhogg/scipy,dominicelse/scipy,pschella/scipy,bkendzior/scipy,ortylp/scipy,WillieMaddox/scipy,futurulus/scipy,gdooper/scipy,FRidh/scipy,person142/scipy,andim/scipy,vberaudi/scipy,anntzer/scipy,cpaulik/scipy,maniteja123/scipy,mingwpy/scipy,mortonjt/scipy,behzadnouri/scipy,gfyoung/scipy,Kamp9/scipy,hainm/scipy,lukauskas/scipy,kleskjr/scipy,newemailjdm/scipy,jonycgn/scipy,mtrbean/scipy,jonycgn/scipy,argriffing/scipy,Kamp9/scipy,aman-iitj/scipy,Newman101/scipy,nmayorov/scipy,piyush0609/scipy,giorgiop/scipy,vhaasteren/scipy,fernand/scipy,rmcgibbo/scipy,mortada/scipy,endolith/scipy,trankmichael/scipy,ilayn/scipy,Eric89GXL/scipy,perimosocordiae/scipy,andyfaff/scipy,hainm/scipy,minhlongdo/scipy,niknow/scipy,piyush0609/scipy,niknow/scipy,jjhelmus/scipy,ndchorley/scipy,minhlongdo/scipy,kalvdans/scipy,Dapid/scipy,andyfaff/scipy,surhudm/scipy,sauliusl/scipy,juliantaylor/scipy,raoulbq/scipy,vhaasteren/scipy,Eric89GXL/scipy,argriffing/scipy,ilayn/scipy,giorgiop/scipy,perimosocordiae/scipy,vigna/scipy,teoliphant/scipy,nvoron23/scipy,andim/scipy,anntzer/scipy,mortonjt/scipy,argriffing/scipy,mikebenfield/scipy,kleskjr/scipy,newemailjdm/scipy,juliantaylor/scipy,mortonjt/scipy,befelix/scipy,mgaitan/scipy,grlee77/scipy,gef756/scipy,sauliusl/scipy,pschella/scipy,jjhelmus/scipy,vigna/scipy,jamestwebber/scipy,nonhermitian/scipy,Newman101/scipy,behzadnouri/scipy,sauliusl/scipy,WarrenWeckesser/scipy,sriki18/scipy,pyramania/scipy,hainm/scipy,gertingold/scipy,jor-/scipy,WarrenWeckesser/scipy,ogrisel/scipy,giorgiop/scipy,dch312/scipy,ortylp/scipy,josephcslater/scipy,e-q/scipy,ChanderG/scipy,Shaswat27/scipy,befelix/scipy,mortonjt/scipy,aarchiba/scipy,jseabold/scipy,zerothi/scipy,trankmichael/scipy,matthewalbani/scipy,andyfaff/scipy,maciejkula/scipy,vanpact/scipy,scipy/scipy,Stefan-Endres/scipy,zxsted/scipy,dch312/scipy,witcxc/scipy,kleskjr/scipy,endolith/scipy,richardotis/scipy,cpaulik/scipy,hainm/scipy,ortylp/scipy,argriffing/scipy,tylerjereddy/scipy,andyfaff/scipy,sonnyhu/scipy,anntzer/scipy,vberaudi/scipy,Dapid/scipy,mdhaber/scipy,jonycgn/scipy,woodscn/scipy,jsilter/scipy,andim/scipy,ndchorley/scipy,endolith/scipy,mtrbean/scipy,Newman101/scipy,chatcannon/scipy,jor-/scipy,matthewalbani/scipy,mingwpy/scipy,teoliphant/scipy,cpaulik/scipy,Dapid/scipy,nonhermitian/scipy,piyush0609/scipy,jamestwebber/scipy,efiring/scipy,kleskjr/scipy,behzadnouri/scipy,pizzathief/scipy,gfyoung/scipy,surhudm/scipy,gfyoung/scipy,e-q/scipy,matthew-brett/scipy,pschella/scipy,aman-iitj/scipy,sargas/scipy,lhilt/scipy,maniteja123/scipy,vhaasteren/scipy,Srisai85/scipy,ortylp/scipy,lhilt/scipy,njwilson23/scipy,perimosocordiae/scipy,Eric89GXL/scipy,bkendzior/scipy,befelix/scipy,aman-iitj/scipy,Srisai85/scipy,zxsted/scipy,raoulbq/scipy,Shaswat27/scipy,lukauskas/scipy,sriki18/scipy,ChanderG/scipy,vanpact/scipy,fernand/scipy,WillieMaddox/scipy,pnedunuri/scipy,rmcgibbo/scipy,WarrenWeckesser/scipy,sonnyhu/scipy,pizzathief/scipy,Stefan-Endres/scipy,chatcannon/scipy,e-q/scipy,mortada/scipy,anntzer/scipy,cpaulik/scipy,rgommers/scipy,mikebenfield/scipy,Dapid/scipy,aarchiba/scipy,josephcslater/scipy,Gillu13/scipy,vhaasteren/scipy,zerothi/scipy,woodscn/scipy,chatcannon/scipy,niknow/scipy,lhilt/scipy,nvoron23/scipy,aman-iitj/scipy,matthewalbani/scipy,pyramania/scipy,gef756/scipy,mdhaber/scipy,Newman101/scipy,scipy/scipy,jakevdp/scipy,matthew-brett/scipy,WillieMaddox/scipy,newemailjdm/scipy,pnedunuri/scipy,njwilson23/scipy,person142/scipy,witcxc/scipy,pyramania/scipy,pyramania/scipy,matthew-brett/scipy,mortonjt/scipy,ortylp/scipy,jseabold/scipy,Gillu13/scipy,Srisai85/scipy,mortada/scipy,Stefan-Endres/scipy,josephcslater/scipy,sriki18/scipy,mgaitan/scipy,jakevdp/scipy,Eric89GXL/scipy | ---
+++
@@ -1,5 +1,11 @@
#! /usr/bin/env python
-# Last Change: Sat Jun 09 10:00 PM 2007 J
+# Last Change: Sun Jul 22 11:00 AM 2007 J
+
+raise ImportError(
+"""pyem has been moved to scikits and renamed to em. Please install
+scikits.learn instead, and change your import to the following:
+
+from scickits.learn.machine import em.""")
from info import __doc__
|
732620e2fa9cb9af11136f11751f1255df9aadf6 | game/itemsets/__init__.py | game/itemsets/__init__.py | # -*- coding: utf-8 -*-
"""
Item Sets
- ItemSet.dbc
"""
from .. import *
from ..globalstrings import *
class ItemSet(Model):
pass
class ItemSetTooltip(Tooltip):
def tooltip(self):
self.append("name", ITEM_SET_NAME % (self.obj.getName(), 0, 0), color=YELLOW)
items = self.obj.getItems()
for item in items:
self.append("item", item.getName())
ret = self.values
self.values = []
return ret
class ItemSetProxy(object):
"""
WDBC proxy for item sets
"""
def __init__(self, cls):
from pywow import wdbc
self.__file = wdbc.get("ItemSet.dbc", build=-1)
def get(self, id):
return self.__file[id]
def getItems(self, row):
from ..items import Item, ItemProxy
Item.initProxy(ItemProxy)
ret = []
for i in range(1, 11):
id = row._raw("item_%i" % (i))
if id:
print id, Item(id)
ret.append(Item(id))
return ret
def getName(self, row):
return row.name_enus
| # -*- coding: utf-8 -*-
"""
Item Sets
- ItemSet.dbc
"""
from .. import *
from ..globalstrings import *
class ItemSet(Model):
pass
class ItemSetTooltip(Tooltip):
def tooltip(self):
items = self.obj.getItems()
maxItems = len(items)
self.append("name", ITEM_SET_NAME % (self.obj.getName(), 0, maxItems), color=YELLOW)
for item in items:
self.append("item", item.getName(), color=GREY)
ret = self.values
self.values = []
return ret
class ItemSetProxy(object):
"""
WDBC proxy for item sets
"""
def __init__(self, cls):
from pywow import wdbc
self.__file = wdbc.get("ItemSet.dbc", build=-1)
def get(self, id):
return self.__file[id]
def getItems(self, row):
from ..items import Item, ItemProxy
Item.initProxy(ItemProxy)
ret = []
for i in range(1, 11):
id = row._raw("item_%i" % (i))
if id:
ret.append(Item(id))
return ret
def getName(self, row):
return row.name_enus
| Add maxItems to ItemSet and remove debug output | game/itemsets: Add maxItems to ItemSet and remove debug output
| Python | cc0-1.0 | jleclanche/pywow,jleclanche/pywow,jleclanche/pywow,jleclanche/pywow,jleclanche/pywow,jleclanche/pywow | ---
+++
@@ -14,11 +14,13 @@
class ItemSetTooltip(Tooltip):
def tooltip(self):
- self.append("name", ITEM_SET_NAME % (self.obj.getName(), 0, 0), color=YELLOW)
+ items = self.obj.getItems()
+ maxItems = len(items)
- items = self.obj.getItems()
+ self.append("name", ITEM_SET_NAME % (self.obj.getName(), 0, maxItems), color=YELLOW)
+
for item in items:
- self.append("item", item.getName())
+ self.append("item", item.getName(), color=GREY)
ret = self.values
self.values = []
@@ -42,7 +44,6 @@
for i in range(1, 11):
id = row._raw("item_%i" % (i))
if id:
- print id, Item(id)
ret.append(Item(id))
return ret |
7bcb8fda2daec22fb627a59306c28e39c8f08a2d | HOME/bin/lib/utils.py | HOME/bin/lib/utils.py | import logging
import re
import subprocess
log = logging.getLogger(__name__)
def run(cmd, check=True, cap=False, input=None, exe='/bin/bash', cwd=None, env=None):
log.debug(f"Executing: {cmd!r}")
shell = isinstance(cmd, str)
result = subprocess.run(
cmd,
check=check,
shell=shell,
capture_output=cap,
executable=exe if shell else None,
input=input.encode() if input else None,
cwd=cwd,
env=env,
)
if cap:
return result.stdout.decode()
else:
return result
def partition(pred, list):
trues, falses = [], []
for item in list:
(trues if pred(item) else falses).append(item)
return trues, falses
def partition_by_regex(regex, list):
r = re.compile(regex or '')
return partition(r.search, list)
| import logging
import re
import subprocess
log = logging.getLogger(__name__)
def run(cmd, check=True, cap=False, input=None, exe='/bin/bash', cwd=None, env=None):
log.debug(f"Executing: {cmd!r}")
shell = isinstance(cmd, str)
result = subprocess.run(
cmd,
check=check,
shell=shell,
stdout=subprocess.PIPE if cap in (True, 'stdout') else None,
stderr=subprocess.PIPE if cap in (True, 'stderr') else None,
executable=exe if shell else None,
input=input.encode() if input else None,
cwd=cwd,
env=env,
)
if cap:
return result.stdout.decode()
else:
return result
def partition(pred, list):
trues, falses = [], []
for item in list:
(trues if pred(item) else falses).append(item)
return trues, falses
def partition_by_regex(regex, list):
r = re.compile(regex or '')
return partition(r.search, list)
| Change 'cap' in 'run' to accept True, 'stdout', or 'stderr' | Change 'cap' in 'run' to accept True, 'stdout', or 'stderr'
In 2e6873e I made cap=True equivalent to the new 'capture_output'.
This was a mistake. Turns out fzf writes its UI to stderr (makes sense),
so that was suppressed. Now you can be specific about what to capture.
| Python | mit | kbd/setup,kbd/setup,kbd/setup,kbd/setup,kbd/setup | ---
+++
@@ -12,7 +12,8 @@
cmd,
check=check,
shell=shell,
- capture_output=cap,
+ stdout=subprocess.PIPE if cap in (True, 'stdout') else None,
+ stderr=subprocess.PIPE if cap in (True, 'stderr') else None,
executable=exe if shell else None,
input=input.encode() if input else None,
cwd=cwd, |
2dc90659a7265740bb5821f0f153f00de1b2205a | pydle/features/__init__.py | pydle/features/__init__.py | from . import rfc1459, account, ctcp, tls, isupport, whox, ircv3
from .rfc1459 import RFC1459Support
from .account import AccountSupport
from .ctcp import CTCPSupport
from .tls import TLSSupport
from .isupport import ISUPPORTSupport
from .whox import WHOXSupport
from .ircv3 import IRCv3Support, IRCv3_1Support, IRCv3_2Support
ALL = [ IRCv3Support, WHOXSupport, ISUPPORTSupport, CTCPSupport, AccountSupport, TLSSupport, RFC1459Support ]
LITE = [ WHOXSupport, ISUPPORTSupport, CTCPSupport, TLSSupport, RFC1459Support ]
| from . import rfc1459, account, ctcp, tls, isupport, whox, ircv3
from .rfc1459 import RFC1459Support
from .account import AccountSupport
from .ctcp import CTCPSupport
from .tls import TLSSupport
from .isupport import ISUPPORTSupport
from .whox import WHOXSupport
from .ircv3 import IRCv3Support, IRCv3_1Support, IRCv3_2Support
from .rpl_whoishost import RplWhoisHostSupport
ALL = [IRCv3Support, WHOXSupport, ISUPPORTSupport, CTCPSupport, AccountSupport, TLSSupport, RFC1459Support,
RplWhoisHostSupport]
LITE = [WHOXSupport, ISUPPORTSupport, CTCPSupport, TLSSupport, RFC1459Support]
| Enable RplWhoisHostSupport implementation (as part of full featured client) | Enable RplWhoisHostSupport implementation (as part of full featured client)
| Python | bsd-3-clause | Shizmob/pydle | ---
+++
@@ -7,6 +7,8 @@
from .isupport import ISUPPORTSupport
from .whox import WHOXSupport
from .ircv3 import IRCv3Support, IRCv3_1Support, IRCv3_2Support
+from .rpl_whoishost import RplWhoisHostSupport
-ALL = [ IRCv3Support, WHOXSupport, ISUPPORTSupport, CTCPSupport, AccountSupport, TLSSupport, RFC1459Support ]
-LITE = [ WHOXSupport, ISUPPORTSupport, CTCPSupport, TLSSupport, RFC1459Support ]
+ALL = [IRCv3Support, WHOXSupport, ISUPPORTSupport, CTCPSupport, AccountSupport, TLSSupport, RFC1459Support,
+ RplWhoisHostSupport]
+LITE = [WHOXSupport, ISUPPORTSupport, CTCPSupport, TLSSupport, RFC1459Support] |
e620068b67f965915cd6713a42380b8136a9feae | build/fbcode_builder_config.py | build/fbcode_builder_config.py | #!/usr/bin/env python
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
'fbcode_builder steps to build & test Bistro'
import specs.fbthrift as fbthrift
import specs.folly as folly
import specs.proxygen as proxygen
from shell_quoting import ShellQuoted
# Since Bistro doesn't presently have an "install" target, there is no
# point in having its spec in the shared spec directory.
def fbcode_builder_spec(builder):
return {
'depends_on': [folly, proxygen, fbthrift],
'steps': [
builder.fb_github_project_workdir('bistro/bistro'),
builder.step('Build bistro', [
# Future: should this share some code with `cmake_install()`?
builder.run(ShellQuoted(
'PATH="$PATH:"{p}/bin '
'TEMPLATES_PATH={p}/include/thrift/templates '
'./cmake/run-cmake.sh Debug -DCMAKE_INSTALL_PREFIX={p}'
).format(p=builder.option('prefix'))),
builder.workdir('cmake/Debug'),
builder.parallel_make(),
]),
builder.step('Run bistro tests', [
builder.run(ShellQuoted('ctest')),
]),
]
}
config = {
'github_project': 'facebook/bistro',
'fbcode_builder_spec': fbcode_builder_spec,
}
| #!/usr/bin/env python
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
'fbcode_builder steps to build & test Bistro'
import specs.fbthrift as fbthrift
import specs.folly as folly
import specs.proxygen as proxygen
from shell_quoting import ShellQuoted
# Since Bistro doesn't presently have an "install" target, there is no
# point in having its spec in the shared spec directory.
def fbcode_builder_spec(builder):
return {
'depends_on': [folly, proxygen, fbthrift],
'steps': [
builder.fb_github_project_workdir('bistro/bistro'),
builder.step('Build bistro', [
# Future: should this share some code with `cmake_install()`?
builder.run(ShellQuoted(
'PATH="$PATH:"{p}/bin '
'TEMPLATES_PATH={p}/include/thrift/templates '
'./cmake/run-cmake.sh Debug -DCMAKE_INSTALL_PREFIX={p}'
).format(p=builder.option('prefix'))),
builder.workdir('cmake/Debug'),
builder.parallel_make(),
]),
builder.step('Run bistro tests', [
builder.run(ShellQuoted('ctest --output-on-failure')),
]),
]
}
config = {
'github_project': 'facebook/bistro',
'fbcode_builder_spec': fbcode_builder_spec,
}
| Make future OSS test failures easier to debug | Make future OSS test failures easier to debug
Summary: Show test output on failure.
Reviewed By: jstrizich
Differential Revision: D6914371
fbshipit-source-id: 668feaefd80c3f0253787b89783cb615fe69bf9b
| Python | mit | facebook/bistro,facebook/bistro,facebook/bistro,facebook/bistro,facebook/bistro,facebook/bistro | ---
+++
@@ -30,7 +30,7 @@
builder.parallel_make(),
]),
builder.step('Run bistro tests', [
- builder.run(ShellQuoted('ctest')),
+ builder.run(ShellQuoted('ctest --output-on-failure')),
]),
]
} |
a7f01689459c167635da206f8d503f2ac86d5503 | hatarake/cli.py | hatarake/cli.py | import logging
import json
import click
import requests
from hatarake.app import CONFIG_PATH, POMODORO_DB
from hatarake.config import Config
from hatarake.models import Pomodoro
from hatarake.report import render_report
@click.group()
def main():
logging.basicConfig(level=logging.DEBUG)
logging.getLogger('gntp').setLevel(logging.INFO)
@main.command()
@click.option('--start', help='start time')
@click.argument('duration', type=int)
@click.argument('title')
def submit(start, duration, title):
config = Config(CONFIG_PATH)
api = config.config.get('server', 'api')
token = config.config.get('server', 'token')
response = requests.post(
api,
headers={'Authorization': 'Token %s' % token},
data={
'created': start,
'duration': duration,
'title': title,
}
)
print response.text
@main.command()
def report():
model = Pomodoro(POMODORO_DB)
config = Config(CONFIG_PATH)
timezone = config.config.get('report', 'timezone', 'UTC')
render_report(model, config, timezone)
| import logging
import click
import requests
from hatarake import USER_AGENT
from hatarake.app import CONFIG_PATH, POMODORO_DB
from hatarake.config import Config
from hatarake.models import Pomodoro
from hatarake.report import render_report
@click.group()
def main():
logging.basicConfig(level=logging.DEBUG)
logging.getLogger('gntp').setLevel(logging.INFO)
@main.command()
@click.option('--start', help='start time')
@click.argument('duration', type=int)
@click.argument('title')
def submit(start, duration, title):
config = Config(CONFIG_PATH)
api = config.config.get('server', 'api')
token = config.config.get('server', 'token')
response = requests.post(
api,
headers={
'Authorization': 'Token %s' % token,
'User-Agent': USER_AGENT,
},
data={
'created': start,
'duration': duration,
'title': title,
}
)
print response.text
@main.command()
def report():
model = Pomodoro(POMODORO_DB)
config = Config(CONFIG_PATH)
timezone = config.config.get('report', 'timezone', 'UTC')
render_report(model, config, timezone)
| Add user agent when submitting | Add user agent when submitting
| Python | mit | kfdm/hatarake | ---
+++
@@ -1,8 +1,9 @@
import logging
-import json
+
import click
import requests
+from hatarake import USER_AGENT
from hatarake.app import CONFIG_PATH, POMODORO_DB
from hatarake.config import Config
from hatarake.models import Pomodoro
@@ -26,7 +27,10 @@
response = requests.post(
api,
- headers={'Authorization': 'Token %s' % token},
+ headers={
+ 'Authorization': 'Token %s' % token,
+ 'User-Agent': USER_AGENT,
+ },
data={
'created': start,
'duration': duration, |
9a18cd0cb6366d45803c19301843ddda3a362cfb | tests/test_publisher.py | tests/test_publisher.py | from lektor.publisher import Command
def test_Command_triggers_no_warnings(recwarn):
# This excercises the issue where publishing via rsync resulted
# in ResourceWarnings about unclosed streams.
# This is essentially how RsyncPublisher runs rsync.
with Command(["echo"]) as client:
for _ in client:
pass
# Delete our reference so that the Command instance gets garbage
# collected here. Otherwise, gc will not happen until after the
# test completes and warnings emitted during gc will not be captured
# by the recwarn fixture.
del client
for warning in recwarn.list:
print(warning)
assert len(recwarn) == 0
| import gc
import warnings
import weakref
import pytest
from lektor.publisher import Command
def test_Command_triggers_no_warnings():
# This excercises the issue where publishing via rsync resulted
# in ResourceWarnings about unclosed streams.
with pytest.warns(None) as record:
# This is essentially how RsyncPublisher runs rsync.
with Command(["echo"]) as client:
for _ in client:
pass
# The ResourceWarnings regarding unclosed files we are checking for
# are issued during finalization. Without this extra effort,
# finalization wouldn't happen until after the test completes.
client_is_alive = weakref.ref(client)
del client
if client_is_alive():
gc.collect()
if client_is_alive():
warnings.warn(
"Unable to trigger garbage collection of Command instance, "
"so unable to check for warnings issued during finalization."
)
for warning in record.list:
print(warning)
assert len(record) == 0
| Reword comment, add check that Command is actually garbage collected | Reword comment, add check that Command is actually garbage collected
| Python | bsd-3-clause | lektor/lektor,lektor/lektor,lektor/lektor,lektor/lektor | ---
+++
@@ -1,20 +1,36 @@
+import gc
+import warnings
+import weakref
+
+import pytest
+
from lektor.publisher import Command
-def test_Command_triggers_no_warnings(recwarn):
+def test_Command_triggers_no_warnings():
# This excercises the issue where publishing via rsync resulted
# in ResourceWarnings about unclosed streams.
- # This is essentially how RsyncPublisher runs rsync.
- with Command(["echo"]) as client:
- for _ in client:
- pass
- # Delete our reference so that the Command instance gets garbage
- # collected here. Otherwise, gc will not happen until after the
- # test completes and warnings emitted during gc will not be captured
- # by the recwarn fixture.
- del client
+ with pytest.warns(None) as record:
+ # This is essentially how RsyncPublisher runs rsync.
+ with Command(["echo"]) as client:
+ for _ in client:
+ pass
- for warning in recwarn.list:
+ # The ResourceWarnings regarding unclosed files we are checking for
+ # are issued during finalization. Without this extra effort,
+ # finalization wouldn't happen until after the test completes.
+ client_is_alive = weakref.ref(client)
+ del client
+ if client_is_alive():
+ gc.collect()
+
+ if client_is_alive():
+ warnings.warn(
+ "Unable to trigger garbage collection of Command instance, "
+ "so unable to check for warnings issued during finalization."
+ )
+
+ for warning in record.list:
print(warning)
- assert len(recwarn) == 0
+ assert len(record) == 0 |
703cdca6725438b55bf544962ce0c554598697be | shoop/admin/templatetags/shoop_admin.py | shoop/admin/templatetags/shoop_admin.py | # -*- coding: utf-8 -*-
# This file is part of Shoop.
#
# Copyright (c) 2012-2015, Shoop Ltd. All rights reserved.
#
# This source code is licensed under the AGPLv3 license found in the
# LICENSE file in the root directory of this source tree.
from bootstrap3.renderers import FormRenderer
from django.utils.safestring import mark_safe
from django_jinja import library
from shoop.admin.template_helpers import shoop_admin as shoop_admin_template_helpers
from shoop.admin.utils.bs3_renderers import AdminFieldRenderer
class Bootstrap3Namespace(object):
def field(self, field, **kwargs):
if not field:
return ""
return mark_safe(AdminFieldRenderer(field, **kwargs).render())
def form(self, form, **kwargs):
return mark_safe(FormRenderer(form, **kwargs).render())
library.global_function(name="shoop_admin", fn=shoop_admin_template_helpers)
library.global_function(name="bs3", fn=Bootstrap3Namespace())
| # -*- coding: utf-8 -*-
# This file is part of Shoop.
#
# Copyright (c) 2012-2015, Shoop Ltd. All rights reserved.
#
# This source code is licensed under the AGPLv3 license found in the
# LICENSE file in the root directory of this source tree.
from bootstrap3.renderers import FormRenderer
from django.utils.safestring import mark_safe
from django_jinja import library
from shoop.admin.template_helpers import shoop_admin as shoop_admin_template_helpers
from shoop.admin.utils.bs3_renderers import AdminFieldRenderer
class Bootstrap3Namespace(object):
def field(self, field, **kwargs):
if not field:
return ""
return mark_safe(AdminFieldRenderer(field, **kwargs).render())
def form(self, form, **kwargs):
return mark_safe(FormRenderer(form, **kwargs).render())
def datetime_field(self, field, **kwargs):
kwargs.setdefault("widget_class", "datetime")
kwargs.setdefault("addon_after", "<span class='fa fa-calendar'></span>")
return self.field(field, **kwargs)
library.global_function(name="shoop_admin", fn=shoop_admin_template_helpers)
library.global_function(name="bs3", fn=Bootstrap3Namespace())
| Add template helper for datetime fields | Admin: Add template helper for datetime fields
Refs SHOOP-1612
| Python | agpl-3.0 | suutari-ai/shoop,shoopio/shoop,suutari/shoop,jorge-marques/shoop,hrayr-artunyan/shuup,taedori81/shoop,shawnadelic/shuup,shawnadelic/shuup,shawnadelic/shuup,suutari-ai/shoop,shoopio/shoop,taedori81/shoop,akx/shoop,suutari/shoop,akx/shoop,shoopio/shoop,hrayr-artunyan/shuup,suutari-ai/shoop,suutari/shoop,hrayr-artunyan/shuup,jorge-marques/shoop,jorge-marques/shoop,taedori81/shoop,akx/shoop | ---
+++
@@ -23,6 +23,11 @@
def form(self, form, **kwargs):
return mark_safe(FormRenderer(form, **kwargs).render())
+ def datetime_field(self, field, **kwargs):
+ kwargs.setdefault("widget_class", "datetime")
+ kwargs.setdefault("addon_after", "<span class='fa fa-calendar'></span>")
+ return self.field(field, **kwargs)
+
library.global_function(name="shoop_admin", fn=shoop_admin_template_helpers)
library.global_function(name="bs3", fn=Bootstrap3Namespace()) |
04cd203c3a137dd7eeece63a38a542fad98c4eb3 | setup.py | setup.py | #!/usr/bin/env python
#coding: utf-8
from setuptools import setup
setup(
name = "foxpath",
author = "Martin Keegan",
author_email = "martin.keegan@okfn.org",
version = "0.9",
license = "GNU Affero General Public License v3.0",
url = "",
download_url = "",
description = "Python library for running FoXPath tests against XML",
py_modules = "",
packages = ["foxpath"],
install_requires = "",
scripts = ""
)
| #!/usr/bin/env python
#coding: utf-8
from setuptools import setup
setup(
name = "foxpath",
author = "Martin Keegan",
author_email = "martin.keegan@okfn.org",
version = "0.91",
license = "GNU Affero General Public License v3.0",
url = "",
download_url = "",
description = "Python library for running FoXPath tests against XML",
py_modules = "",
packages = ["foxpath"],
install_requires = "",
scripts = ""
)
| Add error handling back in | Add error handling back in
| Python | mit | pwyf/foxpath-tools | ---
+++
@@ -7,7 +7,7 @@
name = "foxpath",
author = "Martin Keegan",
author_email = "martin.keegan@okfn.org",
- version = "0.9",
+ version = "0.91",
license = "GNU Affero General Public License v3.0",
url = "",
download_url = "", |
903b33db0df2562df108f827177cb1dc0f39ed24 | setup.py | setup.py | #!/usr/bin/env python
import setuptools
setuptools.setup(
name='systemd-minecraft',
description='A systemd service file for one or more vanilla Minecraft servers',
author='Wurstmineberg',
author_email='mail@wurstmineberg.de',
py_modules=['minecraft'],
install_requires=[
'docopt',
'loops',
'mcrcon',
'more-itertools',
'requests'
],
dependency_links=[
'git+https://github.com/fenhl/python-loops.git#egg=loops'
]
)
| #!/usr/bin/env python
import setuptools
setuptools.setup(
name='systemd-minecraft',
description='A systemd service file for one or more vanilla Minecraft servers',
author='Wurstmineberg',
author_email='mail@wurstmineberg.de',
py_modules=['minecraft'],
install_requires=[
'docopt',
'loops',
'mcrcon',
'more-itertools',
'requests'
],
dependency_links=[
'git+https://github.com/fenhl/python-loops.git#egg=loops',
'git+https://github.com/wurstmineberg/MCRcon.git#egg=mcrcon'
]
)
| Add dependency link for mcrcon | Add dependency link for mcrcon
| Python | mit | wurstmineberg/systemd-minecraft | ---
+++
@@ -16,6 +16,7 @@
'requests'
],
dependency_links=[
- 'git+https://github.com/fenhl/python-loops.git#egg=loops'
+ 'git+https://github.com/fenhl/python-loops.git#egg=loops',
+ 'git+https://github.com/wurstmineberg/MCRcon.git#egg=mcrcon'
]
) |
ca32941b82f1c723465a480355242b37ca19848c | setup.py | setup.py | from datetime import datetime
from distutils.core import setup
import os
import subprocess
if os.path.exists("MANIFEST"):
os.unlink("MANIFEST")
VERSION = ("11", "06", "0", "alpha", "0")
setup(
name='armstrong',
version=".".join(VERSION),
description="Armstrong is an open-source publishing system designed for news organizations that gives your team the technology edge it needs to report in a media-rich environment.",
long_description=open("README.rst").read(),
author='Bay Citizen & Texas Tribune',
author_email='dev@armstrongcms.org',
url='http://github.com/armstrongcms/armstrong/',
packages=["armstrong", ],
namespace_packages=["armstrong", ],
install_requires=[
"armstrong.core.arm_content",
"armstrong.core.arm_sections",
"armstrong.core.arm_wells",
"armstrong.apps.articles",
"armstrong.apps.content",
"armstrong.apps.events",
],
classifiers=[
'Development Status :: 3 - Alpha',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
)
| from datetime import datetime
from distutils.core import setup
import os
import subprocess
if os.path.exists("MANIFEST"):
os.unlink("MANIFEST")
VERSION = ("11", "06", "0", "alpha", "0")
setup(
name='armstrong',
version=".".join(VERSION),
description="Armstrong is an open-source publishing system designed for news organizations that gives your team the technology edge it needs to report in a media-rich environment.",
long_description=open("README.rst").read(),
author='Bay Citizen & Texas Tribune',
author_email='dev@armstrongcms.org',
url='http://github.com/armstrongcms/armstrong/',
packages=["armstrong", ],
namespace_packages=["armstrong", ],
install_requires=[
"armstrong.cli",
"armstrong.core.arm_content",
"armstrong.core.arm_sections",
"armstrong.core.arm_wells",
"armstrong.apps.articles",
"armstrong.apps.content",
"armstrong.apps.events",
],
classifiers=[
'Development Status :: 3 - Alpha',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
)
| Add armstrong.cli to the mix | Add armstrong.cli to the mix
| Python | apache-2.0 | armstrong/armstrong | ---
+++
@@ -19,6 +19,7 @@
packages=["armstrong", ],
namespace_packages=["armstrong", ],
install_requires=[
+ "armstrong.cli",
"armstrong.core.arm_content",
"armstrong.core.arm_sections",
"armstrong.core.arm_wells", |
afddd0f6866cfcb75c0f0109710feec0d25bed94 | setup.py | setup.py | from setuptools import setup, find_packages
setup(name='facebookinsights',
description='A wrapper and command-line interface for the Facebook Insights API.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
#url='http://stdbrouw.github.com/facebook-insights/',
download_url='http://www.github.com/debrouwere/facebook-insights/tarball/master',
version='0.3.3',
license='ISC',
packages=find_packages(),
keywords='data analytics api wrapper facebook insights',
entry_points = {
'console_scripts': [
'insights = insights.commands:main',
],
},
install_requires=[
'click',
'requests',
'rauth',
'facepy',
'python-dateutil',
'pytz',
'addressable',
'flask',
'keyring',
],
# test_suite='facebookinsights.tests',
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Topic :: Scientific/Engineering :: Information Analysis',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
],
) | from setuptools import setup, find_packages
setup(name='facebookinsights',
description='A wrapper and command-line interface for the Facebook Insights API.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
#url='http://stdbrouw.github.com/facebook-insights/',
download_url='http://www.github.com/debrouwere/facebook-insights/tarball/master',
version='0.3.3',
license='ISC',
packages=find_packages(),
include_package_data=True,
keywords='data analytics api wrapper facebook insights',
entry_points = {
'console_scripts': [
'insights = insights.commands:main',
],
},
install_requires=[
'click',
'requests',
'rauth',
'facepy',
'python-dateutil',
'pytz',
'addressable',
'flask',
'keyring',
],
# test_suite='facebookinsights.tests',
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Topic :: Scientific/Engineering :: Information Analysis',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
],
) | Add package data to package. | Add package data to package.
| Python | isc | debrouwere/facebook-insights | ---
+++
@@ -10,6 +10,7 @@
version='0.3.3',
license='ISC',
packages=find_packages(),
+ include_package_data=True,
keywords='data analytics api wrapper facebook insights',
entry_points = {
'console_scripts': [ |
cdd0b19c8827adbdf2f9744acc49d695464c9d41 | setup.py | setup.py | #!/usr/bin/env python
from setuptools import setup
setup(
name='Foodcheck',
version='1.0',
description='OpenShift Python 2.7 Django',
author='Tim Austen, Eileen Lin, Richard Esplin',
author_email='richard-oss@esplins.org',
url='http://www.python.org/sigs/distutils-sig/',
install_requires=['Django<=1.6',
# 'MySQL-python',
# 'pymongo',
# 'psycopg2',
],
)
| #!/usr/bin/env python
from setuptools import setup
setup(
name='Foodcheck',
version='1.0',
description='OpenShift Python 2.7 Django',
author='Tim Austen, Eileen Lin, Richard Esplin',
author_email='richard-oss@esplins.org',
url='http://www.python.org/sigs/distutils-sig/',
install_requires=['Django>=1.5',
# 'MySQL-python',
# 'pymongo',
# 'psycopg2',
],
)
| Check the version correctly this time. | Check the version correctly this time.
| Python | agpl-3.0 | esplinr/foodcheck,esplinr/foodcheck,esplinr/foodcheck,esplinr/foodcheck | ---
+++
@@ -9,7 +9,7 @@
author='Tim Austen, Eileen Lin, Richard Esplin',
author_email='richard-oss@esplins.org',
url='http://www.python.org/sigs/distutils-sig/',
- install_requires=['Django<=1.6',
+ install_requires=['Django>=1.5',
# 'MySQL-python',
# 'pymongo',
# 'psycopg2', |
90364d28522bf368c22f82b404ca79d8211716c7 | setup.py | setup.py | # -*- coding: utf-8 -*-
from setuptools import setup, find_packages
import camplight
setup(name='camplight',
version=camplight.__version__,
author='Mathias Lafeldt',
author_email='mathias.lafeldt@gmail.com',
url='https://github.com/mlafeldt/camplight',
license='MIT',
description='Python implementation of the Campfire API',
long_description=open('README.md').read() + '\n\n' +
open('HISTORY.md').read(),
classifiers=['Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'Natural Language :: English',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python'],
packages=find_packages(),
zip_safe=False,
setup_requires=[],
install_requires=['requests>=0.12.1'],
entry_points="""
# -*- Entry points: -*-
[console_scripts]
camplight=camplight.cli:main
""",
test_suite='test')
| # -*- coding: utf-8 -*-
from setuptools import setup, find_packages
setup(name='camplight',
version='0.3',
author='Mathias Lafeldt',
author_email='mathias.lafeldt@gmail.com',
url='https://github.com/mlafeldt/camplight',
license='MIT',
description='Python implementation of the Campfire API',
long_description=open('README.md').read() + '\n\n' +
open('HISTORY.md').read(),
classifiers=['Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'Natural Language :: English',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python'],
packages=find_packages(),
zip_safe=False,
setup_requires=[],
install_requires=['requests>=0.12.1'],
entry_points="""
# -*- Entry points: -*-
[console_scripts]
camplight=camplight.cli:main
""",
test_suite='test')
| Fix ImportError: No module named requests | Fix ImportError: No module named requests
setup.py imports camplight which ends up importing requests via camplight.api. It
does the import before running setup so you cannot install the package
unless you already have requests installed.
| Python | mit | DataDog/camplight,mlafeldt/camplight | ---
+++
@@ -1,10 +1,9 @@
# -*- coding: utf-8 -*-
from setuptools import setup, find_packages
-import camplight
setup(name='camplight',
- version=camplight.__version__,
+ version='0.3',
author='Mathias Lafeldt',
author_email='mathias.lafeldt@gmail.com',
url='https://github.com/mlafeldt/camplight', |
bf848c8716074314b3b96731228ea28541c54a9f | setup.py | setup.py | from ez_setup import use_setuptools
use_setuptools()
from setuptools import setup, find_packages
setup(name = 'Adafruit_Nokia_LCD',
version = '0.1.0',
author = 'Tony DiCola',
author_email = 'tdicola@adafruit.com',
description = 'Library to display images on the Nokia 5110/3110 LCD.',
license = 'MIT',
url = 'https://github.com/adafruit/Adafruit_Nokia_LCD/',
packages = find_packages())
| from ez_setup import use_setuptools
use_setuptools()
from setuptools import setup, find_packages
setup(name = 'Adafruit_Nokia_LCD',
version = '0.1.0',
author = 'Tony DiCola',
author_email = 'tdicola@adafruit.com',
description = 'Library to display images on the Nokia 5110/3110 LCD.',
license = 'MIT',
url = 'https://github.com/adafruit/Adafruit_Nokia_LCD/',
dependency_links = ['https://github.com/adafruit/Adafruit_Python_GPIO/tarball/master#egg=Adafruit-GPIO-0.1.0'],
install_requires = ['Adafruit-GPIO>=0.1.0'],
packages = find_packages())
| Add dependency on Adafruit-GPIO library. | Add dependency on Adafruit-GPIO library.
| Python | mit | adafruit/Adafruit_Nokia_LCD | ---
+++
@@ -9,4 +9,6 @@
description = 'Library to display images on the Nokia 5110/3110 LCD.',
license = 'MIT',
url = 'https://github.com/adafruit/Adafruit_Nokia_LCD/',
+ dependency_links = ['https://github.com/adafruit/Adafruit_Python_GPIO/tarball/master#egg=Adafruit-GPIO-0.1.0'],
+ install_requires = ['Adafruit-GPIO>=0.1.0'],
packages = find_packages()) |
eaed2b0b37c042f1dbaf5c60163022e2e2605486 | setup.py | setup.py | #!/usr/bin/env python
# encoding: utf-8
"""
setup.py
Setup the Keyring Lib for Python.
"""
import sys
from distutils.core import setup, Extension
from extensions import get_extensions
setup(name = 'keyring',
version = "0.5.1",
description = "Store and access your passwords safely.",
url = "http://home.python-keyring.org/",
keywords = "keyring Keychain GnomeKeyring Kwallet password storage",
maintainer = "Kang Zhang",
maintainer_email = "jobo.zh@gmail.com",
license="PSF",
long_description = open('README.txt').read() + open('CHANGES.txt').read(),
platforms = ["Many"],
packages = ['keyring', 'keyring.tests', 'keyring.util',
'keyring.backends'],
ext_modules = get_extensions()
)
| #!/usr/bin/env python
# encoding: utf-8
"""
setup.py
Setup the Keyring Lib for Python.
"""
import sys
from distutils.core import setup, Extension
from extensions import get_extensions
setup(name = 'keyring',
version = "0.5.1",
description = "Store and access your passwords safely.",
url = "http://home.python-keyring.org/",
keywords = "keyring Keychain GnomeKeyring Kwallet password storage",
maintainer = "Kang Zhang",
maintainer_email = "jobo.zh@gmail.com",
license="PSF",
long_description = open('README').read() + open('CHANGES.txt').read(),
platforms = ["Many"],
packages = ['keyring', 'keyring.tests', 'keyring.util',
'keyring.backends'],
ext_modules = get_extensions()
)
| Use the new README file, which has been renamed in 4ec717c11604 | Use the new README file, which has been renamed in 4ec717c11604
| Python | mit | jaraco/keyring | ---
+++
@@ -19,7 +19,7 @@
maintainer = "Kang Zhang",
maintainer_email = "jobo.zh@gmail.com",
license="PSF",
- long_description = open('README.txt').read() + open('CHANGES.txt').read(),
+ long_description = open('README').read() + open('CHANGES.txt').read(),
platforms = ["Many"],
packages = ['keyring', 'keyring.tests', 'keyring.util',
'keyring.backends'], |
6e004d3ffb2925f5292582e5a4478707863b819e | setup.py | setup.py | from distutils.core import setup, Extension
setup(
name = 'iMX233_GPIO',
version = '0.1.0',
author = 'Stefan Mavrodiev',
author_email = 'support@olimex.com',
url = 'https://www.olimex.com/',
license = 'MIT',
description = 'Control GPIOs on iMX233-OLinuXino.',
long_description = open('README.txt').read() + open('CHANGES.txt').read(),
classifiers = [ 'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Education',
'License :: OSI Approved :: MIT License',
'Operating System :: POSIX :: Linux',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Home Automation',
'Topic :: Software Development :: Embedded Systems'
],
ext_modules = [Extension('iMX233_GPIO', ['source/imx233.c'])],
package_dir={'': 'source'},
packages=[''],
)
| from distutils.core import setup, Extension
setup(
name = 'iMX233_GPIO',
version = '0.1.1',
author = 'Stefan Mavrodiev',
author_email = 'support@olimex.com',
url = 'https://www.olimex.com/',
license = 'MIT',
description = 'Control GPIOs on iMX233-OLinuXino.',
long_description = open('README.txt').read() + open('CHANGES.txt').read(),
classifiers = [ 'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Education',
'License :: OSI Approved :: MIT License',
'Operating System :: POSIX :: Linux',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Home Automation',
'Topic :: Software Development :: Embedded Systems'
],
ext_modules = [Extension('iMX233_GPIO', ['source/imx233.c'])],
package_dir={'': 'source'},
packages=[''],
)
| Update package version to 0.1.1 | Update package version to 0.1.1
| Python | mit | droid4control/iMX233_GPIO,droid4control/iMX233_GPIO | ---
+++
@@ -3,7 +3,7 @@
setup(
name = 'iMX233_GPIO',
- version = '0.1.0',
+ version = '0.1.1',
author = 'Stefan Mavrodiev',
author_email = 'support@olimex.com',
url = 'https://www.olimex.com/', |
72e199f8adc2cb7206c414552c3faa86cc9ed699 | setup.py | setup.py | from ez_setup import use_setuptools
use_setuptools()
from setuptools import setup
setup(
name='urbansim',
version='0.2dev',
description='Tool for modeling metropolitan real estate markets',
author='Synthicity',
author_email='ffoti@berkeley.edu',
license='AGPL',
url='https://github.com/synthicity/urbansim',
classifiers=['Development Status :: 4 - Beta',
'Programming Language :: Python :: 2.7',
('License :: OSI Approved :: '
'GNU Affero General Public License v3')],
packages=['synthicity'],
package_data={'': ['*.py',
'urbanchoice/*.py',
'urbansim/*.py',
'urbansimd/*.py',
'utils/*.py']},
install_requires=['Django>=1.6.2',
'jinja2>=2.7.2',
'numpy>=1.8.0',
'pandas>=0.13.1',
'patsy>=0.2.1',
'pytables>=3.1.0',
'scipy>=0.13.3',
'shapely>=1.3.0',
'simplejson>=3.3.3',
'statsmodels>=0.5.0']
)
| from ez_setup import use_setuptools
use_setuptools()
from setuptools import setup
setup(
name='urbansim',
version='0.2dev',
description='Tool for modeling metropolitan real estate markets',
author='Synthicity',
author_email='ffoti@berkeley.edu',
license='AGPL',
url='https://github.com/synthicity/urbansim',
classifiers=['Development Status :: 4 - Beta',
'Programming Language :: Python :: 2.7',
('License :: OSI Approved :: '
'GNU Affero General Public License v3')],
packages=['synthicity'],
package_data={'': ['*.py',
'urbanchoice/*.py',
'urbansim/*.py',
'urbansimd/*.py',
'utils/*.py']},
install_requires=['Django>=1.6.2',
'jinja2>=2.7.2',
'numpy>=1.8.0',
'pandas>=0.13.1',
'patsy>=0.2.1',
'scipy>=0.13.3',
'shapely>=1.3.0',
'simplejson>=3.3.3',
'statsmodels>=0.5.0',
'tables>=3.1.0']
)
| Fix pytables -> tables for pip. | Fix pytables -> tables for pip.
| Python | bsd-3-clause | apdjustino/urbansim,ual/urbansim,AZMAG/urbansim,SANDAG/urbansim,VladimirTyrin/urbansim,VladimirTyrin/urbansim,SANDAG/urbansim,synthicity/urbansim,apdjustino/urbansim,VladimirTyrin/urbansim,synthicity/urbansim,bricegnichols/urbansim,synthicity/urbansim,AZMAG/urbansim,apdjustino/urbansim,ual/urbansim,waddell/urbansim,UDST/urbansim,AZMAG/urbansim,VladimirTyrin/urbansim,apdjustino/urbansim,waddell/urbansim,waddell/urbansim,SANDAG/urbansim,UDST/urbansim,ual/urbansim,bricegnichols/urbansim,ual/urbansim,UDST/urbansim,AZMAG/urbansim,waddell/urbansim,UDST/urbansim,synthicity/urbansim,SANDAG/urbansim,bricegnichols/urbansim,bricegnichols/urbansim | ---
+++
@@ -26,9 +26,9 @@
'numpy>=1.8.0',
'pandas>=0.13.1',
'patsy>=0.2.1',
- 'pytables>=3.1.0',
'scipy>=0.13.3',
'shapely>=1.3.0',
'simplejson>=3.3.3',
- 'statsmodels>=0.5.0']
+ 'statsmodels>=0.5.0',
+ 'tables>=3.1.0']
) |
ea3df41ae119bf0c1f7097dafde663bd6e98735a | setup.py | setup.py | from distutils.core import setup
setup(
name='udiskie',
version='0.3.3',
description='Removable disk automounter for udisks',
author='Byron Clark',
author_email='byron@theclarkfamily.name',
url='http://bitbucket.org/byronclark/udiskie',
license='MIT',
packages=[
'udiskie',
],
scripts=[
'bin/udiskie',
'bin/udiskie-umount',
],
)
| from distutils.core import setup
setup(
name='udiskie',
version='0.3.4',
description='Removable disk automounter for udisks',
author='Byron Clark',
author_email='byron@theclarkfamily.name',
url='http://bitbucket.org/byronclark/udiskie',
license='MIT',
packages=[
'udiskie',
],
scripts=[
'bin/udiskie',
'bin/udiskie-umount',
],
)
| Bump the version for development. | Bump the version for development.
| Python | mit | coldfix/udiskie,coldfix/udiskie,khardix/udiskie,pstray/udiskie,mathstuf/udiskie,pstray/udiskie | ---
+++
@@ -2,7 +2,7 @@
setup(
name='udiskie',
- version='0.3.3',
+ version='0.3.4',
description='Removable disk automounter for udisks',
author='Byron Clark',
author_email='byron@theclarkfamily.name', |
c49e96114cd675842dc22923f980a2c05c80dcc7 | setup.py | setup.py | # -*- coding: utf-8 -*-
from setuptools import setup
project = "gastosabertos"
setup(
name=project,
version='0.0.1',
url='https://github.com/okfn-brasil/gastos_abertos',
description='Visualization of public spending in Sao Paulo city for Gastos Abertos project',
author='Edgar Zanella Alvarenga',
author_email='e@vaz.io',
packages=["gastosabertos"],
include_package_data=True,
zip_safe=False,
install_requires=[
'Flask>=0.10.1',
'Flask-SQLAlchemy',
'Flask-WTF',
'Flask-Script',
'Flask-Babel',
'Flask-Testing',
'fabric',
'pandas'
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries'
]
)
| # -*- coding: utf-8 -*-
from setuptools import setup
project = "gastosabertos"
setup(
name=project,
version='0.0.1',
url='https://github.com/okfn-brasil/gastos_abertos',
description='Visualization of public spending in Sao Paulo city for Gastos Abertos project',
author='Edgar Zanella Alvarenga',
author_email='e@vaz.io',
packages=["gastosabertos"],
include_package_data=True,
zip_safe=False,
install_requires=[
'Flask>=0.10.1',
'Flask-SQLAlchemy',
'Flask-WTF',
'Flask-Script',
'Flask-Babel',
'Flask-Testing',
'Flask-Restful',
'fabric',
'pandas'
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries'
]
)
| Add Flask-Restfull dependency to python package | Add Flask-Restfull dependency to python package
| Python | agpl-3.0 | okfn-brasil/gastos_abertos,okfn-brasil/gastos_abertos,LuizArmesto/gastos_abertos,nucleo-digital/gastos_abertos,LuizArmesto/gastos_abertos,andresmrm/gastos_abertos,andresmrm/gastos_abertos | ---
+++
@@ -21,6 +21,7 @@
'Flask-Script',
'Flask-Babel',
'Flask-Testing',
+ 'Flask-Restful',
'fabric',
'pandas'
], |
3059f4b36872f2210a4867be804981ca90751194 | setup.py | setup.py | from setuptools import setup
setup(
name='bookmarks',
packages=['bookmarks'],
include_package_data=True,
install_requires=[
'flask',
],
)
| from setuptools import setup
setup(
name='bookmarks',
packages=['bookmarks'],
include_package_data=True,
install_requires=[
'Flask',
'SQLalchemy'
],
)
| Add SQLalchemy to list of install requirements | Add SQLalchemy to list of install requirements
| Python | apache-2.0 | byanofsky/bookmarks,byanofsky/bookmarks,byanofsky/bookmarks | ---
+++
@@ -5,6 +5,7 @@
packages=['bookmarks'],
include_package_data=True,
install_requires=[
- 'flask',
+ 'Flask',
+ 'SQLalchemy'
],
) |
60150b68a254943006611b4300f772e5beb38af0 | setup.py | setup.py | #!/usr/bin/env python
import os
import sys
import io
try:
import setuptools
except ImportError:
from distribute_setup import use_setuptools
use_setuptools()
from setuptools import setup, Extension
from setuptools import find_packages
extra_compile_args = [] if os.name == 'nt' else ["-g", "-O2", "-march=native"]
extra_link_args = [] if os.name == 'nt' else ["-g"]
mod_cv_algorithms = Extension('cv_algorithms._cv_algorithms',
sources=['src/thinning.cpp',
'src/distance.cpp',
'src/grassfire.cpp',
'src/popcount.cpp',
'src/neighbours.cpp'],
extra_compile_args=extra_compile_args,
extra_link_args=extra_link_args)
setup(
name='cv_algorithms',
license='Apache license 2.0',
packages=find_packages(exclude=['tests*']),
install_requires=['cffi>=1.10'],
ext_modules=[mod_cv_algorithms],
test_suite='nose.collector',
tests_require=['nose', 'coverage', 'mock', 'rednose', 'nose-parameterized'],
setup_requires=['nose>=1.0'],
platforms="any",
zip_safe=False,
version='1.0.0',
long_description=io.open("README.rst", encoding="utf-8").read(),
description='Optimized OpenCV extra algorithms for Python',
url="https://github.com/ulikoehler/"
)
| #!/usr/bin/env python
import os
import sys
import io
try:
import setuptools
except ImportError:
from distribute_setup import use_setuptools
use_setuptools()
from setuptools import setup, Extension
from setuptools import find_packages
extra_compile_args = [] if os.name == 'nt' else ["-g", "-O2", "-march=native"]
extra_link_args = [] if os.name == 'nt' else ["-g"]
mod_cv_algorithms = Extension('cv_algorithms._cv_algorithms',
sources=['src/thinning.cpp',
'src/distance.cpp',
'src/grassfire.cpp',
'src/popcount.cpp',
'src/neighbours.cpp'],
extra_compile_args=extra_compile_args,
extra_link_args=extra_link_args)
setup(
name='cv_algorithms',
license='Apache license 2.0',
packages=find_packages(exclude=['tests*']),
install_requires=['cffi>=0.7'],
ext_modules=[mod_cv_algorithms],
test_suite='nose.collector',
tests_require=['nose', 'coverage', 'mock', 'rednose', 'nose-parameterized'],
setup_requires=['nose>=1.0'],
platforms="any",
zip_safe=False,
version='1.0.0',
long_description=io.open("README.rst", encoding="utf-8").read(),
description='Optimized OpenCV extra algorithms for Python',
url="https://github.com/ulikoehler/"
)
| Rollback to supporting older versions of cffi | Rollback to supporting older versions of cffi | Python | apache-2.0 | ulikoehler/cv_algorithms,ulikoehler/cv_algorithms | ---
+++
@@ -26,7 +26,7 @@
name='cv_algorithms',
license='Apache license 2.0',
packages=find_packages(exclude=['tests*']),
- install_requires=['cffi>=1.10'],
+ install_requires=['cffi>=0.7'],
ext_modules=[mod_cv_algorithms],
test_suite='nose.collector',
tests_require=['nose', 'coverage', 'mock', 'rednose', 'nose-parameterized'], |
10e5c50bac81bc072195676f202190de4adc3459 | setup.py | setup.py | import sys
from setuptools import setup
from setuptools.command.test import test as TestCommand
class PyTest(TestCommand):
def finalize_options(self):
TestCommand.finalize_options(self)
self.test_args = []
self.test_suite = True
def run_tests(self):
# import here, cause outside the eggs aren't loaded
import pytest
errno = pytest.main(self.test_args)
sys.exit(errno)
# "import" __version__
for line in open("jack.py"):
if line.startswith("__version__"):
exec(line)
break
setup(
name="JACK-Client",
version=__version__,
py_modules=["jack"],
install_requires=['cffi'],
author="Matthias Geier",
author_email="Matthias.Geier@gmail.com",
description="JACK Audio Connection Kit (JACK) Client for Python",
long_description=open("README.rst").read(),
license="MIT",
keywords="JACK audio low-latency multi-channel".split(),
url="http://jackclient-python.rtfd.org/",
platforms="any",
classifiers=[
"Development Status :: 2 - Pre-Alpha",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Programming Language :: Python :: 2",
"Programming Language :: Python :: 3",
"Topic :: Multimedia :: Sound/Audio",
],
tests_require=['pytest'],
cmdclass={'test': PyTest},
)
| import sys
from setuptools import setup
from setuptools.command.test import test as TestCommand
class PyTest(TestCommand):
def finalize_options(self):
TestCommand.finalize_options(self)
self.test_args = []
self.test_suite = True
def run_tests(self):
# import here, cause outside the eggs aren't loaded
import pytest
errno = pytest.main(self.test_args)
sys.exit(errno)
# "import" __version__
for line in open("jack.py"):
if line.startswith("__version__"):
exec(line)
break
setup(
name="JACK-Client",
version=__version__,
py_modules=["jack"],
install_requires=['cffi'],
author="Matthias Geier",
author_email="Matthias.Geier@gmail.com",
description="JACK Audio Connection Kit (JACK) Client for Python",
long_description=open("README.rst").read(),
license="MIT",
keywords="JACK audio low-latency multi-channel".split(),
url="http://jackclient-python.rtfd.org/",
platforms="any",
classifiers=[
"Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Programming Language :: Python :: 2",
"Programming Language :: Python :: 3",
"Topic :: Multimedia :: Sound/Audio",
],
tests_require=['pytest'],
cmdclass={'test': PyTest},
)
| Change development status from pre-alpha to alpha | Change development status from pre-alpha to alpha
| Python | mit | spatialaudio/jackclient-python,stephenedie/jackclient-python | ---
+++
@@ -35,7 +35,7 @@
url="http://jackclient-python.rtfd.org/",
platforms="any",
classifiers=[
- "Development Status :: 2 - Pre-Alpha",
+ "Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python", |
d4cf6a4737ba6c05c5150e2b892b8b0d01d800e9 | setup.py | setup.py | #!/usr/bin/env python
import setuptools
install_requires = [
'PrettyTable==0.7.2',
'kazoo==1.00',
'simplejson',
'argparse',
'kafka-python'
]
setuptools.setup(
name = 'stormkafkamon',
version = '0.1.0',
license = 'Apache',
description = '''Monitor offsets of a storm kafka spout.''',
author = '',
author_email = '',
url = 'https://github.com/otoolep/stormkafkamon',
platforms = 'any',
packages = ['stormkafkamon'],
zip_safe = True,
verbose = False,
install_requires = install_requires,
dependency_links = ['https://github.com/mumrah/kafka-python/tarball/0.7#egg=kafka-python-0.7.2-0'],
entry_points={
'console_scripts': [
'skmon = stormkafkamon.monitor:main'
]
},
)
| #!/usr/bin/env python
import setuptools
install_requires = [
'PrettyTable==0.7.2',
'kazoo==1.00',
'simplejson',
'argparse',
'kafka-python'
]
setuptools.setup(
name = 'stormkafkamon',
version = '0.1.0',
license = 'Apache',
description = '''Monitor offsets of a storm kafka spout.''',
author = "Philip O'Toole",
author_email = 'philipomailbox-github@yahoo.com',
url = 'https://github.com/otoolep/stormkafkamon',
platforms = 'any',
packages = ['stormkafkamon'],
zip_safe = True,
verbose = False,
install_requires = install_requires,
dependency_links = ['https://github.com/mumrah/kafka-python/tarball/0.7#egg=kafka-python-0.7.2-0'],
entry_points={
'console_scripts': [
'skmon = stormkafkamon.monitor:main'
]
},
)
| Add author and author email. | Add author and author email.
| Python | apache-2.0 | vivekrao1985/stormkafkamon,ianawilson/stormkafkamon,otoolep/stormkafkamon,grue/stormkafkamon | ---
+++
@@ -15,8 +15,8 @@
version = '0.1.0',
license = 'Apache',
description = '''Monitor offsets of a storm kafka spout.''',
- author = '',
- author_email = '',
+ author = "Philip O'Toole",
+ author_email = 'philipomailbox-github@yahoo.com',
url = 'https://github.com/otoolep/stormkafkamon',
platforms = 'any',
packages = ['stormkafkamon'], |
2493bae247c73ad00dd6205bb738bbecf163221f | setup.py | setup.py | #!/usr/bin/env python
from setuptools import setup, find_packages
setup(
name="django-salmonella",
version='0.6',
author='Lincoln Loop: Seth Buntin, Yann Malet',
author_email='info@lincolnloop.com',
description=("raw_id_fields widget replacement that handles display of an object's "
"string value on change and can be overridden via a template."),
packages=find_packages(),
package_data={'salmonella': [
'static/salmonella/js/*.js',
'static/salmonella/img/*.gif',
'templates/salmonella/*.html',
'templates/salmonella/admin/*.html',
'templates/salmonella/admin/widgets/*.html'
]},
url="http://github.com/lincolnloop/django-salmonella/",
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
| #!/usr/bin/env python
from setuptools import setup, find_packages
setup(
name="django-salmonella",
version='0.6',
author='Lincoln Loop: Seth Buntin, Yann Malet',
author_email='info@lincolnloop.com',
description=("raw_id_fields widget replacement that handles display of an object's "
"string value on change and can be overridden via a template."),
packages=find_packages(),
package_data={'salmonella': [
'static/salmonella/js/*.js',
'static/salmonella/img/*.gif',
'templates/salmonella/*.html',
'templates/salmonella/admin/*.html',
'templates/salmonella/admin/widgets/*.html'
]},
include_package_data=True,
url="http://github.com/lincolnloop/django-salmonella/",
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
| Make sure we include package data too. | Make sure we include package data too. | Python | mit | lincolnloop/django-salmonella,Gustavosdo/django-salmonella,lincolnloop/django-salmonella,lincolnloop/django-salmonella,Gustavosdo/django-salmonella,Gustavosdo/django-salmonella,lincolnloop/django-salmonella | ---
+++
@@ -16,6 +16,7 @@
'templates/salmonella/admin/*.html',
'templates/salmonella/admin/widgets/*.html'
]},
+ include_package_data=True,
url="http://github.com/lincolnloop/django-salmonella/",
zip_safe=False,
classifiers=[ |
1e5c0bee1a41befd4e7571fae05efdec2639aba7 | setup.py | setup.py | from setuptools import setup
setup(
name='colorlog',
version='2.6.0',
description='Log formatting with colors!',
long_description=open("README.rst").read(),
author='Sam Clements',
author_email='sam@borntyping.co.uk',
url='https://github.com/borntyping/python-colorlog',
license='MIT License',
packages=[
'colorlog'
],
extras_require={
'windows': [
'colorama'
]
},
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Topic :: Terminals',
'Topic :: Utilities',
],
)
| from setuptools import setup
setup(
name='colorlog',
version='2.6.0',
description='Log formatting with colors!',
long_description=open("README.rst").read(),
author='Sam Clements',
author_email='sam@borntyping.co.uk',
url='https://github.com/borntyping/python-colorlog',
license='MIT License',
packages=[
'colorlog'
],
extras_require={
':sys_platform=="win32"': [
'colorama'
]
},
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Topic :: Terminals',
'Topic :: Utilities',
],
)
| Change platform test according to wheel docs | Change platform test according to wheel docs
| Python | mit | borntyping/python-colorlog | ---
+++
@@ -17,7 +17,7 @@
],
extras_require={
- 'windows': [
+ ':sys_platform=="win32"': [
'colorama'
]
}, |
325ccef4461cdbcc71508e2616b3c8beddc57d8f | setup.py | setup.py | import re
from setuptools import setup
init_py = open('textlines/__init__.py').read()
metadata = dict(re.findall("__([a-z]+)__ = '([^']+)'", init_py))
metadata['doc'] = re.findall('"""(.+)"""', init_py)[0]
setup(
name='textlines',
version=metadata['version'],
description=metadata['doc'],
author=metadata['author'],
author_email=metadata['email'],
url=metadata['url'],
packages=['textlines'],
include_package_data=True,
install_requires=[
'docopt < 1.0.0'
],
entry_points={
'console_scripts': [
'textlines = textlines.cli:main',
],
},
test_suite='nose.collector',
license=open('LICENSE').read(),
)
| import re
from setuptools import setup
init_py = open('textlines/__init__.py').read()
metadata = dict(re.findall("__([a-z]+)__ = '([^']+)'", init_py))
metadata['doc'] = re.findall('"""((.|\n)*)"""', init_py)[0]
setup(
name='textlines',
version=metadata['version'],
description=metadata['doc'],
author=metadata['author'],
author_email=metadata['email'],
url=metadata['url'],
packages=['textlines'],
include_package_data=True,
install_requires=[
'docopt < 1.0.0'
],
entry_points={
'console_scripts': [
'textlines = textlines.cli:main',
],
},
test_suite='nose.collector',
license=open('LICENSE').read(),
)
| Fix re to match newlines as well | Fix re to match newlines as well
| Python | apache-2.0 | michaeljoseph/textlines,michaeljoseph/textlines | ---
+++
@@ -3,7 +3,7 @@
init_py = open('textlines/__init__.py').read()
metadata = dict(re.findall("__([a-z]+)__ = '([^']+)'", init_py))
-metadata['doc'] = re.findall('"""(.+)"""', init_py)[0]
+metadata['doc'] = re.findall('"""((.|\n)*)"""', init_py)[0]
setup(
name='textlines', |
643cfac6270b97706506812e0bb4b070f50b8085 | setup.py | setup.py | #!/usr/bin/env python
from os.path import exists
import re
from setuptools import setup
contents = re.findall(
r'__version__ = \'([.0-9]+)\'',
open('unification/__init__.py', 'r').read())[0]
setup(name='unification',
version=contents,
description='Unification',
url='http://github.com/mrocklin/unification/',
author='https://raw.github.com/mrocklin/unification/master/AUTHORS.md',
maintainer='Matthew Rocklin',
maintainer_email='mrocklin@gmail.com',
license='BSD',
keywords='unification logic-programming dispatch',
packages=['unification'],
install_requires=open('dependencies.txt').read().split('\n'),
long_description=(open('README.rst').read() if exists('README.rst')
else ''),
zip_safe=False)
| #!/usr/bin/env python
from os.path import exists, join
import re
from setuptools import setup
version = re.findall(
r'__version__ = \'([.0-9]+)\'',
open(join('unification', '__init__.py'), 'r').read())[0]
setup(name='unification',
version=version,
description='Unification',
url='http://github.com/mrocklin/unification/',
author='https://raw.github.com/mrocklin/unification/master/AUTHORS.md',
maintainer='Matthew Rocklin',
maintainer_email='mrocklin@gmail.com',
license='BSD',
keywords='unification logic-programming dispatch',
packages=['unification'],
install_requires=open('dependencies.txt').read().split('\n'),
long_description=(open('README.rst').read() if exists('README.rst')
else ''),
zip_safe=False)
| Join paths to avoid hardcoding path separators | Join paths to avoid hardcoding path separators
| Python | bsd-3-clause | mrocklin/unification | ---
+++
@@ -1,15 +1,15 @@
#!/usr/bin/env python
-from os.path import exists
+from os.path import exists, join
import re
from setuptools import setup
-contents = re.findall(
+version = re.findall(
r'__version__ = \'([.0-9]+)\'',
- open('unification/__init__.py', 'r').read())[0]
+ open(join('unification', '__init__.py'), 'r').read())[0]
setup(name='unification',
- version=contents,
+ version=version,
description='Unification',
url='http://github.com/mrocklin/unification/',
author='https://raw.github.com/mrocklin/unification/master/AUTHORS.md', |
4bad234471f1142ad4172a3ec423c90cec512b2b | setup.py | setup.py | #!/usr/bin/env python
# coding: utf-8
import os
import re
from setuptools import setup, find_packages
def load_required_modules():
with open(os.path.join(os.path.dirname(__file__), "requirements.txt")) as f:
return [line.strip() for line in f.read().strip().split(os.linesep) if line.strip()]
setup(
name='dictmixin',
__version__=re.search(
r'__version__\s*=\s*[\'"]([^\'"]*)[\'"]', # It excludes inline comment too
open('dictmixin/__init__.py').read()).group(1),
description='Parsing mixin which converts `data class instance`, `dict object`, and `json string` each other.',
license='MIT',
author='tadashi-aikawa',
author_email='syou.maman@gmail.com',
maintainer='tadashi-aikawa',
maintainer_email='tadashi-aikawa',
url='https://github.com/tadashi-aikawa/dictmixin.git',
keywords='dict json convert parse each other',
packages=find_packages(exclude=['tests*']),
install_requires=load_required_modules(),
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Intended Audience :: Developers',
'Topic :: Software Development :: Build Tools',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
],
)
| #!/usr/bin/env python
# coding: utf-8
import os
import re
from setuptools import setup, find_packages
def load_required_modules():
with open(os.path.join(os.path.dirname(__file__), "requirements.txt")) as f:
return [line.strip() for line in f.readlines() if line.strip()]
setup(
name='dictmixin',
__version__=re.search(
r'__version__\s*=\s*[\'"]([^\'"]*)[\'"]', # It excludes inline comment too
open('dictmixin/__init__.py').read()).group(1),
description='Parsing mixin which converts `data class instance`, `dict object`, and `json string` each other.',
license='MIT',
author='tadashi-aikawa',
author_email='syou.maman@gmail.com',
maintainer='tadashi-aikawa',
maintainer_email='tadashi-aikawa',
url='https://github.com/tadashi-aikawa/dictmixin.git',
keywords='dict json convert parse each other',
packages=find_packages(exclude=['tests*']),
install_requires=load_required_modules(),
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Intended Audience :: Developers',
'Topic :: Software Development :: Build Tools',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
],
)
| Fix bug which load_required_modules always returns list which length is 1 | Fix bug which load_required_modules always returns list which length is 1
| Python | mit | tadashi-aikawa/owlmixin | ---
+++
@@ -8,7 +8,7 @@
def load_required_modules():
with open(os.path.join(os.path.dirname(__file__), "requirements.txt")) as f:
- return [line.strip() for line in f.read().strip().split(os.linesep) if line.strip()]
+ return [line.strip() for line in f.readlines() if line.strip()]
setup( |
9a510242480dfbf01ab5e635ae3eae034a93f279 | setup.py | setup.py | #!/usr/bin/env python
from ez_setup import use_setuptools
use_setuptools()
import os
from setuptools import setup, find_packages
here = os.path.dirname(__file__)
version_file = os.path.join(here, 'src/iptools/__init__.py')
d = {}
execfile(version_file, d)
version = d['__version__']
setup(
name = 'iptools',
version = version,
description = 'Python utilites for manipulating IP addresses',
long_description = "Utilities for manipulating IP addresses including a class that can be used to include CIDR network blocks in Django's INTERNAL_IPS setting.",
url = 'http://python-iptools.googlecode.com',
download_url = '',
author = 'Bryan Davis',
author_email = 'casadebender+iptools@gmail.com',
license = 'BSD',
platforms = ['any',],
package_dir = {'': 'src'},
packages = find_packages('src'),
include_package_data = True,
test_suite='iptools.test_iptools',
classifiers = [
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Utilities',
'Topic :: Internet',
],
zip_safe=False,
)
| #!/usr/bin/env python
from ez_setup import use_setuptools
use_setuptools()
import os
from setuptools import setup, find_packages
here = os.path.dirname(__file__)
version_file = os.path.join(here, 'src/iptools/__init__.py')
d = {}
execfile(version_file, d)
version = d['__version__']
setup(
name = 'iptools',
version = version,
description = 'Python utilites for manipulating IP addresses',
long_description = "Utilities for manipulating IP addresses including a class that can be used to include CIDR network blocks in Django's INTERNAL_IPS setting.",
url = 'http://python-iptools.googlecode.com',
download_url = '',
author = 'Bryan Davis',
author_email = 'casadebender+iptools@gmail.com',
license = 'BSD',
platforms = ['any',],
package_dir = {'': 'src'},
packages = find_packages('src'),
include_package_data = True,
test_suite='iptools.test_iptools',
classifiers = [
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
'Topic :: Utilities',
'Topic :: Internet',
],
zip_safe=False,
)
| Mark as python 3 compatable. | Mark as python 3 compatable.
git-svn-id: c8188841f5432f3fe42d04dee4f87e556eb5cf84@19 99efc558-b41a-11dd-8714-116ca565c52f
| Python | bsd-2-clause | bd808/python-iptools | ---
+++
@@ -35,6 +35,8 @@
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
+ 'Programming Language :: Python :: 2',
+ 'Programming Language :: Python :: 3',
'Topic :: Utilities',
'Topic :: Internet',
], |
a85847e2450929aa1692e78a34337e9ca4408750 | setup.py | setup.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
with open('README.rst') as readme_file:
readme = readme_file.read()
requirements = [
'pyzmq',
'docopt',
'virtualenv',
'requests'
]
setup(
name='tingbot',
version='0.3',
description="Python APIs to write apps for Tingbot",
long_description=readme,
author="Joe Rickerby",
author_email='joerick@mac.com',
url='https://github.com/tingbot/tingbot-python',
packages=[
'tingbot',
'tbtool'
],
package_dir={'tingbot': 'tingbot',
'tbtool': 'tbtool'},
include_package_data=True,
install_requires=requirements,
license="BSD",
zip_safe=False,
keywords='tingbot',
classifiers=[
'Intended Audience :: Developers',
'Natural Language :: English',
"Programming Language :: Python :: 2",
'Programming Language :: Python :: 2.7',
],
entry_points={
'console_scripts': [
'tbtool = tbtool.__main__:main',
],
},
test_suite='tests',
tests_require=['httpretty','mock'],
)
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
with open('README.rst') as readme_file:
readme = readme_file.read()
requirements = [
'pyzmq',
'docopt',
'virtualenv',
'requests',
'Pillow',
]
setup(
name='tingbot',
version='0.3',
description="Python APIs to write apps for Tingbot",
long_description=readme,
author="Joe Rickerby",
author_email='joerick@mac.com',
url='https://github.com/tingbot/tingbot-python',
packages=[
'tingbot',
'tbtool'
],
package_dir={'tingbot': 'tingbot',
'tbtool': 'tbtool'},
include_package_data=True,
install_requires=requirements,
license="BSD",
zip_safe=False,
keywords='tingbot',
classifiers=[
'Intended Audience :: Developers',
'Natural Language :: English',
"Programming Language :: Python :: 2",
'Programming Language :: Python :: 2.7',
],
entry_points={
'console_scripts': [
'tbtool = tbtool.__main__:main',
],
},
test_suite='tests',
tests_require=['httpretty','mock'],
)
| Add missing dependency on Pillow | Add missing dependency on Pillow
| Python | bsd-2-clause | furbrain/tingbot-python | ---
+++
@@ -13,7 +13,8 @@
'pyzmq',
'docopt',
'virtualenv',
- 'requests'
+ 'requests',
+ 'Pillow',
]
setup( |
dcf76b96f6b6e7c504ab49af4a7f72eabda1b4be | setup.py | setup.py | #!/usr/bin/env python3
# encoding: utf-8
from setuptools import setup, find_packages
setup(
name='pyatv',
version='0.1.1',
license='MIT',
url='https://github.com/postlund/pyatv',
author='Pierre Ståhl',
author_email='pierre.staahl@gmail.com',
description='Library for controlling an Apple TV',
packages=find_packages(exclude=['tests', 'tests.*']),
include_package_data=True,
zip_safe=False,
platforms='any',
install_requires=[
'aiohttp==1.2.0',
'zeroconf==0.18.0',
],
test_suite='tests',
keywords=['apple', 'tv'],
tests_require=['tox'],
entry_points={
'console_scripts': [
'atvremote = pyatv.__main__:main'
]
},
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Topic :: Software Development :: Libraries',
'Topic :: Home Automation',
],
)
| #!/usr/bin/env python3
# encoding: utf-8
from setuptools import setup, find_packages
setup(
name='pyatv',
version='0.1.2.dev1',
license='MIT',
url='https://github.com/postlund/pyatv',
author='Pierre Ståhl',
author_email='pierre.staahl@gmail.com',
description='Library for controlling an Apple TV',
packages=find_packages(exclude=['tests', 'tests.*']),
include_package_data=True,
zip_safe=False,
platforms='any',
install_requires=[
'aiohttp==1.2.0',
'zeroconf==0.18.0',
],
test_suite='tests',
keywords=['apple', 'tv'],
tests_require=['tox'],
entry_points={
'console_scripts': [
'atvremote = pyatv.__main__:main'
]
},
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Topic :: Software Development :: Libraries',
'Topic :: Home Automation',
],
)
| Bump to next dev 0.1.2.dev1 | Bump to next dev 0.1.2.dev1
| Python | mit | postlund/pyatv,postlund/pyatv | ---
+++
@@ -5,7 +5,7 @@
setup(
name='pyatv',
- version='0.1.1',
+ version='0.1.2.dev1',
license='MIT',
url='https://github.com/postlund/pyatv',
author='Pierre Ståhl', |
be0042cdc844d0735f05449749735107c5caff16 | setup.py | setup.py | # Copyright 2013-2014 Massachusetts Open Cloud Contributors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the
# License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an "AS
# IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
# express or implied. See the License for the specific language
# governing permissions and limitations under the License.
from setuptools import setup, find_packages
from pip.req import parse_requirements
from os import path
# requirements_file is the relative path from where setup.py is being
# called from, to where requirements.txt resides
requirements_file = path.join(path.dirname(__file__), 'requirements.txt')
requirements = [str(r.req) for r in parse_requirements(requirements_file)]
setup(name='haas',
version='1.0',
url='https://github.com/CCI-MOC/haas',
packages=find_packages(),
scripts=['scripts/haas', 'scripts/create_bridges'],
install_requires=requirements,
)
| # Copyright 2013-2014 Massachusetts Open Cloud Contributors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the
# License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an "AS
# IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
# express or implied. See the License for the specific language
# governing permissions and limitations under the License.
from setuptools import setup, find_packages
from pip.req import parse_requirements
from os import path
# requirements_file is the relative path from where setup.py is being
# called from, to where requirements.txt resides
requirements_file = path.join(path.dirname(__file__), 'requirements.txt')
requirements = [str(r.req) for r in parse_requirements(requirements_file)]
setup(name='haas',
version='0.1',
url='https://github.com/CCI-MOC/haas',
packages=find_packages(),
scripts=['scripts/haas', 'scripts/create_bridges'],
install_requires=requirements,
)
| Change version number to 0.1 | Change version number to 0.1
...in preparation for the upcoming release.
| Python | apache-2.0 | kylehogan/hil,kylehogan/haas,henn/hil,apoorvemohan/haas,meng-sun/hil,SahilTikale/switchHaaS,kylehogan/hil,apoorvemohan/haas,meng-sun/hil,henn/hil_sahil,SahilTikale/haas,henn/hil_sahil,henn/hil,henn/haas,CCI-MOC/haas | ---
+++
@@ -22,7 +22,7 @@
requirements = [str(r.req) for r in parse_requirements(requirements_file)]
setup(name='haas',
- version='1.0',
+ version='0.1',
url='https://github.com/CCI-MOC/haas',
packages=find_packages(),
scripts=['scripts/haas', 'scripts/create_bridges'], |
67a3c5665a2a91d00a8db8dc8a0fcf12935b5040 | setup.py | setup.py | #!/usr/bin/env python
import sys
from setuptools import setup
from shortuuid import __version__
assert sys.version >= "2.5", "Requires Python v2.5 or above."
classifiers = [
"License :: OSI Approved :: BSD License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.5",
"Programming Language :: Python :: 2.6",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.3",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Topic :: Software Development :: Libraries :: Python Modules",
]
setup(
name="shortuuid",
version=__version__,
author="Stochastic Technologies",
author_email="info@stochastictechnologies.com",
url="https://github.com/stochastic-technologies/shortuuid/",
description="A generator library for concise, " "unambiguous and URL-safe UUIDs.",
long_description="A library that generates short, pretty, "
"unambiguous unique IDs "
"by using an extensive, case-sensitive alphabet and omitting "
"similar-looking letters and numbers.",
license="BSD",
classifiers=classifiers,
packages=["shortuuid"],
test_suite="shortuuid.tests",
tests_require=[],
)
| #!/usr/bin/env python
import sys
from setuptools import setup
from shortuuid import __version__
assert sys.version >= "2.5", "Requires Python v2.5 or above."
classifiers = [
"License :: OSI Approved :: BSD License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.5",
"Programming Language :: Python :: 2.6",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.3",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Topic :: Software Development :: Libraries :: Python Modules",
]
setup(
name="shortuuid",
version=__version__,
author="Stochastic Technologies",
author_email="info@stochastictechnologies.com",
url="https://github.com/stochastic-technologies/shortuuid/",
description="A generator library for concise, " "unambiguous and URL-safe UUIDs.",
long_description="A library that generates short, pretty, "
"unambiguous unique IDs "
"by using an extensive, case-sensitive alphabet and omitting "
"similar-looking letters and numbers.",
license="BSD",
classifiers=classifiers,
packages=["shortuuid"],
test_suite="shortuuid.tests",
tests_require=["pre-commit"],
)
| Add pre-commit as a test dependency | Add pre-commit as a test dependency
| Python | bsd-3-clause | stochastic-technologies/shortuuid,skorokithakis/shortuuid | ---
+++
@@ -36,5 +36,5 @@
classifiers=classifiers,
packages=["shortuuid"],
test_suite="shortuuid.tests",
- tests_require=[],
+ tests_require=["pre-commit"],
) |
c4fd6305b543eb04b4555aa2b2fad0d3e7e83ebb | setup.py | setup.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from setuptools import setup, find_packages
setup(name='megacl',
version='0.2.0',
description='mega.co.nz command line client.',
author='Arthibus Gisséhel',
author_email='public-dev-megacl@gissehel.org',
url='https://github.com/gissehel/megacl.git',
packages=['megacllib'],
scripts=['mcl','megacl'],
license='MIT',
keywords='commandline mega.co.nz mega',
long_description=open('README.rst').read(),
install_requires=['supertools','cltools','mega.py'],
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Operating System :: OS Independent',
'Programming Language :: Python :: 2',
'Topic :: Communications',
'Topic :: Internet',
'Topic :: System :: Filesystems',
'Topic :: Utilities',
],
)
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
from setuptools import setup, find_packages
setup(name='megacl',
version='0.2.1',
description='mega.co.nz command line client.',
author='Arthibus Gisséhel',
author_email='public-dev-megacl@gissehel.org',
url='https://github.com/gissehel/megacl.git',
packages=['megacllib'],
scripts=['mcl','megacl'],
license='MIT',
keywords='commandline mega.co.nz mega',
long_description=open('README.rst').read(),
install_requires=['supertools','cltools','mega.py',
'requests', # non declared yet mega.py dependency
'pycrypto', # non declared yet mega.py dependency
],
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Operating System :: OS Independent',
'Programming Language :: Python :: 2',
'Topic :: Communications',
'Topic :: Internet',
'Topic :: System :: Filesystems',
'Topic :: Utilities',
],
)
| Add mega.py dependency as mega.py doesn't explicitly list them yet | Add mega.py dependency as mega.py doesn't explicitly list them yet
| Python | mit | gissehel/megacl | ---
+++
@@ -4,7 +4,7 @@
from setuptools import setup, find_packages
setup(name='megacl',
- version='0.2.0',
+ version='0.2.1',
description='mega.co.nz command line client.',
author='Arthibus Gisséhel',
author_email='public-dev-megacl@gissehel.org',
@@ -14,7 +14,10 @@
license='MIT',
keywords='commandline mega.co.nz mega',
long_description=open('README.rst').read(),
- install_requires=['supertools','cltools','mega.py'],
+ install_requires=['supertools','cltools','mega.py',
+ 'requests', # non declared yet mega.py dependency
+ 'pycrypto', # non declared yet mega.py dependency
+ ],
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console', |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.