commit stringlengths 40 40 | old_file stringlengths 4 118 | new_file stringlengths 4 118 | old_contents stringlengths 0 2.94k | new_contents stringlengths 1 4.43k | subject stringlengths 15 444 | message stringlengths 16 3.45k | lang stringclasses 1 value | license stringclasses 13 values | repos stringlengths 5 43.2k | prompt stringlengths 17 4.58k | response stringlengths 1 4.43k | prompt_tagged stringlengths 58 4.62k | response_tagged stringlengths 1 4.43k | text stringlengths 132 7.29k | text_tagged stringlengths 173 7.33k |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
f1cc57acd774eace69da7ec0ae9a516207d8ff5c | pyrfc3339/__init__.py | pyrfc3339/__init__.py | """
pyRFC3339 parses and generates :RFC:`3339`-compliant timestamps using Python :class:`datetime.datetime` objects.
>>> from pyrfc3339 import generate, parse
>>> from datetime import datetime
>>> import pytz
>>> generate(datetime.utcnow().replace(tzinfo=pytz.utc)) #doctest:+ELLIPSIS
'...T...Z'
>>> parse('2009-01-01T10:01:02Z')
datetime.datetime(2009, 1, 1, 10, 1, 2, tzinfo=<UTC>)
>>> parse('2009-01-01T14:01:02-04:00')
datetime.datetime(2009, 1, 1, 14, 1, 2, tzinfo=<UTC-04:00>)
"""
from generator import generate
from parser import parse
__all__ = ['generate', 'parse']
| """
pyRFC3339 parses and generates :RFC:`3339`-compliant timestamps using Python :class:`datetime.datetime` objects.
>>> from pyrfc3339 import generate, parse
>>> from datetime import datetime
>>> import pytz
>>> generate(datetime.utcnow().replace(tzinfo=pytz.utc)) #doctest:+ELLIPSIS
'...T...Z'
>>> parse('2009-01-01T10:01:02Z')
datetime.datetime(2009, 1, 1, 10, 1, 2, tzinfo=<UTC>)
>>> parse('2009-01-01T14:01:02-04:00')
datetime.datetime(2009, 1, 1, 14, 1, 2, tzinfo=<UTC-04:00>)
"""
from pyrfc3339.generator import generate
from pyrfc3339.parser import parse
__all__ = ['generate', 'parse']
| Fix imports for Python 3 | Fix imports for Python 3
| Python | mit | kurtraschke/pyRFC3339 | """
pyRFC3339 parses and generates :RFC:`3339`-compliant timestamps using Python :class:`datetime.datetime` objects.
>>> from pyrfc3339 import generate, parse
>>> from datetime import datetime
>>> import pytz
>>> generate(datetime.utcnow().replace(tzinfo=pytz.utc)) #doctest:+ELLIPSIS
'...T...Z'
>>> parse('2009-01-01T10:01:02Z')
datetime.datetime(2009, 1, 1, 10, 1, 2, tzinfo=<UTC>)
>>> parse('2009-01-01T14:01:02-04:00')
datetime.datetime(2009, 1, 1, 14, 1, 2, tzinfo=<UTC-04:00>)
"""
from generator import generate
from parser import parse
__all__ = ['generate', 'parse']
Fix imports for Python 3 | """
pyRFC3339 parses and generates :RFC:`3339`-compliant timestamps using Python :class:`datetime.datetime` objects.
>>> from pyrfc3339 import generate, parse
>>> from datetime import datetime
>>> import pytz
>>> generate(datetime.utcnow().replace(tzinfo=pytz.utc)) #doctest:+ELLIPSIS
'...T...Z'
>>> parse('2009-01-01T10:01:02Z')
datetime.datetime(2009, 1, 1, 10, 1, 2, tzinfo=<UTC>)
>>> parse('2009-01-01T14:01:02-04:00')
datetime.datetime(2009, 1, 1, 14, 1, 2, tzinfo=<UTC-04:00>)
"""
from pyrfc3339.generator import generate
from pyrfc3339.parser import parse
__all__ = ['generate', 'parse']
| <commit_before>"""
pyRFC3339 parses and generates :RFC:`3339`-compliant timestamps using Python :class:`datetime.datetime` objects.
>>> from pyrfc3339 import generate, parse
>>> from datetime import datetime
>>> import pytz
>>> generate(datetime.utcnow().replace(tzinfo=pytz.utc)) #doctest:+ELLIPSIS
'...T...Z'
>>> parse('2009-01-01T10:01:02Z')
datetime.datetime(2009, 1, 1, 10, 1, 2, tzinfo=<UTC>)
>>> parse('2009-01-01T14:01:02-04:00')
datetime.datetime(2009, 1, 1, 14, 1, 2, tzinfo=<UTC-04:00>)
"""
from generator import generate
from parser import parse
__all__ = ['generate', 'parse']
<commit_msg>Fix imports for Python 3<commit_after> | """
pyRFC3339 parses and generates :RFC:`3339`-compliant timestamps using Python :class:`datetime.datetime` objects.
>>> from pyrfc3339 import generate, parse
>>> from datetime import datetime
>>> import pytz
>>> generate(datetime.utcnow().replace(tzinfo=pytz.utc)) #doctest:+ELLIPSIS
'...T...Z'
>>> parse('2009-01-01T10:01:02Z')
datetime.datetime(2009, 1, 1, 10, 1, 2, tzinfo=<UTC>)
>>> parse('2009-01-01T14:01:02-04:00')
datetime.datetime(2009, 1, 1, 14, 1, 2, tzinfo=<UTC-04:00>)
"""
from pyrfc3339.generator import generate
from pyrfc3339.parser import parse
__all__ = ['generate', 'parse']
| """
pyRFC3339 parses and generates :RFC:`3339`-compliant timestamps using Python :class:`datetime.datetime` objects.
>>> from pyrfc3339 import generate, parse
>>> from datetime import datetime
>>> import pytz
>>> generate(datetime.utcnow().replace(tzinfo=pytz.utc)) #doctest:+ELLIPSIS
'...T...Z'
>>> parse('2009-01-01T10:01:02Z')
datetime.datetime(2009, 1, 1, 10, 1, 2, tzinfo=<UTC>)
>>> parse('2009-01-01T14:01:02-04:00')
datetime.datetime(2009, 1, 1, 14, 1, 2, tzinfo=<UTC-04:00>)
"""
from generator import generate
from parser import parse
__all__ = ['generate', 'parse']
Fix imports for Python 3"""
pyRFC3339 parses and generates :RFC:`3339`-compliant timestamps using Python :class:`datetime.datetime` objects.
>>> from pyrfc3339 import generate, parse
>>> from datetime import datetime
>>> import pytz
>>> generate(datetime.utcnow().replace(tzinfo=pytz.utc)) #doctest:+ELLIPSIS
'...T...Z'
>>> parse('2009-01-01T10:01:02Z')
datetime.datetime(2009, 1, 1, 10, 1, 2, tzinfo=<UTC>)
>>> parse('2009-01-01T14:01:02-04:00')
datetime.datetime(2009, 1, 1, 14, 1, 2, tzinfo=<UTC-04:00>)
"""
from pyrfc3339.generator import generate
from pyrfc3339.parser import parse
__all__ = ['generate', 'parse']
| <commit_before>"""
pyRFC3339 parses and generates :RFC:`3339`-compliant timestamps using Python :class:`datetime.datetime` objects.
>>> from pyrfc3339 import generate, parse
>>> from datetime import datetime
>>> import pytz
>>> generate(datetime.utcnow().replace(tzinfo=pytz.utc)) #doctest:+ELLIPSIS
'...T...Z'
>>> parse('2009-01-01T10:01:02Z')
datetime.datetime(2009, 1, 1, 10, 1, 2, tzinfo=<UTC>)
>>> parse('2009-01-01T14:01:02-04:00')
datetime.datetime(2009, 1, 1, 14, 1, 2, tzinfo=<UTC-04:00>)
"""
from generator import generate
from parser import parse
__all__ = ['generate', 'parse']
<commit_msg>Fix imports for Python 3<commit_after>"""
pyRFC3339 parses and generates :RFC:`3339`-compliant timestamps using Python :class:`datetime.datetime` objects.
>>> from pyrfc3339 import generate, parse
>>> from datetime import datetime
>>> import pytz
>>> generate(datetime.utcnow().replace(tzinfo=pytz.utc)) #doctest:+ELLIPSIS
'...T...Z'
>>> parse('2009-01-01T10:01:02Z')
datetime.datetime(2009, 1, 1, 10, 1, 2, tzinfo=<UTC>)
>>> parse('2009-01-01T14:01:02-04:00')
datetime.datetime(2009, 1, 1, 14, 1, 2, tzinfo=<UTC-04:00>)
"""
from pyrfc3339.generator import generate
from pyrfc3339.parser import parse
__all__ = ['generate', 'parse']
|
8337575314ae02e99eeded1ffb537a87a423b2c0 | spam/ansiInventory.py | spam/ansiInventory.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
AnsibleInventory:
INTRO:
USAGE:
"""
import os
import ansible.inventory
class AnsibleInventory(object):
'''
Ansible Inventory wrapper class.
'''
def __init__(self, inventory_filename):
'''
Initialize Inventory
'''
if not os.path.exists(inventory_filename):
print "Provide a valid inventory filename"
return
self.inventory = ansible.inventory.InventoryParser(inventory_filename)
def get_hosts(self, group=None):
'''
Get the hosts
'''
if not group:
return self.inventory.hosts.keys()
groupobj = self.inventory.groups.get(group, None)
if not groupobj:
return None
hostobjs = groupobj.get_hosts()
hostlist = []
for host in hostobjs:
hostlist.append(host.name)
return hostlist
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
AnsibleInventory:
INTRO:
USAGE:
"""
import os
import ansible.inventory
class AnsibleInventory(object):
'''
Ansible Inventory wrapper class.
'''
def __init__(self, inventory_filename):
'''
Initialize Inventory
'''
if not os.path.exists(inventory_filename):
print "Provide a valid inventory filename"
return
self.inventory = ansible.inventory.InventoryParser(inventory_filename)
def get_hosts(self, group=None):
'''
Get the hosts
'''
hostlist = []
for group in self.inventory.groups:
groupdict = {}
groupdict['group'] = group
groupdict['hostlist'] = []
groupobj = self.inventory.groups.get(group)
for host in groupobj.get_hosts():
groupdict['hostlist'].append(host.name)
hostlist.append(groupdict)
return hostlist
| Make changes to get_hosts() to return a list of dict | Make changes to get_hosts() to return a list of dict
| Python | apache-2.0 | bdastur/spam,bdastur/spam | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
AnsibleInventory:
INTRO:
USAGE:
"""
import os
import ansible.inventory
class AnsibleInventory(object):
'''
Ansible Inventory wrapper class.
'''
def __init__(self, inventory_filename):
'''
Initialize Inventory
'''
if not os.path.exists(inventory_filename):
print "Provide a valid inventory filename"
return
self.inventory = ansible.inventory.InventoryParser(inventory_filename)
def get_hosts(self, group=None):
'''
Get the hosts
'''
if not group:
return self.inventory.hosts.keys()
groupobj = self.inventory.groups.get(group, None)
if not groupobj:
return None
hostobjs = groupobj.get_hosts()
hostlist = []
for host in hostobjs:
hostlist.append(host.name)
return hostlist
Make changes to get_hosts() to return a list of dict | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
AnsibleInventory:
INTRO:
USAGE:
"""
import os
import ansible.inventory
class AnsibleInventory(object):
'''
Ansible Inventory wrapper class.
'''
def __init__(self, inventory_filename):
'''
Initialize Inventory
'''
if not os.path.exists(inventory_filename):
print "Provide a valid inventory filename"
return
self.inventory = ansible.inventory.InventoryParser(inventory_filename)
def get_hosts(self, group=None):
'''
Get the hosts
'''
hostlist = []
for group in self.inventory.groups:
groupdict = {}
groupdict['group'] = group
groupdict['hostlist'] = []
groupobj = self.inventory.groups.get(group)
for host in groupobj.get_hosts():
groupdict['hostlist'].append(host.name)
hostlist.append(groupdict)
return hostlist
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
AnsibleInventory:
INTRO:
USAGE:
"""
import os
import ansible.inventory
class AnsibleInventory(object):
'''
Ansible Inventory wrapper class.
'''
def __init__(self, inventory_filename):
'''
Initialize Inventory
'''
if not os.path.exists(inventory_filename):
print "Provide a valid inventory filename"
return
self.inventory = ansible.inventory.InventoryParser(inventory_filename)
def get_hosts(self, group=None):
'''
Get the hosts
'''
if not group:
return self.inventory.hosts.keys()
groupobj = self.inventory.groups.get(group, None)
if not groupobj:
return None
hostobjs = groupobj.get_hosts()
hostlist = []
for host in hostobjs:
hostlist.append(host.name)
return hostlist
<commit_msg>Make changes to get_hosts() to return a list of dict<commit_after> | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
AnsibleInventory:
INTRO:
USAGE:
"""
import os
import ansible.inventory
class AnsibleInventory(object):
'''
Ansible Inventory wrapper class.
'''
def __init__(self, inventory_filename):
'''
Initialize Inventory
'''
if not os.path.exists(inventory_filename):
print "Provide a valid inventory filename"
return
self.inventory = ansible.inventory.InventoryParser(inventory_filename)
def get_hosts(self, group=None):
'''
Get the hosts
'''
hostlist = []
for group in self.inventory.groups:
groupdict = {}
groupdict['group'] = group
groupdict['hostlist'] = []
groupobj = self.inventory.groups.get(group)
for host in groupobj.get_hosts():
groupdict['hostlist'].append(host.name)
hostlist.append(groupdict)
return hostlist
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
AnsibleInventory:
INTRO:
USAGE:
"""
import os
import ansible.inventory
class AnsibleInventory(object):
'''
Ansible Inventory wrapper class.
'''
def __init__(self, inventory_filename):
'''
Initialize Inventory
'''
if not os.path.exists(inventory_filename):
print "Provide a valid inventory filename"
return
self.inventory = ansible.inventory.InventoryParser(inventory_filename)
def get_hosts(self, group=None):
'''
Get the hosts
'''
if not group:
return self.inventory.hosts.keys()
groupobj = self.inventory.groups.get(group, None)
if not groupobj:
return None
hostobjs = groupobj.get_hosts()
hostlist = []
for host in hostobjs:
hostlist.append(host.name)
return hostlist
Make changes to get_hosts() to return a list of dict#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
AnsibleInventory:
INTRO:
USAGE:
"""
import os
import ansible.inventory
class AnsibleInventory(object):
'''
Ansible Inventory wrapper class.
'''
def __init__(self, inventory_filename):
'''
Initialize Inventory
'''
if not os.path.exists(inventory_filename):
print "Provide a valid inventory filename"
return
self.inventory = ansible.inventory.InventoryParser(inventory_filename)
def get_hosts(self, group=None):
'''
Get the hosts
'''
hostlist = []
for group in self.inventory.groups:
groupdict = {}
groupdict['group'] = group
groupdict['hostlist'] = []
groupobj = self.inventory.groups.get(group)
for host in groupobj.get_hosts():
groupdict['hostlist'].append(host.name)
hostlist.append(groupdict)
return hostlist
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
AnsibleInventory:
INTRO:
USAGE:
"""
import os
import ansible.inventory
class AnsibleInventory(object):
'''
Ansible Inventory wrapper class.
'''
def __init__(self, inventory_filename):
'''
Initialize Inventory
'''
if not os.path.exists(inventory_filename):
print "Provide a valid inventory filename"
return
self.inventory = ansible.inventory.InventoryParser(inventory_filename)
def get_hosts(self, group=None):
'''
Get the hosts
'''
if not group:
return self.inventory.hosts.keys()
groupobj = self.inventory.groups.get(group, None)
if not groupobj:
return None
hostobjs = groupobj.get_hosts()
hostlist = []
for host in hostobjs:
hostlist.append(host.name)
return hostlist
<commit_msg>Make changes to get_hosts() to return a list of dict<commit_after>#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
AnsibleInventory:
INTRO:
USAGE:
"""
import os
import ansible.inventory
class AnsibleInventory(object):
'''
Ansible Inventory wrapper class.
'''
def __init__(self, inventory_filename):
'''
Initialize Inventory
'''
if not os.path.exists(inventory_filename):
print "Provide a valid inventory filename"
return
self.inventory = ansible.inventory.InventoryParser(inventory_filename)
def get_hosts(self, group=None):
'''
Get the hosts
'''
hostlist = []
for group in self.inventory.groups:
groupdict = {}
groupdict['group'] = group
groupdict['hostlist'] = []
groupobj = self.inventory.groups.get(group)
for host in groupobj.get_hosts():
groupdict['hostlist'].append(host.name)
hostlist.append(groupdict)
return hostlist
|
ffa3d12e5b45cad56367726bdce83de509bc33a7 | state_tracker/state_defs.py | state_tracker/state_defs.py | # Copyright (c) 2001, Stanford University
# All rights reserved.
#
# See the file LICENSE.txt for information on redistributing this software.
import sys
sys.path.append( "../glapi_parser" )
import apiutil
apiutil.CopyrightDef()
print """DESCRIPTION ""
EXPORTS
"""
keys = apiutil.GetDispatchedFunctions("../glapi_parser/APIspec.txt")
for func_name in apiutil.AllSpecials( 'state' ):
print "crState%s" % func_name
for func_name in apiutil.AllSpecials( 'state_feedback' ):
print "crStateFeedback%s" % func_name
for func_name in apiutil.AllSpecials( 'state_select' ):
print "crStateSelect%s" % func_name
print """crStateInit
crStateCreateContext
crStateDestroyContext
crStateDiffContext
crStateSwitchContext
crStateMakeCurrent
crStateSetCurrent
crStateFlushFunc
crStateFlushArg
crStateDiffAPI
crStateSetCurrentPointers
crStateCurrentRecover
crStateTransformUpdateTransform
crStateColorMaterialRecover
crStateError
crStateUpdateColorBits
crStateClientInit
crStateGetCurrent
crStateLimitsInit
crStateMergeExtensions
crStateRasterPosUpdate
crStateTextureCheckDirtyImages
crStateExtensionsInit
crStateSetExtensionString
crStateNativePixelPacking
crStateUseServerArrays
crStateComputeVersion
__currentBits
"""
| # Copyright (c) 2001, Stanford University
# All rights reserved.
#
# See the file LICENSE.txt for information on redistributing this software.
import sys
sys.path.append( "../glapi_parser" )
import apiutil
apiutil.CopyrightDef()
print """DESCRIPTION ""
EXPORTS
"""
keys = apiutil.GetDispatchedFunctions("../glapi_parser/APIspec.txt")
for func_name in apiutil.AllSpecials( 'state' ):
print "crState%s" % func_name
for func_name in apiutil.AllSpecials( 'state_feedback' ):
print "crStateFeedback%s" % func_name
for func_name in apiutil.AllSpecials( 'state_select' ):
print "crStateSelect%s" % func_name
print """crStateInit
crStateReadPixels
crStateGetChromiumParametervCR
crStateCreateContext
crStateDestroyContext
crStateDiffContext
crStateSwitchContext
crStateMakeCurrent
crStateSetCurrent
crStateFlushFunc
crStateFlushArg
crStateDiffAPI
crStateSetCurrentPointers
crStateCurrentRecover
crStateTransformUpdateTransform
crStateColorMaterialRecover
crStateError
crStateUpdateColorBits
crStateClientInit
crStateGetCurrent
crStateLimitsInit
crStateMergeExtensions
crStateRasterPosUpdate
crStateTextureCheckDirtyImages
crStateExtensionsInit
crStateSetExtensionString
crStateNativePixelPacking
crStateUseServerArrays
crStateComputeVersion
__currentBits
"""
| Fix defs for ReadPixels and GetChromiumParametervCR | Fix defs for ReadPixels and GetChromiumParametervCR
| Python | bsd-3-clause | rpavlik/chromium,rpavlik/chromium,rpavlik/chromium,rpavlik/chromium,rpavlik/chromium | # Copyright (c) 2001, Stanford University
# All rights reserved.
#
# See the file LICENSE.txt for information on redistributing this software.
import sys
sys.path.append( "../glapi_parser" )
import apiutil
apiutil.CopyrightDef()
print """DESCRIPTION ""
EXPORTS
"""
keys = apiutil.GetDispatchedFunctions("../glapi_parser/APIspec.txt")
for func_name in apiutil.AllSpecials( 'state' ):
print "crState%s" % func_name
for func_name in apiutil.AllSpecials( 'state_feedback' ):
print "crStateFeedback%s" % func_name
for func_name in apiutil.AllSpecials( 'state_select' ):
print "crStateSelect%s" % func_name
print """crStateInit
crStateCreateContext
crStateDestroyContext
crStateDiffContext
crStateSwitchContext
crStateMakeCurrent
crStateSetCurrent
crStateFlushFunc
crStateFlushArg
crStateDiffAPI
crStateSetCurrentPointers
crStateCurrentRecover
crStateTransformUpdateTransform
crStateColorMaterialRecover
crStateError
crStateUpdateColorBits
crStateClientInit
crStateGetCurrent
crStateLimitsInit
crStateMergeExtensions
crStateRasterPosUpdate
crStateTextureCheckDirtyImages
crStateExtensionsInit
crStateSetExtensionString
crStateNativePixelPacking
crStateUseServerArrays
crStateComputeVersion
__currentBits
"""
Fix defs for ReadPixels and GetChromiumParametervCR | # Copyright (c) 2001, Stanford University
# All rights reserved.
#
# See the file LICENSE.txt for information on redistributing this software.
import sys
sys.path.append( "../glapi_parser" )
import apiutil
apiutil.CopyrightDef()
print """DESCRIPTION ""
EXPORTS
"""
keys = apiutil.GetDispatchedFunctions("../glapi_parser/APIspec.txt")
for func_name in apiutil.AllSpecials( 'state' ):
print "crState%s" % func_name
for func_name in apiutil.AllSpecials( 'state_feedback' ):
print "crStateFeedback%s" % func_name
for func_name in apiutil.AllSpecials( 'state_select' ):
print "crStateSelect%s" % func_name
print """crStateInit
crStateReadPixels
crStateGetChromiumParametervCR
crStateCreateContext
crStateDestroyContext
crStateDiffContext
crStateSwitchContext
crStateMakeCurrent
crStateSetCurrent
crStateFlushFunc
crStateFlushArg
crStateDiffAPI
crStateSetCurrentPointers
crStateCurrentRecover
crStateTransformUpdateTransform
crStateColorMaterialRecover
crStateError
crStateUpdateColorBits
crStateClientInit
crStateGetCurrent
crStateLimitsInit
crStateMergeExtensions
crStateRasterPosUpdate
crStateTextureCheckDirtyImages
crStateExtensionsInit
crStateSetExtensionString
crStateNativePixelPacking
crStateUseServerArrays
crStateComputeVersion
__currentBits
"""
| <commit_before># Copyright (c) 2001, Stanford University
# All rights reserved.
#
# See the file LICENSE.txt for information on redistributing this software.
import sys
sys.path.append( "../glapi_parser" )
import apiutil
apiutil.CopyrightDef()
print """DESCRIPTION ""
EXPORTS
"""
keys = apiutil.GetDispatchedFunctions("../glapi_parser/APIspec.txt")
for func_name in apiutil.AllSpecials( 'state' ):
print "crState%s" % func_name
for func_name in apiutil.AllSpecials( 'state_feedback' ):
print "crStateFeedback%s" % func_name
for func_name in apiutil.AllSpecials( 'state_select' ):
print "crStateSelect%s" % func_name
print """crStateInit
crStateCreateContext
crStateDestroyContext
crStateDiffContext
crStateSwitchContext
crStateMakeCurrent
crStateSetCurrent
crStateFlushFunc
crStateFlushArg
crStateDiffAPI
crStateSetCurrentPointers
crStateCurrentRecover
crStateTransformUpdateTransform
crStateColorMaterialRecover
crStateError
crStateUpdateColorBits
crStateClientInit
crStateGetCurrent
crStateLimitsInit
crStateMergeExtensions
crStateRasterPosUpdate
crStateTextureCheckDirtyImages
crStateExtensionsInit
crStateSetExtensionString
crStateNativePixelPacking
crStateUseServerArrays
crStateComputeVersion
__currentBits
"""
<commit_msg>Fix defs for ReadPixels and GetChromiumParametervCR<commit_after> | # Copyright (c) 2001, Stanford University
# All rights reserved.
#
# See the file LICENSE.txt for information on redistributing this software.
import sys
sys.path.append( "../glapi_parser" )
import apiutil
apiutil.CopyrightDef()
print """DESCRIPTION ""
EXPORTS
"""
keys = apiutil.GetDispatchedFunctions("../glapi_parser/APIspec.txt")
for func_name in apiutil.AllSpecials( 'state' ):
print "crState%s" % func_name
for func_name in apiutil.AllSpecials( 'state_feedback' ):
print "crStateFeedback%s" % func_name
for func_name in apiutil.AllSpecials( 'state_select' ):
print "crStateSelect%s" % func_name
print """crStateInit
crStateReadPixels
crStateGetChromiumParametervCR
crStateCreateContext
crStateDestroyContext
crStateDiffContext
crStateSwitchContext
crStateMakeCurrent
crStateSetCurrent
crStateFlushFunc
crStateFlushArg
crStateDiffAPI
crStateSetCurrentPointers
crStateCurrentRecover
crStateTransformUpdateTransform
crStateColorMaterialRecover
crStateError
crStateUpdateColorBits
crStateClientInit
crStateGetCurrent
crStateLimitsInit
crStateMergeExtensions
crStateRasterPosUpdate
crStateTextureCheckDirtyImages
crStateExtensionsInit
crStateSetExtensionString
crStateNativePixelPacking
crStateUseServerArrays
crStateComputeVersion
__currentBits
"""
| # Copyright (c) 2001, Stanford University
# All rights reserved.
#
# See the file LICENSE.txt for information on redistributing this software.
import sys
sys.path.append( "../glapi_parser" )
import apiutil
apiutil.CopyrightDef()
print """DESCRIPTION ""
EXPORTS
"""
keys = apiutil.GetDispatchedFunctions("../glapi_parser/APIspec.txt")
for func_name in apiutil.AllSpecials( 'state' ):
print "crState%s" % func_name
for func_name in apiutil.AllSpecials( 'state_feedback' ):
print "crStateFeedback%s" % func_name
for func_name in apiutil.AllSpecials( 'state_select' ):
print "crStateSelect%s" % func_name
print """crStateInit
crStateCreateContext
crStateDestroyContext
crStateDiffContext
crStateSwitchContext
crStateMakeCurrent
crStateSetCurrent
crStateFlushFunc
crStateFlushArg
crStateDiffAPI
crStateSetCurrentPointers
crStateCurrentRecover
crStateTransformUpdateTransform
crStateColorMaterialRecover
crStateError
crStateUpdateColorBits
crStateClientInit
crStateGetCurrent
crStateLimitsInit
crStateMergeExtensions
crStateRasterPosUpdate
crStateTextureCheckDirtyImages
crStateExtensionsInit
crStateSetExtensionString
crStateNativePixelPacking
crStateUseServerArrays
crStateComputeVersion
__currentBits
"""
Fix defs for ReadPixels and GetChromiumParametervCR# Copyright (c) 2001, Stanford University
# All rights reserved.
#
# See the file LICENSE.txt for information on redistributing this software.
import sys
sys.path.append( "../glapi_parser" )
import apiutil
apiutil.CopyrightDef()
print """DESCRIPTION ""
EXPORTS
"""
keys = apiutil.GetDispatchedFunctions("../glapi_parser/APIspec.txt")
for func_name in apiutil.AllSpecials( 'state' ):
print "crState%s" % func_name
for func_name in apiutil.AllSpecials( 'state_feedback' ):
print "crStateFeedback%s" % func_name
for func_name in apiutil.AllSpecials( 'state_select' ):
print "crStateSelect%s" % func_name
print """crStateInit
crStateReadPixels
crStateGetChromiumParametervCR
crStateCreateContext
crStateDestroyContext
crStateDiffContext
crStateSwitchContext
crStateMakeCurrent
crStateSetCurrent
crStateFlushFunc
crStateFlushArg
crStateDiffAPI
crStateSetCurrentPointers
crStateCurrentRecover
crStateTransformUpdateTransform
crStateColorMaterialRecover
crStateError
crStateUpdateColorBits
crStateClientInit
crStateGetCurrent
crStateLimitsInit
crStateMergeExtensions
crStateRasterPosUpdate
crStateTextureCheckDirtyImages
crStateExtensionsInit
crStateSetExtensionString
crStateNativePixelPacking
crStateUseServerArrays
crStateComputeVersion
__currentBits
"""
| <commit_before># Copyright (c) 2001, Stanford University
# All rights reserved.
#
# See the file LICENSE.txt for information on redistributing this software.
import sys
sys.path.append( "../glapi_parser" )
import apiutil
apiutil.CopyrightDef()
print """DESCRIPTION ""
EXPORTS
"""
keys = apiutil.GetDispatchedFunctions("../glapi_parser/APIspec.txt")
for func_name in apiutil.AllSpecials( 'state' ):
print "crState%s" % func_name
for func_name in apiutil.AllSpecials( 'state_feedback' ):
print "crStateFeedback%s" % func_name
for func_name in apiutil.AllSpecials( 'state_select' ):
print "crStateSelect%s" % func_name
print """crStateInit
crStateCreateContext
crStateDestroyContext
crStateDiffContext
crStateSwitchContext
crStateMakeCurrent
crStateSetCurrent
crStateFlushFunc
crStateFlushArg
crStateDiffAPI
crStateSetCurrentPointers
crStateCurrentRecover
crStateTransformUpdateTransform
crStateColorMaterialRecover
crStateError
crStateUpdateColorBits
crStateClientInit
crStateGetCurrent
crStateLimitsInit
crStateMergeExtensions
crStateRasterPosUpdate
crStateTextureCheckDirtyImages
crStateExtensionsInit
crStateSetExtensionString
crStateNativePixelPacking
crStateUseServerArrays
crStateComputeVersion
__currentBits
"""
<commit_msg>Fix defs for ReadPixels and GetChromiumParametervCR<commit_after># Copyright (c) 2001, Stanford University
# All rights reserved.
#
# See the file LICENSE.txt for information on redistributing this software.
import sys
sys.path.append( "../glapi_parser" )
import apiutil
apiutil.CopyrightDef()
print """DESCRIPTION ""
EXPORTS
"""
keys = apiutil.GetDispatchedFunctions("../glapi_parser/APIspec.txt")
for func_name in apiutil.AllSpecials( 'state' ):
print "crState%s" % func_name
for func_name in apiutil.AllSpecials( 'state_feedback' ):
print "crStateFeedback%s" % func_name
for func_name in apiutil.AllSpecials( 'state_select' ):
print "crStateSelect%s" % func_name
print """crStateInit
crStateReadPixels
crStateGetChromiumParametervCR
crStateCreateContext
crStateDestroyContext
crStateDiffContext
crStateSwitchContext
crStateMakeCurrent
crStateSetCurrent
crStateFlushFunc
crStateFlushArg
crStateDiffAPI
crStateSetCurrentPointers
crStateCurrentRecover
crStateTransformUpdateTransform
crStateColorMaterialRecover
crStateError
crStateUpdateColorBits
crStateClientInit
crStateGetCurrent
crStateLimitsInit
crStateMergeExtensions
crStateRasterPosUpdate
crStateTextureCheckDirtyImages
crStateExtensionsInit
crStateSetExtensionString
crStateNativePixelPacking
crStateUseServerArrays
crStateComputeVersion
__currentBits
"""
|
2b249d8a81c51d30d9175ac033c7a0b208684d59 | tests/test_basic.py | tests/test_basic.py | import sys
import pubrunner
def test_countwords():
pubrunner.pubrun('examples/CountWords/',True,True)
| import sys
import pubrunner
import pubrunner.command_line
def test_countwords():
#pubrunner.pubrun('examples/CountWords/',True,True)
sys.argv = ['--test','examples/CountWords/']
pubrunner.command_line.main()
| Test case now runs main directly | Test case now runs main directly
| Python | mit | jakelever/pubrunner,jakelever/pubrunner | import sys
import pubrunner
def test_countwords():
pubrunner.pubrun('examples/CountWords/',True,True)
Test case now runs main directly | import sys
import pubrunner
import pubrunner.command_line
def test_countwords():
#pubrunner.pubrun('examples/CountWords/',True,True)
sys.argv = ['--test','examples/CountWords/']
pubrunner.command_line.main()
| <commit_before>import sys
import pubrunner
def test_countwords():
pubrunner.pubrun('examples/CountWords/',True,True)
<commit_msg>Test case now runs main directly<commit_after> | import sys
import pubrunner
import pubrunner.command_line
def test_countwords():
#pubrunner.pubrun('examples/CountWords/',True,True)
sys.argv = ['--test','examples/CountWords/']
pubrunner.command_line.main()
| import sys
import pubrunner
def test_countwords():
pubrunner.pubrun('examples/CountWords/',True,True)
Test case now runs main directlyimport sys
import pubrunner
import pubrunner.command_line
def test_countwords():
#pubrunner.pubrun('examples/CountWords/',True,True)
sys.argv = ['--test','examples/CountWords/']
pubrunner.command_line.main()
| <commit_before>import sys
import pubrunner
def test_countwords():
pubrunner.pubrun('examples/CountWords/',True,True)
<commit_msg>Test case now runs main directly<commit_after>import sys
import pubrunner
import pubrunner.command_line
def test_countwords():
#pubrunner.pubrun('examples/CountWords/',True,True)
sys.argv = ['--test','examples/CountWords/']
pubrunner.command_line.main()
|
806d3293ebbbd0f30f923e73def902e9c14a0879 | tests/test_match.py | tests/test_match.py | import cv2
import numpy
from nose.tools import raises
import stbt
from _stbt.core import _load_template
def black(width=1280, height=720):
return numpy.zeros((height, width, 3), dtype=numpy.uint8)
def test_that_matchresult_image_matches_template_passed_to_match():
assert stbt.match("black.png", frame=black()).image == "black.png"
def test_that_matchresult_str_image_matches_template_passed_to_match():
assert "image=\'black.png\'" in str(stbt.match("black.png", frame=black()))
def test_that_matchresult_str_image_matches_template_passed_to_match_custom():
assert "image=<Custom Image>" in str(
stbt.match(black(30, 30), frame=black()))
@raises(ValueError)
def test_that_match_rejects_greyscale_template():
grey = cv2.cvtColor(_load_template("black.png").image, cv2.cv.CV_BGR2GRAY)
stbt.match(grey, frame=black())
| import cv2
import numpy
from nose.tools import raises
import stbt
from _stbt.core import _load_template
def black(width=1280, height=720):
return numpy.zeros((height, width, 3), dtype=numpy.uint8)
def test_that_matchresult_image_matches_template_passed_to_match():
assert stbt.match("black.png", frame=black()).image == "black.png"
def test_that_matchresult_str_image_matches_template_passed_to_match():
assert "image=\'black.png\'" in str(stbt.match("black.png", frame=black()))
def test_that_matchresult_str_image_matches_template_passed_to_match_custom():
assert "image=<Custom Image>" in str(
stbt.match(black(30, 30), frame=black()))
def test_matchresult_region_when_first_pyramid_level_fails_to_match():
f = _imread("videotestsrc-full-frame.png")
assert stbt.Region(184, 0, width=92, height=160) == stbt.match(
"videotestsrc-redblue-flipped.png", frame=f).region
@raises(ValueError)
def test_that_match_rejects_greyscale_template():
grey = cv2.cvtColor(_load_template("black.png").image, cv2.cv.CV_BGR2GRAY)
stbt.match(grey, frame=black())
| Add test for the region reported for a failed match | tests: Add test for the region reported for a failed match
If the match fails at the first level of the pyramid algorithm (when
we're searching in a down-scaled version of the frame) then we have to
up-scale the region we report for the best (but not good enough) match.
This passes currently, but there was no test for it and I want to make
sure I don't break it when I implement `stbt.match_all`.
| Python | lgpl-2.1 | martynjarvis/stb-tester,LewisHaley/stb-tester,LewisHaley/stb-tester,LewisHaley/stb-tester,martynjarvis/stb-tester,stb-tester/stb-tester,LewisHaley/stb-tester,martynjarvis/stb-tester,LewisHaley/stb-tester,stb-tester/stb-tester,martynjarvis/stb-tester,LewisHaley/stb-tester,martynjarvis/stb-tester,martynjarvis/stb-tester,stb-tester/stb-tester,LewisHaley/stb-tester,martynjarvis/stb-tester,stb-tester/stb-tester | import cv2
import numpy
from nose.tools import raises
import stbt
from _stbt.core import _load_template
def black(width=1280, height=720):
return numpy.zeros((height, width, 3), dtype=numpy.uint8)
def test_that_matchresult_image_matches_template_passed_to_match():
assert stbt.match("black.png", frame=black()).image == "black.png"
def test_that_matchresult_str_image_matches_template_passed_to_match():
assert "image=\'black.png\'" in str(stbt.match("black.png", frame=black()))
def test_that_matchresult_str_image_matches_template_passed_to_match_custom():
assert "image=<Custom Image>" in str(
stbt.match(black(30, 30), frame=black()))
@raises(ValueError)
def test_that_match_rejects_greyscale_template():
grey = cv2.cvtColor(_load_template("black.png").image, cv2.cv.CV_BGR2GRAY)
stbt.match(grey, frame=black())
tests: Add test for the region reported for a failed match
If the match fails at the first level of the pyramid algorithm (when
we're searching in a down-scaled version of the frame) then we have to
up-scale the region we report for the best (but not good enough) match.
This passes currently, but there was no test for it and I want to make
sure I don't break it when I implement `stbt.match_all`. | import cv2
import numpy
from nose.tools import raises
import stbt
from _stbt.core import _load_template
def black(width=1280, height=720):
return numpy.zeros((height, width, 3), dtype=numpy.uint8)
def test_that_matchresult_image_matches_template_passed_to_match():
assert stbt.match("black.png", frame=black()).image == "black.png"
def test_that_matchresult_str_image_matches_template_passed_to_match():
assert "image=\'black.png\'" in str(stbt.match("black.png", frame=black()))
def test_that_matchresult_str_image_matches_template_passed_to_match_custom():
assert "image=<Custom Image>" in str(
stbt.match(black(30, 30), frame=black()))
def test_matchresult_region_when_first_pyramid_level_fails_to_match():
f = _imread("videotestsrc-full-frame.png")
assert stbt.Region(184, 0, width=92, height=160) == stbt.match(
"videotestsrc-redblue-flipped.png", frame=f).region
@raises(ValueError)
def test_that_match_rejects_greyscale_template():
grey = cv2.cvtColor(_load_template("black.png").image, cv2.cv.CV_BGR2GRAY)
stbt.match(grey, frame=black())
| <commit_before>import cv2
import numpy
from nose.tools import raises
import stbt
from _stbt.core import _load_template
def black(width=1280, height=720):
return numpy.zeros((height, width, 3), dtype=numpy.uint8)
def test_that_matchresult_image_matches_template_passed_to_match():
assert stbt.match("black.png", frame=black()).image == "black.png"
def test_that_matchresult_str_image_matches_template_passed_to_match():
assert "image=\'black.png\'" in str(stbt.match("black.png", frame=black()))
def test_that_matchresult_str_image_matches_template_passed_to_match_custom():
assert "image=<Custom Image>" in str(
stbt.match(black(30, 30), frame=black()))
@raises(ValueError)
def test_that_match_rejects_greyscale_template():
grey = cv2.cvtColor(_load_template("black.png").image, cv2.cv.CV_BGR2GRAY)
stbt.match(grey, frame=black())
<commit_msg>tests: Add test for the region reported for a failed match
If the match fails at the first level of the pyramid algorithm (when
we're searching in a down-scaled version of the frame) then we have to
up-scale the region we report for the best (but not good enough) match.
This passes currently, but there was no test for it and I want to make
sure I don't break it when I implement `stbt.match_all`.<commit_after> | import cv2
import numpy
from nose.tools import raises
import stbt
from _stbt.core import _load_template
def black(width=1280, height=720):
return numpy.zeros((height, width, 3), dtype=numpy.uint8)
def test_that_matchresult_image_matches_template_passed_to_match():
assert stbt.match("black.png", frame=black()).image == "black.png"
def test_that_matchresult_str_image_matches_template_passed_to_match():
assert "image=\'black.png\'" in str(stbt.match("black.png", frame=black()))
def test_that_matchresult_str_image_matches_template_passed_to_match_custom():
assert "image=<Custom Image>" in str(
stbt.match(black(30, 30), frame=black()))
def test_matchresult_region_when_first_pyramid_level_fails_to_match():
f = _imread("videotestsrc-full-frame.png")
assert stbt.Region(184, 0, width=92, height=160) == stbt.match(
"videotestsrc-redblue-flipped.png", frame=f).region
@raises(ValueError)
def test_that_match_rejects_greyscale_template():
grey = cv2.cvtColor(_load_template("black.png").image, cv2.cv.CV_BGR2GRAY)
stbt.match(grey, frame=black())
| import cv2
import numpy
from nose.tools import raises
import stbt
from _stbt.core import _load_template
def black(width=1280, height=720):
return numpy.zeros((height, width, 3), dtype=numpy.uint8)
def test_that_matchresult_image_matches_template_passed_to_match():
assert stbt.match("black.png", frame=black()).image == "black.png"
def test_that_matchresult_str_image_matches_template_passed_to_match():
assert "image=\'black.png\'" in str(stbt.match("black.png", frame=black()))
def test_that_matchresult_str_image_matches_template_passed_to_match_custom():
assert "image=<Custom Image>" in str(
stbt.match(black(30, 30), frame=black()))
@raises(ValueError)
def test_that_match_rejects_greyscale_template():
grey = cv2.cvtColor(_load_template("black.png").image, cv2.cv.CV_BGR2GRAY)
stbt.match(grey, frame=black())
tests: Add test for the region reported for a failed match
If the match fails at the first level of the pyramid algorithm (when
we're searching in a down-scaled version of the frame) then we have to
up-scale the region we report for the best (but not good enough) match.
This passes currently, but there was no test for it and I want to make
sure I don't break it when I implement `stbt.match_all`.import cv2
import numpy
from nose.tools import raises
import stbt
from _stbt.core import _load_template
def black(width=1280, height=720):
return numpy.zeros((height, width, 3), dtype=numpy.uint8)
def test_that_matchresult_image_matches_template_passed_to_match():
assert stbt.match("black.png", frame=black()).image == "black.png"
def test_that_matchresult_str_image_matches_template_passed_to_match():
assert "image=\'black.png\'" in str(stbt.match("black.png", frame=black()))
def test_that_matchresult_str_image_matches_template_passed_to_match_custom():
assert "image=<Custom Image>" in str(
stbt.match(black(30, 30), frame=black()))
def test_matchresult_region_when_first_pyramid_level_fails_to_match():
f = _imread("videotestsrc-full-frame.png")
assert stbt.Region(184, 0, width=92, height=160) == stbt.match(
"videotestsrc-redblue-flipped.png", frame=f).region
@raises(ValueError)
def test_that_match_rejects_greyscale_template():
grey = cv2.cvtColor(_load_template("black.png").image, cv2.cv.CV_BGR2GRAY)
stbt.match(grey, frame=black())
| <commit_before>import cv2
import numpy
from nose.tools import raises
import stbt
from _stbt.core import _load_template
def black(width=1280, height=720):
return numpy.zeros((height, width, 3), dtype=numpy.uint8)
def test_that_matchresult_image_matches_template_passed_to_match():
assert stbt.match("black.png", frame=black()).image == "black.png"
def test_that_matchresult_str_image_matches_template_passed_to_match():
assert "image=\'black.png\'" in str(stbt.match("black.png", frame=black()))
def test_that_matchresult_str_image_matches_template_passed_to_match_custom():
assert "image=<Custom Image>" in str(
stbt.match(black(30, 30), frame=black()))
@raises(ValueError)
def test_that_match_rejects_greyscale_template():
grey = cv2.cvtColor(_load_template("black.png").image, cv2.cv.CV_BGR2GRAY)
stbt.match(grey, frame=black())
<commit_msg>tests: Add test for the region reported for a failed match
If the match fails at the first level of the pyramid algorithm (when
we're searching in a down-scaled version of the frame) then we have to
up-scale the region we report for the best (but not good enough) match.
This passes currently, but there was no test for it and I want to make
sure I don't break it when I implement `stbt.match_all`.<commit_after>import cv2
import numpy
from nose.tools import raises
import stbt
from _stbt.core import _load_template
def black(width=1280, height=720):
return numpy.zeros((height, width, 3), dtype=numpy.uint8)
def test_that_matchresult_image_matches_template_passed_to_match():
assert stbt.match("black.png", frame=black()).image == "black.png"
def test_that_matchresult_str_image_matches_template_passed_to_match():
assert "image=\'black.png\'" in str(stbt.match("black.png", frame=black()))
def test_that_matchresult_str_image_matches_template_passed_to_match_custom():
assert "image=<Custom Image>" in str(
stbt.match(black(30, 30), frame=black()))
def test_matchresult_region_when_first_pyramid_level_fails_to_match():
f = _imread("videotestsrc-full-frame.png")
assert stbt.Region(184, 0, width=92, height=160) == stbt.match(
"videotestsrc-redblue-flipped.png", frame=f).region
@raises(ValueError)
def test_that_match_rejects_greyscale_template():
grey = cv2.cvtColor(_load_template("black.png").image, cv2.cv.CV_BGR2GRAY)
stbt.match(grey, frame=black())
|
b72f3f6e180bc332579e71b4abeba2b36319a55e | regscrape/settings.py | regscrape/settings.py | TARGET_SERVER = 'www.regulations.gov'
DEBUG = True
DB_NAME = 'regulations'
ES_HOST = 'thrift://localhost:9500'
DATA_DIR = '/data'
EXTRACTORS = 2
DUMP_START = 0
DUMP_END = 3850000
DUMP_INCREMENT = 50000
MAX_WAIT = 600
CHUNK_SIZE = 10
FILTER = {}
INSTANCES = 2
THREADS_PER_INSTANCE = 2
SITES = ['regsdotgov']
try:
from local_settings import *
except:
pass
| TARGET_SERVER = 'www.regulations.gov'
DEBUG = True
DB_NAME = 'regulations'
ES_HOST = 'thrift://localhost:9500'
DATA_DIR = '/data'
EXTRACTORS = 2
DUMP_START = 0
DUMP_END = 3850000
DUMP_INCREMENT = 10000
MAX_WAIT = 600
CHUNK_SIZE = 10
FILTER = {}
INSTANCES = 2
THREADS_PER_INSTANCE = 2
SITES = ['regsdotgov']
try:
from local_settings import *
except:
pass
| Reduce dump increment to 10k. | Reduce dump increment to 10k.
| Python | bsd-3-clause | sunlightlabs/regulations-scraper,sunlightlabs/regulations-scraper,sunlightlabs/regulations-scraper | TARGET_SERVER = 'www.regulations.gov'
DEBUG = True
DB_NAME = 'regulations'
ES_HOST = 'thrift://localhost:9500'
DATA_DIR = '/data'
EXTRACTORS = 2
DUMP_START = 0
DUMP_END = 3850000
DUMP_INCREMENT = 50000
MAX_WAIT = 600
CHUNK_SIZE = 10
FILTER = {}
INSTANCES = 2
THREADS_PER_INSTANCE = 2
SITES = ['regsdotgov']
try:
from local_settings import *
except:
pass
Reduce dump increment to 10k. | TARGET_SERVER = 'www.regulations.gov'
DEBUG = True
DB_NAME = 'regulations'
ES_HOST = 'thrift://localhost:9500'
DATA_DIR = '/data'
EXTRACTORS = 2
DUMP_START = 0
DUMP_END = 3850000
DUMP_INCREMENT = 10000
MAX_WAIT = 600
CHUNK_SIZE = 10
FILTER = {}
INSTANCES = 2
THREADS_PER_INSTANCE = 2
SITES = ['regsdotgov']
try:
from local_settings import *
except:
pass
| <commit_before>TARGET_SERVER = 'www.regulations.gov'
DEBUG = True
DB_NAME = 'regulations'
ES_HOST = 'thrift://localhost:9500'
DATA_DIR = '/data'
EXTRACTORS = 2
DUMP_START = 0
DUMP_END = 3850000
DUMP_INCREMENT = 50000
MAX_WAIT = 600
CHUNK_SIZE = 10
FILTER = {}
INSTANCES = 2
THREADS_PER_INSTANCE = 2
SITES = ['regsdotgov']
try:
from local_settings import *
except:
pass
<commit_msg>Reduce dump increment to 10k.<commit_after> | TARGET_SERVER = 'www.regulations.gov'
DEBUG = True
DB_NAME = 'regulations'
ES_HOST = 'thrift://localhost:9500'
DATA_DIR = '/data'
EXTRACTORS = 2
DUMP_START = 0
DUMP_END = 3850000
DUMP_INCREMENT = 10000
MAX_WAIT = 600
CHUNK_SIZE = 10
FILTER = {}
INSTANCES = 2
THREADS_PER_INSTANCE = 2
SITES = ['regsdotgov']
try:
from local_settings import *
except:
pass
| TARGET_SERVER = 'www.regulations.gov'
DEBUG = True
DB_NAME = 'regulations'
ES_HOST = 'thrift://localhost:9500'
DATA_DIR = '/data'
EXTRACTORS = 2
DUMP_START = 0
DUMP_END = 3850000
DUMP_INCREMENT = 50000
MAX_WAIT = 600
CHUNK_SIZE = 10
FILTER = {}
INSTANCES = 2
THREADS_PER_INSTANCE = 2
SITES = ['regsdotgov']
try:
from local_settings import *
except:
pass
Reduce dump increment to 10k.TARGET_SERVER = 'www.regulations.gov'
DEBUG = True
DB_NAME = 'regulations'
ES_HOST = 'thrift://localhost:9500'
DATA_DIR = '/data'
EXTRACTORS = 2
DUMP_START = 0
DUMP_END = 3850000
DUMP_INCREMENT = 10000
MAX_WAIT = 600
CHUNK_SIZE = 10
FILTER = {}
INSTANCES = 2
THREADS_PER_INSTANCE = 2
SITES = ['regsdotgov']
try:
from local_settings import *
except:
pass
| <commit_before>TARGET_SERVER = 'www.regulations.gov'
DEBUG = True
DB_NAME = 'regulations'
ES_HOST = 'thrift://localhost:9500'
DATA_DIR = '/data'
EXTRACTORS = 2
DUMP_START = 0
DUMP_END = 3850000
DUMP_INCREMENT = 50000
MAX_WAIT = 600
CHUNK_SIZE = 10
FILTER = {}
INSTANCES = 2
THREADS_PER_INSTANCE = 2
SITES = ['regsdotgov']
try:
from local_settings import *
except:
pass
<commit_msg>Reduce dump increment to 10k.<commit_after>TARGET_SERVER = 'www.regulations.gov'
DEBUG = True
DB_NAME = 'regulations'
ES_HOST = 'thrift://localhost:9500'
DATA_DIR = '/data'
EXTRACTORS = 2
DUMP_START = 0
DUMP_END = 3850000
DUMP_INCREMENT = 10000
MAX_WAIT = 600
CHUNK_SIZE = 10
FILTER = {}
INSTANCES = 2
THREADS_PER_INSTANCE = 2
SITES = ['regsdotgov']
try:
from local_settings import *
except:
pass
|
d254cf6960f2d04e02ed252c4461994483a9d0f5 | launch_control/models/hw_device.py | launch_control/models/hw_device.py | """
Module with the HardwareDevice model.
"""
from launch_control.utils.json import PlainOldData
class HardwareDevice(PlainOldData):
"""
Model representing any HardwareDevice
A device is just a "device_type" attribute with a bag of properties
and a human readable description. Individual device types can be
freely added. For simplicity some common types of devices are
provided as class properties DEVICE_xxx.
Instances will come from a variety of factory classes, each capable
of enumerating devices that it understands. The upside of having a
common class like this is that it's easier to store it in the
database _and_ not have to agree on a common set of properties for,
say, all CPUs.
If you want you can create instances manually, like this:
>>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU)
>>> cpu.desc = "800MHz OMAP3 Processor"
>>> cpu.attributes['machine'] = 'arm'
>>> cpu.attributes['mhz'] = 800
>>> cpu.attributes['vendor'] = 'Texas Instruments'
"""
DEVICE_CPU = "device.cpu"
DEVICE_MEM = "device.mem"
DEVICE_USB = "device.usb"
DEVICE_PCI = "device.pci"
DEVICE_BOARD = "device.board"
__slots__ = ('device_type', 'desc', 'attributes')
def __init__(self, device_type, description, attributes=None):
self.device_type = device_type
self.description = description
self.attributes = attributes or {}
| """
Module with the HardwareDevice model.
"""
from launch_control.utils.json import PlainOldData
class HardwareDevice(PlainOldData):
"""
Model representing any HardwareDevice
A device is just a "device_type" attribute with a bag of properties
and a human readable description. Individual device types can be
freely added. For simplicity some common types of devices are
provided as class properties DEVICE_xxx.
Instances will come from a variety of factory classes, each capable
of enumerating devices that it understands. The upside of having a
common class like this is that it's easier to store it in the
database _and_ not have to agree on a common set of properties for,
say, all CPUs.
If you want you can create instances manually, like this:
>>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU,
... u"800MHz OMAP3 Processor")
>>> cpu.attributes[u'machine'] = u'arm'
>>> cpu.attributes[u'mhz'] = '800'
>>> cpu.attributes[u'vendor'] = u'Texas Instruments'
"""
DEVICE_CPU = "device.cpu"
DEVICE_MEM = "device.mem"
DEVICE_USB = "device.usb"
DEVICE_PCI = "device.pci"
DEVICE_BOARD = "device.board"
__slots__ = ('device_type', 'desc', 'attributes')
def __init__(self, device_type, description, attributes=None):
self.device_type = device_type
self.description = description
self.attributes = attributes or {}
| Fix HardwareDevice docstring to match implementation | Fix HardwareDevice docstring to match implementation
| Python | agpl-3.0 | Linaro/lava-server,OSSystems/lava-server,OSSystems/lava-server,Linaro/lava-server,OSSystems/lava-server,Linaro/lava-server,Linaro/lava-server | """
Module with the HardwareDevice model.
"""
from launch_control.utils.json import PlainOldData
class HardwareDevice(PlainOldData):
"""
Model representing any HardwareDevice
A device is just a "device_type" attribute with a bag of properties
and a human readable description. Individual device types can be
freely added. For simplicity some common types of devices are
provided as class properties DEVICE_xxx.
Instances will come from a variety of factory classes, each capable
of enumerating devices that it understands. The upside of having a
common class like this is that it's easier to store it in the
database _and_ not have to agree on a common set of properties for,
say, all CPUs.
If you want you can create instances manually, like this:
>>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU)
>>> cpu.desc = "800MHz OMAP3 Processor"
>>> cpu.attributes['machine'] = 'arm'
>>> cpu.attributes['mhz'] = 800
>>> cpu.attributes['vendor'] = 'Texas Instruments'
"""
DEVICE_CPU = "device.cpu"
DEVICE_MEM = "device.mem"
DEVICE_USB = "device.usb"
DEVICE_PCI = "device.pci"
DEVICE_BOARD = "device.board"
__slots__ = ('device_type', 'desc', 'attributes')
def __init__(self, device_type, description, attributes=None):
self.device_type = device_type
self.description = description
self.attributes = attributes or {}
Fix HardwareDevice docstring to match implementation | """
Module with the HardwareDevice model.
"""
from launch_control.utils.json import PlainOldData
class HardwareDevice(PlainOldData):
"""
Model representing any HardwareDevice
A device is just a "device_type" attribute with a bag of properties
and a human readable description. Individual device types can be
freely added. For simplicity some common types of devices are
provided as class properties DEVICE_xxx.
Instances will come from a variety of factory classes, each capable
of enumerating devices that it understands. The upside of having a
common class like this is that it's easier to store it in the
database _and_ not have to agree on a common set of properties for,
say, all CPUs.
If you want you can create instances manually, like this:
>>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU,
... u"800MHz OMAP3 Processor")
>>> cpu.attributes[u'machine'] = u'arm'
>>> cpu.attributes[u'mhz'] = '800'
>>> cpu.attributes[u'vendor'] = u'Texas Instruments'
"""
DEVICE_CPU = "device.cpu"
DEVICE_MEM = "device.mem"
DEVICE_USB = "device.usb"
DEVICE_PCI = "device.pci"
DEVICE_BOARD = "device.board"
__slots__ = ('device_type', 'desc', 'attributes')
def __init__(self, device_type, description, attributes=None):
self.device_type = device_type
self.description = description
self.attributes = attributes or {}
| <commit_before>"""
Module with the HardwareDevice model.
"""
from launch_control.utils.json import PlainOldData
class HardwareDevice(PlainOldData):
"""
Model representing any HardwareDevice
A device is just a "device_type" attribute with a bag of properties
and a human readable description. Individual device types can be
freely added. For simplicity some common types of devices are
provided as class properties DEVICE_xxx.
Instances will come from a variety of factory classes, each capable
of enumerating devices that it understands. The upside of having a
common class like this is that it's easier to store it in the
database _and_ not have to agree on a common set of properties for,
say, all CPUs.
If you want you can create instances manually, like this:
>>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU)
>>> cpu.desc = "800MHz OMAP3 Processor"
>>> cpu.attributes['machine'] = 'arm'
>>> cpu.attributes['mhz'] = 800
>>> cpu.attributes['vendor'] = 'Texas Instruments'
"""
DEVICE_CPU = "device.cpu"
DEVICE_MEM = "device.mem"
DEVICE_USB = "device.usb"
DEVICE_PCI = "device.pci"
DEVICE_BOARD = "device.board"
__slots__ = ('device_type', 'desc', 'attributes')
def __init__(self, device_type, description, attributes=None):
self.device_type = device_type
self.description = description
self.attributes = attributes or {}
<commit_msg>Fix HardwareDevice docstring to match implementation<commit_after> | """
Module with the HardwareDevice model.
"""
from launch_control.utils.json import PlainOldData
class HardwareDevice(PlainOldData):
"""
Model representing any HardwareDevice
A device is just a "device_type" attribute with a bag of properties
and a human readable description. Individual device types can be
freely added. For simplicity some common types of devices are
provided as class properties DEVICE_xxx.
Instances will come from a variety of factory classes, each capable
of enumerating devices that it understands. The upside of having a
common class like this is that it's easier to store it in the
database _and_ not have to agree on a common set of properties for,
say, all CPUs.
If you want you can create instances manually, like this:
>>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU,
... u"800MHz OMAP3 Processor")
>>> cpu.attributes[u'machine'] = u'arm'
>>> cpu.attributes[u'mhz'] = '800'
>>> cpu.attributes[u'vendor'] = u'Texas Instruments'
"""
DEVICE_CPU = "device.cpu"
DEVICE_MEM = "device.mem"
DEVICE_USB = "device.usb"
DEVICE_PCI = "device.pci"
DEVICE_BOARD = "device.board"
__slots__ = ('device_type', 'desc', 'attributes')
def __init__(self, device_type, description, attributes=None):
self.device_type = device_type
self.description = description
self.attributes = attributes or {}
| """
Module with the HardwareDevice model.
"""
from launch_control.utils.json import PlainOldData
class HardwareDevice(PlainOldData):
"""
Model representing any HardwareDevice
A device is just a "device_type" attribute with a bag of properties
and a human readable description. Individual device types can be
freely added. For simplicity some common types of devices are
provided as class properties DEVICE_xxx.
Instances will come from a variety of factory classes, each capable
of enumerating devices that it understands. The upside of having a
common class like this is that it's easier to store it in the
database _and_ not have to agree on a common set of properties for,
say, all CPUs.
If you want you can create instances manually, like this:
>>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU)
>>> cpu.desc = "800MHz OMAP3 Processor"
>>> cpu.attributes['machine'] = 'arm'
>>> cpu.attributes['mhz'] = 800
>>> cpu.attributes['vendor'] = 'Texas Instruments'
"""
DEVICE_CPU = "device.cpu"
DEVICE_MEM = "device.mem"
DEVICE_USB = "device.usb"
DEVICE_PCI = "device.pci"
DEVICE_BOARD = "device.board"
__slots__ = ('device_type', 'desc', 'attributes')
def __init__(self, device_type, description, attributes=None):
self.device_type = device_type
self.description = description
self.attributes = attributes or {}
Fix HardwareDevice docstring to match implementation"""
Module with the HardwareDevice model.
"""
from launch_control.utils.json import PlainOldData
class HardwareDevice(PlainOldData):
"""
Model representing any HardwareDevice
A device is just a "device_type" attribute with a bag of properties
and a human readable description. Individual device types can be
freely added. For simplicity some common types of devices are
provided as class properties DEVICE_xxx.
Instances will come from a variety of factory classes, each capable
of enumerating devices that it understands. The upside of having a
common class like this is that it's easier to store it in the
database _and_ not have to agree on a common set of properties for,
say, all CPUs.
If you want you can create instances manually, like this:
>>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU,
... u"800MHz OMAP3 Processor")
>>> cpu.attributes[u'machine'] = u'arm'
>>> cpu.attributes[u'mhz'] = '800'
>>> cpu.attributes[u'vendor'] = u'Texas Instruments'
"""
DEVICE_CPU = "device.cpu"
DEVICE_MEM = "device.mem"
DEVICE_USB = "device.usb"
DEVICE_PCI = "device.pci"
DEVICE_BOARD = "device.board"
__slots__ = ('device_type', 'desc', 'attributes')
def __init__(self, device_type, description, attributes=None):
self.device_type = device_type
self.description = description
self.attributes = attributes or {}
| <commit_before>"""
Module with the HardwareDevice model.
"""
from launch_control.utils.json import PlainOldData
class HardwareDevice(PlainOldData):
"""
Model representing any HardwareDevice
A device is just a "device_type" attribute with a bag of properties
and a human readable description. Individual device types can be
freely added. For simplicity some common types of devices are
provided as class properties DEVICE_xxx.
Instances will come from a variety of factory classes, each capable
of enumerating devices that it understands. The upside of having a
common class like this is that it's easier to store it in the
database _and_ not have to agree on a common set of properties for,
say, all CPUs.
If you want you can create instances manually, like this:
>>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU)
>>> cpu.desc = "800MHz OMAP3 Processor"
>>> cpu.attributes['machine'] = 'arm'
>>> cpu.attributes['mhz'] = 800
>>> cpu.attributes['vendor'] = 'Texas Instruments'
"""
DEVICE_CPU = "device.cpu"
DEVICE_MEM = "device.mem"
DEVICE_USB = "device.usb"
DEVICE_PCI = "device.pci"
DEVICE_BOARD = "device.board"
__slots__ = ('device_type', 'desc', 'attributes')
def __init__(self, device_type, description, attributes=None):
self.device_type = device_type
self.description = description
self.attributes = attributes or {}
<commit_msg>Fix HardwareDevice docstring to match implementation<commit_after>"""
Module with the HardwareDevice model.
"""
from launch_control.utils.json import PlainOldData
class HardwareDevice(PlainOldData):
"""
Model representing any HardwareDevice
A device is just a "device_type" attribute with a bag of properties
and a human readable description. Individual device types can be
freely added. For simplicity some common types of devices are
provided as class properties DEVICE_xxx.
Instances will come from a variety of factory classes, each capable
of enumerating devices that it understands. The upside of having a
common class like this is that it's easier to store it in the
database _and_ not have to agree on a common set of properties for,
say, all CPUs.
If you want you can create instances manually, like this:
>>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU,
... u"800MHz OMAP3 Processor")
>>> cpu.attributes[u'machine'] = u'arm'
>>> cpu.attributes[u'mhz'] = '800'
>>> cpu.attributes[u'vendor'] = u'Texas Instruments'
"""
DEVICE_CPU = "device.cpu"
DEVICE_MEM = "device.mem"
DEVICE_USB = "device.usb"
DEVICE_PCI = "device.pci"
DEVICE_BOARD = "device.board"
__slots__ = ('device_type', 'desc', 'attributes')
def __init__(self, device_type, description, attributes=None):
self.device_type = device_type
self.description = description
self.attributes = attributes or {}
|
b97fd14bba5d45a6e4e3caa02bd947bddfd0ba8b | tools/sniper_stats_jobid.py | tools/sniper_stats_jobid.py | import sniper_stats, intelqueue, iqclient
class SniperStatsJobid(sniper_stats.SniperStatsBase):
def __init__(self, jobid):
self.jobid = jobid
self.ic = iqclient.IntelClient()
def read_metricnames(self):
return self.ic.graphite_dbresults(self.jobid, 'read_metricnames')
def get_snapshots(self):
return self.ic.graphite_dbresults(self.jobid, 'get_snapshots')
def read_snapshot(self, prefix, metrics = None):
return self.ic.graphite_dbresults(self.jobid, 'read_snapshot', {'prefix': prefix, 'metrics': metrics})
def get_topology(self):
return self.ic.graphite_dbresults(self.jobid, 'get_topology')
def get_markers(self):
return self.ic.graphite_dbresults(self.jobid, 'get_markers')
def get_events(self):
return self.ic.graphite_dbresults(self.jobid, 'get_events')
| import sniper_stats, intelqueue, iqclient
class SniperStatsJobid(sniper_stats.SniperStatsBase):
def __init__(self, jobid):
self.jobid = jobid
self.ic = iqclient.IntelClient()
self.names = self.read_metricnames()
def read_metricnames(self):
return self.ic.graphite_dbresults(self.jobid, 'read_metricnames')
def get_snapshots(self):
return self.ic.graphite_dbresults(self.jobid, 'get_snapshots')
def read_snapshot(self, prefix, metrics = None):
return self.ic.graphite_dbresults(self.jobid, 'read_snapshot', {'prefix': prefix, 'metrics': metrics})
def get_topology(self):
return self.ic.graphite_dbresults(self.jobid, 'get_topology')
def get_markers(self):
return self.ic.graphite_dbresults(self.jobid, 'get_markers')
def get_events(self):
return self.ic.graphite_dbresults(self.jobid, 'get_events')
| Read metric names on startup for jobid-based stats so self.names is available as expected | [sniper_stats] Read metric names on startup for jobid-based stats so self.names is available as expected
| Python | mit | abanaiyan/sniper,abanaiyan/sniper,abanaiyan/sniper,abanaiyan/sniper,abanaiyan/sniper | import sniper_stats, intelqueue, iqclient
class SniperStatsJobid(sniper_stats.SniperStatsBase):
def __init__(self, jobid):
self.jobid = jobid
self.ic = iqclient.IntelClient()
def read_metricnames(self):
return self.ic.graphite_dbresults(self.jobid, 'read_metricnames')
def get_snapshots(self):
return self.ic.graphite_dbresults(self.jobid, 'get_snapshots')
def read_snapshot(self, prefix, metrics = None):
return self.ic.graphite_dbresults(self.jobid, 'read_snapshot', {'prefix': prefix, 'metrics': metrics})
def get_topology(self):
return self.ic.graphite_dbresults(self.jobid, 'get_topology')
def get_markers(self):
return self.ic.graphite_dbresults(self.jobid, 'get_markers')
def get_events(self):
return self.ic.graphite_dbresults(self.jobid, 'get_events')
[sniper_stats] Read metric names on startup for jobid-based stats so self.names is available as expected | import sniper_stats, intelqueue, iqclient
class SniperStatsJobid(sniper_stats.SniperStatsBase):
def __init__(self, jobid):
self.jobid = jobid
self.ic = iqclient.IntelClient()
self.names = self.read_metricnames()
def read_metricnames(self):
return self.ic.graphite_dbresults(self.jobid, 'read_metricnames')
def get_snapshots(self):
return self.ic.graphite_dbresults(self.jobid, 'get_snapshots')
def read_snapshot(self, prefix, metrics = None):
return self.ic.graphite_dbresults(self.jobid, 'read_snapshot', {'prefix': prefix, 'metrics': metrics})
def get_topology(self):
return self.ic.graphite_dbresults(self.jobid, 'get_topology')
def get_markers(self):
return self.ic.graphite_dbresults(self.jobid, 'get_markers')
def get_events(self):
return self.ic.graphite_dbresults(self.jobid, 'get_events')
| <commit_before>import sniper_stats, intelqueue, iqclient
class SniperStatsJobid(sniper_stats.SniperStatsBase):
def __init__(self, jobid):
self.jobid = jobid
self.ic = iqclient.IntelClient()
def read_metricnames(self):
return self.ic.graphite_dbresults(self.jobid, 'read_metricnames')
def get_snapshots(self):
return self.ic.graphite_dbresults(self.jobid, 'get_snapshots')
def read_snapshot(self, prefix, metrics = None):
return self.ic.graphite_dbresults(self.jobid, 'read_snapshot', {'prefix': prefix, 'metrics': metrics})
def get_topology(self):
return self.ic.graphite_dbresults(self.jobid, 'get_topology')
def get_markers(self):
return self.ic.graphite_dbresults(self.jobid, 'get_markers')
def get_events(self):
return self.ic.graphite_dbresults(self.jobid, 'get_events')
<commit_msg>[sniper_stats] Read metric names on startup for jobid-based stats so self.names is available as expected<commit_after> | import sniper_stats, intelqueue, iqclient
class SniperStatsJobid(sniper_stats.SniperStatsBase):
def __init__(self, jobid):
self.jobid = jobid
self.ic = iqclient.IntelClient()
self.names = self.read_metricnames()
def read_metricnames(self):
return self.ic.graphite_dbresults(self.jobid, 'read_metricnames')
def get_snapshots(self):
return self.ic.graphite_dbresults(self.jobid, 'get_snapshots')
def read_snapshot(self, prefix, metrics = None):
return self.ic.graphite_dbresults(self.jobid, 'read_snapshot', {'prefix': prefix, 'metrics': metrics})
def get_topology(self):
return self.ic.graphite_dbresults(self.jobid, 'get_topology')
def get_markers(self):
return self.ic.graphite_dbresults(self.jobid, 'get_markers')
def get_events(self):
return self.ic.graphite_dbresults(self.jobid, 'get_events')
| import sniper_stats, intelqueue, iqclient
class SniperStatsJobid(sniper_stats.SniperStatsBase):
def __init__(self, jobid):
self.jobid = jobid
self.ic = iqclient.IntelClient()
def read_metricnames(self):
return self.ic.graphite_dbresults(self.jobid, 'read_metricnames')
def get_snapshots(self):
return self.ic.graphite_dbresults(self.jobid, 'get_snapshots')
def read_snapshot(self, prefix, metrics = None):
return self.ic.graphite_dbresults(self.jobid, 'read_snapshot', {'prefix': prefix, 'metrics': metrics})
def get_topology(self):
return self.ic.graphite_dbresults(self.jobid, 'get_topology')
def get_markers(self):
return self.ic.graphite_dbresults(self.jobid, 'get_markers')
def get_events(self):
return self.ic.graphite_dbresults(self.jobid, 'get_events')
[sniper_stats] Read metric names on startup for jobid-based stats so self.names is available as expectedimport sniper_stats, intelqueue, iqclient
class SniperStatsJobid(sniper_stats.SniperStatsBase):
def __init__(self, jobid):
self.jobid = jobid
self.ic = iqclient.IntelClient()
self.names = self.read_metricnames()
def read_metricnames(self):
return self.ic.graphite_dbresults(self.jobid, 'read_metricnames')
def get_snapshots(self):
return self.ic.graphite_dbresults(self.jobid, 'get_snapshots')
def read_snapshot(self, prefix, metrics = None):
return self.ic.graphite_dbresults(self.jobid, 'read_snapshot', {'prefix': prefix, 'metrics': metrics})
def get_topology(self):
return self.ic.graphite_dbresults(self.jobid, 'get_topology')
def get_markers(self):
return self.ic.graphite_dbresults(self.jobid, 'get_markers')
def get_events(self):
return self.ic.graphite_dbresults(self.jobid, 'get_events')
| <commit_before>import sniper_stats, intelqueue, iqclient
class SniperStatsJobid(sniper_stats.SniperStatsBase):
def __init__(self, jobid):
self.jobid = jobid
self.ic = iqclient.IntelClient()
def read_metricnames(self):
return self.ic.graphite_dbresults(self.jobid, 'read_metricnames')
def get_snapshots(self):
return self.ic.graphite_dbresults(self.jobid, 'get_snapshots')
def read_snapshot(self, prefix, metrics = None):
return self.ic.graphite_dbresults(self.jobid, 'read_snapshot', {'prefix': prefix, 'metrics': metrics})
def get_topology(self):
return self.ic.graphite_dbresults(self.jobid, 'get_topology')
def get_markers(self):
return self.ic.graphite_dbresults(self.jobid, 'get_markers')
def get_events(self):
return self.ic.graphite_dbresults(self.jobid, 'get_events')
<commit_msg>[sniper_stats] Read metric names on startup for jobid-based stats so self.names is available as expected<commit_after>import sniper_stats, intelqueue, iqclient
class SniperStatsJobid(sniper_stats.SniperStatsBase):
def __init__(self, jobid):
self.jobid = jobid
self.ic = iqclient.IntelClient()
self.names = self.read_metricnames()
def read_metricnames(self):
return self.ic.graphite_dbresults(self.jobid, 'read_metricnames')
def get_snapshots(self):
return self.ic.graphite_dbresults(self.jobid, 'get_snapshots')
def read_snapshot(self, prefix, metrics = None):
return self.ic.graphite_dbresults(self.jobid, 'read_snapshot', {'prefix': prefix, 'metrics': metrics})
def get_topology(self):
return self.ic.graphite_dbresults(self.jobid, 'get_topology')
def get_markers(self):
return self.ic.graphite_dbresults(self.jobid, 'get_markers')
def get_events(self):
return self.ic.graphite_dbresults(self.jobid, 'get_events')
|
2a17b9fdb55806d6397f506066a2a7e8c480020b | pylinks/main/tests.py | pylinks/main/tests.py | from django.test import TestCase
from analytics.models import GoogleAnalytics
class GoogleAnalyticsTests(TestCase):
def test_ga_tracking(self):
GoogleAnalytics(site_id=1, web_property_id='12345').save()
response = self.client.get('/')
self.assertContains(response, "_gaq.push(['_setAccount', '12345']);")
| from django.test import TestCase
from analytics.models import GoogleAnalytics
class GoogleAnalyticsTests(TestCase):
def test_ga_tracking(self):
GoogleAnalytics(site_id=1, web_property_id='12345').save()
response = self.client.get('/')
self.assertContains(response, "_gaq.push(['_setAccount', '12345']);")
class AdminTests(TestCase):
def test_admin_login_loads(self):
self.assertEqual(self.client.get('/admin/login/').status_code, 200)
| Add simple admin test just so we catch breakage early | Add simple admin test just so we catch breakage early
| Python | mit | michaelmior/pylinks,michaelmior/pylinks,michaelmior/pylinks | from django.test import TestCase
from analytics.models import GoogleAnalytics
class GoogleAnalyticsTests(TestCase):
def test_ga_tracking(self):
GoogleAnalytics(site_id=1, web_property_id='12345').save()
response = self.client.get('/')
self.assertContains(response, "_gaq.push(['_setAccount', '12345']);")
Add simple admin test just so we catch breakage early | from django.test import TestCase
from analytics.models import GoogleAnalytics
class GoogleAnalyticsTests(TestCase):
def test_ga_tracking(self):
GoogleAnalytics(site_id=1, web_property_id='12345').save()
response = self.client.get('/')
self.assertContains(response, "_gaq.push(['_setAccount', '12345']);")
class AdminTests(TestCase):
def test_admin_login_loads(self):
self.assertEqual(self.client.get('/admin/login/').status_code, 200)
| <commit_before>from django.test import TestCase
from analytics.models import GoogleAnalytics
class GoogleAnalyticsTests(TestCase):
def test_ga_tracking(self):
GoogleAnalytics(site_id=1, web_property_id='12345').save()
response = self.client.get('/')
self.assertContains(response, "_gaq.push(['_setAccount', '12345']);")
<commit_msg>Add simple admin test just so we catch breakage early<commit_after> | from django.test import TestCase
from analytics.models import GoogleAnalytics
class GoogleAnalyticsTests(TestCase):
def test_ga_tracking(self):
GoogleAnalytics(site_id=1, web_property_id='12345').save()
response = self.client.get('/')
self.assertContains(response, "_gaq.push(['_setAccount', '12345']);")
class AdminTests(TestCase):
def test_admin_login_loads(self):
self.assertEqual(self.client.get('/admin/login/').status_code, 200)
| from django.test import TestCase
from analytics.models import GoogleAnalytics
class GoogleAnalyticsTests(TestCase):
def test_ga_tracking(self):
GoogleAnalytics(site_id=1, web_property_id='12345').save()
response = self.client.get('/')
self.assertContains(response, "_gaq.push(['_setAccount', '12345']);")
Add simple admin test just so we catch breakage earlyfrom django.test import TestCase
from analytics.models import GoogleAnalytics
class GoogleAnalyticsTests(TestCase):
def test_ga_tracking(self):
GoogleAnalytics(site_id=1, web_property_id='12345').save()
response = self.client.get('/')
self.assertContains(response, "_gaq.push(['_setAccount', '12345']);")
class AdminTests(TestCase):
def test_admin_login_loads(self):
self.assertEqual(self.client.get('/admin/login/').status_code, 200)
| <commit_before>from django.test import TestCase
from analytics.models import GoogleAnalytics
class GoogleAnalyticsTests(TestCase):
def test_ga_tracking(self):
GoogleAnalytics(site_id=1, web_property_id='12345').save()
response = self.client.get('/')
self.assertContains(response, "_gaq.push(['_setAccount', '12345']);")
<commit_msg>Add simple admin test just so we catch breakage early<commit_after>from django.test import TestCase
from analytics.models import GoogleAnalytics
class GoogleAnalyticsTests(TestCase):
def test_ga_tracking(self):
GoogleAnalytics(site_id=1, web_property_id='12345').save()
response = self.client.get('/')
self.assertContains(response, "_gaq.push(['_setAccount', '12345']);")
class AdminTests(TestCase):
def test_admin_login_loads(self):
self.assertEqual(self.client.get('/admin/login/').status_code, 200)
|
36e8b7f7dd4de93c61f49d65106f2a0410945e2d | pyoracc/model/line.py | pyoracc/model/line.py | from mako.template import Template
class Line(object):
template = Template("""${label}. \\
% for word in words:
${word} \\
% endfor
% if lemmas:
\n#lem: \\
% for lemma in lemmas:
${lemma}; \\
% endfor \n
%endif
""", output_encoding='utf-8')
def __init__(self, label):
self.label = label
self.words = []
self.lemmas = []
self.witnesses = []
self.translation = None
self.notes = []
self.references = []
self.links = []
def __str__(self):
return self.template.render_unicode(**vars(self))
def serialize(self):
return self.template.render_unicode(**vars(self))
| from mako.template import Template
class Line(object):
template = Template("""\n${label}.\t\\
${' '.join(words)}\\
% if references:
% for reference in references:
^${reference}^
% endfor
% endif
% if lemmas:
\n#lem:\\
${'; '.join(lemmas)}\\
% endif
% if notes:
\n
% for note in notes:
${note.serialize()}
% endfor
% endif
% if links:
\n#link: \\
% for link in links:
${link};
% endfor
% endif
""", output_encoding='utf-8')
def __init__(self, label):
self.label = label
self.words = []
self.lemmas = []
self.witnesses = []
self.translation = None
self.notes = []
self.references = []
self.links = []
def __str__(self):
return self.template.render_unicode(**vars(self))
def serialize(self):
return self.template.render_unicode(**vars(self))
| Use join for serializing words and lemmas to avoid printing last ; as required by ATF format. Print also references, notes and links. | Use join for serializing words and lemmas to avoid printing last ; as required by ATF format. Print also references, notes and links.
| Python | mit | UCL/pyoracc | from mako.template import Template
class Line(object):
template = Template("""${label}. \\
% for word in words:
${word} \\
% endfor
% if lemmas:
\n#lem: \\
% for lemma in lemmas:
${lemma}; \\
% endfor \n
%endif
""", output_encoding='utf-8')
def __init__(self, label):
self.label = label
self.words = []
self.lemmas = []
self.witnesses = []
self.translation = None
self.notes = []
self.references = []
self.links = []
def __str__(self):
return self.template.render_unicode(**vars(self))
def serialize(self):
return self.template.render_unicode(**vars(self))
Use join for serializing words and lemmas to avoid printing last ; as required by ATF format. Print also references, notes and links. | from mako.template import Template
class Line(object):
template = Template("""\n${label}.\t\\
${' '.join(words)}\\
% if references:
% for reference in references:
^${reference}^
% endfor
% endif
% if lemmas:
\n#lem:\\
${'; '.join(lemmas)}\\
% endif
% if notes:
\n
% for note in notes:
${note.serialize()}
% endfor
% endif
% if links:
\n#link: \\
% for link in links:
${link};
% endfor
% endif
""", output_encoding='utf-8')
def __init__(self, label):
self.label = label
self.words = []
self.lemmas = []
self.witnesses = []
self.translation = None
self.notes = []
self.references = []
self.links = []
def __str__(self):
return self.template.render_unicode(**vars(self))
def serialize(self):
return self.template.render_unicode(**vars(self))
| <commit_before>from mako.template import Template
class Line(object):
template = Template("""${label}. \\
% for word in words:
${word} \\
% endfor
% if lemmas:
\n#lem: \\
% for lemma in lemmas:
${lemma}; \\
% endfor \n
%endif
""", output_encoding='utf-8')
def __init__(self, label):
self.label = label
self.words = []
self.lemmas = []
self.witnesses = []
self.translation = None
self.notes = []
self.references = []
self.links = []
def __str__(self):
return self.template.render_unicode(**vars(self))
def serialize(self):
return self.template.render_unicode(**vars(self))
<commit_msg>Use join for serializing words and lemmas to avoid printing last ; as required by ATF format. Print also references, notes and links.<commit_after> | from mako.template import Template
class Line(object):
template = Template("""\n${label}.\t\\
${' '.join(words)}\\
% if references:
% for reference in references:
^${reference}^
% endfor
% endif
% if lemmas:
\n#lem:\\
${'; '.join(lemmas)}\\
% endif
% if notes:
\n
% for note in notes:
${note.serialize()}
% endfor
% endif
% if links:
\n#link: \\
% for link in links:
${link};
% endfor
% endif
""", output_encoding='utf-8')
def __init__(self, label):
self.label = label
self.words = []
self.lemmas = []
self.witnesses = []
self.translation = None
self.notes = []
self.references = []
self.links = []
def __str__(self):
return self.template.render_unicode(**vars(self))
def serialize(self):
return self.template.render_unicode(**vars(self))
| from mako.template import Template
class Line(object):
template = Template("""${label}. \\
% for word in words:
${word} \\
% endfor
% if lemmas:
\n#lem: \\
% for lemma in lemmas:
${lemma}; \\
% endfor \n
%endif
""", output_encoding='utf-8')
def __init__(self, label):
self.label = label
self.words = []
self.lemmas = []
self.witnesses = []
self.translation = None
self.notes = []
self.references = []
self.links = []
def __str__(self):
return self.template.render_unicode(**vars(self))
def serialize(self):
return self.template.render_unicode(**vars(self))
Use join for serializing words and lemmas to avoid printing last ; as required by ATF format. Print also references, notes and links.from mako.template import Template
class Line(object):
template = Template("""\n${label}.\t\\
${' '.join(words)}\\
% if references:
% for reference in references:
^${reference}^
% endfor
% endif
% if lemmas:
\n#lem:\\
${'; '.join(lemmas)}\\
% endif
% if notes:
\n
% for note in notes:
${note.serialize()}
% endfor
% endif
% if links:
\n#link: \\
% for link in links:
${link};
% endfor
% endif
""", output_encoding='utf-8')
def __init__(self, label):
self.label = label
self.words = []
self.lemmas = []
self.witnesses = []
self.translation = None
self.notes = []
self.references = []
self.links = []
def __str__(self):
return self.template.render_unicode(**vars(self))
def serialize(self):
return self.template.render_unicode(**vars(self))
| <commit_before>from mako.template import Template
class Line(object):
template = Template("""${label}. \\
% for word in words:
${word} \\
% endfor
% if lemmas:
\n#lem: \\
% for lemma in lemmas:
${lemma}; \\
% endfor \n
%endif
""", output_encoding='utf-8')
def __init__(self, label):
self.label = label
self.words = []
self.lemmas = []
self.witnesses = []
self.translation = None
self.notes = []
self.references = []
self.links = []
def __str__(self):
return self.template.render_unicode(**vars(self))
def serialize(self):
return self.template.render_unicode(**vars(self))
<commit_msg>Use join for serializing words and lemmas to avoid printing last ; as required by ATF format. Print also references, notes and links.<commit_after>from mako.template import Template
class Line(object):
template = Template("""\n${label}.\t\\
${' '.join(words)}\\
% if references:
% for reference in references:
^${reference}^
% endfor
% endif
% if lemmas:
\n#lem:\\
${'; '.join(lemmas)}\\
% endif
% if notes:
\n
% for note in notes:
${note.serialize()}
% endfor
% endif
% if links:
\n#link: \\
% for link in links:
${link};
% endfor
% endif
""", output_encoding='utf-8')
def __init__(self, label):
self.label = label
self.words = []
self.lemmas = []
self.witnesses = []
self.translation = None
self.notes = []
self.references = []
self.links = []
def __str__(self):
return self.template.render_unicode(**vars(self))
def serialize(self):
return self.template.render_unicode(**vars(self))
|
87b1d823f09a20547b08f769636bfc7bcc7f0766 | setup.py | setup.py | from setuptools import setup
setup(
name='rhino',
version='1.0',
packages=['rhino'],
include_package_data=True,
#requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup(<3.1)', 'jinja2'],
#install_requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup<3.1', 'jinja2'],
requires=['Django', 'south', 'jinja2', 'oauth2'],
)
| from setuptools import setup
setup(
name='rhino',
version='1.0',
packages=['rhino'],
include_package_data=True,
#requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup(<3.1)', 'jinja2'],
#install_requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup<3.1', 'jinja2'],
requires=['Django', 'south', 'jinja2', 'oauth2', 'pytidylib', 'remoteobjects'],
)
| Call out these requirements too | Call out these requirements too
| Python | mit | markpasc/leapfrog,markpasc/leapfrog | from setuptools import setup
setup(
name='rhino',
version='1.0',
packages=['rhino'],
include_package_data=True,
#requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup(<3.1)', 'jinja2'],
#install_requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup<3.1', 'jinja2'],
requires=['Django', 'south', 'jinja2', 'oauth2'],
)
Call out these requirements too | from setuptools import setup
setup(
name='rhino',
version='1.0',
packages=['rhino'],
include_package_data=True,
#requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup(<3.1)', 'jinja2'],
#install_requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup<3.1', 'jinja2'],
requires=['Django', 'south', 'jinja2', 'oauth2', 'pytidylib', 'remoteobjects'],
)
| <commit_before>from setuptools import setup
setup(
name='rhino',
version='1.0',
packages=['rhino'],
include_package_data=True,
#requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup(<3.1)', 'jinja2'],
#install_requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup<3.1', 'jinja2'],
requires=['Django', 'south', 'jinja2', 'oauth2'],
)
<commit_msg>Call out these requirements too<commit_after> | from setuptools import setup
setup(
name='rhino',
version='1.0',
packages=['rhino'],
include_package_data=True,
#requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup(<3.1)', 'jinja2'],
#install_requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup<3.1', 'jinja2'],
requires=['Django', 'south', 'jinja2', 'oauth2', 'pytidylib', 'remoteobjects'],
)
| from setuptools import setup
setup(
name='rhino',
version='1.0',
packages=['rhino'],
include_package_data=True,
#requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup(<3.1)', 'jinja2'],
#install_requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup<3.1', 'jinja2'],
requires=['Django', 'south', 'jinja2', 'oauth2'],
)
Call out these requirements toofrom setuptools import setup
setup(
name='rhino',
version='1.0',
packages=['rhino'],
include_package_data=True,
#requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup(<3.1)', 'jinja2'],
#install_requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup<3.1', 'jinja2'],
requires=['Django', 'south', 'jinja2', 'oauth2', 'pytidylib', 'remoteobjects'],
)
| <commit_before>from setuptools import setup
setup(
name='rhino',
version='1.0',
packages=['rhino'],
include_package_data=True,
#requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup(<3.1)', 'jinja2'],
#install_requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup<3.1', 'jinja2'],
requires=['Django', 'south', 'jinja2', 'oauth2'],
)
<commit_msg>Call out these requirements too<commit_after>from setuptools import setup
setup(
name='rhino',
version='1.0',
packages=['rhino'],
include_package_data=True,
#requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup(<3.1)', 'jinja2'],
#install_requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup<3.1', 'jinja2'],
requires=['Django', 'south', 'jinja2', 'oauth2', 'pytidylib', 'remoteobjects'],
)
|
58701c0d750714f8ded53627b0f8c22f256376c6 | setup.py | setup.py | #!/usr/bin/env python
# Copyright (c) 2013 Soren Hansen
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from setuptools import setup, find_packages
setup(
name='basicdb',
version='0.1',
description='Basic database service',
long_description=open('README.rst', 'r').read(),
author='Soren Hansen',
author_email='soren@linux2go.dk',
url='http://github.com/sorenh/basicdb',
packages=find_packages(),
include_package_data=True,
license='Apache 2.0',
keywords='basicdb simpledb')
| #!/usr/bin/env python
# Copyright (c) 2013 Soren Hansen
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from setuptools import setup, find_packages
import pkg_resources
setup(
name='basicdb',
version='0.1',
description='Basic database service',
long_description=pkg_resources.resource_string(__name__, "README.rst"),
author='Soren Hansen',
author_email='soren@linux2go.dk',
url='http://github.com/sorenh/basicdb',
packages=find_packages(),
include_package_data=True,
license='Apache 2.0',
keywords='basicdb simpledb')
| Use pkg_resources to read README.rst | Use pkg_resources to read README.rst
| Python | apache-2.0 | JioCloud/basicdb,varunarya10/basicdb,sorenh/basicdb | #!/usr/bin/env python
# Copyright (c) 2013 Soren Hansen
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from setuptools import setup, find_packages
setup(
name='basicdb',
version='0.1',
description='Basic database service',
long_description=open('README.rst', 'r').read(),
author='Soren Hansen',
author_email='soren@linux2go.dk',
url='http://github.com/sorenh/basicdb',
packages=find_packages(),
include_package_data=True,
license='Apache 2.0',
keywords='basicdb simpledb')
Use pkg_resources to read README.rst | #!/usr/bin/env python
# Copyright (c) 2013 Soren Hansen
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from setuptools import setup, find_packages
import pkg_resources
setup(
name='basicdb',
version='0.1',
description='Basic database service',
long_description=pkg_resources.resource_string(__name__, "README.rst"),
author='Soren Hansen',
author_email='soren@linux2go.dk',
url='http://github.com/sorenh/basicdb',
packages=find_packages(),
include_package_data=True,
license='Apache 2.0',
keywords='basicdb simpledb')
| <commit_before>#!/usr/bin/env python
# Copyright (c) 2013 Soren Hansen
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from setuptools import setup, find_packages
setup(
name='basicdb',
version='0.1',
description='Basic database service',
long_description=open('README.rst', 'r').read(),
author='Soren Hansen',
author_email='soren@linux2go.dk',
url='http://github.com/sorenh/basicdb',
packages=find_packages(),
include_package_data=True,
license='Apache 2.0',
keywords='basicdb simpledb')
<commit_msg>Use pkg_resources to read README.rst<commit_after> | #!/usr/bin/env python
# Copyright (c) 2013 Soren Hansen
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from setuptools import setup, find_packages
import pkg_resources
setup(
name='basicdb',
version='0.1',
description='Basic database service',
long_description=pkg_resources.resource_string(__name__, "README.rst"),
author='Soren Hansen',
author_email='soren@linux2go.dk',
url='http://github.com/sorenh/basicdb',
packages=find_packages(),
include_package_data=True,
license='Apache 2.0',
keywords='basicdb simpledb')
| #!/usr/bin/env python
# Copyright (c) 2013 Soren Hansen
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from setuptools import setup, find_packages
setup(
name='basicdb',
version='0.1',
description='Basic database service',
long_description=open('README.rst', 'r').read(),
author='Soren Hansen',
author_email='soren@linux2go.dk',
url='http://github.com/sorenh/basicdb',
packages=find_packages(),
include_package_data=True,
license='Apache 2.0',
keywords='basicdb simpledb')
Use pkg_resources to read README.rst#!/usr/bin/env python
# Copyright (c) 2013 Soren Hansen
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from setuptools import setup, find_packages
import pkg_resources
setup(
name='basicdb',
version='0.1',
description='Basic database service',
long_description=pkg_resources.resource_string(__name__, "README.rst"),
author='Soren Hansen',
author_email='soren@linux2go.dk',
url='http://github.com/sorenh/basicdb',
packages=find_packages(),
include_package_data=True,
license='Apache 2.0',
keywords='basicdb simpledb')
| <commit_before>#!/usr/bin/env python
# Copyright (c) 2013 Soren Hansen
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from setuptools import setup, find_packages
setup(
name='basicdb',
version='0.1',
description='Basic database service',
long_description=open('README.rst', 'r').read(),
author='Soren Hansen',
author_email='soren@linux2go.dk',
url='http://github.com/sorenh/basicdb',
packages=find_packages(),
include_package_data=True,
license='Apache 2.0',
keywords='basicdb simpledb')
<commit_msg>Use pkg_resources to read README.rst<commit_after>#!/usr/bin/env python
# Copyright (c) 2013 Soren Hansen
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from setuptools import setup, find_packages
import pkg_resources
setup(
name='basicdb',
version='0.1',
description='Basic database service',
long_description=pkg_resources.resource_string(__name__, "README.rst"),
author='Soren Hansen',
author_email='soren@linux2go.dk',
url='http://github.com/sorenh/basicdb',
packages=find_packages(),
include_package_data=True,
license='Apache 2.0',
keywords='basicdb simpledb')
|
7d8e91ce410bf1add9a21777afc0517198c11ced | setup.py | setup.py | from setuptools import setup, find_packages
setup(
name = "OvhApi",
version = "1.0",
description='Python module to manage Ovh API',
author='Guilhem Lettron',
author_email='guilhem.lettron@optiflows.com',
url='https://github.com/optiflows/OvhApi',
packages = find_packages(),
)
| from setuptools import setup, find_packages
setup(
name = "ovhapi",
version = "1.0",
description='Python module to manage Ovh API',
maintainer='Guilhem Lettron',
maintainer_email='guilhem.lettron@optiflows.com',
url='https://github.com/optiflows/OvhApi',
packages = find_packages(),
)
| Change Name to ovhapi Set me as maintainer (not author) | Change Name to ovhapi
Set me as maintainer (not author)
| Python | bsd-3-clause | optiflows/ovhapi | from setuptools import setup, find_packages
setup(
name = "OvhApi",
version = "1.0",
description='Python module to manage Ovh API',
author='Guilhem Lettron',
author_email='guilhem.lettron@optiflows.com',
url='https://github.com/optiflows/OvhApi',
packages = find_packages(),
)
Change Name to ovhapi
Set me as maintainer (not author) | from setuptools import setup, find_packages
setup(
name = "ovhapi",
version = "1.0",
description='Python module to manage Ovh API',
maintainer='Guilhem Lettron',
maintainer_email='guilhem.lettron@optiflows.com',
url='https://github.com/optiflows/OvhApi',
packages = find_packages(),
)
| <commit_before>from setuptools import setup, find_packages
setup(
name = "OvhApi",
version = "1.0",
description='Python module to manage Ovh API',
author='Guilhem Lettron',
author_email='guilhem.lettron@optiflows.com',
url='https://github.com/optiflows/OvhApi',
packages = find_packages(),
)
<commit_msg>Change Name to ovhapi
Set me as maintainer (not author)<commit_after> | from setuptools import setup, find_packages
setup(
name = "ovhapi",
version = "1.0",
description='Python module to manage Ovh API',
maintainer='Guilhem Lettron',
maintainer_email='guilhem.lettron@optiflows.com',
url='https://github.com/optiflows/OvhApi',
packages = find_packages(),
)
| from setuptools import setup, find_packages
setup(
name = "OvhApi",
version = "1.0",
description='Python module to manage Ovh API',
author='Guilhem Lettron',
author_email='guilhem.lettron@optiflows.com',
url='https://github.com/optiflows/OvhApi',
packages = find_packages(),
)
Change Name to ovhapi
Set me as maintainer (not author)from setuptools import setup, find_packages
setup(
name = "ovhapi",
version = "1.0",
description='Python module to manage Ovh API',
maintainer='Guilhem Lettron',
maintainer_email='guilhem.lettron@optiflows.com',
url='https://github.com/optiflows/OvhApi',
packages = find_packages(),
)
| <commit_before>from setuptools import setup, find_packages
setup(
name = "OvhApi",
version = "1.0",
description='Python module to manage Ovh API',
author='Guilhem Lettron',
author_email='guilhem.lettron@optiflows.com',
url='https://github.com/optiflows/OvhApi',
packages = find_packages(),
)
<commit_msg>Change Name to ovhapi
Set me as maintainer (not author)<commit_after>from setuptools import setup, find_packages
setup(
name = "ovhapi",
version = "1.0",
description='Python module to manage Ovh API',
maintainer='Guilhem Lettron',
maintainer_email='guilhem.lettron@optiflows.com',
url='https://github.com/optiflows/OvhApi',
packages = find_packages(),
)
|
360a8c395373bcab0b725b1ac8f8dfd581d6e2b9 | setup.py | setup.py | from setuptools import setup, find_packages
setup(
name='stop-words',
version=__import__("stop_words").get_version(),
description='Get list of common stop words in various languages in Python',
long_description=open('README.rst').read(),
license=open('LICENSE').read(),
author='Alireza Savand',
author_email='alireza.savand@gmail.com',
url='https://github.com/Alir3z4/python-stop-words',
packages=find_packages(),
zip_safe=False,
package_data={
'stop_words': [
'stop-words/*.txt',
'stop-words/languages.json',
]
},
classifiers=[
'Programming Language :: Python',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Topic :: Software Development',
'Development Status :: 6 - Mature',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.6',
'Programming Language :: Python :: 3.7',
'Topic :: Text Processing',
'Topic :: Text Processing :: Filters',
'License :: OSI Approved :: BSD License',
],
)
| from setuptools import setup, find_packages
setup(
name='stop-words',
version=__import__("stop_words").get_version(),
description='Get list of common stop words in various languages in Python',
long_description=open('README.rst').read(),
license=open('LICENSE').read(),
author='Alireza Savand',
author_email='alireza.savand@gmail.com',
url='https://github.com/Alir3z4/python-stop-words',
packages=find_packages(),
zip_safe=False,
package_data={
'stop_words': [
'stop-words/*.txt',
'stop-words/languages.json',
]
},
classifiers=[
'Programming Language :: Python',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Topic :: Software Development',
'Development Status :: 6 - Mature',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: 3.6',
'Programming Language :: Python :: 3.7',
'Topic :: Text Processing',
'Topic :: Text Processing :: Filters',
'License :: OSI Approved :: BSD License',
],
)
| Add Python34 and Python35 to classifiers | Add Python34 and Python35 to classifiers
| Python | bsd-3-clause | Alir3z4/python-stop-words | from setuptools import setup, find_packages
setup(
name='stop-words',
version=__import__("stop_words").get_version(),
description='Get list of common stop words in various languages in Python',
long_description=open('README.rst').read(),
license=open('LICENSE').read(),
author='Alireza Savand',
author_email='alireza.savand@gmail.com',
url='https://github.com/Alir3z4/python-stop-words',
packages=find_packages(),
zip_safe=False,
package_data={
'stop_words': [
'stop-words/*.txt',
'stop-words/languages.json',
]
},
classifiers=[
'Programming Language :: Python',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Topic :: Software Development',
'Development Status :: 6 - Mature',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.6',
'Programming Language :: Python :: 3.7',
'Topic :: Text Processing',
'Topic :: Text Processing :: Filters',
'License :: OSI Approved :: BSD License',
],
)
Add Python34 and Python35 to classifiers | from setuptools import setup, find_packages
setup(
name='stop-words',
version=__import__("stop_words").get_version(),
description='Get list of common stop words in various languages in Python',
long_description=open('README.rst').read(),
license=open('LICENSE').read(),
author='Alireza Savand',
author_email='alireza.savand@gmail.com',
url='https://github.com/Alir3z4/python-stop-words',
packages=find_packages(),
zip_safe=False,
package_data={
'stop_words': [
'stop-words/*.txt',
'stop-words/languages.json',
]
},
classifiers=[
'Programming Language :: Python',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Topic :: Software Development',
'Development Status :: 6 - Mature',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: 3.6',
'Programming Language :: Python :: 3.7',
'Topic :: Text Processing',
'Topic :: Text Processing :: Filters',
'License :: OSI Approved :: BSD License',
],
)
| <commit_before>from setuptools import setup, find_packages
setup(
name='stop-words',
version=__import__("stop_words").get_version(),
description='Get list of common stop words in various languages in Python',
long_description=open('README.rst').read(),
license=open('LICENSE').read(),
author='Alireza Savand',
author_email='alireza.savand@gmail.com',
url='https://github.com/Alir3z4/python-stop-words',
packages=find_packages(),
zip_safe=False,
package_data={
'stop_words': [
'stop-words/*.txt',
'stop-words/languages.json',
]
},
classifiers=[
'Programming Language :: Python',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Topic :: Software Development',
'Development Status :: 6 - Mature',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.6',
'Programming Language :: Python :: 3.7',
'Topic :: Text Processing',
'Topic :: Text Processing :: Filters',
'License :: OSI Approved :: BSD License',
],
)
<commit_msg>Add Python34 and Python35 to classifiers<commit_after> | from setuptools import setup, find_packages
setup(
name='stop-words',
version=__import__("stop_words").get_version(),
description='Get list of common stop words in various languages in Python',
long_description=open('README.rst').read(),
license=open('LICENSE').read(),
author='Alireza Savand',
author_email='alireza.savand@gmail.com',
url='https://github.com/Alir3z4/python-stop-words',
packages=find_packages(),
zip_safe=False,
package_data={
'stop_words': [
'stop-words/*.txt',
'stop-words/languages.json',
]
},
classifiers=[
'Programming Language :: Python',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Topic :: Software Development',
'Development Status :: 6 - Mature',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: 3.6',
'Programming Language :: Python :: 3.7',
'Topic :: Text Processing',
'Topic :: Text Processing :: Filters',
'License :: OSI Approved :: BSD License',
],
)
| from setuptools import setup, find_packages
setup(
name='stop-words',
version=__import__("stop_words").get_version(),
description='Get list of common stop words in various languages in Python',
long_description=open('README.rst').read(),
license=open('LICENSE').read(),
author='Alireza Savand',
author_email='alireza.savand@gmail.com',
url='https://github.com/Alir3z4/python-stop-words',
packages=find_packages(),
zip_safe=False,
package_data={
'stop_words': [
'stop-words/*.txt',
'stop-words/languages.json',
]
},
classifiers=[
'Programming Language :: Python',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Topic :: Software Development',
'Development Status :: 6 - Mature',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.6',
'Programming Language :: Python :: 3.7',
'Topic :: Text Processing',
'Topic :: Text Processing :: Filters',
'License :: OSI Approved :: BSD License',
],
)
Add Python34 and Python35 to classifiersfrom setuptools import setup, find_packages
setup(
name='stop-words',
version=__import__("stop_words").get_version(),
description='Get list of common stop words in various languages in Python',
long_description=open('README.rst').read(),
license=open('LICENSE').read(),
author='Alireza Savand',
author_email='alireza.savand@gmail.com',
url='https://github.com/Alir3z4/python-stop-words',
packages=find_packages(),
zip_safe=False,
package_data={
'stop_words': [
'stop-words/*.txt',
'stop-words/languages.json',
]
},
classifiers=[
'Programming Language :: Python',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Topic :: Software Development',
'Development Status :: 6 - Mature',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: 3.6',
'Programming Language :: Python :: 3.7',
'Topic :: Text Processing',
'Topic :: Text Processing :: Filters',
'License :: OSI Approved :: BSD License',
],
)
| <commit_before>from setuptools import setup, find_packages
setup(
name='stop-words',
version=__import__("stop_words").get_version(),
description='Get list of common stop words in various languages in Python',
long_description=open('README.rst').read(),
license=open('LICENSE').read(),
author='Alireza Savand',
author_email='alireza.savand@gmail.com',
url='https://github.com/Alir3z4/python-stop-words',
packages=find_packages(),
zip_safe=False,
package_data={
'stop_words': [
'stop-words/*.txt',
'stop-words/languages.json',
]
},
classifiers=[
'Programming Language :: Python',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Topic :: Software Development',
'Development Status :: 6 - Mature',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.6',
'Programming Language :: Python :: 3.7',
'Topic :: Text Processing',
'Topic :: Text Processing :: Filters',
'License :: OSI Approved :: BSD License',
],
)
<commit_msg>Add Python34 and Python35 to classifiers<commit_after>from setuptools import setup, find_packages
setup(
name='stop-words',
version=__import__("stop_words").get_version(),
description='Get list of common stop words in various languages in Python',
long_description=open('README.rst').read(),
license=open('LICENSE').read(),
author='Alireza Savand',
author_email='alireza.savand@gmail.com',
url='https://github.com/Alir3z4/python-stop-words',
packages=find_packages(),
zip_safe=False,
package_data={
'stop_words': [
'stop-words/*.txt',
'stop-words/languages.json',
]
},
classifiers=[
'Programming Language :: Python',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Topic :: Software Development',
'Development Status :: 6 - Mature',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: 3.6',
'Programming Language :: Python :: 3.7',
'Topic :: Text Processing',
'Topic :: Text Processing :: Filters',
'License :: OSI Approved :: BSD License',
],
)
|
27434395a599d7e42d2295056396937d89bb53a6 | setup.py | setup.py | import sys
from setuptools import find_packages, setup
VERSION = '2.0.dev0'
install_requires = [
'django-local-settings>=1.0a10',
'stashward',
]
if sys.version_info[:2] < (3, 4):
install_requires.append('enum34')
setup(
name='django-arcutils',
version=VERSION,
url='https://github.com/PSU-OIT-ARC/django-arcutils',
author='PSU - OIT - ARC',
author_email='consultants@pdx.edu',
description='Common utilities used in ARC Django projects',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=install_requires,
extras_require={
'cas': [
'django-cas-client>=1.2.0',
],
'ldap': [
'certifi>=2015.11.20.1',
'ldap3>=1.0.3',
],
'dev': [
'django>=1.7,<1.9',
'flake8',
'ldap3',
'mock',
'model_mommy',
],
},
entry_points="""
[console_scripts]
arcutils = arcutils.__main__:main
""",
classifiers=[
'Development Status :: 3 - Alpha',
'Framework :: Django',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
],
)
| import sys
from setuptools import find_packages, setup
VERSION = '2.0.dev0'
install_requires = [
'django-local-settings>=1.0a10',
'stashward',
]
if sys.version_info[:2] < (3, 4):
install_requires.append('enum34')
setup(
name='django-arcutils',
version=VERSION,
url='https://github.com/PSU-OIT-ARC/django-arcutils',
author='PSU - OIT - ARC',
author_email='consultants@pdx.edu',
description='Common utilities used in ARC Django projects',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=install_requires,
extras_require={
'cas': [
'django-cas-client>=1.2.0',
],
'ldap': [
'certifi>=2015.11.20.1',
'ldap3>=1.0.3',
],
'dev': [
'django>=1.7,<1.9',
'djangorestframework>3.3',
'flake8',
'ldap3',
'mock',
'model_mommy',
],
},
entry_points="""
[console_scripts]
arcutils = arcutils.__main__:main
""",
classifiers=[
'Development Status :: 3 - Alpha',
'Framework :: Django',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
],
)
| Add DRF as a dev dependency | Add DRF as a dev dependency
| Python | mit | PSU-OIT-ARC/django-arcutils,wylee/django-arcutils,wylee/django-arcutils,PSU-OIT-ARC/django-arcutils | import sys
from setuptools import find_packages, setup
VERSION = '2.0.dev0'
install_requires = [
'django-local-settings>=1.0a10',
'stashward',
]
if sys.version_info[:2] < (3, 4):
install_requires.append('enum34')
setup(
name='django-arcutils',
version=VERSION,
url='https://github.com/PSU-OIT-ARC/django-arcutils',
author='PSU - OIT - ARC',
author_email='consultants@pdx.edu',
description='Common utilities used in ARC Django projects',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=install_requires,
extras_require={
'cas': [
'django-cas-client>=1.2.0',
],
'ldap': [
'certifi>=2015.11.20.1',
'ldap3>=1.0.3',
],
'dev': [
'django>=1.7,<1.9',
'flake8',
'ldap3',
'mock',
'model_mommy',
],
},
entry_points="""
[console_scripts]
arcutils = arcutils.__main__:main
""",
classifiers=[
'Development Status :: 3 - Alpha',
'Framework :: Django',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
],
)
Add DRF as a dev dependency | import sys
from setuptools import find_packages, setup
VERSION = '2.0.dev0'
install_requires = [
'django-local-settings>=1.0a10',
'stashward',
]
if sys.version_info[:2] < (3, 4):
install_requires.append('enum34')
setup(
name='django-arcutils',
version=VERSION,
url='https://github.com/PSU-OIT-ARC/django-arcutils',
author='PSU - OIT - ARC',
author_email='consultants@pdx.edu',
description='Common utilities used in ARC Django projects',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=install_requires,
extras_require={
'cas': [
'django-cas-client>=1.2.0',
],
'ldap': [
'certifi>=2015.11.20.1',
'ldap3>=1.0.3',
],
'dev': [
'django>=1.7,<1.9',
'djangorestframework>3.3',
'flake8',
'ldap3',
'mock',
'model_mommy',
],
},
entry_points="""
[console_scripts]
arcutils = arcutils.__main__:main
""",
classifiers=[
'Development Status :: 3 - Alpha',
'Framework :: Django',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
],
)
| <commit_before>import sys
from setuptools import find_packages, setup
VERSION = '2.0.dev0'
install_requires = [
'django-local-settings>=1.0a10',
'stashward',
]
if sys.version_info[:2] < (3, 4):
install_requires.append('enum34')
setup(
name='django-arcutils',
version=VERSION,
url='https://github.com/PSU-OIT-ARC/django-arcutils',
author='PSU - OIT - ARC',
author_email='consultants@pdx.edu',
description='Common utilities used in ARC Django projects',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=install_requires,
extras_require={
'cas': [
'django-cas-client>=1.2.0',
],
'ldap': [
'certifi>=2015.11.20.1',
'ldap3>=1.0.3',
],
'dev': [
'django>=1.7,<1.9',
'flake8',
'ldap3',
'mock',
'model_mommy',
],
},
entry_points="""
[console_scripts]
arcutils = arcutils.__main__:main
""",
classifiers=[
'Development Status :: 3 - Alpha',
'Framework :: Django',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
],
)
<commit_msg>Add DRF as a dev dependency<commit_after> | import sys
from setuptools import find_packages, setup
VERSION = '2.0.dev0'
install_requires = [
'django-local-settings>=1.0a10',
'stashward',
]
if sys.version_info[:2] < (3, 4):
install_requires.append('enum34')
setup(
name='django-arcutils',
version=VERSION,
url='https://github.com/PSU-OIT-ARC/django-arcutils',
author='PSU - OIT - ARC',
author_email='consultants@pdx.edu',
description='Common utilities used in ARC Django projects',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=install_requires,
extras_require={
'cas': [
'django-cas-client>=1.2.0',
],
'ldap': [
'certifi>=2015.11.20.1',
'ldap3>=1.0.3',
],
'dev': [
'django>=1.7,<1.9',
'djangorestframework>3.3',
'flake8',
'ldap3',
'mock',
'model_mommy',
],
},
entry_points="""
[console_scripts]
arcutils = arcutils.__main__:main
""",
classifiers=[
'Development Status :: 3 - Alpha',
'Framework :: Django',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
],
)
| import sys
from setuptools import find_packages, setup
VERSION = '2.0.dev0'
install_requires = [
'django-local-settings>=1.0a10',
'stashward',
]
if sys.version_info[:2] < (3, 4):
install_requires.append('enum34')
setup(
name='django-arcutils',
version=VERSION,
url='https://github.com/PSU-OIT-ARC/django-arcutils',
author='PSU - OIT - ARC',
author_email='consultants@pdx.edu',
description='Common utilities used in ARC Django projects',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=install_requires,
extras_require={
'cas': [
'django-cas-client>=1.2.0',
],
'ldap': [
'certifi>=2015.11.20.1',
'ldap3>=1.0.3',
],
'dev': [
'django>=1.7,<1.9',
'flake8',
'ldap3',
'mock',
'model_mommy',
],
},
entry_points="""
[console_scripts]
arcutils = arcutils.__main__:main
""",
classifiers=[
'Development Status :: 3 - Alpha',
'Framework :: Django',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
],
)
Add DRF as a dev dependencyimport sys
from setuptools import find_packages, setup
VERSION = '2.0.dev0'
install_requires = [
'django-local-settings>=1.0a10',
'stashward',
]
if sys.version_info[:2] < (3, 4):
install_requires.append('enum34')
setup(
name='django-arcutils',
version=VERSION,
url='https://github.com/PSU-OIT-ARC/django-arcutils',
author='PSU - OIT - ARC',
author_email='consultants@pdx.edu',
description='Common utilities used in ARC Django projects',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=install_requires,
extras_require={
'cas': [
'django-cas-client>=1.2.0',
],
'ldap': [
'certifi>=2015.11.20.1',
'ldap3>=1.0.3',
],
'dev': [
'django>=1.7,<1.9',
'djangorestframework>3.3',
'flake8',
'ldap3',
'mock',
'model_mommy',
],
},
entry_points="""
[console_scripts]
arcutils = arcutils.__main__:main
""",
classifiers=[
'Development Status :: 3 - Alpha',
'Framework :: Django',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
],
)
| <commit_before>import sys
from setuptools import find_packages, setup
VERSION = '2.0.dev0'
install_requires = [
'django-local-settings>=1.0a10',
'stashward',
]
if sys.version_info[:2] < (3, 4):
install_requires.append('enum34')
setup(
name='django-arcutils',
version=VERSION,
url='https://github.com/PSU-OIT-ARC/django-arcutils',
author='PSU - OIT - ARC',
author_email='consultants@pdx.edu',
description='Common utilities used in ARC Django projects',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=install_requires,
extras_require={
'cas': [
'django-cas-client>=1.2.0',
],
'ldap': [
'certifi>=2015.11.20.1',
'ldap3>=1.0.3',
],
'dev': [
'django>=1.7,<1.9',
'flake8',
'ldap3',
'mock',
'model_mommy',
],
},
entry_points="""
[console_scripts]
arcutils = arcutils.__main__:main
""",
classifiers=[
'Development Status :: 3 - Alpha',
'Framework :: Django',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
],
)
<commit_msg>Add DRF as a dev dependency<commit_after>import sys
from setuptools import find_packages, setup
VERSION = '2.0.dev0'
install_requires = [
'django-local-settings>=1.0a10',
'stashward',
]
if sys.version_info[:2] < (3, 4):
install_requires.append('enum34')
setup(
name='django-arcutils',
version=VERSION,
url='https://github.com/PSU-OIT-ARC/django-arcutils',
author='PSU - OIT - ARC',
author_email='consultants@pdx.edu',
description='Common utilities used in ARC Django projects',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=install_requires,
extras_require={
'cas': [
'django-cas-client>=1.2.0',
],
'ldap': [
'certifi>=2015.11.20.1',
'ldap3>=1.0.3',
],
'dev': [
'django>=1.7,<1.9',
'djangorestframework>3.3',
'flake8',
'ldap3',
'mock',
'model_mommy',
],
},
entry_points="""
[console_scripts]
arcutils = arcutils.__main__:main
""",
classifiers=[
'Development Status :: 3 - Alpha',
'Framework :: Django',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
],
)
|
5392af2781d6a6a7c71c15ab7699feb3e3b8f2f2 | setup.py | setup.py | __author__ = 'katharine'
import os
import sys
from setuptools import setup, find_packages
requirements_path = os.path.join(os.path.dirname(__file__), 'requirements.txt')
with open(requirements_path) as requirements_file:
requirements = [line.strip() for line in requirements_file.readlines()]
setup(name='pypkjs',
version='3.6',
description='PebbleKit JS in Python!',
url='https://github.com/pebble/pypkjs',
author='Pebble Technology Corporation',
author_email='katharine@pebble.com',
license='MIT',
packages=find_packages(),
install_requires=requirements,
entry_points={
'console_scripts': [
'pypkjs=runner.websocket:run_tool'
],
},
zip_safe=False)
| __author__ = 'katharine'
import os
import sys
from setuptools import setup, find_packages
requirements_path = os.path.join(os.path.dirname(__file__), 'requirements.txt')
with open(requirements_path) as requirements_file:
requirements = [line.strip() for line in requirements_file.readlines()]
setup(name='pypkjs',
version='3.6',
description='PebbleKit JS in Python!',
url='https://github.com/pebble/pypkjs',
author='Pebble Technology Corporation',
author_email='katharine@pebble.com',
license='MIT',
packages=find_packages(),
install_requires=requirements,
package_data={
'javascript.navigator': 'GeoLiteCity.dat'
},
entry_points={
'console_scripts': [
'pypkjs=runner.websocket:run_tool'
],
},
zip_safe=False)
| Add package data for GeoIP DB | Add package data for GeoIP DB
| Python | mit | pebble/pypkjs | __author__ = 'katharine'
import os
import sys
from setuptools import setup, find_packages
requirements_path = os.path.join(os.path.dirname(__file__), 'requirements.txt')
with open(requirements_path) as requirements_file:
requirements = [line.strip() for line in requirements_file.readlines()]
setup(name='pypkjs',
version='3.6',
description='PebbleKit JS in Python!',
url='https://github.com/pebble/pypkjs',
author='Pebble Technology Corporation',
author_email='katharine@pebble.com',
license='MIT',
packages=find_packages(),
install_requires=requirements,
entry_points={
'console_scripts': [
'pypkjs=runner.websocket:run_tool'
],
},
zip_safe=False)
Add package data for GeoIP DB | __author__ = 'katharine'
import os
import sys
from setuptools import setup, find_packages
requirements_path = os.path.join(os.path.dirname(__file__), 'requirements.txt')
with open(requirements_path) as requirements_file:
requirements = [line.strip() for line in requirements_file.readlines()]
setup(name='pypkjs',
version='3.6',
description='PebbleKit JS in Python!',
url='https://github.com/pebble/pypkjs',
author='Pebble Technology Corporation',
author_email='katharine@pebble.com',
license='MIT',
packages=find_packages(),
install_requires=requirements,
package_data={
'javascript.navigator': 'GeoLiteCity.dat'
},
entry_points={
'console_scripts': [
'pypkjs=runner.websocket:run_tool'
],
},
zip_safe=False)
| <commit_before>__author__ = 'katharine'
import os
import sys
from setuptools import setup, find_packages
requirements_path = os.path.join(os.path.dirname(__file__), 'requirements.txt')
with open(requirements_path) as requirements_file:
requirements = [line.strip() for line in requirements_file.readlines()]
setup(name='pypkjs',
version='3.6',
description='PebbleKit JS in Python!',
url='https://github.com/pebble/pypkjs',
author='Pebble Technology Corporation',
author_email='katharine@pebble.com',
license='MIT',
packages=find_packages(),
install_requires=requirements,
entry_points={
'console_scripts': [
'pypkjs=runner.websocket:run_tool'
],
},
zip_safe=False)
<commit_msg>Add package data for GeoIP DB<commit_after> | __author__ = 'katharine'
import os
import sys
from setuptools import setup, find_packages
requirements_path = os.path.join(os.path.dirname(__file__), 'requirements.txt')
with open(requirements_path) as requirements_file:
requirements = [line.strip() for line in requirements_file.readlines()]
setup(name='pypkjs',
version='3.6',
description='PebbleKit JS in Python!',
url='https://github.com/pebble/pypkjs',
author='Pebble Technology Corporation',
author_email='katharine@pebble.com',
license='MIT',
packages=find_packages(),
install_requires=requirements,
package_data={
'javascript.navigator': 'GeoLiteCity.dat'
},
entry_points={
'console_scripts': [
'pypkjs=runner.websocket:run_tool'
],
},
zip_safe=False)
| __author__ = 'katharine'
import os
import sys
from setuptools import setup, find_packages
requirements_path = os.path.join(os.path.dirname(__file__), 'requirements.txt')
with open(requirements_path) as requirements_file:
requirements = [line.strip() for line in requirements_file.readlines()]
setup(name='pypkjs',
version='3.6',
description='PebbleKit JS in Python!',
url='https://github.com/pebble/pypkjs',
author='Pebble Technology Corporation',
author_email='katharine@pebble.com',
license='MIT',
packages=find_packages(),
install_requires=requirements,
entry_points={
'console_scripts': [
'pypkjs=runner.websocket:run_tool'
],
},
zip_safe=False)
Add package data for GeoIP DB__author__ = 'katharine'
import os
import sys
from setuptools import setup, find_packages
requirements_path = os.path.join(os.path.dirname(__file__), 'requirements.txt')
with open(requirements_path) as requirements_file:
requirements = [line.strip() for line in requirements_file.readlines()]
setup(name='pypkjs',
version='3.6',
description='PebbleKit JS in Python!',
url='https://github.com/pebble/pypkjs',
author='Pebble Technology Corporation',
author_email='katharine@pebble.com',
license='MIT',
packages=find_packages(),
install_requires=requirements,
package_data={
'javascript.navigator': 'GeoLiteCity.dat'
},
entry_points={
'console_scripts': [
'pypkjs=runner.websocket:run_tool'
],
},
zip_safe=False)
| <commit_before>__author__ = 'katharine'
import os
import sys
from setuptools import setup, find_packages
requirements_path = os.path.join(os.path.dirname(__file__), 'requirements.txt')
with open(requirements_path) as requirements_file:
requirements = [line.strip() for line in requirements_file.readlines()]
setup(name='pypkjs',
version='3.6',
description='PebbleKit JS in Python!',
url='https://github.com/pebble/pypkjs',
author='Pebble Technology Corporation',
author_email='katharine@pebble.com',
license='MIT',
packages=find_packages(),
install_requires=requirements,
entry_points={
'console_scripts': [
'pypkjs=runner.websocket:run_tool'
],
},
zip_safe=False)
<commit_msg>Add package data for GeoIP DB<commit_after>__author__ = 'katharine'
import os
import sys
from setuptools import setup, find_packages
requirements_path = os.path.join(os.path.dirname(__file__), 'requirements.txt')
with open(requirements_path) as requirements_file:
requirements = [line.strip() for line in requirements_file.readlines()]
setup(name='pypkjs',
version='3.6',
description='PebbleKit JS in Python!',
url='https://github.com/pebble/pypkjs',
author='Pebble Technology Corporation',
author_email='katharine@pebble.com',
license='MIT',
packages=find_packages(),
install_requires=requirements,
package_data={
'javascript.navigator': 'GeoLiteCity.dat'
},
entry_points={
'console_scripts': [
'pypkjs=runner.websocket:run_tool'
],
},
zip_safe=False)
|
e8770775250371766f47317a8aa40e034a5d75de | setup.py | setup.py | #!/usr/bin/env python
from setuptools import setup, find_packages
requires = [
]
setup(name='mirrors',
version='0.1.1',
description='rsync mirror manager',
author='Ethan House',
author_email='ehouse@csh.rit.edu',
packages=find_packages(),
install_requires=requires,
zip_safe=False,
entry_points="""
[console_scripts]
mirrors=mirrors:main
""")
| #!/usr/bin/env python
from setuptools import setup, find_packages
requires = [
]
setup(name='mirrors',
version='0.2.0',
description='rsync mirror manager',
author='Ethan House',
author_email='ehouse@csh.rit.edu',
packages=find_packages(),
install_requires=requires,
zip_safe=False,
entry_points="""
[console_scripts]
mirrors=mirrors:main
""")
| Move release version up to 0.2.0 | Move release version up to 0.2.0
| Python | bsd-2-clause | ehouse/mirrors | #!/usr/bin/env python
from setuptools import setup, find_packages
requires = [
]
setup(name='mirrors',
version='0.1.1',
description='rsync mirror manager',
author='Ethan House',
author_email='ehouse@csh.rit.edu',
packages=find_packages(),
install_requires=requires,
zip_safe=False,
entry_points="""
[console_scripts]
mirrors=mirrors:main
""")
Move release version up to 0.2.0 | #!/usr/bin/env python
from setuptools import setup, find_packages
requires = [
]
setup(name='mirrors',
version='0.2.0',
description='rsync mirror manager',
author='Ethan House',
author_email='ehouse@csh.rit.edu',
packages=find_packages(),
install_requires=requires,
zip_safe=False,
entry_points="""
[console_scripts]
mirrors=mirrors:main
""")
| <commit_before>#!/usr/bin/env python
from setuptools import setup, find_packages
requires = [
]
setup(name='mirrors',
version='0.1.1',
description='rsync mirror manager',
author='Ethan House',
author_email='ehouse@csh.rit.edu',
packages=find_packages(),
install_requires=requires,
zip_safe=False,
entry_points="""
[console_scripts]
mirrors=mirrors:main
""")
<commit_msg>Move release version up to 0.2.0<commit_after> | #!/usr/bin/env python
from setuptools import setup, find_packages
requires = [
]
setup(name='mirrors',
version='0.2.0',
description='rsync mirror manager',
author='Ethan House',
author_email='ehouse@csh.rit.edu',
packages=find_packages(),
install_requires=requires,
zip_safe=False,
entry_points="""
[console_scripts]
mirrors=mirrors:main
""")
| #!/usr/bin/env python
from setuptools import setup, find_packages
requires = [
]
setup(name='mirrors',
version='0.1.1',
description='rsync mirror manager',
author='Ethan House',
author_email='ehouse@csh.rit.edu',
packages=find_packages(),
install_requires=requires,
zip_safe=False,
entry_points="""
[console_scripts]
mirrors=mirrors:main
""")
Move release version up to 0.2.0#!/usr/bin/env python
from setuptools import setup, find_packages
requires = [
]
setup(name='mirrors',
version='0.2.0',
description='rsync mirror manager',
author='Ethan House',
author_email='ehouse@csh.rit.edu',
packages=find_packages(),
install_requires=requires,
zip_safe=False,
entry_points="""
[console_scripts]
mirrors=mirrors:main
""")
| <commit_before>#!/usr/bin/env python
from setuptools import setup, find_packages
requires = [
]
setup(name='mirrors',
version='0.1.1',
description='rsync mirror manager',
author='Ethan House',
author_email='ehouse@csh.rit.edu',
packages=find_packages(),
install_requires=requires,
zip_safe=False,
entry_points="""
[console_scripts]
mirrors=mirrors:main
""")
<commit_msg>Move release version up to 0.2.0<commit_after>#!/usr/bin/env python
from setuptools import setup, find_packages
requires = [
]
setup(name='mirrors',
version='0.2.0',
description='rsync mirror manager',
author='Ethan House',
author_email='ehouse@csh.rit.edu',
packages=find_packages(),
install_requires=requires,
zip_safe=False,
entry_points="""
[console_scripts]
mirrors=mirrors:main
""")
|
a463ac8ae112dd19bfc1c8e2df170023114ded07 | setup.py | setup.py | #!/usr/bin/env python
from setuptools import setup, find_packages
from setuputils import find_version, read
setup(
name='astor',
version=find_version('astor/__init__.py'),
description='Read/rewrite/write Python ASTs',
long_description=read('README.rst'),
author='Patrick Maupin',
author_email='pmaupin@gmail.com',
platforms='Independent',
url='https://github.com/berkerpeksag/astor',
packages=find_packages(exclude=['tests']),
py_modules=['setuputils'],
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: Implementation',
'Programming Language :: Python :: Implementation :: CPython',
'Programming Language :: Python :: Implementation :: PyPy',
'Topic :: Software Development :: Code Generators',
'Topic :: Software Development :: Compilers',
],
keywords='ast, codegen',
)
| #!/usr/bin/env python
from setuptools import setup, find_packages
from setuputils import find_version, read
setup(
name='astor',
version=find_version('astor/__init__.py'),
description='Read/rewrite/write Python ASTs',
long_description=read('README.rst'),
author='Patrick Maupin',
author_email='pmaupin@gmail.com',
platforms='Independent',
url='https://github.com/berkerpeksag/astor',
packages=find_packages(exclude=['tests']),
py_modules=['setuputils'],
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: Implementation',
'Programming Language :: Python :: Implementation :: CPython',
'Programming Language :: Python :: Implementation :: PyPy',
'Topic :: Software Development :: Code Generators',
'Topic :: Software Development :: Compilers',
],
keywords='ast, codegen',
)
| Remove 3.5 for now. It's not added to PyPI yet. | Remove 3.5 for now. It's not added to PyPI yet.
| Python | bsd-3-clause | berkerpeksag/astor | #!/usr/bin/env python
from setuptools import setup, find_packages
from setuputils import find_version, read
setup(
name='astor',
version=find_version('astor/__init__.py'),
description='Read/rewrite/write Python ASTs',
long_description=read('README.rst'),
author='Patrick Maupin',
author_email='pmaupin@gmail.com',
platforms='Independent',
url='https://github.com/berkerpeksag/astor',
packages=find_packages(exclude=['tests']),
py_modules=['setuputils'],
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: Implementation',
'Programming Language :: Python :: Implementation :: CPython',
'Programming Language :: Python :: Implementation :: PyPy',
'Topic :: Software Development :: Code Generators',
'Topic :: Software Development :: Compilers',
],
keywords='ast, codegen',
)
Remove 3.5 for now. It's not added to PyPI yet. | #!/usr/bin/env python
from setuptools import setup, find_packages
from setuputils import find_version, read
setup(
name='astor',
version=find_version('astor/__init__.py'),
description='Read/rewrite/write Python ASTs',
long_description=read('README.rst'),
author='Patrick Maupin',
author_email='pmaupin@gmail.com',
platforms='Independent',
url='https://github.com/berkerpeksag/astor',
packages=find_packages(exclude=['tests']),
py_modules=['setuputils'],
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: Implementation',
'Programming Language :: Python :: Implementation :: CPython',
'Programming Language :: Python :: Implementation :: PyPy',
'Topic :: Software Development :: Code Generators',
'Topic :: Software Development :: Compilers',
],
keywords='ast, codegen',
)
| <commit_before>#!/usr/bin/env python
from setuptools import setup, find_packages
from setuputils import find_version, read
setup(
name='astor',
version=find_version('astor/__init__.py'),
description='Read/rewrite/write Python ASTs',
long_description=read('README.rst'),
author='Patrick Maupin',
author_email='pmaupin@gmail.com',
platforms='Independent',
url='https://github.com/berkerpeksag/astor',
packages=find_packages(exclude=['tests']),
py_modules=['setuputils'],
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: Implementation',
'Programming Language :: Python :: Implementation :: CPython',
'Programming Language :: Python :: Implementation :: PyPy',
'Topic :: Software Development :: Code Generators',
'Topic :: Software Development :: Compilers',
],
keywords='ast, codegen',
)
<commit_msg>Remove 3.5 for now. It's not added to PyPI yet.<commit_after> | #!/usr/bin/env python
from setuptools import setup, find_packages
from setuputils import find_version, read
setup(
name='astor',
version=find_version('astor/__init__.py'),
description='Read/rewrite/write Python ASTs',
long_description=read('README.rst'),
author='Patrick Maupin',
author_email='pmaupin@gmail.com',
platforms='Independent',
url='https://github.com/berkerpeksag/astor',
packages=find_packages(exclude=['tests']),
py_modules=['setuputils'],
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: Implementation',
'Programming Language :: Python :: Implementation :: CPython',
'Programming Language :: Python :: Implementation :: PyPy',
'Topic :: Software Development :: Code Generators',
'Topic :: Software Development :: Compilers',
],
keywords='ast, codegen',
)
| #!/usr/bin/env python
from setuptools import setup, find_packages
from setuputils import find_version, read
setup(
name='astor',
version=find_version('astor/__init__.py'),
description='Read/rewrite/write Python ASTs',
long_description=read('README.rst'),
author='Patrick Maupin',
author_email='pmaupin@gmail.com',
platforms='Independent',
url='https://github.com/berkerpeksag/astor',
packages=find_packages(exclude=['tests']),
py_modules=['setuputils'],
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: Implementation',
'Programming Language :: Python :: Implementation :: CPython',
'Programming Language :: Python :: Implementation :: PyPy',
'Topic :: Software Development :: Code Generators',
'Topic :: Software Development :: Compilers',
],
keywords='ast, codegen',
)
Remove 3.5 for now. It's not added to PyPI yet.#!/usr/bin/env python
from setuptools import setup, find_packages
from setuputils import find_version, read
setup(
name='astor',
version=find_version('astor/__init__.py'),
description='Read/rewrite/write Python ASTs',
long_description=read('README.rst'),
author='Patrick Maupin',
author_email='pmaupin@gmail.com',
platforms='Independent',
url='https://github.com/berkerpeksag/astor',
packages=find_packages(exclude=['tests']),
py_modules=['setuputils'],
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: Implementation',
'Programming Language :: Python :: Implementation :: CPython',
'Programming Language :: Python :: Implementation :: PyPy',
'Topic :: Software Development :: Code Generators',
'Topic :: Software Development :: Compilers',
],
keywords='ast, codegen',
)
| <commit_before>#!/usr/bin/env python
from setuptools import setup, find_packages
from setuputils import find_version, read
setup(
name='astor',
version=find_version('astor/__init__.py'),
description='Read/rewrite/write Python ASTs',
long_description=read('README.rst'),
author='Patrick Maupin',
author_email='pmaupin@gmail.com',
platforms='Independent',
url='https://github.com/berkerpeksag/astor',
packages=find_packages(exclude=['tests']),
py_modules=['setuputils'],
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: Implementation',
'Programming Language :: Python :: Implementation :: CPython',
'Programming Language :: Python :: Implementation :: PyPy',
'Topic :: Software Development :: Code Generators',
'Topic :: Software Development :: Compilers',
],
keywords='ast, codegen',
)
<commit_msg>Remove 3.5 for now. It's not added to PyPI yet.<commit_after>#!/usr/bin/env python
from setuptools import setup, find_packages
from setuputils import find_version, read
setup(
name='astor',
version=find_version('astor/__init__.py'),
description='Read/rewrite/write Python ASTs',
long_description=read('README.rst'),
author='Patrick Maupin',
author_email='pmaupin@gmail.com',
platforms='Independent',
url='https://github.com/berkerpeksag/astor',
packages=find_packages(exclude=['tests']),
py_modules=['setuputils'],
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: Implementation',
'Programming Language :: Python :: Implementation :: CPython',
'Programming Language :: Python :: Implementation :: PyPy',
'Topic :: Software Development :: Code Generators',
'Topic :: Software Development :: Compilers',
],
keywords='ast, codegen',
)
|
9275bc3a9373e453e06f0aebc883a773dfb97627 | setup.py | setup.py | from setuptools import setup, find_packages
from tiddlywebwiki import __version__ as VERSION
setup(
name = 'tiddlywebwiki',
version = VERSION,
description = 'A TiddlyWeb plugin to provide a multi-user TiddlyWiki environment.',
author = 'FND',
author_email = 'FNDo@gmx.net',
packages = find_packages(exclude=['test']),
scripts = ['twinstance'],
platforms = 'Posix; MacOS X; Windows',
install_requires = [
'tiddlyweb>=0.9.96',
'tiddlywebplugins.wikklytextrender',
'tiddlywebplugins.status>=0.5',
'tiddlywebplugins.differ',
'tiddlywebplugins.atom',
'tiddlywebplugins.twimport',
'tiddlywebplugins.utils',
'tiddlywebplugins.instancer>=0.5.5',
'BeautifulSoup',
'wikklytext'],
include_package_data = True,
zip_safe = False
)
| from setuptools import setup, find_packages
from tiddlywebwiki import __version__ as VERSION
setup(
name = 'tiddlywebwiki',
version = VERSION,
description = 'A TiddlyWeb plugin to provide a multi-user TiddlyWiki environment.',
author = 'FND',
author_email = 'FNDo@gmx.net',
packages = find_packages(exclude=['test']),
scripts = ['twinstance'],
platforms = 'Posix; MacOS X; Windows',
install_requires = [
'tiddlyweb>=0.9.96',
'tiddlywebplugins.wikklytextrender',
'tiddlywebplugins.status>=0.5',
'tiddlywebplugins.differ',
'tiddlywebplugins.atom',
'tiddlywebplugins.twimport',
'tiddlywebplugins.utils',
'tiddlywebplugins.instancer>=0.5.5',
'wikklytext'],
include_package_data = True,
zip_safe = False
)
| Remove BeautifulSoup from direct dependency list. | Remove BeautifulSoup from direct dependency list.
| Python | bsd-3-clause | tiddlyweb/tiddlywebwiki,tiddlyweb/tiddlywebwiki,tiddlyweb/tiddlywebwiki | from setuptools import setup, find_packages
from tiddlywebwiki import __version__ as VERSION
setup(
name = 'tiddlywebwiki',
version = VERSION,
description = 'A TiddlyWeb plugin to provide a multi-user TiddlyWiki environment.',
author = 'FND',
author_email = 'FNDo@gmx.net',
packages = find_packages(exclude=['test']),
scripts = ['twinstance'],
platforms = 'Posix; MacOS X; Windows',
install_requires = [
'tiddlyweb>=0.9.96',
'tiddlywebplugins.wikklytextrender',
'tiddlywebplugins.status>=0.5',
'tiddlywebplugins.differ',
'tiddlywebplugins.atom',
'tiddlywebplugins.twimport',
'tiddlywebplugins.utils',
'tiddlywebplugins.instancer>=0.5.5',
'BeautifulSoup',
'wikklytext'],
include_package_data = True,
zip_safe = False
)
Remove BeautifulSoup from direct dependency list. | from setuptools import setup, find_packages
from tiddlywebwiki import __version__ as VERSION
setup(
name = 'tiddlywebwiki',
version = VERSION,
description = 'A TiddlyWeb plugin to provide a multi-user TiddlyWiki environment.',
author = 'FND',
author_email = 'FNDo@gmx.net',
packages = find_packages(exclude=['test']),
scripts = ['twinstance'],
platforms = 'Posix; MacOS X; Windows',
install_requires = [
'tiddlyweb>=0.9.96',
'tiddlywebplugins.wikklytextrender',
'tiddlywebplugins.status>=0.5',
'tiddlywebplugins.differ',
'tiddlywebplugins.atom',
'tiddlywebplugins.twimport',
'tiddlywebplugins.utils',
'tiddlywebplugins.instancer>=0.5.5',
'wikklytext'],
include_package_data = True,
zip_safe = False
)
| <commit_before>from setuptools import setup, find_packages
from tiddlywebwiki import __version__ as VERSION
setup(
name = 'tiddlywebwiki',
version = VERSION,
description = 'A TiddlyWeb plugin to provide a multi-user TiddlyWiki environment.',
author = 'FND',
author_email = 'FNDo@gmx.net',
packages = find_packages(exclude=['test']),
scripts = ['twinstance'],
platforms = 'Posix; MacOS X; Windows',
install_requires = [
'tiddlyweb>=0.9.96',
'tiddlywebplugins.wikklytextrender',
'tiddlywebplugins.status>=0.5',
'tiddlywebplugins.differ',
'tiddlywebplugins.atom',
'tiddlywebplugins.twimport',
'tiddlywebplugins.utils',
'tiddlywebplugins.instancer>=0.5.5',
'BeautifulSoup',
'wikklytext'],
include_package_data = True,
zip_safe = False
)
<commit_msg>Remove BeautifulSoup from direct dependency list.<commit_after> | from setuptools import setup, find_packages
from tiddlywebwiki import __version__ as VERSION
setup(
name = 'tiddlywebwiki',
version = VERSION,
description = 'A TiddlyWeb plugin to provide a multi-user TiddlyWiki environment.',
author = 'FND',
author_email = 'FNDo@gmx.net',
packages = find_packages(exclude=['test']),
scripts = ['twinstance'],
platforms = 'Posix; MacOS X; Windows',
install_requires = [
'tiddlyweb>=0.9.96',
'tiddlywebplugins.wikklytextrender',
'tiddlywebplugins.status>=0.5',
'tiddlywebplugins.differ',
'tiddlywebplugins.atom',
'tiddlywebplugins.twimport',
'tiddlywebplugins.utils',
'tiddlywebplugins.instancer>=0.5.5',
'wikklytext'],
include_package_data = True,
zip_safe = False
)
| from setuptools import setup, find_packages
from tiddlywebwiki import __version__ as VERSION
setup(
name = 'tiddlywebwiki',
version = VERSION,
description = 'A TiddlyWeb plugin to provide a multi-user TiddlyWiki environment.',
author = 'FND',
author_email = 'FNDo@gmx.net',
packages = find_packages(exclude=['test']),
scripts = ['twinstance'],
platforms = 'Posix; MacOS X; Windows',
install_requires = [
'tiddlyweb>=0.9.96',
'tiddlywebplugins.wikklytextrender',
'tiddlywebplugins.status>=0.5',
'tiddlywebplugins.differ',
'tiddlywebplugins.atom',
'tiddlywebplugins.twimport',
'tiddlywebplugins.utils',
'tiddlywebplugins.instancer>=0.5.5',
'BeautifulSoup',
'wikklytext'],
include_package_data = True,
zip_safe = False
)
Remove BeautifulSoup from direct dependency list.from setuptools import setup, find_packages
from tiddlywebwiki import __version__ as VERSION
setup(
name = 'tiddlywebwiki',
version = VERSION,
description = 'A TiddlyWeb plugin to provide a multi-user TiddlyWiki environment.',
author = 'FND',
author_email = 'FNDo@gmx.net',
packages = find_packages(exclude=['test']),
scripts = ['twinstance'],
platforms = 'Posix; MacOS X; Windows',
install_requires = [
'tiddlyweb>=0.9.96',
'tiddlywebplugins.wikklytextrender',
'tiddlywebplugins.status>=0.5',
'tiddlywebplugins.differ',
'tiddlywebplugins.atom',
'tiddlywebplugins.twimport',
'tiddlywebplugins.utils',
'tiddlywebplugins.instancer>=0.5.5',
'wikklytext'],
include_package_data = True,
zip_safe = False
)
| <commit_before>from setuptools import setup, find_packages
from tiddlywebwiki import __version__ as VERSION
setup(
name = 'tiddlywebwiki',
version = VERSION,
description = 'A TiddlyWeb plugin to provide a multi-user TiddlyWiki environment.',
author = 'FND',
author_email = 'FNDo@gmx.net',
packages = find_packages(exclude=['test']),
scripts = ['twinstance'],
platforms = 'Posix; MacOS X; Windows',
install_requires = [
'tiddlyweb>=0.9.96',
'tiddlywebplugins.wikklytextrender',
'tiddlywebplugins.status>=0.5',
'tiddlywebplugins.differ',
'tiddlywebplugins.atom',
'tiddlywebplugins.twimport',
'tiddlywebplugins.utils',
'tiddlywebplugins.instancer>=0.5.5',
'BeautifulSoup',
'wikklytext'],
include_package_data = True,
zip_safe = False
)
<commit_msg>Remove BeautifulSoup from direct dependency list.<commit_after>from setuptools import setup, find_packages
from tiddlywebwiki import __version__ as VERSION
setup(
name = 'tiddlywebwiki',
version = VERSION,
description = 'A TiddlyWeb plugin to provide a multi-user TiddlyWiki environment.',
author = 'FND',
author_email = 'FNDo@gmx.net',
packages = find_packages(exclude=['test']),
scripts = ['twinstance'],
platforms = 'Posix; MacOS X; Windows',
install_requires = [
'tiddlyweb>=0.9.96',
'tiddlywebplugins.wikklytextrender',
'tiddlywebplugins.status>=0.5',
'tiddlywebplugins.differ',
'tiddlywebplugins.atom',
'tiddlywebplugins.twimport',
'tiddlywebplugins.utils',
'tiddlywebplugins.instancer>=0.5.5',
'wikklytext'],
include_package_data = True,
zip_safe = False
)
|
7d547301e047556a7f95a76e80c0cf1fde5aa960 | setup.py | setup.py | import os
import re
from setuptools import setup
HERE = os.path.dirname(os.path.abspath(__file__))
def get_version():
filename = os.path.join(HERE, 'transfluent.py')
contents = open(filename).read()
pattern = r"^__version__ = '(.*?)'$"
return re.search(pattern, contents, re.MULTILINE).group(1)
setup(
name='transfluent',
version=get_version(),
description='A Python wrapper for Transfluent API',
long_description=(
open('README.rst').read() + '\n' +
open('CHANGES.rst').read()
),
author='Janne Vanhala',
author_email='janne.vanhala@gmail.com',
url='http://github.com/jpvanhal/python-transfluent',
py_modules=['transfluent'],
license=open('LICENSE').read(),
platforms='any',
install_requires=[
'requests>=1.0',
],
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: Implementation :: PyPy',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
| import os
import re
from setuptools import setup
HERE = os.path.dirname(os.path.abspath(__file__))
def get_version():
filename = os.path.join(HERE, 'transfluent.py')
contents = open(filename).read()
pattern = r"^__version__ = '(.*?)'$"
return re.search(pattern, contents, re.MULTILINE).group(1)
setup(
name='transfluent',
version=get_version(),
description='A Python wrapper for Transfluent API',
long_description=(
open('README.rst').read() + '\n' +
open('CHANGES.rst').read()
),
author='Janne Vanhala',
author_email='janne.vanhala@gmail.com',
url='http://github.com/jpvanhal/python-transfluent',
py_modules=['transfluent'],
license=open('LICENSE').read(),
platforms='any',
install_requires=[
'requests>=1.0',
],
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: Implementation :: PyPy',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
| Add Python 3.5 to trove classifiers | Add Python 3.5 to trove classifiers
| Python | bsd-3-clause | jpvanhal/python-transfluent | import os
import re
from setuptools import setup
HERE = os.path.dirname(os.path.abspath(__file__))
def get_version():
filename = os.path.join(HERE, 'transfluent.py')
contents = open(filename).read()
pattern = r"^__version__ = '(.*?)'$"
return re.search(pattern, contents, re.MULTILINE).group(1)
setup(
name='transfluent',
version=get_version(),
description='A Python wrapper for Transfluent API',
long_description=(
open('README.rst').read() + '\n' +
open('CHANGES.rst').read()
),
author='Janne Vanhala',
author_email='janne.vanhala@gmail.com',
url='http://github.com/jpvanhal/python-transfluent',
py_modules=['transfluent'],
license=open('LICENSE').read(),
platforms='any',
install_requires=[
'requests>=1.0',
],
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: Implementation :: PyPy',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
Add Python 3.5 to trove classifiers | import os
import re
from setuptools import setup
HERE = os.path.dirname(os.path.abspath(__file__))
def get_version():
filename = os.path.join(HERE, 'transfluent.py')
contents = open(filename).read()
pattern = r"^__version__ = '(.*?)'$"
return re.search(pattern, contents, re.MULTILINE).group(1)
setup(
name='transfluent',
version=get_version(),
description='A Python wrapper for Transfluent API',
long_description=(
open('README.rst').read() + '\n' +
open('CHANGES.rst').read()
),
author='Janne Vanhala',
author_email='janne.vanhala@gmail.com',
url='http://github.com/jpvanhal/python-transfluent',
py_modules=['transfluent'],
license=open('LICENSE').read(),
platforms='any',
install_requires=[
'requests>=1.0',
],
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: Implementation :: PyPy',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
| <commit_before>import os
import re
from setuptools import setup
HERE = os.path.dirname(os.path.abspath(__file__))
def get_version():
filename = os.path.join(HERE, 'transfluent.py')
contents = open(filename).read()
pattern = r"^__version__ = '(.*?)'$"
return re.search(pattern, contents, re.MULTILINE).group(1)
setup(
name='transfluent',
version=get_version(),
description='A Python wrapper for Transfluent API',
long_description=(
open('README.rst').read() + '\n' +
open('CHANGES.rst').read()
),
author='Janne Vanhala',
author_email='janne.vanhala@gmail.com',
url='http://github.com/jpvanhal/python-transfluent',
py_modules=['transfluent'],
license=open('LICENSE').read(),
platforms='any',
install_requires=[
'requests>=1.0',
],
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: Implementation :: PyPy',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
<commit_msg>Add Python 3.5 to trove classifiers<commit_after> | import os
import re
from setuptools import setup
HERE = os.path.dirname(os.path.abspath(__file__))
def get_version():
filename = os.path.join(HERE, 'transfluent.py')
contents = open(filename).read()
pattern = r"^__version__ = '(.*?)'$"
return re.search(pattern, contents, re.MULTILINE).group(1)
setup(
name='transfluent',
version=get_version(),
description='A Python wrapper for Transfluent API',
long_description=(
open('README.rst').read() + '\n' +
open('CHANGES.rst').read()
),
author='Janne Vanhala',
author_email='janne.vanhala@gmail.com',
url='http://github.com/jpvanhal/python-transfluent',
py_modules=['transfluent'],
license=open('LICENSE').read(),
platforms='any',
install_requires=[
'requests>=1.0',
],
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: Implementation :: PyPy',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
| import os
import re
from setuptools import setup
HERE = os.path.dirname(os.path.abspath(__file__))
def get_version():
filename = os.path.join(HERE, 'transfluent.py')
contents = open(filename).read()
pattern = r"^__version__ = '(.*?)'$"
return re.search(pattern, contents, re.MULTILINE).group(1)
setup(
name='transfluent',
version=get_version(),
description='A Python wrapper for Transfluent API',
long_description=(
open('README.rst').read() + '\n' +
open('CHANGES.rst').read()
),
author='Janne Vanhala',
author_email='janne.vanhala@gmail.com',
url='http://github.com/jpvanhal/python-transfluent',
py_modules=['transfluent'],
license=open('LICENSE').read(),
platforms='any',
install_requires=[
'requests>=1.0',
],
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: Implementation :: PyPy',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
Add Python 3.5 to trove classifiersimport os
import re
from setuptools import setup
HERE = os.path.dirname(os.path.abspath(__file__))
def get_version():
filename = os.path.join(HERE, 'transfluent.py')
contents = open(filename).read()
pattern = r"^__version__ = '(.*?)'$"
return re.search(pattern, contents, re.MULTILINE).group(1)
setup(
name='transfluent',
version=get_version(),
description='A Python wrapper for Transfluent API',
long_description=(
open('README.rst').read() + '\n' +
open('CHANGES.rst').read()
),
author='Janne Vanhala',
author_email='janne.vanhala@gmail.com',
url='http://github.com/jpvanhal/python-transfluent',
py_modules=['transfluent'],
license=open('LICENSE').read(),
platforms='any',
install_requires=[
'requests>=1.0',
],
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: Implementation :: PyPy',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
| <commit_before>import os
import re
from setuptools import setup
HERE = os.path.dirname(os.path.abspath(__file__))
def get_version():
filename = os.path.join(HERE, 'transfluent.py')
contents = open(filename).read()
pattern = r"^__version__ = '(.*?)'$"
return re.search(pattern, contents, re.MULTILINE).group(1)
setup(
name='transfluent',
version=get_version(),
description='A Python wrapper for Transfluent API',
long_description=(
open('README.rst').read() + '\n' +
open('CHANGES.rst').read()
),
author='Janne Vanhala',
author_email='janne.vanhala@gmail.com',
url='http://github.com/jpvanhal/python-transfluent',
py_modules=['transfluent'],
license=open('LICENSE').read(),
platforms='any',
install_requires=[
'requests>=1.0',
],
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: Implementation :: PyPy',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
<commit_msg>Add Python 3.5 to trove classifiers<commit_after>import os
import re
from setuptools import setup
HERE = os.path.dirname(os.path.abspath(__file__))
def get_version():
filename = os.path.join(HERE, 'transfluent.py')
contents = open(filename).read()
pattern = r"^__version__ = '(.*?)'$"
return re.search(pattern, contents, re.MULTILINE).group(1)
setup(
name='transfluent',
version=get_version(),
description='A Python wrapper for Transfluent API',
long_description=(
open('README.rst').read() + '\n' +
open('CHANGES.rst').read()
),
author='Janne Vanhala',
author_email='janne.vanhala@gmail.com',
url='http://github.com/jpvanhal/python-transfluent',
py_modules=['transfluent'],
license=open('LICENSE').read(),
platforms='any',
install_requires=[
'requests>=1.0',
],
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: Implementation :: PyPy',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
|
aa647c7e9a35b2293739c55e353319d2626f1f33 | setup.py | setup.py | #! /usr/bin/env python3
# -*- coding:Utf8 -*-
#--------------------------------------------------------------------------------------------------------------
# All necessary import:
#--------------------------------------------------------------------------------------------------------------
import os, sys, glob
import setuptools as st
from distutils.core import setup
from distutils.command.install_data import install_data
packages = st.find_packages()
#--------------------------------------------------------------------------------------------------------------
# Call the setup function:
#--------------------------------------------------------------------------------------------------------------
setup(
name = 'kmsi',
version = '0.1',
description = 'Python Module for analysis gadget simulation.',
author = 'Guillaume Plum',
packages = packages,
cmdclass = {'install_data': install_data},
# data_files = [
# ('share/LibThese/animation-plugins', ["share/LibThese/animation-plugins/__init__.py"]), #glob.glob("share/LibThese/animation-plugins/*.py")),
# ],
# scripts = [
# 'scripts/animationv2.py',
# 'scripts/models_plot.py',
# ],
)
#vim:spelllang=
| #! /usr/bin/env python3
# -*- coding:Utf8 -*-
#--------------------------------------------------------------------------------------------------------------
# All necessary import:
#--------------------------------------------------------------------------------------------------------------
import os, sys, glob
import setuptools as st
from distutils.core import setup
from distutils.command.install_data import install_data
packages = st.find_packages()
#--------------------------------------------------------------------------------------------------------------
# Call the setup function:
#--------------------------------------------------------------------------------------------------------------
setup(
name = 'kmsi',
version = '0.1',
description = 'Python Module for analysis gadget simulation.',
author = 'Guillaume Plum',
packages = packages,
cmdclass = {'install_data': install_data},
data_files = [
('share/kmsi', ["share/kmsi/config.yaml"]), #glob.glob("share/LibThese/animation-plugins/*.py")),
],
scripts = [
'scripts/kmsi',
],
)
#vim:spelllang=
| Add the kmsi script to the install files. | Add the kmsi script to the install files.
| Python | mit | ElricleNecro/kmsi | #! /usr/bin/env python3
# -*- coding:Utf8 -*-
#--------------------------------------------------------------------------------------------------------------
# All necessary import:
#--------------------------------------------------------------------------------------------------------------
import os, sys, glob
import setuptools as st
from distutils.core import setup
from distutils.command.install_data import install_data
packages = st.find_packages()
#--------------------------------------------------------------------------------------------------------------
# Call the setup function:
#--------------------------------------------------------------------------------------------------------------
setup(
name = 'kmsi',
version = '0.1',
description = 'Python Module for analysis gadget simulation.',
author = 'Guillaume Plum',
packages = packages,
cmdclass = {'install_data': install_data},
# data_files = [
# ('share/LibThese/animation-plugins', ["share/LibThese/animation-plugins/__init__.py"]), #glob.glob("share/LibThese/animation-plugins/*.py")),
# ],
# scripts = [
# 'scripts/animationv2.py',
# 'scripts/models_plot.py',
# ],
)
#vim:spelllang=
Add the kmsi script to the install files. | #! /usr/bin/env python3
# -*- coding:Utf8 -*-
#--------------------------------------------------------------------------------------------------------------
# All necessary import:
#--------------------------------------------------------------------------------------------------------------
import os, sys, glob
import setuptools as st
from distutils.core import setup
from distutils.command.install_data import install_data
packages = st.find_packages()
#--------------------------------------------------------------------------------------------------------------
# Call the setup function:
#--------------------------------------------------------------------------------------------------------------
setup(
name = 'kmsi',
version = '0.1',
description = 'Python Module for analysis gadget simulation.',
author = 'Guillaume Plum',
packages = packages,
cmdclass = {'install_data': install_data},
data_files = [
('share/kmsi', ["share/kmsi/config.yaml"]), #glob.glob("share/LibThese/animation-plugins/*.py")),
],
scripts = [
'scripts/kmsi',
],
)
#vim:spelllang=
| <commit_before>#! /usr/bin/env python3
# -*- coding:Utf8 -*-
#--------------------------------------------------------------------------------------------------------------
# All necessary import:
#--------------------------------------------------------------------------------------------------------------
import os, sys, glob
import setuptools as st
from distutils.core import setup
from distutils.command.install_data import install_data
packages = st.find_packages()
#--------------------------------------------------------------------------------------------------------------
# Call the setup function:
#--------------------------------------------------------------------------------------------------------------
setup(
name = 'kmsi',
version = '0.1',
description = 'Python Module for analysis gadget simulation.',
author = 'Guillaume Plum',
packages = packages,
cmdclass = {'install_data': install_data},
# data_files = [
# ('share/LibThese/animation-plugins', ["share/LibThese/animation-plugins/__init__.py"]), #glob.glob("share/LibThese/animation-plugins/*.py")),
# ],
# scripts = [
# 'scripts/animationv2.py',
# 'scripts/models_plot.py',
# ],
)
#vim:spelllang=
<commit_msg>Add the kmsi script to the install files.<commit_after> | #! /usr/bin/env python3
# -*- coding:Utf8 -*-
#--------------------------------------------------------------------------------------------------------------
# All necessary import:
#--------------------------------------------------------------------------------------------------------------
import os, sys, glob
import setuptools as st
from distutils.core import setup
from distutils.command.install_data import install_data
packages = st.find_packages()
#--------------------------------------------------------------------------------------------------------------
# Call the setup function:
#--------------------------------------------------------------------------------------------------------------
setup(
name = 'kmsi',
version = '0.1',
description = 'Python Module for analysis gadget simulation.',
author = 'Guillaume Plum',
packages = packages,
cmdclass = {'install_data': install_data},
data_files = [
('share/kmsi', ["share/kmsi/config.yaml"]), #glob.glob("share/LibThese/animation-plugins/*.py")),
],
scripts = [
'scripts/kmsi',
],
)
#vim:spelllang=
| #! /usr/bin/env python3
# -*- coding:Utf8 -*-
#--------------------------------------------------------------------------------------------------------------
# All necessary import:
#--------------------------------------------------------------------------------------------------------------
import os, sys, glob
import setuptools as st
from distutils.core import setup
from distutils.command.install_data import install_data
packages = st.find_packages()
#--------------------------------------------------------------------------------------------------------------
# Call the setup function:
#--------------------------------------------------------------------------------------------------------------
setup(
name = 'kmsi',
version = '0.1',
description = 'Python Module for analysis gadget simulation.',
author = 'Guillaume Plum',
packages = packages,
cmdclass = {'install_data': install_data},
# data_files = [
# ('share/LibThese/animation-plugins', ["share/LibThese/animation-plugins/__init__.py"]), #glob.glob("share/LibThese/animation-plugins/*.py")),
# ],
# scripts = [
# 'scripts/animationv2.py',
# 'scripts/models_plot.py',
# ],
)
#vim:spelllang=
Add the kmsi script to the install files.#! /usr/bin/env python3
# -*- coding:Utf8 -*-
#--------------------------------------------------------------------------------------------------------------
# All necessary import:
#--------------------------------------------------------------------------------------------------------------
import os, sys, glob
import setuptools as st
from distutils.core import setup
from distutils.command.install_data import install_data
packages = st.find_packages()
#--------------------------------------------------------------------------------------------------------------
# Call the setup function:
#--------------------------------------------------------------------------------------------------------------
setup(
name = 'kmsi',
version = '0.1',
description = 'Python Module for analysis gadget simulation.',
author = 'Guillaume Plum',
packages = packages,
cmdclass = {'install_data': install_data},
data_files = [
('share/kmsi', ["share/kmsi/config.yaml"]), #glob.glob("share/LibThese/animation-plugins/*.py")),
],
scripts = [
'scripts/kmsi',
],
)
#vim:spelllang=
| <commit_before>#! /usr/bin/env python3
# -*- coding:Utf8 -*-
#--------------------------------------------------------------------------------------------------------------
# All necessary import:
#--------------------------------------------------------------------------------------------------------------
import os, sys, glob
import setuptools as st
from distutils.core import setup
from distutils.command.install_data import install_data
packages = st.find_packages()
#--------------------------------------------------------------------------------------------------------------
# Call the setup function:
#--------------------------------------------------------------------------------------------------------------
setup(
name = 'kmsi',
version = '0.1',
description = 'Python Module for analysis gadget simulation.',
author = 'Guillaume Plum',
packages = packages,
cmdclass = {'install_data': install_data},
# data_files = [
# ('share/LibThese/animation-plugins', ["share/LibThese/animation-plugins/__init__.py"]), #glob.glob("share/LibThese/animation-plugins/*.py")),
# ],
# scripts = [
# 'scripts/animationv2.py',
# 'scripts/models_plot.py',
# ],
)
#vim:spelllang=
<commit_msg>Add the kmsi script to the install files.<commit_after>#! /usr/bin/env python3
# -*- coding:Utf8 -*-
#--------------------------------------------------------------------------------------------------------------
# All necessary import:
#--------------------------------------------------------------------------------------------------------------
import os, sys, glob
import setuptools as st
from distutils.core import setup
from distutils.command.install_data import install_data
packages = st.find_packages()
#--------------------------------------------------------------------------------------------------------------
# Call the setup function:
#--------------------------------------------------------------------------------------------------------------
setup(
name = 'kmsi',
version = '0.1',
description = 'Python Module for analysis gadget simulation.',
author = 'Guillaume Plum',
packages = packages,
cmdclass = {'install_data': install_data},
data_files = [
('share/kmsi', ["share/kmsi/config.yaml"]), #glob.glob("share/LibThese/animation-plugins/*.py")),
],
scripts = [
'scripts/kmsi',
],
)
#vim:spelllang=
|
f92c37200bb889188af21f0280c908f1bc2bcbff | setup.py | setup.py | from setuptools import setup, find_packages
setup(
name='incuna-pigeon',
version='0.0.0',
description='Notification management',
url='https://github.com/incuna/incuna-pigeon',
author='Incuna',
author_email='admin@incuna.com',
license='BSD',
classifiers=[
'Development Status :: 1 - Planning',
'Intended Audience :: Developers',
'Topic :: Communications',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
],
keywords='notifications',
packages=find_packages(),
)
| from setuptools import setup, find_packages
setup(
name='incuna-pigeon',
version='0.0.0',
description='Notification management',
url='https://github.com/incuna/incuna-pigeon',
author='Incuna',
author_email='admin@incuna.com',
license='BSD',
classifiers=[
'Development Status :: 1 - Planning',
'Intended Audience :: Developers',
'Topic :: Communications',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3 :: Only',
'Programming Language :: Python :: 3.4',
],
keywords='notifications',
packages=find_packages(),
)
| Add Python 3 Only trove classifier | Add Python 3 Only trove classifier
| Python | bsd-2-clause | incuna/incuna-pigeon | from setuptools import setup, find_packages
setup(
name='incuna-pigeon',
version='0.0.0',
description='Notification management',
url='https://github.com/incuna/incuna-pigeon',
author='Incuna',
author_email='admin@incuna.com',
license='BSD',
classifiers=[
'Development Status :: 1 - Planning',
'Intended Audience :: Developers',
'Topic :: Communications',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
],
keywords='notifications',
packages=find_packages(),
)
Add Python 3 Only trove classifier | from setuptools import setup, find_packages
setup(
name='incuna-pigeon',
version='0.0.0',
description='Notification management',
url='https://github.com/incuna/incuna-pigeon',
author='Incuna',
author_email='admin@incuna.com',
license='BSD',
classifiers=[
'Development Status :: 1 - Planning',
'Intended Audience :: Developers',
'Topic :: Communications',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3 :: Only',
'Programming Language :: Python :: 3.4',
],
keywords='notifications',
packages=find_packages(),
)
| <commit_before>from setuptools import setup, find_packages
setup(
name='incuna-pigeon',
version='0.0.0',
description='Notification management',
url='https://github.com/incuna/incuna-pigeon',
author='Incuna',
author_email='admin@incuna.com',
license='BSD',
classifiers=[
'Development Status :: 1 - Planning',
'Intended Audience :: Developers',
'Topic :: Communications',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
],
keywords='notifications',
packages=find_packages(),
)
<commit_msg>Add Python 3 Only trove classifier<commit_after> | from setuptools import setup, find_packages
setup(
name='incuna-pigeon',
version='0.0.0',
description='Notification management',
url='https://github.com/incuna/incuna-pigeon',
author='Incuna',
author_email='admin@incuna.com',
license='BSD',
classifiers=[
'Development Status :: 1 - Planning',
'Intended Audience :: Developers',
'Topic :: Communications',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3 :: Only',
'Programming Language :: Python :: 3.4',
],
keywords='notifications',
packages=find_packages(),
)
| from setuptools import setup, find_packages
setup(
name='incuna-pigeon',
version='0.0.0',
description='Notification management',
url='https://github.com/incuna/incuna-pigeon',
author='Incuna',
author_email='admin@incuna.com',
license='BSD',
classifiers=[
'Development Status :: 1 - Planning',
'Intended Audience :: Developers',
'Topic :: Communications',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
],
keywords='notifications',
packages=find_packages(),
)
Add Python 3 Only trove classifierfrom setuptools import setup, find_packages
setup(
name='incuna-pigeon',
version='0.0.0',
description='Notification management',
url='https://github.com/incuna/incuna-pigeon',
author='Incuna',
author_email='admin@incuna.com',
license='BSD',
classifiers=[
'Development Status :: 1 - Planning',
'Intended Audience :: Developers',
'Topic :: Communications',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3 :: Only',
'Programming Language :: Python :: 3.4',
],
keywords='notifications',
packages=find_packages(),
)
| <commit_before>from setuptools import setup, find_packages
setup(
name='incuna-pigeon',
version='0.0.0',
description='Notification management',
url='https://github.com/incuna/incuna-pigeon',
author='Incuna',
author_email='admin@incuna.com',
license='BSD',
classifiers=[
'Development Status :: 1 - Planning',
'Intended Audience :: Developers',
'Topic :: Communications',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
],
keywords='notifications',
packages=find_packages(),
)
<commit_msg>Add Python 3 Only trove classifier<commit_after>from setuptools import setup, find_packages
setup(
name='incuna-pigeon',
version='0.0.0',
description='Notification management',
url='https://github.com/incuna/incuna-pigeon',
author='Incuna',
author_email='admin@incuna.com',
license='BSD',
classifiers=[
'Development Status :: 1 - Planning',
'Intended Audience :: Developers',
'Topic :: Communications',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3 :: Only',
'Programming Language :: Python :: 3.4',
],
keywords='notifications',
packages=find_packages(),
)
|
51c65f37ea5f0d2cd98de8e63f541d533e1f8a65 | setup.py | setup.py | #!/usr/bin/env python
from codecs import open
from setuptools import find_packages, setup
with open('README.rst', 'r', 'utf-8') as f:
readme = f.read()
setup(
name='django-paginationlinks',
version='0.1.1',
description='Django Pagination Links',
long_description=readme,
url='https://github.com/blancltd/django-paginationlinks',
maintainer='Blanc Ltd',
maintainer_email='studio@blanc.ltd.uk',
platforms=['any'],
packages=find_packages(),
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
license='BSD',
)
| #!/usr/bin/env python
from codecs import open
from setuptools import find_packages, setup
with open('README.rst', 'r', 'utf-8') as f:
readme = f.read()
setup(
name='django-paginationlinks',
version='0.1.1',
description='Django Pagination Links',
long_description=readme,
url='https://github.com/developersociety/django-paginationlinks',
maintainer='Blanc Ltd',
maintainer_email='studio@blanc.ltd.uk',
platforms=['any'],
packages=find_packages(),
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
license='BSD',
)
| Update GitHub repos from blancltd to developersociety | Update GitHub repos from blancltd to developersociety
| Python | bsd-3-clause | blancltd/django-paginationlinks | #!/usr/bin/env python
from codecs import open
from setuptools import find_packages, setup
with open('README.rst', 'r', 'utf-8') as f:
readme = f.read()
setup(
name='django-paginationlinks',
version='0.1.1',
description='Django Pagination Links',
long_description=readme,
url='https://github.com/blancltd/django-paginationlinks',
maintainer='Blanc Ltd',
maintainer_email='studio@blanc.ltd.uk',
platforms=['any'],
packages=find_packages(),
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
license='BSD',
)
Update GitHub repos from blancltd to developersociety | #!/usr/bin/env python
from codecs import open
from setuptools import find_packages, setup
with open('README.rst', 'r', 'utf-8') as f:
readme = f.read()
setup(
name='django-paginationlinks',
version='0.1.1',
description='Django Pagination Links',
long_description=readme,
url='https://github.com/developersociety/django-paginationlinks',
maintainer='Blanc Ltd',
maintainer_email='studio@blanc.ltd.uk',
platforms=['any'],
packages=find_packages(),
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
license='BSD',
)
| <commit_before>#!/usr/bin/env python
from codecs import open
from setuptools import find_packages, setup
with open('README.rst', 'r', 'utf-8') as f:
readme = f.read()
setup(
name='django-paginationlinks',
version='0.1.1',
description='Django Pagination Links',
long_description=readme,
url='https://github.com/blancltd/django-paginationlinks',
maintainer='Blanc Ltd',
maintainer_email='studio@blanc.ltd.uk',
platforms=['any'],
packages=find_packages(),
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
license='BSD',
)
<commit_msg>Update GitHub repos from blancltd to developersociety<commit_after> | #!/usr/bin/env python
from codecs import open
from setuptools import find_packages, setup
with open('README.rst', 'r', 'utf-8') as f:
readme = f.read()
setup(
name='django-paginationlinks',
version='0.1.1',
description='Django Pagination Links',
long_description=readme,
url='https://github.com/developersociety/django-paginationlinks',
maintainer='Blanc Ltd',
maintainer_email='studio@blanc.ltd.uk',
platforms=['any'],
packages=find_packages(),
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
license='BSD',
)
| #!/usr/bin/env python
from codecs import open
from setuptools import find_packages, setup
with open('README.rst', 'r', 'utf-8') as f:
readme = f.read()
setup(
name='django-paginationlinks',
version='0.1.1',
description='Django Pagination Links',
long_description=readme,
url='https://github.com/blancltd/django-paginationlinks',
maintainer='Blanc Ltd',
maintainer_email='studio@blanc.ltd.uk',
platforms=['any'],
packages=find_packages(),
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
license='BSD',
)
Update GitHub repos from blancltd to developersociety#!/usr/bin/env python
from codecs import open
from setuptools import find_packages, setup
with open('README.rst', 'r', 'utf-8') as f:
readme = f.read()
setup(
name='django-paginationlinks',
version='0.1.1',
description='Django Pagination Links',
long_description=readme,
url='https://github.com/developersociety/django-paginationlinks',
maintainer='Blanc Ltd',
maintainer_email='studio@blanc.ltd.uk',
platforms=['any'],
packages=find_packages(),
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
license='BSD',
)
| <commit_before>#!/usr/bin/env python
from codecs import open
from setuptools import find_packages, setup
with open('README.rst', 'r', 'utf-8') as f:
readme = f.read()
setup(
name='django-paginationlinks',
version='0.1.1',
description='Django Pagination Links',
long_description=readme,
url='https://github.com/blancltd/django-paginationlinks',
maintainer='Blanc Ltd',
maintainer_email='studio@blanc.ltd.uk',
platforms=['any'],
packages=find_packages(),
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
license='BSD',
)
<commit_msg>Update GitHub repos from blancltd to developersociety<commit_after>#!/usr/bin/env python
from codecs import open
from setuptools import find_packages, setup
with open('README.rst', 'r', 'utf-8') as f:
readme = f.read()
setup(
name='django-paginationlinks',
version='0.1.1',
description='Django Pagination Links',
long_description=readme,
url='https://github.com/developersociety/django-paginationlinks',
maintainer='Blanc Ltd',
maintainer_email='studio@blanc.ltd.uk',
platforms=['any'],
packages=find_packages(),
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
license='BSD',
)
|
e8efd3b22d7b301771d72b2d7b9ca4cca474e302 | setup.py | setup.py |
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
config = {
'description': 'pelican-do',
'author': 'Commands to automate common pelican tasks',
'url': '',
'download_url': '',
'author_email': 'gustavoajz@gmail.com',
'version': '0.1',
'install_requires': [
'click==6.2',
'Jinja2==2.8',
'awesome-slugify==1.6.5',
],
'extras_require': {
'development': [
],
},
'setup_requires': [
'pytest-runner',
],
'tests_require': [
'pytest==2.8.5',
'pytest-cov==2.2.0'
],
'packages': ['pelican_do'],
'scripts': [],
'name': 'pelican-do',
'entry_points': {
'console_scripts': ['pelican-do=pelican_do.main:main']
}
}
setup(**config)
|
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
config = {
'description': 'pelican-do',
'author': 'Commands to automate common pelican tasks',
'url': '',
'download_url': '',
'author_email': 'gustavoajz@gmail.com',
'version': '0.1',
'install_requires': [
'click==6.2',
'Jinja2==2.8',
'awesome-slugify==1.6.5',
],
'extras_require': {
'development': [
],
},
'setup_requires': [
'pytest-runner',
],
'tests_require': [
'pytest>=2.6.4',
'pytest-cov==2.2.0'
],
'packages': ['pelican_do'],
'scripts': [],
'name': 'pelican-do',
'entry_points': {
'console_scripts': ['pelican-do=pelican_do.main:main']
}
}
setup(**config)
| Downgrade pytest version to be able to use default shippable minion | Downgrade pytest version to be able to use default shippable minion
| Python | mit | gusajz/pelican-do,gusajz/pelican-do |
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
config = {
'description': 'pelican-do',
'author': 'Commands to automate common pelican tasks',
'url': '',
'download_url': '',
'author_email': 'gustavoajz@gmail.com',
'version': '0.1',
'install_requires': [
'click==6.2',
'Jinja2==2.8',
'awesome-slugify==1.6.5',
],
'extras_require': {
'development': [
],
},
'setup_requires': [
'pytest-runner',
],
'tests_require': [
'pytest==2.8.5',
'pytest-cov==2.2.0'
],
'packages': ['pelican_do'],
'scripts': [],
'name': 'pelican-do',
'entry_points': {
'console_scripts': ['pelican-do=pelican_do.main:main']
}
}
setup(**config)
Downgrade pytest version to be able to use default shippable minion |
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
config = {
'description': 'pelican-do',
'author': 'Commands to automate common pelican tasks',
'url': '',
'download_url': '',
'author_email': 'gustavoajz@gmail.com',
'version': '0.1',
'install_requires': [
'click==6.2',
'Jinja2==2.8',
'awesome-slugify==1.6.5',
],
'extras_require': {
'development': [
],
},
'setup_requires': [
'pytest-runner',
],
'tests_require': [
'pytest>=2.6.4',
'pytest-cov==2.2.0'
],
'packages': ['pelican_do'],
'scripts': [],
'name': 'pelican-do',
'entry_points': {
'console_scripts': ['pelican-do=pelican_do.main:main']
}
}
setup(**config)
| <commit_before>
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
config = {
'description': 'pelican-do',
'author': 'Commands to automate common pelican tasks',
'url': '',
'download_url': '',
'author_email': 'gustavoajz@gmail.com',
'version': '0.1',
'install_requires': [
'click==6.2',
'Jinja2==2.8',
'awesome-slugify==1.6.5',
],
'extras_require': {
'development': [
],
},
'setup_requires': [
'pytest-runner',
],
'tests_require': [
'pytest==2.8.5',
'pytest-cov==2.2.0'
],
'packages': ['pelican_do'],
'scripts': [],
'name': 'pelican-do',
'entry_points': {
'console_scripts': ['pelican-do=pelican_do.main:main']
}
}
setup(**config)
<commit_msg>Downgrade pytest version to be able to use default shippable minion<commit_after> |
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
config = {
'description': 'pelican-do',
'author': 'Commands to automate common pelican tasks',
'url': '',
'download_url': '',
'author_email': 'gustavoajz@gmail.com',
'version': '0.1',
'install_requires': [
'click==6.2',
'Jinja2==2.8',
'awesome-slugify==1.6.5',
],
'extras_require': {
'development': [
],
},
'setup_requires': [
'pytest-runner',
],
'tests_require': [
'pytest>=2.6.4',
'pytest-cov==2.2.0'
],
'packages': ['pelican_do'],
'scripts': [],
'name': 'pelican-do',
'entry_points': {
'console_scripts': ['pelican-do=pelican_do.main:main']
}
}
setup(**config)
|
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
config = {
'description': 'pelican-do',
'author': 'Commands to automate common pelican tasks',
'url': '',
'download_url': '',
'author_email': 'gustavoajz@gmail.com',
'version': '0.1',
'install_requires': [
'click==6.2',
'Jinja2==2.8',
'awesome-slugify==1.6.5',
],
'extras_require': {
'development': [
],
},
'setup_requires': [
'pytest-runner',
],
'tests_require': [
'pytest==2.8.5',
'pytest-cov==2.2.0'
],
'packages': ['pelican_do'],
'scripts': [],
'name': 'pelican-do',
'entry_points': {
'console_scripts': ['pelican-do=pelican_do.main:main']
}
}
setup(**config)
Downgrade pytest version to be able to use default shippable minion
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
config = {
'description': 'pelican-do',
'author': 'Commands to automate common pelican tasks',
'url': '',
'download_url': '',
'author_email': 'gustavoajz@gmail.com',
'version': '0.1',
'install_requires': [
'click==6.2',
'Jinja2==2.8',
'awesome-slugify==1.6.5',
],
'extras_require': {
'development': [
],
},
'setup_requires': [
'pytest-runner',
],
'tests_require': [
'pytest>=2.6.4',
'pytest-cov==2.2.0'
],
'packages': ['pelican_do'],
'scripts': [],
'name': 'pelican-do',
'entry_points': {
'console_scripts': ['pelican-do=pelican_do.main:main']
}
}
setup(**config)
| <commit_before>
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
config = {
'description': 'pelican-do',
'author': 'Commands to automate common pelican tasks',
'url': '',
'download_url': '',
'author_email': 'gustavoajz@gmail.com',
'version': '0.1',
'install_requires': [
'click==6.2',
'Jinja2==2.8',
'awesome-slugify==1.6.5',
],
'extras_require': {
'development': [
],
},
'setup_requires': [
'pytest-runner',
],
'tests_require': [
'pytest==2.8.5',
'pytest-cov==2.2.0'
],
'packages': ['pelican_do'],
'scripts': [],
'name': 'pelican-do',
'entry_points': {
'console_scripts': ['pelican-do=pelican_do.main:main']
}
}
setup(**config)
<commit_msg>Downgrade pytest version to be able to use default shippable minion<commit_after>
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
config = {
'description': 'pelican-do',
'author': 'Commands to automate common pelican tasks',
'url': '',
'download_url': '',
'author_email': 'gustavoajz@gmail.com',
'version': '0.1',
'install_requires': [
'click==6.2',
'Jinja2==2.8',
'awesome-slugify==1.6.5',
],
'extras_require': {
'development': [
],
},
'setup_requires': [
'pytest-runner',
],
'tests_require': [
'pytest>=2.6.4',
'pytest-cov==2.2.0'
],
'packages': ['pelican_do'],
'scripts': [],
'name': 'pelican-do',
'entry_points': {
'console_scripts': ['pelican-do=pelican_do.main:main']
}
}
setup(**config)
|
b7acc8ca9c6c41aff7ffb419125f54d21da09652 | setup.py | setup.py | #!/usr/bin/env python
from distutils.core import setup
setup(name='plyprotobuf',
version='1.0',
description='Protobuf Parsing Library that uses ply',
author='Dusan Klinec',
url='https://github.com/sb98052/plyprotobuf',
packages=['plyproto'],
)
| #!/usr/bin/env python
from distutils.core import setup
setup(name='plyprotobuf',
version='1.0',
description='Protobuf Parsing Library that uses ply',
author='Dusan Klinec',
url='https://github.com/sb98052/plyprotobuf',
packages=['plyproto'],
install_requires=['ply']
)
| Add dependency to ply package | Add dependency to ply package
| Python | apache-2.0 | sb98052/plyprotobuf | #!/usr/bin/env python
from distutils.core import setup
setup(name='plyprotobuf',
version='1.0',
description='Protobuf Parsing Library that uses ply',
author='Dusan Klinec',
url='https://github.com/sb98052/plyprotobuf',
packages=['plyproto'],
)
Add dependency to ply package | #!/usr/bin/env python
from distutils.core import setup
setup(name='plyprotobuf',
version='1.0',
description='Protobuf Parsing Library that uses ply',
author='Dusan Klinec',
url='https://github.com/sb98052/plyprotobuf',
packages=['plyproto'],
install_requires=['ply']
)
| <commit_before>#!/usr/bin/env python
from distutils.core import setup
setup(name='plyprotobuf',
version='1.0',
description='Protobuf Parsing Library that uses ply',
author='Dusan Klinec',
url='https://github.com/sb98052/plyprotobuf',
packages=['plyproto'],
)
<commit_msg>Add dependency to ply package<commit_after> | #!/usr/bin/env python
from distutils.core import setup
setup(name='plyprotobuf',
version='1.0',
description='Protobuf Parsing Library that uses ply',
author='Dusan Klinec',
url='https://github.com/sb98052/plyprotobuf',
packages=['plyproto'],
install_requires=['ply']
)
| #!/usr/bin/env python
from distutils.core import setup
setup(name='plyprotobuf',
version='1.0',
description='Protobuf Parsing Library that uses ply',
author='Dusan Klinec',
url='https://github.com/sb98052/plyprotobuf',
packages=['plyproto'],
)
Add dependency to ply package#!/usr/bin/env python
from distutils.core import setup
setup(name='plyprotobuf',
version='1.0',
description='Protobuf Parsing Library that uses ply',
author='Dusan Klinec',
url='https://github.com/sb98052/plyprotobuf',
packages=['plyproto'],
install_requires=['ply']
)
| <commit_before>#!/usr/bin/env python
from distutils.core import setup
setup(name='plyprotobuf',
version='1.0',
description='Protobuf Parsing Library that uses ply',
author='Dusan Klinec',
url='https://github.com/sb98052/plyprotobuf',
packages=['plyproto'],
)
<commit_msg>Add dependency to ply package<commit_after>#!/usr/bin/env python
from distutils.core import setup
setup(name='plyprotobuf',
version='1.0',
description='Protobuf Parsing Library that uses ply',
author='Dusan Klinec',
url='https://github.com/sb98052/plyprotobuf',
packages=['plyproto'],
install_requires=['ply']
)
|
08b4cc4e065e63eef522756888fa8a75d9bf6ddb | setup.py | setup.py | from setuptools import setup, find_packages
setup(
name='django-nap',
version='0.3',
description='A light REST tool for Django',
author='Curtis Maloney',
author_email='curtis@tinbrain.net',
url='http://github.com/funkybob/django-nap',
keywords=['django', 'json', 'rest'],
packages = find_packages(),
zip_safe=False,
classifiers = [
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
requires = [
'Django (>=1.4)',
],
)
| from setuptools import setup, find_packages
setup(
name='django-nap',
version='0.4',
description='A light REST tool for Django',
author='Curtis Maloney',
author_email='curtis@tinbrain.net',
url='http://github.com/funkybob/django-nap',
keywords=['django', 'json', 'rest'],
packages = find_packages(excludes=['test.*']),
zip_safe=False,
classifiers = [
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
requires = [
'Django (>=1.4)',
],
)
| Fix find packages Bump version | Fix find packages
Bump version
| Python | bsd-3-clause | limbera/django-nap,MarkusH/django-nap | from setuptools import setup, find_packages
setup(
name='django-nap',
version='0.3',
description='A light REST tool for Django',
author='Curtis Maloney',
author_email='curtis@tinbrain.net',
url='http://github.com/funkybob/django-nap',
keywords=['django', 'json', 'rest'],
packages = find_packages(),
zip_safe=False,
classifiers = [
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
requires = [
'Django (>=1.4)',
],
)
Fix find packages
Bump version | from setuptools import setup, find_packages
setup(
name='django-nap',
version='0.4',
description='A light REST tool for Django',
author='Curtis Maloney',
author_email='curtis@tinbrain.net',
url='http://github.com/funkybob/django-nap',
keywords=['django', 'json', 'rest'],
packages = find_packages(excludes=['test.*']),
zip_safe=False,
classifiers = [
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
requires = [
'Django (>=1.4)',
],
)
| <commit_before>from setuptools import setup, find_packages
setup(
name='django-nap',
version='0.3',
description='A light REST tool for Django',
author='Curtis Maloney',
author_email='curtis@tinbrain.net',
url='http://github.com/funkybob/django-nap',
keywords=['django', 'json', 'rest'],
packages = find_packages(),
zip_safe=False,
classifiers = [
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
requires = [
'Django (>=1.4)',
],
)
<commit_msg>Fix find packages
Bump version<commit_after> | from setuptools import setup, find_packages
setup(
name='django-nap',
version='0.4',
description='A light REST tool for Django',
author='Curtis Maloney',
author_email='curtis@tinbrain.net',
url='http://github.com/funkybob/django-nap',
keywords=['django', 'json', 'rest'],
packages = find_packages(excludes=['test.*']),
zip_safe=False,
classifiers = [
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
requires = [
'Django (>=1.4)',
],
)
| from setuptools import setup, find_packages
setup(
name='django-nap',
version='0.3',
description='A light REST tool for Django',
author='Curtis Maloney',
author_email='curtis@tinbrain.net',
url='http://github.com/funkybob/django-nap',
keywords=['django', 'json', 'rest'],
packages = find_packages(),
zip_safe=False,
classifiers = [
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
requires = [
'Django (>=1.4)',
],
)
Fix find packages
Bump versionfrom setuptools import setup, find_packages
setup(
name='django-nap',
version='0.4',
description='A light REST tool for Django',
author='Curtis Maloney',
author_email='curtis@tinbrain.net',
url='http://github.com/funkybob/django-nap',
keywords=['django', 'json', 'rest'],
packages = find_packages(excludes=['test.*']),
zip_safe=False,
classifiers = [
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
requires = [
'Django (>=1.4)',
],
)
| <commit_before>from setuptools import setup, find_packages
setup(
name='django-nap',
version='0.3',
description='A light REST tool for Django',
author='Curtis Maloney',
author_email='curtis@tinbrain.net',
url='http://github.com/funkybob/django-nap',
keywords=['django', 'json', 'rest'],
packages = find_packages(),
zip_safe=False,
classifiers = [
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
requires = [
'Django (>=1.4)',
],
)
<commit_msg>Fix find packages
Bump version<commit_after>from setuptools import setup, find_packages
setup(
name='django-nap',
version='0.4',
description='A light REST tool for Django',
author='Curtis Maloney',
author_email='curtis@tinbrain.net',
url='http://github.com/funkybob/django-nap',
keywords=['django', 'json', 'rest'],
packages = find_packages(excludes=['test.*']),
zip_safe=False,
classifiers = [
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
requires = [
'Django (>=1.4)',
],
)
|
1bb67543b4187ee05c616afb731c229aaa94fdd3 | setup.py | setup.py | import os
import sys
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
requirements = ['httplib2', 'argparse', 'prettytable']
if sys.version_info < (2, 6):
requirements.append('simplejson')
setup(
name = "python-keystoneclient",
version = "2.7",
description = "Client library for OpenStack Keystone API",
long_description = read('README.rst'),
url = 'https://github.com/4P/python-keystoneclient',
license = 'Apache',
author = 'Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss',
author_email = 'gabriel.hurley@nebula.com',
packages = find_packages(exclude=['tests', 'tests.*']),
classifiers = [
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
install_requires = requirements,
tests_require = ["nose", "mock", "mox"],
test_suite = "nose.collector",
entry_points = {
'console_scripts': ['keystone = keystoneclient.shell:main']
}
)
| import os
import sys
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
requirements = ['httplib2', 'argparse', 'prettytable']
if sys.version_info < (2, 6):
requirements.append('simplejson')
setup(
name = "python-keystoneclient",
version = "2012.1",
description = "Client library for OpenStack Keystone API",
long_description = read('README.rst'),
url = 'https://github.com/openstack/python-keystoneclient',
license = 'Apache',
author = 'Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss',
author_email = 'gabriel.hurley@nebula.com',
packages = find_packages(exclude=['tests', 'tests.*']),
classifiers = [
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
install_requires = requirements,
tests_require = ["nose", "mock", "mox"],
test_suite = "nose.collector",
entry_points = {
'console_scripts': ['keystone = keystoneclient.shell:main']
}
)
| Adjust version number to match other deliveries | Adjust version number to match other deliveries
Set version from 2.7 to 2012.1 to match the other OpenStack
Keystone deliveries (python-keystoneclient will be released
as part of Keystone 2012.1~e3). Also adjusted the location
of the git repository to match new location. Fixes bug 917656.
Change-Id: I4d8d071e3cdc5665e29a89067958f5f1e8964221
| Python | apache-2.0 | citrix-openstack-build/keystoneauth,jamielennox/keystoneauth,sileht/keystoneauth | import os
import sys
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
requirements = ['httplib2', 'argparse', 'prettytable']
if sys.version_info < (2, 6):
requirements.append('simplejson')
setup(
name = "python-keystoneclient",
version = "2.7",
description = "Client library for OpenStack Keystone API",
long_description = read('README.rst'),
url = 'https://github.com/4P/python-keystoneclient',
license = 'Apache',
author = 'Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss',
author_email = 'gabriel.hurley@nebula.com',
packages = find_packages(exclude=['tests', 'tests.*']),
classifiers = [
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
install_requires = requirements,
tests_require = ["nose", "mock", "mox"],
test_suite = "nose.collector",
entry_points = {
'console_scripts': ['keystone = keystoneclient.shell:main']
}
)
Adjust version number to match other deliveries
Set version from 2.7 to 2012.1 to match the other OpenStack
Keystone deliveries (python-keystoneclient will be released
as part of Keystone 2012.1~e3). Also adjusted the location
of the git repository to match new location. Fixes bug 917656.
Change-Id: I4d8d071e3cdc5665e29a89067958f5f1e8964221 | import os
import sys
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
requirements = ['httplib2', 'argparse', 'prettytable']
if sys.version_info < (2, 6):
requirements.append('simplejson')
setup(
name = "python-keystoneclient",
version = "2012.1",
description = "Client library for OpenStack Keystone API",
long_description = read('README.rst'),
url = 'https://github.com/openstack/python-keystoneclient',
license = 'Apache',
author = 'Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss',
author_email = 'gabriel.hurley@nebula.com',
packages = find_packages(exclude=['tests', 'tests.*']),
classifiers = [
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
install_requires = requirements,
tests_require = ["nose", "mock", "mox"],
test_suite = "nose.collector",
entry_points = {
'console_scripts': ['keystone = keystoneclient.shell:main']
}
)
| <commit_before>import os
import sys
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
requirements = ['httplib2', 'argparse', 'prettytable']
if sys.version_info < (2, 6):
requirements.append('simplejson')
setup(
name = "python-keystoneclient",
version = "2.7",
description = "Client library for OpenStack Keystone API",
long_description = read('README.rst'),
url = 'https://github.com/4P/python-keystoneclient',
license = 'Apache',
author = 'Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss',
author_email = 'gabriel.hurley@nebula.com',
packages = find_packages(exclude=['tests', 'tests.*']),
classifiers = [
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
install_requires = requirements,
tests_require = ["nose", "mock", "mox"],
test_suite = "nose.collector",
entry_points = {
'console_scripts': ['keystone = keystoneclient.shell:main']
}
)
<commit_msg>Adjust version number to match other deliveries
Set version from 2.7 to 2012.1 to match the other OpenStack
Keystone deliveries (python-keystoneclient will be released
as part of Keystone 2012.1~e3). Also adjusted the location
of the git repository to match new location. Fixes bug 917656.
Change-Id: I4d8d071e3cdc5665e29a89067958f5f1e8964221<commit_after> | import os
import sys
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
requirements = ['httplib2', 'argparse', 'prettytable']
if sys.version_info < (2, 6):
requirements.append('simplejson')
setup(
name = "python-keystoneclient",
version = "2012.1",
description = "Client library for OpenStack Keystone API",
long_description = read('README.rst'),
url = 'https://github.com/openstack/python-keystoneclient',
license = 'Apache',
author = 'Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss',
author_email = 'gabriel.hurley@nebula.com',
packages = find_packages(exclude=['tests', 'tests.*']),
classifiers = [
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
install_requires = requirements,
tests_require = ["nose", "mock", "mox"],
test_suite = "nose.collector",
entry_points = {
'console_scripts': ['keystone = keystoneclient.shell:main']
}
)
| import os
import sys
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
requirements = ['httplib2', 'argparse', 'prettytable']
if sys.version_info < (2, 6):
requirements.append('simplejson')
setup(
name = "python-keystoneclient",
version = "2.7",
description = "Client library for OpenStack Keystone API",
long_description = read('README.rst'),
url = 'https://github.com/4P/python-keystoneclient',
license = 'Apache',
author = 'Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss',
author_email = 'gabriel.hurley@nebula.com',
packages = find_packages(exclude=['tests', 'tests.*']),
classifiers = [
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
install_requires = requirements,
tests_require = ["nose", "mock", "mox"],
test_suite = "nose.collector",
entry_points = {
'console_scripts': ['keystone = keystoneclient.shell:main']
}
)
Adjust version number to match other deliveries
Set version from 2.7 to 2012.1 to match the other OpenStack
Keystone deliveries (python-keystoneclient will be released
as part of Keystone 2012.1~e3). Also adjusted the location
of the git repository to match new location. Fixes bug 917656.
Change-Id: I4d8d071e3cdc5665e29a89067958f5f1e8964221import os
import sys
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
requirements = ['httplib2', 'argparse', 'prettytable']
if sys.version_info < (2, 6):
requirements.append('simplejson')
setup(
name = "python-keystoneclient",
version = "2012.1",
description = "Client library for OpenStack Keystone API",
long_description = read('README.rst'),
url = 'https://github.com/openstack/python-keystoneclient',
license = 'Apache',
author = 'Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss',
author_email = 'gabriel.hurley@nebula.com',
packages = find_packages(exclude=['tests', 'tests.*']),
classifiers = [
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
install_requires = requirements,
tests_require = ["nose", "mock", "mox"],
test_suite = "nose.collector",
entry_points = {
'console_scripts': ['keystone = keystoneclient.shell:main']
}
)
| <commit_before>import os
import sys
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
requirements = ['httplib2', 'argparse', 'prettytable']
if sys.version_info < (2, 6):
requirements.append('simplejson')
setup(
name = "python-keystoneclient",
version = "2.7",
description = "Client library for OpenStack Keystone API",
long_description = read('README.rst'),
url = 'https://github.com/4P/python-keystoneclient',
license = 'Apache',
author = 'Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss',
author_email = 'gabriel.hurley@nebula.com',
packages = find_packages(exclude=['tests', 'tests.*']),
classifiers = [
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
install_requires = requirements,
tests_require = ["nose", "mock", "mox"],
test_suite = "nose.collector",
entry_points = {
'console_scripts': ['keystone = keystoneclient.shell:main']
}
)
<commit_msg>Adjust version number to match other deliveries
Set version from 2.7 to 2012.1 to match the other OpenStack
Keystone deliveries (python-keystoneclient will be released
as part of Keystone 2012.1~e3). Also adjusted the location
of the git repository to match new location. Fixes bug 917656.
Change-Id: I4d8d071e3cdc5665e29a89067958f5f1e8964221<commit_after>import os
import sys
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
requirements = ['httplib2', 'argparse', 'prettytable']
if sys.version_info < (2, 6):
requirements.append('simplejson')
setup(
name = "python-keystoneclient",
version = "2012.1",
description = "Client library for OpenStack Keystone API",
long_description = read('README.rst'),
url = 'https://github.com/openstack/python-keystoneclient',
license = 'Apache',
author = 'Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss',
author_email = 'gabriel.hurley@nebula.com',
packages = find_packages(exclude=['tests', 'tests.*']),
classifiers = [
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
install_requires = requirements,
tests_require = ["nose", "mock", "mox"],
test_suite = "nose.collector",
entry_points = {
'console_scripts': ['keystone = keystoneclient.shell:main']
}
)
|
710d41a1c1b328793224975ee5afc4aebe462f28 | setup.py | setup.py | from setuptools import setup, find_packages
import os
version = '0.4'
here = os.path.abspath(os.path.dirname(__file__))
try:
README = open(os.path.join(here, 'README.txt')).read()
CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read()
except IOError:
README = CHANGES = ''
setup(name='tgext.admin',
version=version,
description="Admin Controller add-on for basic TG identity model.",
long_description=README + "\n" +
CHANGES,
# Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers
classifiers=[
"Programming Language :: Python",
"Topic :: Software Development :: Libraries :: Python Modules",
],
keywords='TG2, TG, sprox, Rest, internet, adminn',
author='Christopher Perkins',
author_email='chris@percious.com',
url='tgtools.googlecode.com',
license='MIT',
packages=find_packages(exclude=['ez_setup']),
namespace_packages=['tgext'],
include_package_data=True,
zip_safe=True,
install_requires=[
'setuptools',
'tgext.crud>=0.4',
# -*- Extra requirements: -*-
],
entry_points="""
# -*- Entry points: -*-
""",
)
| from setuptools import setup, find_packages
import os
version = '0.5'
here = os.path.abspath(os.path.dirname(__file__))
try:
README = open(os.path.join(here, 'README.txt')).read()
CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read()
except IOError:
README = CHANGES = ''
setup(name='tgext.admin',
version=version,
description="Admin Controller add-on for basic TG identity model.",
long_description=README + "\n" +
CHANGES,
# Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers
classifiers=[
"Programming Language :: Python",
"Topic :: Software Development :: Libraries :: Python Modules",
],
keywords='turbogears2.extension, TG2, TG, sprox, Rest, internet, admin',
author='Christopher Perkins',
author_email='chris@percious.com',
url='tgtools.googlecode.com',
license='MIT',
packages=find_packages(exclude=['ez_setup']),
namespace_packages=['tgext'],
include_package_data=True,
zip_safe=True,
install_requires=[
'setuptools',
'tgext.crud>=0.4',
# -*- Extra requirements: -*-
],
entry_points="""
# -*- Entry points: -*-
""",
)
| Raise version and change keywords for upcoming release | Raise version and change keywords for upcoming release
| Python | mit | TurboGears/tgext.admin,TurboGears/tgext.admin | from setuptools import setup, find_packages
import os
version = '0.4'
here = os.path.abspath(os.path.dirname(__file__))
try:
README = open(os.path.join(here, 'README.txt')).read()
CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read()
except IOError:
README = CHANGES = ''
setup(name='tgext.admin',
version=version,
description="Admin Controller add-on for basic TG identity model.",
long_description=README + "\n" +
CHANGES,
# Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers
classifiers=[
"Programming Language :: Python",
"Topic :: Software Development :: Libraries :: Python Modules",
],
keywords='TG2, TG, sprox, Rest, internet, adminn',
author='Christopher Perkins',
author_email='chris@percious.com',
url='tgtools.googlecode.com',
license='MIT',
packages=find_packages(exclude=['ez_setup']),
namespace_packages=['tgext'],
include_package_data=True,
zip_safe=True,
install_requires=[
'setuptools',
'tgext.crud>=0.4',
# -*- Extra requirements: -*-
],
entry_points="""
# -*- Entry points: -*-
""",
)
Raise version and change keywords for upcoming release | from setuptools import setup, find_packages
import os
version = '0.5'
here = os.path.abspath(os.path.dirname(__file__))
try:
README = open(os.path.join(here, 'README.txt')).read()
CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read()
except IOError:
README = CHANGES = ''
setup(name='tgext.admin',
version=version,
description="Admin Controller add-on for basic TG identity model.",
long_description=README + "\n" +
CHANGES,
# Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers
classifiers=[
"Programming Language :: Python",
"Topic :: Software Development :: Libraries :: Python Modules",
],
keywords='turbogears2.extension, TG2, TG, sprox, Rest, internet, admin',
author='Christopher Perkins',
author_email='chris@percious.com',
url='tgtools.googlecode.com',
license='MIT',
packages=find_packages(exclude=['ez_setup']),
namespace_packages=['tgext'],
include_package_data=True,
zip_safe=True,
install_requires=[
'setuptools',
'tgext.crud>=0.4',
# -*- Extra requirements: -*-
],
entry_points="""
# -*- Entry points: -*-
""",
)
| <commit_before>from setuptools import setup, find_packages
import os
version = '0.4'
here = os.path.abspath(os.path.dirname(__file__))
try:
README = open(os.path.join(here, 'README.txt')).read()
CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read()
except IOError:
README = CHANGES = ''
setup(name='tgext.admin',
version=version,
description="Admin Controller add-on for basic TG identity model.",
long_description=README + "\n" +
CHANGES,
# Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers
classifiers=[
"Programming Language :: Python",
"Topic :: Software Development :: Libraries :: Python Modules",
],
keywords='TG2, TG, sprox, Rest, internet, adminn',
author='Christopher Perkins',
author_email='chris@percious.com',
url='tgtools.googlecode.com',
license='MIT',
packages=find_packages(exclude=['ez_setup']),
namespace_packages=['tgext'],
include_package_data=True,
zip_safe=True,
install_requires=[
'setuptools',
'tgext.crud>=0.4',
# -*- Extra requirements: -*-
],
entry_points="""
# -*- Entry points: -*-
""",
)
<commit_msg>Raise version and change keywords for upcoming release<commit_after> | from setuptools import setup, find_packages
import os
version = '0.5'
here = os.path.abspath(os.path.dirname(__file__))
try:
README = open(os.path.join(here, 'README.txt')).read()
CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read()
except IOError:
README = CHANGES = ''
setup(name='tgext.admin',
version=version,
description="Admin Controller add-on for basic TG identity model.",
long_description=README + "\n" +
CHANGES,
# Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers
classifiers=[
"Programming Language :: Python",
"Topic :: Software Development :: Libraries :: Python Modules",
],
keywords='turbogears2.extension, TG2, TG, sprox, Rest, internet, admin',
author='Christopher Perkins',
author_email='chris@percious.com',
url='tgtools.googlecode.com',
license='MIT',
packages=find_packages(exclude=['ez_setup']),
namespace_packages=['tgext'],
include_package_data=True,
zip_safe=True,
install_requires=[
'setuptools',
'tgext.crud>=0.4',
# -*- Extra requirements: -*-
],
entry_points="""
# -*- Entry points: -*-
""",
)
| from setuptools import setup, find_packages
import os
version = '0.4'
here = os.path.abspath(os.path.dirname(__file__))
try:
README = open(os.path.join(here, 'README.txt')).read()
CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read()
except IOError:
README = CHANGES = ''
setup(name='tgext.admin',
version=version,
description="Admin Controller add-on for basic TG identity model.",
long_description=README + "\n" +
CHANGES,
# Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers
classifiers=[
"Programming Language :: Python",
"Topic :: Software Development :: Libraries :: Python Modules",
],
keywords='TG2, TG, sprox, Rest, internet, adminn',
author='Christopher Perkins',
author_email='chris@percious.com',
url='tgtools.googlecode.com',
license='MIT',
packages=find_packages(exclude=['ez_setup']),
namespace_packages=['tgext'],
include_package_data=True,
zip_safe=True,
install_requires=[
'setuptools',
'tgext.crud>=0.4',
# -*- Extra requirements: -*-
],
entry_points="""
# -*- Entry points: -*-
""",
)
Raise version and change keywords for upcoming releasefrom setuptools import setup, find_packages
import os
version = '0.5'
here = os.path.abspath(os.path.dirname(__file__))
try:
README = open(os.path.join(here, 'README.txt')).read()
CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read()
except IOError:
README = CHANGES = ''
setup(name='tgext.admin',
version=version,
description="Admin Controller add-on for basic TG identity model.",
long_description=README + "\n" +
CHANGES,
# Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers
classifiers=[
"Programming Language :: Python",
"Topic :: Software Development :: Libraries :: Python Modules",
],
keywords='turbogears2.extension, TG2, TG, sprox, Rest, internet, admin',
author='Christopher Perkins',
author_email='chris@percious.com',
url='tgtools.googlecode.com',
license='MIT',
packages=find_packages(exclude=['ez_setup']),
namespace_packages=['tgext'],
include_package_data=True,
zip_safe=True,
install_requires=[
'setuptools',
'tgext.crud>=0.4',
# -*- Extra requirements: -*-
],
entry_points="""
# -*- Entry points: -*-
""",
)
| <commit_before>from setuptools import setup, find_packages
import os
version = '0.4'
here = os.path.abspath(os.path.dirname(__file__))
try:
README = open(os.path.join(here, 'README.txt')).read()
CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read()
except IOError:
README = CHANGES = ''
setup(name='tgext.admin',
version=version,
description="Admin Controller add-on for basic TG identity model.",
long_description=README + "\n" +
CHANGES,
# Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers
classifiers=[
"Programming Language :: Python",
"Topic :: Software Development :: Libraries :: Python Modules",
],
keywords='TG2, TG, sprox, Rest, internet, adminn',
author='Christopher Perkins',
author_email='chris@percious.com',
url='tgtools.googlecode.com',
license='MIT',
packages=find_packages(exclude=['ez_setup']),
namespace_packages=['tgext'],
include_package_data=True,
zip_safe=True,
install_requires=[
'setuptools',
'tgext.crud>=0.4',
# -*- Extra requirements: -*-
],
entry_points="""
# -*- Entry points: -*-
""",
)
<commit_msg>Raise version and change keywords for upcoming release<commit_after>from setuptools import setup, find_packages
import os
version = '0.5'
here = os.path.abspath(os.path.dirname(__file__))
try:
README = open(os.path.join(here, 'README.txt')).read()
CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read()
except IOError:
README = CHANGES = ''
setup(name='tgext.admin',
version=version,
description="Admin Controller add-on for basic TG identity model.",
long_description=README + "\n" +
CHANGES,
# Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers
classifiers=[
"Programming Language :: Python",
"Topic :: Software Development :: Libraries :: Python Modules",
],
keywords='turbogears2.extension, TG2, TG, sprox, Rest, internet, admin',
author='Christopher Perkins',
author_email='chris@percious.com',
url='tgtools.googlecode.com',
license='MIT',
packages=find_packages(exclude=['ez_setup']),
namespace_packages=['tgext'],
include_package_data=True,
zip_safe=True,
install_requires=[
'setuptools',
'tgext.crud>=0.4',
# -*- Extra requirements: -*-
],
entry_points="""
# -*- Entry points: -*-
""",
)
|
c983c4dc27c547dab0c5f4afd5438231d26ce840 | setup.py | setup.py | from setuptools import setup
setup(
name='ncdjango',
description='A map server for NetCDF data',
keywords='netcdf,django,map server',
version='0.4.0',
packages=[
'ncdjango', 'ncdjango.geoprocessing', 'ncdjango.migrations', 'ncdjango.interfaces',
'ncdjango.interfaces.arcgis', 'ncdjango.interfaces.arcgis_extended', 'ncdjango.interfaces.data'
],
install_requires=[
'six', 'requests', 'Django>=1.7.0,<1.9.0', 'Pillow>=2.9.0', 'Shapely>=1.3.2', 'GDAL>=1.11.0',
'django-tastypie>=0.11.1', 'netCDF4>=1.1.6', 'numpy>=1.8.1', 'pyproj>=1.9.4', 'fiona', 'rasterio>=0.28.0',
'clover', 'ply>=3.8', 'celery>=3.1.19', 'djangorestframework'
],
dependency_links=[
'git+https://github.com/consbio/clover.git'
],
url='https://github.com/consbio/ncdjango',
license='BSD',
)
| from setuptools import setup
setup(
name='ncdjango',
description='A map server for NetCDF data',
keywords='netcdf,django,map server',
version='0.4.0',
packages=[
'ncdjango', 'ncdjango.geoprocessing', 'ncdjango.migrations', 'ncdjango.interfaces',
'ncdjango.interfaces.arcgis', 'ncdjango.interfaces.arcgis_extended', 'ncdjango.interfaces.data'
],
install_requires=[
'six', 'requests', 'Django>=1.7.0,<1.9.0', 'Pillow>=2.9.0', 'Shapely>=1.3.2', 'django-tastypie>=0.11.1',
'netCDF4>=1.1.6', 'numpy>=1.8.1', 'pyproj>=1.9.4', 'fiona', 'rasterio>=0.28.0', 'clover', 'ply>=3.8',
'celery>=3.1.19', 'djangorestframework'
],
dependency_links=[
'git+https://github.com/consbio/clover.git'
],
url='https://github.com/consbio/ncdjango',
license='BSD',
)
| Remove GDAL requirement (it's not a direct dependency) | Remove GDAL requirement (it's not a direct dependency)
| Python | bsd-3-clause | consbio/ncdjango,consbio/ncdjango | from setuptools import setup
setup(
name='ncdjango',
description='A map server for NetCDF data',
keywords='netcdf,django,map server',
version='0.4.0',
packages=[
'ncdjango', 'ncdjango.geoprocessing', 'ncdjango.migrations', 'ncdjango.interfaces',
'ncdjango.interfaces.arcgis', 'ncdjango.interfaces.arcgis_extended', 'ncdjango.interfaces.data'
],
install_requires=[
'six', 'requests', 'Django>=1.7.0,<1.9.0', 'Pillow>=2.9.0', 'Shapely>=1.3.2', 'GDAL>=1.11.0',
'django-tastypie>=0.11.1', 'netCDF4>=1.1.6', 'numpy>=1.8.1', 'pyproj>=1.9.4', 'fiona', 'rasterio>=0.28.0',
'clover', 'ply>=3.8', 'celery>=3.1.19', 'djangorestframework'
],
dependency_links=[
'git+https://github.com/consbio/clover.git'
],
url='https://github.com/consbio/ncdjango',
license='BSD',
)
Remove GDAL requirement (it's not a direct dependency) | from setuptools import setup
setup(
name='ncdjango',
description='A map server for NetCDF data',
keywords='netcdf,django,map server',
version='0.4.0',
packages=[
'ncdjango', 'ncdjango.geoprocessing', 'ncdjango.migrations', 'ncdjango.interfaces',
'ncdjango.interfaces.arcgis', 'ncdjango.interfaces.arcgis_extended', 'ncdjango.interfaces.data'
],
install_requires=[
'six', 'requests', 'Django>=1.7.0,<1.9.0', 'Pillow>=2.9.0', 'Shapely>=1.3.2', 'django-tastypie>=0.11.1',
'netCDF4>=1.1.6', 'numpy>=1.8.1', 'pyproj>=1.9.4', 'fiona', 'rasterio>=0.28.0', 'clover', 'ply>=3.8',
'celery>=3.1.19', 'djangorestframework'
],
dependency_links=[
'git+https://github.com/consbio/clover.git'
],
url='https://github.com/consbio/ncdjango',
license='BSD',
)
| <commit_before>from setuptools import setup
setup(
name='ncdjango',
description='A map server for NetCDF data',
keywords='netcdf,django,map server',
version='0.4.0',
packages=[
'ncdjango', 'ncdjango.geoprocessing', 'ncdjango.migrations', 'ncdjango.interfaces',
'ncdjango.interfaces.arcgis', 'ncdjango.interfaces.arcgis_extended', 'ncdjango.interfaces.data'
],
install_requires=[
'six', 'requests', 'Django>=1.7.0,<1.9.0', 'Pillow>=2.9.0', 'Shapely>=1.3.2', 'GDAL>=1.11.0',
'django-tastypie>=0.11.1', 'netCDF4>=1.1.6', 'numpy>=1.8.1', 'pyproj>=1.9.4', 'fiona', 'rasterio>=0.28.0',
'clover', 'ply>=3.8', 'celery>=3.1.19', 'djangorestframework'
],
dependency_links=[
'git+https://github.com/consbio/clover.git'
],
url='https://github.com/consbio/ncdjango',
license='BSD',
)
<commit_msg>Remove GDAL requirement (it's not a direct dependency)<commit_after> | from setuptools import setup
setup(
name='ncdjango',
description='A map server for NetCDF data',
keywords='netcdf,django,map server',
version='0.4.0',
packages=[
'ncdjango', 'ncdjango.geoprocessing', 'ncdjango.migrations', 'ncdjango.interfaces',
'ncdjango.interfaces.arcgis', 'ncdjango.interfaces.arcgis_extended', 'ncdjango.interfaces.data'
],
install_requires=[
'six', 'requests', 'Django>=1.7.0,<1.9.0', 'Pillow>=2.9.0', 'Shapely>=1.3.2', 'django-tastypie>=0.11.1',
'netCDF4>=1.1.6', 'numpy>=1.8.1', 'pyproj>=1.9.4', 'fiona', 'rasterio>=0.28.0', 'clover', 'ply>=3.8',
'celery>=3.1.19', 'djangorestframework'
],
dependency_links=[
'git+https://github.com/consbio/clover.git'
],
url='https://github.com/consbio/ncdjango',
license='BSD',
)
| from setuptools import setup
setup(
name='ncdjango',
description='A map server for NetCDF data',
keywords='netcdf,django,map server',
version='0.4.0',
packages=[
'ncdjango', 'ncdjango.geoprocessing', 'ncdjango.migrations', 'ncdjango.interfaces',
'ncdjango.interfaces.arcgis', 'ncdjango.interfaces.arcgis_extended', 'ncdjango.interfaces.data'
],
install_requires=[
'six', 'requests', 'Django>=1.7.0,<1.9.0', 'Pillow>=2.9.0', 'Shapely>=1.3.2', 'GDAL>=1.11.0',
'django-tastypie>=0.11.1', 'netCDF4>=1.1.6', 'numpy>=1.8.1', 'pyproj>=1.9.4', 'fiona', 'rasterio>=0.28.0',
'clover', 'ply>=3.8', 'celery>=3.1.19', 'djangorestframework'
],
dependency_links=[
'git+https://github.com/consbio/clover.git'
],
url='https://github.com/consbio/ncdjango',
license='BSD',
)
Remove GDAL requirement (it's not a direct dependency)from setuptools import setup
setup(
name='ncdjango',
description='A map server for NetCDF data',
keywords='netcdf,django,map server',
version='0.4.0',
packages=[
'ncdjango', 'ncdjango.geoprocessing', 'ncdjango.migrations', 'ncdjango.interfaces',
'ncdjango.interfaces.arcgis', 'ncdjango.interfaces.arcgis_extended', 'ncdjango.interfaces.data'
],
install_requires=[
'six', 'requests', 'Django>=1.7.0,<1.9.0', 'Pillow>=2.9.0', 'Shapely>=1.3.2', 'django-tastypie>=0.11.1',
'netCDF4>=1.1.6', 'numpy>=1.8.1', 'pyproj>=1.9.4', 'fiona', 'rasterio>=0.28.0', 'clover', 'ply>=3.8',
'celery>=3.1.19', 'djangorestframework'
],
dependency_links=[
'git+https://github.com/consbio/clover.git'
],
url='https://github.com/consbio/ncdjango',
license='BSD',
)
| <commit_before>from setuptools import setup
setup(
name='ncdjango',
description='A map server for NetCDF data',
keywords='netcdf,django,map server',
version='0.4.0',
packages=[
'ncdjango', 'ncdjango.geoprocessing', 'ncdjango.migrations', 'ncdjango.interfaces',
'ncdjango.interfaces.arcgis', 'ncdjango.interfaces.arcgis_extended', 'ncdjango.interfaces.data'
],
install_requires=[
'six', 'requests', 'Django>=1.7.0,<1.9.0', 'Pillow>=2.9.0', 'Shapely>=1.3.2', 'GDAL>=1.11.0',
'django-tastypie>=0.11.1', 'netCDF4>=1.1.6', 'numpy>=1.8.1', 'pyproj>=1.9.4', 'fiona', 'rasterio>=0.28.0',
'clover', 'ply>=3.8', 'celery>=3.1.19', 'djangorestframework'
],
dependency_links=[
'git+https://github.com/consbio/clover.git'
],
url='https://github.com/consbio/ncdjango',
license='BSD',
)
<commit_msg>Remove GDAL requirement (it's not a direct dependency)<commit_after>from setuptools import setup
setup(
name='ncdjango',
description='A map server for NetCDF data',
keywords='netcdf,django,map server',
version='0.4.0',
packages=[
'ncdjango', 'ncdjango.geoprocessing', 'ncdjango.migrations', 'ncdjango.interfaces',
'ncdjango.interfaces.arcgis', 'ncdjango.interfaces.arcgis_extended', 'ncdjango.interfaces.data'
],
install_requires=[
'six', 'requests', 'Django>=1.7.0,<1.9.0', 'Pillow>=2.9.0', 'Shapely>=1.3.2', 'django-tastypie>=0.11.1',
'netCDF4>=1.1.6', 'numpy>=1.8.1', 'pyproj>=1.9.4', 'fiona', 'rasterio>=0.28.0', 'clover', 'ply>=3.8',
'celery>=3.1.19', 'djangorestframework'
],
dependency_links=[
'git+https://github.com/consbio/clover.git'
],
url='https://github.com/consbio/ncdjango',
license='BSD',
)
|
0a33b7d8df544226df711db33a27f45421c19290 | setup.py | setup.py | from setuptools import setup
version = '2.0.0'
setup(name='pyactiveresource',
version=version,
description='ActiveResource for Python',
author='Shopify',
author_email='developers@shopify.com',
url='https://github.com/Shopify/pyactiveresource/',
packages=['pyactiveresource', 'pyactiveresource/testing'],
license='MIT License',
test_suite='test',
tests_require=[
'python-dateutil<2.0', # >= 2.0 is for python>=3.0
'PyYAML',
],
platforms=['any'],
classifiers=['Development Status :: 5 - Production/Stable',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Software Development',
'Topic :: Software Development :: Libraries',
'Topic :: Software Development :: Libraries :: Python Modules']
)
| from setuptools import setup
import sys
version = '2.0.0'
if sys.version_info >= (3,):
python_dateutils_version = 'python-dateutil>=2.0'
else:
python_dateutils_version = 'python-dateutil<2.0'
setup(name='pyactiveresource',
version=version,
description='ActiveResource for Python',
author='Shopify',
author_email='developers@shopify.com',
url='https://github.com/Shopify/pyactiveresource/',
packages=['pyactiveresource', 'pyactiveresource/testing'],
license='MIT License',
test_suite='test',
tests_require=[
python_dateutils_version,
'PyYAML',
],
platforms=['any'],
classifiers=['Development Status :: 5 - Production/Stable',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Software Development',
'Topic :: Software Development :: Libraries',
'Topic :: Software Development :: Libraries :: Python Modules']
)
| Use the right version of python-dateutils when using python 3. | Use the right version of python-dateutils when using python 3.
| Python | mit | metric-collective/pyactiveresource,piran/pyactiveresource,varesa/pyactiveresource,hockeybuggy/pyactiveresource | from setuptools import setup
version = '2.0.0'
setup(name='pyactiveresource',
version=version,
description='ActiveResource for Python',
author='Shopify',
author_email='developers@shopify.com',
url='https://github.com/Shopify/pyactiveresource/',
packages=['pyactiveresource', 'pyactiveresource/testing'],
license='MIT License',
test_suite='test',
tests_require=[
'python-dateutil<2.0', # >= 2.0 is for python>=3.0
'PyYAML',
],
platforms=['any'],
classifiers=['Development Status :: 5 - Production/Stable',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Software Development',
'Topic :: Software Development :: Libraries',
'Topic :: Software Development :: Libraries :: Python Modules']
)
Use the right version of python-dateutils when using python 3. | from setuptools import setup
import sys
version = '2.0.0'
if sys.version_info >= (3,):
python_dateutils_version = 'python-dateutil>=2.0'
else:
python_dateutils_version = 'python-dateutil<2.0'
setup(name='pyactiveresource',
version=version,
description='ActiveResource for Python',
author='Shopify',
author_email='developers@shopify.com',
url='https://github.com/Shopify/pyactiveresource/',
packages=['pyactiveresource', 'pyactiveresource/testing'],
license='MIT License',
test_suite='test',
tests_require=[
python_dateutils_version,
'PyYAML',
],
platforms=['any'],
classifiers=['Development Status :: 5 - Production/Stable',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Software Development',
'Topic :: Software Development :: Libraries',
'Topic :: Software Development :: Libraries :: Python Modules']
)
| <commit_before>from setuptools import setup
version = '2.0.0'
setup(name='pyactiveresource',
version=version,
description='ActiveResource for Python',
author='Shopify',
author_email='developers@shopify.com',
url='https://github.com/Shopify/pyactiveresource/',
packages=['pyactiveresource', 'pyactiveresource/testing'],
license='MIT License',
test_suite='test',
tests_require=[
'python-dateutil<2.0', # >= 2.0 is for python>=3.0
'PyYAML',
],
platforms=['any'],
classifiers=['Development Status :: 5 - Production/Stable',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Software Development',
'Topic :: Software Development :: Libraries',
'Topic :: Software Development :: Libraries :: Python Modules']
)
<commit_msg>Use the right version of python-dateutils when using python 3.<commit_after> | from setuptools import setup
import sys
version = '2.0.0'
if sys.version_info >= (3,):
python_dateutils_version = 'python-dateutil>=2.0'
else:
python_dateutils_version = 'python-dateutil<2.0'
setup(name='pyactiveresource',
version=version,
description='ActiveResource for Python',
author='Shopify',
author_email='developers@shopify.com',
url='https://github.com/Shopify/pyactiveresource/',
packages=['pyactiveresource', 'pyactiveresource/testing'],
license='MIT License',
test_suite='test',
tests_require=[
python_dateutils_version,
'PyYAML',
],
platforms=['any'],
classifiers=['Development Status :: 5 - Production/Stable',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Software Development',
'Topic :: Software Development :: Libraries',
'Topic :: Software Development :: Libraries :: Python Modules']
)
| from setuptools import setup
version = '2.0.0'
setup(name='pyactiveresource',
version=version,
description='ActiveResource for Python',
author='Shopify',
author_email='developers@shopify.com',
url='https://github.com/Shopify/pyactiveresource/',
packages=['pyactiveresource', 'pyactiveresource/testing'],
license='MIT License',
test_suite='test',
tests_require=[
'python-dateutil<2.0', # >= 2.0 is for python>=3.0
'PyYAML',
],
platforms=['any'],
classifiers=['Development Status :: 5 - Production/Stable',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Software Development',
'Topic :: Software Development :: Libraries',
'Topic :: Software Development :: Libraries :: Python Modules']
)
Use the right version of python-dateutils when using python 3.from setuptools import setup
import sys
version = '2.0.0'
if sys.version_info >= (3,):
python_dateutils_version = 'python-dateutil>=2.0'
else:
python_dateutils_version = 'python-dateutil<2.0'
setup(name='pyactiveresource',
version=version,
description='ActiveResource for Python',
author='Shopify',
author_email='developers@shopify.com',
url='https://github.com/Shopify/pyactiveresource/',
packages=['pyactiveresource', 'pyactiveresource/testing'],
license='MIT License',
test_suite='test',
tests_require=[
python_dateutils_version,
'PyYAML',
],
platforms=['any'],
classifiers=['Development Status :: 5 - Production/Stable',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Software Development',
'Topic :: Software Development :: Libraries',
'Topic :: Software Development :: Libraries :: Python Modules']
)
| <commit_before>from setuptools import setup
version = '2.0.0'
setup(name='pyactiveresource',
version=version,
description='ActiveResource for Python',
author='Shopify',
author_email='developers@shopify.com',
url='https://github.com/Shopify/pyactiveresource/',
packages=['pyactiveresource', 'pyactiveresource/testing'],
license='MIT License',
test_suite='test',
tests_require=[
'python-dateutil<2.0', # >= 2.0 is for python>=3.0
'PyYAML',
],
platforms=['any'],
classifiers=['Development Status :: 5 - Production/Stable',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Software Development',
'Topic :: Software Development :: Libraries',
'Topic :: Software Development :: Libraries :: Python Modules']
)
<commit_msg>Use the right version of python-dateutils when using python 3.<commit_after>from setuptools import setup
import sys
version = '2.0.0'
if sys.version_info >= (3,):
python_dateutils_version = 'python-dateutil>=2.0'
else:
python_dateutils_version = 'python-dateutil<2.0'
setup(name='pyactiveresource',
version=version,
description='ActiveResource for Python',
author='Shopify',
author_email='developers@shopify.com',
url='https://github.com/Shopify/pyactiveresource/',
packages=['pyactiveresource', 'pyactiveresource/testing'],
license='MIT License',
test_suite='test',
tests_require=[
python_dateutils_version,
'PyYAML',
],
platforms=['any'],
classifiers=['Development Status :: 5 - Production/Stable',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Software Development',
'Topic :: Software Development :: Libraries',
'Topic :: Software Development :: Libraries :: Python Modules']
)
|
973473bd7ba0e19e8e86492d4e15b272c441b278 | setup.py | setup.py | """Setuptools configuration for engine."""
from setuptools import setup
from setuptools import find_packages
with open('README.rst', 'r') as readmefile:
README = readmefile.read()
setup(
name='asyncdef.engine',
version='0.1.0',
url='https://github.com/asyncdef/engine',
description='Core event loop implementation.',
author="Kevin Conway",
author_email="kevinjacobconway@gmail.com",
long_description=README,
license='Apache 2.0',
packages=find_packages(exclude=['build', 'dist', 'docs']),
install_requires=[
'iface<2.0.0',
'asyncdef.interfaces<2.0.0',
],
extras_require={
'testing': [
'pep257',
'pep8',
'pyenchant',
'pyflakes',
'pylint',
'pytest',
'pytest-cov',
],
},
entry_points={
'console_scripts': [
],
},
include_package_data=True,
zip_safe=False,
)
| """Setuptools configuration for engine."""
from setuptools import setup
from setuptools import find_packages
with open('README.rst', 'r') as readmefile:
README = readmefile.read()
setup(
name='asyncdef.engine',
version='0.1.0',
url='https://github.com/asyncdef/engine',
description='Core event loop implementation.',
author="Kevin Conway",
author_email="kevinjacobconway@gmail.com",
long_description=README,
license='Apache 2.0',
packages=[
'asyncdef',
'asyncdef.engine',
'asyncdef.engine.processors',
],
install_requires=[
'iface<2.0.0',
'asyncdef.interfaces<2.0.0',
],
extras_require={
'testing': [
'pep257',
'pep8',
'pyenchant',
'pyflakes',
'pylint',
'pytest',
'pytest-cov',
],
},
entry_points={
'console_scripts': [
],
},
include_package_data=True,
zip_safe=False,
)
| Swap find_packages for a manual list | Swap find_packages for a manual list
This is required to support PEP420 namespace packages.
| Python | apache-2.0 | asyncdef/engine | """Setuptools configuration for engine."""
from setuptools import setup
from setuptools import find_packages
with open('README.rst', 'r') as readmefile:
README = readmefile.read()
setup(
name='asyncdef.engine',
version='0.1.0',
url='https://github.com/asyncdef/engine',
description='Core event loop implementation.',
author="Kevin Conway",
author_email="kevinjacobconway@gmail.com",
long_description=README,
license='Apache 2.0',
packages=find_packages(exclude=['build', 'dist', 'docs']),
install_requires=[
'iface<2.0.0',
'asyncdef.interfaces<2.0.0',
],
extras_require={
'testing': [
'pep257',
'pep8',
'pyenchant',
'pyflakes',
'pylint',
'pytest',
'pytest-cov',
],
},
entry_points={
'console_scripts': [
],
},
include_package_data=True,
zip_safe=False,
)
Swap find_packages for a manual list
This is required to support PEP420 namespace packages. | """Setuptools configuration for engine."""
from setuptools import setup
from setuptools import find_packages
with open('README.rst', 'r') as readmefile:
README = readmefile.read()
setup(
name='asyncdef.engine',
version='0.1.0',
url='https://github.com/asyncdef/engine',
description='Core event loop implementation.',
author="Kevin Conway",
author_email="kevinjacobconway@gmail.com",
long_description=README,
license='Apache 2.0',
packages=[
'asyncdef',
'asyncdef.engine',
'asyncdef.engine.processors',
],
install_requires=[
'iface<2.0.0',
'asyncdef.interfaces<2.0.0',
],
extras_require={
'testing': [
'pep257',
'pep8',
'pyenchant',
'pyflakes',
'pylint',
'pytest',
'pytest-cov',
],
},
entry_points={
'console_scripts': [
],
},
include_package_data=True,
zip_safe=False,
)
| <commit_before>"""Setuptools configuration for engine."""
from setuptools import setup
from setuptools import find_packages
with open('README.rst', 'r') as readmefile:
README = readmefile.read()
setup(
name='asyncdef.engine',
version='0.1.0',
url='https://github.com/asyncdef/engine',
description='Core event loop implementation.',
author="Kevin Conway",
author_email="kevinjacobconway@gmail.com",
long_description=README,
license='Apache 2.0',
packages=find_packages(exclude=['build', 'dist', 'docs']),
install_requires=[
'iface<2.0.0',
'asyncdef.interfaces<2.0.0',
],
extras_require={
'testing': [
'pep257',
'pep8',
'pyenchant',
'pyflakes',
'pylint',
'pytest',
'pytest-cov',
],
},
entry_points={
'console_scripts': [
],
},
include_package_data=True,
zip_safe=False,
)
<commit_msg>Swap find_packages for a manual list
This is required to support PEP420 namespace packages.<commit_after> | """Setuptools configuration for engine."""
from setuptools import setup
from setuptools import find_packages
with open('README.rst', 'r') as readmefile:
README = readmefile.read()
setup(
name='asyncdef.engine',
version='0.1.0',
url='https://github.com/asyncdef/engine',
description='Core event loop implementation.',
author="Kevin Conway",
author_email="kevinjacobconway@gmail.com",
long_description=README,
license='Apache 2.0',
packages=[
'asyncdef',
'asyncdef.engine',
'asyncdef.engine.processors',
],
install_requires=[
'iface<2.0.0',
'asyncdef.interfaces<2.0.0',
],
extras_require={
'testing': [
'pep257',
'pep8',
'pyenchant',
'pyflakes',
'pylint',
'pytest',
'pytest-cov',
],
},
entry_points={
'console_scripts': [
],
},
include_package_data=True,
zip_safe=False,
)
| """Setuptools configuration for engine."""
from setuptools import setup
from setuptools import find_packages
with open('README.rst', 'r') as readmefile:
README = readmefile.read()
setup(
name='asyncdef.engine',
version='0.1.0',
url='https://github.com/asyncdef/engine',
description='Core event loop implementation.',
author="Kevin Conway",
author_email="kevinjacobconway@gmail.com",
long_description=README,
license='Apache 2.0',
packages=find_packages(exclude=['build', 'dist', 'docs']),
install_requires=[
'iface<2.0.0',
'asyncdef.interfaces<2.0.0',
],
extras_require={
'testing': [
'pep257',
'pep8',
'pyenchant',
'pyflakes',
'pylint',
'pytest',
'pytest-cov',
],
},
entry_points={
'console_scripts': [
],
},
include_package_data=True,
zip_safe=False,
)
Swap find_packages for a manual list
This is required to support PEP420 namespace packages."""Setuptools configuration for engine."""
from setuptools import setup
from setuptools import find_packages
with open('README.rst', 'r') as readmefile:
README = readmefile.read()
setup(
name='asyncdef.engine',
version='0.1.0',
url='https://github.com/asyncdef/engine',
description='Core event loop implementation.',
author="Kevin Conway",
author_email="kevinjacobconway@gmail.com",
long_description=README,
license='Apache 2.0',
packages=[
'asyncdef',
'asyncdef.engine',
'asyncdef.engine.processors',
],
install_requires=[
'iface<2.0.0',
'asyncdef.interfaces<2.0.0',
],
extras_require={
'testing': [
'pep257',
'pep8',
'pyenchant',
'pyflakes',
'pylint',
'pytest',
'pytest-cov',
],
},
entry_points={
'console_scripts': [
],
},
include_package_data=True,
zip_safe=False,
)
| <commit_before>"""Setuptools configuration for engine."""
from setuptools import setup
from setuptools import find_packages
with open('README.rst', 'r') as readmefile:
README = readmefile.read()
setup(
name='asyncdef.engine',
version='0.1.0',
url='https://github.com/asyncdef/engine',
description='Core event loop implementation.',
author="Kevin Conway",
author_email="kevinjacobconway@gmail.com",
long_description=README,
license='Apache 2.0',
packages=find_packages(exclude=['build', 'dist', 'docs']),
install_requires=[
'iface<2.0.0',
'asyncdef.interfaces<2.0.0',
],
extras_require={
'testing': [
'pep257',
'pep8',
'pyenchant',
'pyflakes',
'pylint',
'pytest',
'pytest-cov',
],
},
entry_points={
'console_scripts': [
],
},
include_package_data=True,
zip_safe=False,
)
<commit_msg>Swap find_packages for a manual list
This is required to support PEP420 namespace packages.<commit_after>"""Setuptools configuration for engine."""
from setuptools import setup
from setuptools import find_packages
with open('README.rst', 'r') as readmefile:
README = readmefile.read()
setup(
name='asyncdef.engine',
version='0.1.0',
url='https://github.com/asyncdef/engine',
description='Core event loop implementation.',
author="Kevin Conway",
author_email="kevinjacobconway@gmail.com",
long_description=README,
license='Apache 2.0',
packages=[
'asyncdef',
'asyncdef.engine',
'asyncdef.engine.processors',
],
install_requires=[
'iface<2.0.0',
'asyncdef.interfaces<2.0.0',
],
extras_require={
'testing': [
'pep257',
'pep8',
'pyenchant',
'pyflakes',
'pylint',
'pytest',
'pytest-cov',
],
},
entry_points={
'console_scripts': [
],
},
include_package_data=True,
zip_safe=False,
)
|
5acb2258164ce5f8d7c8c297b32534547e861450 | lms/djangoapps/debug/views.py | lms/djangoapps/debug/views.py | """Views for debugging and diagnostics"""
import pprint
from django.http import Http404
from django.contrib.auth.decorators import login_required
from django_future.csrf import ensure_csrf_cookie, csrf_exempt
from mitxmako.shortcuts import render_to_response
from codejail.safe_exec import safe_exec
@login_required
@ensure_csrf_cookie
def run_python(request):
if not request.user.is_staff:
raise Http404
c = {}
c['code'] = ''
c['results'] = None
if request.method == 'POST':
py_code = c['code'] = request.POST.get('code')
g = {}
try:
safe_exec(py_code, g)
except Exception as e:
c['results'] = str(e)
else:
c['results'] = pprint.pformat(g)
return render_to_response("debug/run_python_form.html", c)
| """Views for debugging and diagnostics"""
import pprint
import traceback
from django.http import Http404
from django.contrib.auth.decorators import login_required
from django_future.csrf import ensure_csrf_cookie, csrf_exempt
from mitxmako.shortcuts import render_to_response
from codejail.safe_exec import safe_exec
@login_required
@ensure_csrf_cookie
def run_python(request):
"""A page to allow testing the Python sandbox on a production server."""
if not request.user.is_staff:
raise Http404
c = {}
c['code'] = ''
c['results'] = None
if request.method == 'POST':
py_code = c['code'] = request.POST.get('code')
g = {}
try:
safe_exec(py_code, g)
except Exception as e:
c['results'] = traceback.format_exc()
else:
c['results'] = pprint.pformat(g)
return render_to_response("debug/run_python_form.html", c)
| Print the full traceback when execution fails. | Print the full traceback when execution fails.
| Python | agpl-3.0 | cpennington/edx-platform,Shrhawk/edx-platform,chrisndodge/edx-platform,xinjiguaike/edx-platform,leansoft/edx-platform,jamiefolsom/edx-platform,chand3040/cloud_that,nttks/jenkins-test,PepperPD/edx-pepper-platform,edry/edx-platform,naresh21/synergetics-edx-platform,doganov/edx-platform,adoosii/edx-platform,eduNEXT/edx-platform,analyseuc3m/ANALYSE-v1,shubhdev/edxOnBaadal,vikas1885/test1,motion2015/edx-platform,ubc/edx-platform,benpatterson/edx-platform,JioEducation/edx-platform,praveen-pal/edx-platform,olexiim/edx-platform,xuxiao19910803/edx,carsongee/edx-platform,shurihell/testasia,lduarte1991/edx-platform,waheedahmed/edx-platform,kxliugang/edx-platform,dsajkl/123,Lektorium-LLC/edx-platform,ferabra/edx-platform,nanolearningllc/edx-platform-cypress-2,nikolas/edx-platform,amir-qayyum-khan/edx-platform,torchingloom/edx-platform,TeachAtTUM/edx-platform,Semi-global/edx-platform,mushtaqak/edx-platform,mcgachey/edx-platform,ampax/edx-platform,OmarIthawi/edx-platform,10clouds/edx-platform,nanolearningllc/edx-platform-cypress,jamiefolsom/edx-platform,procangroup/edx-platform,Edraak/edx-platform,shashank971/edx-platform,yokose-ks/edx-platform,hkawasaki/kawasaki-aio8-1,simbs/edx-platform,jbzdak/edx-platform,nttks/edx-platform,utecuy/edx-platform,jazztpt/edx-platform,msegado/edx-platform,philanthropy-u/edx-platform,ahmedaljazzar/edx-platform,DefyVentures/edx-platform,ZLLab-Mooc/edx-platform,hkawasaki/kawasaki-aio8-2,hamzehd/edx-platform,jswope00/griffinx,cecep-edu/edx-platform,chand3040/cloud_that,bigdatauniversity/edx-platform,shubhdev/edx-platform,OmarIthawi/edx-platform,zerobatu/edx-platform,auferack08/edx-platform,Shrhawk/edx-platform,alexthered/kienhoc-platform,jazkarta/edx-platform,LICEF/edx-platform,waheedahmed/edx-platform,amir-qayyum-khan/edx-platform,wwj718/ANALYSE,adoosii/edx-platform,sudheerchintala/LearnEraPlatForm,etzhou/edx-platform,eestay/edx-platform,zerobatu/edx-platform,Edraak/edraak-platform,ferabra/edx-platform,tiagochiavericosta/edx-platform,OmarIthawi/edx-platform,vismartltd/edx-platform,EduPepperPDTesting/pepper2013-testing,utecuy/edx-platform,nanolearningllc/edx-platform-cypress,4eek/edx-platform,zadgroup/edx-platform,stvstnfrd/edx-platform,solashirai/edx-platform,caesar2164/edx-platform,CourseTalk/edx-platform,Livit/Livit.Learn.EdX,EduPepperPD/pepper2013,eduNEXT/edunext-platform,rue89-tech/edx-platform,pepeportela/edx-platform,antoviaque/edx-platform,J861449197/edx-platform,raccoongang/edx-platform,cselis86/edx-platform,MSOpenTech/edx-platform,zerobatu/edx-platform,eduNEXT/edx-platform,eduNEXT/edunext-platform,synergeticsedx/deployment-wipro,nttks/edx-platform,olexiim/edx-platform,nagyistoce/edx-platform,morpheby/levelup-by,tanmaykm/edx-platform,torchingloom/edx-platform,gymnasium/edx-platform,gsehub/edx-platform,chand3040/cloud_that,cognitiveclass/edx-platform,DefyVentures/edx-platform,fly19890211/edx-platform,jruiperezv/ANALYSE,jazkarta/edx-platform,mjirayu/sit_academy,hmcmooc/muddx-platform,franosincic/edx-platform,shashank971/edx-platform,rhndg/openedx,tanmaykm/edx-platform,mbareta/edx-platform-ft,auferack08/edx-platform,eestay/edx-platform,cognitiveclass/edx-platform,carsongee/edx-platform,martynovp/edx-platform,zadgroup/edx-platform,bdero/edx-platform,Unow/edx-platform,Livit/Livit.Learn.EdX,hkawasaki/kawasaki-aio8-1,beni55/edx-platform,B-MOOC/edx-platform,rationalAgent/edx-platform-custom,playm2mboy/edx-platform,fintech-circle/edx-platform,edry/edx-platform,Edraak/edx-platform,devs1991/test_edx_docmode,atsolakid/edx-platform,morenopc/edx-platform,Lektorium-LLC/edx-platform,proversity-org/edx-platform,rue89-tech/edx-platform,knehez/edx-platform,zerobatu/edx-platform,ahmadio/edx-platform,mahendra-r/edx-platform,don-github/edx-platform,Edraak/circleci-edx-platform,LearnEra/LearnEraPlaftform,ampax/edx-platform-backup,zubair-arbi/edx-platform,UXE/local-edx,olexiim/edx-platform,defance/edx-platform,prarthitm/edxplatform,franosincic/edx-platform,morpheby/levelup-by,defance/edx-platform,shashank971/edx-platform,miptliot/edx-platform,eemirtekin/edx-platform,martynovp/edx-platform,shabab12/edx-platform,Kalyzee/edx-platform,eduNEXT/edx-platform,10clouds/edx-platform,sameetb-cuelogic/edx-platform-test,arifsetiawan/edx-platform,caesar2164/edx-platform,zofuthan/edx-platform,beacloudgenius/edx-platform,IndonesiaX/edx-platform,defance/edx-platform,vasyarv/edx-platform,LICEF/edx-platform,nttks/edx-platform,RPI-OPENEDX/edx-platform,edx-solutions/edx-platform,eemirtekin/edx-platform,xuxiao19910803/edx,pepeportela/edx-platform,Endika/edx-platform,arifsetiawan/edx-platform,leansoft/edx-platform,antonve/s4-project-mooc,yokose-ks/edx-platform,nanolearning/edx-platform,EduPepperPD/pepper2013,DNFcode/edx-platform,alexthered/kienhoc-platform,nikolas/edx-platform,chauhanhardik/populo_2,itsjeyd/edx-platform,procangroup/edx-platform,devs1991/test_edx_docmode,dkarakats/edx-platform,PepperPD/edx-pepper-platform,yokose-ks/edx-platform,ESOedX/edx-platform,solashirai/edx-platform,pabloborrego93/edx-platform,playm2mboy/edx-platform,peterm-itr/edx-platform,doismellburning/edx-platform,valtech-mooc/edx-platform,DNFcode/edx-platform,lduarte1991/edx-platform,mjirayu/sit_academy,ahmadiga/min_edx,BehavioralInsightsTeam/edx-platform,nanolearningllc/edx-platform-cypress-2,motion2015/a3,ampax/edx-platform-backup,ovnicraft/edx-platform,jswope00/GAI,atsolakid/edx-platform,zofuthan/edx-platform,tiagochiavericosta/edx-platform,jbzdak/edx-platform,xuxiao19910803/edx-platform,jolyonb/edx-platform,fintech-circle/edx-platform,RPI-OPENEDX/edx-platform,IONISx/edx-platform,J861449197/edx-platform,antonve/s4-project-mooc,ovnicraft/edx-platform,chand3040/cloud_that,mahendra-r/edx-platform,marcore/edx-platform,mitocw/edx-platform,hkawasaki/kawasaki-aio8-0,Unow/edx-platform,jruiperezv/ANALYSE,ubc/edx-platform,longmen21/edx-platform,shurihell/testasia,wwj718/ANALYSE,appsembler/edx-platform,morenopc/edx-platform,bigdatauniversity/edx-platform,syjeon/new_edx,eemirtekin/edx-platform,ahmadio/edx-platform,chauhanhardik/populo,fintech-circle/edx-platform,IITBinterns13/edx-platform-dev,kalebhartje/schoolboost,jamesblunt/edx-platform,nttks/edx-platform,bdero/edx-platform,B-MOOC/edx-platform,don-github/edx-platform,MakeHer/edx-platform,arbrandes/edx-platform,IndonesiaX/edx-platform,simbs/edx-platform,apigee/edx-platform,teltek/edx-platform,shubhdev/openedx,zofuthan/edx-platform,lduarte1991/edx-platform,AkA84/edx-platform,martynovp/edx-platform,Softmotions/edx-platform,defance/edx-platform,miptliot/edx-platform,SravanthiSinha/edx-platform,nikolas/edx-platform,kursitet/edx-platform,nanolearningllc/edx-platform-cypress-2,cyanna/edx-platform,cecep-edu/edx-platform,mahendra-r/edx-platform,abdoosh00/edx-rtl-final,ferabra/edx-platform,kxliugang/edx-platform,dkarakats/edx-platform,benpatterson/edx-platform,raccoongang/edx-platform,dcosentino/edx-platform,jbassen/edx-platform,mtlchun/edx,iivic/BoiseStateX,bitifirefly/edx-platform,ahmadio/edx-platform,mbareta/edx-platform-ft,chrisndodge/edx-platform,kursitet/edx-platform,bitifirefly/edx-platform,mtlchun/edx,chudaol/edx-platform,rue89-tech/edx-platform,mjg2203/edx-platform-seas,ubc/edx-platform,LearnEra/LearnEraPlaftform,10clouds/edx-platform,longmen21/edx-platform,marcore/edx-platform,carsongee/edx-platform,pdehaye/theming-edx-platform,Unow/edx-platform,chand3040/cloud_that,TsinghuaX/edx-platform,kamalx/edx-platform,WatanabeYasumasa/edx-platform,benpatterson/edx-platform,pdehaye/theming-edx-platform,beacloudgenius/edx-platform,gsehub/edx-platform,hkawasaki/kawasaki-aio8-1,syjeon/new_edx,xuxiao19910803/edx-platform,jswope00/GAI,proversity-org/edx-platform,PepperPD/edx-pepper-platform,EduPepperPD/pepper2013,playm2mboy/edx-platform,torchingloom/edx-platform,fly19890211/edx-platform,chauhanhardik/populo_2,ak2703/edx-platform,dsajkl/reqiop,ampax/edx-platform,mitocw/edx-platform,shubhdev/openedx,hmcmooc/muddx-platform,RPI-OPENEDX/edx-platform,mahendra-r/edx-platform,dsajkl/123,chauhanhardik/populo,rhndg/openedx,Softmotions/edx-platform,halvertoluke/edx-platform,miptliot/edx-platform,Edraak/circleci-edx-platform,rhndg/openedx,mcgachey/edx-platform,jazztpt/edx-platform,Ayub-Khan/edx-platform,pku9104038/edx-platform,vasyarv/edx-platform,ampax/edx-platform,jelugbo/tundex,y12uc231/edx-platform,chauhanhardik/populo,jamesblunt/edx-platform,pku9104038/edx-platform,kalebhartje/schoolboost,hamzehd/edx-platform,pelikanchik/edx-platform,WatanabeYasumasa/edx-platform,UOMx/edx-platform,DNFcode/edx-platform,ZLLab-Mooc/edx-platform,appsembler/edx-platform,jelugbo/tundex,longmen21/edx-platform,ovnicraft/edx-platform,Shrhawk/edx-platform,pomegranited/edx-platform,romain-li/edx-platform,simbs/edx-platform,philanthropy-u/edx-platform,doismellburning/edx-platform,wwj718/ANALYSE,IONISx/edx-platform,praveen-pal/edx-platform,franosincic/edx-platform,MakeHer/edx-platform,unicri/edx-platform,Softmotions/edx-platform,xuxiao19910803/edx-platform,IONISx/edx-platform,SivilTaram/edx-platform,mushtaqak/edx-platform,eestay/edx-platform,zhenzhai/edx-platform,waheedahmed/edx-platform,apigee/edx-platform,JCBarahona/edX,stvstnfrd/edx-platform,dsajkl/123,deepsrijit1105/edx-platform,polimediaupv/edx-platform,jzoldak/edx-platform,Semi-global/edx-platform,zubair-arbi/edx-platform,ampax/edx-platform-backup,pabloborrego93/edx-platform,ZLLab-Mooc/edx-platform,ovnicraft/edx-platform,chudaol/edx-platform,arbrandes/edx-platform,chauhanhardik/populo,Edraak/edx-platform,cecep-edu/edx-platform,ahmadiga/min_edx,Kalyzee/edx-platform,mtlchun/edx,etzhou/edx-platform,mjirayu/sit_academy,JCBarahona/edX,UXE/local-edx,antoviaque/edx-platform,jazztpt/edx-platform,hkawasaki/kawasaki-aio8-0,shubhdev/openedx,EduPepperPDTesting/pepper2013-testing,Edraak/edraak-platform,fly19890211/edx-platform,bdero/edx-platform,UOMx/edx-platform,AkA84/edx-platform,DefyVentures/edx-platform,fly19890211/edx-platform,RPI-OPENEDX/edx-platform,beni55/edx-platform,analyseuc3m/ANALYSE-v1,kmoocdev2/edx-platform,J861449197/edx-platform,cselis86/edx-platform,abdoosh00/edraak,antoviaque/edx-platform,PepperPD/edx-pepper-platform,shubhdev/openedx,mbareta/edx-platform-ft,TsinghuaX/edx-platform,morenopc/edx-platform,edx-solutions/edx-platform,kmoocdev/edx-platform,pepeportela/edx-platform,peterm-itr/edx-platform,halvertoluke/edx-platform,valtech-mooc/edx-platform,mjg2203/edx-platform-seas,xinjiguaike/edx-platform,angelapper/edx-platform,openfun/edx-platform,jruiperezv/ANALYSE,olexiim/edx-platform,CredoReference/edx-platform,auferack08/edx-platform,edx/edx-platform,IndonesiaX/edx-platform,shurihell/testasia,SravanthiSinha/edx-platform,msegado/edx-platform,jjmiranda/edx-platform,bitifirefly/edx-platform,JioEducation/edx-platform,mtlchun/edx,4eek/edx-platform,kxliugang/edx-platform,kmoocdev/edx-platform,Endika/edx-platform,y12uc231/edx-platform,vikas1885/test1,jazkarta/edx-platform,dsajkl/123,yokose-ks/edx-platform,appsembler/edx-platform,cognitiveclass/edx-platform,analyseuc3m/ANALYSE-v1,arbrandes/edx-platform,cecep-edu/edx-platform,caesar2164/edx-platform,jolyonb/edx-platform,Endika/edx-platform,dsajkl/reqiop,wwj718/edx-platform,dkarakats/edx-platform,WatanabeYasumasa/edx-platform,cselis86/edx-platform,chauhanhardik/populo,nagyistoce/edx-platform,torchingloom/edx-platform,shashank971/edx-platform,jonathan-beard/edx-platform,shurihell/testasia,Stanford-Online/edx-platform,rismalrv/edx-platform,auferack08/edx-platform,jswope00/griffinx,valtech-mooc/edx-platform,motion2015/edx-platform,stvstnfrd/edx-platform,stvstnfrd/edx-platform,doismellburning/edx-platform,xingyepei/edx-platform,longmen21/edx-platform,openfun/edx-platform,xinjiguaike/edx-platform,rationalAgent/edx-platform-custom,cselis86/edx-platform,JCBarahona/edX,kxliugang/edx-platform,10clouds/edx-platform,amir-qayyum-khan/edx-platform,syjeon/new_edx,zadgroup/edx-platform,pelikanchik/edx-platform,ahmedaljazzar/edx-platform,ferabra/edx-platform,prarthitm/edxplatform,sudheerchintala/LearnEraPlatForm,hmcmooc/muddx-platform,dcosentino/edx-platform,wwj718/edx-platform,CourseTalk/edx-platform,romain-li/edx-platform,louyihua/edx-platform,jazkarta/edx-platform,nanolearning/edx-platform,AkA84/edx-platform,yokose-ks/edx-platform,ahmadio/edx-platform,nanolearning/edx-platform,pelikanchik/edx-platform,gymnasium/edx-platform,solashirai/edx-platform,Ayub-Khan/edx-platform,polimediaupv/edx-platform,xinjiguaike/edx-platform,BehavioralInsightsTeam/edx-platform,EduPepperPDTesting/pepper2013-testing,cselis86/edx-platform,jbzdak/edx-platform,romain-li/edx-platform,motion2015/a3,hkawasaki/kawasaki-aio8-0,jelugbo/tundex,EduPepperPD/pepper2013,doganov/edx-platform,vikas1885/test1,kmoocdev2/edx-platform,wwj718/ANALYSE,knehez/edx-platform,IITBinterns13/edx-platform-dev,IndonesiaX/edx-platform,don-github/edx-platform,xuxiao19910803/edx-platform,y12uc231/edx-platform,JCBarahona/edX,ubc/edx-platform,J861449197/edx-platform,abdoosh00/edraak,Semi-global/edx-platform,tiagochiavericosta/edx-platform,SivilTaram/edx-platform,DNFcode/edx-platform,edry/edx-platform,franosincic/edx-platform,edx-solutions/edx-platform,jbassen/edx-platform,bigdatauniversity/edx-platform,EDUlib/edx-platform,abdoosh00/edraak,pomegranited/edx-platform,jswope00/griffinx,Lektorium-LLC/edx-platform,raccoongang/edx-platform,MSOpenTech/edx-platform,sameetb-cuelogic/edx-platform-test,iivic/BoiseStateX,xingyepei/edx-platform,beacloudgenius/edx-platform,etzhou/edx-platform,tiagochiavericosta/edx-platform,jonathan-beard/edx-platform,apigee/edx-platform,MSOpenTech/edx-platform,simbs/edx-platform,ampax/edx-platform-backup,arbrandes/edx-platform,devs1991/test_edx_docmode,MakeHer/edx-platform,deepsrijit1105/edx-platform,UXE/local-edx,Livit/Livit.Learn.EdX,ZLLab-Mooc/edx-platform,ferabra/edx-platform,eduNEXT/edunext-platform,proversity-org/edx-platform,angelapper/edx-platform,CourseTalk/edx-platform,leansoft/edx-platform,alexthered/kienhoc-platform,chudaol/edx-platform,angelapper/edx-platform,mitocw/edx-platform,wwj718/ANALYSE,morenopc/edx-platform,iivic/BoiseStateX,kxliugang/edx-platform,prarthitm/edxplatform,pabloborrego93/edx-platform,jazkarta/edx-platform-for-isc,romain-li/edx-platform,xingyepei/edx-platform,miptliot/edx-platform,morenopc/edx-platform,kmoocdev/edx-platform,devs1991/test_edx_docmode,peterm-itr/edx-platform,TeachAtTUM/edx-platform,shashank971/edx-platform,jonathan-beard/edx-platform,DNFcode/edx-platform,jbassen/edx-platform,knehez/edx-platform,jazztpt/edx-platform,solashirai/edx-platform,Unow/edx-platform,ovnicraft/edx-platform,cyanna/edx-platform,zofuthan/edx-platform,doganov/edx-platform,ak2703/edx-platform,antonve/s4-project-mooc,kalebhartje/schoolboost,edx-solutions/edx-platform,B-MOOC/edx-platform,rue89-tech/edx-platform,nttks/jenkins-test,Kalyzee/edx-platform,motion2015/a3,hastexo/edx-platform,EDUlib/edx-platform,DefyVentures/edx-platform,don-github/edx-platform,jazkarta/edx-platform-for-isc,unicri/edx-platform,mushtaqak/edx-platform,eduNEXT/edx-platform,Stanford-Online/edx-platform,cpennington/edx-platform,RPI-OPENEDX/edx-platform,IONISx/edx-platform,zerobatu/edx-platform,chauhanhardik/populo_2,synergeticsedx/deployment-wipro,alexthered/kienhoc-platform,shabab12/edx-platform,arifsetiawan/edx-platform,jjmiranda/edx-platform,synergeticsedx/deployment-wipro,jonathan-beard/edx-platform,zadgroup/edx-platform,unicri/edx-platform,deepsrijit1105/edx-platform,beacloudgenius/edx-platform,adoosii/edx-platform,appliedx/edx-platform,J861449197/edx-platform,zhenzhai/edx-platform,mjg2203/edx-platform-seas,nagyistoce/edx-platform,zofuthan/edx-platform,bigdatauniversity/edx-platform,dkarakats/edx-platform,kamalx/edx-platform,devs1991/test_edx_docmode,abdoosh00/edx-rtl-final,Lektorium-LLC/edx-platform,Edraak/circleci-edx-platform,pdehaye/theming-edx-platform,jolyonb/edx-platform,halvertoluke/edx-platform,jzoldak/edx-platform,ampax/edx-platform-backup,shubhdev/edxOnBaadal,cognitiveclass/edx-platform,chudaol/edx-platform,kmoocdev2/edx-platform,mitocw/edx-platform,Semi-global/edx-platform,utecuy/edx-platform,Kalyzee/edx-platform,gymnasium/edx-platform,edry/edx-platform,a-parhom/edx-platform,nttks/edx-platform,polimediaupv/edx-platform,fly19890211/edx-platform,procangroup/edx-platform,jswope00/griffinx,devs1991/test_edx_docmode,vasyarv/edx-platform,msegado/edx-platform,knehez/edx-platform,jazztpt/edx-platform,vismartltd/edx-platform,DefyVentures/edx-platform,alu042/edx-platform,atsolakid/edx-platform,naresh21/synergetics-edx-platform,Shrhawk/edx-platform,rationalAgent/edx-platform-custom,hastexo/edx-platform,marcore/edx-platform,chrisndodge/edx-platform,cyanna/edx-platform,MakeHer/edx-platform,nikolas/edx-platform,pdehaye/theming-edx-platform,jazkarta/edx-platform-for-isc,rismalrv/edx-platform,beni55/edx-platform,JCBarahona/edX,BehavioralInsightsTeam/edx-platform,LearnEra/LearnEraPlaftform,angelapper/edx-platform,syjeon/new_edx,doganov/edx-platform,torchingloom/edx-platform,valtech-mooc/edx-platform,unicri/edx-platform,a-parhom/edx-platform,motion2015/edx-platform,cyanna/edx-platform,philanthropy-u/edx-platform,ESOedX/edx-platform,morpheby/levelup-by,alexthered/kienhoc-platform,EduPepperPD/pepper2013,rhndg/openedx,dcosentino/edx-platform,analyseuc3m/ANALYSE-v1,dkarakats/edx-platform,EduPepperPDTesting/pepper2013-testing,longmen21/edx-platform,LICEF/edx-platform,zadgroup/edx-platform,jazkarta/edx-platform-for-isc,kamalx/edx-platform,wwj718/edx-platform,motion2015/edx-platform,bdero/edx-platform,kursitet/edx-platform,kmoocdev/edx-platform,chrisndodge/edx-platform,inares/edx-platform,prarthitm/edxplatform,sudheerchintala/LearnEraPlatForm,Kalyzee/edx-platform,beni55/edx-platform,kursitet/edx-platform,SravanthiSinha/edx-platform,IONISx/edx-platform,antonve/s4-project-mooc,mushtaqak/edx-platform,alu042/edx-platform,mcgachey/edx-platform,appliedx/edx-platform,procangroup/edx-platform,xingyepei/edx-platform,shubhdev/edxOnBaadal,WatanabeYasumasa/edx-platform,shubhdev/edx-platform,jruiperezv/ANALYSE,mjg2203/edx-platform-seas,amir-qayyum-khan/edx-platform,Softmotions/edx-platform,praveen-pal/edx-platform,jamiefolsom/edx-platform,Ayub-Khan/edx-platform,jamesblunt/edx-platform,halvertoluke/edx-platform,MSOpenTech/edx-platform,waheedahmed/edx-platform,caesar2164/edx-platform,alu042/edx-platform,iivic/BoiseStateX,louyihua/edx-platform,zhenzhai/edx-platform,antoviaque/edx-platform,jbassen/edx-platform,Edraak/circleci-edx-platform,BehavioralInsightsTeam/edx-platform,itsjeyd/edx-platform,leansoft/edx-platform,martynovp/edx-platform,AkA84/edx-platform,beni55/edx-platform,gsehub/edx-platform,abdoosh00/edx-rtl-final,louyihua/edx-platform,doismellburning/edx-platform,jelugbo/tundex,rhndg/openedx,LearnEra/LearnEraPlaftform,vikas1885/test1,ZLLab-Mooc/edx-platform,IITBinterns13/edx-platform-dev,EduPepperPDTesting/pepper2013-testing,PepperPD/edx-pepper-platform,simbs/edx-platform,peterm-itr/edx-platform,hmcmooc/muddx-platform,shabab12/edx-platform,appsembler/edx-platform,ahmedaljazzar/edx-platform,gsehub/edx-platform,jjmiranda/edx-platform,Semi-global/edx-platform,xinjiguaike/edx-platform,devs1991/test_edx_docmode,shubhdev/edx-platform,vikas1885/test1,TsinghuaX/edx-platform,ahmadiga/min_edx,bitifirefly/edx-platform,JioEducation/edx-platform,arifsetiawan/edx-platform,abdoosh00/edx-rtl-final,nanolearning/edx-platform,jruiperezv/ANALYSE,synergeticsedx/deployment-wipro,vismartltd/edx-platform,LICEF/edx-platform,naresh21/synergetics-edx-platform,valtech-mooc/edx-platform,ESOedX/edx-platform,utecuy/edx-platform,Ayub-Khan/edx-platform,motion2015/a3,sudheerchintala/LearnEraPlatForm,hamzehd/edx-platform,rismalrv/edx-platform,edx/edx-platform,SivilTaram/edx-platform,kmoocdev2/edx-platform,shubhdev/edxOnBaadal,vismartltd/edx-platform,philanthropy-u/edx-platform,Livit/Livit.Learn.EdX,EDUlib/edx-platform,motion2015/a3,benpatterson/edx-platform,iivic/BoiseStateX,adoosii/edx-platform,tiagochiavericosta/edx-platform,ahmadiga/min_edx,xuxiao19910803/edx,inares/edx-platform,ak2703/edx-platform,polimediaupv/edx-platform,waheedahmed/edx-platform,teltek/edx-platform,mjirayu/sit_academy,etzhou/edx-platform,UXE/local-edx,shabab12/edx-platform,leansoft/edx-platform,ahmadiga/min_edx,pomegranited/edx-platform,jelugbo/tundex,zhenzhai/edx-platform,y12uc231/edx-platform,ahmedaljazzar/edx-platform,fintech-circle/edx-platform,jazkarta/edx-platform,hkawasaki/kawasaki-aio8-2,SravanthiSinha/edx-platform,lduarte1991/edx-platform,franosincic/edx-platform,UOMx/edx-platform,kalebhartje/schoolboost,chudaol/edx-platform,pepeportela/edx-platform,rationalAgent/edx-platform-custom,hkawasaki/kawasaki-aio8-2,Endika/edx-platform,hkawasaki/kawasaki-aio8-0,mtlchun/edx,itsjeyd/edx-platform,jswope00/GAI,apigee/edx-platform,shubhdev/edxOnBaadal,msegado/edx-platform,jswope00/griffinx,shubhdev/edx-platform,UOMx/edx-platform,tanmaykm/edx-platform,openfun/edx-platform,polimediaupv/edx-platform,B-MOOC/edx-platform,wwj718/edx-platform,rationalAgent/edx-platform-custom,vasyarv/edx-platform,shubhdev/edx-platform,edry/edx-platform,martynovp/edx-platform,TsinghuaX/edx-platform,LICEF/edx-platform,pabloborrego93/edx-platform,hkawasaki/kawasaki-aio8-1,TeachAtTUM/edx-platform,xingyepei/edx-platform,Shrhawk/edx-platform,pomegranited/edx-platform,carsongee/edx-platform,Edraak/edx-platform,jjmiranda/edx-platform,sameetb-cuelogic/edx-platform-test,xuxiao19910803/edx,ampax/edx-platform,devs1991/test_edx_docmode,y12uc231/edx-platform,CredoReference/edx-platform,dsajkl/123,alu042/edx-platform,andyzsf/edx,inares/edx-platform,ubc/edx-platform,openfun/edx-platform,nanolearning/edx-platform,praveen-pal/edx-platform,pku9104038/edx-platform,cyanna/edx-platform,deepsrijit1105/edx-platform,Edraak/edraak-platform,rue89-tech/edx-platform,benpatterson/edx-platform,appliedx/edx-platform,itsjeyd/edx-platform,cpennington/edx-platform,ak2703/edx-platform,a-parhom/edx-platform,B-MOOC/edx-platform,jamesblunt/edx-platform,jamesblunt/edx-platform,eestay/edx-platform,arifsetiawan/edx-platform,edx/edx-platform,nttks/jenkins-test,chauhanhardik/populo_2,eduNEXT/edunext-platform,kmoocdev2/edx-platform,naresh21/synergetics-edx-platform,andyzsf/edx,dsajkl/reqiop,edx/edx-platform,CourseTalk/edx-platform,andyzsf/edx,nttks/jenkins-test,vismartltd/edx-platform,dcosentino/edx-platform,jbzdak/edx-platform,AkA84/edx-platform,cpennington/edx-platform,teltek/edx-platform,Stanford-Online/edx-platform,doganov/edx-platform,antonve/s4-project-mooc,abdoosh00/edraak,mjirayu/sit_academy,4eek/edx-platform,chauhanhardik/populo_2,cognitiveclass/edx-platform,kmoocdev/edx-platform,nanolearningllc/edx-platform-cypress,inares/edx-platform,kamalx/edx-platform,jbzdak/edx-platform,EduPepperPDTesting/pepper2013-testing,doismellburning/edx-platform,jbassen/edx-platform,jamiefolsom/edx-platform,kursitet/edx-platform,knehez/edx-platform,teltek/edx-platform,CredoReference/edx-platform,vasyarv/edx-platform,louyihua/edx-platform,EDUlib/edx-platform,hamzehd/edx-platform,hamzehd/edx-platform,atsolakid/edx-platform,shurihell/testasia,mushtaqak/edx-platform,nanolearningllc/edx-platform-cypress-2,pomegranited/edx-platform,MSOpenTech/edx-platform,sameetb-cuelogic/edx-platform-test,Stanford-Online/edx-platform,beacloudgenius/edx-platform,jolyonb/edx-platform,Edraak/edx-platform,solashirai/edx-platform,tanmaykm/edx-platform,CredoReference/edx-platform,4eek/edx-platform,utecuy/edx-platform,jonathan-beard/edx-platform,jzoldak/edx-platform,SivilTaram/edx-platform,Ayub-Khan/edx-platform,jamiefolsom/edx-platform,andyzsf/edx,hastexo/edx-platform,Edraak/edraak-platform,halvertoluke/edx-platform,nanolearningllc/edx-platform-cypress-2,adoosii/edx-platform,don-github/edx-platform,dsajkl/reqiop,marcore/edx-platform,hkawasaki/kawasaki-aio8-2,appliedx/edx-platform,nagyistoce/edx-platform,romain-li/edx-platform,playm2mboy/edx-platform,mcgachey/edx-platform,wwj718/edx-platform,mahendra-r/edx-platform,cecep-edu/edx-platform,proversity-org/edx-platform,jswope00/GAI,rismalrv/edx-platform,nttks/jenkins-test,bitifirefly/edx-platform,raccoongang/edx-platform,ahmadio/edx-platform,IndonesiaX/edx-platform,appliedx/edx-platform,inares/edx-platform,bigdatauniversity/edx-platform,morpheby/levelup-by,IITBinterns13/edx-platform-dev,sameetb-cuelogic/edx-platform-test,kamalx/edx-platform,MakeHer/edx-platform,xuxiao19910803/edx,zubair-arbi/edx-platform,xuxiao19910803/edx-platform,dcosentino/edx-platform,rismalrv/edx-platform,gymnasium/edx-platform,playm2mboy/edx-platform,Edraak/circleci-edx-platform,a-parhom/edx-platform,SravanthiSinha/edx-platform,zubair-arbi/edx-platform,JioEducation/edx-platform,TeachAtTUM/edx-platform,pelikanchik/edx-platform,openfun/edx-platform,nanolearningllc/edx-platform-cypress,nanolearningllc/edx-platform-cypress,etzhou/edx-platform,jazkarta/edx-platform-for-isc,4eek/edx-platform,kalebhartje/schoolboost,mcgachey/edx-platform,pku9104038/edx-platform,atsolakid/edx-platform,mbareta/edx-platform-ft,msegado/edx-platform,Softmotions/edx-platform,eemirtekin/edx-platform,nikolas/edx-platform,hastexo/edx-platform,eestay/edx-platform,ESOedX/edx-platform,OmarIthawi/edx-platform,zhenzhai/edx-platform,nagyistoce/edx-platform,shubhdev/openedx,olexiim/edx-platform,eemirtekin/edx-platform,SivilTaram/edx-platform,motion2015/edx-platform,unicri/edx-platform,zubair-arbi/edx-platform,jzoldak/edx-platform,ak2703/edx-platform | """Views for debugging and diagnostics"""
import pprint
from django.http import Http404
from django.contrib.auth.decorators import login_required
from django_future.csrf import ensure_csrf_cookie, csrf_exempt
from mitxmako.shortcuts import render_to_response
from codejail.safe_exec import safe_exec
@login_required
@ensure_csrf_cookie
def run_python(request):
if not request.user.is_staff:
raise Http404
c = {}
c['code'] = ''
c['results'] = None
if request.method == 'POST':
py_code = c['code'] = request.POST.get('code')
g = {}
try:
safe_exec(py_code, g)
except Exception as e:
c['results'] = str(e)
else:
c['results'] = pprint.pformat(g)
return render_to_response("debug/run_python_form.html", c)
Print the full traceback when execution fails. | """Views for debugging and diagnostics"""
import pprint
import traceback
from django.http import Http404
from django.contrib.auth.decorators import login_required
from django_future.csrf import ensure_csrf_cookie, csrf_exempt
from mitxmako.shortcuts import render_to_response
from codejail.safe_exec import safe_exec
@login_required
@ensure_csrf_cookie
def run_python(request):
"""A page to allow testing the Python sandbox on a production server."""
if not request.user.is_staff:
raise Http404
c = {}
c['code'] = ''
c['results'] = None
if request.method == 'POST':
py_code = c['code'] = request.POST.get('code')
g = {}
try:
safe_exec(py_code, g)
except Exception as e:
c['results'] = traceback.format_exc()
else:
c['results'] = pprint.pformat(g)
return render_to_response("debug/run_python_form.html", c)
| <commit_before>"""Views for debugging and diagnostics"""
import pprint
from django.http import Http404
from django.contrib.auth.decorators import login_required
from django_future.csrf import ensure_csrf_cookie, csrf_exempt
from mitxmako.shortcuts import render_to_response
from codejail.safe_exec import safe_exec
@login_required
@ensure_csrf_cookie
def run_python(request):
if not request.user.is_staff:
raise Http404
c = {}
c['code'] = ''
c['results'] = None
if request.method == 'POST':
py_code = c['code'] = request.POST.get('code')
g = {}
try:
safe_exec(py_code, g)
except Exception as e:
c['results'] = str(e)
else:
c['results'] = pprint.pformat(g)
return render_to_response("debug/run_python_form.html", c)
<commit_msg>Print the full traceback when execution fails.<commit_after> | """Views for debugging and diagnostics"""
import pprint
import traceback
from django.http import Http404
from django.contrib.auth.decorators import login_required
from django_future.csrf import ensure_csrf_cookie, csrf_exempt
from mitxmako.shortcuts import render_to_response
from codejail.safe_exec import safe_exec
@login_required
@ensure_csrf_cookie
def run_python(request):
"""A page to allow testing the Python sandbox on a production server."""
if not request.user.is_staff:
raise Http404
c = {}
c['code'] = ''
c['results'] = None
if request.method == 'POST':
py_code = c['code'] = request.POST.get('code')
g = {}
try:
safe_exec(py_code, g)
except Exception as e:
c['results'] = traceback.format_exc()
else:
c['results'] = pprint.pformat(g)
return render_to_response("debug/run_python_form.html", c)
| """Views for debugging and diagnostics"""
import pprint
from django.http import Http404
from django.contrib.auth.decorators import login_required
from django_future.csrf import ensure_csrf_cookie, csrf_exempt
from mitxmako.shortcuts import render_to_response
from codejail.safe_exec import safe_exec
@login_required
@ensure_csrf_cookie
def run_python(request):
if not request.user.is_staff:
raise Http404
c = {}
c['code'] = ''
c['results'] = None
if request.method == 'POST':
py_code = c['code'] = request.POST.get('code')
g = {}
try:
safe_exec(py_code, g)
except Exception as e:
c['results'] = str(e)
else:
c['results'] = pprint.pformat(g)
return render_to_response("debug/run_python_form.html", c)
Print the full traceback when execution fails."""Views for debugging and diagnostics"""
import pprint
import traceback
from django.http import Http404
from django.contrib.auth.decorators import login_required
from django_future.csrf import ensure_csrf_cookie, csrf_exempt
from mitxmako.shortcuts import render_to_response
from codejail.safe_exec import safe_exec
@login_required
@ensure_csrf_cookie
def run_python(request):
"""A page to allow testing the Python sandbox on a production server."""
if not request.user.is_staff:
raise Http404
c = {}
c['code'] = ''
c['results'] = None
if request.method == 'POST':
py_code = c['code'] = request.POST.get('code')
g = {}
try:
safe_exec(py_code, g)
except Exception as e:
c['results'] = traceback.format_exc()
else:
c['results'] = pprint.pformat(g)
return render_to_response("debug/run_python_form.html", c)
| <commit_before>"""Views for debugging and diagnostics"""
import pprint
from django.http import Http404
from django.contrib.auth.decorators import login_required
from django_future.csrf import ensure_csrf_cookie, csrf_exempt
from mitxmako.shortcuts import render_to_response
from codejail.safe_exec import safe_exec
@login_required
@ensure_csrf_cookie
def run_python(request):
if not request.user.is_staff:
raise Http404
c = {}
c['code'] = ''
c['results'] = None
if request.method == 'POST':
py_code = c['code'] = request.POST.get('code')
g = {}
try:
safe_exec(py_code, g)
except Exception as e:
c['results'] = str(e)
else:
c['results'] = pprint.pformat(g)
return render_to_response("debug/run_python_form.html", c)
<commit_msg>Print the full traceback when execution fails.<commit_after>"""Views for debugging and diagnostics"""
import pprint
import traceback
from django.http import Http404
from django.contrib.auth.decorators import login_required
from django_future.csrf import ensure_csrf_cookie, csrf_exempt
from mitxmako.shortcuts import render_to_response
from codejail.safe_exec import safe_exec
@login_required
@ensure_csrf_cookie
def run_python(request):
"""A page to allow testing the Python sandbox on a production server."""
if not request.user.is_staff:
raise Http404
c = {}
c['code'] = ''
c['results'] = None
if request.method == 'POST':
py_code = c['code'] = request.POST.get('code')
g = {}
try:
safe_exec(py_code, g)
except Exception as e:
c['results'] = traceback.format_exc()
else:
c['results'] = pprint.pformat(g)
return render_to_response("debug/run_python_form.html", c)
|
4b03c2c39c90bd1563954df16f353348f20d7280 | setup.py | setup.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""The Screen class lets you to do positioned writes to the dos terminal.
The Screen class also allows you to specify the colors for foreground and
background, to the extent the dos terminal allows.
"""
classifiers = """\
Development Status :: 3 - Alpha
Environment :: Win32 (MS Windows)
Intended Audience :: Developers
License :: OSI Approved :: BSD License
Operating System :: Microsoft :: MS-DOS
Programming Language :: Python
Topic :: Software Development :: Libraries
Topic :: System :: Shells
Topic :: Terminals
"""
from distutils.core import setup
doclines = __doc__.split('\n')
setup(
name='dosbox-screen',
version='0.0.1',
requires=['ctypes', 'colorama'],
description=doclines[0],
classifiers=[line for line in classifiers.split('\n') if line],
long_description=' '.join(doclines),
license="BSD",
#platform='win32',
author='Bjorn Pettersen',
author_email='bjorn@tkbe.org',
url='https://github.com/thebjorn/doscmd-screen',
download_url='https://github.com/thebjorn/doscmd-screen',
py_modules=['screen']
)
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""The Screen class lets you to do positioned writes to the dos terminal.
The Screen class also allows you to specify the colors for foreground and
background, to the extent the dos terminal allows.
"""
classifiers = """\
Development Status :: 3 - Alpha
Environment :: Win32 (MS Windows)
Intended Audience :: Developers
License :: OSI Approved :: BSD License
Operating System :: Microsoft :: MS-DOS
Programming Language :: Python
Topic :: Software Development :: Libraries
Topic :: System :: Shells
Topic :: Terminals
"""
from distutils.core import setup
doclines = __doc__.split('\n')
setup(
name='dosbox-screen',
version='0.0.3',
requires=['colorama'],
install_requires=['colorama'],
description=doclines[0],
classifiers=[line for line in classifiers.split('\n') if line],
long_description=' '.join(doclines),
license="BSD",
#platform='win32',
author='Bjorn Pettersen',
author_email='bjorn@tkbe.org',
url='https://github.com/thebjorn/doscmd-screen',
download_url='https://github.com/thebjorn/doscmd-screen',
py_modules=['screen']
)
| Add install_requires and remove ctypes from requirements. | Add install_requires and remove ctypes from requirements.
| Python | mit | thebjorn/doscmd-screen | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""The Screen class lets you to do positioned writes to the dos terminal.
The Screen class also allows you to specify the colors for foreground and
background, to the extent the dos terminal allows.
"""
classifiers = """\
Development Status :: 3 - Alpha
Environment :: Win32 (MS Windows)
Intended Audience :: Developers
License :: OSI Approved :: BSD License
Operating System :: Microsoft :: MS-DOS
Programming Language :: Python
Topic :: Software Development :: Libraries
Topic :: System :: Shells
Topic :: Terminals
"""
from distutils.core import setup
doclines = __doc__.split('\n')
setup(
name='dosbox-screen',
version='0.0.1',
requires=['ctypes', 'colorama'],
description=doclines[0],
classifiers=[line for line in classifiers.split('\n') if line],
long_description=' '.join(doclines),
license="BSD",
#platform='win32',
author='Bjorn Pettersen',
author_email='bjorn@tkbe.org',
url='https://github.com/thebjorn/doscmd-screen',
download_url='https://github.com/thebjorn/doscmd-screen',
py_modules=['screen']
)
Add install_requires and remove ctypes from requirements. | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""The Screen class lets you to do positioned writes to the dos terminal.
The Screen class also allows you to specify the colors for foreground and
background, to the extent the dos terminal allows.
"""
classifiers = """\
Development Status :: 3 - Alpha
Environment :: Win32 (MS Windows)
Intended Audience :: Developers
License :: OSI Approved :: BSD License
Operating System :: Microsoft :: MS-DOS
Programming Language :: Python
Topic :: Software Development :: Libraries
Topic :: System :: Shells
Topic :: Terminals
"""
from distutils.core import setup
doclines = __doc__.split('\n')
setup(
name='dosbox-screen',
version='0.0.3',
requires=['colorama'],
install_requires=['colorama'],
description=doclines[0],
classifiers=[line for line in classifiers.split('\n') if line],
long_description=' '.join(doclines),
license="BSD",
#platform='win32',
author='Bjorn Pettersen',
author_email='bjorn@tkbe.org',
url='https://github.com/thebjorn/doscmd-screen',
download_url='https://github.com/thebjorn/doscmd-screen',
py_modules=['screen']
)
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""The Screen class lets you to do positioned writes to the dos terminal.
The Screen class also allows you to specify the colors for foreground and
background, to the extent the dos terminal allows.
"""
classifiers = """\
Development Status :: 3 - Alpha
Environment :: Win32 (MS Windows)
Intended Audience :: Developers
License :: OSI Approved :: BSD License
Operating System :: Microsoft :: MS-DOS
Programming Language :: Python
Topic :: Software Development :: Libraries
Topic :: System :: Shells
Topic :: Terminals
"""
from distutils.core import setup
doclines = __doc__.split('\n')
setup(
name='dosbox-screen',
version='0.0.1',
requires=['ctypes', 'colorama'],
description=doclines[0],
classifiers=[line for line in classifiers.split('\n') if line],
long_description=' '.join(doclines),
license="BSD",
#platform='win32',
author='Bjorn Pettersen',
author_email='bjorn@tkbe.org',
url='https://github.com/thebjorn/doscmd-screen',
download_url='https://github.com/thebjorn/doscmd-screen',
py_modules=['screen']
)
<commit_msg>Add install_requires and remove ctypes from requirements.<commit_after> | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""The Screen class lets you to do positioned writes to the dos terminal.
The Screen class also allows you to specify the colors for foreground and
background, to the extent the dos terminal allows.
"""
classifiers = """\
Development Status :: 3 - Alpha
Environment :: Win32 (MS Windows)
Intended Audience :: Developers
License :: OSI Approved :: BSD License
Operating System :: Microsoft :: MS-DOS
Programming Language :: Python
Topic :: Software Development :: Libraries
Topic :: System :: Shells
Topic :: Terminals
"""
from distutils.core import setup
doclines = __doc__.split('\n')
setup(
name='dosbox-screen',
version='0.0.3',
requires=['colorama'],
install_requires=['colorama'],
description=doclines[0],
classifiers=[line for line in classifiers.split('\n') if line],
long_description=' '.join(doclines),
license="BSD",
#platform='win32',
author='Bjorn Pettersen',
author_email='bjorn@tkbe.org',
url='https://github.com/thebjorn/doscmd-screen',
download_url='https://github.com/thebjorn/doscmd-screen',
py_modules=['screen']
)
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""The Screen class lets you to do positioned writes to the dos terminal.
The Screen class also allows you to specify the colors for foreground and
background, to the extent the dos terminal allows.
"""
classifiers = """\
Development Status :: 3 - Alpha
Environment :: Win32 (MS Windows)
Intended Audience :: Developers
License :: OSI Approved :: BSD License
Operating System :: Microsoft :: MS-DOS
Programming Language :: Python
Topic :: Software Development :: Libraries
Topic :: System :: Shells
Topic :: Terminals
"""
from distutils.core import setup
doclines = __doc__.split('\n')
setup(
name='dosbox-screen',
version='0.0.1',
requires=['ctypes', 'colorama'],
description=doclines[0],
classifiers=[line for line in classifiers.split('\n') if line],
long_description=' '.join(doclines),
license="BSD",
#platform='win32',
author='Bjorn Pettersen',
author_email='bjorn@tkbe.org',
url='https://github.com/thebjorn/doscmd-screen',
download_url='https://github.com/thebjorn/doscmd-screen',
py_modules=['screen']
)
Add install_requires and remove ctypes from requirements.#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""The Screen class lets you to do positioned writes to the dos terminal.
The Screen class also allows you to specify the colors for foreground and
background, to the extent the dos terminal allows.
"""
classifiers = """\
Development Status :: 3 - Alpha
Environment :: Win32 (MS Windows)
Intended Audience :: Developers
License :: OSI Approved :: BSD License
Operating System :: Microsoft :: MS-DOS
Programming Language :: Python
Topic :: Software Development :: Libraries
Topic :: System :: Shells
Topic :: Terminals
"""
from distutils.core import setup
doclines = __doc__.split('\n')
setup(
name='dosbox-screen',
version='0.0.3',
requires=['colorama'],
install_requires=['colorama'],
description=doclines[0],
classifiers=[line for line in classifiers.split('\n') if line],
long_description=' '.join(doclines),
license="BSD",
#platform='win32',
author='Bjorn Pettersen',
author_email='bjorn@tkbe.org',
url='https://github.com/thebjorn/doscmd-screen',
download_url='https://github.com/thebjorn/doscmd-screen',
py_modules=['screen']
)
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""The Screen class lets you to do positioned writes to the dos terminal.
The Screen class also allows you to specify the colors for foreground and
background, to the extent the dos terminal allows.
"""
classifiers = """\
Development Status :: 3 - Alpha
Environment :: Win32 (MS Windows)
Intended Audience :: Developers
License :: OSI Approved :: BSD License
Operating System :: Microsoft :: MS-DOS
Programming Language :: Python
Topic :: Software Development :: Libraries
Topic :: System :: Shells
Topic :: Terminals
"""
from distutils.core import setup
doclines = __doc__.split('\n')
setup(
name='dosbox-screen',
version='0.0.1',
requires=['ctypes', 'colorama'],
description=doclines[0],
classifiers=[line for line in classifiers.split('\n') if line],
long_description=' '.join(doclines),
license="BSD",
#platform='win32',
author='Bjorn Pettersen',
author_email='bjorn@tkbe.org',
url='https://github.com/thebjorn/doscmd-screen',
download_url='https://github.com/thebjorn/doscmd-screen',
py_modules=['screen']
)
<commit_msg>Add install_requires and remove ctypes from requirements.<commit_after>#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""The Screen class lets you to do positioned writes to the dos terminal.
The Screen class also allows you to specify the colors for foreground and
background, to the extent the dos terminal allows.
"""
classifiers = """\
Development Status :: 3 - Alpha
Environment :: Win32 (MS Windows)
Intended Audience :: Developers
License :: OSI Approved :: BSD License
Operating System :: Microsoft :: MS-DOS
Programming Language :: Python
Topic :: Software Development :: Libraries
Topic :: System :: Shells
Topic :: Terminals
"""
from distutils.core import setup
doclines = __doc__.split('\n')
setup(
name='dosbox-screen',
version='0.0.3',
requires=['colorama'],
install_requires=['colorama'],
description=doclines[0],
classifiers=[line for line in classifiers.split('\n') if line],
long_description=' '.join(doclines),
license="BSD",
#platform='win32',
author='Bjorn Pettersen',
author_email='bjorn@tkbe.org',
url='https://github.com/thebjorn/doscmd-screen',
download_url='https://github.com/thebjorn/doscmd-screen',
py_modules=['screen']
)
|
7182f52f495174dc7a9689100f5298e848b8229c | setup.py | setup.py | """Cloud browser package."""
from setuptools import setup, find_packages
from cloud_browser import __version__
# Base packages.
MOD_NAME = "cloud_browser"
PKGS = [x for x in find_packages() if x.split('.')[0] == MOD_NAME]
setup(
name="django-cloud-browser",
version=__version__,
description="Django Cloud Browser application.",
long_description="Browser for cloud datastores (Rackspace, AWS, etc.).",
url="https://github.com/ryan-roemer/django-cloud-browser",
author="Ryan Roemer",
author_email="ryan@loose-bits.com",
classifiers=[
"Development Status :: 3 - Alpha",
"Environment :: Web Environment",
"Framework :: Django",
"Intended Audience :: Developers",
"Intended Audience :: System Administrators",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Topic :: Internet",
"Topic :: Internet :: WWW/HTTP :: Site Management",
],
install_requires=[
"distribute",
],
packages=PKGS,
include_package_data=True,
)
| """Cloud browser package."""
from __future__ import with_statement
import os
from setuptools import setup, find_packages
from cloud_browser import __version__
###############################################################################
# Base packages.
###############################################################################
MOD_NAME = "cloud_browser"
PKGS = [x for x in find_packages() if x.split('.')[0] == MOD_NAME]
###############################################################################
# Helpers.
###############################################################################
def read_file(name):
"""Read file name (without extension) to string."""
cur_path = os.path.dirname(__file__)
exts = ('txt', 'rst')
for ext in exts:
path = os.path.join(cur_path, '.'.join((name, ext)))
if os.path.exists(path):
with open(path, 'rb') as file_obj:
return file_obj.read()
return ''
###############################################################################
# Setup.
###############################################################################
setup(
name="django-cloud-browser",
version=__version__,
description="Django Cloud Browser application.",
long_description=read_file("README"),
url="https://github.com/ryan-roemer/django-cloud-browser",
author="Ryan Roemer",
author_email="ryan@loose-bits.com",
classifiers=[
"Development Status :: 3 - Alpha",
"Environment :: Web Environment",
"Framework :: Django",
"Intended Audience :: Developers",
"Intended Audience :: System Administrators",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Topic :: Internet",
"Topic :: Internet :: WWW/HTTP :: Site Management",
],
install_requires=[
"distribute",
],
packages=PKGS,
include_package_data=True,
)
| Add long description read straight from README. | Setup: Add long description read straight from README.
| Python | mit | ryan-roemer/django-cloud-browser,ryan-roemer/django-cloud-browser,UrbanDaddy/django-cloud-browser,UrbanDaddy/django-cloud-browser,ryan-roemer/django-cloud-browser | """Cloud browser package."""
from setuptools import setup, find_packages
from cloud_browser import __version__
# Base packages.
MOD_NAME = "cloud_browser"
PKGS = [x for x in find_packages() if x.split('.')[0] == MOD_NAME]
setup(
name="django-cloud-browser",
version=__version__,
description="Django Cloud Browser application.",
long_description="Browser for cloud datastores (Rackspace, AWS, etc.).",
url="https://github.com/ryan-roemer/django-cloud-browser",
author="Ryan Roemer",
author_email="ryan@loose-bits.com",
classifiers=[
"Development Status :: 3 - Alpha",
"Environment :: Web Environment",
"Framework :: Django",
"Intended Audience :: Developers",
"Intended Audience :: System Administrators",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Topic :: Internet",
"Topic :: Internet :: WWW/HTTP :: Site Management",
],
install_requires=[
"distribute",
],
packages=PKGS,
include_package_data=True,
)
Setup: Add long description read straight from README. | """Cloud browser package."""
from __future__ import with_statement
import os
from setuptools import setup, find_packages
from cloud_browser import __version__
###############################################################################
# Base packages.
###############################################################################
MOD_NAME = "cloud_browser"
PKGS = [x for x in find_packages() if x.split('.')[0] == MOD_NAME]
###############################################################################
# Helpers.
###############################################################################
def read_file(name):
"""Read file name (without extension) to string."""
cur_path = os.path.dirname(__file__)
exts = ('txt', 'rst')
for ext in exts:
path = os.path.join(cur_path, '.'.join((name, ext)))
if os.path.exists(path):
with open(path, 'rb') as file_obj:
return file_obj.read()
return ''
###############################################################################
# Setup.
###############################################################################
setup(
name="django-cloud-browser",
version=__version__,
description="Django Cloud Browser application.",
long_description=read_file("README"),
url="https://github.com/ryan-roemer/django-cloud-browser",
author="Ryan Roemer",
author_email="ryan@loose-bits.com",
classifiers=[
"Development Status :: 3 - Alpha",
"Environment :: Web Environment",
"Framework :: Django",
"Intended Audience :: Developers",
"Intended Audience :: System Administrators",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Topic :: Internet",
"Topic :: Internet :: WWW/HTTP :: Site Management",
],
install_requires=[
"distribute",
],
packages=PKGS,
include_package_data=True,
)
| <commit_before>"""Cloud browser package."""
from setuptools import setup, find_packages
from cloud_browser import __version__
# Base packages.
MOD_NAME = "cloud_browser"
PKGS = [x for x in find_packages() if x.split('.')[0] == MOD_NAME]
setup(
name="django-cloud-browser",
version=__version__,
description="Django Cloud Browser application.",
long_description="Browser for cloud datastores (Rackspace, AWS, etc.).",
url="https://github.com/ryan-roemer/django-cloud-browser",
author="Ryan Roemer",
author_email="ryan@loose-bits.com",
classifiers=[
"Development Status :: 3 - Alpha",
"Environment :: Web Environment",
"Framework :: Django",
"Intended Audience :: Developers",
"Intended Audience :: System Administrators",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Topic :: Internet",
"Topic :: Internet :: WWW/HTTP :: Site Management",
],
install_requires=[
"distribute",
],
packages=PKGS,
include_package_data=True,
)
<commit_msg>Setup: Add long description read straight from README.<commit_after> | """Cloud browser package."""
from __future__ import with_statement
import os
from setuptools import setup, find_packages
from cloud_browser import __version__
###############################################################################
# Base packages.
###############################################################################
MOD_NAME = "cloud_browser"
PKGS = [x for x in find_packages() if x.split('.')[0] == MOD_NAME]
###############################################################################
# Helpers.
###############################################################################
def read_file(name):
"""Read file name (without extension) to string."""
cur_path = os.path.dirname(__file__)
exts = ('txt', 'rst')
for ext in exts:
path = os.path.join(cur_path, '.'.join((name, ext)))
if os.path.exists(path):
with open(path, 'rb') as file_obj:
return file_obj.read()
return ''
###############################################################################
# Setup.
###############################################################################
setup(
name="django-cloud-browser",
version=__version__,
description="Django Cloud Browser application.",
long_description=read_file("README"),
url="https://github.com/ryan-roemer/django-cloud-browser",
author="Ryan Roemer",
author_email="ryan@loose-bits.com",
classifiers=[
"Development Status :: 3 - Alpha",
"Environment :: Web Environment",
"Framework :: Django",
"Intended Audience :: Developers",
"Intended Audience :: System Administrators",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Topic :: Internet",
"Topic :: Internet :: WWW/HTTP :: Site Management",
],
install_requires=[
"distribute",
],
packages=PKGS,
include_package_data=True,
)
| """Cloud browser package."""
from setuptools import setup, find_packages
from cloud_browser import __version__
# Base packages.
MOD_NAME = "cloud_browser"
PKGS = [x for x in find_packages() if x.split('.')[0] == MOD_NAME]
setup(
name="django-cloud-browser",
version=__version__,
description="Django Cloud Browser application.",
long_description="Browser for cloud datastores (Rackspace, AWS, etc.).",
url="https://github.com/ryan-roemer/django-cloud-browser",
author="Ryan Roemer",
author_email="ryan@loose-bits.com",
classifiers=[
"Development Status :: 3 - Alpha",
"Environment :: Web Environment",
"Framework :: Django",
"Intended Audience :: Developers",
"Intended Audience :: System Administrators",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Topic :: Internet",
"Topic :: Internet :: WWW/HTTP :: Site Management",
],
install_requires=[
"distribute",
],
packages=PKGS,
include_package_data=True,
)
Setup: Add long description read straight from README."""Cloud browser package."""
from __future__ import with_statement
import os
from setuptools import setup, find_packages
from cloud_browser import __version__
###############################################################################
# Base packages.
###############################################################################
MOD_NAME = "cloud_browser"
PKGS = [x for x in find_packages() if x.split('.')[0] == MOD_NAME]
###############################################################################
# Helpers.
###############################################################################
def read_file(name):
"""Read file name (without extension) to string."""
cur_path = os.path.dirname(__file__)
exts = ('txt', 'rst')
for ext in exts:
path = os.path.join(cur_path, '.'.join((name, ext)))
if os.path.exists(path):
with open(path, 'rb') as file_obj:
return file_obj.read()
return ''
###############################################################################
# Setup.
###############################################################################
setup(
name="django-cloud-browser",
version=__version__,
description="Django Cloud Browser application.",
long_description=read_file("README"),
url="https://github.com/ryan-roemer/django-cloud-browser",
author="Ryan Roemer",
author_email="ryan@loose-bits.com",
classifiers=[
"Development Status :: 3 - Alpha",
"Environment :: Web Environment",
"Framework :: Django",
"Intended Audience :: Developers",
"Intended Audience :: System Administrators",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Topic :: Internet",
"Topic :: Internet :: WWW/HTTP :: Site Management",
],
install_requires=[
"distribute",
],
packages=PKGS,
include_package_data=True,
)
| <commit_before>"""Cloud browser package."""
from setuptools import setup, find_packages
from cloud_browser import __version__
# Base packages.
MOD_NAME = "cloud_browser"
PKGS = [x for x in find_packages() if x.split('.')[0] == MOD_NAME]
setup(
name="django-cloud-browser",
version=__version__,
description="Django Cloud Browser application.",
long_description="Browser for cloud datastores (Rackspace, AWS, etc.).",
url="https://github.com/ryan-roemer/django-cloud-browser",
author="Ryan Roemer",
author_email="ryan@loose-bits.com",
classifiers=[
"Development Status :: 3 - Alpha",
"Environment :: Web Environment",
"Framework :: Django",
"Intended Audience :: Developers",
"Intended Audience :: System Administrators",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Topic :: Internet",
"Topic :: Internet :: WWW/HTTP :: Site Management",
],
install_requires=[
"distribute",
],
packages=PKGS,
include_package_data=True,
)
<commit_msg>Setup: Add long description read straight from README.<commit_after>"""Cloud browser package."""
from __future__ import with_statement
import os
from setuptools import setup, find_packages
from cloud_browser import __version__
###############################################################################
# Base packages.
###############################################################################
MOD_NAME = "cloud_browser"
PKGS = [x for x in find_packages() if x.split('.')[0] == MOD_NAME]
###############################################################################
# Helpers.
###############################################################################
def read_file(name):
"""Read file name (without extension) to string."""
cur_path = os.path.dirname(__file__)
exts = ('txt', 'rst')
for ext in exts:
path = os.path.join(cur_path, '.'.join((name, ext)))
if os.path.exists(path):
with open(path, 'rb') as file_obj:
return file_obj.read()
return ''
###############################################################################
# Setup.
###############################################################################
setup(
name="django-cloud-browser",
version=__version__,
description="Django Cloud Browser application.",
long_description=read_file("README"),
url="https://github.com/ryan-roemer/django-cloud-browser",
author="Ryan Roemer",
author_email="ryan@loose-bits.com",
classifiers=[
"Development Status :: 3 - Alpha",
"Environment :: Web Environment",
"Framework :: Django",
"Intended Audience :: Developers",
"Intended Audience :: System Administrators",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Topic :: Internet",
"Topic :: Internet :: WWW/HTTP :: Site Management",
],
install_requires=[
"distribute",
],
packages=PKGS,
include_package_data=True,
)
|
18e07203967dc2ee53a992e61ab709fc0a58d882 | setup.py | setup.py | import io
import os
import re
from setuptools import setup, find_packages
def find_version():
file_dir = os.path.dirname(__file__)
with io.open(os.path.join(file_dir, 'auth0', '__init__.py')) as f:
version = re.search(r'^__version__ = [\'"]([^\'"]*)[\'"]', f.read())
if version:
return version.group(1)
else:
raise RuntimeError("Unable to find version string.")
with open('README.rst') as f:
long_description = f.read()
setup(
name='auth0-python',
version=find_version(),
description='Auth0 Python SDK',
long_description=long_description,
author='Auth0',
author_email='support@auth0.com',
license='MIT',
packages=find_packages(),
install_requires=['requests'],
extras_require={'test': ['mock']},
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
url='https://github.com/auth0/auth0-python',
)
| import io
import os
import re
from setuptools import setup, find_packages
def find_version():
file_dir = os.path.dirname(__file__)
with io.open(os.path.join(file_dir, 'auth0', '__init__.py')) as f:
version = re.search(r'^__version__ = [\'"]([^\'"]*)[\'"]', f.read())
if version:
return version.group(1)
else:
raise RuntimeError("Unable to find version string.")
with io.open('README.rst', encoding='utf-8') as f:
long_description = f.read()
setup(
name='auth0-python',
version=find_version(),
description='Auth0 Python SDK',
long_description=long_description,
author='Auth0',
author_email='support@auth0.com',
license='MIT',
packages=find_packages(),
install_requires=['requests'],
extras_require={'test': ['mock']},
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
url='https://github.com/auth0/auth0-python',
)
| Use io.open with UTF-8 encoding to avoid UnicodeDecodeError | Use io.open with UTF-8 encoding to avoid UnicodeDecodeError
| Python | mit | auth0/auth0-python,auth0/auth0-python | import io
import os
import re
from setuptools import setup, find_packages
def find_version():
file_dir = os.path.dirname(__file__)
with io.open(os.path.join(file_dir, 'auth0', '__init__.py')) as f:
version = re.search(r'^__version__ = [\'"]([^\'"]*)[\'"]', f.read())
if version:
return version.group(1)
else:
raise RuntimeError("Unable to find version string.")
with open('README.rst') as f:
long_description = f.read()
setup(
name='auth0-python',
version=find_version(),
description='Auth0 Python SDK',
long_description=long_description,
author='Auth0',
author_email='support@auth0.com',
license='MIT',
packages=find_packages(),
install_requires=['requests'],
extras_require={'test': ['mock']},
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
url='https://github.com/auth0/auth0-python',
)
Use io.open with UTF-8 encoding to avoid UnicodeDecodeError | import io
import os
import re
from setuptools import setup, find_packages
def find_version():
file_dir = os.path.dirname(__file__)
with io.open(os.path.join(file_dir, 'auth0', '__init__.py')) as f:
version = re.search(r'^__version__ = [\'"]([^\'"]*)[\'"]', f.read())
if version:
return version.group(1)
else:
raise RuntimeError("Unable to find version string.")
with io.open('README.rst', encoding='utf-8') as f:
long_description = f.read()
setup(
name='auth0-python',
version=find_version(),
description='Auth0 Python SDK',
long_description=long_description,
author='Auth0',
author_email='support@auth0.com',
license='MIT',
packages=find_packages(),
install_requires=['requests'],
extras_require={'test': ['mock']},
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
url='https://github.com/auth0/auth0-python',
)
| <commit_before>import io
import os
import re
from setuptools import setup, find_packages
def find_version():
file_dir = os.path.dirname(__file__)
with io.open(os.path.join(file_dir, 'auth0', '__init__.py')) as f:
version = re.search(r'^__version__ = [\'"]([^\'"]*)[\'"]', f.read())
if version:
return version.group(1)
else:
raise RuntimeError("Unable to find version string.")
with open('README.rst') as f:
long_description = f.read()
setup(
name='auth0-python',
version=find_version(),
description='Auth0 Python SDK',
long_description=long_description,
author='Auth0',
author_email='support@auth0.com',
license='MIT',
packages=find_packages(),
install_requires=['requests'],
extras_require={'test': ['mock']},
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
url='https://github.com/auth0/auth0-python',
)
<commit_msg>Use io.open with UTF-8 encoding to avoid UnicodeDecodeError<commit_after> | import io
import os
import re
from setuptools import setup, find_packages
def find_version():
file_dir = os.path.dirname(__file__)
with io.open(os.path.join(file_dir, 'auth0', '__init__.py')) as f:
version = re.search(r'^__version__ = [\'"]([^\'"]*)[\'"]', f.read())
if version:
return version.group(1)
else:
raise RuntimeError("Unable to find version string.")
with io.open('README.rst', encoding='utf-8') as f:
long_description = f.read()
setup(
name='auth0-python',
version=find_version(),
description='Auth0 Python SDK',
long_description=long_description,
author='Auth0',
author_email='support@auth0.com',
license='MIT',
packages=find_packages(),
install_requires=['requests'],
extras_require={'test': ['mock']},
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
url='https://github.com/auth0/auth0-python',
)
| import io
import os
import re
from setuptools import setup, find_packages
def find_version():
file_dir = os.path.dirname(__file__)
with io.open(os.path.join(file_dir, 'auth0', '__init__.py')) as f:
version = re.search(r'^__version__ = [\'"]([^\'"]*)[\'"]', f.read())
if version:
return version.group(1)
else:
raise RuntimeError("Unable to find version string.")
with open('README.rst') as f:
long_description = f.read()
setup(
name='auth0-python',
version=find_version(),
description='Auth0 Python SDK',
long_description=long_description,
author='Auth0',
author_email='support@auth0.com',
license='MIT',
packages=find_packages(),
install_requires=['requests'],
extras_require={'test': ['mock']},
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
url='https://github.com/auth0/auth0-python',
)
Use io.open with UTF-8 encoding to avoid UnicodeDecodeErrorimport io
import os
import re
from setuptools import setup, find_packages
def find_version():
file_dir = os.path.dirname(__file__)
with io.open(os.path.join(file_dir, 'auth0', '__init__.py')) as f:
version = re.search(r'^__version__ = [\'"]([^\'"]*)[\'"]', f.read())
if version:
return version.group(1)
else:
raise RuntimeError("Unable to find version string.")
with io.open('README.rst', encoding='utf-8') as f:
long_description = f.read()
setup(
name='auth0-python',
version=find_version(),
description='Auth0 Python SDK',
long_description=long_description,
author='Auth0',
author_email='support@auth0.com',
license='MIT',
packages=find_packages(),
install_requires=['requests'],
extras_require={'test': ['mock']},
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
url='https://github.com/auth0/auth0-python',
)
| <commit_before>import io
import os
import re
from setuptools import setup, find_packages
def find_version():
file_dir = os.path.dirname(__file__)
with io.open(os.path.join(file_dir, 'auth0', '__init__.py')) as f:
version = re.search(r'^__version__ = [\'"]([^\'"]*)[\'"]', f.read())
if version:
return version.group(1)
else:
raise RuntimeError("Unable to find version string.")
with open('README.rst') as f:
long_description = f.read()
setup(
name='auth0-python',
version=find_version(),
description='Auth0 Python SDK',
long_description=long_description,
author='Auth0',
author_email='support@auth0.com',
license='MIT',
packages=find_packages(),
install_requires=['requests'],
extras_require={'test': ['mock']},
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
url='https://github.com/auth0/auth0-python',
)
<commit_msg>Use io.open with UTF-8 encoding to avoid UnicodeDecodeError<commit_after>import io
import os
import re
from setuptools import setup, find_packages
def find_version():
file_dir = os.path.dirname(__file__)
with io.open(os.path.join(file_dir, 'auth0', '__init__.py')) as f:
version = re.search(r'^__version__ = [\'"]([^\'"]*)[\'"]', f.read())
if version:
return version.group(1)
else:
raise RuntimeError("Unable to find version string.")
with io.open('README.rst', encoding='utf-8') as f:
long_description = f.read()
setup(
name='auth0-python',
version=find_version(),
description='Auth0 Python SDK',
long_description=long_description,
author='Auth0',
author_email='support@auth0.com',
license='MIT',
packages=find_packages(),
install_requires=['requests'],
extras_require={'test': ['mock']},
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
url='https://github.com/auth0/auth0-python',
)
|
8a3d4015136d6aebccb092208f7c6d02b5c93e13 | setup.py | setup.py | from setuptools import setup, find_packages
setup(
name='ssbio',
version='0.1',
author='Nathan Mih',
author_email='nmih@ucsd.edu',
license='MIT',
url='http://github.com/nmih/ssbio',
description='Various tools and functions to enable structural systems biology',
packages=find_packages(),
package_dir={'ssbio': 'ssbio'},
package_data={'ssbio': ['ssbio/etc/*']},
scripts = ['ssbio/scripts/cleanpdb', 'ssbio/scripts/aggprop', 'ssbio/scripts/thermostability', 'ssbio/structure/mutatepdb.py', 'ssbio/structure/tleap.py',
'ssbio/databases/drugbank.py', 'ssbio/structure/properties/msmsprops.py',
'ssbio/dock/dock.py'],
long_description=open('README.md').read(),
install_requires=['biopython',
'numpy',
'tqdm',
'pandas',
'requests',
'cachetools',
'bioservices',
'prody',
'xmltodict']
) | from setuptools import setup, find_packages
setup(
name='ssbio',
version='0.1',
author='Nathan Mih',
author_email='nmih@ucsd.edu',
license='MIT',
url='http://github.com/nmih/ssbio',
description='Various tools and functions to enable structural systems biology',
packages=find_packages(),
package_dir={'ssbio': 'ssbio'},
package_data={'ssbio': ['ssbio/etc/*']},
scripts = ['ssbio/scripts/cleanpdb', 'ssbio/scripts/aggprop', 'ssbio/scripts/thermostability', 'ssbio/structure/mutatepdb.py', 'ssbio/structure/tleap.py',
'ssbio/databases/drugbank.py', 'ssbio/structure/properties/msmsprops.py',
'ssbio/dock/dock.py'],
long_description=open('README.md').read(),
install_requires=['biopython',
'numpy',
'tqdm',
'pandas',
'requests',
'cachetools',
'bioservices',
'prody',
'xmltodict',
'bidict']
) | Add bidict to required packages | Add bidict to required packages
| Python | mit | SBRG/ssbio,nmih/ssbio,nmih/ssbio,SBRG/ssbio | from setuptools import setup, find_packages
setup(
name='ssbio',
version='0.1',
author='Nathan Mih',
author_email='nmih@ucsd.edu',
license='MIT',
url='http://github.com/nmih/ssbio',
description='Various tools and functions to enable structural systems biology',
packages=find_packages(),
package_dir={'ssbio': 'ssbio'},
package_data={'ssbio': ['ssbio/etc/*']},
scripts = ['ssbio/scripts/cleanpdb', 'ssbio/scripts/aggprop', 'ssbio/scripts/thermostability', 'ssbio/structure/mutatepdb.py', 'ssbio/structure/tleap.py',
'ssbio/databases/drugbank.py', 'ssbio/structure/properties/msmsprops.py',
'ssbio/dock/dock.py'],
long_description=open('README.md').read(),
install_requires=['biopython',
'numpy',
'tqdm',
'pandas',
'requests',
'cachetools',
'bioservices',
'prody',
'xmltodict']
)Add bidict to required packages | from setuptools import setup, find_packages
setup(
name='ssbio',
version='0.1',
author='Nathan Mih',
author_email='nmih@ucsd.edu',
license='MIT',
url='http://github.com/nmih/ssbio',
description='Various tools and functions to enable structural systems biology',
packages=find_packages(),
package_dir={'ssbio': 'ssbio'},
package_data={'ssbio': ['ssbio/etc/*']},
scripts = ['ssbio/scripts/cleanpdb', 'ssbio/scripts/aggprop', 'ssbio/scripts/thermostability', 'ssbio/structure/mutatepdb.py', 'ssbio/structure/tleap.py',
'ssbio/databases/drugbank.py', 'ssbio/structure/properties/msmsprops.py',
'ssbio/dock/dock.py'],
long_description=open('README.md').read(),
install_requires=['biopython',
'numpy',
'tqdm',
'pandas',
'requests',
'cachetools',
'bioservices',
'prody',
'xmltodict',
'bidict']
) | <commit_before>from setuptools import setup, find_packages
setup(
name='ssbio',
version='0.1',
author='Nathan Mih',
author_email='nmih@ucsd.edu',
license='MIT',
url='http://github.com/nmih/ssbio',
description='Various tools and functions to enable structural systems biology',
packages=find_packages(),
package_dir={'ssbio': 'ssbio'},
package_data={'ssbio': ['ssbio/etc/*']},
scripts = ['ssbio/scripts/cleanpdb', 'ssbio/scripts/aggprop', 'ssbio/scripts/thermostability', 'ssbio/structure/mutatepdb.py', 'ssbio/structure/tleap.py',
'ssbio/databases/drugbank.py', 'ssbio/structure/properties/msmsprops.py',
'ssbio/dock/dock.py'],
long_description=open('README.md').read(),
install_requires=['biopython',
'numpy',
'tqdm',
'pandas',
'requests',
'cachetools',
'bioservices',
'prody',
'xmltodict']
)<commit_msg>Add bidict to required packages<commit_after> | from setuptools import setup, find_packages
setup(
name='ssbio',
version='0.1',
author='Nathan Mih',
author_email='nmih@ucsd.edu',
license='MIT',
url='http://github.com/nmih/ssbio',
description='Various tools and functions to enable structural systems biology',
packages=find_packages(),
package_dir={'ssbio': 'ssbio'},
package_data={'ssbio': ['ssbio/etc/*']},
scripts = ['ssbio/scripts/cleanpdb', 'ssbio/scripts/aggprop', 'ssbio/scripts/thermostability', 'ssbio/structure/mutatepdb.py', 'ssbio/structure/tleap.py',
'ssbio/databases/drugbank.py', 'ssbio/structure/properties/msmsprops.py',
'ssbio/dock/dock.py'],
long_description=open('README.md').read(),
install_requires=['biopython',
'numpy',
'tqdm',
'pandas',
'requests',
'cachetools',
'bioservices',
'prody',
'xmltodict',
'bidict']
) | from setuptools import setup, find_packages
setup(
name='ssbio',
version='0.1',
author='Nathan Mih',
author_email='nmih@ucsd.edu',
license='MIT',
url='http://github.com/nmih/ssbio',
description='Various tools and functions to enable structural systems biology',
packages=find_packages(),
package_dir={'ssbio': 'ssbio'},
package_data={'ssbio': ['ssbio/etc/*']},
scripts = ['ssbio/scripts/cleanpdb', 'ssbio/scripts/aggprop', 'ssbio/scripts/thermostability', 'ssbio/structure/mutatepdb.py', 'ssbio/structure/tleap.py',
'ssbio/databases/drugbank.py', 'ssbio/structure/properties/msmsprops.py',
'ssbio/dock/dock.py'],
long_description=open('README.md').read(),
install_requires=['biopython',
'numpy',
'tqdm',
'pandas',
'requests',
'cachetools',
'bioservices',
'prody',
'xmltodict']
)Add bidict to required packagesfrom setuptools import setup, find_packages
setup(
name='ssbio',
version='0.1',
author='Nathan Mih',
author_email='nmih@ucsd.edu',
license='MIT',
url='http://github.com/nmih/ssbio',
description='Various tools and functions to enable structural systems biology',
packages=find_packages(),
package_dir={'ssbio': 'ssbio'},
package_data={'ssbio': ['ssbio/etc/*']},
scripts = ['ssbio/scripts/cleanpdb', 'ssbio/scripts/aggprop', 'ssbio/scripts/thermostability', 'ssbio/structure/mutatepdb.py', 'ssbio/structure/tleap.py',
'ssbio/databases/drugbank.py', 'ssbio/structure/properties/msmsprops.py',
'ssbio/dock/dock.py'],
long_description=open('README.md').read(),
install_requires=['biopython',
'numpy',
'tqdm',
'pandas',
'requests',
'cachetools',
'bioservices',
'prody',
'xmltodict',
'bidict']
) | <commit_before>from setuptools import setup, find_packages
setup(
name='ssbio',
version='0.1',
author='Nathan Mih',
author_email='nmih@ucsd.edu',
license='MIT',
url='http://github.com/nmih/ssbio',
description='Various tools and functions to enable structural systems biology',
packages=find_packages(),
package_dir={'ssbio': 'ssbio'},
package_data={'ssbio': ['ssbio/etc/*']},
scripts = ['ssbio/scripts/cleanpdb', 'ssbio/scripts/aggprop', 'ssbio/scripts/thermostability', 'ssbio/structure/mutatepdb.py', 'ssbio/structure/tleap.py',
'ssbio/databases/drugbank.py', 'ssbio/structure/properties/msmsprops.py',
'ssbio/dock/dock.py'],
long_description=open('README.md').read(),
install_requires=['biopython',
'numpy',
'tqdm',
'pandas',
'requests',
'cachetools',
'bioservices',
'prody',
'xmltodict']
)<commit_msg>Add bidict to required packages<commit_after>from setuptools import setup, find_packages
setup(
name='ssbio',
version='0.1',
author='Nathan Mih',
author_email='nmih@ucsd.edu',
license='MIT',
url='http://github.com/nmih/ssbio',
description='Various tools and functions to enable structural systems biology',
packages=find_packages(),
package_dir={'ssbio': 'ssbio'},
package_data={'ssbio': ['ssbio/etc/*']},
scripts = ['ssbio/scripts/cleanpdb', 'ssbio/scripts/aggprop', 'ssbio/scripts/thermostability', 'ssbio/structure/mutatepdb.py', 'ssbio/structure/tleap.py',
'ssbio/databases/drugbank.py', 'ssbio/structure/properties/msmsprops.py',
'ssbio/dock/dock.py'],
long_description=open('README.md').read(),
install_requires=['biopython',
'numpy',
'tqdm',
'pandas',
'requests',
'cachetools',
'bioservices',
'prody',
'xmltodict',
'bidict']
) |
695743c3a887224c212b434de7d6e2ccf08d4620 | setup.py | setup.py | #!/usr/bin/env python
from distutils.core import setup
setup(
name='wagtail_embed_videos',
version='0.0.5',
description='Embed Videos for Wagtail CMS.',
long_description=README,
author='Diogo Marques',
author_email='doriva.marques.29@gmail.com',
maintainer='Diogo Marques',
maintainer_email='doriva.marques.29@gmail.com',
url='https://github.com/infoportugal/wagtail-embedvideos',
packages=['wagtail_embed_videos', 'wagtail_embed_videos.views'],
package_data={'wagtail_embed_videos': ['static/wagtail_embed_videos/js/*.js']},
requires=['django(>=1.7)', 'wagtail(>=1.0)', 'django-embed-video(>=1.0)'],
install_requires=['wagtail', 'django-embed-video']
classifiers=[
'Programming Language :: Python',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Operating System :: OS Independent',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'Framework :: Django',
'Framework :: Wagtail CMS',
'License :: OSI Approved :: BSD License'],
license='New BSD',
)
| #!/usr/bin/env python
from distutils.core import setup
setup(
name='wagtail_embed_videos',
version='0.0.5',
description='Embed Videos for Wagtail CMS.',
long_description=README,
author='Diogo Marques',
author_email='doriva.marques.29@gmail.com',
maintainer='Diogo Marques',
maintainer_email='doriva.marques.29@gmail.com',
url='https://github.com/infoportugal/wagtail-embedvideos',
packages=['wagtail_embed_videos', 'wagtail_embed_videos.views'],
package_data={'wagtail_embed_videos': ['static/wagtail_embed_videos/js/*.js']},
requires=['django(>=1.7)', 'wagtail(>=1.0)', 'django-embed-video(>=1.0)'],
install_requires=['wagtail', 'django-embed-video'],
classifiers=[
'Programming Language :: Python',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Operating System :: OS Independent',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'Framework :: Django',
'Framework :: Wagtail CMS',
'License :: OSI Approved :: BSD License'],
license='New BSD',
)
| Fix other SyntaxError: invalid syntax | Fix other SyntaxError: invalid syntax
Fix
Traceback (most recent call last):
File "<string>", line 20, in <module>
File "/tmp/pip-7dj0aup6-build/setup.py", line 19
classifiers=[
^
SyntaxError: invalid syntax | Python | bsd-3-clause | SalahAdDin/wagtail-embedvideos,SalahAdDin/wagtail-embedvideos,infoportugal/wagtail-embedvideos,infoportugal/wagtail-embedvideos,SalahAdDin/wagtail-embedvideos,infoportugal/wagtail-embedvideos | #!/usr/bin/env python
from distutils.core import setup
setup(
name='wagtail_embed_videos',
version='0.0.5',
description='Embed Videos for Wagtail CMS.',
long_description=README,
author='Diogo Marques',
author_email='doriva.marques.29@gmail.com',
maintainer='Diogo Marques',
maintainer_email='doriva.marques.29@gmail.com',
url='https://github.com/infoportugal/wagtail-embedvideos',
packages=['wagtail_embed_videos', 'wagtail_embed_videos.views'],
package_data={'wagtail_embed_videos': ['static/wagtail_embed_videos/js/*.js']},
requires=['django(>=1.7)', 'wagtail(>=1.0)', 'django-embed-video(>=1.0)'],
install_requires=['wagtail', 'django-embed-video']
classifiers=[
'Programming Language :: Python',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Operating System :: OS Independent',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'Framework :: Django',
'Framework :: Wagtail CMS',
'License :: OSI Approved :: BSD License'],
license='New BSD',
)
Fix other SyntaxError: invalid syntax
Fix
Traceback (most recent call last):
File "<string>", line 20, in <module>
File "/tmp/pip-7dj0aup6-build/setup.py", line 19
classifiers=[
^
SyntaxError: invalid syntax | #!/usr/bin/env python
from distutils.core import setup
setup(
name='wagtail_embed_videos',
version='0.0.5',
description='Embed Videos for Wagtail CMS.',
long_description=README,
author='Diogo Marques',
author_email='doriva.marques.29@gmail.com',
maintainer='Diogo Marques',
maintainer_email='doriva.marques.29@gmail.com',
url='https://github.com/infoportugal/wagtail-embedvideos',
packages=['wagtail_embed_videos', 'wagtail_embed_videos.views'],
package_data={'wagtail_embed_videos': ['static/wagtail_embed_videos/js/*.js']},
requires=['django(>=1.7)', 'wagtail(>=1.0)', 'django-embed-video(>=1.0)'],
install_requires=['wagtail', 'django-embed-video'],
classifiers=[
'Programming Language :: Python',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Operating System :: OS Independent',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'Framework :: Django',
'Framework :: Wagtail CMS',
'License :: OSI Approved :: BSD License'],
license='New BSD',
)
| <commit_before>#!/usr/bin/env python
from distutils.core import setup
setup(
name='wagtail_embed_videos',
version='0.0.5',
description='Embed Videos for Wagtail CMS.',
long_description=README,
author='Diogo Marques',
author_email='doriva.marques.29@gmail.com',
maintainer='Diogo Marques',
maintainer_email='doriva.marques.29@gmail.com',
url='https://github.com/infoportugal/wagtail-embedvideos',
packages=['wagtail_embed_videos', 'wagtail_embed_videos.views'],
package_data={'wagtail_embed_videos': ['static/wagtail_embed_videos/js/*.js']},
requires=['django(>=1.7)', 'wagtail(>=1.0)', 'django-embed-video(>=1.0)'],
install_requires=['wagtail', 'django-embed-video']
classifiers=[
'Programming Language :: Python',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Operating System :: OS Independent',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'Framework :: Django',
'Framework :: Wagtail CMS',
'License :: OSI Approved :: BSD License'],
license='New BSD',
)
<commit_msg>Fix other SyntaxError: invalid syntax
Fix
Traceback (most recent call last):
File "<string>", line 20, in <module>
File "/tmp/pip-7dj0aup6-build/setup.py", line 19
classifiers=[
^
SyntaxError: invalid syntax<commit_after> | #!/usr/bin/env python
from distutils.core import setup
setup(
name='wagtail_embed_videos',
version='0.0.5',
description='Embed Videos for Wagtail CMS.',
long_description=README,
author='Diogo Marques',
author_email='doriva.marques.29@gmail.com',
maintainer='Diogo Marques',
maintainer_email='doriva.marques.29@gmail.com',
url='https://github.com/infoportugal/wagtail-embedvideos',
packages=['wagtail_embed_videos', 'wagtail_embed_videos.views'],
package_data={'wagtail_embed_videos': ['static/wagtail_embed_videos/js/*.js']},
requires=['django(>=1.7)', 'wagtail(>=1.0)', 'django-embed-video(>=1.0)'],
install_requires=['wagtail', 'django-embed-video'],
classifiers=[
'Programming Language :: Python',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Operating System :: OS Independent',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'Framework :: Django',
'Framework :: Wagtail CMS',
'License :: OSI Approved :: BSD License'],
license='New BSD',
)
| #!/usr/bin/env python
from distutils.core import setup
setup(
name='wagtail_embed_videos',
version='0.0.5',
description='Embed Videos for Wagtail CMS.',
long_description=README,
author='Diogo Marques',
author_email='doriva.marques.29@gmail.com',
maintainer='Diogo Marques',
maintainer_email='doriva.marques.29@gmail.com',
url='https://github.com/infoportugal/wagtail-embedvideos',
packages=['wagtail_embed_videos', 'wagtail_embed_videos.views'],
package_data={'wagtail_embed_videos': ['static/wagtail_embed_videos/js/*.js']},
requires=['django(>=1.7)', 'wagtail(>=1.0)', 'django-embed-video(>=1.0)'],
install_requires=['wagtail', 'django-embed-video']
classifiers=[
'Programming Language :: Python',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Operating System :: OS Independent',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'Framework :: Django',
'Framework :: Wagtail CMS',
'License :: OSI Approved :: BSD License'],
license='New BSD',
)
Fix other SyntaxError: invalid syntax
Fix
Traceback (most recent call last):
File "<string>", line 20, in <module>
File "/tmp/pip-7dj0aup6-build/setup.py", line 19
classifiers=[
^
SyntaxError: invalid syntax#!/usr/bin/env python
from distutils.core import setup
setup(
name='wagtail_embed_videos',
version='0.0.5',
description='Embed Videos for Wagtail CMS.',
long_description=README,
author='Diogo Marques',
author_email='doriva.marques.29@gmail.com',
maintainer='Diogo Marques',
maintainer_email='doriva.marques.29@gmail.com',
url='https://github.com/infoportugal/wagtail-embedvideos',
packages=['wagtail_embed_videos', 'wagtail_embed_videos.views'],
package_data={'wagtail_embed_videos': ['static/wagtail_embed_videos/js/*.js']},
requires=['django(>=1.7)', 'wagtail(>=1.0)', 'django-embed-video(>=1.0)'],
install_requires=['wagtail', 'django-embed-video'],
classifiers=[
'Programming Language :: Python',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Operating System :: OS Independent',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'Framework :: Django',
'Framework :: Wagtail CMS',
'License :: OSI Approved :: BSD License'],
license='New BSD',
)
| <commit_before>#!/usr/bin/env python
from distutils.core import setup
setup(
name='wagtail_embed_videos',
version='0.0.5',
description='Embed Videos for Wagtail CMS.',
long_description=README,
author='Diogo Marques',
author_email='doriva.marques.29@gmail.com',
maintainer='Diogo Marques',
maintainer_email='doriva.marques.29@gmail.com',
url='https://github.com/infoportugal/wagtail-embedvideos',
packages=['wagtail_embed_videos', 'wagtail_embed_videos.views'],
package_data={'wagtail_embed_videos': ['static/wagtail_embed_videos/js/*.js']},
requires=['django(>=1.7)', 'wagtail(>=1.0)', 'django-embed-video(>=1.0)'],
install_requires=['wagtail', 'django-embed-video']
classifiers=[
'Programming Language :: Python',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Operating System :: OS Independent',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'Framework :: Django',
'Framework :: Wagtail CMS',
'License :: OSI Approved :: BSD License'],
license='New BSD',
)
<commit_msg>Fix other SyntaxError: invalid syntax
Fix
Traceback (most recent call last):
File "<string>", line 20, in <module>
File "/tmp/pip-7dj0aup6-build/setup.py", line 19
classifiers=[
^
SyntaxError: invalid syntax<commit_after>#!/usr/bin/env python
from distutils.core import setup
setup(
name='wagtail_embed_videos',
version='0.0.5',
description='Embed Videos for Wagtail CMS.',
long_description=README,
author='Diogo Marques',
author_email='doriva.marques.29@gmail.com',
maintainer='Diogo Marques',
maintainer_email='doriva.marques.29@gmail.com',
url='https://github.com/infoportugal/wagtail-embedvideos',
packages=['wagtail_embed_videos', 'wagtail_embed_videos.views'],
package_data={'wagtail_embed_videos': ['static/wagtail_embed_videos/js/*.js']},
requires=['django(>=1.7)', 'wagtail(>=1.0)', 'django-embed-video(>=1.0)'],
install_requires=['wagtail', 'django-embed-video'],
classifiers=[
'Programming Language :: Python',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Operating System :: OS Independent',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'Framework :: Django',
'Framework :: Wagtail CMS',
'License :: OSI Approved :: BSD License'],
license='New BSD',
)
|
55745f668715c294cd5662712b2d1ccb7726f125 | setup.py | setup.py | from distutils.core import setup
import multi_schema
setup(
name = "django-multi-schema",
version = multi_schema.__version__,
description = "Postgres schema support in django.",
url = "http://hg.schinckel.net/django-multi-schema",
author = "Matthew Schinckel",
author_email = "matt@schinckel.net",
packages = [
"multi_schema",
],
classifiers = [
'Programming Language :: Python',
'Operating System :: OS Independent',
'Framework :: Django',
],
)
| from distutils.core import setup
import multi_schema
setup(
name = "django-multi-schema",
version = multi_schema.__version__,
description = "Postgres schema support in django.",
url = "http://hg.schinckel.net/django-multi-schema",
author = "Matthew Schinckel",
author_email = "matt@schinckel.net",
packages = [
"multi_schema",
],
install_requires = [
'south==0.7.4',
],
classifiers = [
'Programming Language :: Python',
'Operating System :: OS Independent',
'Framework :: Django',
],
)
| Add south as a dependency, so we can apply a version. Does not need to be installed in INSTALLED_APPS. | Add south as a dependency, so we can apply a version.
Does not need to be installed in INSTALLED_APPS.
| Python | bsd-3-clause | schinckel/django-boardinghouse,schinckel/django-boardinghouse,schinckel/django-boardinghouse | from distutils.core import setup
import multi_schema
setup(
name = "django-multi-schema",
version = multi_schema.__version__,
description = "Postgres schema support in django.",
url = "http://hg.schinckel.net/django-multi-schema",
author = "Matthew Schinckel",
author_email = "matt@schinckel.net",
packages = [
"multi_schema",
],
classifiers = [
'Programming Language :: Python',
'Operating System :: OS Independent',
'Framework :: Django',
],
)
Add south as a dependency, so we can apply a version.
Does not need to be installed in INSTALLED_APPS. | from distutils.core import setup
import multi_schema
setup(
name = "django-multi-schema",
version = multi_schema.__version__,
description = "Postgres schema support in django.",
url = "http://hg.schinckel.net/django-multi-schema",
author = "Matthew Schinckel",
author_email = "matt@schinckel.net",
packages = [
"multi_schema",
],
install_requires = [
'south==0.7.4',
],
classifiers = [
'Programming Language :: Python',
'Operating System :: OS Independent',
'Framework :: Django',
],
)
| <commit_before>from distutils.core import setup
import multi_schema
setup(
name = "django-multi-schema",
version = multi_schema.__version__,
description = "Postgres schema support in django.",
url = "http://hg.schinckel.net/django-multi-schema",
author = "Matthew Schinckel",
author_email = "matt@schinckel.net",
packages = [
"multi_schema",
],
classifiers = [
'Programming Language :: Python',
'Operating System :: OS Independent',
'Framework :: Django',
],
)
<commit_msg>Add south as a dependency, so we can apply a version.
Does not need to be installed in INSTALLED_APPS.<commit_after> | from distutils.core import setup
import multi_schema
setup(
name = "django-multi-schema",
version = multi_schema.__version__,
description = "Postgres schema support in django.",
url = "http://hg.schinckel.net/django-multi-schema",
author = "Matthew Schinckel",
author_email = "matt@schinckel.net",
packages = [
"multi_schema",
],
install_requires = [
'south==0.7.4',
],
classifiers = [
'Programming Language :: Python',
'Operating System :: OS Independent',
'Framework :: Django',
],
)
| from distutils.core import setup
import multi_schema
setup(
name = "django-multi-schema",
version = multi_schema.__version__,
description = "Postgres schema support in django.",
url = "http://hg.schinckel.net/django-multi-schema",
author = "Matthew Schinckel",
author_email = "matt@schinckel.net",
packages = [
"multi_schema",
],
classifiers = [
'Programming Language :: Python',
'Operating System :: OS Independent',
'Framework :: Django',
],
)
Add south as a dependency, so we can apply a version.
Does not need to be installed in INSTALLED_APPS.from distutils.core import setup
import multi_schema
setup(
name = "django-multi-schema",
version = multi_schema.__version__,
description = "Postgres schema support in django.",
url = "http://hg.schinckel.net/django-multi-schema",
author = "Matthew Schinckel",
author_email = "matt@schinckel.net",
packages = [
"multi_schema",
],
install_requires = [
'south==0.7.4',
],
classifiers = [
'Programming Language :: Python',
'Operating System :: OS Independent',
'Framework :: Django',
],
)
| <commit_before>from distutils.core import setup
import multi_schema
setup(
name = "django-multi-schema",
version = multi_schema.__version__,
description = "Postgres schema support in django.",
url = "http://hg.schinckel.net/django-multi-schema",
author = "Matthew Schinckel",
author_email = "matt@schinckel.net",
packages = [
"multi_schema",
],
classifiers = [
'Programming Language :: Python',
'Operating System :: OS Independent',
'Framework :: Django',
],
)
<commit_msg>Add south as a dependency, so we can apply a version.
Does not need to be installed in INSTALLED_APPS.<commit_after>from distutils.core import setup
import multi_schema
setup(
name = "django-multi-schema",
version = multi_schema.__version__,
description = "Postgres schema support in django.",
url = "http://hg.schinckel.net/django-multi-schema",
author = "Matthew Schinckel",
author_email = "matt@schinckel.net",
packages = [
"multi_schema",
],
install_requires = [
'south==0.7.4',
],
classifiers = [
'Programming Language :: Python',
'Operating System :: OS Independent',
'Framework :: Django',
],
)
|
a84eca4bffd30c068194cc6c6d8176178fe26e78 | setup.py | setup.py | #!/usr/bin/env python
from setuptools import setup, find_packages
from os.path import dirname, join
import sys, os
# When creating the sdist, make sure the django.mo file also exists:
if 'sdist' in sys.argv:
try:
os.chdir('fluent_comments')
from django.core.management.commands.compilemessages import compile_messages
compile_messages(sys.stderr)
finally:
os.chdir('..')
setup(
name='django-fluent-comments',
version='0.8.0',
license='Apache License, Version 2.0',
install_requires=[
'Django>=1.2.0',
'django-crispy-forms>=1.1.1',
'akismet>=0.2',
],
description='A modern, ajax-based appearance for django.contrib.comments',
long_description=open('README.rst').read(),
author='Diederik van der Boor',
author_email='opensource@edoburu.nl',
url='https://github.com/edoburu/django-fluent-comments',
download_url='https://github.com/edoburu/django-fluent-comments/zipball/master',
packages=find_packages(exclude=('example*',)),
include_package_data=True,
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
]
)
| #!/usr/bin/env python
from setuptools import setup, find_packages
from os.path import dirname, join
import sys, os
# When creating the sdist, make sure the django.mo file also exists:
if 'sdist' in sys.argv:
try:
os.chdir('fluent_comments')
from django.core.management.commands.compilemessages import compile_messages
compile_messages(sys.stderr)
finally:
os.chdir('..')
setup(
name='django-fluent-comments',
version='0.8.0',
license='Apache License, Version 2.0',
install_requires=[
'django-crispy-forms>=1.1.1',
'akismet>=0.2',
],
requires=[
'Django (>=1.2)',
],
description='A modern, ajax-based appearance for django.contrib.comments',
long_description=open('README.rst').read(),
author='Diederik van der Boor',
author_email='opensource@edoburu.nl',
url='https://github.com/edoburu/django-fluent-comments',
download_url='https://github.com/edoburu/django-fluent-comments/zipball/master',
packages=find_packages(exclude=('example*',)),
include_package_data=True,
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
]
)
| Move Django out of `install_requires` to to `requires` block. | Move Django out of `install_requires` to to `requires` block.
| Python | apache-2.0 | Afnarel/django-fluent-comments,BangorUniversity/django-fluent-comments,mgpyh/django-fluent-comments,akszydelko/django-fluent-comments,PetrDlouhy/django-fluent-comments,django-fluent/django-fluent-comments,mgpyh/django-fluent-comments,PetrDlouhy/django-fluent-comments,PetrDlouhy/django-fluent-comments,Afnarel/django-fluent-comments,edoburu/django-fluent-comments,edoburu/django-fluent-comments,django-fluent/django-fluent-comments,BangorUniversity/django-fluent-comments,akszydelko/django-fluent-comments,akszydelko/django-fluent-comments,django-fluent/django-fluent-comments,mgpyh/django-fluent-comments,Afnarel/django-fluent-comments,edoburu/django-fluent-comments,django-fluent/django-fluent-comments,BangorUniversity/django-fluent-comments | #!/usr/bin/env python
from setuptools import setup, find_packages
from os.path import dirname, join
import sys, os
# When creating the sdist, make sure the django.mo file also exists:
if 'sdist' in sys.argv:
try:
os.chdir('fluent_comments')
from django.core.management.commands.compilemessages import compile_messages
compile_messages(sys.stderr)
finally:
os.chdir('..')
setup(
name='django-fluent-comments',
version='0.8.0',
license='Apache License, Version 2.0',
install_requires=[
'Django>=1.2.0',
'django-crispy-forms>=1.1.1',
'akismet>=0.2',
],
description='A modern, ajax-based appearance for django.contrib.comments',
long_description=open('README.rst').read(),
author='Diederik van der Boor',
author_email='opensource@edoburu.nl',
url='https://github.com/edoburu/django-fluent-comments',
download_url='https://github.com/edoburu/django-fluent-comments/zipball/master',
packages=find_packages(exclude=('example*',)),
include_package_data=True,
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
]
)
Move Django out of `install_requires` to to `requires` block. | #!/usr/bin/env python
from setuptools import setup, find_packages
from os.path import dirname, join
import sys, os
# When creating the sdist, make sure the django.mo file also exists:
if 'sdist' in sys.argv:
try:
os.chdir('fluent_comments')
from django.core.management.commands.compilemessages import compile_messages
compile_messages(sys.stderr)
finally:
os.chdir('..')
setup(
name='django-fluent-comments',
version='0.8.0',
license='Apache License, Version 2.0',
install_requires=[
'django-crispy-forms>=1.1.1',
'akismet>=0.2',
],
requires=[
'Django (>=1.2)',
],
description='A modern, ajax-based appearance for django.contrib.comments',
long_description=open('README.rst').read(),
author='Diederik van der Boor',
author_email='opensource@edoburu.nl',
url='https://github.com/edoburu/django-fluent-comments',
download_url='https://github.com/edoburu/django-fluent-comments/zipball/master',
packages=find_packages(exclude=('example*',)),
include_package_data=True,
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
]
)
| <commit_before>#!/usr/bin/env python
from setuptools import setup, find_packages
from os.path import dirname, join
import sys, os
# When creating the sdist, make sure the django.mo file also exists:
if 'sdist' in sys.argv:
try:
os.chdir('fluent_comments')
from django.core.management.commands.compilemessages import compile_messages
compile_messages(sys.stderr)
finally:
os.chdir('..')
setup(
name='django-fluent-comments',
version='0.8.0',
license='Apache License, Version 2.0',
install_requires=[
'Django>=1.2.0',
'django-crispy-forms>=1.1.1',
'akismet>=0.2',
],
description='A modern, ajax-based appearance for django.contrib.comments',
long_description=open('README.rst').read(),
author='Diederik van der Boor',
author_email='opensource@edoburu.nl',
url='https://github.com/edoburu/django-fluent-comments',
download_url='https://github.com/edoburu/django-fluent-comments/zipball/master',
packages=find_packages(exclude=('example*',)),
include_package_data=True,
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
]
)
<commit_msg>Move Django out of `install_requires` to to `requires` block.<commit_after> | #!/usr/bin/env python
from setuptools import setup, find_packages
from os.path import dirname, join
import sys, os
# When creating the sdist, make sure the django.mo file also exists:
if 'sdist' in sys.argv:
try:
os.chdir('fluent_comments')
from django.core.management.commands.compilemessages import compile_messages
compile_messages(sys.stderr)
finally:
os.chdir('..')
setup(
name='django-fluent-comments',
version='0.8.0',
license='Apache License, Version 2.0',
install_requires=[
'django-crispy-forms>=1.1.1',
'akismet>=0.2',
],
requires=[
'Django (>=1.2)',
],
description='A modern, ajax-based appearance for django.contrib.comments',
long_description=open('README.rst').read(),
author='Diederik van der Boor',
author_email='opensource@edoburu.nl',
url='https://github.com/edoburu/django-fluent-comments',
download_url='https://github.com/edoburu/django-fluent-comments/zipball/master',
packages=find_packages(exclude=('example*',)),
include_package_data=True,
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
]
)
| #!/usr/bin/env python
from setuptools import setup, find_packages
from os.path import dirname, join
import sys, os
# When creating the sdist, make sure the django.mo file also exists:
if 'sdist' in sys.argv:
try:
os.chdir('fluent_comments')
from django.core.management.commands.compilemessages import compile_messages
compile_messages(sys.stderr)
finally:
os.chdir('..')
setup(
name='django-fluent-comments',
version='0.8.0',
license='Apache License, Version 2.0',
install_requires=[
'Django>=1.2.0',
'django-crispy-forms>=1.1.1',
'akismet>=0.2',
],
description='A modern, ajax-based appearance for django.contrib.comments',
long_description=open('README.rst').read(),
author='Diederik van der Boor',
author_email='opensource@edoburu.nl',
url='https://github.com/edoburu/django-fluent-comments',
download_url='https://github.com/edoburu/django-fluent-comments/zipball/master',
packages=find_packages(exclude=('example*',)),
include_package_data=True,
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
]
)
Move Django out of `install_requires` to to `requires` block.#!/usr/bin/env python
from setuptools import setup, find_packages
from os.path import dirname, join
import sys, os
# When creating the sdist, make sure the django.mo file also exists:
if 'sdist' in sys.argv:
try:
os.chdir('fluent_comments')
from django.core.management.commands.compilemessages import compile_messages
compile_messages(sys.stderr)
finally:
os.chdir('..')
setup(
name='django-fluent-comments',
version='0.8.0',
license='Apache License, Version 2.0',
install_requires=[
'django-crispy-forms>=1.1.1',
'akismet>=0.2',
],
requires=[
'Django (>=1.2)',
],
description='A modern, ajax-based appearance for django.contrib.comments',
long_description=open('README.rst').read(),
author='Diederik van der Boor',
author_email='opensource@edoburu.nl',
url='https://github.com/edoburu/django-fluent-comments',
download_url='https://github.com/edoburu/django-fluent-comments/zipball/master',
packages=find_packages(exclude=('example*',)),
include_package_data=True,
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
]
)
| <commit_before>#!/usr/bin/env python
from setuptools import setup, find_packages
from os.path import dirname, join
import sys, os
# When creating the sdist, make sure the django.mo file also exists:
if 'sdist' in sys.argv:
try:
os.chdir('fluent_comments')
from django.core.management.commands.compilemessages import compile_messages
compile_messages(sys.stderr)
finally:
os.chdir('..')
setup(
name='django-fluent-comments',
version='0.8.0',
license='Apache License, Version 2.0',
install_requires=[
'Django>=1.2.0',
'django-crispy-forms>=1.1.1',
'akismet>=0.2',
],
description='A modern, ajax-based appearance for django.contrib.comments',
long_description=open('README.rst').read(),
author='Diederik van der Boor',
author_email='opensource@edoburu.nl',
url='https://github.com/edoburu/django-fluent-comments',
download_url='https://github.com/edoburu/django-fluent-comments/zipball/master',
packages=find_packages(exclude=('example*',)),
include_package_data=True,
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
]
)
<commit_msg>Move Django out of `install_requires` to to `requires` block.<commit_after>#!/usr/bin/env python
from setuptools import setup, find_packages
from os.path import dirname, join
import sys, os
# When creating the sdist, make sure the django.mo file also exists:
if 'sdist' in sys.argv:
try:
os.chdir('fluent_comments')
from django.core.management.commands.compilemessages import compile_messages
compile_messages(sys.stderr)
finally:
os.chdir('..')
setup(
name='django-fluent-comments',
version='0.8.0',
license='Apache License, Version 2.0',
install_requires=[
'django-crispy-forms>=1.1.1',
'akismet>=0.2',
],
requires=[
'Django (>=1.2)',
],
description='A modern, ajax-based appearance for django.contrib.comments',
long_description=open('README.rst').read(),
author='Diederik van der Boor',
author_email='opensource@edoburu.nl',
url='https://github.com/edoburu/django-fluent-comments',
download_url='https://github.com/edoburu/django-fluent-comments/zipball/master',
packages=find_packages(exclude=('example*',)),
include_package_data=True,
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
]
)
|
b588213ccf61a45a1e0e96f626c528483b3aea7e | setup.py | setup.py | """
Flask-Static
---------------
Generates a static website from a Flask application.
"""
from setuptools import setup
setup(
name='Flask-Static',
version='0.1',
url='http://exyr.org/Flask-Static/',
license='BSD',
author='Simon Sapin',
author_email='simon.sapin@exyr.org',
description='Generates a static website from a Flask application',
long_description=__doc__,
packages=['flaskext'],
namespace_packages=['flaskext'],
test_suite='test_flaskstatic',
zip_safe=False,
platforms='any',
install_requires=[
'Flask',
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
| """
Flask-Static
---------------
Generates a static website from a Flask application.
"""
from setuptools import setup
setup(
name='Flask-Static',
version='0.1dev',
url='http://exyr.org/Flask-Static/',
license='BSD',
author='Simon Sapin',
author_email='simon.sapin@exyr.org',
description='Generates a static website from a Flask application',
long_description=__doc__,
packages=['flaskext'],
namespace_packages=['flaskext'],
test_suite='test_flaskstatic',
zip_safe=False,
platforms='any',
install_requires=[
'Flask',
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
| Change version number to 0.1dev | Change version number to 0.1dev
| Python | bsd-3-clause | SimonSapin/Frozen-Flask,SimonSapin/Frozen-Flask,SimonSapin/Frozen-Flask | """
Flask-Static
---------------
Generates a static website from a Flask application.
"""
from setuptools import setup
setup(
name='Flask-Static',
version='0.1',
url='http://exyr.org/Flask-Static/',
license='BSD',
author='Simon Sapin',
author_email='simon.sapin@exyr.org',
description='Generates a static website from a Flask application',
long_description=__doc__,
packages=['flaskext'],
namespace_packages=['flaskext'],
test_suite='test_flaskstatic',
zip_safe=False,
platforms='any',
install_requires=[
'Flask',
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
Change version number to 0.1dev | """
Flask-Static
---------------
Generates a static website from a Flask application.
"""
from setuptools import setup
setup(
name='Flask-Static',
version='0.1dev',
url='http://exyr.org/Flask-Static/',
license='BSD',
author='Simon Sapin',
author_email='simon.sapin@exyr.org',
description='Generates a static website from a Flask application',
long_description=__doc__,
packages=['flaskext'],
namespace_packages=['flaskext'],
test_suite='test_flaskstatic',
zip_safe=False,
platforms='any',
install_requires=[
'Flask',
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
| <commit_before>"""
Flask-Static
---------------
Generates a static website from a Flask application.
"""
from setuptools import setup
setup(
name='Flask-Static',
version='0.1',
url='http://exyr.org/Flask-Static/',
license='BSD',
author='Simon Sapin',
author_email='simon.sapin@exyr.org',
description='Generates a static website from a Flask application',
long_description=__doc__,
packages=['flaskext'],
namespace_packages=['flaskext'],
test_suite='test_flaskstatic',
zip_safe=False,
platforms='any',
install_requires=[
'Flask',
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
<commit_msg>Change version number to 0.1dev<commit_after> | """
Flask-Static
---------------
Generates a static website from a Flask application.
"""
from setuptools import setup
setup(
name='Flask-Static',
version='0.1dev',
url='http://exyr.org/Flask-Static/',
license='BSD',
author='Simon Sapin',
author_email='simon.sapin@exyr.org',
description='Generates a static website from a Flask application',
long_description=__doc__,
packages=['flaskext'],
namespace_packages=['flaskext'],
test_suite='test_flaskstatic',
zip_safe=False,
platforms='any',
install_requires=[
'Flask',
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
| """
Flask-Static
---------------
Generates a static website from a Flask application.
"""
from setuptools import setup
setup(
name='Flask-Static',
version='0.1',
url='http://exyr.org/Flask-Static/',
license='BSD',
author='Simon Sapin',
author_email='simon.sapin@exyr.org',
description='Generates a static website from a Flask application',
long_description=__doc__,
packages=['flaskext'],
namespace_packages=['flaskext'],
test_suite='test_flaskstatic',
zip_safe=False,
platforms='any',
install_requires=[
'Flask',
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
Change version number to 0.1dev"""
Flask-Static
---------------
Generates a static website from a Flask application.
"""
from setuptools import setup
setup(
name='Flask-Static',
version='0.1dev',
url='http://exyr.org/Flask-Static/',
license='BSD',
author='Simon Sapin',
author_email='simon.sapin@exyr.org',
description='Generates a static website from a Flask application',
long_description=__doc__,
packages=['flaskext'],
namespace_packages=['flaskext'],
test_suite='test_flaskstatic',
zip_safe=False,
platforms='any',
install_requires=[
'Flask',
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
| <commit_before>"""
Flask-Static
---------------
Generates a static website from a Flask application.
"""
from setuptools import setup
setup(
name='Flask-Static',
version='0.1',
url='http://exyr.org/Flask-Static/',
license='BSD',
author='Simon Sapin',
author_email='simon.sapin@exyr.org',
description='Generates a static website from a Flask application',
long_description=__doc__,
packages=['flaskext'],
namespace_packages=['flaskext'],
test_suite='test_flaskstatic',
zip_safe=False,
platforms='any',
install_requires=[
'Flask',
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
<commit_msg>Change version number to 0.1dev<commit_after>"""
Flask-Static
---------------
Generates a static website from a Flask application.
"""
from setuptools import setup
setup(
name='Flask-Static',
version='0.1dev',
url='http://exyr.org/Flask-Static/',
license='BSD',
author='Simon Sapin',
author_email='simon.sapin@exyr.org',
description='Generates a static website from a Flask application',
long_description=__doc__,
packages=['flaskext'],
namespace_packages=['flaskext'],
test_suite='test_flaskstatic',
zip_safe=False,
platforms='any',
install_requires=[
'Flask',
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
|
53440a46cb38194e12c383a718f392b9da2175b6 | setup.py | setup.py | #!/usr/bin/python
import setuptools
import numpy
from setuptools import Extension
setuptools.setup(
name='mrpt',
version='0.1',
url='http://github.com/teemupitkanen/mrpt',
install_requires=[],
packages={ '.': 'mrpt' },
zip_safe=False,
test_suite='py.test',
entry_points='',
ext_modules = [
Extension('mrptlib',
sources = [
'cpp/mrptmodule.cpp',
],
extra_compile_args=['-std=c++11', '-O3', '-march=native', '-ffast-math', '-s',
'-fno-rtti', '-fopenmp', '-DNDEBUG'],
extra_link_args=['-lgomp'],
libraries = ['stdc++'],
include_dirs = ['cpp/lib', numpy.get_include()]
)
]
)
| #!/usr/bin/python
import setuptools
import numpy
from setuptools import Extension
# Not all CPUs have march as a tuning parameter
import platform
cputune = ['-march=native',]
if platform.machine() == "ppc64le":
cputune = ['-mcpu=native',]
setuptools.setup(
name='mrpt',
version='0.1',
url='http://github.com/teemupitkanen/mrpt',
install_requires=[],
packages={ '.': 'mrpt' },
zip_safe=False,
test_suite='py.test',
entry_points='',
ext_modules = [
Extension('mrptlib',
sources = [
'cpp/mrptmodule.cpp',
],
extra_compile_args=['-std=c++11', '-O3', '-ffast-math', '-s',
'-fno-rtti', '-fopenmp', '-DNDEBUG'] + cputune,
extra_link_args=['-lgomp'],
libraries = ['stdc++'],
include_dirs = ['cpp/lib', numpy.get_include()]
)
]
)
| Solve ppc64le build due to unexisting -march | Solve ppc64le build due to unexisting -march
On ppc64le (e.g: POWER8 machines), the -mcpu should be used instead.
| Python | mit | teemupitkanen/mrpt,teemupitkanen/mrpt,teemupitkanen/mrpt | #!/usr/bin/python
import setuptools
import numpy
from setuptools import Extension
setuptools.setup(
name='mrpt',
version='0.1',
url='http://github.com/teemupitkanen/mrpt',
install_requires=[],
packages={ '.': 'mrpt' },
zip_safe=False,
test_suite='py.test',
entry_points='',
ext_modules = [
Extension('mrptlib',
sources = [
'cpp/mrptmodule.cpp',
],
extra_compile_args=['-std=c++11', '-O3', '-march=native', '-ffast-math', '-s',
'-fno-rtti', '-fopenmp', '-DNDEBUG'],
extra_link_args=['-lgomp'],
libraries = ['stdc++'],
include_dirs = ['cpp/lib', numpy.get_include()]
)
]
)
Solve ppc64le build due to unexisting -march
On ppc64le (e.g: POWER8 machines), the -mcpu should be used instead. | #!/usr/bin/python
import setuptools
import numpy
from setuptools import Extension
# Not all CPUs have march as a tuning parameter
import platform
cputune = ['-march=native',]
if platform.machine() == "ppc64le":
cputune = ['-mcpu=native',]
setuptools.setup(
name='mrpt',
version='0.1',
url='http://github.com/teemupitkanen/mrpt',
install_requires=[],
packages={ '.': 'mrpt' },
zip_safe=False,
test_suite='py.test',
entry_points='',
ext_modules = [
Extension('mrptlib',
sources = [
'cpp/mrptmodule.cpp',
],
extra_compile_args=['-std=c++11', '-O3', '-ffast-math', '-s',
'-fno-rtti', '-fopenmp', '-DNDEBUG'] + cputune,
extra_link_args=['-lgomp'],
libraries = ['stdc++'],
include_dirs = ['cpp/lib', numpy.get_include()]
)
]
)
| <commit_before>#!/usr/bin/python
import setuptools
import numpy
from setuptools import Extension
setuptools.setup(
name='mrpt',
version='0.1',
url='http://github.com/teemupitkanen/mrpt',
install_requires=[],
packages={ '.': 'mrpt' },
zip_safe=False,
test_suite='py.test',
entry_points='',
ext_modules = [
Extension('mrptlib',
sources = [
'cpp/mrptmodule.cpp',
],
extra_compile_args=['-std=c++11', '-O3', '-march=native', '-ffast-math', '-s',
'-fno-rtti', '-fopenmp', '-DNDEBUG'],
extra_link_args=['-lgomp'],
libraries = ['stdc++'],
include_dirs = ['cpp/lib', numpy.get_include()]
)
]
)
<commit_msg>Solve ppc64le build due to unexisting -march
On ppc64le (e.g: POWER8 machines), the -mcpu should be used instead.<commit_after> | #!/usr/bin/python
import setuptools
import numpy
from setuptools import Extension
# Not all CPUs have march as a tuning parameter
import platform
cputune = ['-march=native',]
if platform.machine() == "ppc64le":
cputune = ['-mcpu=native',]
setuptools.setup(
name='mrpt',
version='0.1',
url='http://github.com/teemupitkanen/mrpt',
install_requires=[],
packages={ '.': 'mrpt' },
zip_safe=False,
test_suite='py.test',
entry_points='',
ext_modules = [
Extension('mrptlib',
sources = [
'cpp/mrptmodule.cpp',
],
extra_compile_args=['-std=c++11', '-O3', '-ffast-math', '-s',
'-fno-rtti', '-fopenmp', '-DNDEBUG'] + cputune,
extra_link_args=['-lgomp'],
libraries = ['stdc++'],
include_dirs = ['cpp/lib', numpy.get_include()]
)
]
)
| #!/usr/bin/python
import setuptools
import numpy
from setuptools import Extension
setuptools.setup(
name='mrpt',
version='0.1',
url='http://github.com/teemupitkanen/mrpt',
install_requires=[],
packages={ '.': 'mrpt' },
zip_safe=False,
test_suite='py.test',
entry_points='',
ext_modules = [
Extension('mrptlib',
sources = [
'cpp/mrptmodule.cpp',
],
extra_compile_args=['-std=c++11', '-O3', '-march=native', '-ffast-math', '-s',
'-fno-rtti', '-fopenmp', '-DNDEBUG'],
extra_link_args=['-lgomp'],
libraries = ['stdc++'],
include_dirs = ['cpp/lib', numpy.get_include()]
)
]
)
Solve ppc64le build due to unexisting -march
On ppc64le (e.g: POWER8 machines), the -mcpu should be used instead.#!/usr/bin/python
import setuptools
import numpy
from setuptools import Extension
# Not all CPUs have march as a tuning parameter
import platform
cputune = ['-march=native',]
if platform.machine() == "ppc64le":
cputune = ['-mcpu=native',]
setuptools.setup(
name='mrpt',
version='0.1',
url='http://github.com/teemupitkanen/mrpt',
install_requires=[],
packages={ '.': 'mrpt' },
zip_safe=False,
test_suite='py.test',
entry_points='',
ext_modules = [
Extension('mrptlib',
sources = [
'cpp/mrptmodule.cpp',
],
extra_compile_args=['-std=c++11', '-O3', '-ffast-math', '-s',
'-fno-rtti', '-fopenmp', '-DNDEBUG'] + cputune,
extra_link_args=['-lgomp'],
libraries = ['stdc++'],
include_dirs = ['cpp/lib', numpy.get_include()]
)
]
)
| <commit_before>#!/usr/bin/python
import setuptools
import numpy
from setuptools import Extension
setuptools.setup(
name='mrpt',
version='0.1',
url='http://github.com/teemupitkanen/mrpt',
install_requires=[],
packages={ '.': 'mrpt' },
zip_safe=False,
test_suite='py.test',
entry_points='',
ext_modules = [
Extension('mrptlib',
sources = [
'cpp/mrptmodule.cpp',
],
extra_compile_args=['-std=c++11', '-O3', '-march=native', '-ffast-math', '-s',
'-fno-rtti', '-fopenmp', '-DNDEBUG'],
extra_link_args=['-lgomp'],
libraries = ['stdc++'],
include_dirs = ['cpp/lib', numpy.get_include()]
)
]
)
<commit_msg>Solve ppc64le build due to unexisting -march
On ppc64le (e.g: POWER8 machines), the -mcpu should be used instead.<commit_after>#!/usr/bin/python
import setuptools
import numpy
from setuptools import Extension
# Not all CPUs have march as a tuning parameter
import platform
cputune = ['-march=native',]
if platform.machine() == "ppc64le":
cputune = ['-mcpu=native',]
setuptools.setup(
name='mrpt',
version='0.1',
url='http://github.com/teemupitkanen/mrpt',
install_requires=[],
packages={ '.': 'mrpt' },
zip_safe=False,
test_suite='py.test',
entry_points='',
ext_modules = [
Extension('mrptlib',
sources = [
'cpp/mrptmodule.cpp',
],
extra_compile_args=['-std=c++11', '-O3', '-ffast-math', '-s',
'-fno-rtti', '-fopenmp', '-DNDEBUG'] + cputune,
extra_link_args=['-lgomp'],
libraries = ['stdc++'],
include_dirs = ['cpp/lib', numpy.get_include()]
)
]
)
|
a714511115bfee0fbdc6c70bd0abfceaa08384f6 | idlk/__init__.py | idlk/__init__.py | from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
import os
import sys
import unicodedata
import idlk.base41
if sys.version_info[0] == 3:
_get_byte = lambda c: c
else:
_get_byte = ord
def hash_macroman(data):
h = 0
for c in data:
h = ((h << 8) + h) + _get_byte(c)
return h % 0xFFFEECED
def idlk(filename):
# Normalize to NFC.
filename = unicodedata.normalize('NFC', filename)
# Convert to lowercase first.
filename = filename.lower()
# The original algorithm seems to prefer Mac Roman encoding as long as
# there are no non-mappable characters in the file name.
try:
macroman_name = filename.encode("macroman")
except UnicodeEncodeError:
pass
else:
hashed = base41.encode(hash_macroman(macroman_name))
base, ext = os.path.splitext(macroman_name)
return "~{:s}~{:s}.idlk".format(base[0:18].decode("macroman"), hashed)
# Regrettably the encoding / hashing algorithm for unicode filenames is
# not currently known. Please file a feature request/patch if you
# discover a working implementation.
return False
| """
A lock filename generator for idlk files used by a well known DTP suite.
"""
from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
import os
import sys
import unicodedata
from idlk import base41
if sys.version_info[0] == 3:
_get_byte = lambda c: c
else:
_get_byte = ord
def hash_macroman(data):
"""
Compute the hash for the given byte string.
"""
result = 0
for char in data:
result = ((result << 8) + result) + _get_byte(char)
return result % 0xFFFEECED
def idlk(filename):
"""
Generate the lock file name for the given file.
"""
# Normalize to NFC.
filename = unicodedata.normalize('NFC', filename)
# Convert to lowercase first.
filename = filename.lower()
# The original algorithm seems to prefer Mac Roman encoding as long as
# there are no non-mappable characters in the file name.
try:
macroman_name = filename.encode("macroman")
except UnicodeEncodeError:
pass
else:
hashed = base41.encode(hash_macroman(macroman_name))
base = os.path.splitext(macroman_name)[0]
return "~{:s}~{:s}.idlk".format(base[0:18].decode("macroman"), hashed)
# Regrettably the encoding / hashing algorithm for unicode filenames is
# not currently known. Please file a feature request/patch if you
# discover a working implementation.
return False
| Fix issues reported by pylint | Fix issues reported by pylint
| Python | mit | znerol/py-idlk | from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
import os
import sys
import unicodedata
import idlk.base41
if sys.version_info[0] == 3:
_get_byte = lambda c: c
else:
_get_byte = ord
def hash_macroman(data):
h = 0
for c in data:
h = ((h << 8) + h) + _get_byte(c)
return h % 0xFFFEECED
def idlk(filename):
# Normalize to NFC.
filename = unicodedata.normalize('NFC', filename)
# Convert to lowercase first.
filename = filename.lower()
# The original algorithm seems to prefer Mac Roman encoding as long as
# there are no non-mappable characters in the file name.
try:
macroman_name = filename.encode("macroman")
except UnicodeEncodeError:
pass
else:
hashed = base41.encode(hash_macroman(macroman_name))
base, ext = os.path.splitext(macroman_name)
return "~{:s}~{:s}.idlk".format(base[0:18].decode("macroman"), hashed)
# Regrettably the encoding / hashing algorithm for unicode filenames is
# not currently known. Please file a feature request/patch if you
# discover a working implementation.
return False
Fix issues reported by pylint | """
A lock filename generator for idlk files used by a well known DTP suite.
"""
from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
import os
import sys
import unicodedata
from idlk import base41
if sys.version_info[0] == 3:
_get_byte = lambda c: c
else:
_get_byte = ord
def hash_macroman(data):
"""
Compute the hash for the given byte string.
"""
result = 0
for char in data:
result = ((result << 8) + result) + _get_byte(char)
return result % 0xFFFEECED
def idlk(filename):
"""
Generate the lock file name for the given file.
"""
# Normalize to NFC.
filename = unicodedata.normalize('NFC', filename)
# Convert to lowercase first.
filename = filename.lower()
# The original algorithm seems to prefer Mac Roman encoding as long as
# there are no non-mappable characters in the file name.
try:
macroman_name = filename.encode("macroman")
except UnicodeEncodeError:
pass
else:
hashed = base41.encode(hash_macroman(macroman_name))
base = os.path.splitext(macroman_name)[0]
return "~{:s}~{:s}.idlk".format(base[0:18].decode("macroman"), hashed)
# Regrettably the encoding / hashing algorithm for unicode filenames is
# not currently known. Please file a feature request/patch if you
# discover a working implementation.
return False
| <commit_before>from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
import os
import sys
import unicodedata
import idlk.base41
if sys.version_info[0] == 3:
_get_byte = lambda c: c
else:
_get_byte = ord
def hash_macroman(data):
h = 0
for c in data:
h = ((h << 8) + h) + _get_byte(c)
return h % 0xFFFEECED
def idlk(filename):
# Normalize to NFC.
filename = unicodedata.normalize('NFC', filename)
# Convert to lowercase first.
filename = filename.lower()
# The original algorithm seems to prefer Mac Roman encoding as long as
# there are no non-mappable characters in the file name.
try:
macroman_name = filename.encode("macroman")
except UnicodeEncodeError:
pass
else:
hashed = base41.encode(hash_macroman(macroman_name))
base, ext = os.path.splitext(macroman_name)
return "~{:s}~{:s}.idlk".format(base[0:18].decode("macroman"), hashed)
# Regrettably the encoding / hashing algorithm for unicode filenames is
# not currently known. Please file a feature request/patch if you
# discover a working implementation.
return False
<commit_msg>Fix issues reported by pylint<commit_after> | """
A lock filename generator for idlk files used by a well known DTP suite.
"""
from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
import os
import sys
import unicodedata
from idlk import base41
if sys.version_info[0] == 3:
_get_byte = lambda c: c
else:
_get_byte = ord
def hash_macroman(data):
"""
Compute the hash for the given byte string.
"""
result = 0
for char in data:
result = ((result << 8) + result) + _get_byte(char)
return result % 0xFFFEECED
def idlk(filename):
"""
Generate the lock file name for the given file.
"""
# Normalize to NFC.
filename = unicodedata.normalize('NFC', filename)
# Convert to lowercase first.
filename = filename.lower()
# The original algorithm seems to prefer Mac Roman encoding as long as
# there are no non-mappable characters in the file name.
try:
macroman_name = filename.encode("macroman")
except UnicodeEncodeError:
pass
else:
hashed = base41.encode(hash_macroman(macroman_name))
base = os.path.splitext(macroman_name)[0]
return "~{:s}~{:s}.idlk".format(base[0:18].decode("macroman"), hashed)
# Regrettably the encoding / hashing algorithm for unicode filenames is
# not currently known. Please file a feature request/patch if you
# discover a working implementation.
return False
| from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
import os
import sys
import unicodedata
import idlk.base41
if sys.version_info[0] == 3:
_get_byte = lambda c: c
else:
_get_byte = ord
def hash_macroman(data):
h = 0
for c in data:
h = ((h << 8) + h) + _get_byte(c)
return h % 0xFFFEECED
def idlk(filename):
# Normalize to NFC.
filename = unicodedata.normalize('NFC', filename)
# Convert to lowercase first.
filename = filename.lower()
# The original algorithm seems to prefer Mac Roman encoding as long as
# there are no non-mappable characters in the file name.
try:
macroman_name = filename.encode("macroman")
except UnicodeEncodeError:
pass
else:
hashed = base41.encode(hash_macroman(macroman_name))
base, ext = os.path.splitext(macroman_name)
return "~{:s}~{:s}.idlk".format(base[0:18].decode("macroman"), hashed)
# Regrettably the encoding / hashing algorithm for unicode filenames is
# not currently known. Please file a feature request/patch if you
# discover a working implementation.
return False
Fix issues reported by pylint"""
A lock filename generator for idlk files used by a well known DTP suite.
"""
from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
import os
import sys
import unicodedata
from idlk import base41
if sys.version_info[0] == 3:
_get_byte = lambda c: c
else:
_get_byte = ord
def hash_macroman(data):
"""
Compute the hash for the given byte string.
"""
result = 0
for char in data:
result = ((result << 8) + result) + _get_byte(char)
return result % 0xFFFEECED
def idlk(filename):
"""
Generate the lock file name for the given file.
"""
# Normalize to NFC.
filename = unicodedata.normalize('NFC', filename)
# Convert to lowercase first.
filename = filename.lower()
# The original algorithm seems to prefer Mac Roman encoding as long as
# there are no non-mappable characters in the file name.
try:
macroman_name = filename.encode("macroman")
except UnicodeEncodeError:
pass
else:
hashed = base41.encode(hash_macroman(macroman_name))
base = os.path.splitext(macroman_name)[0]
return "~{:s}~{:s}.idlk".format(base[0:18].decode("macroman"), hashed)
# Regrettably the encoding / hashing algorithm for unicode filenames is
# not currently known. Please file a feature request/patch if you
# discover a working implementation.
return False
| <commit_before>from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
import os
import sys
import unicodedata
import idlk.base41
if sys.version_info[0] == 3:
_get_byte = lambda c: c
else:
_get_byte = ord
def hash_macroman(data):
h = 0
for c in data:
h = ((h << 8) + h) + _get_byte(c)
return h % 0xFFFEECED
def idlk(filename):
# Normalize to NFC.
filename = unicodedata.normalize('NFC', filename)
# Convert to lowercase first.
filename = filename.lower()
# The original algorithm seems to prefer Mac Roman encoding as long as
# there are no non-mappable characters in the file name.
try:
macroman_name = filename.encode("macroman")
except UnicodeEncodeError:
pass
else:
hashed = base41.encode(hash_macroman(macroman_name))
base, ext = os.path.splitext(macroman_name)
return "~{:s}~{:s}.idlk".format(base[0:18].decode("macroman"), hashed)
# Regrettably the encoding / hashing algorithm for unicode filenames is
# not currently known. Please file a feature request/patch if you
# discover a working implementation.
return False
<commit_msg>Fix issues reported by pylint<commit_after>"""
A lock filename generator for idlk files used by a well known DTP suite.
"""
from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
import os
import sys
import unicodedata
from idlk import base41
if sys.version_info[0] == 3:
_get_byte = lambda c: c
else:
_get_byte = ord
def hash_macroman(data):
"""
Compute the hash for the given byte string.
"""
result = 0
for char in data:
result = ((result << 8) + result) + _get_byte(char)
return result % 0xFFFEECED
def idlk(filename):
"""
Generate the lock file name for the given file.
"""
# Normalize to NFC.
filename = unicodedata.normalize('NFC', filename)
# Convert to lowercase first.
filename = filename.lower()
# The original algorithm seems to prefer Mac Roman encoding as long as
# there are no non-mappable characters in the file name.
try:
macroman_name = filename.encode("macroman")
except UnicodeEncodeError:
pass
else:
hashed = base41.encode(hash_macroman(macroman_name))
base = os.path.splitext(macroman_name)[0]
return "~{:s}~{:s}.idlk".format(base[0:18].decode("macroman"), hashed)
# Regrettably the encoding / hashing algorithm for unicode filenames is
# not currently known. Please file a feature request/patch if you
# discover a working implementation.
return False
|
3bc0876e7bae2cfb62724f1e5dce1a93f71b7252 | docstring_parser/parser/__init__.py | docstring_parser/parser/__init__.py | """Docstring parsing."""
from . import rest
from .common import ParseError
_styles = {"rest": rest.parse}
def parse(text: str, style: str = "auto"):
"""
Parse the docstring into its components.
:param str text: docstring text to parse
:param text style: docstring style, choose from: 'rest', 'auto'
:returns: parsed docstring
"""
if style == "auto":
rets = []
for _parse in _styles.values():
try:
rets.append(_parse(text))
except ParseError as e:
exc = e
if not rets:
raise exc
return sorted(rets, key=lambda ret: len(ret.meta), reverse=True)[0]
else:
return _styles[style]
| """Docstring parsing."""
from . import rest
from .common import ParseError, Docstring
_styles = {"rest": rest.parse}
def _parse_score(docstring: Docstring) -> int:
"""
Produce a score for the parsing.
:param Docstring docstring: parsed docstring representation
:returns int: parse score, higher is better
"""
score = 0
if docstring.short_description:
score += 1
if docstring.long_description:
score += docstring.long_description.count('\n')
score += len(docstring.params)
score += len(docstring.raises)
if docstring.returns:
score += 2
return score
def parse(text: str, style: str = 'auto') -> Docstring:
"""
Parse the docstring into its components.
:param str text: docstring text to parse
:param str style: docstring style, choose from: 'rest', 'auto'
:returns Docstring: parsed docstring representation
"""
if style == 'auto':
rets = []
for _parse in _styles.values():
try:
rets.append(_parse(text))
except ParseError as e:
exc = e
if not rets:
raise exc
return sorted(rets, key=_parse_score, reverse=True)[0]
else:
return _styles[style](text)
| Fix parsing when style specified, add 'auto' score | Fix parsing when style specified, add 'auto' score
| Python | mit | rr-/docstring_parser | """Docstring parsing."""
from . import rest
from .common import ParseError
_styles = {"rest": rest.parse}
def parse(text: str, style: str = "auto"):
"""
Parse the docstring into its components.
:param str text: docstring text to parse
:param text style: docstring style, choose from: 'rest', 'auto'
:returns: parsed docstring
"""
if style == "auto":
rets = []
for _parse in _styles.values():
try:
rets.append(_parse(text))
except ParseError as e:
exc = e
if not rets:
raise exc
return sorted(rets, key=lambda ret: len(ret.meta), reverse=True)[0]
else:
return _styles[style]
Fix parsing when style specified, add 'auto' score | """Docstring parsing."""
from . import rest
from .common import ParseError, Docstring
_styles = {"rest": rest.parse}
def _parse_score(docstring: Docstring) -> int:
"""
Produce a score for the parsing.
:param Docstring docstring: parsed docstring representation
:returns int: parse score, higher is better
"""
score = 0
if docstring.short_description:
score += 1
if docstring.long_description:
score += docstring.long_description.count('\n')
score += len(docstring.params)
score += len(docstring.raises)
if docstring.returns:
score += 2
return score
def parse(text: str, style: str = 'auto') -> Docstring:
"""
Parse the docstring into its components.
:param str text: docstring text to parse
:param str style: docstring style, choose from: 'rest', 'auto'
:returns Docstring: parsed docstring representation
"""
if style == 'auto':
rets = []
for _parse in _styles.values():
try:
rets.append(_parse(text))
except ParseError as e:
exc = e
if not rets:
raise exc
return sorted(rets, key=_parse_score, reverse=True)[0]
else:
return _styles[style](text)
| <commit_before>"""Docstring parsing."""
from . import rest
from .common import ParseError
_styles = {"rest": rest.parse}
def parse(text: str, style: str = "auto"):
"""
Parse the docstring into its components.
:param str text: docstring text to parse
:param text style: docstring style, choose from: 'rest', 'auto'
:returns: parsed docstring
"""
if style == "auto":
rets = []
for _parse in _styles.values():
try:
rets.append(_parse(text))
except ParseError as e:
exc = e
if not rets:
raise exc
return sorted(rets, key=lambda ret: len(ret.meta), reverse=True)[0]
else:
return _styles[style]
<commit_msg>Fix parsing when style specified, add 'auto' score<commit_after> | """Docstring parsing."""
from . import rest
from .common import ParseError, Docstring
_styles = {"rest": rest.parse}
def _parse_score(docstring: Docstring) -> int:
"""
Produce a score for the parsing.
:param Docstring docstring: parsed docstring representation
:returns int: parse score, higher is better
"""
score = 0
if docstring.short_description:
score += 1
if docstring.long_description:
score += docstring.long_description.count('\n')
score += len(docstring.params)
score += len(docstring.raises)
if docstring.returns:
score += 2
return score
def parse(text: str, style: str = 'auto') -> Docstring:
"""
Parse the docstring into its components.
:param str text: docstring text to parse
:param str style: docstring style, choose from: 'rest', 'auto'
:returns Docstring: parsed docstring representation
"""
if style == 'auto':
rets = []
for _parse in _styles.values():
try:
rets.append(_parse(text))
except ParseError as e:
exc = e
if not rets:
raise exc
return sorted(rets, key=_parse_score, reverse=True)[0]
else:
return _styles[style](text)
| """Docstring parsing."""
from . import rest
from .common import ParseError
_styles = {"rest": rest.parse}
def parse(text: str, style: str = "auto"):
"""
Parse the docstring into its components.
:param str text: docstring text to parse
:param text style: docstring style, choose from: 'rest', 'auto'
:returns: parsed docstring
"""
if style == "auto":
rets = []
for _parse in _styles.values():
try:
rets.append(_parse(text))
except ParseError as e:
exc = e
if not rets:
raise exc
return sorted(rets, key=lambda ret: len(ret.meta), reverse=True)[0]
else:
return _styles[style]
Fix parsing when style specified, add 'auto' score"""Docstring parsing."""
from . import rest
from .common import ParseError, Docstring
_styles = {"rest": rest.parse}
def _parse_score(docstring: Docstring) -> int:
"""
Produce a score for the parsing.
:param Docstring docstring: parsed docstring representation
:returns int: parse score, higher is better
"""
score = 0
if docstring.short_description:
score += 1
if docstring.long_description:
score += docstring.long_description.count('\n')
score += len(docstring.params)
score += len(docstring.raises)
if docstring.returns:
score += 2
return score
def parse(text: str, style: str = 'auto') -> Docstring:
"""
Parse the docstring into its components.
:param str text: docstring text to parse
:param str style: docstring style, choose from: 'rest', 'auto'
:returns Docstring: parsed docstring representation
"""
if style == 'auto':
rets = []
for _parse in _styles.values():
try:
rets.append(_parse(text))
except ParseError as e:
exc = e
if not rets:
raise exc
return sorted(rets, key=_parse_score, reverse=True)[0]
else:
return _styles[style](text)
| <commit_before>"""Docstring parsing."""
from . import rest
from .common import ParseError
_styles = {"rest": rest.parse}
def parse(text: str, style: str = "auto"):
"""
Parse the docstring into its components.
:param str text: docstring text to parse
:param text style: docstring style, choose from: 'rest', 'auto'
:returns: parsed docstring
"""
if style == "auto":
rets = []
for _parse in _styles.values():
try:
rets.append(_parse(text))
except ParseError as e:
exc = e
if not rets:
raise exc
return sorted(rets, key=lambda ret: len(ret.meta), reverse=True)[0]
else:
return _styles[style]
<commit_msg>Fix parsing when style specified, add 'auto' score<commit_after>"""Docstring parsing."""
from . import rest
from .common import ParseError, Docstring
_styles = {"rest": rest.parse}
def _parse_score(docstring: Docstring) -> int:
"""
Produce a score for the parsing.
:param Docstring docstring: parsed docstring representation
:returns int: parse score, higher is better
"""
score = 0
if docstring.short_description:
score += 1
if docstring.long_description:
score += docstring.long_description.count('\n')
score += len(docstring.params)
score += len(docstring.raises)
if docstring.returns:
score += 2
return score
def parse(text: str, style: str = 'auto') -> Docstring:
"""
Parse the docstring into its components.
:param str text: docstring text to parse
:param str style: docstring style, choose from: 'rest', 'auto'
:returns Docstring: parsed docstring representation
"""
if style == 'auto':
rets = []
for _parse in _styles.values():
try:
rets.append(_parse(text))
except ParseError as e:
exc = e
if not rets:
raise exc
return sorted(rets, key=_parse_score, reverse=True)[0]
else:
return _styles[style](text)
|
0e1bdcb4e6d2404bb832ab86ec7bf526c1c90bbb | teami18n/teami18n/models.py | teami18n/teami18n/models.py | from django.db import models
from django_countries import countries
class Country(models.Model):
code = models.CharField(max_length=2, choices=tuple(countries),
unique=True)
class Podcast(models.Model):
story_id = models.CharField(max_length=16, unique=True)
link = models.URLField()
title = models.TextField()
teaser = models.TextField()
program_name = models.TextField(blank=True)
show_date = models.DateTimeField(null=True, blank=True)
image_link = models.URLField(null=True, blank=True)
countries = models.ManyToManyField(Country, related_name="podcasts")
| from django.db import models
from django_countries import countries
class Country(models.Model):
code = models.CharField(max_length=2, choices=tuple(countries),
unique=True)
def __unicode__(self):
return self.code
class Podcast(models.Model):
story_id = models.CharField(max_length=16, unique=True)
link = models.URLField()
title = models.TextField()
teaser = models.TextField()
program_name = models.TextField(blank=True)
show_date = models.DateTimeField(null=True, blank=True)
image_link = models.URLField(null=True, blank=True)
countries = models.ManyToManyField(Country, related_name="podcasts")
def __unicode__(self):
return self.title
| Add nice name for working in the shell | Add nice name for working in the shell
| Python | mit | team-i18n/hackaway,team-i18n/hackaway,team-i18n/hackaway | from django.db import models
from django_countries import countries
class Country(models.Model):
code = models.CharField(max_length=2, choices=tuple(countries),
unique=True)
class Podcast(models.Model):
story_id = models.CharField(max_length=16, unique=True)
link = models.URLField()
title = models.TextField()
teaser = models.TextField()
program_name = models.TextField(blank=True)
show_date = models.DateTimeField(null=True, blank=True)
image_link = models.URLField(null=True, blank=True)
countries = models.ManyToManyField(Country, related_name="podcasts")
Add nice name for working in the shell | from django.db import models
from django_countries import countries
class Country(models.Model):
code = models.CharField(max_length=2, choices=tuple(countries),
unique=True)
def __unicode__(self):
return self.code
class Podcast(models.Model):
story_id = models.CharField(max_length=16, unique=True)
link = models.URLField()
title = models.TextField()
teaser = models.TextField()
program_name = models.TextField(blank=True)
show_date = models.DateTimeField(null=True, blank=True)
image_link = models.URLField(null=True, blank=True)
countries = models.ManyToManyField(Country, related_name="podcasts")
def __unicode__(self):
return self.title
| <commit_before>from django.db import models
from django_countries import countries
class Country(models.Model):
code = models.CharField(max_length=2, choices=tuple(countries),
unique=True)
class Podcast(models.Model):
story_id = models.CharField(max_length=16, unique=True)
link = models.URLField()
title = models.TextField()
teaser = models.TextField()
program_name = models.TextField(blank=True)
show_date = models.DateTimeField(null=True, blank=True)
image_link = models.URLField(null=True, blank=True)
countries = models.ManyToManyField(Country, related_name="podcasts")
<commit_msg>Add nice name for working in the shell<commit_after> | from django.db import models
from django_countries import countries
class Country(models.Model):
code = models.CharField(max_length=2, choices=tuple(countries),
unique=True)
def __unicode__(self):
return self.code
class Podcast(models.Model):
story_id = models.CharField(max_length=16, unique=True)
link = models.URLField()
title = models.TextField()
teaser = models.TextField()
program_name = models.TextField(blank=True)
show_date = models.DateTimeField(null=True, blank=True)
image_link = models.URLField(null=True, blank=True)
countries = models.ManyToManyField(Country, related_name="podcasts")
def __unicode__(self):
return self.title
| from django.db import models
from django_countries import countries
class Country(models.Model):
code = models.CharField(max_length=2, choices=tuple(countries),
unique=True)
class Podcast(models.Model):
story_id = models.CharField(max_length=16, unique=True)
link = models.URLField()
title = models.TextField()
teaser = models.TextField()
program_name = models.TextField(blank=True)
show_date = models.DateTimeField(null=True, blank=True)
image_link = models.URLField(null=True, blank=True)
countries = models.ManyToManyField(Country, related_name="podcasts")
Add nice name for working in the shellfrom django.db import models
from django_countries import countries
class Country(models.Model):
code = models.CharField(max_length=2, choices=tuple(countries),
unique=True)
def __unicode__(self):
return self.code
class Podcast(models.Model):
story_id = models.CharField(max_length=16, unique=True)
link = models.URLField()
title = models.TextField()
teaser = models.TextField()
program_name = models.TextField(blank=True)
show_date = models.DateTimeField(null=True, blank=True)
image_link = models.URLField(null=True, blank=True)
countries = models.ManyToManyField(Country, related_name="podcasts")
def __unicode__(self):
return self.title
| <commit_before>from django.db import models
from django_countries import countries
class Country(models.Model):
code = models.CharField(max_length=2, choices=tuple(countries),
unique=True)
class Podcast(models.Model):
story_id = models.CharField(max_length=16, unique=True)
link = models.URLField()
title = models.TextField()
teaser = models.TextField()
program_name = models.TextField(blank=True)
show_date = models.DateTimeField(null=True, blank=True)
image_link = models.URLField(null=True, blank=True)
countries = models.ManyToManyField(Country, related_name="podcasts")
<commit_msg>Add nice name for working in the shell<commit_after>from django.db import models
from django_countries import countries
class Country(models.Model):
code = models.CharField(max_length=2, choices=tuple(countries),
unique=True)
def __unicode__(self):
return self.code
class Podcast(models.Model):
story_id = models.CharField(max_length=16, unique=True)
link = models.URLField()
title = models.TextField()
teaser = models.TextField()
program_name = models.TextField(blank=True)
show_date = models.DateTimeField(null=True, blank=True)
image_link = models.URLField(null=True, blank=True)
countries = models.ManyToManyField(Country, related_name="podcasts")
def __unicode__(self):
return self.title
|
e1a0e3e6895ce14822b111ee17b182a79b7b28c9 | miniraf/calc.py | miniraf/calc.py | def create_parser(subparsers):
pass
| import argparse
from astropy.io import fits
import sys
OP_MAP = {"+": lambda x, y: x + y,
"-": lambda x, y: x - y,
"*": lambda x, y: x * y,
"/": lambda x, y: x / y}
def create_parser(subparsers):
parser_calc = subparsers.add_parser("calc", help="calc help")
parser_calc.add_argument("-o", "--output", metavar="OUTFILE", default=sys.stdout.buffer)
parser_calc.add_argument("file1")
parser_calc.add_argument("op", choices=["+", "-", "*", "/"])
parser_calc.add_argument("file2")
parser_calc.set_defaults(func=main)
def load_fits_data(filename):
with fits.open(filename) as f:
data = f[0].data
return data
def main(args):
a, b = load_fits_data(args.file1), load_fits_data(args.file2)
result = OP_MAP[args.op](a, b)
hdu = fits.PrimaryHDU(result)
hdu.writeto(args.output)
| Add simple four-function output option | Add simple four-function output option
Signed-off-by: Lizhou Sha <d6acb26e253550574bc1141efa0eb5e6de15daeb@mit.edu>
| Python | mit | vulpicastor/miniraf | def create_parser(subparsers):
pass
Add simple four-function output option
Signed-off-by: Lizhou Sha <d6acb26e253550574bc1141efa0eb5e6de15daeb@mit.edu> | import argparse
from astropy.io import fits
import sys
OP_MAP = {"+": lambda x, y: x + y,
"-": lambda x, y: x - y,
"*": lambda x, y: x * y,
"/": lambda x, y: x / y}
def create_parser(subparsers):
parser_calc = subparsers.add_parser("calc", help="calc help")
parser_calc.add_argument("-o", "--output", metavar="OUTFILE", default=sys.stdout.buffer)
parser_calc.add_argument("file1")
parser_calc.add_argument("op", choices=["+", "-", "*", "/"])
parser_calc.add_argument("file2")
parser_calc.set_defaults(func=main)
def load_fits_data(filename):
with fits.open(filename) as f:
data = f[0].data
return data
def main(args):
a, b = load_fits_data(args.file1), load_fits_data(args.file2)
result = OP_MAP[args.op](a, b)
hdu = fits.PrimaryHDU(result)
hdu.writeto(args.output)
| <commit_before>def create_parser(subparsers):
pass
<commit_msg>Add simple four-function output option
Signed-off-by: Lizhou Sha <d6acb26e253550574bc1141efa0eb5e6de15daeb@mit.edu><commit_after> | import argparse
from astropy.io import fits
import sys
OP_MAP = {"+": lambda x, y: x + y,
"-": lambda x, y: x - y,
"*": lambda x, y: x * y,
"/": lambda x, y: x / y}
def create_parser(subparsers):
parser_calc = subparsers.add_parser("calc", help="calc help")
parser_calc.add_argument("-o", "--output", metavar="OUTFILE", default=sys.stdout.buffer)
parser_calc.add_argument("file1")
parser_calc.add_argument("op", choices=["+", "-", "*", "/"])
parser_calc.add_argument("file2")
parser_calc.set_defaults(func=main)
def load_fits_data(filename):
with fits.open(filename) as f:
data = f[0].data
return data
def main(args):
a, b = load_fits_data(args.file1), load_fits_data(args.file2)
result = OP_MAP[args.op](a, b)
hdu = fits.PrimaryHDU(result)
hdu.writeto(args.output)
| def create_parser(subparsers):
pass
Add simple four-function output option
Signed-off-by: Lizhou Sha <d6acb26e253550574bc1141efa0eb5e6de15daeb@mit.edu>import argparse
from astropy.io import fits
import sys
OP_MAP = {"+": lambda x, y: x + y,
"-": lambda x, y: x - y,
"*": lambda x, y: x * y,
"/": lambda x, y: x / y}
def create_parser(subparsers):
parser_calc = subparsers.add_parser("calc", help="calc help")
parser_calc.add_argument("-o", "--output", metavar="OUTFILE", default=sys.stdout.buffer)
parser_calc.add_argument("file1")
parser_calc.add_argument("op", choices=["+", "-", "*", "/"])
parser_calc.add_argument("file2")
parser_calc.set_defaults(func=main)
def load_fits_data(filename):
with fits.open(filename) as f:
data = f[0].data
return data
def main(args):
a, b = load_fits_data(args.file1), load_fits_data(args.file2)
result = OP_MAP[args.op](a, b)
hdu = fits.PrimaryHDU(result)
hdu.writeto(args.output)
| <commit_before>def create_parser(subparsers):
pass
<commit_msg>Add simple four-function output option
Signed-off-by: Lizhou Sha <d6acb26e253550574bc1141efa0eb5e6de15daeb@mit.edu><commit_after>import argparse
from astropy.io import fits
import sys
OP_MAP = {"+": lambda x, y: x + y,
"-": lambda x, y: x - y,
"*": lambda x, y: x * y,
"/": lambda x, y: x / y}
def create_parser(subparsers):
parser_calc = subparsers.add_parser("calc", help="calc help")
parser_calc.add_argument("-o", "--output", metavar="OUTFILE", default=sys.stdout.buffer)
parser_calc.add_argument("file1")
parser_calc.add_argument("op", choices=["+", "-", "*", "/"])
parser_calc.add_argument("file2")
parser_calc.set_defaults(func=main)
def load_fits_data(filename):
with fits.open(filename) as f:
data = f[0].data
return data
def main(args):
a, b = load_fits_data(args.file1), load_fits_data(args.file2)
result = OP_MAP[args.op](a, b)
hdu = fits.PrimaryHDU(result)
hdu.writeto(args.output)
|
8ca16832b54c887e6e3a84d7018181bf7e55fba0 | comrade/core/context_processors.py | comrade/core/context_processors.py | from django.conf import settings
from django.contrib.sites.models import Site
from settings import DeploymentType
def default(request):
context = {}
context['DEPLOYMENT'] = settings.DEPLOYMENT
context['site'] = Site.objects.get_current()
if settings.DEPLOYMENT != DeploymentType.PRODUCTION:
context['GIT_COMMIT'] = settings.GIT_COMMIT
return context
| from django.conf import settings
from django.contrib.sites.models import Site
from settings import DeploymentType
def default(request):
context = {}
context['DEPLOYMENT'] = settings.DEPLOYMENT
context['site'] = Site.objects.get_current()
if settings.DEPLOYMENT != DeploymentType.PRODUCTION:
context['GIT_COMMIT'] = settings.GIT_COMMIT
return context
def ssl_media(request):
if request.is_secure():
ssl_media_url = settings.MEDIA_URL.replace('http://','https://')
else:
ssl_media_url = settings.MEDIA_URL
return {'MEDIA_URL': ssl_media_url}
| Add SSL media context processor. | Add SSL media context processor.
| Python | mit | bueda/django-comrade | from django.conf import settings
from django.contrib.sites.models import Site
from settings import DeploymentType
def default(request):
context = {}
context['DEPLOYMENT'] = settings.DEPLOYMENT
context['site'] = Site.objects.get_current()
if settings.DEPLOYMENT != DeploymentType.PRODUCTION:
context['GIT_COMMIT'] = settings.GIT_COMMIT
return context
Add SSL media context processor. | from django.conf import settings
from django.contrib.sites.models import Site
from settings import DeploymentType
def default(request):
context = {}
context['DEPLOYMENT'] = settings.DEPLOYMENT
context['site'] = Site.objects.get_current()
if settings.DEPLOYMENT != DeploymentType.PRODUCTION:
context['GIT_COMMIT'] = settings.GIT_COMMIT
return context
def ssl_media(request):
if request.is_secure():
ssl_media_url = settings.MEDIA_URL.replace('http://','https://')
else:
ssl_media_url = settings.MEDIA_URL
return {'MEDIA_URL': ssl_media_url}
| <commit_before>from django.conf import settings
from django.contrib.sites.models import Site
from settings import DeploymentType
def default(request):
context = {}
context['DEPLOYMENT'] = settings.DEPLOYMENT
context['site'] = Site.objects.get_current()
if settings.DEPLOYMENT != DeploymentType.PRODUCTION:
context['GIT_COMMIT'] = settings.GIT_COMMIT
return context
<commit_msg>Add SSL media context processor.<commit_after> | from django.conf import settings
from django.contrib.sites.models import Site
from settings import DeploymentType
def default(request):
context = {}
context['DEPLOYMENT'] = settings.DEPLOYMENT
context['site'] = Site.objects.get_current()
if settings.DEPLOYMENT != DeploymentType.PRODUCTION:
context['GIT_COMMIT'] = settings.GIT_COMMIT
return context
def ssl_media(request):
if request.is_secure():
ssl_media_url = settings.MEDIA_URL.replace('http://','https://')
else:
ssl_media_url = settings.MEDIA_URL
return {'MEDIA_URL': ssl_media_url}
| from django.conf import settings
from django.contrib.sites.models import Site
from settings import DeploymentType
def default(request):
context = {}
context['DEPLOYMENT'] = settings.DEPLOYMENT
context['site'] = Site.objects.get_current()
if settings.DEPLOYMENT != DeploymentType.PRODUCTION:
context['GIT_COMMIT'] = settings.GIT_COMMIT
return context
Add SSL media context processor.from django.conf import settings
from django.contrib.sites.models import Site
from settings import DeploymentType
def default(request):
context = {}
context['DEPLOYMENT'] = settings.DEPLOYMENT
context['site'] = Site.objects.get_current()
if settings.DEPLOYMENT != DeploymentType.PRODUCTION:
context['GIT_COMMIT'] = settings.GIT_COMMIT
return context
def ssl_media(request):
if request.is_secure():
ssl_media_url = settings.MEDIA_URL.replace('http://','https://')
else:
ssl_media_url = settings.MEDIA_URL
return {'MEDIA_URL': ssl_media_url}
| <commit_before>from django.conf import settings
from django.contrib.sites.models import Site
from settings import DeploymentType
def default(request):
context = {}
context['DEPLOYMENT'] = settings.DEPLOYMENT
context['site'] = Site.objects.get_current()
if settings.DEPLOYMENT != DeploymentType.PRODUCTION:
context['GIT_COMMIT'] = settings.GIT_COMMIT
return context
<commit_msg>Add SSL media context processor.<commit_after>from django.conf import settings
from django.contrib.sites.models import Site
from settings import DeploymentType
def default(request):
context = {}
context['DEPLOYMENT'] = settings.DEPLOYMENT
context['site'] = Site.objects.get_current()
if settings.DEPLOYMENT != DeploymentType.PRODUCTION:
context['GIT_COMMIT'] = settings.GIT_COMMIT
return context
def ssl_media(request):
if request.is_secure():
ssl_media_url = settings.MEDIA_URL.replace('http://','https://')
else:
ssl_media_url = settings.MEDIA_URL
return {'MEDIA_URL': ssl_media_url}
|
7a85762ead43d8ba75547488eecda120417e8c2a | lib/python/opendiamond/helpers.py | lib/python/opendiamond/helpers.py | #
# The OpenDiamond Platform for Interactive Search
# Version 4
#
# Copyright (c) 2009 Carnegie Mellon University
# All rights reserved.
#
# This software is distributed under the terms of the Eclipse Public
# License, Version 1.0 which can be found in the file named LICENSE.
# ANY USE, REPRODUCTION OR DISTRIBUTION OF THIS SOFTWARE CONSTITUTES
# RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT
#
from subprocess import Popen, PIPE
def _GenerateCookie(scopelist, servers):
cmd = ["cookiecutter"]
for server in servers:
cmd.extend(['-s', server])
for url in scopelist:
cmd.extend(['-u', url])
return Popen(cmd, stdout=PIPE).stdout.read()
def GenerateCookie(scopelist, servers, proxies=None):
if not proxies:
return _GenerateCookie(scopelist, servers)
cookie = []
n = len(proxies)
for i in range(n):
scope = [ '/proxy/%dof%d/%s:5873%s' % (i+1, n, server, scope)
for scope in scopelist for server in servers ]
cookie.append(_GenerateCookie(scope, (proxies[i],)))
return ''.join(cookie)
| #
# The OpenDiamond Platform for Interactive Search
# Version 4
#
# Copyright (c) 2009 Carnegie Mellon University
# All rights reserved.
#
# This software is distributed under the terms of the Eclipse Public
# License, Version 1.0 which can be found in the file named LICENSE.
# ANY USE, REPRODUCTION OR DISTRIBUTION OF THIS SOFTWARE CONSTITUTES
# RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT
#
| Remove Python wrapper function for executing cookiecutter program | Remove Python wrapper function for executing cookiecutter program
| Python | epl-1.0 | cmusatyalab/opendiamond,cmusatyalab/opendiamond,cmusatyalab/opendiamond,cmusatyalab/opendiamond,cmusatyalab/opendiamond | #
# The OpenDiamond Platform for Interactive Search
# Version 4
#
# Copyright (c) 2009 Carnegie Mellon University
# All rights reserved.
#
# This software is distributed under the terms of the Eclipse Public
# License, Version 1.0 which can be found in the file named LICENSE.
# ANY USE, REPRODUCTION OR DISTRIBUTION OF THIS SOFTWARE CONSTITUTES
# RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT
#
from subprocess import Popen, PIPE
def _GenerateCookie(scopelist, servers):
cmd = ["cookiecutter"]
for server in servers:
cmd.extend(['-s', server])
for url in scopelist:
cmd.extend(['-u', url])
return Popen(cmd, stdout=PIPE).stdout.read()
def GenerateCookie(scopelist, servers, proxies=None):
if not proxies:
return _GenerateCookie(scopelist, servers)
cookie = []
n = len(proxies)
for i in range(n):
scope = [ '/proxy/%dof%d/%s:5873%s' % (i+1, n, server, scope)
for scope in scopelist for server in servers ]
cookie.append(_GenerateCookie(scope, (proxies[i],)))
return ''.join(cookie)
Remove Python wrapper function for executing cookiecutter program | #
# The OpenDiamond Platform for Interactive Search
# Version 4
#
# Copyright (c) 2009 Carnegie Mellon University
# All rights reserved.
#
# This software is distributed under the terms of the Eclipse Public
# License, Version 1.0 which can be found in the file named LICENSE.
# ANY USE, REPRODUCTION OR DISTRIBUTION OF THIS SOFTWARE CONSTITUTES
# RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT
#
| <commit_before>#
# The OpenDiamond Platform for Interactive Search
# Version 4
#
# Copyright (c) 2009 Carnegie Mellon University
# All rights reserved.
#
# This software is distributed under the terms of the Eclipse Public
# License, Version 1.0 which can be found in the file named LICENSE.
# ANY USE, REPRODUCTION OR DISTRIBUTION OF THIS SOFTWARE CONSTITUTES
# RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT
#
from subprocess import Popen, PIPE
def _GenerateCookie(scopelist, servers):
cmd = ["cookiecutter"]
for server in servers:
cmd.extend(['-s', server])
for url in scopelist:
cmd.extend(['-u', url])
return Popen(cmd, stdout=PIPE).stdout.read()
def GenerateCookie(scopelist, servers, proxies=None):
if not proxies:
return _GenerateCookie(scopelist, servers)
cookie = []
n = len(proxies)
for i in range(n):
scope = [ '/proxy/%dof%d/%s:5873%s' % (i+1, n, server, scope)
for scope in scopelist for server in servers ]
cookie.append(_GenerateCookie(scope, (proxies[i],)))
return ''.join(cookie)
<commit_msg>Remove Python wrapper function for executing cookiecutter program<commit_after> | #
# The OpenDiamond Platform for Interactive Search
# Version 4
#
# Copyright (c) 2009 Carnegie Mellon University
# All rights reserved.
#
# This software is distributed under the terms of the Eclipse Public
# License, Version 1.0 which can be found in the file named LICENSE.
# ANY USE, REPRODUCTION OR DISTRIBUTION OF THIS SOFTWARE CONSTITUTES
# RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT
#
| #
# The OpenDiamond Platform for Interactive Search
# Version 4
#
# Copyright (c) 2009 Carnegie Mellon University
# All rights reserved.
#
# This software is distributed under the terms of the Eclipse Public
# License, Version 1.0 which can be found in the file named LICENSE.
# ANY USE, REPRODUCTION OR DISTRIBUTION OF THIS SOFTWARE CONSTITUTES
# RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT
#
from subprocess import Popen, PIPE
def _GenerateCookie(scopelist, servers):
cmd = ["cookiecutter"]
for server in servers:
cmd.extend(['-s', server])
for url in scopelist:
cmd.extend(['-u', url])
return Popen(cmd, stdout=PIPE).stdout.read()
def GenerateCookie(scopelist, servers, proxies=None):
if not proxies:
return _GenerateCookie(scopelist, servers)
cookie = []
n = len(proxies)
for i in range(n):
scope = [ '/proxy/%dof%d/%s:5873%s' % (i+1, n, server, scope)
for scope in scopelist for server in servers ]
cookie.append(_GenerateCookie(scope, (proxies[i],)))
return ''.join(cookie)
Remove Python wrapper function for executing cookiecutter program#
# The OpenDiamond Platform for Interactive Search
# Version 4
#
# Copyright (c) 2009 Carnegie Mellon University
# All rights reserved.
#
# This software is distributed under the terms of the Eclipse Public
# License, Version 1.0 which can be found in the file named LICENSE.
# ANY USE, REPRODUCTION OR DISTRIBUTION OF THIS SOFTWARE CONSTITUTES
# RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT
#
| <commit_before>#
# The OpenDiamond Platform for Interactive Search
# Version 4
#
# Copyright (c) 2009 Carnegie Mellon University
# All rights reserved.
#
# This software is distributed under the terms of the Eclipse Public
# License, Version 1.0 which can be found in the file named LICENSE.
# ANY USE, REPRODUCTION OR DISTRIBUTION OF THIS SOFTWARE CONSTITUTES
# RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT
#
from subprocess import Popen, PIPE
def _GenerateCookie(scopelist, servers):
cmd = ["cookiecutter"]
for server in servers:
cmd.extend(['-s', server])
for url in scopelist:
cmd.extend(['-u', url])
return Popen(cmd, stdout=PIPE).stdout.read()
def GenerateCookie(scopelist, servers, proxies=None):
if not proxies:
return _GenerateCookie(scopelist, servers)
cookie = []
n = len(proxies)
for i in range(n):
scope = [ '/proxy/%dof%d/%s:5873%s' % (i+1, n, server, scope)
for scope in scopelist for server in servers ]
cookie.append(_GenerateCookie(scope, (proxies[i],)))
return ''.join(cookie)
<commit_msg>Remove Python wrapper function for executing cookiecutter program<commit_after>#
# The OpenDiamond Platform for Interactive Search
# Version 4
#
# Copyright (c) 2009 Carnegie Mellon University
# All rights reserved.
#
# This software is distributed under the terms of the Eclipse Public
# License, Version 1.0 which can be found in the file named LICENSE.
# ANY USE, REPRODUCTION OR DISTRIBUTION OF THIS SOFTWARE CONSTITUTES
# RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT
#
|
41b241de6f2afa94b442007518d481526bfb66ae | linked-list/remove-k-from-list.py | linked-list/remove-k-from-list.py | # Given a singly linked list of integers l and an integer k, remove all elements from list l that have a value equal to k
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
| # Given a singly linked list of integers l and an integer k, remove all elements from list l that have a value equal to k
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
class LinkedList(object):
def __init__(self, head=None):
self.head = head
| Add initialization to linked list class | Add initialization to linked list class
| Python | mit | derekmpham/interview-prep,derekmpham/interview-prep | # Given a singly linked list of integers l and an integer k, remove all elements from list l that have a value equal to k
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
Add initialization to linked list class | # Given a singly linked list of integers l and an integer k, remove all elements from list l that have a value equal to k
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
class LinkedList(object):
def __init__(self, head=None):
self.head = head
| <commit_before># Given a singly linked list of integers l and an integer k, remove all elements from list l that have a value equal to k
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
<commit_msg>Add initialization to linked list class<commit_after> | # Given a singly linked list of integers l and an integer k, remove all elements from list l that have a value equal to k
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
class LinkedList(object):
def __init__(self, head=None):
self.head = head
| # Given a singly linked list of integers l and an integer k, remove all elements from list l that have a value equal to k
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
Add initialization to linked list class# Given a singly linked list of integers l and an integer k, remove all elements from list l that have a value equal to k
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
class LinkedList(object):
def __init__(self, head=None):
self.head = head
| <commit_before># Given a singly linked list of integers l and an integer k, remove all elements from list l that have a value equal to k
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
<commit_msg>Add initialization to linked list class<commit_after># Given a singly linked list of integers l and an integer k, remove all elements from list l that have a value equal to k
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
class LinkedList(object):
def __init__(self, head=None):
self.head = head
|
da50d1b66f662f5e3e1b89fd88632f7076c32084 | apps/careers/models.py | apps/careers/models.py | from cms import sitemaps
from cms.apps.pages.models import ContentBase
from cms.models import HtmlField, PageBase
from django.db import models
from historylinks import shortcuts as historylinks
class Careers(ContentBase):
classifier = 'apps'
urlconf = '{{ project_name }}.apps.careers.urls'
per_page = models.PositiveIntegerField(
'careers per page',
default=10,
blank=True,
null=True
)
def __str__(self):
return self.page.title
class Career(PageBase):
page = models.ForeignKey(
Careers
)
title = models.CharField(
max_length=256,
)
slug = models.CharField(
max_length=256,
unique=True
)
location = models.CharField(
max_length=256,
blank=True,
null=True
)
summary = models.TextField(
blank=True,
null=True
)
description = HtmlField()
email_address = models.EmailField()
order = models.PositiveIntegerField(
default=0
)
class Meta:
ordering = ['order']
def __str__(self):
return self.title
def get_absolute_url(self):
return self.page.page.reverse('career_detail', kwargs={
'slug': self.slug,
})
historylinks.register(Career)
sitemaps.register(Career)
| from cms import sitemaps
from cms.apps.pages.models import ContentBase
from cms.models import HtmlField, PageBase
from django.db import models
from historylinks import shortcuts as historylinks
class Careers(ContentBase):
classifier = 'apps'
urlconf = '{{ project_name }}.apps.careers.urls'
per_page = models.PositiveIntegerField(
'careers per page',
default=10,
blank=True,
null=True
)
def __str__(self):
return self.page.title
class Career(PageBase):
page = models.ForeignKey(
Careers
)
location = models.CharField(
max_length=256,
blank=True,
null=True
)
summary = models.TextField(
blank=True,
null=True
)
description = HtmlField()
email_address = models.EmailField()
order = models.PositiveIntegerField(
default=0
)
class Meta:
ordering = ['order']
def __str__(self):
return self.title
def get_absolute_url(self):
return self.page.page.reverse('career_detail', kwargs={
'slug': self.slug,
})
historylinks.register(Career)
sitemaps.register(Career)
| Remove duplicate fields from Career | Remove duplicate fields from Career
| Python | mit | onespacemedia/cms-jobs,onespacemedia/cms-jobs | from cms import sitemaps
from cms.apps.pages.models import ContentBase
from cms.models import HtmlField, PageBase
from django.db import models
from historylinks import shortcuts as historylinks
class Careers(ContentBase):
classifier = 'apps'
urlconf = '{{ project_name }}.apps.careers.urls'
per_page = models.PositiveIntegerField(
'careers per page',
default=10,
blank=True,
null=True
)
def __str__(self):
return self.page.title
class Career(PageBase):
page = models.ForeignKey(
Careers
)
title = models.CharField(
max_length=256,
)
slug = models.CharField(
max_length=256,
unique=True
)
location = models.CharField(
max_length=256,
blank=True,
null=True
)
summary = models.TextField(
blank=True,
null=True
)
description = HtmlField()
email_address = models.EmailField()
order = models.PositiveIntegerField(
default=0
)
class Meta:
ordering = ['order']
def __str__(self):
return self.title
def get_absolute_url(self):
return self.page.page.reverse('career_detail', kwargs={
'slug': self.slug,
})
historylinks.register(Career)
sitemaps.register(Career)
Remove duplicate fields from Career | from cms import sitemaps
from cms.apps.pages.models import ContentBase
from cms.models import HtmlField, PageBase
from django.db import models
from historylinks import shortcuts as historylinks
class Careers(ContentBase):
classifier = 'apps'
urlconf = '{{ project_name }}.apps.careers.urls'
per_page = models.PositiveIntegerField(
'careers per page',
default=10,
blank=True,
null=True
)
def __str__(self):
return self.page.title
class Career(PageBase):
page = models.ForeignKey(
Careers
)
location = models.CharField(
max_length=256,
blank=True,
null=True
)
summary = models.TextField(
blank=True,
null=True
)
description = HtmlField()
email_address = models.EmailField()
order = models.PositiveIntegerField(
default=0
)
class Meta:
ordering = ['order']
def __str__(self):
return self.title
def get_absolute_url(self):
return self.page.page.reverse('career_detail', kwargs={
'slug': self.slug,
})
historylinks.register(Career)
sitemaps.register(Career)
| <commit_before>from cms import sitemaps
from cms.apps.pages.models import ContentBase
from cms.models import HtmlField, PageBase
from django.db import models
from historylinks import shortcuts as historylinks
class Careers(ContentBase):
classifier = 'apps'
urlconf = '{{ project_name }}.apps.careers.urls'
per_page = models.PositiveIntegerField(
'careers per page',
default=10,
blank=True,
null=True
)
def __str__(self):
return self.page.title
class Career(PageBase):
page = models.ForeignKey(
Careers
)
title = models.CharField(
max_length=256,
)
slug = models.CharField(
max_length=256,
unique=True
)
location = models.CharField(
max_length=256,
blank=True,
null=True
)
summary = models.TextField(
blank=True,
null=True
)
description = HtmlField()
email_address = models.EmailField()
order = models.PositiveIntegerField(
default=0
)
class Meta:
ordering = ['order']
def __str__(self):
return self.title
def get_absolute_url(self):
return self.page.page.reverse('career_detail', kwargs={
'slug': self.slug,
})
historylinks.register(Career)
sitemaps.register(Career)
<commit_msg>Remove duplicate fields from Career<commit_after> | from cms import sitemaps
from cms.apps.pages.models import ContentBase
from cms.models import HtmlField, PageBase
from django.db import models
from historylinks import shortcuts as historylinks
class Careers(ContentBase):
classifier = 'apps'
urlconf = '{{ project_name }}.apps.careers.urls'
per_page = models.PositiveIntegerField(
'careers per page',
default=10,
blank=True,
null=True
)
def __str__(self):
return self.page.title
class Career(PageBase):
page = models.ForeignKey(
Careers
)
location = models.CharField(
max_length=256,
blank=True,
null=True
)
summary = models.TextField(
blank=True,
null=True
)
description = HtmlField()
email_address = models.EmailField()
order = models.PositiveIntegerField(
default=0
)
class Meta:
ordering = ['order']
def __str__(self):
return self.title
def get_absolute_url(self):
return self.page.page.reverse('career_detail', kwargs={
'slug': self.slug,
})
historylinks.register(Career)
sitemaps.register(Career)
| from cms import sitemaps
from cms.apps.pages.models import ContentBase
from cms.models import HtmlField, PageBase
from django.db import models
from historylinks import shortcuts as historylinks
class Careers(ContentBase):
classifier = 'apps'
urlconf = '{{ project_name }}.apps.careers.urls'
per_page = models.PositiveIntegerField(
'careers per page',
default=10,
blank=True,
null=True
)
def __str__(self):
return self.page.title
class Career(PageBase):
page = models.ForeignKey(
Careers
)
title = models.CharField(
max_length=256,
)
slug = models.CharField(
max_length=256,
unique=True
)
location = models.CharField(
max_length=256,
blank=True,
null=True
)
summary = models.TextField(
blank=True,
null=True
)
description = HtmlField()
email_address = models.EmailField()
order = models.PositiveIntegerField(
default=0
)
class Meta:
ordering = ['order']
def __str__(self):
return self.title
def get_absolute_url(self):
return self.page.page.reverse('career_detail', kwargs={
'slug': self.slug,
})
historylinks.register(Career)
sitemaps.register(Career)
Remove duplicate fields from Careerfrom cms import sitemaps
from cms.apps.pages.models import ContentBase
from cms.models import HtmlField, PageBase
from django.db import models
from historylinks import shortcuts as historylinks
class Careers(ContentBase):
classifier = 'apps'
urlconf = '{{ project_name }}.apps.careers.urls'
per_page = models.PositiveIntegerField(
'careers per page',
default=10,
blank=True,
null=True
)
def __str__(self):
return self.page.title
class Career(PageBase):
page = models.ForeignKey(
Careers
)
location = models.CharField(
max_length=256,
blank=True,
null=True
)
summary = models.TextField(
blank=True,
null=True
)
description = HtmlField()
email_address = models.EmailField()
order = models.PositiveIntegerField(
default=0
)
class Meta:
ordering = ['order']
def __str__(self):
return self.title
def get_absolute_url(self):
return self.page.page.reverse('career_detail', kwargs={
'slug': self.slug,
})
historylinks.register(Career)
sitemaps.register(Career)
| <commit_before>from cms import sitemaps
from cms.apps.pages.models import ContentBase
from cms.models import HtmlField, PageBase
from django.db import models
from historylinks import shortcuts as historylinks
class Careers(ContentBase):
classifier = 'apps'
urlconf = '{{ project_name }}.apps.careers.urls'
per_page = models.PositiveIntegerField(
'careers per page',
default=10,
blank=True,
null=True
)
def __str__(self):
return self.page.title
class Career(PageBase):
page = models.ForeignKey(
Careers
)
title = models.CharField(
max_length=256,
)
slug = models.CharField(
max_length=256,
unique=True
)
location = models.CharField(
max_length=256,
blank=True,
null=True
)
summary = models.TextField(
blank=True,
null=True
)
description = HtmlField()
email_address = models.EmailField()
order = models.PositiveIntegerField(
default=0
)
class Meta:
ordering = ['order']
def __str__(self):
return self.title
def get_absolute_url(self):
return self.page.page.reverse('career_detail', kwargs={
'slug': self.slug,
})
historylinks.register(Career)
sitemaps.register(Career)
<commit_msg>Remove duplicate fields from Career<commit_after>from cms import sitemaps
from cms.apps.pages.models import ContentBase
from cms.models import HtmlField, PageBase
from django.db import models
from historylinks import shortcuts as historylinks
class Careers(ContentBase):
classifier = 'apps'
urlconf = '{{ project_name }}.apps.careers.urls'
per_page = models.PositiveIntegerField(
'careers per page',
default=10,
blank=True,
null=True
)
def __str__(self):
return self.page.title
class Career(PageBase):
page = models.ForeignKey(
Careers
)
location = models.CharField(
max_length=256,
blank=True,
null=True
)
summary = models.TextField(
blank=True,
null=True
)
description = HtmlField()
email_address = models.EmailField()
order = models.PositiveIntegerField(
default=0
)
class Meta:
ordering = ['order']
def __str__(self):
return self.title
def get_absolute_url(self):
return self.page.page.reverse('career_detail', kwargs={
'slug': self.slug,
})
historylinks.register(Career)
sitemaps.register(Career)
|
8416a3ed1a6af2d0037f77744d809441591086cd | mrp_bom_location/models/mrp_bom.py | mrp_bom_location/models/mrp_bom.py | # Copyright 2017 Eficent Business and IT Consulting Services S.L.
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from odoo import fields, models
class MrpBom(models.Model):
_inherit = "mrp.bom"
location_id = fields.Many2one(
related='picking_type_id.default_location_dest_id',
store=True,
)
class MrpBomLine(models.Model):
_inherit = "mrp.bom.line"
location_id = fields.Many2one(
related='bom_id.picking_type_id.default_location_src_id',
store=True,
)
| # Copyright 2017 Eficent Business and IT Consulting Services S.L.
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from odoo import fields, models
class MrpBom(models.Model):
_inherit = "mrp.bom"
location_id = fields.Many2one(
related='picking_type_id.default_location_dest_id',
readonly=True,
store=True,
)
class MrpBomLine(models.Model):
_inherit = "mrp.bom.line"
location_id = fields.Many2one(
related='bom_id.picking_type_id.default_location_src_id',
readonly=True,
store=True,
)
| Make the related location readonly | [IMP] Make the related location readonly
| Python | agpl-3.0 | OCA/manufacture,OCA/manufacture | # Copyright 2017 Eficent Business and IT Consulting Services S.L.
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from odoo import fields, models
class MrpBom(models.Model):
_inherit = "mrp.bom"
location_id = fields.Many2one(
related='picking_type_id.default_location_dest_id',
store=True,
)
class MrpBomLine(models.Model):
_inherit = "mrp.bom.line"
location_id = fields.Many2one(
related='bom_id.picking_type_id.default_location_src_id',
store=True,
)
[IMP] Make the related location readonly | # Copyright 2017 Eficent Business and IT Consulting Services S.L.
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from odoo import fields, models
class MrpBom(models.Model):
_inherit = "mrp.bom"
location_id = fields.Many2one(
related='picking_type_id.default_location_dest_id',
readonly=True,
store=True,
)
class MrpBomLine(models.Model):
_inherit = "mrp.bom.line"
location_id = fields.Many2one(
related='bom_id.picking_type_id.default_location_src_id',
readonly=True,
store=True,
)
| <commit_before># Copyright 2017 Eficent Business and IT Consulting Services S.L.
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from odoo import fields, models
class MrpBom(models.Model):
_inherit = "mrp.bom"
location_id = fields.Many2one(
related='picking_type_id.default_location_dest_id',
store=True,
)
class MrpBomLine(models.Model):
_inherit = "mrp.bom.line"
location_id = fields.Many2one(
related='bom_id.picking_type_id.default_location_src_id',
store=True,
)
<commit_msg>[IMP] Make the related location readonly<commit_after> | # Copyright 2017 Eficent Business and IT Consulting Services S.L.
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from odoo import fields, models
class MrpBom(models.Model):
_inherit = "mrp.bom"
location_id = fields.Many2one(
related='picking_type_id.default_location_dest_id',
readonly=True,
store=True,
)
class MrpBomLine(models.Model):
_inherit = "mrp.bom.line"
location_id = fields.Many2one(
related='bom_id.picking_type_id.default_location_src_id',
readonly=True,
store=True,
)
| # Copyright 2017 Eficent Business and IT Consulting Services S.L.
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from odoo import fields, models
class MrpBom(models.Model):
_inherit = "mrp.bom"
location_id = fields.Many2one(
related='picking_type_id.default_location_dest_id',
store=True,
)
class MrpBomLine(models.Model):
_inherit = "mrp.bom.line"
location_id = fields.Many2one(
related='bom_id.picking_type_id.default_location_src_id',
store=True,
)
[IMP] Make the related location readonly# Copyright 2017 Eficent Business and IT Consulting Services S.L.
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from odoo import fields, models
class MrpBom(models.Model):
_inherit = "mrp.bom"
location_id = fields.Many2one(
related='picking_type_id.default_location_dest_id',
readonly=True,
store=True,
)
class MrpBomLine(models.Model):
_inherit = "mrp.bom.line"
location_id = fields.Many2one(
related='bom_id.picking_type_id.default_location_src_id',
readonly=True,
store=True,
)
| <commit_before># Copyright 2017 Eficent Business and IT Consulting Services S.L.
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from odoo import fields, models
class MrpBom(models.Model):
_inherit = "mrp.bom"
location_id = fields.Many2one(
related='picking_type_id.default_location_dest_id',
store=True,
)
class MrpBomLine(models.Model):
_inherit = "mrp.bom.line"
location_id = fields.Many2one(
related='bom_id.picking_type_id.default_location_src_id',
store=True,
)
<commit_msg>[IMP] Make the related location readonly<commit_after># Copyright 2017 Eficent Business and IT Consulting Services S.L.
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from odoo import fields, models
class MrpBom(models.Model):
_inherit = "mrp.bom"
location_id = fields.Many2one(
related='picking_type_id.default_location_dest_id',
readonly=True,
store=True,
)
class MrpBomLine(models.Model):
_inherit = "mrp.bom.line"
location_id = fields.Many2one(
related='bom_id.picking_type_id.default_location_src_id',
readonly=True,
store=True,
)
|
57a7651ba9583830ab32fae0bb8d790bb2bdb6a8 | setup.py | setup.py | #!/usr/bin/env python
from setuptools import setup
setup(name='programmabletuple',
version='0.2.0',
description='Python metaclass for making named tuples with programmability',
long_description=open('README.rst').read(),
author='Tschijnmo TSCHAU',
author_email='tschijnmotschau@gmail.com',
url='https://github.com/tschijnmo/programmabletuple',
license='MIT',
packages=['programmabletuple', ],
classifiers=[
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
| #!/usr/bin/env python
from setuptools import setup
setup(name='programmabletuple',
version='0.3.0',
description='Python metaclass for making named tuples with programmability',
long_description=open('README.rst').read(),
author='Tschijnmo TSCHAU',
author_email='tschijnmotschau@gmail.com',
url='https://github.com/tschijnmo/programmabletuple',
license='MIT',
packages=['programmabletuple', ],
classifiers=[
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
| Change the version number to 3.0 | Change the version number to 3.0
The most notable change from 2.0 is the new initializer.
| Python | mit | tschijnmo/programmabletuple | #!/usr/bin/env python
from setuptools import setup
setup(name='programmabletuple',
version='0.2.0',
description='Python metaclass for making named tuples with programmability',
long_description=open('README.rst').read(),
author='Tschijnmo TSCHAU',
author_email='tschijnmotschau@gmail.com',
url='https://github.com/tschijnmo/programmabletuple',
license='MIT',
packages=['programmabletuple', ],
classifiers=[
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
Change the version number to 3.0
The most notable change from 2.0 is the new initializer. | #!/usr/bin/env python
from setuptools import setup
setup(name='programmabletuple',
version='0.3.0',
description='Python metaclass for making named tuples with programmability',
long_description=open('README.rst').read(),
author='Tschijnmo TSCHAU',
author_email='tschijnmotschau@gmail.com',
url='https://github.com/tschijnmo/programmabletuple',
license='MIT',
packages=['programmabletuple', ],
classifiers=[
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
| <commit_before>#!/usr/bin/env python
from setuptools import setup
setup(name='programmabletuple',
version='0.2.0',
description='Python metaclass for making named tuples with programmability',
long_description=open('README.rst').read(),
author='Tschijnmo TSCHAU',
author_email='tschijnmotschau@gmail.com',
url='https://github.com/tschijnmo/programmabletuple',
license='MIT',
packages=['programmabletuple', ],
classifiers=[
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
<commit_msg>Change the version number to 3.0
The most notable change from 2.0 is the new initializer.<commit_after> | #!/usr/bin/env python
from setuptools import setup
setup(name='programmabletuple',
version='0.3.0',
description='Python metaclass for making named tuples with programmability',
long_description=open('README.rst').read(),
author='Tschijnmo TSCHAU',
author_email='tschijnmotschau@gmail.com',
url='https://github.com/tschijnmo/programmabletuple',
license='MIT',
packages=['programmabletuple', ],
classifiers=[
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
| #!/usr/bin/env python
from setuptools import setup
setup(name='programmabletuple',
version='0.2.0',
description='Python metaclass for making named tuples with programmability',
long_description=open('README.rst').read(),
author='Tschijnmo TSCHAU',
author_email='tschijnmotschau@gmail.com',
url='https://github.com/tschijnmo/programmabletuple',
license='MIT',
packages=['programmabletuple', ],
classifiers=[
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
Change the version number to 3.0
The most notable change from 2.0 is the new initializer.#!/usr/bin/env python
from setuptools import setup
setup(name='programmabletuple',
version='0.3.0',
description='Python metaclass for making named tuples with programmability',
long_description=open('README.rst').read(),
author='Tschijnmo TSCHAU',
author_email='tschijnmotschau@gmail.com',
url='https://github.com/tschijnmo/programmabletuple',
license='MIT',
packages=['programmabletuple', ],
classifiers=[
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
| <commit_before>#!/usr/bin/env python
from setuptools import setup
setup(name='programmabletuple',
version='0.2.0',
description='Python metaclass for making named tuples with programmability',
long_description=open('README.rst').read(),
author='Tschijnmo TSCHAU',
author_email='tschijnmotschau@gmail.com',
url='https://github.com/tschijnmo/programmabletuple',
license='MIT',
packages=['programmabletuple', ],
classifiers=[
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
<commit_msg>Change the version number to 3.0
The most notable change from 2.0 is the new initializer.<commit_after>#!/usr/bin/env python
from setuptools import setup
setup(name='programmabletuple',
version='0.3.0',
description='Python metaclass for making named tuples with programmability',
long_description=open('README.rst').read(),
author='Tschijnmo TSCHAU',
author_email='tschijnmotschau@gmail.com',
url='https://github.com/tschijnmo/programmabletuple',
license='MIT',
packages=['programmabletuple', ],
classifiers=[
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
|
21afbaab7deb874703f4968ea1337b59120f0ad0 | music-stream.py | music-stream.py | import urllib.request
import subprocess
LIMIT = 10
PLAYER = 'vlc'
url = 'http://streams.twitch.tv/kraken/streams?limit='+str(LIMIT)+'&offset=0&game=Music&broadcaster_language=&on_site=1'
with urllib.request.urlopen(url) as response:
html = response.read().decode('utf8')
i = 0
urls = []
for line in html.split(','):
if 'status' in line:
status = line.split('"')[-2]
status = ''.join(i for i in status if ord(i)<128) #filter non ascii characters
if 'display_name' in line:
name = line.split('"')[-2]
print(str(i) + ') ' + name + ' : ' + status)
i += 1
if 'url' in line:
url = line.split('"')[-2]
urls.append(url)
choice = LIMIT
while (choice >= LIMIT):
choice = int(input('Choose a stream\n'))
cmd = ['livestreamer', urls[choice], 'audio']
if PLAYER != 'vlc':
cmd.append('-p')
cmd.append(PLAYER)
subprocess.Popen(cmd, shell=False)
| import urllib.request
import subprocess
LIMIT = 10
PLAYER = 'vlc'
STREAMS_URL = 'http://streams.twitch.tv/kraken/streams?limit='+str(LIMIT)+'&offset=0&game=Music&broadcaster_language=&on_site=1'
while True:
with urllib.request.urlopen(STREAMS_URL) as response:
html = response.read().decode('utf8')
i = 0
urls = []
for line in html.split(','):
if 'status' in line:
status = line.split('"')[-2]
status = ''.join(i for i in status if ord(i)<128) #filter non ascii characters
if 'display_name' in line:
name = line.split('"')[-2]
print(str(i) + ') ' + name + ' : ' + status)
i += 1
if 'url' in line:
url = line.split('"')[-2]
urls.append(url)
choice = LIMIT
while (choice >= LIMIT):
choice = int(input('Choose a stream\n'))
cmd = ['livestreamer', urls[choice], 'audio']
if PLAYER != 'vlc':
cmd.append('-p')
cmd.append(PLAYER)
subprocess.call(cmd, shell=False)
print('\n\n\n')
| Refresh streams list when player is closed | Refresh streams list when player is closed | Python | mit | GaudyZircon/music-stream | import urllib.request
import subprocess
LIMIT = 10
PLAYER = 'vlc'
url = 'http://streams.twitch.tv/kraken/streams?limit='+str(LIMIT)+'&offset=0&game=Music&broadcaster_language=&on_site=1'
with urllib.request.urlopen(url) as response:
html = response.read().decode('utf8')
i = 0
urls = []
for line in html.split(','):
if 'status' in line:
status = line.split('"')[-2]
status = ''.join(i for i in status if ord(i)<128) #filter non ascii characters
if 'display_name' in line:
name = line.split('"')[-2]
print(str(i) + ') ' + name + ' : ' + status)
i += 1
if 'url' in line:
url = line.split('"')[-2]
urls.append(url)
choice = LIMIT
while (choice >= LIMIT):
choice = int(input('Choose a stream\n'))
cmd = ['livestreamer', urls[choice], 'audio']
if PLAYER != 'vlc':
cmd.append('-p')
cmd.append(PLAYER)
subprocess.Popen(cmd, shell=False)
Refresh streams list when player is closed | import urllib.request
import subprocess
LIMIT = 10
PLAYER = 'vlc'
STREAMS_URL = 'http://streams.twitch.tv/kraken/streams?limit='+str(LIMIT)+'&offset=0&game=Music&broadcaster_language=&on_site=1'
while True:
with urllib.request.urlopen(STREAMS_URL) as response:
html = response.read().decode('utf8')
i = 0
urls = []
for line in html.split(','):
if 'status' in line:
status = line.split('"')[-2]
status = ''.join(i for i in status if ord(i)<128) #filter non ascii characters
if 'display_name' in line:
name = line.split('"')[-2]
print(str(i) + ') ' + name + ' : ' + status)
i += 1
if 'url' in line:
url = line.split('"')[-2]
urls.append(url)
choice = LIMIT
while (choice >= LIMIT):
choice = int(input('Choose a stream\n'))
cmd = ['livestreamer', urls[choice], 'audio']
if PLAYER != 'vlc':
cmd.append('-p')
cmd.append(PLAYER)
subprocess.call(cmd, shell=False)
print('\n\n\n')
| <commit_before>import urllib.request
import subprocess
LIMIT = 10
PLAYER = 'vlc'
url = 'http://streams.twitch.tv/kraken/streams?limit='+str(LIMIT)+'&offset=0&game=Music&broadcaster_language=&on_site=1'
with urllib.request.urlopen(url) as response:
html = response.read().decode('utf8')
i = 0
urls = []
for line in html.split(','):
if 'status' in line:
status = line.split('"')[-2]
status = ''.join(i for i in status if ord(i)<128) #filter non ascii characters
if 'display_name' in line:
name = line.split('"')[-2]
print(str(i) + ') ' + name + ' : ' + status)
i += 1
if 'url' in line:
url = line.split('"')[-2]
urls.append(url)
choice = LIMIT
while (choice >= LIMIT):
choice = int(input('Choose a stream\n'))
cmd = ['livestreamer', urls[choice], 'audio']
if PLAYER != 'vlc':
cmd.append('-p')
cmd.append(PLAYER)
subprocess.Popen(cmd, shell=False)
<commit_msg>Refresh streams list when player is closed<commit_after> | import urllib.request
import subprocess
LIMIT = 10
PLAYER = 'vlc'
STREAMS_URL = 'http://streams.twitch.tv/kraken/streams?limit='+str(LIMIT)+'&offset=0&game=Music&broadcaster_language=&on_site=1'
while True:
with urllib.request.urlopen(STREAMS_URL) as response:
html = response.read().decode('utf8')
i = 0
urls = []
for line in html.split(','):
if 'status' in line:
status = line.split('"')[-2]
status = ''.join(i for i in status if ord(i)<128) #filter non ascii characters
if 'display_name' in line:
name = line.split('"')[-2]
print(str(i) + ') ' + name + ' : ' + status)
i += 1
if 'url' in line:
url = line.split('"')[-2]
urls.append(url)
choice = LIMIT
while (choice >= LIMIT):
choice = int(input('Choose a stream\n'))
cmd = ['livestreamer', urls[choice], 'audio']
if PLAYER != 'vlc':
cmd.append('-p')
cmd.append(PLAYER)
subprocess.call(cmd, shell=False)
print('\n\n\n')
| import urllib.request
import subprocess
LIMIT = 10
PLAYER = 'vlc'
url = 'http://streams.twitch.tv/kraken/streams?limit='+str(LIMIT)+'&offset=0&game=Music&broadcaster_language=&on_site=1'
with urllib.request.urlopen(url) as response:
html = response.read().decode('utf8')
i = 0
urls = []
for line in html.split(','):
if 'status' in line:
status = line.split('"')[-2]
status = ''.join(i for i in status if ord(i)<128) #filter non ascii characters
if 'display_name' in line:
name = line.split('"')[-2]
print(str(i) + ') ' + name + ' : ' + status)
i += 1
if 'url' in line:
url = line.split('"')[-2]
urls.append(url)
choice = LIMIT
while (choice >= LIMIT):
choice = int(input('Choose a stream\n'))
cmd = ['livestreamer', urls[choice], 'audio']
if PLAYER != 'vlc':
cmd.append('-p')
cmd.append(PLAYER)
subprocess.Popen(cmd, shell=False)
Refresh streams list when player is closedimport urllib.request
import subprocess
LIMIT = 10
PLAYER = 'vlc'
STREAMS_URL = 'http://streams.twitch.tv/kraken/streams?limit='+str(LIMIT)+'&offset=0&game=Music&broadcaster_language=&on_site=1'
while True:
with urllib.request.urlopen(STREAMS_URL) as response:
html = response.read().decode('utf8')
i = 0
urls = []
for line in html.split(','):
if 'status' in line:
status = line.split('"')[-2]
status = ''.join(i for i in status if ord(i)<128) #filter non ascii characters
if 'display_name' in line:
name = line.split('"')[-2]
print(str(i) + ') ' + name + ' : ' + status)
i += 1
if 'url' in line:
url = line.split('"')[-2]
urls.append(url)
choice = LIMIT
while (choice >= LIMIT):
choice = int(input('Choose a stream\n'))
cmd = ['livestreamer', urls[choice], 'audio']
if PLAYER != 'vlc':
cmd.append('-p')
cmd.append(PLAYER)
subprocess.call(cmd, shell=False)
print('\n\n\n')
| <commit_before>import urllib.request
import subprocess
LIMIT = 10
PLAYER = 'vlc'
url = 'http://streams.twitch.tv/kraken/streams?limit='+str(LIMIT)+'&offset=0&game=Music&broadcaster_language=&on_site=1'
with urllib.request.urlopen(url) as response:
html = response.read().decode('utf8')
i = 0
urls = []
for line in html.split(','):
if 'status' in line:
status = line.split('"')[-2]
status = ''.join(i for i in status if ord(i)<128) #filter non ascii characters
if 'display_name' in line:
name = line.split('"')[-2]
print(str(i) + ') ' + name + ' : ' + status)
i += 1
if 'url' in line:
url = line.split('"')[-2]
urls.append(url)
choice = LIMIT
while (choice >= LIMIT):
choice = int(input('Choose a stream\n'))
cmd = ['livestreamer', urls[choice], 'audio']
if PLAYER != 'vlc':
cmd.append('-p')
cmd.append(PLAYER)
subprocess.Popen(cmd, shell=False)
<commit_msg>Refresh streams list when player is closed<commit_after>import urllib.request
import subprocess
LIMIT = 10
PLAYER = 'vlc'
STREAMS_URL = 'http://streams.twitch.tv/kraken/streams?limit='+str(LIMIT)+'&offset=0&game=Music&broadcaster_language=&on_site=1'
while True:
with urllib.request.urlopen(STREAMS_URL) as response:
html = response.read().decode('utf8')
i = 0
urls = []
for line in html.split(','):
if 'status' in line:
status = line.split('"')[-2]
status = ''.join(i for i in status if ord(i)<128) #filter non ascii characters
if 'display_name' in line:
name = line.split('"')[-2]
print(str(i) + ') ' + name + ' : ' + status)
i += 1
if 'url' in line:
url = line.split('"')[-2]
urls.append(url)
choice = LIMIT
while (choice >= LIMIT):
choice = int(input('Choose a stream\n'))
cmd = ['livestreamer', urls[choice], 'audio']
if PLAYER != 'vlc':
cmd.append('-p')
cmd.append(PLAYER)
subprocess.call(cmd, shell=False)
print('\n\n\n')
|
9d9827721e3d4c45f8917662d2f59759fb4ecd66 | muv/__init__.py | muv/__init__.py | """
Miscellaneous utilities.
"""
import numpy as np
def kennard_stone(d, k):
"""
Use the Kennard-Stone algorithm to select k maximally separated
examples from a dataset.
See R. W. Kennard and L. A. Stone (1969): Computer Aided Design of
Experiments, Technometrics, 11:1, 137-148.
Algorithm
---------
1. Choose the two examples separated by the largest distance. In the
case of a tie, use the first examples returned by np.where.
2. For the remaining k - 2 selections, choose the example with the
greatest distance to the closest example among all previously
chosen points.
Parameters
----------
d : ndarray
Pairwise distance matrix between dataset examples.
k : int
Number of examples to select.
"""
assert 1 < k < d.shape[0]
chosen = []
# choose initial points
first = np.where(d == np.amax(d))
chosen.append(first[0][0])
chosen.append(first[1][0])
d = np.ma.array(d, mask=np.ones_like(d, dtype=bool))
# choose remaining points
while len(chosen) < k:
d.mask[:, chosen] = False
d.mask[chosen] = True
print d
p = np.ma.argmax(np.ma.amin(d, axis=1))
chosen.append(p)
return chosen
| """
Miscellaneous utilities.
"""
import numpy as np
class MUV(object):
"""
Generate maximum unbiased validation (MUV) datasets for virtual
screening as described in Rohrer and Baumann, J. Chem. Inf. Model.
2009, 49, 169-184.
"""
def kennard_stone(d, k):
"""
Use the Kennard-Stone algorithm to select k maximally separated
examples from a dataset.
See Kennard and Stone, Technometrics 1969, 11, 137-148.
Algorithm
---------
1. Choose the two examples separated by the largest distance. In the
case of a tie, use the first examples returned by np.where.
2. For the remaining k - 2 selections, choose the example with the
greatest distance to the closest example among all previously
chosen points.
Parameters
----------
d : ndarray
Pairwise distance matrix between dataset examples.
k : int
Number of examples to select.
"""
assert 1 < k < d.shape[0]
chosen = []
# choose initial points
first = np.where(d == np.amax(d))
chosen.append(first[0][0])
chosen.append(first[1][0])
d = np.ma.array(d, mask=np.ones_like(d, dtype=bool))
# choose remaining points
while len(chosen) < k:
d.mask[:, chosen] = False
d.mask[chosen] = True
print d
p = np.ma.argmax(np.ma.amin(d, axis=1))
chosen.append(p)
return chosen
| Fix reference and add MUV class | Fix reference and add MUV class
| Python | bsd-3-clause | skearnes/muv | """
Miscellaneous utilities.
"""
import numpy as np
def kennard_stone(d, k):
"""
Use the Kennard-Stone algorithm to select k maximally separated
examples from a dataset.
See R. W. Kennard and L. A. Stone (1969): Computer Aided Design of
Experiments, Technometrics, 11:1, 137-148.
Algorithm
---------
1. Choose the two examples separated by the largest distance. In the
case of a tie, use the first examples returned by np.where.
2. For the remaining k - 2 selections, choose the example with the
greatest distance to the closest example among all previously
chosen points.
Parameters
----------
d : ndarray
Pairwise distance matrix between dataset examples.
k : int
Number of examples to select.
"""
assert 1 < k < d.shape[0]
chosen = []
# choose initial points
first = np.where(d == np.amax(d))
chosen.append(first[0][0])
chosen.append(first[1][0])
d = np.ma.array(d, mask=np.ones_like(d, dtype=bool))
# choose remaining points
while len(chosen) < k:
d.mask[:, chosen] = False
d.mask[chosen] = True
print d
p = np.ma.argmax(np.ma.amin(d, axis=1))
chosen.append(p)
return chosen
Fix reference and add MUV class | """
Miscellaneous utilities.
"""
import numpy as np
class MUV(object):
"""
Generate maximum unbiased validation (MUV) datasets for virtual
screening as described in Rohrer and Baumann, J. Chem. Inf. Model.
2009, 49, 169-184.
"""
def kennard_stone(d, k):
"""
Use the Kennard-Stone algorithm to select k maximally separated
examples from a dataset.
See Kennard and Stone, Technometrics 1969, 11, 137-148.
Algorithm
---------
1. Choose the two examples separated by the largest distance. In the
case of a tie, use the first examples returned by np.where.
2. For the remaining k - 2 selections, choose the example with the
greatest distance to the closest example among all previously
chosen points.
Parameters
----------
d : ndarray
Pairwise distance matrix between dataset examples.
k : int
Number of examples to select.
"""
assert 1 < k < d.shape[0]
chosen = []
# choose initial points
first = np.where(d == np.amax(d))
chosen.append(first[0][0])
chosen.append(first[1][0])
d = np.ma.array(d, mask=np.ones_like(d, dtype=bool))
# choose remaining points
while len(chosen) < k:
d.mask[:, chosen] = False
d.mask[chosen] = True
print d
p = np.ma.argmax(np.ma.amin(d, axis=1))
chosen.append(p)
return chosen
| <commit_before>"""
Miscellaneous utilities.
"""
import numpy as np
def kennard_stone(d, k):
"""
Use the Kennard-Stone algorithm to select k maximally separated
examples from a dataset.
See R. W. Kennard and L. A. Stone (1969): Computer Aided Design of
Experiments, Technometrics, 11:1, 137-148.
Algorithm
---------
1. Choose the two examples separated by the largest distance. In the
case of a tie, use the first examples returned by np.where.
2. For the remaining k - 2 selections, choose the example with the
greatest distance to the closest example among all previously
chosen points.
Parameters
----------
d : ndarray
Pairwise distance matrix between dataset examples.
k : int
Number of examples to select.
"""
assert 1 < k < d.shape[0]
chosen = []
# choose initial points
first = np.where(d == np.amax(d))
chosen.append(first[0][0])
chosen.append(first[1][0])
d = np.ma.array(d, mask=np.ones_like(d, dtype=bool))
# choose remaining points
while len(chosen) < k:
d.mask[:, chosen] = False
d.mask[chosen] = True
print d
p = np.ma.argmax(np.ma.amin(d, axis=1))
chosen.append(p)
return chosen
<commit_msg>Fix reference and add MUV class<commit_after> | """
Miscellaneous utilities.
"""
import numpy as np
class MUV(object):
"""
Generate maximum unbiased validation (MUV) datasets for virtual
screening as described in Rohrer and Baumann, J. Chem. Inf. Model.
2009, 49, 169-184.
"""
def kennard_stone(d, k):
"""
Use the Kennard-Stone algorithm to select k maximally separated
examples from a dataset.
See Kennard and Stone, Technometrics 1969, 11, 137-148.
Algorithm
---------
1. Choose the two examples separated by the largest distance. In the
case of a tie, use the first examples returned by np.where.
2. For the remaining k - 2 selections, choose the example with the
greatest distance to the closest example among all previously
chosen points.
Parameters
----------
d : ndarray
Pairwise distance matrix between dataset examples.
k : int
Number of examples to select.
"""
assert 1 < k < d.shape[0]
chosen = []
# choose initial points
first = np.where(d == np.amax(d))
chosen.append(first[0][0])
chosen.append(first[1][0])
d = np.ma.array(d, mask=np.ones_like(d, dtype=bool))
# choose remaining points
while len(chosen) < k:
d.mask[:, chosen] = False
d.mask[chosen] = True
print d
p = np.ma.argmax(np.ma.amin(d, axis=1))
chosen.append(p)
return chosen
| """
Miscellaneous utilities.
"""
import numpy as np
def kennard_stone(d, k):
"""
Use the Kennard-Stone algorithm to select k maximally separated
examples from a dataset.
See R. W. Kennard and L. A. Stone (1969): Computer Aided Design of
Experiments, Technometrics, 11:1, 137-148.
Algorithm
---------
1. Choose the two examples separated by the largest distance. In the
case of a tie, use the first examples returned by np.where.
2. For the remaining k - 2 selections, choose the example with the
greatest distance to the closest example among all previously
chosen points.
Parameters
----------
d : ndarray
Pairwise distance matrix between dataset examples.
k : int
Number of examples to select.
"""
assert 1 < k < d.shape[0]
chosen = []
# choose initial points
first = np.where(d == np.amax(d))
chosen.append(first[0][0])
chosen.append(first[1][0])
d = np.ma.array(d, mask=np.ones_like(d, dtype=bool))
# choose remaining points
while len(chosen) < k:
d.mask[:, chosen] = False
d.mask[chosen] = True
print d
p = np.ma.argmax(np.ma.amin(d, axis=1))
chosen.append(p)
return chosen
Fix reference and add MUV class"""
Miscellaneous utilities.
"""
import numpy as np
class MUV(object):
"""
Generate maximum unbiased validation (MUV) datasets for virtual
screening as described in Rohrer and Baumann, J. Chem. Inf. Model.
2009, 49, 169-184.
"""
def kennard_stone(d, k):
"""
Use the Kennard-Stone algorithm to select k maximally separated
examples from a dataset.
See Kennard and Stone, Technometrics 1969, 11, 137-148.
Algorithm
---------
1. Choose the two examples separated by the largest distance. In the
case of a tie, use the first examples returned by np.where.
2. For the remaining k - 2 selections, choose the example with the
greatest distance to the closest example among all previously
chosen points.
Parameters
----------
d : ndarray
Pairwise distance matrix between dataset examples.
k : int
Number of examples to select.
"""
assert 1 < k < d.shape[0]
chosen = []
# choose initial points
first = np.where(d == np.amax(d))
chosen.append(first[0][0])
chosen.append(first[1][0])
d = np.ma.array(d, mask=np.ones_like(d, dtype=bool))
# choose remaining points
while len(chosen) < k:
d.mask[:, chosen] = False
d.mask[chosen] = True
print d
p = np.ma.argmax(np.ma.amin(d, axis=1))
chosen.append(p)
return chosen
| <commit_before>"""
Miscellaneous utilities.
"""
import numpy as np
def kennard_stone(d, k):
"""
Use the Kennard-Stone algorithm to select k maximally separated
examples from a dataset.
See R. W. Kennard and L. A. Stone (1969): Computer Aided Design of
Experiments, Technometrics, 11:1, 137-148.
Algorithm
---------
1. Choose the two examples separated by the largest distance. In the
case of a tie, use the first examples returned by np.where.
2. For the remaining k - 2 selections, choose the example with the
greatest distance to the closest example among all previously
chosen points.
Parameters
----------
d : ndarray
Pairwise distance matrix between dataset examples.
k : int
Number of examples to select.
"""
assert 1 < k < d.shape[0]
chosen = []
# choose initial points
first = np.where(d == np.amax(d))
chosen.append(first[0][0])
chosen.append(first[1][0])
d = np.ma.array(d, mask=np.ones_like(d, dtype=bool))
# choose remaining points
while len(chosen) < k:
d.mask[:, chosen] = False
d.mask[chosen] = True
print d
p = np.ma.argmax(np.ma.amin(d, axis=1))
chosen.append(p)
return chosen
<commit_msg>Fix reference and add MUV class<commit_after>"""
Miscellaneous utilities.
"""
import numpy as np
class MUV(object):
"""
Generate maximum unbiased validation (MUV) datasets for virtual
screening as described in Rohrer and Baumann, J. Chem. Inf. Model.
2009, 49, 169-184.
"""
def kennard_stone(d, k):
"""
Use the Kennard-Stone algorithm to select k maximally separated
examples from a dataset.
See Kennard and Stone, Technometrics 1969, 11, 137-148.
Algorithm
---------
1. Choose the two examples separated by the largest distance. In the
case of a tie, use the first examples returned by np.where.
2. For the remaining k - 2 selections, choose the example with the
greatest distance to the closest example among all previously
chosen points.
Parameters
----------
d : ndarray
Pairwise distance matrix between dataset examples.
k : int
Number of examples to select.
"""
assert 1 < k < d.shape[0]
chosen = []
# choose initial points
first = np.where(d == np.amax(d))
chosen.append(first[0][0])
chosen.append(first[1][0])
d = np.ma.array(d, mask=np.ones_like(d, dtype=bool))
# choose remaining points
while len(chosen) < k:
d.mask[:, chosen] = False
d.mask[chosen] = True
print d
p = np.ma.argmax(np.ma.amin(d, axis=1))
chosen.append(p)
return chosen
|
50a1f9cbd5e9ab3a279e195bba06708aae58d3c2 | setup.py | setup.py | from setuptools import setup, find_packages
setup(name='notation',
description='Better human-readable numbers.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
url='https://github.com/debrouwere/python-notation/',
download_url='https://www.github.com/debrouwere/python-notation/tarball/master',
version='1.0.0',
license='ISC',
packages=find_packages(),
keywords='human numbers format notation scientific engineering',
install_requires=[],
test_suite='notation.tests',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: ISC License',
'Operating System :: OS Independent',
'Topic :: Scientific/Engineering :: Information Analysis',
'Topic :: Utilities',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
],
)
| from setuptools import setup, find_packages
setup(name='notation',
description='Better human-readable numbers.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
url='https://github.com/debrouwere/python-notation/',
download_url='https://www.github.com/debrouwere/python-notation/tarball/master',
version='1.0.0',
license='ISC',
packages=find_packages(),
keywords='human numbers format notation scientific engineering',
install_requires=[],
test_suite='notation.tests',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: ISC License (ISCL)',
'Operating System :: OS Independent',
'Topic :: Scientific/Engineering :: Information Analysis',
'Topic :: Utilities',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
],
)
| Fix reference to ISC license. | Fix reference to ISC license.
| Python | isc | debrouwere/python-ballpark | from setuptools import setup, find_packages
setup(name='notation',
description='Better human-readable numbers.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
url='https://github.com/debrouwere/python-notation/',
download_url='https://www.github.com/debrouwere/python-notation/tarball/master',
version='1.0.0',
license='ISC',
packages=find_packages(),
keywords='human numbers format notation scientific engineering',
install_requires=[],
test_suite='notation.tests',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: ISC License',
'Operating System :: OS Independent',
'Topic :: Scientific/Engineering :: Information Analysis',
'Topic :: Utilities',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
],
)
Fix reference to ISC license. | from setuptools import setup, find_packages
setup(name='notation',
description='Better human-readable numbers.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
url='https://github.com/debrouwere/python-notation/',
download_url='https://www.github.com/debrouwere/python-notation/tarball/master',
version='1.0.0',
license='ISC',
packages=find_packages(),
keywords='human numbers format notation scientific engineering',
install_requires=[],
test_suite='notation.tests',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: ISC License (ISCL)',
'Operating System :: OS Independent',
'Topic :: Scientific/Engineering :: Information Analysis',
'Topic :: Utilities',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
],
)
| <commit_before>from setuptools import setup, find_packages
setup(name='notation',
description='Better human-readable numbers.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
url='https://github.com/debrouwere/python-notation/',
download_url='https://www.github.com/debrouwere/python-notation/tarball/master',
version='1.0.0',
license='ISC',
packages=find_packages(),
keywords='human numbers format notation scientific engineering',
install_requires=[],
test_suite='notation.tests',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: ISC License',
'Operating System :: OS Independent',
'Topic :: Scientific/Engineering :: Information Analysis',
'Topic :: Utilities',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
],
)
<commit_msg>Fix reference to ISC license.<commit_after> | from setuptools import setup, find_packages
setup(name='notation',
description='Better human-readable numbers.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
url='https://github.com/debrouwere/python-notation/',
download_url='https://www.github.com/debrouwere/python-notation/tarball/master',
version='1.0.0',
license='ISC',
packages=find_packages(),
keywords='human numbers format notation scientific engineering',
install_requires=[],
test_suite='notation.tests',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: ISC License (ISCL)',
'Operating System :: OS Independent',
'Topic :: Scientific/Engineering :: Information Analysis',
'Topic :: Utilities',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
],
)
| from setuptools import setup, find_packages
setup(name='notation',
description='Better human-readable numbers.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
url='https://github.com/debrouwere/python-notation/',
download_url='https://www.github.com/debrouwere/python-notation/tarball/master',
version='1.0.0',
license='ISC',
packages=find_packages(),
keywords='human numbers format notation scientific engineering',
install_requires=[],
test_suite='notation.tests',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: ISC License',
'Operating System :: OS Independent',
'Topic :: Scientific/Engineering :: Information Analysis',
'Topic :: Utilities',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
],
)
Fix reference to ISC license.from setuptools import setup, find_packages
setup(name='notation',
description='Better human-readable numbers.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
url='https://github.com/debrouwere/python-notation/',
download_url='https://www.github.com/debrouwere/python-notation/tarball/master',
version='1.0.0',
license='ISC',
packages=find_packages(),
keywords='human numbers format notation scientific engineering',
install_requires=[],
test_suite='notation.tests',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: ISC License (ISCL)',
'Operating System :: OS Independent',
'Topic :: Scientific/Engineering :: Information Analysis',
'Topic :: Utilities',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
],
)
| <commit_before>from setuptools import setup, find_packages
setup(name='notation',
description='Better human-readable numbers.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
url='https://github.com/debrouwere/python-notation/',
download_url='https://www.github.com/debrouwere/python-notation/tarball/master',
version='1.0.0',
license='ISC',
packages=find_packages(),
keywords='human numbers format notation scientific engineering',
install_requires=[],
test_suite='notation.tests',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: ISC License',
'Operating System :: OS Independent',
'Topic :: Scientific/Engineering :: Information Analysis',
'Topic :: Utilities',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
],
)
<commit_msg>Fix reference to ISC license.<commit_after>from setuptools import setup, find_packages
setup(name='notation',
description='Better human-readable numbers.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
url='https://github.com/debrouwere/python-notation/',
download_url='https://www.github.com/debrouwere/python-notation/tarball/master',
version='1.0.0',
license='ISC',
packages=find_packages(),
keywords='human numbers format notation scientific engineering',
install_requires=[],
test_suite='notation.tests',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: ISC License (ISCL)',
'Operating System :: OS Independent',
'Topic :: Scientific/Engineering :: Information Analysis',
'Topic :: Utilities',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
],
)
|
06e2d872bda93ed3b520e6ab7ee646aa2f7a0fcd | setup.py | setup.py | #####################################################
# Copyright (c) Xuanyi Dong [GitHub D-X-Y], 2019.08 #
#####################################################
# [2020.03.09] Upgrade to v1.2
import os
from setuptools import setup
def read(fname='README.md'):
with open(os.path.join(os.path.dirname(__file__), fname), encoding='utf-8') as cfile:
return cfile.read()
setup(
name = "nas_bench_201",
version = "1.2",
author = "Xuanyi Dong",
author_email = "dongxuanyi888@gmail.com",
description = "API for NAS-Bench-201 (a benchmark for neural architecture search).",
license = "MIT",
keywords = "NAS Dataset API DeepLearning",
url = "https://github.com/D-X-Y/NAS-Bench-201",
packages=['nas_201_api'],
long_description=read('README.md'),
long_description_content_type='text/markdown',
classifiers=[
"Programming Language :: Python",
"Topic :: Database",
"Topic :: Scientific/Engineering :: Artificial Intelligence",
"License :: OSI Approved :: MIT License",
],
)
| #####################################################
# Copyright (c) Xuanyi Dong [GitHub D-X-Y], 2019.08 #
#####################################################
# [2020.02.25] Initialize the API as v1.1
# [2020.03.09] Upgrade the API to v1.2
# [2020.03.16] Upgrade the API to v1.3
import os
from setuptools import setup
def read(fname='README.md'):
with open(os.path.join(os.path.dirname(__file__), fname), encoding='utf-8') as cfile:
return cfile.read()
setup(
name = "nas_bench_201",
version = "1.3",
author = "Xuanyi Dong",
author_email = "dongxuanyi888@gmail.com",
description = "API for NAS-Bench-201 (a benchmark for neural architecture search).",
license = "MIT",
keywords = "NAS Dataset API DeepLearning",
url = "https://github.com/D-X-Y/NAS-Bench-201",
packages=['nas_201_api'],
long_description=read('README.md'),
long_description_content_type='text/markdown',
classifiers=[
"Programming Language :: Python",
"Topic :: Database",
"Topic :: Scientific/Engineering :: Artificial Intelligence",
"License :: OSI Approved :: MIT License",
],
)
| Upgrade API to v1.3 and benchmark file to v1.1 | Upgrade API to v1.3 and benchmark file to v1.1
| Python | mit | D-X-Y/ResNeXt,D-X-Y/ResNeXt | #####################################################
# Copyright (c) Xuanyi Dong [GitHub D-X-Y], 2019.08 #
#####################################################
# [2020.03.09] Upgrade to v1.2
import os
from setuptools import setup
def read(fname='README.md'):
with open(os.path.join(os.path.dirname(__file__), fname), encoding='utf-8') as cfile:
return cfile.read()
setup(
name = "nas_bench_201",
version = "1.2",
author = "Xuanyi Dong",
author_email = "dongxuanyi888@gmail.com",
description = "API for NAS-Bench-201 (a benchmark for neural architecture search).",
license = "MIT",
keywords = "NAS Dataset API DeepLearning",
url = "https://github.com/D-X-Y/NAS-Bench-201",
packages=['nas_201_api'],
long_description=read('README.md'),
long_description_content_type='text/markdown',
classifiers=[
"Programming Language :: Python",
"Topic :: Database",
"Topic :: Scientific/Engineering :: Artificial Intelligence",
"License :: OSI Approved :: MIT License",
],
)
Upgrade API to v1.3 and benchmark file to v1.1 | #####################################################
# Copyright (c) Xuanyi Dong [GitHub D-X-Y], 2019.08 #
#####################################################
# [2020.02.25] Initialize the API as v1.1
# [2020.03.09] Upgrade the API to v1.2
# [2020.03.16] Upgrade the API to v1.3
import os
from setuptools import setup
def read(fname='README.md'):
with open(os.path.join(os.path.dirname(__file__), fname), encoding='utf-8') as cfile:
return cfile.read()
setup(
name = "nas_bench_201",
version = "1.3",
author = "Xuanyi Dong",
author_email = "dongxuanyi888@gmail.com",
description = "API for NAS-Bench-201 (a benchmark for neural architecture search).",
license = "MIT",
keywords = "NAS Dataset API DeepLearning",
url = "https://github.com/D-X-Y/NAS-Bench-201",
packages=['nas_201_api'],
long_description=read('README.md'),
long_description_content_type='text/markdown',
classifiers=[
"Programming Language :: Python",
"Topic :: Database",
"Topic :: Scientific/Engineering :: Artificial Intelligence",
"License :: OSI Approved :: MIT License",
],
)
| <commit_before>#####################################################
# Copyright (c) Xuanyi Dong [GitHub D-X-Y], 2019.08 #
#####################################################
# [2020.03.09] Upgrade to v1.2
import os
from setuptools import setup
def read(fname='README.md'):
with open(os.path.join(os.path.dirname(__file__), fname), encoding='utf-8') as cfile:
return cfile.read()
setup(
name = "nas_bench_201",
version = "1.2",
author = "Xuanyi Dong",
author_email = "dongxuanyi888@gmail.com",
description = "API for NAS-Bench-201 (a benchmark for neural architecture search).",
license = "MIT",
keywords = "NAS Dataset API DeepLearning",
url = "https://github.com/D-X-Y/NAS-Bench-201",
packages=['nas_201_api'],
long_description=read('README.md'),
long_description_content_type='text/markdown',
classifiers=[
"Programming Language :: Python",
"Topic :: Database",
"Topic :: Scientific/Engineering :: Artificial Intelligence",
"License :: OSI Approved :: MIT License",
],
)
<commit_msg>Upgrade API to v1.3 and benchmark file to v1.1<commit_after> | #####################################################
# Copyright (c) Xuanyi Dong [GitHub D-X-Y], 2019.08 #
#####################################################
# [2020.02.25] Initialize the API as v1.1
# [2020.03.09] Upgrade the API to v1.2
# [2020.03.16] Upgrade the API to v1.3
import os
from setuptools import setup
def read(fname='README.md'):
with open(os.path.join(os.path.dirname(__file__), fname), encoding='utf-8') as cfile:
return cfile.read()
setup(
name = "nas_bench_201",
version = "1.3",
author = "Xuanyi Dong",
author_email = "dongxuanyi888@gmail.com",
description = "API for NAS-Bench-201 (a benchmark for neural architecture search).",
license = "MIT",
keywords = "NAS Dataset API DeepLearning",
url = "https://github.com/D-X-Y/NAS-Bench-201",
packages=['nas_201_api'],
long_description=read('README.md'),
long_description_content_type='text/markdown',
classifiers=[
"Programming Language :: Python",
"Topic :: Database",
"Topic :: Scientific/Engineering :: Artificial Intelligence",
"License :: OSI Approved :: MIT License",
],
)
| #####################################################
# Copyright (c) Xuanyi Dong [GitHub D-X-Y], 2019.08 #
#####################################################
# [2020.03.09] Upgrade to v1.2
import os
from setuptools import setup
def read(fname='README.md'):
with open(os.path.join(os.path.dirname(__file__), fname), encoding='utf-8') as cfile:
return cfile.read()
setup(
name = "nas_bench_201",
version = "1.2",
author = "Xuanyi Dong",
author_email = "dongxuanyi888@gmail.com",
description = "API for NAS-Bench-201 (a benchmark for neural architecture search).",
license = "MIT",
keywords = "NAS Dataset API DeepLearning",
url = "https://github.com/D-X-Y/NAS-Bench-201",
packages=['nas_201_api'],
long_description=read('README.md'),
long_description_content_type='text/markdown',
classifiers=[
"Programming Language :: Python",
"Topic :: Database",
"Topic :: Scientific/Engineering :: Artificial Intelligence",
"License :: OSI Approved :: MIT License",
],
)
Upgrade API to v1.3 and benchmark file to v1.1#####################################################
# Copyright (c) Xuanyi Dong [GitHub D-X-Y], 2019.08 #
#####################################################
# [2020.02.25] Initialize the API as v1.1
# [2020.03.09] Upgrade the API to v1.2
# [2020.03.16] Upgrade the API to v1.3
import os
from setuptools import setup
def read(fname='README.md'):
with open(os.path.join(os.path.dirname(__file__), fname), encoding='utf-8') as cfile:
return cfile.read()
setup(
name = "nas_bench_201",
version = "1.3",
author = "Xuanyi Dong",
author_email = "dongxuanyi888@gmail.com",
description = "API for NAS-Bench-201 (a benchmark for neural architecture search).",
license = "MIT",
keywords = "NAS Dataset API DeepLearning",
url = "https://github.com/D-X-Y/NAS-Bench-201",
packages=['nas_201_api'],
long_description=read('README.md'),
long_description_content_type='text/markdown',
classifiers=[
"Programming Language :: Python",
"Topic :: Database",
"Topic :: Scientific/Engineering :: Artificial Intelligence",
"License :: OSI Approved :: MIT License",
],
)
| <commit_before>#####################################################
# Copyright (c) Xuanyi Dong [GitHub D-X-Y], 2019.08 #
#####################################################
# [2020.03.09] Upgrade to v1.2
import os
from setuptools import setup
def read(fname='README.md'):
with open(os.path.join(os.path.dirname(__file__), fname), encoding='utf-8') as cfile:
return cfile.read()
setup(
name = "nas_bench_201",
version = "1.2",
author = "Xuanyi Dong",
author_email = "dongxuanyi888@gmail.com",
description = "API for NAS-Bench-201 (a benchmark for neural architecture search).",
license = "MIT",
keywords = "NAS Dataset API DeepLearning",
url = "https://github.com/D-X-Y/NAS-Bench-201",
packages=['nas_201_api'],
long_description=read('README.md'),
long_description_content_type='text/markdown',
classifiers=[
"Programming Language :: Python",
"Topic :: Database",
"Topic :: Scientific/Engineering :: Artificial Intelligence",
"License :: OSI Approved :: MIT License",
],
)
<commit_msg>Upgrade API to v1.3 and benchmark file to v1.1<commit_after>#####################################################
# Copyright (c) Xuanyi Dong [GitHub D-X-Y], 2019.08 #
#####################################################
# [2020.02.25] Initialize the API as v1.1
# [2020.03.09] Upgrade the API to v1.2
# [2020.03.16] Upgrade the API to v1.3
import os
from setuptools import setup
def read(fname='README.md'):
with open(os.path.join(os.path.dirname(__file__), fname), encoding='utf-8') as cfile:
return cfile.read()
setup(
name = "nas_bench_201",
version = "1.3",
author = "Xuanyi Dong",
author_email = "dongxuanyi888@gmail.com",
description = "API for NAS-Bench-201 (a benchmark for neural architecture search).",
license = "MIT",
keywords = "NAS Dataset API DeepLearning",
url = "https://github.com/D-X-Y/NAS-Bench-201",
packages=['nas_201_api'],
long_description=read('README.md'),
long_description_content_type='text/markdown',
classifiers=[
"Programming Language :: Python",
"Topic :: Database",
"Topic :: Scientific/Engineering :: Artificial Intelligence",
"License :: OSI Approved :: MIT License",
],
)
|
2bcff1f482fe4209dcf1bd53f9b535fefdd82aa3 | setup.py | setup.py | from setuptools import setup, find_packages
setup(name='googleanalytics',
description='A wrapper for the Google Analytics API.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
url='https://github.com/debrouwere/google-analytics/',
download_url='http://www.github.com/debrouwere/google-analytics/tarball/master',
version='0.8.4',
license='ISC',
packages=find_packages(),
keywords='data analytics api wrapper google',
scripts=[
'bin/googleanalytics'
],
install_requires=[
'oauth2client==1.3',
'google-api-python-client==1.3',
'python-dateutil==1.5',
'addressable>=1',
'inspector>=0.2',
'flask==0.10',
'keyring==4',
'click==3.3',
'pyyaml>=3'
],
test_suite='googleanalytics.tests',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Scientific/Engineering :: Information Analysis',
],
) | from setuptools import setup, find_packages
setup(name='googleanalytics',
description='A wrapper for the Google Analytics API.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
url='https://github.com/debrouwere/google-analytics/',
download_url='http://www.github.com/debrouwere/google-analytics/tarball/master',
version='0.8.5',
license='ISC',
packages=find_packages(),
keywords='data analytics api wrapper google',
scripts=[
'bin/googleanalytics'
],
install_requires=[
'oauth2client==1.3',
'google-api-python-client==1.3',
'python-dateutil==1.5',
'addressable>=1',
'inspect-it>=0.2',
'flask==0.10',
'keyring==4',
'click==3.3',
'pyyaml>=3'
],
test_suite='googleanalytics.tests',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Scientific/Engineering :: Information Analysis',
],
) | Deal with dependency rename: inspector to inspect-it. | Deal with dependency rename: inspector to inspect-it.
| Python | isc | debrouwere/google-analytics | from setuptools import setup, find_packages
setup(name='googleanalytics',
description='A wrapper for the Google Analytics API.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
url='https://github.com/debrouwere/google-analytics/',
download_url='http://www.github.com/debrouwere/google-analytics/tarball/master',
version='0.8.4',
license='ISC',
packages=find_packages(),
keywords='data analytics api wrapper google',
scripts=[
'bin/googleanalytics'
],
install_requires=[
'oauth2client==1.3',
'google-api-python-client==1.3',
'python-dateutil==1.5',
'addressable>=1',
'inspector>=0.2',
'flask==0.10',
'keyring==4',
'click==3.3',
'pyyaml>=3'
],
test_suite='googleanalytics.tests',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Scientific/Engineering :: Information Analysis',
],
)Deal with dependency rename: inspector to inspect-it. | from setuptools import setup, find_packages
setup(name='googleanalytics',
description='A wrapper for the Google Analytics API.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
url='https://github.com/debrouwere/google-analytics/',
download_url='http://www.github.com/debrouwere/google-analytics/tarball/master',
version='0.8.5',
license='ISC',
packages=find_packages(),
keywords='data analytics api wrapper google',
scripts=[
'bin/googleanalytics'
],
install_requires=[
'oauth2client==1.3',
'google-api-python-client==1.3',
'python-dateutil==1.5',
'addressable>=1',
'inspect-it>=0.2',
'flask==0.10',
'keyring==4',
'click==3.3',
'pyyaml>=3'
],
test_suite='googleanalytics.tests',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Scientific/Engineering :: Information Analysis',
],
) | <commit_before>from setuptools import setup, find_packages
setup(name='googleanalytics',
description='A wrapper for the Google Analytics API.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
url='https://github.com/debrouwere/google-analytics/',
download_url='http://www.github.com/debrouwere/google-analytics/tarball/master',
version='0.8.4',
license='ISC',
packages=find_packages(),
keywords='data analytics api wrapper google',
scripts=[
'bin/googleanalytics'
],
install_requires=[
'oauth2client==1.3',
'google-api-python-client==1.3',
'python-dateutil==1.5',
'addressable>=1',
'inspector>=0.2',
'flask==0.10',
'keyring==4',
'click==3.3',
'pyyaml>=3'
],
test_suite='googleanalytics.tests',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Scientific/Engineering :: Information Analysis',
],
)<commit_msg>Deal with dependency rename: inspector to inspect-it.<commit_after> | from setuptools import setup, find_packages
setup(name='googleanalytics',
description='A wrapper for the Google Analytics API.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
url='https://github.com/debrouwere/google-analytics/',
download_url='http://www.github.com/debrouwere/google-analytics/tarball/master',
version='0.8.5',
license='ISC',
packages=find_packages(),
keywords='data analytics api wrapper google',
scripts=[
'bin/googleanalytics'
],
install_requires=[
'oauth2client==1.3',
'google-api-python-client==1.3',
'python-dateutil==1.5',
'addressable>=1',
'inspect-it>=0.2',
'flask==0.10',
'keyring==4',
'click==3.3',
'pyyaml>=3'
],
test_suite='googleanalytics.tests',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Scientific/Engineering :: Information Analysis',
],
) | from setuptools import setup, find_packages
setup(name='googleanalytics',
description='A wrapper for the Google Analytics API.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
url='https://github.com/debrouwere/google-analytics/',
download_url='http://www.github.com/debrouwere/google-analytics/tarball/master',
version='0.8.4',
license='ISC',
packages=find_packages(),
keywords='data analytics api wrapper google',
scripts=[
'bin/googleanalytics'
],
install_requires=[
'oauth2client==1.3',
'google-api-python-client==1.3',
'python-dateutil==1.5',
'addressable>=1',
'inspector>=0.2',
'flask==0.10',
'keyring==4',
'click==3.3',
'pyyaml>=3'
],
test_suite='googleanalytics.tests',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Scientific/Engineering :: Information Analysis',
],
)Deal with dependency rename: inspector to inspect-it.from setuptools import setup, find_packages
setup(name='googleanalytics',
description='A wrapper for the Google Analytics API.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
url='https://github.com/debrouwere/google-analytics/',
download_url='http://www.github.com/debrouwere/google-analytics/tarball/master',
version='0.8.5',
license='ISC',
packages=find_packages(),
keywords='data analytics api wrapper google',
scripts=[
'bin/googleanalytics'
],
install_requires=[
'oauth2client==1.3',
'google-api-python-client==1.3',
'python-dateutil==1.5',
'addressable>=1',
'inspect-it>=0.2',
'flask==0.10',
'keyring==4',
'click==3.3',
'pyyaml>=3'
],
test_suite='googleanalytics.tests',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Scientific/Engineering :: Information Analysis',
],
) | <commit_before>from setuptools import setup, find_packages
setup(name='googleanalytics',
description='A wrapper for the Google Analytics API.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
url='https://github.com/debrouwere/google-analytics/',
download_url='http://www.github.com/debrouwere/google-analytics/tarball/master',
version='0.8.4',
license='ISC',
packages=find_packages(),
keywords='data analytics api wrapper google',
scripts=[
'bin/googleanalytics'
],
install_requires=[
'oauth2client==1.3',
'google-api-python-client==1.3',
'python-dateutil==1.5',
'addressable>=1',
'inspector>=0.2',
'flask==0.10',
'keyring==4',
'click==3.3',
'pyyaml>=3'
],
test_suite='googleanalytics.tests',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Scientific/Engineering :: Information Analysis',
],
)<commit_msg>Deal with dependency rename: inspector to inspect-it.<commit_after>from setuptools import setup, find_packages
setup(name='googleanalytics',
description='A wrapper for the Google Analytics API.',
long_description=open('README.rst').read(),
author='Stijn Debrouwere',
author_email='stijn@debrouwere.org',
url='https://github.com/debrouwere/google-analytics/',
download_url='http://www.github.com/debrouwere/google-analytics/tarball/master',
version='0.8.5',
license='ISC',
packages=find_packages(),
keywords='data analytics api wrapper google',
scripts=[
'bin/googleanalytics'
],
install_requires=[
'oauth2client==1.3',
'google-api-python-client==1.3',
'python-dateutil==1.5',
'addressable>=1',
'inspect-it>=0.2',
'flask==0.10',
'keyring==4',
'click==3.3',
'pyyaml>=3'
],
test_suite='googleanalytics.tests',
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Scientific/Engineering :: Information Analysis',
],
) |
f7459258c0d19de00c042768c68aeeb699c4f04c | setup.py | setup.py | from setuptools import setup, find_packages
import pygout
readme = open('README.rst', 'r').read()
DESCRIPTION = readme.split('\n')[0]
LONG_DESCRIPTION = readme
INSTALL_REQUIRES = [
'Pygments == 1.5',
]
setup(
name = 'PygOut',
version = pygout.__version__,
url = 'http://github.com/alanbriolat/PygOut',
license = 'BSD License',
author = 'Alan Briolat, Helen Gray',
description = DESCRIPTION,
long_descrption = LONG_DESCRIPTION,
packages = find_packages(),
platforms = 'any',
install_requires = INSTALL_REQUIRES,
entry_points = {
'console_scripts': ['pygout = pygout.cmdline:main'],
},
classifiers = [
'Development Status :: 2 - Pre-Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: End Users/Desktop',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
],
)
| import sys
from setuptools import setup, find_packages
import pygout
readme = open('README.rst', 'r').read()
DESCRIPTION = readme.split('\n')[0]
LONG_DESCRIPTION = readme
INSTALL_REQUIRES = [
'Pygments == 1.5',
]
if sys.version_info < (2, 7):
INSTALL_REQUIRES.append('argparse >= 1.1')
if sys.version_info < (3, 0):
INSTALL_REQUIRES.append('configparser >= 3.0')
setup(
name = 'PygOut',
version = pygout.__version__,
url = 'http://github.com/alanbriolat/PygOut',
license = 'BSD License',
author = 'Alan Briolat, Helen M. Gray',
description = DESCRIPTION,
long_descrption = LONG_DESCRIPTION,
packages = find_packages(),
platforms = 'any',
install_requires = INSTALL_REQUIRES,
entry_points = {
'console_scripts': ['pygout = pygout.cmdline:main'],
},
classifiers = [
'Development Status :: 2 - Pre-Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: End Users/Desktop',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
],
)
| Add dependency on argparse and new configparser | Add dependency on argparse and new configparser
| Python | bsd-3-clause | alanbriolat/PygOut | from setuptools import setup, find_packages
import pygout
readme = open('README.rst', 'r').read()
DESCRIPTION = readme.split('\n')[0]
LONG_DESCRIPTION = readme
INSTALL_REQUIRES = [
'Pygments == 1.5',
]
setup(
name = 'PygOut',
version = pygout.__version__,
url = 'http://github.com/alanbriolat/PygOut',
license = 'BSD License',
author = 'Alan Briolat, Helen Gray',
description = DESCRIPTION,
long_descrption = LONG_DESCRIPTION,
packages = find_packages(),
platforms = 'any',
install_requires = INSTALL_REQUIRES,
entry_points = {
'console_scripts': ['pygout = pygout.cmdline:main'],
},
classifiers = [
'Development Status :: 2 - Pre-Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: End Users/Desktop',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
],
)
Add dependency on argparse and new configparser | import sys
from setuptools import setup, find_packages
import pygout
readme = open('README.rst', 'r').read()
DESCRIPTION = readme.split('\n')[0]
LONG_DESCRIPTION = readme
INSTALL_REQUIRES = [
'Pygments == 1.5',
]
if sys.version_info < (2, 7):
INSTALL_REQUIRES.append('argparse >= 1.1')
if sys.version_info < (3, 0):
INSTALL_REQUIRES.append('configparser >= 3.0')
setup(
name = 'PygOut',
version = pygout.__version__,
url = 'http://github.com/alanbriolat/PygOut',
license = 'BSD License',
author = 'Alan Briolat, Helen M. Gray',
description = DESCRIPTION,
long_descrption = LONG_DESCRIPTION,
packages = find_packages(),
platforms = 'any',
install_requires = INSTALL_REQUIRES,
entry_points = {
'console_scripts': ['pygout = pygout.cmdline:main'],
},
classifiers = [
'Development Status :: 2 - Pre-Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: End Users/Desktop',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
],
)
| <commit_before>from setuptools import setup, find_packages
import pygout
readme = open('README.rst', 'r').read()
DESCRIPTION = readme.split('\n')[0]
LONG_DESCRIPTION = readme
INSTALL_REQUIRES = [
'Pygments == 1.5',
]
setup(
name = 'PygOut',
version = pygout.__version__,
url = 'http://github.com/alanbriolat/PygOut',
license = 'BSD License',
author = 'Alan Briolat, Helen Gray',
description = DESCRIPTION,
long_descrption = LONG_DESCRIPTION,
packages = find_packages(),
platforms = 'any',
install_requires = INSTALL_REQUIRES,
entry_points = {
'console_scripts': ['pygout = pygout.cmdline:main'],
},
classifiers = [
'Development Status :: 2 - Pre-Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: End Users/Desktop',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
],
)
<commit_msg>Add dependency on argparse and new configparser<commit_after> | import sys
from setuptools import setup, find_packages
import pygout
readme = open('README.rst', 'r').read()
DESCRIPTION = readme.split('\n')[0]
LONG_DESCRIPTION = readme
INSTALL_REQUIRES = [
'Pygments == 1.5',
]
if sys.version_info < (2, 7):
INSTALL_REQUIRES.append('argparse >= 1.1')
if sys.version_info < (3, 0):
INSTALL_REQUIRES.append('configparser >= 3.0')
setup(
name = 'PygOut',
version = pygout.__version__,
url = 'http://github.com/alanbriolat/PygOut',
license = 'BSD License',
author = 'Alan Briolat, Helen M. Gray',
description = DESCRIPTION,
long_descrption = LONG_DESCRIPTION,
packages = find_packages(),
platforms = 'any',
install_requires = INSTALL_REQUIRES,
entry_points = {
'console_scripts': ['pygout = pygout.cmdline:main'],
},
classifiers = [
'Development Status :: 2 - Pre-Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: End Users/Desktop',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
],
)
| from setuptools import setup, find_packages
import pygout
readme = open('README.rst', 'r').read()
DESCRIPTION = readme.split('\n')[0]
LONG_DESCRIPTION = readme
INSTALL_REQUIRES = [
'Pygments == 1.5',
]
setup(
name = 'PygOut',
version = pygout.__version__,
url = 'http://github.com/alanbriolat/PygOut',
license = 'BSD License',
author = 'Alan Briolat, Helen Gray',
description = DESCRIPTION,
long_descrption = LONG_DESCRIPTION,
packages = find_packages(),
platforms = 'any',
install_requires = INSTALL_REQUIRES,
entry_points = {
'console_scripts': ['pygout = pygout.cmdline:main'],
},
classifiers = [
'Development Status :: 2 - Pre-Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: End Users/Desktop',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
],
)
Add dependency on argparse and new configparserimport sys
from setuptools import setup, find_packages
import pygout
readme = open('README.rst', 'r').read()
DESCRIPTION = readme.split('\n')[0]
LONG_DESCRIPTION = readme
INSTALL_REQUIRES = [
'Pygments == 1.5',
]
if sys.version_info < (2, 7):
INSTALL_REQUIRES.append('argparse >= 1.1')
if sys.version_info < (3, 0):
INSTALL_REQUIRES.append('configparser >= 3.0')
setup(
name = 'PygOut',
version = pygout.__version__,
url = 'http://github.com/alanbriolat/PygOut',
license = 'BSD License',
author = 'Alan Briolat, Helen M. Gray',
description = DESCRIPTION,
long_descrption = LONG_DESCRIPTION,
packages = find_packages(),
platforms = 'any',
install_requires = INSTALL_REQUIRES,
entry_points = {
'console_scripts': ['pygout = pygout.cmdline:main'],
},
classifiers = [
'Development Status :: 2 - Pre-Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: End Users/Desktop',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
],
)
| <commit_before>from setuptools import setup, find_packages
import pygout
readme = open('README.rst', 'r').read()
DESCRIPTION = readme.split('\n')[0]
LONG_DESCRIPTION = readme
INSTALL_REQUIRES = [
'Pygments == 1.5',
]
setup(
name = 'PygOut',
version = pygout.__version__,
url = 'http://github.com/alanbriolat/PygOut',
license = 'BSD License',
author = 'Alan Briolat, Helen Gray',
description = DESCRIPTION,
long_descrption = LONG_DESCRIPTION,
packages = find_packages(),
platforms = 'any',
install_requires = INSTALL_REQUIRES,
entry_points = {
'console_scripts': ['pygout = pygout.cmdline:main'],
},
classifiers = [
'Development Status :: 2 - Pre-Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: End Users/Desktop',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
],
)
<commit_msg>Add dependency on argparse and new configparser<commit_after>import sys
from setuptools import setup, find_packages
import pygout
readme = open('README.rst', 'r').read()
DESCRIPTION = readme.split('\n')[0]
LONG_DESCRIPTION = readme
INSTALL_REQUIRES = [
'Pygments == 1.5',
]
if sys.version_info < (2, 7):
INSTALL_REQUIRES.append('argparse >= 1.1')
if sys.version_info < (3, 0):
INSTALL_REQUIRES.append('configparser >= 3.0')
setup(
name = 'PygOut',
version = pygout.__version__,
url = 'http://github.com/alanbriolat/PygOut',
license = 'BSD License',
author = 'Alan Briolat, Helen M. Gray',
description = DESCRIPTION,
long_descrption = LONG_DESCRIPTION,
packages = find_packages(),
platforms = 'any',
install_requires = INSTALL_REQUIRES,
entry_points = {
'console_scripts': ['pygout = pygout.cmdline:main'],
},
classifiers = [
'Development Status :: 2 - Pre-Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: End Users/Desktop',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
],
)
|
98a58ee53e918b0587e1a792c5d08c3c005a7770 | setup.py | setup.py | import sys
from setuptools import setup
meta = dict(
name="stacklogger",
version="0.1.0",
description="A stack-aware logging extension",
author="Will Maier",
author_email="willmaier@ml1.net",
py_modules=["stacklogger"],
test_suite="tests.py",
install_requires=["setuptools"],
keywords="logging stack frame",
url="http://packages.python.org/stacklogger",
classifiers=[
"Development Status :: 3 - Alpha",
"Intended Audience :: Developers",
"License :: OSI Approved :: BSD License",
"License :: OSI Approved :: MIT License",
"Natural Language :: English",
"Programming Language :: Python"
"Topic :: System :: Logging",
],
)
# Automatic conversion for Python 3 requires distribute.
if False and sys.version_info >= (3,):
meta.update(dict(
use_2to3=True,
))
setup(**meta)
| import sys
from setuptools import setup
meta = dict(
name="stacklogger",
version="0.1.0",
description="A stack-aware logging extension",
author="Will Maier",
author_email="willmaier@ml1.net",
py_modules=["stacklogger"],
test_suite="tests",
install_requires=["setuptools"],
keywords="logging stack frame",
url="http://packages.python.org/stacklogger",
classifiers=[
"Development Status :: 3 - Alpha",
"Intended Audience :: Developers",
"License :: OSI Approved :: BSD License",
"License :: OSI Approved :: MIT License",
"Natural Language :: English",
"Programming Language :: Python"
"Topic :: System :: Logging",
],
)
# Automatic conversion for Python 3 requires distribute.
if False and sys.version_info >= (3,):
meta.update(dict(
use_2to3=True,
))
setup(**meta)
| Fix test_suite (uses module, not file, name). | Fix test_suite (uses module, not file, name).
| Python | isc | whilp/stacklogger | import sys
from setuptools import setup
meta = dict(
name="stacklogger",
version="0.1.0",
description="A stack-aware logging extension",
author="Will Maier",
author_email="willmaier@ml1.net",
py_modules=["stacklogger"],
test_suite="tests.py",
install_requires=["setuptools"],
keywords="logging stack frame",
url="http://packages.python.org/stacklogger",
classifiers=[
"Development Status :: 3 - Alpha",
"Intended Audience :: Developers",
"License :: OSI Approved :: BSD License",
"License :: OSI Approved :: MIT License",
"Natural Language :: English",
"Programming Language :: Python"
"Topic :: System :: Logging",
],
)
# Automatic conversion for Python 3 requires distribute.
if False and sys.version_info >= (3,):
meta.update(dict(
use_2to3=True,
))
setup(**meta)
Fix test_suite (uses module, not file, name). | import sys
from setuptools import setup
meta = dict(
name="stacklogger",
version="0.1.0",
description="A stack-aware logging extension",
author="Will Maier",
author_email="willmaier@ml1.net",
py_modules=["stacklogger"],
test_suite="tests",
install_requires=["setuptools"],
keywords="logging stack frame",
url="http://packages.python.org/stacklogger",
classifiers=[
"Development Status :: 3 - Alpha",
"Intended Audience :: Developers",
"License :: OSI Approved :: BSD License",
"License :: OSI Approved :: MIT License",
"Natural Language :: English",
"Programming Language :: Python"
"Topic :: System :: Logging",
],
)
# Automatic conversion for Python 3 requires distribute.
if False and sys.version_info >= (3,):
meta.update(dict(
use_2to3=True,
))
setup(**meta)
| <commit_before>import sys
from setuptools import setup
meta = dict(
name="stacklogger",
version="0.1.0",
description="A stack-aware logging extension",
author="Will Maier",
author_email="willmaier@ml1.net",
py_modules=["stacklogger"],
test_suite="tests.py",
install_requires=["setuptools"],
keywords="logging stack frame",
url="http://packages.python.org/stacklogger",
classifiers=[
"Development Status :: 3 - Alpha",
"Intended Audience :: Developers",
"License :: OSI Approved :: BSD License",
"License :: OSI Approved :: MIT License",
"Natural Language :: English",
"Programming Language :: Python"
"Topic :: System :: Logging",
],
)
# Automatic conversion for Python 3 requires distribute.
if False and sys.version_info >= (3,):
meta.update(dict(
use_2to3=True,
))
setup(**meta)
<commit_msg>Fix test_suite (uses module, not file, name).<commit_after> | import sys
from setuptools import setup
meta = dict(
name="stacklogger",
version="0.1.0",
description="A stack-aware logging extension",
author="Will Maier",
author_email="willmaier@ml1.net",
py_modules=["stacklogger"],
test_suite="tests",
install_requires=["setuptools"],
keywords="logging stack frame",
url="http://packages.python.org/stacklogger",
classifiers=[
"Development Status :: 3 - Alpha",
"Intended Audience :: Developers",
"License :: OSI Approved :: BSD License",
"License :: OSI Approved :: MIT License",
"Natural Language :: English",
"Programming Language :: Python"
"Topic :: System :: Logging",
],
)
# Automatic conversion for Python 3 requires distribute.
if False and sys.version_info >= (3,):
meta.update(dict(
use_2to3=True,
))
setup(**meta)
| import sys
from setuptools import setup
meta = dict(
name="stacklogger",
version="0.1.0",
description="A stack-aware logging extension",
author="Will Maier",
author_email="willmaier@ml1.net",
py_modules=["stacklogger"],
test_suite="tests.py",
install_requires=["setuptools"],
keywords="logging stack frame",
url="http://packages.python.org/stacklogger",
classifiers=[
"Development Status :: 3 - Alpha",
"Intended Audience :: Developers",
"License :: OSI Approved :: BSD License",
"License :: OSI Approved :: MIT License",
"Natural Language :: English",
"Programming Language :: Python"
"Topic :: System :: Logging",
],
)
# Automatic conversion for Python 3 requires distribute.
if False and sys.version_info >= (3,):
meta.update(dict(
use_2to3=True,
))
setup(**meta)
Fix test_suite (uses module, not file, name).import sys
from setuptools import setup
meta = dict(
name="stacklogger",
version="0.1.0",
description="A stack-aware logging extension",
author="Will Maier",
author_email="willmaier@ml1.net",
py_modules=["stacklogger"],
test_suite="tests",
install_requires=["setuptools"],
keywords="logging stack frame",
url="http://packages.python.org/stacklogger",
classifiers=[
"Development Status :: 3 - Alpha",
"Intended Audience :: Developers",
"License :: OSI Approved :: BSD License",
"License :: OSI Approved :: MIT License",
"Natural Language :: English",
"Programming Language :: Python"
"Topic :: System :: Logging",
],
)
# Automatic conversion for Python 3 requires distribute.
if False and sys.version_info >= (3,):
meta.update(dict(
use_2to3=True,
))
setup(**meta)
| <commit_before>import sys
from setuptools import setup
meta = dict(
name="stacklogger",
version="0.1.0",
description="A stack-aware logging extension",
author="Will Maier",
author_email="willmaier@ml1.net",
py_modules=["stacklogger"],
test_suite="tests.py",
install_requires=["setuptools"],
keywords="logging stack frame",
url="http://packages.python.org/stacklogger",
classifiers=[
"Development Status :: 3 - Alpha",
"Intended Audience :: Developers",
"License :: OSI Approved :: BSD License",
"License :: OSI Approved :: MIT License",
"Natural Language :: English",
"Programming Language :: Python"
"Topic :: System :: Logging",
],
)
# Automatic conversion for Python 3 requires distribute.
if False and sys.version_info >= (3,):
meta.update(dict(
use_2to3=True,
))
setup(**meta)
<commit_msg>Fix test_suite (uses module, not file, name).<commit_after>import sys
from setuptools import setup
meta = dict(
name="stacklogger",
version="0.1.0",
description="A stack-aware logging extension",
author="Will Maier",
author_email="willmaier@ml1.net",
py_modules=["stacklogger"],
test_suite="tests",
install_requires=["setuptools"],
keywords="logging stack frame",
url="http://packages.python.org/stacklogger",
classifiers=[
"Development Status :: 3 - Alpha",
"Intended Audience :: Developers",
"License :: OSI Approved :: BSD License",
"License :: OSI Approved :: MIT License",
"Natural Language :: English",
"Programming Language :: Python"
"Topic :: System :: Logging",
],
)
# Automatic conversion for Python 3 requires distribute.
if False and sys.version_info >= (3,):
meta.update(dict(
use_2to3=True,
))
setup(**meta)
|
dc76cf3a5fcd60223ac903900a56945bb6ee607a | setup.py | setup.py | import sys
import os
from setuptools import setup
long_description = open('README.rst').read()
classifiers = [
'Development Status :: 4 - Beta',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: 3.6',
]
setup_kwargs = dict(
name='powershift-cli',
version='1.1.0',
description='Pluggable command line client for OpenShift.',
long_description=long_description,
url='https://github.com/getwarped/powershift-cli',
author='Graham Dumpleton',
author_email='Graham.Dumpleton@gmail.com',
license='BSD',
classifiers=classifiers,
keywords='openshift kubernetes',
packages=['powershift', 'powershift.cli'],
package_dir={'powershift': 'src/powershift'},
package_data={'powershift.cli': ['completion-bash.sh']},
entry_points = {'console_scripts':['powershift = powershift.cli:main']},
install_requires=['click'],
)
setup(**setup_kwargs)
| import sys
import os
from setuptools import setup
long_description = open('README.rst').read()
classifiers = [
'Development Status :: 4 - Beta',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: 3.6',
]
setup_kwargs = dict(
name='powershift-cli',
version='1.1.0',
description='Pluggable command line client for OpenShift.',
long_description=long_description,
url='https://github.com/getwarped/powershift-cli',
author='Graham Dumpleton',
author_email='Graham.Dumpleton@gmail.com',
license='BSD',
classifiers=classifiers,
keywords='openshift kubernetes',
packages=['powershift', 'powershift.cli'],
package_dir={'powershift': 'src/powershift'},
package_data={'powershift.cli': ['completion-bash.sh']},
entry_points = {'console_scripts':['powershift = powershift.cli:main']},
install_requires=['click'],
extra_requires={'cluster': ['powershift-cluster>=1.1.0']},
)
setup(**setup_kwargs)
| Allow pulling in of cluster plugin using powershift-cli[cluster] as install target. | Allow pulling in of cluster plugin using powershift-cli[cluster] as install target.
| Python | bsd-2-clause | getwarped/powershift-cli,getwarped/powershift-cli | import sys
import os
from setuptools import setup
long_description = open('README.rst').read()
classifiers = [
'Development Status :: 4 - Beta',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: 3.6',
]
setup_kwargs = dict(
name='powershift-cli',
version='1.1.0',
description='Pluggable command line client for OpenShift.',
long_description=long_description,
url='https://github.com/getwarped/powershift-cli',
author='Graham Dumpleton',
author_email='Graham.Dumpleton@gmail.com',
license='BSD',
classifiers=classifiers,
keywords='openshift kubernetes',
packages=['powershift', 'powershift.cli'],
package_dir={'powershift': 'src/powershift'},
package_data={'powershift.cli': ['completion-bash.sh']},
entry_points = {'console_scripts':['powershift = powershift.cli:main']},
install_requires=['click'],
)
setup(**setup_kwargs)
Allow pulling in of cluster plugin using powershift-cli[cluster] as install target. | import sys
import os
from setuptools import setup
long_description = open('README.rst').read()
classifiers = [
'Development Status :: 4 - Beta',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: 3.6',
]
setup_kwargs = dict(
name='powershift-cli',
version='1.1.0',
description='Pluggable command line client for OpenShift.',
long_description=long_description,
url='https://github.com/getwarped/powershift-cli',
author='Graham Dumpleton',
author_email='Graham.Dumpleton@gmail.com',
license='BSD',
classifiers=classifiers,
keywords='openshift kubernetes',
packages=['powershift', 'powershift.cli'],
package_dir={'powershift': 'src/powershift'},
package_data={'powershift.cli': ['completion-bash.sh']},
entry_points = {'console_scripts':['powershift = powershift.cli:main']},
install_requires=['click'],
extra_requires={'cluster': ['powershift-cluster>=1.1.0']},
)
setup(**setup_kwargs)
| <commit_before>import sys
import os
from setuptools import setup
long_description = open('README.rst').read()
classifiers = [
'Development Status :: 4 - Beta',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: 3.6',
]
setup_kwargs = dict(
name='powershift-cli',
version='1.1.0',
description='Pluggable command line client for OpenShift.',
long_description=long_description,
url='https://github.com/getwarped/powershift-cli',
author='Graham Dumpleton',
author_email='Graham.Dumpleton@gmail.com',
license='BSD',
classifiers=classifiers,
keywords='openshift kubernetes',
packages=['powershift', 'powershift.cli'],
package_dir={'powershift': 'src/powershift'},
package_data={'powershift.cli': ['completion-bash.sh']},
entry_points = {'console_scripts':['powershift = powershift.cli:main']},
install_requires=['click'],
)
setup(**setup_kwargs)
<commit_msg>Allow pulling in of cluster plugin using powershift-cli[cluster] as install target.<commit_after> | import sys
import os
from setuptools import setup
long_description = open('README.rst').read()
classifiers = [
'Development Status :: 4 - Beta',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: 3.6',
]
setup_kwargs = dict(
name='powershift-cli',
version='1.1.0',
description='Pluggable command line client for OpenShift.',
long_description=long_description,
url='https://github.com/getwarped/powershift-cli',
author='Graham Dumpleton',
author_email='Graham.Dumpleton@gmail.com',
license='BSD',
classifiers=classifiers,
keywords='openshift kubernetes',
packages=['powershift', 'powershift.cli'],
package_dir={'powershift': 'src/powershift'},
package_data={'powershift.cli': ['completion-bash.sh']},
entry_points = {'console_scripts':['powershift = powershift.cli:main']},
install_requires=['click'],
extra_requires={'cluster': ['powershift-cluster>=1.1.0']},
)
setup(**setup_kwargs)
| import sys
import os
from setuptools import setup
long_description = open('README.rst').read()
classifiers = [
'Development Status :: 4 - Beta',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: 3.6',
]
setup_kwargs = dict(
name='powershift-cli',
version='1.1.0',
description='Pluggable command line client for OpenShift.',
long_description=long_description,
url='https://github.com/getwarped/powershift-cli',
author='Graham Dumpleton',
author_email='Graham.Dumpleton@gmail.com',
license='BSD',
classifiers=classifiers,
keywords='openshift kubernetes',
packages=['powershift', 'powershift.cli'],
package_dir={'powershift': 'src/powershift'},
package_data={'powershift.cli': ['completion-bash.sh']},
entry_points = {'console_scripts':['powershift = powershift.cli:main']},
install_requires=['click'],
)
setup(**setup_kwargs)
Allow pulling in of cluster plugin using powershift-cli[cluster] as install target.import sys
import os
from setuptools import setup
long_description = open('README.rst').read()
classifiers = [
'Development Status :: 4 - Beta',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: 3.6',
]
setup_kwargs = dict(
name='powershift-cli',
version='1.1.0',
description='Pluggable command line client for OpenShift.',
long_description=long_description,
url='https://github.com/getwarped/powershift-cli',
author='Graham Dumpleton',
author_email='Graham.Dumpleton@gmail.com',
license='BSD',
classifiers=classifiers,
keywords='openshift kubernetes',
packages=['powershift', 'powershift.cli'],
package_dir={'powershift': 'src/powershift'},
package_data={'powershift.cli': ['completion-bash.sh']},
entry_points = {'console_scripts':['powershift = powershift.cli:main']},
install_requires=['click'],
extra_requires={'cluster': ['powershift-cluster>=1.1.0']},
)
setup(**setup_kwargs)
| <commit_before>import sys
import os
from setuptools import setup
long_description = open('README.rst').read()
classifiers = [
'Development Status :: 4 - Beta',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: 3.6',
]
setup_kwargs = dict(
name='powershift-cli',
version='1.1.0',
description='Pluggable command line client for OpenShift.',
long_description=long_description,
url='https://github.com/getwarped/powershift-cli',
author='Graham Dumpleton',
author_email='Graham.Dumpleton@gmail.com',
license='BSD',
classifiers=classifiers,
keywords='openshift kubernetes',
packages=['powershift', 'powershift.cli'],
package_dir={'powershift': 'src/powershift'},
package_data={'powershift.cli': ['completion-bash.sh']},
entry_points = {'console_scripts':['powershift = powershift.cli:main']},
install_requires=['click'],
)
setup(**setup_kwargs)
<commit_msg>Allow pulling in of cluster plugin using powershift-cli[cluster] as install target.<commit_after>import sys
import os
from setuptools import setup
long_description = open('README.rst').read()
classifiers = [
'Development Status :: 4 - Beta',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: Python :: 3.6',
]
setup_kwargs = dict(
name='powershift-cli',
version='1.1.0',
description='Pluggable command line client for OpenShift.',
long_description=long_description,
url='https://github.com/getwarped/powershift-cli',
author='Graham Dumpleton',
author_email='Graham.Dumpleton@gmail.com',
license='BSD',
classifiers=classifiers,
keywords='openshift kubernetes',
packages=['powershift', 'powershift.cli'],
package_dir={'powershift': 'src/powershift'},
package_data={'powershift.cli': ['completion-bash.sh']},
entry_points = {'console_scripts':['powershift = powershift.cli:main']},
install_requires=['click'],
extra_requires={'cluster': ['powershift-cluster>=1.1.0']},
)
setup(**setup_kwargs)
|
b7eccddf76f484da10340f56215c0d07237d34ce | setup.py | setup.py | from distutils.core import setup
setup(
name='Pykka',
version='0.1',
author='Stein Magnus Jodal',
author_email='stein.magnus@jodal.no',
packages=['pykka'],
url='http://github.com/jodal/pykka',
license='Apache License, Version 2.0',
description='Pykka makes actors look like regular objects',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 1 - Planning',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Topic :: Software Development :: Libraries',
],
)
| from distutils.core import setup
setup(
name='Pykka',
version='0.1',
author='Stein Magnus Jodal',
author_email='stein.magnus@jodal.no',
packages=['pykka'],
url='http://github.com/jodal/pykka',
license='Apache License, Version 2.0',
description='Pykka makes actors look like regular objects',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Topic :: Software Development :: Libraries',
],
)
| Change dev status to pre-alpha | Change dev status to pre-alpha
| Python | apache-2.0 | tamland/pykka,jodal/pykka,tempbottle/pykka | from distutils.core import setup
setup(
name='Pykka',
version='0.1',
author='Stein Magnus Jodal',
author_email='stein.magnus@jodal.no',
packages=['pykka'],
url='http://github.com/jodal/pykka',
license='Apache License, Version 2.0',
description='Pykka makes actors look like regular objects',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 1 - Planning',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Topic :: Software Development :: Libraries',
],
)
Change dev status to pre-alpha | from distutils.core import setup
setup(
name='Pykka',
version='0.1',
author='Stein Magnus Jodal',
author_email='stein.magnus@jodal.no',
packages=['pykka'],
url='http://github.com/jodal/pykka',
license='Apache License, Version 2.0',
description='Pykka makes actors look like regular objects',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Topic :: Software Development :: Libraries',
],
)
| <commit_before>from distutils.core import setup
setup(
name='Pykka',
version='0.1',
author='Stein Magnus Jodal',
author_email='stein.magnus@jodal.no',
packages=['pykka'],
url='http://github.com/jodal/pykka',
license='Apache License, Version 2.0',
description='Pykka makes actors look like regular objects',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 1 - Planning',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Topic :: Software Development :: Libraries',
],
)
<commit_msg>Change dev status to pre-alpha<commit_after> | from distutils.core import setup
setup(
name='Pykka',
version='0.1',
author='Stein Magnus Jodal',
author_email='stein.magnus@jodal.no',
packages=['pykka'],
url='http://github.com/jodal/pykka',
license='Apache License, Version 2.0',
description='Pykka makes actors look like regular objects',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Topic :: Software Development :: Libraries',
],
)
| from distutils.core import setup
setup(
name='Pykka',
version='0.1',
author='Stein Magnus Jodal',
author_email='stein.magnus@jodal.no',
packages=['pykka'],
url='http://github.com/jodal/pykka',
license='Apache License, Version 2.0',
description='Pykka makes actors look like regular objects',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 1 - Planning',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Topic :: Software Development :: Libraries',
],
)
Change dev status to pre-alphafrom distutils.core import setup
setup(
name='Pykka',
version='0.1',
author='Stein Magnus Jodal',
author_email='stein.magnus@jodal.no',
packages=['pykka'],
url='http://github.com/jodal/pykka',
license='Apache License, Version 2.0',
description='Pykka makes actors look like regular objects',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Topic :: Software Development :: Libraries',
],
)
| <commit_before>from distutils.core import setup
setup(
name='Pykka',
version='0.1',
author='Stein Magnus Jodal',
author_email='stein.magnus@jodal.no',
packages=['pykka'],
url='http://github.com/jodal/pykka',
license='Apache License, Version 2.0',
description='Pykka makes actors look like regular objects',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 1 - Planning',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Topic :: Software Development :: Libraries',
],
)
<commit_msg>Change dev status to pre-alpha<commit_after>from distutils.core import setup
setup(
name='Pykka',
version='0.1',
author='Stein Magnus Jodal',
author_email='stein.magnus@jodal.no',
packages=['pykka'],
url='http://github.com/jodal/pykka',
license='Apache License, Version 2.0',
description='Pykka makes actors look like regular objects',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Topic :: Software Development :: Libraries',
],
)
|
83aa3020a585d41ae2d7f2a7759e42389c22bc65 | setup.py | setup.py | from setuptools import find_packages, setup
setup(name='satnogsclient',
packages=find_packages(),
version='0.2.4',
author='SatNOGS team',
author_email='client-dev@satnogs.org',
url='https://github.com/satnogs/satnogs-client/',
description='SatNOGS Client',
install_requires=['APScheduler',
'SQLAlchemy',
'requests',
'validators',
'python-dateutil',
'ephem',
'pytz'],
dependency_links=[
'git+https://github.com/brandon-rhodes/pyephem.git@47d0ba3616ee6c308f2eed319af3901592d00f70#egg=ephem'
],
scripts=['satnogsclient/bin/satnogs-poller',
'satnogsclient/bin/satnogs-task'])
| from setuptools import find_packages, setup
setup(name='satnogsclient',
packages=find_packages(),
version='0.2.4',
author='SatNOGS team',
author_email='client-dev@satnogs.org',
url='https://github.com/satnogs/satnogs-client/',
description='SatNOGS Client',
install_requires=['APScheduler',
'SQLAlchemy',
'requests',
'validators',
'python-dateutil',
'ephem',
'pytz'],
scripts=['satnogsclient/bin/satnogs-poller',
'satnogsclient/bin/satnogs-task'])
| Upgrade pyephem to latest version. | Upgrade pyephem to latest version.
* Pyephem v3.7.6.0 now includes this required patch:
https://github.com/brandon-rhodes/pyephem/commit/47d0ba3616ee6c308f2eed319af3901592d00f70
* No need to specify a custom dependency in our setup.py.
| Python | agpl-3.0 | adamkalis/satnogs-client,cshields/satnogs-client,cshields/satnogs-client,adamkalis/satnogs-client | from setuptools import find_packages, setup
setup(name='satnogsclient',
packages=find_packages(),
version='0.2.4',
author='SatNOGS team',
author_email='client-dev@satnogs.org',
url='https://github.com/satnogs/satnogs-client/',
description='SatNOGS Client',
install_requires=['APScheduler',
'SQLAlchemy',
'requests',
'validators',
'python-dateutil',
'ephem',
'pytz'],
dependency_links=[
'git+https://github.com/brandon-rhodes/pyephem.git@47d0ba3616ee6c308f2eed319af3901592d00f70#egg=ephem'
],
scripts=['satnogsclient/bin/satnogs-poller',
'satnogsclient/bin/satnogs-task'])
Upgrade pyephem to latest version.
* Pyephem v3.7.6.0 now includes this required patch:
https://github.com/brandon-rhodes/pyephem/commit/47d0ba3616ee6c308f2eed319af3901592d00f70
* No need to specify a custom dependency in our setup.py. | from setuptools import find_packages, setup
setup(name='satnogsclient',
packages=find_packages(),
version='0.2.4',
author='SatNOGS team',
author_email='client-dev@satnogs.org',
url='https://github.com/satnogs/satnogs-client/',
description='SatNOGS Client',
install_requires=['APScheduler',
'SQLAlchemy',
'requests',
'validators',
'python-dateutil',
'ephem',
'pytz'],
scripts=['satnogsclient/bin/satnogs-poller',
'satnogsclient/bin/satnogs-task'])
| <commit_before>from setuptools import find_packages, setup
setup(name='satnogsclient',
packages=find_packages(),
version='0.2.4',
author='SatNOGS team',
author_email='client-dev@satnogs.org',
url='https://github.com/satnogs/satnogs-client/',
description='SatNOGS Client',
install_requires=['APScheduler',
'SQLAlchemy',
'requests',
'validators',
'python-dateutil',
'ephem',
'pytz'],
dependency_links=[
'git+https://github.com/brandon-rhodes/pyephem.git@47d0ba3616ee6c308f2eed319af3901592d00f70#egg=ephem'
],
scripts=['satnogsclient/bin/satnogs-poller',
'satnogsclient/bin/satnogs-task'])
<commit_msg>Upgrade pyephem to latest version.
* Pyephem v3.7.6.0 now includes this required patch:
https://github.com/brandon-rhodes/pyephem/commit/47d0ba3616ee6c308f2eed319af3901592d00f70
* No need to specify a custom dependency in our setup.py.<commit_after> | from setuptools import find_packages, setup
setup(name='satnogsclient',
packages=find_packages(),
version='0.2.4',
author='SatNOGS team',
author_email='client-dev@satnogs.org',
url='https://github.com/satnogs/satnogs-client/',
description='SatNOGS Client',
install_requires=['APScheduler',
'SQLAlchemy',
'requests',
'validators',
'python-dateutil',
'ephem',
'pytz'],
scripts=['satnogsclient/bin/satnogs-poller',
'satnogsclient/bin/satnogs-task'])
| from setuptools import find_packages, setup
setup(name='satnogsclient',
packages=find_packages(),
version='0.2.4',
author='SatNOGS team',
author_email='client-dev@satnogs.org',
url='https://github.com/satnogs/satnogs-client/',
description='SatNOGS Client',
install_requires=['APScheduler',
'SQLAlchemy',
'requests',
'validators',
'python-dateutil',
'ephem',
'pytz'],
dependency_links=[
'git+https://github.com/brandon-rhodes/pyephem.git@47d0ba3616ee6c308f2eed319af3901592d00f70#egg=ephem'
],
scripts=['satnogsclient/bin/satnogs-poller',
'satnogsclient/bin/satnogs-task'])
Upgrade pyephem to latest version.
* Pyephem v3.7.6.0 now includes this required patch:
https://github.com/brandon-rhodes/pyephem/commit/47d0ba3616ee6c308f2eed319af3901592d00f70
* No need to specify a custom dependency in our setup.py.from setuptools import find_packages, setup
setup(name='satnogsclient',
packages=find_packages(),
version='0.2.4',
author='SatNOGS team',
author_email='client-dev@satnogs.org',
url='https://github.com/satnogs/satnogs-client/',
description='SatNOGS Client',
install_requires=['APScheduler',
'SQLAlchemy',
'requests',
'validators',
'python-dateutil',
'ephem',
'pytz'],
scripts=['satnogsclient/bin/satnogs-poller',
'satnogsclient/bin/satnogs-task'])
| <commit_before>from setuptools import find_packages, setup
setup(name='satnogsclient',
packages=find_packages(),
version='0.2.4',
author='SatNOGS team',
author_email='client-dev@satnogs.org',
url='https://github.com/satnogs/satnogs-client/',
description='SatNOGS Client',
install_requires=['APScheduler',
'SQLAlchemy',
'requests',
'validators',
'python-dateutil',
'ephem',
'pytz'],
dependency_links=[
'git+https://github.com/brandon-rhodes/pyephem.git@47d0ba3616ee6c308f2eed319af3901592d00f70#egg=ephem'
],
scripts=['satnogsclient/bin/satnogs-poller',
'satnogsclient/bin/satnogs-task'])
<commit_msg>Upgrade pyephem to latest version.
* Pyephem v3.7.6.0 now includes this required patch:
https://github.com/brandon-rhodes/pyephem/commit/47d0ba3616ee6c308f2eed319af3901592d00f70
* No need to specify a custom dependency in our setup.py.<commit_after>from setuptools import find_packages, setup
setup(name='satnogsclient',
packages=find_packages(),
version='0.2.4',
author='SatNOGS team',
author_email='client-dev@satnogs.org',
url='https://github.com/satnogs/satnogs-client/',
description='SatNOGS Client',
install_requires=['APScheduler',
'SQLAlchemy',
'requests',
'validators',
'python-dateutil',
'ephem',
'pytz'],
scripts=['satnogsclient/bin/satnogs-poller',
'satnogsclient/bin/satnogs-task'])
|
b86612bdefb58dd7f1f930cfb756dcd16f77c770 | setup.py | setup.py | from setuptools import setup
def readme():
with open("README.rst", 'r') as f:
return f.read()
setup(
name="asciigraf",
version="0.3.0",
packages=["asciigraf"],
description="A python library for making ascii-art into network graphs.",
classifiers=[
"Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Topic :: Software Development :: Libraries :: Python Modules",
],
long_description=readme(),
author="Opus One Solutions",
author_email="rnd@opusonesolutions.com",
url="https://github.com/AnjoMan/asciigraf",
keywords=["graph", "network", "testing", "parser"],
license="MIT",
install_requires=[
'networkx==1.11',
],
extras_require={
"test": ["pytest", "pytest-cov"],
},
)
| from setuptools import setup
def readme():
with open("README.rst", 'r') as f:
return f.read()
setup(
name="asciigraf",
version="0.3.0",
packages=["asciigraf"],
description="A python library for making ascii-art into network graphs.",
classifiers=[
"Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Topic :: Software Development :: Libraries :: Python Modules",
],
long_description=readme(),
author="Opus One Solutions",
author_email="rnd@opusonesolutions.com",
url="https://github.com/opusonesolutions/asciigraf",
keywords=["graph", "network", "testing", "parser"],
license="MIT",
install_requires=[
'networkx==1.11',
],
extras_require={
"test": ["pytest", "pytest-cov"],
},
)
| Change package url to point to opusonesolutions group | Change package url to point to opusonesolutions group | Python | mit | AnjoMan/asciigraf | from setuptools import setup
def readme():
with open("README.rst", 'r') as f:
return f.read()
setup(
name="asciigraf",
version="0.3.0",
packages=["asciigraf"],
description="A python library for making ascii-art into network graphs.",
classifiers=[
"Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Topic :: Software Development :: Libraries :: Python Modules",
],
long_description=readme(),
author="Opus One Solutions",
author_email="rnd@opusonesolutions.com",
url="https://github.com/AnjoMan/asciigraf",
keywords=["graph", "network", "testing", "parser"],
license="MIT",
install_requires=[
'networkx==1.11',
],
extras_require={
"test": ["pytest", "pytest-cov"],
},
)
Change package url to point to opusonesolutions group | from setuptools import setup
def readme():
with open("README.rst", 'r') as f:
return f.read()
setup(
name="asciigraf",
version="0.3.0",
packages=["asciigraf"],
description="A python library for making ascii-art into network graphs.",
classifiers=[
"Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Topic :: Software Development :: Libraries :: Python Modules",
],
long_description=readme(),
author="Opus One Solutions",
author_email="rnd@opusonesolutions.com",
url="https://github.com/opusonesolutions/asciigraf",
keywords=["graph", "network", "testing", "parser"],
license="MIT",
install_requires=[
'networkx==1.11',
],
extras_require={
"test": ["pytest", "pytest-cov"],
},
)
| <commit_before>from setuptools import setup
def readme():
with open("README.rst", 'r') as f:
return f.read()
setup(
name="asciigraf",
version="0.3.0",
packages=["asciigraf"],
description="A python library for making ascii-art into network graphs.",
classifiers=[
"Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Topic :: Software Development :: Libraries :: Python Modules",
],
long_description=readme(),
author="Opus One Solutions",
author_email="rnd@opusonesolutions.com",
url="https://github.com/AnjoMan/asciigraf",
keywords=["graph", "network", "testing", "parser"],
license="MIT",
install_requires=[
'networkx==1.11',
],
extras_require={
"test": ["pytest", "pytest-cov"],
},
)
<commit_msg>Change package url to point to opusonesolutions group<commit_after> | from setuptools import setup
def readme():
with open("README.rst", 'r') as f:
return f.read()
setup(
name="asciigraf",
version="0.3.0",
packages=["asciigraf"],
description="A python library for making ascii-art into network graphs.",
classifiers=[
"Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Topic :: Software Development :: Libraries :: Python Modules",
],
long_description=readme(),
author="Opus One Solutions",
author_email="rnd@opusonesolutions.com",
url="https://github.com/opusonesolutions/asciigraf",
keywords=["graph", "network", "testing", "parser"],
license="MIT",
install_requires=[
'networkx==1.11',
],
extras_require={
"test": ["pytest", "pytest-cov"],
},
)
| from setuptools import setup
def readme():
with open("README.rst", 'r') as f:
return f.read()
setup(
name="asciigraf",
version="0.3.0",
packages=["asciigraf"],
description="A python library for making ascii-art into network graphs.",
classifiers=[
"Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Topic :: Software Development :: Libraries :: Python Modules",
],
long_description=readme(),
author="Opus One Solutions",
author_email="rnd@opusonesolutions.com",
url="https://github.com/AnjoMan/asciigraf",
keywords=["graph", "network", "testing", "parser"],
license="MIT",
install_requires=[
'networkx==1.11',
],
extras_require={
"test": ["pytest", "pytest-cov"],
},
)
Change package url to point to opusonesolutions groupfrom setuptools import setup
def readme():
with open("README.rst", 'r') as f:
return f.read()
setup(
name="asciigraf",
version="0.3.0",
packages=["asciigraf"],
description="A python library for making ascii-art into network graphs.",
classifiers=[
"Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Topic :: Software Development :: Libraries :: Python Modules",
],
long_description=readme(),
author="Opus One Solutions",
author_email="rnd@opusonesolutions.com",
url="https://github.com/opusonesolutions/asciigraf",
keywords=["graph", "network", "testing", "parser"],
license="MIT",
install_requires=[
'networkx==1.11',
],
extras_require={
"test": ["pytest", "pytest-cov"],
},
)
| <commit_before>from setuptools import setup
def readme():
with open("README.rst", 'r') as f:
return f.read()
setup(
name="asciigraf",
version="0.3.0",
packages=["asciigraf"],
description="A python library for making ascii-art into network graphs.",
classifiers=[
"Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Topic :: Software Development :: Libraries :: Python Modules",
],
long_description=readme(),
author="Opus One Solutions",
author_email="rnd@opusonesolutions.com",
url="https://github.com/AnjoMan/asciigraf",
keywords=["graph", "network", "testing", "parser"],
license="MIT",
install_requires=[
'networkx==1.11',
],
extras_require={
"test": ["pytest", "pytest-cov"],
},
)
<commit_msg>Change package url to point to opusonesolutions group<commit_after>from setuptools import setup
def readme():
with open("README.rst", 'r') as f:
return f.read()
setup(
name="asciigraf",
version="0.3.0",
packages=["asciigraf"],
description="A python library for making ascii-art into network graphs.",
classifiers=[
"Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Topic :: Software Development :: Libraries :: Python Modules",
],
long_description=readme(),
author="Opus One Solutions",
author_email="rnd@opusonesolutions.com",
url="https://github.com/opusonesolutions/asciigraf",
keywords=["graph", "network", "testing", "parser"],
license="MIT",
install_requires=[
'networkx==1.11',
],
extras_require={
"test": ["pytest", "pytest-cov"],
},
)
|
d616642e11c0151f44cdae6038d8cdae07abdf8c | setup.py | setup.py | from distutils.core import setup
setup(
name='Getty Art',
version='0.0.1',
author='Clemens Wolff',
author_email='clemens.wolff+pypi@gmail.com',
packages=['getty_art'],
url='https://github.com/c-w/GettyArt',
download_url='http://pypi.python.org/pypi/GettyArt',
license='LICENSE.txt',
description='Scraper for art available from getty.edu',
long_description=open('README.rst').read(),
)
| from distutils.core import setup
setup(
name='Getty Art',
version='0.0.1',
author='Clemens Wolff',
author_email='clemens.wolff+pypi@gmail.com',
packages=['getty_art'],
url='https://github.com/c-w/GettyArt',
download_url='http://pypi.python.org/pypi/GettyArt',
license='LICENSE.txt',
description='Scraper for the fine art on getty.edu',
long_description=open('README.rst').read(),
)
| Make tag-line consistent with GitHub | Make tag-line consistent with GitHub
| Python | mit | c-w/GettyArt | from distutils.core import setup
setup(
name='Getty Art',
version='0.0.1',
author='Clemens Wolff',
author_email='clemens.wolff+pypi@gmail.com',
packages=['getty_art'],
url='https://github.com/c-w/GettyArt',
download_url='http://pypi.python.org/pypi/GettyArt',
license='LICENSE.txt',
description='Scraper for art available from getty.edu',
long_description=open('README.rst').read(),
)
Make tag-line consistent with GitHub | from distutils.core import setup
setup(
name='Getty Art',
version='0.0.1',
author='Clemens Wolff',
author_email='clemens.wolff+pypi@gmail.com',
packages=['getty_art'],
url='https://github.com/c-w/GettyArt',
download_url='http://pypi.python.org/pypi/GettyArt',
license='LICENSE.txt',
description='Scraper for the fine art on getty.edu',
long_description=open('README.rst').read(),
)
| <commit_before>from distutils.core import setup
setup(
name='Getty Art',
version='0.0.1',
author='Clemens Wolff',
author_email='clemens.wolff+pypi@gmail.com',
packages=['getty_art'],
url='https://github.com/c-w/GettyArt',
download_url='http://pypi.python.org/pypi/GettyArt',
license='LICENSE.txt',
description='Scraper for art available from getty.edu',
long_description=open('README.rst').read(),
)
<commit_msg>Make tag-line consistent with GitHub<commit_after> | from distutils.core import setup
setup(
name='Getty Art',
version='0.0.1',
author='Clemens Wolff',
author_email='clemens.wolff+pypi@gmail.com',
packages=['getty_art'],
url='https://github.com/c-w/GettyArt',
download_url='http://pypi.python.org/pypi/GettyArt',
license='LICENSE.txt',
description='Scraper for the fine art on getty.edu',
long_description=open('README.rst').read(),
)
| from distutils.core import setup
setup(
name='Getty Art',
version='0.0.1',
author='Clemens Wolff',
author_email='clemens.wolff+pypi@gmail.com',
packages=['getty_art'],
url='https://github.com/c-w/GettyArt',
download_url='http://pypi.python.org/pypi/GettyArt',
license='LICENSE.txt',
description='Scraper for art available from getty.edu',
long_description=open('README.rst').read(),
)
Make tag-line consistent with GitHubfrom distutils.core import setup
setup(
name='Getty Art',
version='0.0.1',
author='Clemens Wolff',
author_email='clemens.wolff+pypi@gmail.com',
packages=['getty_art'],
url='https://github.com/c-w/GettyArt',
download_url='http://pypi.python.org/pypi/GettyArt',
license='LICENSE.txt',
description='Scraper for the fine art on getty.edu',
long_description=open('README.rst').read(),
)
| <commit_before>from distutils.core import setup
setup(
name='Getty Art',
version='0.0.1',
author='Clemens Wolff',
author_email='clemens.wolff+pypi@gmail.com',
packages=['getty_art'],
url='https://github.com/c-w/GettyArt',
download_url='http://pypi.python.org/pypi/GettyArt',
license='LICENSE.txt',
description='Scraper for art available from getty.edu',
long_description=open('README.rst').read(),
)
<commit_msg>Make tag-line consistent with GitHub<commit_after>from distutils.core import setup
setup(
name='Getty Art',
version='0.0.1',
author='Clemens Wolff',
author_email='clemens.wolff+pypi@gmail.com',
packages=['getty_art'],
url='https://github.com/c-w/GettyArt',
download_url='http://pypi.python.org/pypi/GettyArt',
license='LICENSE.txt',
description='Scraper for the fine art on getty.edu',
long_description=open('README.rst').read(),
)
|
ec25b097520930b19a8ae63a1bab23efcc3ba752 | setup.py | setup.py | #
# Copyright 2017 Google Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import setuptools
def readme():
with open('docs/README.md') as f:
return f.read()
setuptools.setup(
name='pstar',
description='pstar: better python collections',
long_description=readme(),
version='0.1.0',
url='https://github.com/iansf/pstar',
download_url='https://github.com/iansf/pstar/archive/0.1.0.tar.gz',
author='Ian Fischer, Google',
author_email='iansf@google.com',
packages=['pstar'],
license='Apache 2.0',
install_requires=[],
test_suite='nose.collector',
tests_require=['nose'],
)
| #
# Copyright 2017 Google Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import setuptools
def readme():
return '`pstar` documentation and source code can be found at https://github.com/iansf/pstar.'
def version():
return '0.1.1'
setuptools.setup(
name='pstar',
description='pstar: numpy for arbitrary data',
long_description=readme(),
version=version(),
url='https://github.com/iansf/pstar',
download_url='https://github.com/iansf/pstar/archive/%s.tar.gz' % version(),
author='Ian Fischer, Google',
author_email='iansf@google.com',
packages=['pstar'],
license='Apache 2.0',
install_requires=[],
test_suite='nose.collector',
tests_require=['nose'],
)
| Fix bug with pip install. Update version to 0.1.1. | Fix bug with pip install. Update version to 0.1.1. | Python | apache-2.0 | iansf/pstar | #
# Copyright 2017 Google Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import setuptools
def readme():
with open('docs/README.md') as f:
return f.read()
setuptools.setup(
name='pstar',
description='pstar: better python collections',
long_description=readme(),
version='0.1.0',
url='https://github.com/iansf/pstar',
download_url='https://github.com/iansf/pstar/archive/0.1.0.tar.gz',
author='Ian Fischer, Google',
author_email='iansf@google.com',
packages=['pstar'],
license='Apache 2.0',
install_requires=[],
test_suite='nose.collector',
tests_require=['nose'],
)
Fix bug with pip install. Update version to 0.1.1. | #
# Copyright 2017 Google Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import setuptools
def readme():
return '`pstar` documentation and source code can be found at https://github.com/iansf/pstar.'
def version():
return '0.1.1'
setuptools.setup(
name='pstar',
description='pstar: numpy for arbitrary data',
long_description=readme(),
version=version(),
url='https://github.com/iansf/pstar',
download_url='https://github.com/iansf/pstar/archive/%s.tar.gz' % version(),
author='Ian Fischer, Google',
author_email='iansf@google.com',
packages=['pstar'],
license='Apache 2.0',
install_requires=[],
test_suite='nose.collector',
tests_require=['nose'],
)
| <commit_before>#
# Copyright 2017 Google Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import setuptools
def readme():
with open('docs/README.md') as f:
return f.read()
setuptools.setup(
name='pstar',
description='pstar: better python collections',
long_description=readme(),
version='0.1.0',
url='https://github.com/iansf/pstar',
download_url='https://github.com/iansf/pstar/archive/0.1.0.tar.gz',
author='Ian Fischer, Google',
author_email='iansf@google.com',
packages=['pstar'],
license='Apache 2.0',
install_requires=[],
test_suite='nose.collector',
tests_require=['nose'],
)
<commit_msg>Fix bug with pip install. Update version to 0.1.1.<commit_after> | #
# Copyright 2017 Google Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import setuptools
def readme():
return '`pstar` documentation and source code can be found at https://github.com/iansf/pstar.'
def version():
return '0.1.1'
setuptools.setup(
name='pstar',
description='pstar: numpy for arbitrary data',
long_description=readme(),
version=version(),
url='https://github.com/iansf/pstar',
download_url='https://github.com/iansf/pstar/archive/%s.tar.gz' % version(),
author='Ian Fischer, Google',
author_email='iansf@google.com',
packages=['pstar'],
license='Apache 2.0',
install_requires=[],
test_suite='nose.collector',
tests_require=['nose'],
)
| #
# Copyright 2017 Google Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import setuptools
def readme():
with open('docs/README.md') as f:
return f.read()
setuptools.setup(
name='pstar',
description='pstar: better python collections',
long_description=readme(),
version='0.1.0',
url='https://github.com/iansf/pstar',
download_url='https://github.com/iansf/pstar/archive/0.1.0.tar.gz',
author='Ian Fischer, Google',
author_email='iansf@google.com',
packages=['pstar'],
license='Apache 2.0',
install_requires=[],
test_suite='nose.collector',
tests_require=['nose'],
)
Fix bug with pip install. Update version to 0.1.1.#
# Copyright 2017 Google Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import setuptools
def readme():
return '`pstar` documentation and source code can be found at https://github.com/iansf/pstar.'
def version():
return '0.1.1'
setuptools.setup(
name='pstar',
description='pstar: numpy for arbitrary data',
long_description=readme(),
version=version(),
url='https://github.com/iansf/pstar',
download_url='https://github.com/iansf/pstar/archive/%s.tar.gz' % version(),
author='Ian Fischer, Google',
author_email='iansf@google.com',
packages=['pstar'],
license='Apache 2.0',
install_requires=[],
test_suite='nose.collector',
tests_require=['nose'],
)
| <commit_before>#
# Copyright 2017 Google Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import setuptools
def readme():
with open('docs/README.md') as f:
return f.read()
setuptools.setup(
name='pstar',
description='pstar: better python collections',
long_description=readme(),
version='0.1.0',
url='https://github.com/iansf/pstar',
download_url='https://github.com/iansf/pstar/archive/0.1.0.tar.gz',
author='Ian Fischer, Google',
author_email='iansf@google.com',
packages=['pstar'],
license='Apache 2.0',
install_requires=[],
test_suite='nose.collector',
tests_require=['nose'],
)
<commit_msg>Fix bug with pip install. Update version to 0.1.1.<commit_after>#
# Copyright 2017 Google Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import setuptools
def readme():
return '`pstar` documentation and source code can be found at https://github.com/iansf/pstar.'
def version():
return '0.1.1'
setuptools.setup(
name='pstar',
description='pstar: numpy for arbitrary data',
long_description=readme(),
version=version(),
url='https://github.com/iansf/pstar',
download_url='https://github.com/iansf/pstar/archive/%s.tar.gz' % version(),
author='Ian Fischer, Google',
author_email='iansf@google.com',
packages=['pstar'],
license='Apache 2.0',
install_requires=[],
test_suite='nose.collector',
tests_require=['nose'],
)
|
63349f4a73c8cb99f495cc15621ecf946e9e652e | setup.py | setup.py | # -*- coding: utf-8 -*-
VERSION = '0.2'
from setuptools import setup
setup(
name='nutshell',
packages=["nutshell"],
version=VERSION,
description='A minimal python library to access Nutshell CRM:s JSON-RPC API.',
author=u'Emil Stenström',
author_email='em@kth.se',
url='https://github.com/EmilStenstrom/python-nutshell',
download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/' + VERSION,
install_requires=["requests>=2.9.1", "six>=1.10.0"],
tests_require=["mock>=1.0.1", "nose>=1.3.7"],
test_suite="nose.collector",
keywords=['nutshell', 'nutshell-crm', 'json-rpc'],
classifiers=[],
)
| # -*- coding: utf-8 -*-
VERSION = '0.2'
from setuptools import setup
setup(
name='nutshell',
packages=["nutshell"],
version=VERSION,
description='A minimal python library to access Nutshell CRM:s JSON-RPC API.',
author=u'Emil Stenström',
author_email='em@kth.se',
url='https://github.com/EmilStenstrom/python-nutshell',
download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/' + VERSION,
install_requires=["requests>=2.9.1", "six>=1.10.0"],
tests_require=["mock>=1.0.1", "nose>=1.3.7"],
test_suite="nose.collector",
keywords=['nutshell', 'nutshell-crm', 'json-rpc'],
classifiers=[
"Programming Language :: Python",
"Programming Language :: Python :: 2.6",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.3",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Operating System :: OS Independent",
])
| Add proper Python version classifiers. | Add proper Python version classifiers.
| Python | mit | EmilStenstrom/python-nutshell | # -*- coding: utf-8 -*-
VERSION = '0.2'
from setuptools import setup
setup(
name='nutshell',
packages=["nutshell"],
version=VERSION,
description='A minimal python library to access Nutshell CRM:s JSON-RPC API.',
author=u'Emil Stenström',
author_email='em@kth.se',
url='https://github.com/EmilStenstrom/python-nutshell',
download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/' + VERSION,
install_requires=["requests>=2.9.1", "six>=1.10.0"],
tests_require=["mock>=1.0.1", "nose>=1.3.7"],
test_suite="nose.collector",
keywords=['nutshell', 'nutshell-crm', 'json-rpc'],
classifiers=[],
)
Add proper Python version classifiers. | # -*- coding: utf-8 -*-
VERSION = '0.2'
from setuptools import setup
setup(
name='nutshell',
packages=["nutshell"],
version=VERSION,
description='A minimal python library to access Nutshell CRM:s JSON-RPC API.',
author=u'Emil Stenström',
author_email='em@kth.se',
url='https://github.com/EmilStenstrom/python-nutshell',
download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/' + VERSION,
install_requires=["requests>=2.9.1", "six>=1.10.0"],
tests_require=["mock>=1.0.1", "nose>=1.3.7"],
test_suite="nose.collector",
keywords=['nutshell', 'nutshell-crm', 'json-rpc'],
classifiers=[
"Programming Language :: Python",
"Programming Language :: Python :: 2.6",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.3",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Operating System :: OS Independent",
])
| <commit_before># -*- coding: utf-8 -*-
VERSION = '0.2'
from setuptools import setup
setup(
name='nutshell',
packages=["nutshell"],
version=VERSION,
description='A minimal python library to access Nutshell CRM:s JSON-RPC API.',
author=u'Emil Stenström',
author_email='em@kth.se',
url='https://github.com/EmilStenstrom/python-nutshell',
download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/' + VERSION,
install_requires=["requests>=2.9.1", "six>=1.10.0"],
tests_require=["mock>=1.0.1", "nose>=1.3.7"],
test_suite="nose.collector",
keywords=['nutshell', 'nutshell-crm', 'json-rpc'],
classifiers=[],
)
<commit_msg>Add proper Python version classifiers.<commit_after> | # -*- coding: utf-8 -*-
VERSION = '0.2'
from setuptools import setup
setup(
name='nutshell',
packages=["nutshell"],
version=VERSION,
description='A minimal python library to access Nutshell CRM:s JSON-RPC API.',
author=u'Emil Stenström',
author_email='em@kth.se',
url='https://github.com/EmilStenstrom/python-nutshell',
download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/' + VERSION,
install_requires=["requests>=2.9.1", "six>=1.10.0"],
tests_require=["mock>=1.0.1", "nose>=1.3.7"],
test_suite="nose.collector",
keywords=['nutshell', 'nutshell-crm', 'json-rpc'],
classifiers=[
"Programming Language :: Python",
"Programming Language :: Python :: 2.6",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.3",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Operating System :: OS Independent",
])
| # -*- coding: utf-8 -*-
VERSION = '0.2'
from setuptools import setup
setup(
name='nutshell',
packages=["nutshell"],
version=VERSION,
description='A minimal python library to access Nutshell CRM:s JSON-RPC API.',
author=u'Emil Stenström',
author_email='em@kth.se',
url='https://github.com/EmilStenstrom/python-nutshell',
download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/' + VERSION,
install_requires=["requests>=2.9.1", "six>=1.10.0"],
tests_require=["mock>=1.0.1", "nose>=1.3.7"],
test_suite="nose.collector",
keywords=['nutshell', 'nutshell-crm', 'json-rpc'],
classifiers=[],
)
Add proper Python version classifiers.# -*- coding: utf-8 -*-
VERSION = '0.2'
from setuptools import setup
setup(
name='nutshell',
packages=["nutshell"],
version=VERSION,
description='A minimal python library to access Nutshell CRM:s JSON-RPC API.',
author=u'Emil Stenström',
author_email='em@kth.se',
url='https://github.com/EmilStenstrom/python-nutshell',
download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/' + VERSION,
install_requires=["requests>=2.9.1", "six>=1.10.0"],
tests_require=["mock>=1.0.1", "nose>=1.3.7"],
test_suite="nose.collector",
keywords=['nutshell', 'nutshell-crm', 'json-rpc'],
classifiers=[
"Programming Language :: Python",
"Programming Language :: Python :: 2.6",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.3",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Operating System :: OS Independent",
])
| <commit_before># -*- coding: utf-8 -*-
VERSION = '0.2'
from setuptools import setup
setup(
name='nutshell',
packages=["nutshell"],
version=VERSION,
description='A minimal python library to access Nutshell CRM:s JSON-RPC API.',
author=u'Emil Stenström',
author_email='em@kth.se',
url='https://github.com/EmilStenstrom/python-nutshell',
download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/' + VERSION,
install_requires=["requests>=2.9.1", "six>=1.10.0"],
tests_require=["mock>=1.0.1", "nose>=1.3.7"],
test_suite="nose.collector",
keywords=['nutshell', 'nutshell-crm', 'json-rpc'],
classifiers=[],
)
<commit_msg>Add proper Python version classifiers.<commit_after># -*- coding: utf-8 -*-
VERSION = '0.2'
from setuptools import setup
setup(
name='nutshell',
packages=["nutshell"],
version=VERSION,
description='A minimal python library to access Nutshell CRM:s JSON-RPC API.',
author=u'Emil Stenström',
author_email='em@kth.se',
url='https://github.com/EmilStenstrom/python-nutshell',
download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/' + VERSION,
install_requires=["requests>=2.9.1", "six>=1.10.0"],
tests_require=["mock>=1.0.1", "nose>=1.3.7"],
test_suite="nose.collector",
keywords=['nutshell', 'nutshell-crm', 'json-rpc'],
classifiers=[
"Programming Language :: Python",
"Programming Language :: Python :: 2.6",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.3",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Operating System :: OS Independent",
])
|
d363631106c95365e138bea08ca0f8811c1dba67 | setup.py | setup.py | #!/usr/bin/env python
from setuptools import setup, find_packages
setup(
name='django-glitter-news',
version='0.1',
description='Django Glitter News for Django',
long_description=open('README.rst').read(),
url='https://github.com/blancltd/django-glitter-news',
maintainer='Blanc Ltd',
maintainer_email='studio@blanc.ltd.uk',
platforms=['any'],
packages=find_packages(),
install_requires=[
'django-glitter',
],
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
],
license='BSD 3-Clause',
)
| #!/usr/bin/env python
from setuptools import setup, find_packages
setup(
name='django-glitter-news',
version='0.1',
description='Django Glitter News for Django',
long_description=open('README.rst').read(),
url='https://github.com/blancltd/django-glitter-news',
maintainer='Blanc Ltd',
maintainer_email='studio@blanc.ltd.uk',
platforms=['any'],
packages=find_packages(),
install_requires=[
'django-glitter',
],
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
license='BSD 3-Clause',
)
| Update python versions we care about | Update python versions we care about
| Python | bsd-2-clause | blancltd/glitter-news | #!/usr/bin/env python
from setuptools import setup, find_packages
setup(
name='django-glitter-news',
version='0.1',
description='Django Glitter News for Django',
long_description=open('README.rst').read(),
url='https://github.com/blancltd/django-glitter-news',
maintainer='Blanc Ltd',
maintainer_email='studio@blanc.ltd.uk',
platforms=['any'],
packages=find_packages(),
install_requires=[
'django-glitter',
],
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
],
license='BSD 3-Clause',
)
Update python versions we care about | #!/usr/bin/env python
from setuptools import setup, find_packages
setup(
name='django-glitter-news',
version='0.1',
description='Django Glitter News for Django',
long_description=open('README.rst').read(),
url='https://github.com/blancltd/django-glitter-news',
maintainer='Blanc Ltd',
maintainer_email='studio@blanc.ltd.uk',
platforms=['any'],
packages=find_packages(),
install_requires=[
'django-glitter',
],
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
license='BSD 3-Clause',
)
| <commit_before>#!/usr/bin/env python
from setuptools import setup, find_packages
setup(
name='django-glitter-news',
version='0.1',
description='Django Glitter News for Django',
long_description=open('README.rst').read(),
url='https://github.com/blancltd/django-glitter-news',
maintainer='Blanc Ltd',
maintainer_email='studio@blanc.ltd.uk',
platforms=['any'],
packages=find_packages(),
install_requires=[
'django-glitter',
],
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
],
license='BSD 3-Clause',
)
<commit_msg>Update python versions we care about<commit_after> | #!/usr/bin/env python
from setuptools import setup, find_packages
setup(
name='django-glitter-news',
version='0.1',
description='Django Glitter News for Django',
long_description=open('README.rst').read(),
url='https://github.com/blancltd/django-glitter-news',
maintainer='Blanc Ltd',
maintainer_email='studio@blanc.ltd.uk',
platforms=['any'],
packages=find_packages(),
install_requires=[
'django-glitter',
],
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
license='BSD 3-Clause',
)
| #!/usr/bin/env python
from setuptools import setup, find_packages
setup(
name='django-glitter-news',
version='0.1',
description='Django Glitter News for Django',
long_description=open('README.rst').read(),
url='https://github.com/blancltd/django-glitter-news',
maintainer='Blanc Ltd',
maintainer_email='studio@blanc.ltd.uk',
platforms=['any'],
packages=find_packages(),
install_requires=[
'django-glitter',
],
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
],
license='BSD 3-Clause',
)
Update python versions we care about#!/usr/bin/env python
from setuptools import setup, find_packages
setup(
name='django-glitter-news',
version='0.1',
description='Django Glitter News for Django',
long_description=open('README.rst').read(),
url='https://github.com/blancltd/django-glitter-news',
maintainer='Blanc Ltd',
maintainer_email='studio@blanc.ltd.uk',
platforms=['any'],
packages=find_packages(),
install_requires=[
'django-glitter',
],
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
license='BSD 3-Clause',
)
| <commit_before>#!/usr/bin/env python
from setuptools import setup, find_packages
setup(
name='django-glitter-news',
version='0.1',
description='Django Glitter News for Django',
long_description=open('README.rst').read(),
url='https://github.com/blancltd/django-glitter-news',
maintainer='Blanc Ltd',
maintainer_email='studio@blanc.ltd.uk',
platforms=['any'],
packages=find_packages(),
install_requires=[
'django-glitter',
],
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
],
license='BSD 3-Clause',
)
<commit_msg>Update python versions we care about<commit_after>#!/usr/bin/env python
from setuptools import setup, find_packages
setup(
name='django-glitter-news',
version='0.1',
description='Django Glitter News for Django',
long_description=open('README.rst').read(),
url='https://github.com/blancltd/django-glitter-news',
maintainer='Blanc Ltd',
maintainer_email='studio@blanc.ltd.uk',
platforms=['any'],
packages=find_packages(),
install_requires=[
'django-glitter',
],
classifiers=[
'Environment :: Web Environment',
'Framework :: Django',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
],
license='BSD 3-Clause',
)
|
34e04331fae60e63d0bc0daf47161bc8507835b8 | setup.py | setup.py | #!/usr/bin/env python
import os
import sys
from serfclient import __version__
try:
from setuptools import setup
from setuptools.command.test import test as TestCommand
class PyTest(TestCommand):
def finalize_options(self):
TestCommand.finalize_options(self)
self.test_args = []
self.test_suite = True
def run_tests(self):
import pytest
errno = pytest.main(self.test_args)
sys.exit(errno)
except ImportError:
from distutils.core import setup
PyTest = lambda x: x
try:
long_description = open(os.path.join(os.path.dirname(__file__),
'README.rst')).read()
except:
long_description = None
setup(
name='serfclient',
version=__version__,
description='Python client for the Serf orchestration tool',
long_description=long_description,
url='https://github.com/KushalP/serfclient-py',
author='Kushal Pisavadia',
author_email='kushal@violentlymild.com',
maintainer='Kushal Pisavadia',
maintainer_email='kushal@violentlymild.com',
keywords=['Serf', 'orchestration', 'service discovery'],
license='MIT',
packages=['serfclient'],
install_requires=['msgpack-python >= 0.4.0'],
tests_require=['pytest >= 2.5.2', 'pytest-cov >= 1.6'],
cmdclass={'test': PyTest}
)
| #!/usr/bin/env python
import os
import sys
from serfclient import __version__
try:
from setuptools import setup
from setuptools.command.test import test as TestCommand
class PyTest(TestCommand):
def finalize_options(self):
TestCommand.finalize_options(self)
self.test_args = []
self.test_suite = True
def run_tests(self):
import pytest
errno = pytest.main(self.test_args)
sys.exit(errno)
except ImportError:
from distutils.core import setup
PyTest = lambda x: x
try:
long_description = open(os.path.join(os.path.dirname(__file__),
'README.rst')).read()
except:
long_description = None
setup(
name='serfclient',
version=__version__,
description='Python client for the Serf orchestration tool',
long_description=long_description,
url='https://github.com/KushalP/serfclient-py',
author='Kushal Pisavadia',
author_email='kushal@violentlymild.com',
maintainer='Kushal Pisavadia',
maintainer_email='kushal@violentlymild.com',
keywords=['Serf', 'orchestration', 'service discovery'],
license='MIT',
packages=['serfclient'],
install_requires=['msgpack-python >= 0.4.0'],
tests_require=['pytest >= 2.5.2',
'pytest-cov >= 1.6',
'python-coveralls >= 2.4.2'],
cmdclass={'test': PyTest}
)
| Add python-coveralls as a test dependency | Add python-coveralls as a test dependency
This is so that we can push our coverage stats to coverage.io.
| Python | mit | charleswhchan/serfclient-py,KushalP/serfclient-py | #!/usr/bin/env python
import os
import sys
from serfclient import __version__
try:
from setuptools import setup
from setuptools.command.test import test as TestCommand
class PyTest(TestCommand):
def finalize_options(self):
TestCommand.finalize_options(self)
self.test_args = []
self.test_suite = True
def run_tests(self):
import pytest
errno = pytest.main(self.test_args)
sys.exit(errno)
except ImportError:
from distutils.core import setup
PyTest = lambda x: x
try:
long_description = open(os.path.join(os.path.dirname(__file__),
'README.rst')).read()
except:
long_description = None
setup(
name='serfclient',
version=__version__,
description='Python client for the Serf orchestration tool',
long_description=long_description,
url='https://github.com/KushalP/serfclient-py',
author='Kushal Pisavadia',
author_email='kushal@violentlymild.com',
maintainer='Kushal Pisavadia',
maintainer_email='kushal@violentlymild.com',
keywords=['Serf', 'orchestration', 'service discovery'],
license='MIT',
packages=['serfclient'],
install_requires=['msgpack-python >= 0.4.0'],
tests_require=['pytest >= 2.5.2', 'pytest-cov >= 1.6'],
cmdclass={'test': PyTest}
)
Add python-coveralls as a test dependency
This is so that we can push our coverage stats to coverage.io. | #!/usr/bin/env python
import os
import sys
from serfclient import __version__
try:
from setuptools import setup
from setuptools.command.test import test as TestCommand
class PyTest(TestCommand):
def finalize_options(self):
TestCommand.finalize_options(self)
self.test_args = []
self.test_suite = True
def run_tests(self):
import pytest
errno = pytest.main(self.test_args)
sys.exit(errno)
except ImportError:
from distutils.core import setup
PyTest = lambda x: x
try:
long_description = open(os.path.join(os.path.dirname(__file__),
'README.rst')).read()
except:
long_description = None
setup(
name='serfclient',
version=__version__,
description='Python client for the Serf orchestration tool',
long_description=long_description,
url='https://github.com/KushalP/serfclient-py',
author='Kushal Pisavadia',
author_email='kushal@violentlymild.com',
maintainer='Kushal Pisavadia',
maintainer_email='kushal@violentlymild.com',
keywords=['Serf', 'orchestration', 'service discovery'],
license='MIT',
packages=['serfclient'],
install_requires=['msgpack-python >= 0.4.0'],
tests_require=['pytest >= 2.5.2',
'pytest-cov >= 1.6',
'python-coveralls >= 2.4.2'],
cmdclass={'test': PyTest}
)
| <commit_before>#!/usr/bin/env python
import os
import sys
from serfclient import __version__
try:
from setuptools import setup
from setuptools.command.test import test as TestCommand
class PyTest(TestCommand):
def finalize_options(self):
TestCommand.finalize_options(self)
self.test_args = []
self.test_suite = True
def run_tests(self):
import pytest
errno = pytest.main(self.test_args)
sys.exit(errno)
except ImportError:
from distutils.core import setup
PyTest = lambda x: x
try:
long_description = open(os.path.join(os.path.dirname(__file__),
'README.rst')).read()
except:
long_description = None
setup(
name='serfclient',
version=__version__,
description='Python client for the Serf orchestration tool',
long_description=long_description,
url='https://github.com/KushalP/serfclient-py',
author='Kushal Pisavadia',
author_email='kushal@violentlymild.com',
maintainer='Kushal Pisavadia',
maintainer_email='kushal@violentlymild.com',
keywords=['Serf', 'orchestration', 'service discovery'],
license='MIT',
packages=['serfclient'],
install_requires=['msgpack-python >= 0.4.0'],
tests_require=['pytest >= 2.5.2', 'pytest-cov >= 1.6'],
cmdclass={'test': PyTest}
)
<commit_msg>Add python-coveralls as a test dependency
This is so that we can push our coverage stats to coverage.io.<commit_after> | #!/usr/bin/env python
import os
import sys
from serfclient import __version__
try:
from setuptools import setup
from setuptools.command.test import test as TestCommand
class PyTest(TestCommand):
def finalize_options(self):
TestCommand.finalize_options(self)
self.test_args = []
self.test_suite = True
def run_tests(self):
import pytest
errno = pytest.main(self.test_args)
sys.exit(errno)
except ImportError:
from distutils.core import setup
PyTest = lambda x: x
try:
long_description = open(os.path.join(os.path.dirname(__file__),
'README.rst')).read()
except:
long_description = None
setup(
name='serfclient',
version=__version__,
description='Python client for the Serf orchestration tool',
long_description=long_description,
url='https://github.com/KushalP/serfclient-py',
author='Kushal Pisavadia',
author_email='kushal@violentlymild.com',
maintainer='Kushal Pisavadia',
maintainer_email='kushal@violentlymild.com',
keywords=['Serf', 'orchestration', 'service discovery'],
license='MIT',
packages=['serfclient'],
install_requires=['msgpack-python >= 0.4.0'],
tests_require=['pytest >= 2.5.2',
'pytest-cov >= 1.6',
'python-coveralls >= 2.4.2'],
cmdclass={'test': PyTest}
)
| #!/usr/bin/env python
import os
import sys
from serfclient import __version__
try:
from setuptools import setup
from setuptools.command.test import test as TestCommand
class PyTest(TestCommand):
def finalize_options(self):
TestCommand.finalize_options(self)
self.test_args = []
self.test_suite = True
def run_tests(self):
import pytest
errno = pytest.main(self.test_args)
sys.exit(errno)
except ImportError:
from distutils.core import setup
PyTest = lambda x: x
try:
long_description = open(os.path.join(os.path.dirname(__file__),
'README.rst')).read()
except:
long_description = None
setup(
name='serfclient',
version=__version__,
description='Python client for the Serf orchestration tool',
long_description=long_description,
url='https://github.com/KushalP/serfclient-py',
author='Kushal Pisavadia',
author_email='kushal@violentlymild.com',
maintainer='Kushal Pisavadia',
maintainer_email='kushal@violentlymild.com',
keywords=['Serf', 'orchestration', 'service discovery'],
license='MIT',
packages=['serfclient'],
install_requires=['msgpack-python >= 0.4.0'],
tests_require=['pytest >= 2.5.2', 'pytest-cov >= 1.6'],
cmdclass={'test': PyTest}
)
Add python-coveralls as a test dependency
This is so that we can push our coverage stats to coverage.io.#!/usr/bin/env python
import os
import sys
from serfclient import __version__
try:
from setuptools import setup
from setuptools.command.test import test as TestCommand
class PyTest(TestCommand):
def finalize_options(self):
TestCommand.finalize_options(self)
self.test_args = []
self.test_suite = True
def run_tests(self):
import pytest
errno = pytest.main(self.test_args)
sys.exit(errno)
except ImportError:
from distutils.core import setup
PyTest = lambda x: x
try:
long_description = open(os.path.join(os.path.dirname(__file__),
'README.rst')).read()
except:
long_description = None
setup(
name='serfclient',
version=__version__,
description='Python client for the Serf orchestration tool',
long_description=long_description,
url='https://github.com/KushalP/serfclient-py',
author='Kushal Pisavadia',
author_email='kushal@violentlymild.com',
maintainer='Kushal Pisavadia',
maintainer_email='kushal@violentlymild.com',
keywords=['Serf', 'orchestration', 'service discovery'],
license='MIT',
packages=['serfclient'],
install_requires=['msgpack-python >= 0.4.0'],
tests_require=['pytest >= 2.5.2',
'pytest-cov >= 1.6',
'python-coveralls >= 2.4.2'],
cmdclass={'test': PyTest}
)
| <commit_before>#!/usr/bin/env python
import os
import sys
from serfclient import __version__
try:
from setuptools import setup
from setuptools.command.test import test as TestCommand
class PyTest(TestCommand):
def finalize_options(self):
TestCommand.finalize_options(self)
self.test_args = []
self.test_suite = True
def run_tests(self):
import pytest
errno = pytest.main(self.test_args)
sys.exit(errno)
except ImportError:
from distutils.core import setup
PyTest = lambda x: x
try:
long_description = open(os.path.join(os.path.dirname(__file__),
'README.rst')).read()
except:
long_description = None
setup(
name='serfclient',
version=__version__,
description='Python client for the Serf orchestration tool',
long_description=long_description,
url='https://github.com/KushalP/serfclient-py',
author='Kushal Pisavadia',
author_email='kushal@violentlymild.com',
maintainer='Kushal Pisavadia',
maintainer_email='kushal@violentlymild.com',
keywords=['Serf', 'orchestration', 'service discovery'],
license='MIT',
packages=['serfclient'],
install_requires=['msgpack-python >= 0.4.0'],
tests_require=['pytest >= 2.5.2', 'pytest-cov >= 1.6'],
cmdclass={'test': PyTest}
)
<commit_msg>Add python-coveralls as a test dependency
This is so that we can push our coverage stats to coverage.io.<commit_after>#!/usr/bin/env python
import os
import sys
from serfclient import __version__
try:
from setuptools import setup
from setuptools.command.test import test as TestCommand
class PyTest(TestCommand):
def finalize_options(self):
TestCommand.finalize_options(self)
self.test_args = []
self.test_suite = True
def run_tests(self):
import pytest
errno = pytest.main(self.test_args)
sys.exit(errno)
except ImportError:
from distutils.core import setup
PyTest = lambda x: x
try:
long_description = open(os.path.join(os.path.dirname(__file__),
'README.rst')).read()
except:
long_description = None
setup(
name='serfclient',
version=__version__,
description='Python client for the Serf orchestration tool',
long_description=long_description,
url='https://github.com/KushalP/serfclient-py',
author='Kushal Pisavadia',
author_email='kushal@violentlymild.com',
maintainer='Kushal Pisavadia',
maintainer_email='kushal@violentlymild.com',
keywords=['Serf', 'orchestration', 'service discovery'],
license='MIT',
packages=['serfclient'],
install_requires=['msgpack-python >= 0.4.0'],
tests_require=['pytest >= 2.5.2',
'pytest-cov >= 1.6',
'python-coveralls >= 2.4.2'],
cmdclass={'test': PyTest}
)
|
302086ac111798bb3a5a977ed443ab274ee28dec | setup.py | setup.py | import os
import sys
import setuptools
from keystoneclient.openstack.common import setup
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
requires = setup.parse_requirements()
depend_links = setup.parse_dependency_links()
tests_require = setup.parse_requirements(['tools/test-requires'])
setuptools.setup(
name="python-keystoneclient",
version=setup.get_post_version('keystoneclient'),
description="Client library for OpenStack Keystone API",
long_description=read('README.rst'),
url='https://github.com/openstack/python-keystoneclient',
license='Apache',
author='Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss',
author_email='gabriel.hurley@nebula.com',
packages=setuptools.find_packages(exclude=['tests', 'tests.*']),
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
install_requires=requires,
dependency_links=depend_links,
cmdclass=setup.get_cmdclass(),
tests_require=tests_require,
test_suite="nose.collector",
entry_points={
'console_scripts': ['keystone = keystoneclient.shell:main']
}
)
| import os
import sys
import setuptools
from keystoneclient.openstack.common import setup
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
requires = setup.parse_requirements()
depend_links = setup.parse_dependency_links()
tests_require = setup.parse_requirements(['tools/test-requires'])
setuptools.setup(
name="python-keystoneclient",
version=setup.get_post_version('keystoneclient'),
description="Client library for OpenStack Identity API (Keystone)",
long_description=read('README.rst'),
url='https://github.com/openstack/python-keystoneclient',
license='Apache',
author='Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss',
author_email='gabriel.hurley@nebula.com',
packages=setuptools.find_packages(exclude=['tests', 'tests.*']),
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
install_requires=requires,
dependency_links=depend_links,
cmdclass=setup.get_cmdclass(),
tests_require=tests_require,
test_suite="nose.collector",
entry_points={
'console_scripts': ['keystone = keystoneclient.shell:main']
}
)
| Replace refs to 'Keystone API' with 'Identity API' | Replace refs to 'Keystone API' with 'Identity API'
Formally, OpenStack Keystone implements the OpenStack Identity API, and
this is a client to the API, not to Keystone itself.
Change-Id: If568866221a29ba041f0f2cd56dc81deeb9ebc00
| Python | apache-2.0 | sileht/keystoneauth,jamielennox/keystoneauth,citrix-openstack-build/keystoneauth | import os
import sys
import setuptools
from keystoneclient.openstack.common import setup
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
requires = setup.parse_requirements()
depend_links = setup.parse_dependency_links()
tests_require = setup.parse_requirements(['tools/test-requires'])
setuptools.setup(
name="python-keystoneclient",
version=setup.get_post_version('keystoneclient'),
description="Client library for OpenStack Keystone API",
long_description=read('README.rst'),
url='https://github.com/openstack/python-keystoneclient',
license='Apache',
author='Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss',
author_email='gabriel.hurley@nebula.com',
packages=setuptools.find_packages(exclude=['tests', 'tests.*']),
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
install_requires=requires,
dependency_links=depend_links,
cmdclass=setup.get_cmdclass(),
tests_require=tests_require,
test_suite="nose.collector",
entry_points={
'console_scripts': ['keystone = keystoneclient.shell:main']
}
)
Replace refs to 'Keystone API' with 'Identity API'
Formally, OpenStack Keystone implements the OpenStack Identity API, and
this is a client to the API, not to Keystone itself.
Change-Id: If568866221a29ba041f0f2cd56dc81deeb9ebc00 | import os
import sys
import setuptools
from keystoneclient.openstack.common import setup
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
requires = setup.parse_requirements()
depend_links = setup.parse_dependency_links()
tests_require = setup.parse_requirements(['tools/test-requires'])
setuptools.setup(
name="python-keystoneclient",
version=setup.get_post_version('keystoneclient'),
description="Client library for OpenStack Identity API (Keystone)",
long_description=read('README.rst'),
url='https://github.com/openstack/python-keystoneclient',
license='Apache',
author='Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss',
author_email='gabriel.hurley@nebula.com',
packages=setuptools.find_packages(exclude=['tests', 'tests.*']),
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
install_requires=requires,
dependency_links=depend_links,
cmdclass=setup.get_cmdclass(),
tests_require=tests_require,
test_suite="nose.collector",
entry_points={
'console_scripts': ['keystone = keystoneclient.shell:main']
}
)
| <commit_before>import os
import sys
import setuptools
from keystoneclient.openstack.common import setup
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
requires = setup.parse_requirements()
depend_links = setup.parse_dependency_links()
tests_require = setup.parse_requirements(['tools/test-requires'])
setuptools.setup(
name="python-keystoneclient",
version=setup.get_post_version('keystoneclient'),
description="Client library for OpenStack Keystone API",
long_description=read('README.rst'),
url='https://github.com/openstack/python-keystoneclient',
license='Apache',
author='Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss',
author_email='gabriel.hurley@nebula.com',
packages=setuptools.find_packages(exclude=['tests', 'tests.*']),
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
install_requires=requires,
dependency_links=depend_links,
cmdclass=setup.get_cmdclass(),
tests_require=tests_require,
test_suite="nose.collector",
entry_points={
'console_scripts': ['keystone = keystoneclient.shell:main']
}
)
<commit_msg>Replace refs to 'Keystone API' with 'Identity API'
Formally, OpenStack Keystone implements the OpenStack Identity API, and
this is a client to the API, not to Keystone itself.
Change-Id: If568866221a29ba041f0f2cd56dc81deeb9ebc00<commit_after> | import os
import sys
import setuptools
from keystoneclient.openstack.common import setup
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
requires = setup.parse_requirements()
depend_links = setup.parse_dependency_links()
tests_require = setup.parse_requirements(['tools/test-requires'])
setuptools.setup(
name="python-keystoneclient",
version=setup.get_post_version('keystoneclient'),
description="Client library for OpenStack Identity API (Keystone)",
long_description=read('README.rst'),
url='https://github.com/openstack/python-keystoneclient',
license='Apache',
author='Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss',
author_email='gabriel.hurley@nebula.com',
packages=setuptools.find_packages(exclude=['tests', 'tests.*']),
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
install_requires=requires,
dependency_links=depend_links,
cmdclass=setup.get_cmdclass(),
tests_require=tests_require,
test_suite="nose.collector",
entry_points={
'console_scripts': ['keystone = keystoneclient.shell:main']
}
)
| import os
import sys
import setuptools
from keystoneclient.openstack.common import setup
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
requires = setup.parse_requirements()
depend_links = setup.parse_dependency_links()
tests_require = setup.parse_requirements(['tools/test-requires'])
setuptools.setup(
name="python-keystoneclient",
version=setup.get_post_version('keystoneclient'),
description="Client library for OpenStack Keystone API",
long_description=read('README.rst'),
url='https://github.com/openstack/python-keystoneclient',
license='Apache',
author='Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss',
author_email='gabriel.hurley@nebula.com',
packages=setuptools.find_packages(exclude=['tests', 'tests.*']),
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
install_requires=requires,
dependency_links=depend_links,
cmdclass=setup.get_cmdclass(),
tests_require=tests_require,
test_suite="nose.collector",
entry_points={
'console_scripts': ['keystone = keystoneclient.shell:main']
}
)
Replace refs to 'Keystone API' with 'Identity API'
Formally, OpenStack Keystone implements the OpenStack Identity API, and
this is a client to the API, not to Keystone itself.
Change-Id: If568866221a29ba041f0f2cd56dc81deeb9ebc00import os
import sys
import setuptools
from keystoneclient.openstack.common import setup
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
requires = setup.parse_requirements()
depend_links = setup.parse_dependency_links()
tests_require = setup.parse_requirements(['tools/test-requires'])
setuptools.setup(
name="python-keystoneclient",
version=setup.get_post_version('keystoneclient'),
description="Client library for OpenStack Identity API (Keystone)",
long_description=read('README.rst'),
url='https://github.com/openstack/python-keystoneclient',
license='Apache',
author='Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss',
author_email='gabriel.hurley@nebula.com',
packages=setuptools.find_packages(exclude=['tests', 'tests.*']),
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
install_requires=requires,
dependency_links=depend_links,
cmdclass=setup.get_cmdclass(),
tests_require=tests_require,
test_suite="nose.collector",
entry_points={
'console_scripts': ['keystone = keystoneclient.shell:main']
}
)
| <commit_before>import os
import sys
import setuptools
from keystoneclient.openstack.common import setup
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
requires = setup.parse_requirements()
depend_links = setup.parse_dependency_links()
tests_require = setup.parse_requirements(['tools/test-requires'])
setuptools.setup(
name="python-keystoneclient",
version=setup.get_post_version('keystoneclient'),
description="Client library for OpenStack Keystone API",
long_description=read('README.rst'),
url='https://github.com/openstack/python-keystoneclient',
license='Apache',
author='Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss',
author_email='gabriel.hurley@nebula.com',
packages=setuptools.find_packages(exclude=['tests', 'tests.*']),
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
install_requires=requires,
dependency_links=depend_links,
cmdclass=setup.get_cmdclass(),
tests_require=tests_require,
test_suite="nose.collector",
entry_points={
'console_scripts': ['keystone = keystoneclient.shell:main']
}
)
<commit_msg>Replace refs to 'Keystone API' with 'Identity API'
Formally, OpenStack Keystone implements the OpenStack Identity API, and
this is a client to the API, not to Keystone itself.
Change-Id: If568866221a29ba041f0f2cd56dc81deeb9ebc00<commit_after>import os
import sys
import setuptools
from keystoneclient.openstack.common import setup
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
requires = setup.parse_requirements()
depend_links = setup.parse_dependency_links()
tests_require = setup.parse_requirements(['tools/test-requires'])
setuptools.setup(
name="python-keystoneclient",
version=setup.get_post_version('keystoneclient'),
description="Client library for OpenStack Identity API (Keystone)",
long_description=read('README.rst'),
url='https://github.com/openstack/python-keystoneclient',
license='Apache',
author='Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss',
author_email='gabriel.hurley@nebula.com',
packages=setuptools.find_packages(exclude=['tests', 'tests.*']),
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'License :: OSI Approved :: Apache Software License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
install_requires=requires,
dependency_links=depend_links,
cmdclass=setup.get_cmdclass(),
tests_require=tests_require,
test_suite="nose.collector",
entry_points={
'console_scripts': ['keystone = keystoneclient.shell:main']
}
)
|
4f2a4ad90c85fb50eb02127676dec3e257e2b874 | setup.py | setup.py | # -*- coding: utf-8 -*-
"""Package setup for heroku-tools CLI application."""
import os
from setuptools import setup
dependencies = ['click', 'sarge', 'pyyaml', 'requests', 'dateutils']
README = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read()
setup(
name='heroku-tools',
version='0.2.1',
url='https://github.com/yunojuno/heroku-tools',
license='MIT',
author='Hugo Rodger-Brown',
author_email='hugo@yunojuno.com',
description=(
"Command line application for managing Heroku applications."
),
long_description=README,
include_package_data=True,
packages=[
'heroku_tools',
],
install_requires=dependencies,
entry_points={
'console_scripts': [
'heroku-tools = heroku_tools:entry_point',
],
},
classifiers=[
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
]
)
| # -*- coding: utf-8 -*-
"""Package setup for heroku-tools CLI application."""
import os
from setuptools import setup
dependencies = ['click', 'sarge', 'pyyaml', 'requests', 'dateutils']
README = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read()
setup(
name='heroku-tools',
version='0.2.2',
url='https://github.com/yunojuno/heroku-tools',
license='MIT',
author='Hugo Rodger-Brown',
author_email='hugo@yunojuno.com',
description=(
"Command line application for managing Heroku applications."
),
long_description=README,
include_package_data=True,
packages=[
'heroku_tools',
'heroku_tools.settings',
],
install_requires=dependencies,
entry_points={
'console_scripts': [
'heroku-tools = heroku_tools:entry_point',
],
},
classifiers=[
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
]
)
| Add missing settings package to release (0.2.2) | Add missing settings package to release (0.2.2)
| Python | mit | yunojuno/heroku-tools | # -*- coding: utf-8 -*-
"""Package setup for heroku-tools CLI application."""
import os
from setuptools import setup
dependencies = ['click', 'sarge', 'pyyaml', 'requests', 'dateutils']
README = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read()
setup(
name='heroku-tools',
version='0.2.1',
url='https://github.com/yunojuno/heroku-tools',
license='MIT',
author='Hugo Rodger-Brown',
author_email='hugo@yunojuno.com',
description=(
"Command line application for managing Heroku applications."
),
long_description=README,
include_package_data=True,
packages=[
'heroku_tools',
],
install_requires=dependencies,
entry_points={
'console_scripts': [
'heroku-tools = heroku_tools:entry_point',
],
},
classifiers=[
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
]
)
Add missing settings package to release (0.2.2) | # -*- coding: utf-8 -*-
"""Package setup for heroku-tools CLI application."""
import os
from setuptools import setup
dependencies = ['click', 'sarge', 'pyyaml', 'requests', 'dateutils']
README = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read()
setup(
name='heroku-tools',
version='0.2.2',
url='https://github.com/yunojuno/heroku-tools',
license='MIT',
author='Hugo Rodger-Brown',
author_email='hugo@yunojuno.com',
description=(
"Command line application for managing Heroku applications."
),
long_description=README,
include_package_data=True,
packages=[
'heroku_tools',
'heroku_tools.settings',
],
install_requires=dependencies,
entry_points={
'console_scripts': [
'heroku-tools = heroku_tools:entry_point',
],
},
classifiers=[
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
]
)
| <commit_before># -*- coding: utf-8 -*-
"""Package setup for heroku-tools CLI application."""
import os
from setuptools import setup
dependencies = ['click', 'sarge', 'pyyaml', 'requests', 'dateutils']
README = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read()
setup(
name='heroku-tools',
version='0.2.1',
url='https://github.com/yunojuno/heroku-tools',
license='MIT',
author='Hugo Rodger-Brown',
author_email='hugo@yunojuno.com',
description=(
"Command line application for managing Heroku applications."
),
long_description=README,
include_package_data=True,
packages=[
'heroku_tools',
],
install_requires=dependencies,
entry_points={
'console_scripts': [
'heroku-tools = heroku_tools:entry_point',
],
},
classifiers=[
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
]
)
<commit_msg>Add missing settings package to release (0.2.2)<commit_after> | # -*- coding: utf-8 -*-
"""Package setup for heroku-tools CLI application."""
import os
from setuptools import setup
dependencies = ['click', 'sarge', 'pyyaml', 'requests', 'dateutils']
README = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read()
setup(
name='heroku-tools',
version='0.2.2',
url='https://github.com/yunojuno/heroku-tools',
license='MIT',
author='Hugo Rodger-Brown',
author_email='hugo@yunojuno.com',
description=(
"Command line application for managing Heroku applications."
),
long_description=README,
include_package_data=True,
packages=[
'heroku_tools',
'heroku_tools.settings',
],
install_requires=dependencies,
entry_points={
'console_scripts': [
'heroku-tools = heroku_tools:entry_point',
],
},
classifiers=[
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
]
)
| # -*- coding: utf-8 -*-
"""Package setup for heroku-tools CLI application."""
import os
from setuptools import setup
dependencies = ['click', 'sarge', 'pyyaml', 'requests', 'dateutils']
README = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read()
setup(
name='heroku-tools',
version='0.2.1',
url='https://github.com/yunojuno/heroku-tools',
license='MIT',
author='Hugo Rodger-Brown',
author_email='hugo@yunojuno.com',
description=(
"Command line application for managing Heroku applications."
),
long_description=README,
include_package_data=True,
packages=[
'heroku_tools',
],
install_requires=dependencies,
entry_points={
'console_scripts': [
'heroku-tools = heroku_tools:entry_point',
],
},
classifiers=[
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
]
)
Add missing settings package to release (0.2.2)# -*- coding: utf-8 -*-
"""Package setup for heroku-tools CLI application."""
import os
from setuptools import setup
dependencies = ['click', 'sarge', 'pyyaml', 'requests', 'dateutils']
README = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read()
setup(
name='heroku-tools',
version='0.2.2',
url='https://github.com/yunojuno/heroku-tools',
license='MIT',
author='Hugo Rodger-Brown',
author_email='hugo@yunojuno.com',
description=(
"Command line application for managing Heroku applications."
),
long_description=README,
include_package_data=True,
packages=[
'heroku_tools',
'heroku_tools.settings',
],
install_requires=dependencies,
entry_points={
'console_scripts': [
'heroku-tools = heroku_tools:entry_point',
],
},
classifiers=[
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
]
)
| <commit_before># -*- coding: utf-8 -*-
"""Package setup for heroku-tools CLI application."""
import os
from setuptools import setup
dependencies = ['click', 'sarge', 'pyyaml', 'requests', 'dateutils']
README = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read()
setup(
name='heroku-tools',
version='0.2.1',
url='https://github.com/yunojuno/heroku-tools',
license='MIT',
author='Hugo Rodger-Brown',
author_email='hugo@yunojuno.com',
description=(
"Command line application for managing Heroku applications."
),
long_description=README,
include_package_data=True,
packages=[
'heroku_tools',
],
install_requires=dependencies,
entry_points={
'console_scripts': [
'heroku-tools = heroku_tools:entry_point',
],
},
classifiers=[
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
]
)
<commit_msg>Add missing settings package to release (0.2.2)<commit_after># -*- coding: utf-8 -*-
"""Package setup for heroku-tools CLI application."""
import os
from setuptools import setup
dependencies = ['click', 'sarge', 'pyyaml', 'requests', 'dateutils']
README = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read()
setup(
name='heroku-tools',
version='0.2.2',
url='https://github.com/yunojuno/heroku-tools',
license='MIT',
author='Hugo Rodger-Brown',
author_email='hugo@yunojuno.com',
description=(
"Command line application for managing Heroku applications."
),
long_description=README,
include_package_data=True,
packages=[
'heroku_tools',
'heroku_tools.settings',
],
install_requires=dependencies,
entry_points={
'console_scripts': [
'heroku-tools = heroku_tools:entry_point',
],
},
classifiers=[
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
]
)
|
21e99ff4617d4307d2223b79c25d3642132e27b3 | setup.py | setup.py | from setuptools import setup
setup(name='journal_dates',
packages=[''],
version='0.1',
description='Prints a monthly journal template',
url='http://github.com/bzamecnik/journal_dates',
author='Bohumir Zamecnik',
author_email='bohumir.zamecnik@gmail.com',
license='MIT',
install_requires=['arrow'],
zip_safe=False,
entry_points={
'console_scripts': [
'journal_dates=journal_dates:main',
],
})
| from setuptools import setup
setup(name='journal_dates',
packages=[''],
version='0.1',
description='Prints a monthly journal template',
url='http://github.com/bzamecnik/journal_dates.py',
author='Bohumir Zamecnik',
author_email='bohumir.zamecnik@gmail.com',
license='MIT',
install_requires=['arrow'],
zip_safe=False,
entry_points={
'console_scripts': [
'journal_dates=journal_dates:main',
],
})
| Fix the github repo name. | Fix the github repo name.
| Python | mit | bzamecnik/journal_dates | from setuptools import setup
setup(name='journal_dates',
packages=[''],
version='0.1',
description='Prints a monthly journal template',
url='http://github.com/bzamecnik/journal_dates',
author='Bohumir Zamecnik',
author_email='bohumir.zamecnik@gmail.com',
license='MIT',
install_requires=['arrow'],
zip_safe=False,
entry_points={
'console_scripts': [
'journal_dates=journal_dates:main',
],
})
Fix the github repo name. | from setuptools import setup
setup(name='journal_dates',
packages=[''],
version='0.1',
description='Prints a monthly journal template',
url='http://github.com/bzamecnik/journal_dates.py',
author='Bohumir Zamecnik',
author_email='bohumir.zamecnik@gmail.com',
license='MIT',
install_requires=['arrow'],
zip_safe=False,
entry_points={
'console_scripts': [
'journal_dates=journal_dates:main',
],
})
| <commit_before>from setuptools import setup
setup(name='journal_dates',
packages=[''],
version='0.1',
description='Prints a monthly journal template',
url='http://github.com/bzamecnik/journal_dates',
author='Bohumir Zamecnik',
author_email='bohumir.zamecnik@gmail.com',
license='MIT',
install_requires=['arrow'],
zip_safe=False,
entry_points={
'console_scripts': [
'journal_dates=journal_dates:main',
],
})
<commit_msg>Fix the github repo name.<commit_after> | from setuptools import setup
setup(name='journal_dates',
packages=[''],
version='0.1',
description='Prints a monthly journal template',
url='http://github.com/bzamecnik/journal_dates.py',
author='Bohumir Zamecnik',
author_email='bohumir.zamecnik@gmail.com',
license='MIT',
install_requires=['arrow'],
zip_safe=False,
entry_points={
'console_scripts': [
'journal_dates=journal_dates:main',
],
})
| from setuptools import setup
setup(name='journal_dates',
packages=[''],
version='0.1',
description='Prints a monthly journal template',
url='http://github.com/bzamecnik/journal_dates',
author='Bohumir Zamecnik',
author_email='bohumir.zamecnik@gmail.com',
license='MIT',
install_requires=['arrow'],
zip_safe=False,
entry_points={
'console_scripts': [
'journal_dates=journal_dates:main',
],
})
Fix the github repo name.from setuptools import setup
setup(name='journal_dates',
packages=[''],
version='0.1',
description='Prints a monthly journal template',
url='http://github.com/bzamecnik/journal_dates.py',
author='Bohumir Zamecnik',
author_email='bohumir.zamecnik@gmail.com',
license='MIT',
install_requires=['arrow'],
zip_safe=False,
entry_points={
'console_scripts': [
'journal_dates=journal_dates:main',
],
})
| <commit_before>from setuptools import setup
setup(name='journal_dates',
packages=[''],
version='0.1',
description='Prints a monthly journal template',
url='http://github.com/bzamecnik/journal_dates',
author='Bohumir Zamecnik',
author_email='bohumir.zamecnik@gmail.com',
license='MIT',
install_requires=['arrow'],
zip_safe=False,
entry_points={
'console_scripts': [
'journal_dates=journal_dates:main',
],
})
<commit_msg>Fix the github repo name.<commit_after>from setuptools import setup
setup(name='journal_dates',
packages=[''],
version='0.1',
description='Prints a monthly journal template',
url='http://github.com/bzamecnik/journal_dates.py',
author='Bohumir Zamecnik',
author_email='bohumir.zamecnik@gmail.com',
license='MIT',
install_requires=['arrow'],
zip_safe=False,
entry_points={
'console_scripts': [
'journal_dates=journal_dates:main',
],
})
|
86fc1b91aeb9dc17b5776ef372050c707c64fc30 | setup.py | setup.py | #!/usr/bin/env python
"""
properties: Fancy properties for Python.
"""
from distutils.core import setup
from setuptools import find_packages
CLASSIFIERS = [
'Development Status :: 4 - Beta',
'Programming Language :: Python',
'Topic :: Scientific/Engineering',
'Topic :: Scientific/Engineering :: Mathematics',
'Topic :: Scientific/Engineering :: Physics',
'Operating System :: Microsoft :: Windows',
'Operating System :: POSIX',
'Operating System :: Unix',
'Operating System :: MacOS',
'Natural Language :: English',
]
with open("README.rst") as f:
LONG_DESCRIPTION = ''.join(f.readlines())
setup(
name="properties",
version="0.2.3",
packages=find_packages(exclude=('tests',)),
install_requires=[
'future',
'numpy>=1.7',
'six',
'vectormath>=0.1.0',
],
author="3point Science",
author_email="info@3ptscience.com",
description="properties",
long_description=LONG_DESCRIPTION,
keywords="property",
url="http://steno3d.com/",
download_url="http://github.com/3ptscience/properties",
classifiers=CLASSIFIERS,
platforms=["Windows", "Linux", "Solaris", "Mac OS-X", "Unix"],
use_2to3=False,
)
| #!/usr/bin/env python
"""
properties: Fancy properties for Python.
"""
from distutils.core import setup
from setuptools import find_packages
CLASSIFIERS = [
'Development Status :: 4 - Beta',
'Programming Language :: Python',
'Topic :: Scientific/Engineering',
'Topic :: Scientific/Engineering :: Mathematics',
'Topic :: Scientific/Engineering :: Physics',
'Operating System :: Microsoft :: Windows',
'Operating System :: POSIX',
'Operating System :: Unix',
'Operating System :: MacOS',
'Natural Language :: English',
]
with open("README.rst") as f:
LONG_DESCRIPTION = ''.join(f.readlines())
setup(
name="properties",
version="0.2.3",
packages=find_packages(exclude=('tests',)),
install_requires=[
'numpy>=1.7',
'six',
'vectormath>=0.1.1',
],
author="3point Science",
author_email="info@3ptscience.com",
description="properties",
long_description=LONG_DESCRIPTION,
keywords="property",
url="http://steno3d.com/",
download_url="http://github.com/3ptscience/properties",
classifiers=CLASSIFIERS,
platforms=["Windows", "Linux", "Solaris", "Mac OS-X", "Unix"],
use_2to3=False,
)
| Remove future, bump vectormath dependencies | Remove future, bump vectormath dependencies
| Python | mit | 3ptscience/properties,aranzgeo/properties | #!/usr/bin/env python
"""
properties: Fancy properties for Python.
"""
from distutils.core import setup
from setuptools import find_packages
CLASSIFIERS = [
'Development Status :: 4 - Beta',
'Programming Language :: Python',
'Topic :: Scientific/Engineering',
'Topic :: Scientific/Engineering :: Mathematics',
'Topic :: Scientific/Engineering :: Physics',
'Operating System :: Microsoft :: Windows',
'Operating System :: POSIX',
'Operating System :: Unix',
'Operating System :: MacOS',
'Natural Language :: English',
]
with open("README.rst") as f:
LONG_DESCRIPTION = ''.join(f.readlines())
setup(
name="properties",
version="0.2.3",
packages=find_packages(exclude=('tests',)),
install_requires=[
'future',
'numpy>=1.7',
'six',
'vectormath>=0.1.0',
],
author="3point Science",
author_email="info@3ptscience.com",
description="properties",
long_description=LONG_DESCRIPTION,
keywords="property",
url="http://steno3d.com/",
download_url="http://github.com/3ptscience/properties",
classifiers=CLASSIFIERS,
platforms=["Windows", "Linux", "Solaris", "Mac OS-X", "Unix"],
use_2to3=False,
)
Remove future, bump vectormath dependencies | #!/usr/bin/env python
"""
properties: Fancy properties for Python.
"""
from distutils.core import setup
from setuptools import find_packages
CLASSIFIERS = [
'Development Status :: 4 - Beta',
'Programming Language :: Python',
'Topic :: Scientific/Engineering',
'Topic :: Scientific/Engineering :: Mathematics',
'Topic :: Scientific/Engineering :: Physics',
'Operating System :: Microsoft :: Windows',
'Operating System :: POSIX',
'Operating System :: Unix',
'Operating System :: MacOS',
'Natural Language :: English',
]
with open("README.rst") as f:
LONG_DESCRIPTION = ''.join(f.readlines())
setup(
name="properties",
version="0.2.3",
packages=find_packages(exclude=('tests',)),
install_requires=[
'numpy>=1.7',
'six',
'vectormath>=0.1.1',
],
author="3point Science",
author_email="info@3ptscience.com",
description="properties",
long_description=LONG_DESCRIPTION,
keywords="property",
url="http://steno3d.com/",
download_url="http://github.com/3ptscience/properties",
classifiers=CLASSIFIERS,
platforms=["Windows", "Linux", "Solaris", "Mac OS-X", "Unix"],
use_2to3=False,
)
| <commit_before>#!/usr/bin/env python
"""
properties: Fancy properties for Python.
"""
from distutils.core import setup
from setuptools import find_packages
CLASSIFIERS = [
'Development Status :: 4 - Beta',
'Programming Language :: Python',
'Topic :: Scientific/Engineering',
'Topic :: Scientific/Engineering :: Mathematics',
'Topic :: Scientific/Engineering :: Physics',
'Operating System :: Microsoft :: Windows',
'Operating System :: POSIX',
'Operating System :: Unix',
'Operating System :: MacOS',
'Natural Language :: English',
]
with open("README.rst") as f:
LONG_DESCRIPTION = ''.join(f.readlines())
setup(
name="properties",
version="0.2.3",
packages=find_packages(exclude=('tests',)),
install_requires=[
'future',
'numpy>=1.7',
'six',
'vectormath>=0.1.0',
],
author="3point Science",
author_email="info@3ptscience.com",
description="properties",
long_description=LONG_DESCRIPTION,
keywords="property",
url="http://steno3d.com/",
download_url="http://github.com/3ptscience/properties",
classifiers=CLASSIFIERS,
platforms=["Windows", "Linux", "Solaris", "Mac OS-X", "Unix"],
use_2to3=False,
)
<commit_msg>Remove future, bump vectormath dependencies<commit_after> | #!/usr/bin/env python
"""
properties: Fancy properties for Python.
"""
from distutils.core import setup
from setuptools import find_packages
CLASSIFIERS = [
'Development Status :: 4 - Beta',
'Programming Language :: Python',
'Topic :: Scientific/Engineering',
'Topic :: Scientific/Engineering :: Mathematics',
'Topic :: Scientific/Engineering :: Physics',
'Operating System :: Microsoft :: Windows',
'Operating System :: POSIX',
'Operating System :: Unix',
'Operating System :: MacOS',
'Natural Language :: English',
]
with open("README.rst") as f:
LONG_DESCRIPTION = ''.join(f.readlines())
setup(
name="properties",
version="0.2.3",
packages=find_packages(exclude=('tests',)),
install_requires=[
'numpy>=1.7',
'six',
'vectormath>=0.1.1',
],
author="3point Science",
author_email="info@3ptscience.com",
description="properties",
long_description=LONG_DESCRIPTION,
keywords="property",
url="http://steno3d.com/",
download_url="http://github.com/3ptscience/properties",
classifiers=CLASSIFIERS,
platforms=["Windows", "Linux", "Solaris", "Mac OS-X", "Unix"],
use_2to3=False,
)
| #!/usr/bin/env python
"""
properties: Fancy properties for Python.
"""
from distutils.core import setup
from setuptools import find_packages
CLASSIFIERS = [
'Development Status :: 4 - Beta',
'Programming Language :: Python',
'Topic :: Scientific/Engineering',
'Topic :: Scientific/Engineering :: Mathematics',
'Topic :: Scientific/Engineering :: Physics',
'Operating System :: Microsoft :: Windows',
'Operating System :: POSIX',
'Operating System :: Unix',
'Operating System :: MacOS',
'Natural Language :: English',
]
with open("README.rst") as f:
LONG_DESCRIPTION = ''.join(f.readlines())
setup(
name="properties",
version="0.2.3",
packages=find_packages(exclude=('tests',)),
install_requires=[
'future',
'numpy>=1.7',
'six',
'vectormath>=0.1.0',
],
author="3point Science",
author_email="info@3ptscience.com",
description="properties",
long_description=LONG_DESCRIPTION,
keywords="property",
url="http://steno3d.com/",
download_url="http://github.com/3ptscience/properties",
classifiers=CLASSIFIERS,
platforms=["Windows", "Linux", "Solaris", "Mac OS-X", "Unix"],
use_2to3=False,
)
Remove future, bump vectormath dependencies#!/usr/bin/env python
"""
properties: Fancy properties for Python.
"""
from distutils.core import setup
from setuptools import find_packages
CLASSIFIERS = [
'Development Status :: 4 - Beta',
'Programming Language :: Python',
'Topic :: Scientific/Engineering',
'Topic :: Scientific/Engineering :: Mathematics',
'Topic :: Scientific/Engineering :: Physics',
'Operating System :: Microsoft :: Windows',
'Operating System :: POSIX',
'Operating System :: Unix',
'Operating System :: MacOS',
'Natural Language :: English',
]
with open("README.rst") as f:
LONG_DESCRIPTION = ''.join(f.readlines())
setup(
name="properties",
version="0.2.3",
packages=find_packages(exclude=('tests',)),
install_requires=[
'numpy>=1.7',
'six',
'vectormath>=0.1.1',
],
author="3point Science",
author_email="info@3ptscience.com",
description="properties",
long_description=LONG_DESCRIPTION,
keywords="property",
url="http://steno3d.com/",
download_url="http://github.com/3ptscience/properties",
classifiers=CLASSIFIERS,
platforms=["Windows", "Linux", "Solaris", "Mac OS-X", "Unix"],
use_2to3=False,
)
| <commit_before>#!/usr/bin/env python
"""
properties: Fancy properties for Python.
"""
from distutils.core import setup
from setuptools import find_packages
CLASSIFIERS = [
'Development Status :: 4 - Beta',
'Programming Language :: Python',
'Topic :: Scientific/Engineering',
'Topic :: Scientific/Engineering :: Mathematics',
'Topic :: Scientific/Engineering :: Physics',
'Operating System :: Microsoft :: Windows',
'Operating System :: POSIX',
'Operating System :: Unix',
'Operating System :: MacOS',
'Natural Language :: English',
]
with open("README.rst") as f:
LONG_DESCRIPTION = ''.join(f.readlines())
setup(
name="properties",
version="0.2.3",
packages=find_packages(exclude=('tests',)),
install_requires=[
'future',
'numpy>=1.7',
'six',
'vectormath>=0.1.0',
],
author="3point Science",
author_email="info@3ptscience.com",
description="properties",
long_description=LONG_DESCRIPTION,
keywords="property",
url="http://steno3d.com/",
download_url="http://github.com/3ptscience/properties",
classifiers=CLASSIFIERS,
platforms=["Windows", "Linux", "Solaris", "Mac OS-X", "Unix"],
use_2to3=False,
)
<commit_msg>Remove future, bump vectormath dependencies<commit_after>#!/usr/bin/env python
"""
properties: Fancy properties for Python.
"""
from distutils.core import setup
from setuptools import find_packages
CLASSIFIERS = [
'Development Status :: 4 - Beta',
'Programming Language :: Python',
'Topic :: Scientific/Engineering',
'Topic :: Scientific/Engineering :: Mathematics',
'Topic :: Scientific/Engineering :: Physics',
'Operating System :: Microsoft :: Windows',
'Operating System :: POSIX',
'Operating System :: Unix',
'Operating System :: MacOS',
'Natural Language :: English',
]
with open("README.rst") as f:
LONG_DESCRIPTION = ''.join(f.readlines())
setup(
name="properties",
version="0.2.3",
packages=find_packages(exclude=('tests',)),
install_requires=[
'numpy>=1.7',
'six',
'vectormath>=0.1.1',
],
author="3point Science",
author_email="info@3ptscience.com",
description="properties",
long_description=LONG_DESCRIPTION,
keywords="property",
url="http://steno3d.com/",
download_url="http://github.com/3ptscience/properties",
classifiers=CLASSIFIERS,
platforms=["Windows", "Linux", "Solaris", "Mac OS-X", "Unix"],
use_2to3=False,
)
|
c4b4a7ab03728b907d0c64e3322fd08a56f5c9ce | suave.py | suave.py | #!/usr/bin/env python
import curses
import os
import time
from box import Box
from utils import load_yaml
def main(screen):
"""
Draws and redraws the screen.
"""
# Hide the cursor.
curses.curs_set(0)
# Load config from file.
config = load_yaml(os.path.expanduser('~/.suave/config.yml'))
# Create boxes from config.
boxes = []
for box in config:
boxes.append(
Box(
screen=screen,
rows=box['rows'],
columns=box['columns'],
rows_offset=box['rows-offset'],
columns_offset=box['columns-offset'],
command=box['command'],
interval=box['interval'],
)
)
while True:
# Re/draw the screen and boxes.
screen.erase()
screen.refresh()
[box.refresh() for box in boxes]
# Wait before redrawing again.
time.sleep(1)
curses.wrapper(main)
| #!/usr/bin/env python
import curses
import os
import time
from box import Box
from utils import load_yaml
def main(screen):
"""
Draws and redraws the screen.
"""
# Hide the cursor.
curses.curs_set(0)
# Load config from file.
config = load_yaml(os.path.expanduser('~/.suave/config.yml'))
# Create boxes from config.
boxes = []
for box in config:
boxes.append(
Box(
screen=screen,
rows=box['rows'],
columns=box['columns'],
rows_offset=box['rows-offset'],
columns_offset=box['columns-offset'],
command=box['command'],
interval=box['interval'],
)
)
while True:
# Re/draw the screen and boxes.
screen.clear()
screen.refresh()
[box.refresh() for box in boxes]
# Wait before redrawing again.
time.sleep(1)
curses.wrapper(main)
| Clear the screen rather than erasing it before refreshing | Clear the screen rather than erasing it before refreshing
| Python | mit | countermeasure/suave | #!/usr/bin/env python
import curses
import os
import time
from box import Box
from utils import load_yaml
def main(screen):
"""
Draws and redraws the screen.
"""
# Hide the cursor.
curses.curs_set(0)
# Load config from file.
config = load_yaml(os.path.expanduser('~/.suave/config.yml'))
# Create boxes from config.
boxes = []
for box in config:
boxes.append(
Box(
screen=screen,
rows=box['rows'],
columns=box['columns'],
rows_offset=box['rows-offset'],
columns_offset=box['columns-offset'],
command=box['command'],
interval=box['interval'],
)
)
while True:
# Re/draw the screen and boxes.
screen.erase()
screen.refresh()
[box.refresh() for box in boxes]
# Wait before redrawing again.
time.sleep(1)
curses.wrapper(main)
Clear the screen rather than erasing it before refreshing | #!/usr/bin/env python
import curses
import os
import time
from box import Box
from utils import load_yaml
def main(screen):
"""
Draws and redraws the screen.
"""
# Hide the cursor.
curses.curs_set(0)
# Load config from file.
config = load_yaml(os.path.expanduser('~/.suave/config.yml'))
# Create boxes from config.
boxes = []
for box in config:
boxes.append(
Box(
screen=screen,
rows=box['rows'],
columns=box['columns'],
rows_offset=box['rows-offset'],
columns_offset=box['columns-offset'],
command=box['command'],
interval=box['interval'],
)
)
while True:
# Re/draw the screen and boxes.
screen.clear()
screen.refresh()
[box.refresh() for box in boxes]
# Wait before redrawing again.
time.sleep(1)
curses.wrapper(main)
| <commit_before>#!/usr/bin/env python
import curses
import os
import time
from box import Box
from utils import load_yaml
def main(screen):
"""
Draws and redraws the screen.
"""
# Hide the cursor.
curses.curs_set(0)
# Load config from file.
config = load_yaml(os.path.expanduser('~/.suave/config.yml'))
# Create boxes from config.
boxes = []
for box in config:
boxes.append(
Box(
screen=screen,
rows=box['rows'],
columns=box['columns'],
rows_offset=box['rows-offset'],
columns_offset=box['columns-offset'],
command=box['command'],
interval=box['interval'],
)
)
while True:
# Re/draw the screen and boxes.
screen.erase()
screen.refresh()
[box.refresh() for box in boxes]
# Wait before redrawing again.
time.sleep(1)
curses.wrapper(main)
<commit_msg>Clear the screen rather than erasing it before refreshing<commit_after> | #!/usr/bin/env python
import curses
import os
import time
from box import Box
from utils import load_yaml
def main(screen):
"""
Draws and redraws the screen.
"""
# Hide the cursor.
curses.curs_set(0)
# Load config from file.
config = load_yaml(os.path.expanduser('~/.suave/config.yml'))
# Create boxes from config.
boxes = []
for box in config:
boxes.append(
Box(
screen=screen,
rows=box['rows'],
columns=box['columns'],
rows_offset=box['rows-offset'],
columns_offset=box['columns-offset'],
command=box['command'],
interval=box['interval'],
)
)
while True:
# Re/draw the screen and boxes.
screen.clear()
screen.refresh()
[box.refresh() for box in boxes]
# Wait before redrawing again.
time.sleep(1)
curses.wrapper(main)
| #!/usr/bin/env python
import curses
import os
import time
from box import Box
from utils import load_yaml
def main(screen):
"""
Draws and redraws the screen.
"""
# Hide the cursor.
curses.curs_set(0)
# Load config from file.
config = load_yaml(os.path.expanduser('~/.suave/config.yml'))
# Create boxes from config.
boxes = []
for box in config:
boxes.append(
Box(
screen=screen,
rows=box['rows'],
columns=box['columns'],
rows_offset=box['rows-offset'],
columns_offset=box['columns-offset'],
command=box['command'],
interval=box['interval'],
)
)
while True:
# Re/draw the screen and boxes.
screen.erase()
screen.refresh()
[box.refresh() for box in boxes]
# Wait before redrawing again.
time.sleep(1)
curses.wrapper(main)
Clear the screen rather than erasing it before refreshing#!/usr/bin/env python
import curses
import os
import time
from box import Box
from utils import load_yaml
def main(screen):
"""
Draws and redraws the screen.
"""
# Hide the cursor.
curses.curs_set(0)
# Load config from file.
config = load_yaml(os.path.expanduser('~/.suave/config.yml'))
# Create boxes from config.
boxes = []
for box in config:
boxes.append(
Box(
screen=screen,
rows=box['rows'],
columns=box['columns'],
rows_offset=box['rows-offset'],
columns_offset=box['columns-offset'],
command=box['command'],
interval=box['interval'],
)
)
while True:
# Re/draw the screen and boxes.
screen.clear()
screen.refresh()
[box.refresh() for box in boxes]
# Wait before redrawing again.
time.sleep(1)
curses.wrapper(main)
| <commit_before>#!/usr/bin/env python
import curses
import os
import time
from box import Box
from utils import load_yaml
def main(screen):
"""
Draws and redraws the screen.
"""
# Hide the cursor.
curses.curs_set(0)
# Load config from file.
config = load_yaml(os.path.expanduser('~/.suave/config.yml'))
# Create boxes from config.
boxes = []
for box in config:
boxes.append(
Box(
screen=screen,
rows=box['rows'],
columns=box['columns'],
rows_offset=box['rows-offset'],
columns_offset=box['columns-offset'],
command=box['command'],
interval=box['interval'],
)
)
while True:
# Re/draw the screen and boxes.
screen.erase()
screen.refresh()
[box.refresh() for box in boxes]
# Wait before redrawing again.
time.sleep(1)
curses.wrapper(main)
<commit_msg>Clear the screen rather than erasing it before refreshing<commit_after>#!/usr/bin/env python
import curses
import os
import time
from box import Box
from utils import load_yaml
def main(screen):
"""
Draws and redraws the screen.
"""
# Hide the cursor.
curses.curs_set(0)
# Load config from file.
config = load_yaml(os.path.expanduser('~/.suave/config.yml'))
# Create boxes from config.
boxes = []
for box in config:
boxes.append(
Box(
screen=screen,
rows=box['rows'],
columns=box['columns'],
rows_offset=box['rows-offset'],
columns_offset=box['columns-offset'],
command=box['command'],
interval=box['interval'],
)
)
while True:
# Re/draw the screen and boxes.
screen.clear()
screen.refresh()
[box.refresh() for box in boxes]
# Wait before redrawing again.
time.sleep(1)
curses.wrapper(main)
|
4ee409a5635b1d027f5d3c68fb2a62f554c9c801 | ib_insync/__init__.py | ib_insync/__init__.py | import sys
import importlib
from .version import __version__, __version_info__ # noqa
from . import util
if sys.version_info < (3, 6, 0):
raise RuntimeError('ib_insync requires Python 3.6 or higher')
try:
import ibapi
except ImportError:
raise RuntimeError(
'IB API from http://interactivebrokers.github.io is required')
if util.ibapiVersionInfo() < (9, 73, 6):
raise RuntimeError(
f'Old version ({ibapi.__version__}) of ibapi package detected. '
'The newest version from http://interactivebrokers.github.io '
'is required')
from .version import __version__, __version_info__ # noqa
from .objects import * # noqa
from .event import * # noqa
from .contract import * # noqa
from .order import * # noqa
from .ticker import * # noqa
from .ib import * # noqa
from .client import * # noqa
from .wrapper import * # noqa
from .flexreport import * # noqa
from .ibcontroller import * # noqa
from . import util # noqa
__all__ = ['util']
for _m in (
objects, event, contract, order, ticker, ib, # noqa
client, wrapper, flexreport, ibcontroller): # noqa
__all__ += _m.__all__
del sys
del importlib
del ibapi
| import sys
import importlib
if sys.version_info < (3, 6, 0):
raise RuntimeError('ib_insync requires Python 3.6 or higher')
try:
import ibapi
except ImportError:
raise RuntimeError(
'IB API from http://interactivebrokers.github.io is required')
from . import util # noqa
if util.ibapiVersionInfo() < (9, 73, 6):
raise RuntimeError(
f'Old version ({ibapi.__version__}) of ibapi package detected. '
'The newest version from http://interactivebrokers.github.io '
'is required')
from .version import __version__, __version_info__ # noqa
from .objects import * # noqa
from .event import * # noqa
from .contract import * # noqa
from .order import * # noqa
from .ticker import * # noqa
from .ib import * # noqa
from .client import * # noqa
from .wrapper import * # noqa
from .flexreport import * # noqa
from .ibcontroller import * # noqa
__all__ = ['util']
for _m in (
objects, event, contract, order, ticker, ib, # noqa
client, wrapper, flexreport, ibcontroller): # noqa
__all__ += _m.__all__
del sys
del importlib
del ibapi
| Fix explicit check for presence of ibapi package | Fix explicit check for presence of ibapi package
| Python | bsd-2-clause | erdewit/ib_insync,erdewit/ib_insync | import sys
import importlib
from .version import __version__, __version_info__ # noqa
from . import util
if sys.version_info < (3, 6, 0):
raise RuntimeError('ib_insync requires Python 3.6 or higher')
try:
import ibapi
except ImportError:
raise RuntimeError(
'IB API from http://interactivebrokers.github.io is required')
if util.ibapiVersionInfo() < (9, 73, 6):
raise RuntimeError(
f'Old version ({ibapi.__version__}) of ibapi package detected. '
'The newest version from http://interactivebrokers.github.io '
'is required')
from .version import __version__, __version_info__ # noqa
from .objects import * # noqa
from .event import * # noqa
from .contract import * # noqa
from .order import * # noqa
from .ticker import * # noqa
from .ib import * # noqa
from .client import * # noqa
from .wrapper import * # noqa
from .flexreport import * # noqa
from .ibcontroller import * # noqa
from . import util # noqa
__all__ = ['util']
for _m in (
objects, event, contract, order, ticker, ib, # noqa
client, wrapper, flexreport, ibcontroller): # noqa
__all__ += _m.__all__
del sys
del importlib
del ibapi
Fix explicit check for presence of ibapi package | import sys
import importlib
if sys.version_info < (3, 6, 0):
raise RuntimeError('ib_insync requires Python 3.6 or higher')
try:
import ibapi
except ImportError:
raise RuntimeError(
'IB API from http://interactivebrokers.github.io is required')
from . import util # noqa
if util.ibapiVersionInfo() < (9, 73, 6):
raise RuntimeError(
f'Old version ({ibapi.__version__}) of ibapi package detected. '
'The newest version from http://interactivebrokers.github.io '
'is required')
from .version import __version__, __version_info__ # noqa
from .objects import * # noqa
from .event import * # noqa
from .contract import * # noqa
from .order import * # noqa
from .ticker import * # noqa
from .ib import * # noqa
from .client import * # noqa
from .wrapper import * # noqa
from .flexreport import * # noqa
from .ibcontroller import * # noqa
__all__ = ['util']
for _m in (
objects, event, contract, order, ticker, ib, # noqa
client, wrapper, flexreport, ibcontroller): # noqa
__all__ += _m.__all__
del sys
del importlib
del ibapi
| <commit_before>import sys
import importlib
from .version import __version__, __version_info__ # noqa
from . import util
if sys.version_info < (3, 6, 0):
raise RuntimeError('ib_insync requires Python 3.6 or higher')
try:
import ibapi
except ImportError:
raise RuntimeError(
'IB API from http://interactivebrokers.github.io is required')
if util.ibapiVersionInfo() < (9, 73, 6):
raise RuntimeError(
f'Old version ({ibapi.__version__}) of ibapi package detected. '
'The newest version from http://interactivebrokers.github.io '
'is required')
from .version import __version__, __version_info__ # noqa
from .objects import * # noqa
from .event import * # noqa
from .contract import * # noqa
from .order import * # noqa
from .ticker import * # noqa
from .ib import * # noqa
from .client import * # noqa
from .wrapper import * # noqa
from .flexreport import * # noqa
from .ibcontroller import * # noqa
from . import util # noqa
__all__ = ['util']
for _m in (
objects, event, contract, order, ticker, ib, # noqa
client, wrapper, flexreport, ibcontroller): # noqa
__all__ += _m.__all__
del sys
del importlib
del ibapi
<commit_msg>Fix explicit check for presence of ibapi package<commit_after> | import sys
import importlib
if sys.version_info < (3, 6, 0):
raise RuntimeError('ib_insync requires Python 3.6 or higher')
try:
import ibapi
except ImportError:
raise RuntimeError(
'IB API from http://interactivebrokers.github.io is required')
from . import util # noqa
if util.ibapiVersionInfo() < (9, 73, 6):
raise RuntimeError(
f'Old version ({ibapi.__version__}) of ibapi package detected. '
'The newest version from http://interactivebrokers.github.io '
'is required')
from .version import __version__, __version_info__ # noqa
from .objects import * # noqa
from .event import * # noqa
from .contract import * # noqa
from .order import * # noqa
from .ticker import * # noqa
from .ib import * # noqa
from .client import * # noqa
from .wrapper import * # noqa
from .flexreport import * # noqa
from .ibcontroller import * # noqa
__all__ = ['util']
for _m in (
objects, event, contract, order, ticker, ib, # noqa
client, wrapper, flexreport, ibcontroller): # noqa
__all__ += _m.__all__
del sys
del importlib
del ibapi
| import sys
import importlib
from .version import __version__, __version_info__ # noqa
from . import util
if sys.version_info < (3, 6, 0):
raise RuntimeError('ib_insync requires Python 3.6 or higher')
try:
import ibapi
except ImportError:
raise RuntimeError(
'IB API from http://interactivebrokers.github.io is required')
if util.ibapiVersionInfo() < (9, 73, 6):
raise RuntimeError(
f'Old version ({ibapi.__version__}) of ibapi package detected. '
'The newest version from http://interactivebrokers.github.io '
'is required')
from .version import __version__, __version_info__ # noqa
from .objects import * # noqa
from .event import * # noqa
from .contract import * # noqa
from .order import * # noqa
from .ticker import * # noqa
from .ib import * # noqa
from .client import * # noqa
from .wrapper import * # noqa
from .flexreport import * # noqa
from .ibcontroller import * # noqa
from . import util # noqa
__all__ = ['util']
for _m in (
objects, event, contract, order, ticker, ib, # noqa
client, wrapper, flexreport, ibcontroller): # noqa
__all__ += _m.__all__
del sys
del importlib
del ibapi
Fix explicit check for presence of ibapi packageimport sys
import importlib
if sys.version_info < (3, 6, 0):
raise RuntimeError('ib_insync requires Python 3.6 or higher')
try:
import ibapi
except ImportError:
raise RuntimeError(
'IB API from http://interactivebrokers.github.io is required')
from . import util # noqa
if util.ibapiVersionInfo() < (9, 73, 6):
raise RuntimeError(
f'Old version ({ibapi.__version__}) of ibapi package detected. '
'The newest version from http://interactivebrokers.github.io '
'is required')
from .version import __version__, __version_info__ # noqa
from .objects import * # noqa
from .event import * # noqa
from .contract import * # noqa
from .order import * # noqa
from .ticker import * # noqa
from .ib import * # noqa
from .client import * # noqa
from .wrapper import * # noqa
from .flexreport import * # noqa
from .ibcontroller import * # noqa
__all__ = ['util']
for _m in (
objects, event, contract, order, ticker, ib, # noqa
client, wrapper, flexreport, ibcontroller): # noqa
__all__ += _m.__all__
del sys
del importlib
del ibapi
| <commit_before>import sys
import importlib
from .version import __version__, __version_info__ # noqa
from . import util
if sys.version_info < (3, 6, 0):
raise RuntimeError('ib_insync requires Python 3.6 or higher')
try:
import ibapi
except ImportError:
raise RuntimeError(
'IB API from http://interactivebrokers.github.io is required')
if util.ibapiVersionInfo() < (9, 73, 6):
raise RuntimeError(
f'Old version ({ibapi.__version__}) of ibapi package detected. '
'The newest version from http://interactivebrokers.github.io '
'is required')
from .version import __version__, __version_info__ # noqa
from .objects import * # noqa
from .event import * # noqa
from .contract import * # noqa
from .order import * # noqa
from .ticker import * # noqa
from .ib import * # noqa
from .client import * # noqa
from .wrapper import * # noqa
from .flexreport import * # noqa
from .ibcontroller import * # noqa
from . import util # noqa
__all__ = ['util']
for _m in (
objects, event, contract, order, ticker, ib, # noqa
client, wrapper, flexreport, ibcontroller): # noqa
__all__ += _m.__all__
del sys
del importlib
del ibapi
<commit_msg>Fix explicit check for presence of ibapi package<commit_after>import sys
import importlib
if sys.version_info < (3, 6, 0):
raise RuntimeError('ib_insync requires Python 3.6 or higher')
try:
import ibapi
except ImportError:
raise RuntimeError(
'IB API from http://interactivebrokers.github.io is required')
from . import util # noqa
if util.ibapiVersionInfo() < (9, 73, 6):
raise RuntimeError(
f'Old version ({ibapi.__version__}) of ibapi package detected. '
'The newest version from http://interactivebrokers.github.io '
'is required')
from .version import __version__, __version_info__ # noqa
from .objects import * # noqa
from .event import * # noqa
from .contract import * # noqa
from .order import * # noqa
from .ticker import * # noqa
from .ib import * # noqa
from .client import * # noqa
from .wrapper import * # noqa
from .flexreport import * # noqa
from .ibcontroller import * # noqa
__all__ = ['util']
for _m in (
objects, event, contract, order, ticker, ib, # noqa
client, wrapper, flexreport, ibcontroller): # noqa
__all__ += _m.__all__
del sys
del importlib
del ibapi
|
50395649bf65b2ee7ef31fc8fedf62b9a66e96bf | metaci/cumulusci/utils.py | metaci/cumulusci/utils.py | from cumulusci.core.config import ConnectedAppOAuthConfig
from django.conf import settings
def get_connected_app():
return ConnectedAppOAuthConfig({
'callback_url': settings.CONNECTED_APP_CALLBACK_URL,
'client_id': settings.CONNECTED_APP_CLIENT_ID,
'client_secret': settings.CONNECTED_APP_CLIENT_SECRET,
})
| from cumulusci.core.config import ConnectedAppOAuthConfig
from django.conf import settings
def get_connected_app():
return ConnectedAppOAuthConfig({
'callback_url': settings.CONNECTED_APP_CALLBACK_URL,
'client_id': settings.CONNECTED_APP_CLIENT_ID,
'client_secret': settings.CONNECTED_APP_CLIENT_SECRET,
})
| Revert file to original state | Revert file to original state | Python | bsd-3-clause | SalesforceFoundation/mrbelvedereci,SalesforceFoundation/mrbelvedereci,SalesforceFoundation/mrbelvedereci,SalesforceFoundation/mrbelvedereci | from cumulusci.core.config import ConnectedAppOAuthConfig
from django.conf import settings
def get_connected_app():
return ConnectedAppOAuthConfig({
'callback_url': settings.CONNECTED_APP_CALLBACK_URL,
'client_id': settings.CONNECTED_APP_CLIENT_ID,
'client_secret': settings.CONNECTED_APP_CLIENT_SECRET,
})
Revert file to original state | from cumulusci.core.config import ConnectedAppOAuthConfig
from django.conf import settings
def get_connected_app():
return ConnectedAppOAuthConfig({
'callback_url': settings.CONNECTED_APP_CALLBACK_URL,
'client_id': settings.CONNECTED_APP_CLIENT_ID,
'client_secret': settings.CONNECTED_APP_CLIENT_SECRET,
})
| <commit_before>from cumulusci.core.config import ConnectedAppOAuthConfig
from django.conf import settings
def get_connected_app():
return ConnectedAppOAuthConfig({
'callback_url': settings.CONNECTED_APP_CALLBACK_URL,
'client_id': settings.CONNECTED_APP_CLIENT_ID,
'client_secret': settings.CONNECTED_APP_CLIENT_SECRET,
})
<commit_msg>Revert file to original state<commit_after> | from cumulusci.core.config import ConnectedAppOAuthConfig
from django.conf import settings
def get_connected_app():
return ConnectedAppOAuthConfig({
'callback_url': settings.CONNECTED_APP_CALLBACK_URL,
'client_id': settings.CONNECTED_APP_CLIENT_ID,
'client_secret': settings.CONNECTED_APP_CLIENT_SECRET,
})
| from cumulusci.core.config import ConnectedAppOAuthConfig
from django.conf import settings
def get_connected_app():
return ConnectedAppOAuthConfig({
'callback_url': settings.CONNECTED_APP_CALLBACK_URL,
'client_id': settings.CONNECTED_APP_CLIENT_ID,
'client_secret': settings.CONNECTED_APP_CLIENT_SECRET,
})
Revert file to original statefrom cumulusci.core.config import ConnectedAppOAuthConfig
from django.conf import settings
def get_connected_app():
return ConnectedAppOAuthConfig({
'callback_url': settings.CONNECTED_APP_CALLBACK_URL,
'client_id': settings.CONNECTED_APP_CLIENT_ID,
'client_secret': settings.CONNECTED_APP_CLIENT_SECRET,
})
| <commit_before>from cumulusci.core.config import ConnectedAppOAuthConfig
from django.conf import settings
def get_connected_app():
return ConnectedAppOAuthConfig({
'callback_url': settings.CONNECTED_APP_CALLBACK_URL,
'client_id': settings.CONNECTED_APP_CLIENT_ID,
'client_secret': settings.CONNECTED_APP_CLIENT_SECRET,
})
<commit_msg>Revert file to original state<commit_after>from cumulusci.core.config import ConnectedAppOAuthConfig
from django.conf import settings
def get_connected_app():
return ConnectedAppOAuthConfig({
'callback_url': settings.CONNECTED_APP_CALLBACK_URL,
'client_id': settings.CONNECTED_APP_CLIENT_ID,
'client_secret': settings.CONNECTED_APP_CLIENT_SECRET,
})
|
525c224080b3ac13864fbd3b5b9db2e884691edf | polyaxon/sidecar/sidecar/sidecar/monitor.py | polyaxon/sidecar/sidecar/sidecar/monitor.py | import ocular
from polyaxon_schemas.pod import PodLifeCycle
def is_container_terminated(event, container_id):
statuses_by_name = ocular.processor.get_container_statuses_by_name(
event.status.to_dict().get('container_statuses', []))
statuses = ocular.processor.get_container_status(statuses_by_name, (container_id,))
return statuses.get('state', {}).get('terminated')
def is_pod_running(k8s_manager, pod_id, container_id):
event = k8s_manager.k8s_api.read_namespaced_pod_status(pod_id, k8s_manager.namespace)
is_terminated = is_container_terminated(event=event, container_id=container_id)
return (
event.status.phase in {PodLifeCycle.RUNNING,
PodLifeCycle.PENDING,
PodLifeCycle.CONTAINER_CREATING}and
not is_terminated
)
| import ocular
from polyaxon_schemas.pod import PodLifeCycle
def is_container_terminated(event, container_id):
container_statuses = event.status.to_dict().get('container_statuses') or []
statuses_by_name = ocular.processor.get_container_statuses_by_name(container_statuses)
statuses = ocular.processor.get_container_status(statuses_by_name, (container_id,))
statuses = statuses or {}
return statuses.get('state', {}).get('terminated')
def is_pod_running(k8s_manager, pod_id, container_id):
event = k8s_manager.k8s_api.read_namespaced_pod_status(pod_id, k8s_manager.namespace)
is_terminated = is_container_terminated(event=event, container_id=container_id)
return (
event.status.phase in {PodLifeCycle.RUNNING,
PodLifeCycle.PENDING,
PodLifeCycle.CONTAINER_CREATING} and
not is_terminated
)
| Fix sidecar check for terminated containers | Fix sidecar check for terminated containers
| Python | apache-2.0 | polyaxon/polyaxon,polyaxon/polyaxon,polyaxon/polyaxon | import ocular
from polyaxon_schemas.pod import PodLifeCycle
def is_container_terminated(event, container_id):
statuses_by_name = ocular.processor.get_container_statuses_by_name(
event.status.to_dict().get('container_statuses', []))
statuses = ocular.processor.get_container_status(statuses_by_name, (container_id,))
return statuses.get('state', {}).get('terminated')
def is_pod_running(k8s_manager, pod_id, container_id):
event = k8s_manager.k8s_api.read_namespaced_pod_status(pod_id, k8s_manager.namespace)
is_terminated = is_container_terminated(event=event, container_id=container_id)
return (
event.status.phase in {PodLifeCycle.RUNNING,
PodLifeCycle.PENDING,
PodLifeCycle.CONTAINER_CREATING}and
not is_terminated
)
Fix sidecar check for terminated containers | import ocular
from polyaxon_schemas.pod import PodLifeCycle
def is_container_terminated(event, container_id):
container_statuses = event.status.to_dict().get('container_statuses') or []
statuses_by_name = ocular.processor.get_container_statuses_by_name(container_statuses)
statuses = ocular.processor.get_container_status(statuses_by_name, (container_id,))
statuses = statuses or {}
return statuses.get('state', {}).get('terminated')
def is_pod_running(k8s_manager, pod_id, container_id):
event = k8s_manager.k8s_api.read_namespaced_pod_status(pod_id, k8s_manager.namespace)
is_terminated = is_container_terminated(event=event, container_id=container_id)
return (
event.status.phase in {PodLifeCycle.RUNNING,
PodLifeCycle.PENDING,
PodLifeCycle.CONTAINER_CREATING} and
not is_terminated
)
| <commit_before>import ocular
from polyaxon_schemas.pod import PodLifeCycle
def is_container_terminated(event, container_id):
statuses_by_name = ocular.processor.get_container_statuses_by_name(
event.status.to_dict().get('container_statuses', []))
statuses = ocular.processor.get_container_status(statuses_by_name, (container_id,))
return statuses.get('state', {}).get('terminated')
def is_pod_running(k8s_manager, pod_id, container_id):
event = k8s_manager.k8s_api.read_namespaced_pod_status(pod_id, k8s_manager.namespace)
is_terminated = is_container_terminated(event=event, container_id=container_id)
return (
event.status.phase in {PodLifeCycle.RUNNING,
PodLifeCycle.PENDING,
PodLifeCycle.CONTAINER_CREATING}and
not is_terminated
)
<commit_msg>Fix sidecar check for terminated containers<commit_after> | import ocular
from polyaxon_schemas.pod import PodLifeCycle
def is_container_terminated(event, container_id):
container_statuses = event.status.to_dict().get('container_statuses') or []
statuses_by_name = ocular.processor.get_container_statuses_by_name(container_statuses)
statuses = ocular.processor.get_container_status(statuses_by_name, (container_id,))
statuses = statuses or {}
return statuses.get('state', {}).get('terminated')
def is_pod_running(k8s_manager, pod_id, container_id):
event = k8s_manager.k8s_api.read_namespaced_pod_status(pod_id, k8s_manager.namespace)
is_terminated = is_container_terminated(event=event, container_id=container_id)
return (
event.status.phase in {PodLifeCycle.RUNNING,
PodLifeCycle.PENDING,
PodLifeCycle.CONTAINER_CREATING} and
not is_terminated
)
| import ocular
from polyaxon_schemas.pod import PodLifeCycle
def is_container_terminated(event, container_id):
statuses_by_name = ocular.processor.get_container_statuses_by_name(
event.status.to_dict().get('container_statuses', []))
statuses = ocular.processor.get_container_status(statuses_by_name, (container_id,))
return statuses.get('state', {}).get('terminated')
def is_pod_running(k8s_manager, pod_id, container_id):
event = k8s_manager.k8s_api.read_namespaced_pod_status(pod_id, k8s_manager.namespace)
is_terminated = is_container_terminated(event=event, container_id=container_id)
return (
event.status.phase in {PodLifeCycle.RUNNING,
PodLifeCycle.PENDING,
PodLifeCycle.CONTAINER_CREATING}and
not is_terminated
)
Fix sidecar check for terminated containersimport ocular
from polyaxon_schemas.pod import PodLifeCycle
def is_container_terminated(event, container_id):
container_statuses = event.status.to_dict().get('container_statuses') or []
statuses_by_name = ocular.processor.get_container_statuses_by_name(container_statuses)
statuses = ocular.processor.get_container_status(statuses_by_name, (container_id,))
statuses = statuses or {}
return statuses.get('state', {}).get('terminated')
def is_pod_running(k8s_manager, pod_id, container_id):
event = k8s_manager.k8s_api.read_namespaced_pod_status(pod_id, k8s_manager.namespace)
is_terminated = is_container_terminated(event=event, container_id=container_id)
return (
event.status.phase in {PodLifeCycle.RUNNING,
PodLifeCycle.PENDING,
PodLifeCycle.CONTAINER_CREATING} and
not is_terminated
)
| <commit_before>import ocular
from polyaxon_schemas.pod import PodLifeCycle
def is_container_terminated(event, container_id):
statuses_by_name = ocular.processor.get_container_statuses_by_name(
event.status.to_dict().get('container_statuses', []))
statuses = ocular.processor.get_container_status(statuses_by_name, (container_id,))
return statuses.get('state', {}).get('terminated')
def is_pod_running(k8s_manager, pod_id, container_id):
event = k8s_manager.k8s_api.read_namespaced_pod_status(pod_id, k8s_manager.namespace)
is_terminated = is_container_terminated(event=event, container_id=container_id)
return (
event.status.phase in {PodLifeCycle.RUNNING,
PodLifeCycle.PENDING,
PodLifeCycle.CONTAINER_CREATING}and
not is_terminated
)
<commit_msg>Fix sidecar check for terminated containers<commit_after>import ocular
from polyaxon_schemas.pod import PodLifeCycle
def is_container_terminated(event, container_id):
container_statuses = event.status.to_dict().get('container_statuses') or []
statuses_by_name = ocular.processor.get_container_statuses_by_name(container_statuses)
statuses = ocular.processor.get_container_status(statuses_by_name, (container_id,))
statuses = statuses or {}
return statuses.get('state', {}).get('terminated')
def is_pod_running(k8s_manager, pod_id, container_id):
event = k8s_manager.k8s_api.read_namespaced_pod_status(pod_id, k8s_manager.namespace)
is_terminated = is_container_terminated(event=event, container_id=container_id)
return (
event.status.phase in {PodLifeCycle.RUNNING,
PodLifeCycle.PENDING,
PodLifeCycle.CONTAINER_CREATING} and
not is_terminated
)
|
d387ab6335ba73a0ecbc1ffa55e9b35ff119bd58 | journal/views.py | journal/views.py | import json
from django.utils.decorators import method_decorator
from django.views.decorators.csrf import csrf_exempt
from django.views.generic import View
from django.http import JsonResponse
from .models import Entry
@method_decorator(csrf_exempt, name='dispatch')
class RestView(View):
def get(self, request):
last = request.GET.get('last', None)
if last is None:
entries = Entry.objects.all()
else:
entries = Entry.objects.filter(id__gt=last)
ret = map(lambda x: {'id': x.id, 'content': x.content.decode()},
entries)
return JsonResponse({'entries': list(ret)})
@csrf_exempt
def put(self, request):
entries = json.loads(request.body.decode())
for entry in entries:
Entry(content=entry.encode()).save()
res = JsonResponse({'ok': 1})
res.status_code = 201
return res
| import json
from django.utils.decorators import method_decorator
from django.views.decorators.csrf import csrf_exempt
from django.views.generic import View
from django.http import JsonResponse
from .models import Entry
@method_decorator(csrf_exempt, name='dispatch')
class RestView(View):
def get(self, request):
last = request.GET.get('last', None)
if last is None:
entries = Entry.objects.all()
else:
entries = Entry.objects.filter(id__gt=last)
ret = map(lambda x: {'id': x.id, 'content': x.content.decode()},
entries)
return JsonResponse({'entries': list(ret)})
@csrf_exempt
def put(self, request):
entries = json.loads(request.body.decode())
for entry in entries['entries']:
Entry(content=entry['content'].encode()).save()
res = JsonResponse({'ok': 1})
res.status_code = 201
return res
| Update the protocol to mirror the return result of get. | Update the protocol to mirror the return result of get.
| Python | agpl-3.0 | etesync/journal-manager | import json
from django.utils.decorators import method_decorator
from django.views.decorators.csrf import csrf_exempt
from django.views.generic import View
from django.http import JsonResponse
from .models import Entry
@method_decorator(csrf_exempt, name='dispatch')
class RestView(View):
def get(self, request):
last = request.GET.get('last', None)
if last is None:
entries = Entry.objects.all()
else:
entries = Entry.objects.filter(id__gt=last)
ret = map(lambda x: {'id': x.id, 'content': x.content.decode()},
entries)
return JsonResponse({'entries': list(ret)})
@csrf_exempt
def put(self, request):
entries = json.loads(request.body.decode())
for entry in entries:
Entry(content=entry.encode()).save()
res = JsonResponse({'ok': 1})
res.status_code = 201
return res
Update the protocol to mirror the return result of get. | import json
from django.utils.decorators import method_decorator
from django.views.decorators.csrf import csrf_exempt
from django.views.generic import View
from django.http import JsonResponse
from .models import Entry
@method_decorator(csrf_exempt, name='dispatch')
class RestView(View):
def get(self, request):
last = request.GET.get('last', None)
if last is None:
entries = Entry.objects.all()
else:
entries = Entry.objects.filter(id__gt=last)
ret = map(lambda x: {'id': x.id, 'content': x.content.decode()},
entries)
return JsonResponse({'entries': list(ret)})
@csrf_exempt
def put(self, request):
entries = json.loads(request.body.decode())
for entry in entries['entries']:
Entry(content=entry['content'].encode()).save()
res = JsonResponse({'ok': 1})
res.status_code = 201
return res
| <commit_before>import json
from django.utils.decorators import method_decorator
from django.views.decorators.csrf import csrf_exempt
from django.views.generic import View
from django.http import JsonResponse
from .models import Entry
@method_decorator(csrf_exempt, name='dispatch')
class RestView(View):
def get(self, request):
last = request.GET.get('last', None)
if last is None:
entries = Entry.objects.all()
else:
entries = Entry.objects.filter(id__gt=last)
ret = map(lambda x: {'id': x.id, 'content': x.content.decode()},
entries)
return JsonResponse({'entries': list(ret)})
@csrf_exempt
def put(self, request):
entries = json.loads(request.body.decode())
for entry in entries:
Entry(content=entry.encode()).save()
res = JsonResponse({'ok': 1})
res.status_code = 201
return res
<commit_msg>Update the protocol to mirror the return result of get.<commit_after> | import json
from django.utils.decorators import method_decorator
from django.views.decorators.csrf import csrf_exempt
from django.views.generic import View
from django.http import JsonResponse
from .models import Entry
@method_decorator(csrf_exempt, name='dispatch')
class RestView(View):
def get(self, request):
last = request.GET.get('last', None)
if last is None:
entries = Entry.objects.all()
else:
entries = Entry.objects.filter(id__gt=last)
ret = map(lambda x: {'id': x.id, 'content': x.content.decode()},
entries)
return JsonResponse({'entries': list(ret)})
@csrf_exempt
def put(self, request):
entries = json.loads(request.body.decode())
for entry in entries['entries']:
Entry(content=entry['content'].encode()).save()
res = JsonResponse({'ok': 1})
res.status_code = 201
return res
| import json
from django.utils.decorators import method_decorator
from django.views.decorators.csrf import csrf_exempt
from django.views.generic import View
from django.http import JsonResponse
from .models import Entry
@method_decorator(csrf_exempt, name='dispatch')
class RestView(View):
def get(self, request):
last = request.GET.get('last', None)
if last is None:
entries = Entry.objects.all()
else:
entries = Entry.objects.filter(id__gt=last)
ret = map(lambda x: {'id': x.id, 'content': x.content.decode()},
entries)
return JsonResponse({'entries': list(ret)})
@csrf_exempt
def put(self, request):
entries = json.loads(request.body.decode())
for entry in entries:
Entry(content=entry.encode()).save()
res = JsonResponse({'ok': 1})
res.status_code = 201
return res
Update the protocol to mirror the return result of get.import json
from django.utils.decorators import method_decorator
from django.views.decorators.csrf import csrf_exempt
from django.views.generic import View
from django.http import JsonResponse
from .models import Entry
@method_decorator(csrf_exempt, name='dispatch')
class RestView(View):
def get(self, request):
last = request.GET.get('last', None)
if last is None:
entries = Entry.objects.all()
else:
entries = Entry.objects.filter(id__gt=last)
ret = map(lambda x: {'id': x.id, 'content': x.content.decode()},
entries)
return JsonResponse({'entries': list(ret)})
@csrf_exempt
def put(self, request):
entries = json.loads(request.body.decode())
for entry in entries['entries']:
Entry(content=entry['content'].encode()).save()
res = JsonResponse({'ok': 1})
res.status_code = 201
return res
| <commit_before>import json
from django.utils.decorators import method_decorator
from django.views.decorators.csrf import csrf_exempt
from django.views.generic import View
from django.http import JsonResponse
from .models import Entry
@method_decorator(csrf_exempt, name='dispatch')
class RestView(View):
def get(self, request):
last = request.GET.get('last', None)
if last is None:
entries = Entry.objects.all()
else:
entries = Entry.objects.filter(id__gt=last)
ret = map(lambda x: {'id': x.id, 'content': x.content.decode()},
entries)
return JsonResponse({'entries': list(ret)})
@csrf_exempt
def put(self, request):
entries = json.loads(request.body.decode())
for entry in entries:
Entry(content=entry.encode()).save()
res = JsonResponse({'ok': 1})
res.status_code = 201
return res
<commit_msg>Update the protocol to mirror the return result of get.<commit_after>import json
from django.utils.decorators import method_decorator
from django.views.decorators.csrf import csrf_exempt
from django.views.generic import View
from django.http import JsonResponse
from .models import Entry
@method_decorator(csrf_exempt, name='dispatch')
class RestView(View):
def get(self, request):
last = request.GET.get('last', None)
if last is None:
entries = Entry.objects.all()
else:
entries = Entry.objects.filter(id__gt=last)
ret = map(lambda x: {'id': x.id, 'content': x.content.decode()},
entries)
return JsonResponse({'entries': list(ret)})
@csrf_exempt
def put(self, request):
entries = json.loads(request.body.decode())
for entry in entries['entries']:
Entry(content=entry['content'].encode()).save()
res = JsonResponse({'ok': 1})
res.status_code = 201
return res
|
c87c4a972f0f2d4966142fa666a900112762ed76 | scipy/constants/tests/test_codata.py | scipy/constants/tests/test_codata.py |
import warnings
from scipy.constants import find
from numpy.testing import assert_equal
def test_find():
warnings.simplefilter('ignore', DeprecationWarning)
keys = find('weak mixing', disp=False)
assert_equal(keys, ['weak mixing angle'])
keys = find('qwertyuiop', disp=False)
assert_equal(keys, [])
keys = find('natural unit', disp=False)
assert_equal(keys, sorted(['natural unit of velocity',
'natural unit of action',
'natural unit of action in eV s',
'natural unit of mass',
'natural unit of energy',
'natural unit of energy in MeV',
'natural unit of momentum',
'natural unit of momentum in MeV/c',
'natural unit of length',
'natural unit of time']))
|
import warnings
from scipy.constants import find
from numpy.testing import assert_equal, run_module_suite
def test_find():
warnings.simplefilter('ignore', DeprecationWarning)
keys = find('weak mixing', disp=False)
assert_equal(keys, ['weak mixing angle'])
keys = find('qwertyuiop', disp=False)
assert_equal(keys, [])
keys = find('natural unit', disp=False)
assert_equal(keys, sorted(['natural unit of velocity',
'natural unit of action',
'natural unit of action in eV s',
'natural unit of mass',
'natural unit of energy',
'natural unit of energy in MeV',
'natural unit of momentum',
'natural unit of momentum in MeV/c',
'natural unit of length',
'natural unit of time']))
if __name__ == "__main__":
run_module_suite()
| Allow codata tests to be run as script. | ENH: Allow codata tests to be run as script.
| Python | bsd-3-clause | zerothi/scipy,zxsted/scipy,josephcslater/scipy,rgommers/scipy,grlee77/scipy,sargas/scipy,dch312/scipy,ilayn/scipy,apbard/scipy,jakevdp/scipy,niknow/scipy,vanpact/scipy,jakevdp/scipy,rmcgibbo/scipy,zxsted/scipy,pnedunuri/scipy,raoulbq/scipy,lhilt/scipy,mgaitan/scipy,mingwpy/scipy,maciejkula/scipy,njwilson23/scipy,Dapid/scipy,woodscn/scipy,perimosocordiae/scipy,aarchiba/scipy,pbrod/scipy,aarchiba/scipy,trankmichael/scipy,Srisai85/scipy,aarchiba/scipy,pschella/scipy,tylerjereddy/scipy,minhlongdo/scipy,mortonjt/scipy,dominicelse/scipy,jsilter/scipy,sargas/scipy,vanpact/scipy,anielsen001/scipy,mingwpy/scipy,matthew-brett/scipy,aeklant/scipy,mikebenfield/scipy,Srisai85/scipy,sauliusl/scipy,FRidh/scipy,andim/scipy,ortylp/scipy,josephcslater/scipy,mhogg/scipy,WarrenWeckesser/scipy,ndchorley/scipy,josephcslater/scipy,mortonjt/scipy,felipebetancur/scipy,jonycgn/scipy,mtrbean/scipy,juliantaylor/scipy,mdhaber/scipy,mdhaber/scipy,kleskjr/scipy,ogrisel/scipy,teoliphant/scipy,petebachant/scipy,lukauskas/scipy,Gillu13/scipy,sonnyhu/scipy,WillieMaddox/scipy,arokem/scipy,jonycgn/scipy,gdooper/scipy,ChanderG/scipy,pbrod/scipy,surhudm/scipy,rgommers/scipy,anntzer/scipy,gfyoung/scipy,sriki18/scipy,nmayorov/scipy,ogrisel/scipy,nvoron23/scipy,giorgiop/scipy,trankmichael/scipy,scipy/scipy,dominicelse/scipy,petebachant/scipy,ChanderG/scipy,ndchorley/scipy,scipy/scipy,Stefan-Endres/scipy,futurulus/scipy,giorgiop/scipy,behzadnouri/scipy,sargas/scipy,haudren/scipy,Gillu13/scipy,teoliphant/scipy,WillieMaddox/scipy,cpaulik/scipy,juliantaylor/scipy,sriki18/scipy,ortylp/scipy,zerothi/scipy,gdooper/scipy,befelix/scipy,raoulbq/scipy,sargas/scipy,newemailjdm/scipy,piyush0609/scipy,vhaasteren/scipy,sriki18/scipy,sauliusl/scipy,rgommers/scipy,zaxliu/scipy,Stefan-Endres/scipy,dch312/scipy,scipy/scipy,aeklant/scipy,cpaulik/scipy,maniteja123/scipy,zerothi/scipy,cpaulik/scipy,Eric89GXL/scipy,mgaitan/scipy,bkendzior/scipy,jonycgn/scipy,juliantaylor/scipy,felipebetancur/scipy,giorgiop/scipy,gertingold/scipy,befelix/scipy,maciejkula/scipy,vberaudi/scipy,lhilt/scipy,ChanderG/scipy,mgaitan/scipy,hainm/scipy,endolith/scipy,lukauskas/scipy,gdooper/scipy,kalvdans/scipy,richardotis/scipy,jor-/scipy,mgaitan/scipy,mtrbean/scipy,Gillu13/scipy,raoulbq/scipy,lukauskas/scipy,tylerjereddy/scipy,person142/scipy,mortada/scipy,vhaasteren/scipy,Shaswat27/scipy,vigna/scipy,pschella/scipy,surhudm/scipy,kalvdans/scipy,witcxc/scipy,piyush0609/scipy,woodscn/scipy,matthew-brett/scipy,nvoron23/scipy,matthewalbani/scipy,mdhaber/scipy,gfyoung/scipy,pnedunuri/scipy,vanpact/scipy,ilayn/scipy,FRidh/scipy,mingwpy/scipy,efiring/scipy,jakevdp/scipy,Shaswat27/scipy,matthewalbani/scipy,pizzathief/scipy,mdhaber/scipy,jonycgn/scipy,pyramania/scipy,apbard/scipy,Dapid/scipy,rmcgibbo/scipy,vhaasteren/scipy,niknow/scipy,apbard/scipy,sonnyhu/scipy,larsmans/scipy,bkendzior/scipy,perimosocordiae/scipy,maniteja123/scipy,mortada/scipy,felipebetancur/scipy,andyfaff/scipy,mikebenfield/scipy,jseabold/scipy,mtrbean/scipy,jamestwebber/scipy,anielsen001/scipy,ortylp/scipy,Gillu13/scipy,woodscn/scipy,witcxc/scipy,haudren/scipy,witcxc/scipy,andyfaff/scipy,newemailjdm/scipy,vhaasteren/scipy,njwilson23/scipy,grlee77/scipy,surhudm/scipy,zaxliu/scipy,ortylp/scipy,njwilson23/scipy,dch312/scipy,fredrikw/scipy,pnedunuri/scipy,mortada/scipy,mortada/scipy,fredrikw/scipy,gef756/scipy,zaxliu/scipy,chatcannon/scipy,matthew-brett/scipy,jsilter/scipy,ilayn/scipy,mortada/scipy,Stefan-Endres/scipy,jseabold/scipy,minhlongdo/scipy,anntzer/scipy,andyfaff/scipy,futurulus/scipy,sauliusl/scipy,e-q/scipy,mingwpy/scipy,rmcgibbo/scipy,Dapid/scipy,vigna/scipy,minhlongdo/scipy,mdhaber/scipy,raoulbq/scipy,pyramania/scipy,richardotis/scipy,jor-/scipy,befelix/scipy,sriki18/scipy,mortonjt/scipy,endolith/scipy,haudren/scipy,juliantaylor/scipy,hainm/scipy,e-q/scipy,pizzathief/scipy,endolith/scipy,FRidh/scipy,argriffing/scipy,perimosocordiae/scipy,tylerjereddy/scipy,fernand/scipy,tylerjereddy/scipy,argriffing/scipy,vigna/scipy,mhogg/scipy,vanpact/scipy,woodscn/scipy,Eric89GXL/scipy,vberaudi/scipy,ChanderG/scipy,raoulbq/scipy,gef756/scipy,sauliusl/scipy,anntzer/scipy,jseabold/scipy,pbrod/scipy,argriffing/scipy,nvoron23/scipy,ales-erjavec/scipy,ortylp/scipy,Newman101/scipy,jjhelmus/scipy,petebachant/scipy,mingwpy/scipy,chatcannon/scipy,Eric89GXL/scipy,Shaswat27/scipy,ogrisel/scipy,anntzer/scipy,andyfaff/scipy,aman-iitj/scipy,cpaulik/scipy,sauliusl/scipy,zerothi/scipy,hainm/scipy,Srisai85/scipy,scipy/scipy,mtrbean/scipy,fredrikw/scipy,endolith/scipy,Dapid/scipy,ndchorley/scipy,giorgiop/scipy,felipebetancur/scipy,Gillu13/scipy,maciejkula/scipy,jonycgn/scipy,vanpact/scipy,fernand/scipy,argriffing/scipy,aarchiba/scipy,jjhelmus/scipy,newemailjdm/scipy,matthewalbani/scipy,behzadnouri/scipy,jamestwebber/scipy,behzadnouri/scipy,andim/scipy,pschella/scipy,nvoron23/scipy,ndchorley/scipy,Shaswat27/scipy,argriffing/scipy,andyfaff/scipy,juliantaylor/scipy,felipebetancur/scipy,mtrbean/scipy,fredrikw/scipy,sonnyhu/scipy,efiring/scipy,e-q/scipy,jsilter/scipy,mtrbean/scipy,behzadnouri/scipy,grlee77/scipy,ChanderG/scipy,maciejkula/scipy,ogrisel/scipy,maciejkula/scipy,Srisai85/scipy,jamestwebber/scipy,woodscn/scipy,pschella/scipy,vigna/scipy,dch312/scipy,lhilt/scipy,haudren/scipy,nmayorov/scipy,futurulus/scipy,Newman101/scipy,larsmans/scipy,Eric89GXL/scipy,larsmans/scipy,maniteja123/scipy,maniteja123/scipy,gfyoung/scipy,fernand/scipy,sargas/scipy,nmayorov/scipy,lhilt/scipy,richardotis/scipy,vanpact/scipy,teoliphant/scipy,arokem/scipy,aeklant/scipy,pyramania/scipy,jamestwebber/scipy,vberaudi/scipy,dominicelse/scipy,efiring/scipy,rmcgibbo/scipy,ales-erjavec/scipy,Eric89GXL/scipy,niknow/scipy,zaxliu/scipy,surhudm/scipy,maniteja123/scipy,mikebenfield/scipy,gef756/scipy,vberaudi/scipy,efiring/scipy,sriki18/scipy,FRidh/scipy,gdooper/scipy,futurulus/scipy,cpaulik/scipy,WarrenWeckesser/scipy,larsmans/scipy,person142/scipy,teoliphant/scipy,kleskjr/scipy,befelix/scipy,arokem/scipy,Kamp9/scipy,kleskjr/scipy,teoliphant/scipy,kleskjr/scipy,dominicelse/scipy,ndchorley/scipy,anntzer/scipy,jjhelmus/scipy,pbrod/scipy,ilayn/scipy,Shaswat27/scipy,efiring/scipy,njwilson23/scipy,niknow/scipy,dominicelse/scipy,behzadnouri/scipy,mingwpy/scipy,rgommers/scipy,FRidh/scipy,person142/scipy,mortonjt/scipy,vhaasteren/scipy,pbrod/scipy,nvoron23/scipy,jor-/scipy,niknow/scipy,trankmichael/scipy,fernand/scipy,haudren/scipy,njwilson23/scipy,matthew-brett/scipy,nvoron23/scipy,endolith/scipy,e-q/scipy,jonycgn/scipy,dch312/scipy,WillieMaddox/scipy,rmcgibbo/scipy,lukauskas/scipy,andim/scipy,jakevdp/scipy,zerothi/scipy,witcxc/scipy,ales-erjavec/scipy,ales-erjavec/scipy,jor-/scipy,grlee77/scipy,nonhermitian/scipy,aman-iitj/scipy,vhaasteren/scipy,chatcannon/scipy,arokem/scipy,giorgiop/scipy,futurulus/scipy,e-q/scipy,ChanderG/scipy,ilayn/scipy,WarrenWeckesser/scipy,mikebenfield/scipy,Newman101/scipy,richardotis/scipy,jsilter/scipy,perimosocordiae/scipy,gfyoung/scipy,scipy/scipy,kalvdans/scipy,sonnyhu/scipy,lukauskas/scipy,WillieMaddox/scipy,aman-iitj/scipy,Newman101/scipy,zerothi/scipy,apbard/scipy,jsilter/scipy,hainm/scipy,Stefan-Endres/scipy,maniteja123/scipy,petebachant/scipy,efiring/scipy,mhogg/scipy,Dapid/scipy,nmayorov/scipy,anielsen001/scipy,Dapid/scipy,aeklant/scipy,nonhermitian/scipy,perimosocordiae/scipy,gertingold/scipy,chatcannon/scipy,zxsted/scipy,trankmichael/scipy,behzadnouri/scipy,WillieMaddox/scipy,newemailjdm/scipy,bkendzior/scipy,vberaudi/scipy,anielsen001/scipy,argriffing/scipy,minhlongdo/scipy,aman-iitj/scipy,futurulus/scipy,pizzathief/scipy,aeklant/scipy,anielsen001/scipy,cpaulik/scipy,njwilson23/scipy,zaxliu/scipy,surhudm/scipy,Kamp9/scipy,Eric89GXL/scipy,mhogg/scipy,scipy/scipy,gfyoung/scipy,befelix/scipy,Newman101/scipy,ales-erjavec/scipy,raoulbq/scipy,zaxliu/scipy,pyramania/scipy,larsmans/scipy,mortada/scipy,chatcannon/scipy,grlee77/scipy,person142/scipy,Kamp9/scipy,giorgiop/scipy,kalvdans/scipy,Stefan-Endres/scipy,mhogg/scipy,mortonjt/scipy,Kamp9/scipy,zxsted/scipy,hainm/scipy,lhilt/scipy,haudren/scipy,sauliusl/scipy,aman-iitj/scipy,aarchiba/scipy,kleskjr/scipy,andyfaff/scipy,gef756/scipy,WarrenWeckesser/scipy,fredrikw/scipy,gertingold/scipy,petebachant/scipy,petebachant/scipy,endolith/scipy,Newman101/scipy,sonnyhu/scipy,trankmichael/scipy,josephcslater/scipy,vigna/scipy,kleskjr/scipy,matthewalbani/scipy,ortylp/scipy,Gillu13/scipy,gdooper/scipy,Kamp9/scipy,larsmans/scipy,mikebenfield/scipy,matthewalbani/scipy,Shaswat27/scipy,lukauskas/scipy,pizzathief/scipy,josephcslater/scipy,pnedunuri/scipy,matthew-brett/scipy,chatcannon/scipy,bkendzior/scipy,ilayn/scipy,jjhelmus/scipy,perimosocordiae/scipy,apbard/scipy,ogrisel/scipy,anntzer/scipy,rmcgibbo/scipy,surhudm/scipy,mdhaber/scipy,bkendzior/scipy,jseabold/scipy,sriki18/scipy,arokem/scipy,kalvdans/scipy,Kamp9/scipy,Srisai85/scipy,pnedunuri/scipy,pyramania/scipy,andim/scipy,fernand/scipy,gef756/scipy,trankmichael/scipy,nonhermitian/scipy,pnedunuri/scipy,niknow/scipy,piyush0609/scipy,richardotis/scipy,jamestwebber/scipy,Stefan-Endres/scipy,woodscn/scipy,minhlongdo/scipy,fredrikw/scipy,newemailjdm/scipy,WarrenWeckesser/scipy,ales-erjavec/scipy,andim/scipy,richardotis/scipy,piyush0609/scipy,andim/scipy,jor-/scipy,FRidh/scipy,nonhermitian/scipy,tylerjereddy/scipy,jseabold/scipy,gertingold/scipy,pizzathief/scipy,jseabold/scipy,zxsted/scipy,rgommers/scipy,gertingold/scipy,WarrenWeckesser/scipy,jakevdp/scipy,mortonjt/scipy,hainm/scipy,jjhelmus/scipy,piyush0609/scipy,WillieMaddox/scipy,ndchorley/scipy,zxsted/scipy,mgaitan/scipy,person142/scipy,minhlongdo/scipy,fernand/scipy,gef756/scipy,pbrod/scipy,pschella/scipy,anielsen001/scipy,vberaudi/scipy,aman-iitj/scipy,mhogg/scipy,witcxc/scipy,sonnyhu/scipy,piyush0609/scipy,mgaitan/scipy,nmayorov/scipy,newemailjdm/scipy,felipebetancur/scipy,Srisai85/scipy,nonhermitian/scipy |
import warnings
from scipy.constants import find
from numpy.testing import assert_equal
def test_find():
warnings.simplefilter('ignore', DeprecationWarning)
keys = find('weak mixing', disp=False)
assert_equal(keys, ['weak mixing angle'])
keys = find('qwertyuiop', disp=False)
assert_equal(keys, [])
keys = find('natural unit', disp=False)
assert_equal(keys, sorted(['natural unit of velocity',
'natural unit of action',
'natural unit of action in eV s',
'natural unit of mass',
'natural unit of energy',
'natural unit of energy in MeV',
'natural unit of momentum',
'natural unit of momentum in MeV/c',
'natural unit of length',
'natural unit of time']))
ENH: Allow codata tests to be run as script. |
import warnings
from scipy.constants import find
from numpy.testing import assert_equal, run_module_suite
def test_find():
warnings.simplefilter('ignore', DeprecationWarning)
keys = find('weak mixing', disp=False)
assert_equal(keys, ['weak mixing angle'])
keys = find('qwertyuiop', disp=False)
assert_equal(keys, [])
keys = find('natural unit', disp=False)
assert_equal(keys, sorted(['natural unit of velocity',
'natural unit of action',
'natural unit of action in eV s',
'natural unit of mass',
'natural unit of energy',
'natural unit of energy in MeV',
'natural unit of momentum',
'natural unit of momentum in MeV/c',
'natural unit of length',
'natural unit of time']))
if __name__ == "__main__":
run_module_suite()
| <commit_before>
import warnings
from scipy.constants import find
from numpy.testing import assert_equal
def test_find():
warnings.simplefilter('ignore', DeprecationWarning)
keys = find('weak mixing', disp=False)
assert_equal(keys, ['weak mixing angle'])
keys = find('qwertyuiop', disp=False)
assert_equal(keys, [])
keys = find('natural unit', disp=False)
assert_equal(keys, sorted(['natural unit of velocity',
'natural unit of action',
'natural unit of action in eV s',
'natural unit of mass',
'natural unit of energy',
'natural unit of energy in MeV',
'natural unit of momentum',
'natural unit of momentum in MeV/c',
'natural unit of length',
'natural unit of time']))
<commit_msg>ENH: Allow codata tests to be run as script.<commit_after> |
import warnings
from scipy.constants import find
from numpy.testing import assert_equal, run_module_suite
def test_find():
warnings.simplefilter('ignore', DeprecationWarning)
keys = find('weak mixing', disp=False)
assert_equal(keys, ['weak mixing angle'])
keys = find('qwertyuiop', disp=False)
assert_equal(keys, [])
keys = find('natural unit', disp=False)
assert_equal(keys, sorted(['natural unit of velocity',
'natural unit of action',
'natural unit of action in eV s',
'natural unit of mass',
'natural unit of energy',
'natural unit of energy in MeV',
'natural unit of momentum',
'natural unit of momentum in MeV/c',
'natural unit of length',
'natural unit of time']))
if __name__ == "__main__":
run_module_suite()
|
import warnings
from scipy.constants import find
from numpy.testing import assert_equal
def test_find():
warnings.simplefilter('ignore', DeprecationWarning)
keys = find('weak mixing', disp=False)
assert_equal(keys, ['weak mixing angle'])
keys = find('qwertyuiop', disp=False)
assert_equal(keys, [])
keys = find('natural unit', disp=False)
assert_equal(keys, sorted(['natural unit of velocity',
'natural unit of action',
'natural unit of action in eV s',
'natural unit of mass',
'natural unit of energy',
'natural unit of energy in MeV',
'natural unit of momentum',
'natural unit of momentum in MeV/c',
'natural unit of length',
'natural unit of time']))
ENH: Allow codata tests to be run as script.
import warnings
from scipy.constants import find
from numpy.testing import assert_equal, run_module_suite
def test_find():
warnings.simplefilter('ignore', DeprecationWarning)
keys = find('weak mixing', disp=False)
assert_equal(keys, ['weak mixing angle'])
keys = find('qwertyuiop', disp=False)
assert_equal(keys, [])
keys = find('natural unit', disp=False)
assert_equal(keys, sorted(['natural unit of velocity',
'natural unit of action',
'natural unit of action in eV s',
'natural unit of mass',
'natural unit of energy',
'natural unit of energy in MeV',
'natural unit of momentum',
'natural unit of momentum in MeV/c',
'natural unit of length',
'natural unit of time']))
if __name__ == "__main__":
run_module_suite()
| <commit_before>
import warnings
from scipy.constants import find
from numpy.testing import assert_equal
def test_find():
warnings.simplefilter('ignore', DeprecationWarning)
keys = find('weak mixing', disp=False)
assert_equal(keys, ['weak mixing angle'])
keys = find('qwertyuiop', disp=False)
assert_equal(keys, [])
keys = find('natural unit', disp=False)
assert_equal(keys, sorted(['natural unit of velocity',
'natural unit of action',
'natural unit of action in eV s',
'natural unit of mass',
'natural unit of energy',
'natural unit of energy in MeV',
'natural unit of momentum',
'natural unit of momentum in MeV/c',
'natural unit of length',
'natural unit of time']))
<commit_msg>ENH: Allow codata tests to be run as script.<commit_after>
import warnings
from scipy.constants import find
from numpy.testing import assert_equal, run_module_suite
def test_find():
warnings.simplefilter('ignore', DeprecationWarning)
keys = find('weak mixing', disp=False)
assert_equal(keys, ['weak mixing angle'])
keys = find('qwertyuiop', disp=False)
assert_equal(keys, [])
keys = find('natural unit', disp=False)
assert_equal(keys, sorted(['natural unit of velocity',
'natural unit of action',
'natural unit of action in eV s',
'natural unit of mass',
'natural unit of energy',
'natural unit of energy in MeV',
'natural unit of momentum',
'natural unit of momentum in MeV/c',
'natural unit of length',
'natural unit of time']))
if __name__ == "__main__":
run_module_suite()
|
3a2f4940ff83d3d2645505b82d1207a96f6d209e | linked-list/is-list-palindrome.py | linked-list/is-list-palindrome.py | # Given a singly linked list of integers, determine whether or not it's a palindrome
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
def is_list_palindrome(l):
if l.value is None:
return True
# find center of list
fast = l
slow = l
while fast.next and fast.next.next:
fast = fast.next.next
slow = slow.next
# reverse second half of list
p = slow.next
current_node = None
while p:
next = p.next
p.next = current_node
current_node = p
p = next
def create_nodes(l):
root = Node(-1)
current_node = root
for value in l:
current_node.next = Node(value)
current_node = current_node.next
return root.next
is_list_palindrome(create_nodes([1, 2, 3, 4]))
| # Given a singly linked list of integers, determine whether or not it's a palindrome
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
def is_list_palindrome(l):
if l is None or l.next is None:
return True
# find center of list
fast = l
slow = l
while fast.next and fast.next.next:
fast = fast.next.next
slow = slow.next
# reverse second half of list
p = slow.next
current_node = None
while p:
next = p.next
p.next = current_node
current_node = p
p = next
# check for palindrome
part_one = current_node
part_two = l
while part_one and part_one.value == part_two.value:
part_one = part_one.next
part_two = part_two.next
return part_one is None
def create_nodes(l):
root = Node(-1)
current_node = root
for value in l:
current_node.next = Node(value)
current_node = current_node.next
return root.next
print is_list_palindrome(create_nodes([1, 2, 3, 4]))
| Add check for palindrome component of method | Add check for palindrome component of method
| Python | mit | derekmpham/interview-prep,derekmpham/interview-prep | # Given a singly linked list of integers, determine whether or not it's a palindrome
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
def is_list_palindrome(l):
if l.value is None:
return True
# find center of list
fast = l
slow = l
while fast.next and fast.next.next:
fast = fast.next.next
slow = slow.next
# reverse second half of list
p = slow.next
current_node = None
while p:
next = p.next
p.next = current_node
current_node = p
p = next
def create_nodes(l):
root = Node(-1)
current_node = root
for value in l:
current_node.next = Node(value)
current_node = current_node.next
return root.next
is_list_palindrome(create_nodes([1, 2, 3, 4]))
Add check for palindrome component of method | # Given a singly linked list of integers, determine whether or not it's a palindrome
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
def is_list_palindrome(l):
if l is None or l.next is None:
return True
# find center of list
fast = l
slow = l
while fast.next and fast.next.next:
fast = fast.next.next
slow = slow.next
# reverse second half of list
p = slow.next
current_node = None
while p:
next = p.next
p.next = current_node
current_node = p
p = next
# check for palindrome
part_one = current_node
part_two = l
while part_one and part_one.value == part_two.value:
part_one = part_one.next
part_two = part_two.next
return part_one is None
def create_nodes(l):
root = Node(-1)
current_node = root
for value in l:
current_node.next = Node(value)
current_node = current_node.next
return root.next
print is_list_palindrome(create_nodes([1, 2, 3, 4]))
| <commit_before># Given a singly linked list of integers, determine whether or not it's a palindrome
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
def is_list_palindrome(l):
if l.value is None:
return True
# find center of list
fast = l
slow = l
while fast.next and fast.next.next:
fast = fast.next.next
slow = slow.next
# reverse second half of list
p = slow.next
current_node = None
while p:
next = p.next
p.next = current_node
current_node = p
p = next
def create_nodes(l):
root = Node(-1)
current_node = root
for value in l:
current_node.next = Node(value)
current_node = current_node.next
return root.next
is_list_palindrome(create_nodes([1, 2, 3, 4]))
<commit_msg>Add check for palindrome component of method<commit_after> | # Given a singly linked list of integers, determine whether or not it's a palindrome
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
def is_list_palindrome(l):
if l is None or l.next is None:
return True
# find center of list
fast = l
slow = l
while fast.next and fast.next.next:
fast = fast.next.next
slow = slow.next
# reverse second half of list
p = slow.next
current_node = None
while p:
next = p.next
p.next = current_node
current_node = p
p = next
# check for palindrome
part_one = current_node
part_two = l
while part_one and part_one.value == part_two.value:
part_one = part_one.next
part_two = part_two.next
return part_one is None
def create_nodes(l):
root = Node(-1)
current_node = root
for value in l:
current_node.next = Node(value)
current_node = current_node.next
return root.next
print is_list_palindrome(create_nodes([1, 2, 3, 4]))
| # Given a singly linked list of integers, determine whether or not it's a palindrome
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
def is_list_palindrome(l):
if l.value is None:
return True
# find center of list
fast = l
slow = l
while fast.next and fast.next.next:
fast = fast.next.next
slow = slow.next
# reverse second half of list
p = slow.next
current_node = None
while p:
next = p.next
p.next = current_node
current_node = p
p = next
def create_nodes(l):
root = Node(-1)
current_node = root
for value in l:
current_node.next = Node(value)
current_node = current_node.next
return root.next
is_list_palindrome(create_nodes([1, 2, 3, 4]))
Add check for palindrome component of method# Given a singly linked list of integers, determine whether or not it's a palindrome
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
def is_list_palindrome(l):
if l is None or l.next is None:
return True
# find center of list
fast = l
slow = l
while fast.next and fast.next.next:
fast = fast.next.next
slow = slow.next
# reverse second half of list
p = slow.next
current_node = None
while p:
next = p.next
p.next = current_node
current_node = p
p = next
# check for palindrome
part_one = current_node
part_two = l
while part_one and part_one.value == part_two.value:
part_one = part_one.next
part_two = part_two.next
return part_one is None
def create_nodes(l):
root = Node(-1)
current_node = root
for value in l:
current_node.next = Node(value)
current_node = current_node.next
return root.next
print is_list_palindrome(create_nodes([1, 2, 3, 4]))
| <commit_before># Given a singly linked list of integers, determine whether or not it's a palindrome
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
def is_list_palindrome(l):
if l.value is None:
return True
# find center of list
fast = l
slow = l
while fast.next and fast.next.next:
fast = fast.next.next
slow = slow.next
# reverse second half of list
p = slow.next
current_node = None
while p:
next = p.next
p.next = current_node
current_node = p
p = next
def create_nodes(l):
root = Node(-1)
current_node = root
for value in l:
current_node.next = Node(value)
current_node = current_node.next
return root.next
is_list_palindrome(create_nodes([1, 2, 3, 4]))
<commit_msg>Add check for palindrome component of method<commit_after># Given a singly linked list of integers, determine whether or not it's a palindrome
class Node(object): # define constructor
def __init__(self, value):
self.value = value
self.next = None
def is_list_palindrome(l):
if l is None or l.next is None:
return True
# find center of list
fast = l
slow = l
while fast.next and fast.next.next:
fast = fast.next.next
slow = slow.next
# reverse second half of list
p = slow.next
current_node = None
while p:
next = p.next
p.next = current_node
current_node = p
p = next
# check for palindrome
part_one = current_node
part_two = l
while part_one and part_one.value == part_two.value:
part_one = part_one.next
part_two = part_two.next
return part_one is None
def create_nodes(l):
root = Node(-1)
current_node = root
for value in l:
current_node.next = Node(value)
current_node = current_node.next
return root.next
print is_list_palindrome(create_nodes([1, 2, 3, 4]))
|
b43dfa19979dc74efb27e56771535b102547e792 | utils.py | utils.py | import sqlite3
import shelve
def connect_db(name):
"""
Open a connection to the database used to store quotes.
:param name: (str) Name of database file
:return: (shelve.DbfilenameShelf)
"""
try:
return shelve.open(name)
except Exception:
raise Exception('Unable to connect to database with name {}'.format(name))
class DBClient(object):
"""Client for interacting with database for the application"""
def __init__(self, database_name):
self.conn = sqlite3.connect(database_name)
self._create_quotes_table()
def _create_quotes_table(self):
"""
Create the table used for storing quotes if it does not exist already
"""
with self.conn:
self.conn.execute('''
CREATE TABLE IF NOT EXISTS quotes (
author TEXT
quote TEXT
created_at TEXT
)
''')
def close_connection(self):
"""
Close connection to the database
"""
self.conn.close()
| import sqlite3
import shelve
def connect_db(name):
"""
Open a connection to the database used to store quotes.
:param name: (str) Name of database file
:return: (shelve.DbfilenameShelf)
"""
try:
return shelve.open(name)
except Exception:
raise Exception('Unable to connect to database with name {}'.format(name))
class DBClient(object):
"""Client for interacting with database for the application"""
def __init__(self, database_name: str):
self.conn = sqlite3.connect(database_name)
self._create_quotes_table()
def _create_quotes_table(self):
"""
Create the table used for storing quotes if it does not exist already
"""
with self.conn:
self.conn.execute('''
CREATE TABLE IF NOT EXISTS quotes (
author TEXT,
quote TEXT,
created_at TEXT
);
''')
def close_connection(self):
"""
Close connection to the database
"""
self.conn.close()
def insert_quote(self, author: str, quote: str, created_at: str):
"""
Insert a quote into the database
:param author: (str) Name of the author that said the quote
:param quote: (str) The quote for the author
:param created_at: (str) Timestamp for when the quote was saved to database
"""
with self.conn:
self.conn.execute('''
INSERT INTO quotes VALUES (?, ?, ?)
''', (author, quote, created_at))
| Add method to insert quotes into database | Add method to insert quotes into database
Fix schema for quotes table
| Python | mit | nickdibari/Get-Quote | import sqlite3
import shelve
def connect_db(name):
"""
Open a connection to the database used to store quotes.
:param name: (str) Name of database file
:return: (shelve.DbfilenameShelf)
"""
try:
return shelve.open(name)
except Exception:
raise Exception('Unable to connect to database with name {}'.format(name))
class DBClient(object):
"""Client for interacting with database for the application"""
def __init__(self, database_name):
self.conn = sqlite3.connect(database_name)
self._create_quotes_table()
def _create_quotes_table(self):
"""
Create the table used for storing quotes if it does not exist already
"""
with self.conn:
self.conn.execute('''
CREATE TABLE IF NOT EXISTS quotes (
author TEXT
quote TEXT
created_at TEXT
)
''')
def close_connection(self):
"""
Close connection to the database
"""
self.conn.close()
Add method to insert quotes into database
Fix schema for quotes table | import sqlite3
import shelve
def connect_db(name):
"""
Open a connection to the database used to store quotes.
:param name: (str) Name of database file
:return: (shelve.DbfilenameShelf)
"""
try:
return shelve.open(name)
except Exception:
raise Exception('Unable to connect to database with name {}'.format(name))
class DBClient(object):
"""Client for interacting with database for the application"""
def __init__(self, database_name: str):
self.conn = sqlite3.connect(database_name)
self._create_quotes_table()
def _create_quotes_table(self):
"""
Create the table used for storing quotes if it does not exist already
"""
with self.conn:
self.conn.execute('''
CREATE TABLE IF NOT EXISTS quotes (
author TEXT,
quote TEXT,
created_at TEXT
);
''')
def close_connection(self):
"""
Close connection to the database
"""
self.conn.close()
def insert_quote(self, author: str, quote: str, created_at: str):
"""
Insert a quote into the database
:param author: (str) Name of the author that said the quote
:param quote: (str) The quote for the author
:param created_at: (str) Timestamp for when the quote was saved to database
"""
with self.conn:
self.conn.execute('''
INSERT INTO quotes VALUES (?, ?, ?)
''', (author, quote, created_at))
| <commit_before>import sqlite3
import shelve
def connect_db(name):
"""
Open a connection to the database used to store quotes.
:param name: (str) Name of database file
:return: (shelve.DbfilenameShelf)
"""
try:
return shelve.open(name)
except Exception:
raise Exception('Unable to connect to database with name {}'.format(name))
class DBClient(object):
"""Client for interacting with database for the application"""
def __init__(self, database_name):
self.conn = sqlite3.connect(database_name)
self._create_quotes_table()
def _create_quotes_table(self):
"""
Create the table used for storing quotes if it does not exist already
"""
with self.conn:
self.conn.execute('''
CREATE TABLE IF NOT EXISTS quotes (
author TEXT
quote TEXT
created_at TEXT
)
''')
def close_connection(self):
"""
Close connection to the database
"""
self.conn.close()
<commit_msg>Add method to insert quotes into database
Fix schema for quotes table<commit_after> | import sqlite3
import shelve
def connect_db(name):
"""
Open a connection to the database used to store quotes.
:param name: (str) Name of database file
:return: (shelve.DbfilenameShelf)
"""
try:
return shelve.open(name)
except Exception:
raise Exception('Unable to connect to database with name {}'.format(name))
class DBClient(object):
"""Client for interacting with database for the application"""
def __init__(self, database_name: str):
self.conn = sqlite3.connect(database_name)
self._create_quotes_table()
def _create_quotes_table(self):
"""
Create the table used for storing quotes if it does not exist already
"""
with self.conn:
self.conn.execute('''
CREATE TABLE IF NOT EXISTS quotes (
author TEXT,
quote TEXT,
created_at TEXT
);
''')
def close_connection(self):
"""
Close connection to the database
"""
self.conn.close()
def insert_quote(self, author: str, quote: str, created_at: str):
"""
Insert a quote into the database
:param author: (str) Name of the author that said the quote
:param quote: (str) The quote for the author
:param created_at: (str) Timestamp for when the quote was saved to database
"""
with self.conn:
self.conn.execute('''
INSERT INTO quotes VALUES (?, ?, ?)
''', (author, quote, created_at))
| import sqlite3
import shelve
def connect_db(name):
"""
Open a connection to the database used to store quotes.
:param name: (str) Name of database file
:return: (shelve.DbfilenameShelf)
"""
try:
return shelve.open(name)
except Exception:
raise Exception('Unable to connect to database with name {}'.format(name))
class DBClient(object):
"""Client for interacting with database for the application"""
def __init__(self, database_name):
self.conn = sqlite3.connect(database_name)
self._create_quotes_table()
def _create_quotes_table(self):
"""
Create the table used for storing quotes if it does not exist already
"""
with self.conn:
self.conn.execute('''
CREATE TABLE IF NOT EXISTS quotes (
author TEXT
quote TEXT
created_at TEXT
)
''')
def close_connection(self):
"""
Close connection to the database
"""
self.conn.close()
Add method to insert quotes into database
Fix schema for quotes tableimport sqlite3
import shelve
def connect_db(name):
"""
Open a connection to the database used to store quotes.
:param name: (str) Name of database file
:return: (shelve.DbfilenameShelf)
"""
try:
return shelve.open(name)
except Exception:
raise Exception('Unable to connect to database with name {}'.format(name))
class DBClient(object):
"""Client for interacting with database for the application"""
def __init__(self, database_name: str):
self.conn = sqlite3.connect(database_name)
self._create_quotes_table()
def _create_quotes_table(self):
"""
Create the table used for storing quotes if it does not exist already
"""
with self.conn:
self.conn.execute('''
CREATE TABLE IF NOT EXISTS quotes (
author TEXT,
quote TEXT,
created_at TEXT
);
''')
def close_connection(self):
"""
Close connection to the database
"""
self.conn.close()
def insert_quote(self, author: str, quote: str, created_at: str):
"""
Insert a quote into the database
:param author: (str) Name of the author that said the quote
:param quote: (str) The quote for the author
:param created_at: (str) Timestamp for when the quote was saved to database
"""
with self.conn:
self.conn.execute('''
INSERT INTO quotes VALUES (?, ?, ?)
''', (author, quote, created_at))
| <commit_before>import sqlite3
import shelve
def connect_db(name):
"""
Open a connection to the database used to store quotes.
:param name: (str) Name of database file
:return: (shelve.DbfilenameShelf)
"""
try:
return shelve.open(name)
except Exception:
raise Exception('Unable to connect to database with name {}'.format(name))
class DBClient(object):
"""Client for interacting with database for the application"""
def __init__(self, database_name):
self.conn = sqlite3.connect(database_name)
self._create_quotes_table()
def _create_quotes_table(self):
"""
Create the table used for storing quotes if it does not exist already
"""
with self.conn:
self.conn.execute('''
CREATE TABLE IF NOT EXISTS quotes (
author TEXT
quote TEXT
created_at TEXT
)
''')
def close_connection(self):
"""
Close connection to the database
"""
self.conn.close()
<commit_msg>Add method to insert quotes into database
Fix schema for quotes table<commit_after>import sqlite3
import shelve
def connect_db(name):
"""
Open a connection to the database used to store quotes.
:param name: (str) Name of database file
:return: (shelve.DbfilenameShelf)
"""
try:
return shelve.open(name)
except Exception:
raise Exception('Unable to connect to database with name {}'.format(name))
class DBClient(object):
"""Client for interacting with database for the application"""
def __init__(self, database_name: str):
self.conn = sqlite3.connect(database_name)
self._create_quotes_table()
def _create_quotes_table(self):
"""
Create the table used for storing quotes if it does not exist already
"""
with self.conn:
self.conn.execute('''
CREATE TABLE IF NOT EXISTS quotes (
author TEXT,
quote TEXT,
created_at TEXT
);
''')
def close_connection(self):
"""
Close connection to the database
"""
self.conn.close()
def insert_quote(self, author: str, quote: str, created_at: str):
"""
Insert a quote into the database
:param author: (str) Name of the author that said the quote
:param quote: (str) The quote for the author
:param created_at: (str) Timestamp for when the quote was saved to database
"""
with self.conn:
self.conn.execute('''
INSERT INTO quotes VALUES (?, ?, ?)
''', (author, quote, created_at))
|
44b3f1d2c3e5a31887454aa82b324f15898122bc | zazu/cli.py | zazu/cli.py | # -*- coding: utf-8 -*-
"""entry point for zazu"""
__author__ = "Nicholas Wiles"
__copyright__ = "Copyright 2016, Lily Robotics"
import click
import git_helper
import subprocess
import zazu.build
import zazu.config
import zazu.dev.commands
import zazu.repo.commands
import zazu.style
import zazu.upgrade
@click.group()
@click.version_option(version=zazu.__version__)
@click.pass_context
def cli(ctx):
try:
ctx.obj = zazu.config.Config(git_helper.get_root_path())
required_zazu_version = ctx.obj.zazu_version_required()
if required_zazu_version and required_zazu_version != zazu.__version__:
click.echo('Warning: this repo has requested zazu {}, which doesn\'t match the installed version ({}). \
Use "zazu upgrade" to fix this'.format(ctx.obj.zazu_version_required(), __version__))
except subprocess.CalledProcessError:
pass
cli.add_command(zazu.upgrade.upgrade)
cli.add_command(zazu.style.style)
cli.add_command(zazu.build.build)
cli.add_command(zazu.dev.commands.dev)
cli.add_command(zazu.repo.commands.repo)
| # -*- coding: utf-8 -*-
"""entry point for zazu"""
__author__ = "Nicholas Wiles"
__copyright__ = "Copyright 2016, Lily Robotics"
import click
import git_helper
import subprocess
import zazu.build
import zazu.config
import zazu.dev.commands
import zazu.repo.commands
import zazu.style
import zazu.upgrade
@click.group()
@click.version_option(version=zazu.__version__)
@click.pass_context
def cli(ctx):
try:
ctx.obj = zazu.config.Config(git_helper.get_root_path())
required_zazu_version = ctx.obj.zazu_version_required()
if required_zazu_version and required_zazu_version != zazu.__version__:
click.echo('Warning: this repo has requested zazu {}, which doesn\'t match the installed version ({}). \
Use "zazu upgrade" to fix this'.format(ctx.obj.zazu_version_required(), zazu.__version__))
except subprocess.CalledProcessError:
pass
cli.add_command(zazu.upgrade.upgrade)
cli.add_command(zazu.style.style)
cli.add_command(zazu.build.build)
cli.add_command(zazu.dev.commands.dev)
cli.add_command(zazu.repo.commands.repo)
| Fix missing module on __version__ | Fix missing module on __version__
(develop)
| Python | mit | stopthatcow/zazu,stopthatcow/zazu | # -*- coding: utf-8 -*-
"""entry point for zazu"""
__author__ = "Nicholas Wiles"
__copyright__ = "Copyright 2016, Lily Robotics"
import click
import git_helper
import subprocess
import zazu.build
import zazu.config
import zazu.dev.commands
import zazu.repo.commands
import zazu.style
import zazu.upgrade
@click.group()
@click.version_option(version=zazu.__version__)
@click.pass_context
def cli(ctx):
try:
ctx.obj = zazu.config.Config(git_helper.get_root_path())
required_zazu_version = ctx.obj.zazu_version_required()
if required_zazu_version and required_zazu_version != zazu.__version__:
click.echo('Warning: this repo has requested zazu {}, which doesn\'t match the installed version ({}). \
Use "zazu upgrade" to fix this'.format(ctx.obj.zazu_version_required(), __version__))
except subprocess.CalledProcessError:
pass
cli.add_command(zazu.upgrade.upgrade)
cli.add_command(zazu.style.style)
cli.add_command(zazu.build.build)
cli.add_command(zazu.dev.commands.dev)
cli.add_command(zazu.repo.commands.repo)
Fix missing module on __version__
(develop) | # -*- coding: utf-8 -*-
"""entry point for zazu"""
__author__ = "Nicholas Wiles"
__copyright__ = "Copyright 2016, Lily Robotics"
import click
import git_helper
import subprocess
import zazu.build
import zazu.config
import zazu.dev.commands
import zazu.repo.commands
import zazu.style
import zazu.upgrade
@click.group()
@click.version_option(version=zazu.__version__)
@click.pass_context
def cli(ctx):
try:
ctx.obj = zazu.config.Config(git_helper.get_root_path())
required_zazu_version = ctx.obj.zazu_version_required()
if required_zazu_version and required_zazu_version != zazu.__version__:
click.echo('Warning: this repo has requested zazu {}, which doesn\'t match the installed version ({}). \
Use "zazu upgrade" to fix this'.format(ctx.obj.zazu_version_required(), zazu.__version__))
except subprocess.CalledProcessError:
pass
cli.add_command(zazu.upgrade.upgrade)
cli.add_command(zazu.style.style)
cli.add_command(zazu.build.build)
cli.add_command(zazu.dev.commands.dev)
cli.add_command(zazu.repo.commands.repo)
| <commit_before># -*- coding: utf-8 -*-
"""entry point for zazu"""
__author__ = "Nicholas Wiles"
__copyright__ = "Copyright 2016, Lily Robotics"
import click
import git_helper
import subprocess
import zazu.build
import zazu.config
import zazu.dev.commands
import zazu.repo.commands
import zazu.style
import zazu.upgrade
@click.group()
@click.version_option(version=zazu.__version__)
@click.pass_context
def cli(ctx):
try:
ctx.obj = zazu.config.Config(git_helper.get_root_path())
required_zazu_version = ctx.obj.zazu_version_required()
if required_zazu_version and required_zazu_version != zazu.__version__:
click.echo('Warning: this repo has requested zazu {}, which doesn\'t match the installed version ({}). \
Use "zazu upgrade" to fix this'.format(ctx.obj.zazu_version_required(), __version__))
except subprocess.CalledProcessError:
pass
cli.add_command(zazu.upgrade.upgrade)
cli.add_command(zazu.style.style)
cli.add_command(zazu.build.build)
cli.add_command(zazu.dev.commands.dev)
cli.add_command(zazu.repo.commands.repo)
<commit_msg>Fix missing module on __version__
(develop)<commit_after> | # -*- coding: utf-8 -*-
"""entry point for zazu"""
__author__ = "Nicholas Wiles"
__copyright__ = "Copyright 2016, Lily Robotics"
import click
import git_helper
import subprocess
import zazu.build
import zazu.config
import zazu.dev.commands
import zazu.repo.commands
import zazu.style
import zazu.upgrade
@click.group()
@click.version_option(version=zazu.__version__)
@click.pass_context
def cli(ctx):
try:
ctx.obj = zazu.config.Config(git_helper.get_root_path())
required_zazu_version = ctx.obj.zazu_version_required()
if required_zazu_version and required_zazu_version != zazu.__version__:
click.echo('Warning: this repo has requested zazu {}, which doesn\'t match the installed version ({}). \
Use "zazu upgrade" to fix this'.format(ctx.obj.zazu_version_required(), zazu.__version__))
except subprocess.CalledProcessError:
pass
cli.add_command(zazu.upgrade.upgrade)
cli.add_command(zazu.style.style)
cli.add_command(zazu.build.build)
cli.add_command(zazu.dev.commands.dev)
cli.add_command(zazu.repo.commands.repo)
| # -*- coding: utf-8 -*-
"""entry point for zazu"""
__author__ = "Nicholas Wiles"
__copyright__ = "Copyright 2016, Lily Robotics"
import click
import git_helper
import subprocess
import zazu.build
import zazu.config
import zazu.dev.commands
import zazu.repo.commands
import zazu.style
import zazu.upgrade
@click.group()
@click.version_option(version=zazu.__version__)
@click.pass_context
def cli(ctx):
try:
ctx.obj = zazu.config.Config(git_helper.get_root_path())
required_zazu_version = ctx.obj.zazu_version_required()
if required_zazu_version and required_zazu_version != zazu.__version__:
click.echo('Warning: this repo has requested zazu {}, which doesn\'t match the installed version ({}). \
Use "zazu upgrade" to fix this'.format(ctx.obj.zazu_version_required(), __version__))
except subprocess.CalledProcessError:
pass
cli.add_command(zazu.upgrade.upgrade)
cli.add_command(zazu.style.style)
cli.add_command(zazu.build.build)
cli.add_command(zazu.dev.commands.dev)
cli.add_command(zazu.repo.commands.repo)
Fix missing module on __version__
(develop)# -*- coding: utf-8 -*-
"""entry point for zazu"""
__author__ = "Nicholas Wiles"
__copyright__ = "Copyright 2016, Lily Robotics"
import click
import git_helper
import subprocess
import zazu.build
import zazu.config
import zazu.dev.commands
import zazu.repo.commands
import zazu.style
import zazu.upgrade
@click.group()
@click.version_option(version=zazu.__version__)
@click.pass_context
def cli(ctx):
try:
ctx.obj = zazu.config.Config(git_helper.get_root_path())
required_zazu_version = ctx.obj.zazu_version_required()
if required_zazu_version and required_zazu_version != zazu.__version__:
click.echo('Warning: this repo has requested zazu {}, which doesn\'t match the installed version ({}). \
Use "zazu upgrade" to fix this'.format(ctx.obj.zazu_version_required(), zazu.__version__))
except subprocess.CalledProcessError:
pass
cli.add_command(zazu.upgrade.upgrade)
cli.add_command(zazu.style.style)
cli.add_command(zazu.build.build)
cli.add_command(zazu.dev.commands.dev)
cli.add_command(zazu.repo.commands.repo)
| <commit_before># -*- coding: utf-8 -*-
"""entry point for zazu"""
__author__ = "Nicholas Wiles"
__copyright__ = "Copyright 2016, Lily Robotics"
import click
import git_helper
import subprocess
import zazu.build
import zazu.config
import zazu.dev.commands
import zazu.repo.commands
import zazu.style
import zazu.upgrade
@click.group()
@click.version_option(version=zazu.__version__)
@click.pass_context
def cli(ctx):
try:
ctx.obj = zazu.config.Config(git_helper.get_root_path())
required_zazu_version = ctx.obj.zazu_version_required()
if required_zazu_version and required_zazu_version != zazu.__version__:
click.echo('Warning: this repo has requested zazu {}, which doesn\'t match the installed version ({}). \
Use "zazu upgrade" to fix this'.format(ctx.obj.zazu_version_required(), __version__))
except subprocess.CalledProcessError:
pass
cli.add_command(zazu.upgrade.upgrade)
cli.add_command(zazu.style.style)
cli.add_command(zazu.build.build)
cli.add_command(zazu.dev.commands.dev)
cli.add_command(zazu.repo.commands.repo)
<commit_msg>Fix missing module on __version__
(develop)<commit_after># -*- coding: utf-8 -*-
"""entry point for zazu"""
__author__ = "Nicholas Wiles"
__copyright__ = "Copyright 2016, Lily Robotics"
import click
import git_helper
import subprocess
import zazu.build
import zazu.config
import zazu.dev.commands
import zazu.repo.commands
import zazu.style
import zazu.upgrade
@click.group()
@click.version_option(version=zazu.__version__)
@click.pass_context
def cli(ctx):
try:
ctx.obj = zazu.config.Config(git_helper.get_root_path())
required_zazu_version = ctx.obj.zazu_version_required()
if required_zazu_version and required_zazu_version != zazu.__version__:
click.echo('Warning: this repo has requested zazu {}, which doesn\'t match the installed version ({}). \
Use "zazu upgrade" to fix this'.format(ctx.obj.zazu_version_required(), zazu.__version__))
except subprocess.CalledProcessError:
pass
cli.add_command(zazu.upgrade.upgrade)
cli.add_command(zazu.style.style)
cli.add_command(zazu.build.build)
cli.add_command(zazu.dev.commands.dev)
cli.add_command(zazu.repo.commands.repo)
|
f0b54d2f706912fd4a1e21117d1534170a9ce125 | Config.py | Config.py | # Web
domain = 'mockbox.io'
useSSL = False
keyFile = 'keys/privkey.pem'
certFile = 'keys/cacert.pem'
httpPort = 81
httpsPort = 443
# SMTP
bindingPort = 587
bindingIP = '0.0.0.0'
# dropSize in bytes
dropSize = 10000000
# General
# warning! This can fill your disk up really quickly.
logEmail = False
| # Web
domain = 'mockbox.io'
useSSL = False
keyFile = 'keys/privkey.pem'
certFile = 'keys/cacert.pem'
httpPort = 80
httpsPort = 443
# SMTP
bindingPort = 587
bindingIP = '0.0.0.0'
# dropSize in bytes
dropSize = 10000000
# General
# warning! This can fill your disk up really quickly.
logEmail = False
| Move default port back to 80 | Move default port back to 80
| Python | mit | jkingsman/Mockbox,jkingsman/Mockbox,jkingsman/Mockbox | # Web
domain = 'mockbox.io'
useSSL = False
keyFile = 'keys/privkey.pem'
certFile = 'keys/cacert.pem'
httpPort = 81
httpsPort = 443
# SMTP
bindingPort = 587
bindingIP = '0.0.0.0'
# dropSize in bytes
dropSize = 10000000
# General
# warning! This can fill your disk up really quickly.
logEmail = False
Move default port back to 80 | # Web
domain = 'mockbox.io'
useSSL = False
keyFile = 'keys/privkey.pem'
certFile = 'keys/cacert.pem'
httpPort = 80
httpsPort = 443
# SMTP
bindingPort = 587
bindingIP = '0.0.0.0'
# dropSize in bytes
dropSize = 10000000
# General
# warning! This can fill your disk up really quickly.
logEmail = False
| <commit_before># Web
domain = 'mockbox.io'
useSSL = False
keyFile = 'keys/privkey.pem'
certFile = 'keys/cacert.pem'
httpPort = 81
httpsPort = 443
# SMTP
bindingPort = 587
bindingIP = '0.0.0.0'
# dropSize in bytes
dropSize = 10000000
# General
# warning! This can fill your disk up really quickly.
logEmail = False
<commit_msg>Move default port back to 80<commit_after> | # Web
domain = 'mockbox.io'
useSSL = False
keyFile = 'keys/privkey.pem'
certFile = 'keys/cacert.pem'
httpPort = 80
httpsPort = 443
# SMTP
bindingPort = 587
bindingIP = '0.0.0.0'
# dropSize in bytes
dropSize = 10000000
# General
# warning! This can fill your disk up really quickly.
logEmail = False
| # Web
domain = 'mockbox.io'
useSSL = False
keyFile = 'keys/privkey.pem'
certFile = 'keys/cacert.pem'
httpPort = 81
httpsPort = 443
# SMTP
bindingPort = 587
bindingIP = '0.0.0.0'
# dropSize in bytes
dropSize = 10000000
# General
# warning! This can fill your disk up really quickly.
logEmail = False
Move default port back to 80# Web
domain = 'mockbox.io'
useSSL = False
keyFile = 'keys/privkey.pem'
certFile = 'keys/cacert.pem'
httpPort = 80
httpsPort = 443
# SMTP
bindingPort = 587
bindingIP = '0.0.0.0'
# dropSize in bytes
dropSize = 10000000
# General
# warning! This can fill your disk up really quickly.
logEmail = False
| <commit_before># Web
domain = 'mockbox.io'
useSSL = False
keyFile = 'keys/privkey.pem'
certFile = 'keys/cacert.pem'
httpPort = 81
httpsPort = 443
# SMTP
bindingPort = 587
bindingIP = '0.0.0.0'
# dropSize in bytes
dropSize = 10000000
# General
# warning! This can fill your disk up really quickly.
logEmail = False
<commit_msg>Move default port back to 80<commit_after># Web
domain = 'mockbox.io'
useSSL = False
keyFile = 'keys/privkey.pem'
certFile = 'keys/cacert.pem'
httpPort = 80
httpsPort = 443
# SMTP
bindingPort = 587
bindingIP = '0.0.0.0'
# dropSize in bytes
dropSize = 10000000
# General
# warning! This can fill your disk up really quickly.
logEmail = False
|
918b001cb6d9743d3d2ee1b2bab8f14c90e1adf7 | src/ice/rom_finder.py | src/ice/rom_finder.py |
from console import Console
from rom import ROM
from functools import reduce
class ROMFinder(object):
def __init__(self, filesystem):
self.filesystem = filesystem
def roms_for_console(self, console):
"""
@param console - A console object
@returns A list of ROM objects representing all of the valid ROMs for a
given console.
Valid ROMs are defined as ROMs for which `console`'s `is_valid_rom` method
returns True.
Returns an empty list if `console` is not enabled
"""
if not console.is_enabled():
return []
paths = self.filesystem.files_in_directory(console.roms_directory())
valid_rom_paths = filter(console.is_valid_rom, paths)
return map(lambda path: ROM(path, console), valid_rom_paths)
def roms_for_consoles(self, consoles):
"""
@param consoles - An iterable list of consoles
@returns A list of all of the ROMs for all of the consoles in `consoles`
Equivalent to calling `roms_for_console` on every element of `consoles`
and combining the results
"""
assert hasattr(
consoles, '__iter__'), "Expecting an iterable list of consoles"
def rom_collector(roms, console):
roms.extend(self.roms_for_console(console))
return roms
return reduce(rom_collector, consoles, [])
|
from console import Console
from rom import ROM
from functools import reduce
class ROMFinder(object):
def __init__(self, filesystem):
self.filesystem = filesystem
def roms_for_console(self, console):
"""
@param console - A console object
@returns A list of ROM objects representing all of the valid ROMs for a
given console.
Valid ROMs are defined as ROMs for which `console`'s `is_valid_rom` method
returns True.
Returns an empty list if `console` is not enabled
"""
if not console.is_enabled():
return []
paths = self.filesystem.files_in_directory(console.roms_directory())
valid_rom_paths = filter(console.is_valid_rom, paths)
return map(lambda path: ROM(path, console), valid_rom_paths)
def roms_for_consoles(self, consoles):
"""
@param consoles - An iterable list of consoles
@returns A list of all of the ROMs for all of the consoles in `consoles`
Equivalent to calling `roms_for_console` on every element of `consoles`
and combining the results
"""
return reduce(lambda roms, console: roms + self.roms_for_console(console), consoles, [])
| Replace 'list.extend' call with '+' operator | [Cleanup] Replace 'list.extend' call with '+' operator
I knew there had to be an easier way for merging lists other than `extend`. Turns out the plus operator does exactly what I need.
| Python | mit | rdoyle1978/Ice,scottrice/Ice |
from console import Console
from rom import ROM
from functools import reduce
class ROMFinder(object):
def __init__(self, filesystem):
self.filesystem = filesystem
def roms_for_console(self, console):
"""
@param console - A console object
@returns A list of ROM objects representing all of the valid ROMs for a
given console.
Valid ROMs are defined as ROMs for which `console`'s `is_valid_rom` method
returns True.
Returns an empty list if `console` is not enabled
"""
if not console.is_enabled():
return []
paths = self.filesystem.files_in_directory(console.roms_directory())
valid_rom_paths = filter(console.is_valid_rom, paths)
return map(lambda path: ROM(path, console), valid_rom_paths)
def roms_for_consoles(self, consoles):
"""
@param consoles - An iterable list of consoles
@returns A list of all of the ROMs for all of the consoles in `consoles`
Equivalent to calling `roms_for_console` on every element of `consoles`
and combining the results
"""
assert hasattr(
consoles, '__iter__'), "Expecting an iterable list of consoles"
def rom_collector(roms, console):
roms.extend(self.roms_for_console(console))
return roms
return reduce(rom_collector, consoles, [])
[Cleanup] Replace 'list.extend' call with '+' operator
I knew there had to be an easier way for merging lists other than `extend`. Turns out the plus operator does exactly what I need. |
from console import Console
from rom import ROM
from functools import reduce
class ROMFinder(object):
def __init__(self, filesystem):
self.filesystem = filesystem
def roms_for_console(self, console):
"""
@param console - A console object
@returns A list of ROM objects representing all of the valid ROMs for a
given console.
Valid ROMs are defined as ROMs for which `console`'s `is_valid_rom` method
returns True.
Returns an empty list if `console` is not enabled
"""
if not console.is_enabled():
return []
paths = self.filesystem.files_in_directory(console.roms_directory())
valid_rom_paths = filter(console.is_valid_rom, paths)
return map(lambda path: ROM(path, console), valid_rom_paths)
def roms_for_consoles(self, consoles):
"""
@param consoles - An iterable list of consoles
@returns A list of all of the ROMs for all of the consoles in `consoles`
Equivalent to calling `roms_for_console` on every element of `consoles`
and combining the results
"""
return reduce(lambda roms, console: roms + self.roms_for_console(console), consoles, [])
| <commit_before>
from console import Console
from rom import ROM
from functools import reduce
class ROMFinder(object):
def __init__(self, filesystem):
self.filesystem = filesystem
def roms_for_console(self, console):
"""
@param console - A console object
@returns A list of ROM objects representing all of the valid ROMs for a
given console.
Valid ROMs are defined as ROMs for which `console`'s `is_valid_rom` method
returns True.
Returns an empty list if `console` is not enabled
"""
if not console.is_enabled():
return []
paths = self.filesystem.files_in_directory(console.roms_directory())
valid_rom_paths = filter(console.is_valid_rom, paths)
return map(lambda path: ROM(path, console), valid_rom_paths)
def roms_for_consoles(self, consoles):
"""
@param consoles - An iterable list of consoles
@returns A list of all of the ROMs for all of the consoles in `consoles`
Equivalent to calling `roms_for_console` on every element of `consoles`
and combining the results
"""
assert hasattr(
consoles, '__iter__'), "Expecting an iterable list of consoles"
def rom_collector(roms, console):
roms.extend(self.roms_for_console(console))
return roms
return reduce(rom_collector, consoles, [])
<commit_msg>[Cleanup] Replace 'list.extend' call with '+' operator
I knew there had to be an easier way for merging lists other than `extend`. Turns out the plus operator does exactly what I need.<commit_after> |
from console import Console
from rom import ROM
from functools import reduce
class ROMFinder(object):
def __init__(self, filesystem):
self.filesystem = filesystem
def roms_for_console(self, console):
"""
@param console - A console object
@returns A list of ROM objects representing all of the valid ROMs for a
given console.
Valid ROMs are defined as ROMs for which `console`'s `is_valid_rom` method
returns True.
Returns an empty list if `console` is not enabled
"""
if not console.is_enabled():
return []
paths = self.filesystem.files_in_directory(console.roms_directory())
valid_rom_paths = filter(console.is_valid_rom, paths)
return map(lambda path: ROM(path, console), valid_rom_paths)
def roms_for_consoles(self, consoles):
"""
@param consoles - An iterable list of consoles
@returns A list of all of the ROMs for all of the consoles in `consoles`
Equivalent to calling `roms_for_console` on every element of `consoles`
and combining the results
"""
return reduce(lambda roms, console: roms + self.roms_for_console(console), consoles, [])
|
from console import Console
from rom import ROM
from functools import reduce
class ROMFinder(object):
def __init__(self, filesystem):
self.filesystem = filesystem
def roms_for_console(self, console):
"""
@param console - A console object
@returns A list of ROM objects representing all of the valid ROMs for a
given console.
Valid ROMs are defined as ROMs for which `console`'s `is_valid_rom` method
returns True.
Returns an empty list if `console` is not enabled
"""
if not console.is_enabled():
return []
paths = self.filesystem.files_in_directory(console.roms_directory())
valid_rom_paths = filter(console.is_valid_rom, paths)
return map(lambda path: ROM(path, console), valid_rom_paths)
def roms_for_consoles(self, consoles):
"""
@param consoles - An iterable list of consoles
@returns A list of all of the ROMs for all of the consoles in `consoles`
Equivalent to calling `roms_for_console` on every element of `consoles`
and combining the results
"""
assert hasattr(
consoles, '__iter__'), "Expecting an iterable list of consoles"
def rom_collector(roms, console):
roms.extend(self.roms_for_console(console))
return roms
return reduce(rom_collector, consoles, [])
[Cleanup] Replace 'list.extend' call with '+' operator
I knew there had to be an easier way for merging lists other than `extend`. Turns out the plus operator does exactly what I need.
from console import Console
from rom import ROM
from functools import reduce
class ROMFinder(object):
def __init__(self, filesystem):
self.filesystem = filesystem
def roms_for_console(self, console):
"""
@param console - A console object
@returns A list of ROM objects representing all of the valid ROMs for a
given console.
Valid ROMs are defined as ROMs for which `console`'s `is_valid_rom` method
returns True.
Returns an empty list if `console` is not enabled
"""
if not console.is_enabled():
return []
paths = self.filesystem.files_in_directory(console.roms_directory())
valid_rom_paths = filter(console.is_valid_rom, paths)
return map(lambda path: ROM(path, console), valid_rom_paths)
def roms_for_consoles(self, consoles):
"""
@param consoles - An iterable list of consoles
@returns A list of all of the ROMs for all of the consoles in `consoles`
Equivalent to calling `roms_for_console` on every element of `consoles`
and combining the results
"""
return reduce(lambda roms, console: roms + self.roms_for_console(console), consoles, [])
| <commit_before>
from console import Console
from rom import ROM
from functools import reduce
class ROMFinder(object):
def __init__(self, filesystem):
self.filesystem = filesystem
def roms_for_console(self, console):
"""
@param console - A console object
@returns A list of ROM objects representing all of the valid ROMs for a
given console.
Valid ROMs are defined as ROMs for which `console`'s `is_valid_rom` method
returns True.
Returns an empty list if `console` is not enabled
"""
if not console.is_enabled():
return []
paths = self.filesystem.files_in_directory(console.roms_directory())
valid_rom_paths = filter(console.is_valid_rom, paths)
return map(lambda path: ROM(path, console), valid_rom_paths)
def roms_for_consoles(self, consoles):
"""
@param consoles - An iterable list of consoles
@returns A list of all of the ROMs for all of the consoles in `consoles`
Equivalent to calling `roms_for_console` on every element of `consoles`
and combining the results
"""
assert hasattr(
consoles, '__iter__'), "Expecting an iterable list of consoles"
def rom_collector(roms, console):
roms.extend(self.roms_for_console(console))
return roms
return reduce(rom_collector, consoles, [])
<commit_msg>[Cleanup] Replace 'list.extend' call with '+' operator
I knew there had to be an easier way for merging lists other than `extend`. Turns out the plus operator does exactly what I need.<commit_after>
from console import Console
from rom import ROM
from functools import reduce
class ROMFinder(object):
def __init__(self, filesystem):
self.filesystem = filesystem
def roms_for_console(self, console):
"""
@param console - A console object
@returns A list of ROM objects representing all of the valid ROMs for a
given console.
Valid ROMs are defined as ROMs for which `console`'s `is_valid_rom` method
returns True.
Returns an empty list if `console` is not enabled
"""
if not console.is_enabled():
return []
paths = self.filesystem.files_in_directory(console.roms_directory())
valid_rom_paths = filter(console.is_valid_rom, paths)
return map(lambda path: ROM(path, console), valid_rom_paths)
def roms_for_consoles(self, consoles):
"""
@param consoles - An iterable list of consoles
@returns A list of all of the ROMs for all of the consoles in `consoles`
Equivalent to calling `roms_for_console` on every element of `consoles`
and combining the results
"""
return reduce(lambda roms, console: roms + self.roms_for_console(console), consoles, [])
|
bb48fbaa362367c117501953f3f1ba7500ff5735 | bqueryd/__init__.py | bqueryd/__init__.py | __version__ = 0.6
import os
import logging
logger = logging.getLogger('bqueryd')
ch = logging.StreamHandler()
formatter = logging.Formatter('%(asctime)s %(name)s %(message)s', datefmt="%Y-%m-%d %H:%M:%S")
ch.setFormatter(formatter)
logger.addHandler(ch)
DEFAULT_DATA_DIR = '/srv/bcolz/'
INCOMING = os.path.join(DEFAULT_DATA_DIR, 'incoming')
if not os.path.exists(INCOMING):
os.mkdir(INCOMING)
REDIS_SET_KEY = 'bqueryd_controllers'
from rpc import RPC, RPCError
from controller import ControllerNode
from worker import WorkerNode
| __version__ = 0.8
import os
import logging
logger = logging.getLogger('bqueryd')
ch = logging.StreamHandler()
formatter = logging.Formatter('%(asctime)s %(name)s %(message)s', datefmt="%Y-%m-%d %H:%M:%S")
ch.setFormatter(formatter)
logger.addHandler(ch)
DEFAULT_DATA_DIR = '/srv/bcolz/'
INCOMING = os.path.join(DEFAULT_DATA_DIR, 'incoming')
if not os.path.exists(INCOMING):
os.makedirs(INCOMING)
REDIS_SET_KEY = 'bqueryd_controllers'
from rpc import RPC, RPCError
from controller import ControllerNode
from worker import WorkerNode
| Use os.makedirs in stead of os.mkdir to handle intermediate paths | Use os.makedirs in stead of os.mkdir to handle intermediate paths
| Python | bsd-3-clause | visualfabriq/bqueryd | __version__ = 0.6
import os
import logging
logger = logging.getLogger('bqueryd')
ch = logging.StreamHandler()
formatter = logging.Formatter('%(asctime)s %(name)s %(message)s', datefmt="%Y-%m-%d %H:%M:%S")
ch.setFormatter(formatter)
logger.addHandler(ch)
DEFAULT_DATA_DIR = '/srv/bcolz/'
INCOMING = os.path.join(DEFAULT_DATA_DIR, 'incoming')
if not os.path.exists(INCOMING):
os.mkdir(INCOMING)
REDIS_SET_KEY = 'bqueryd_controllers'
from rpc import RPC, RPCError
from controller import ControllerNode
from worker import WorkerNode
Use os.makedirs in stead of os.mkdir to handle intermediate paths | __version__ = 0.8
import os
import logging
logger = logging.getLogger('bqueryd')
ch = logging.StreamHandler()
formatter = logging.Formatter('%(asctime)s %(name)s %(message)s', datefmt="%Y-%m-%d %H:%M:%S")
ch.setFormatter(formatter)
logger.addHandler(ch)
DEFAULT_DATA_DIR = '/srv/bcolz/'
INCOMING = os.path.join(DEFAULT_DATA_DIR, 'incoming')
if not os.path.exists(INCOMING):
os.makedirs(INCOMING)
REDIS_SET_KEY = 'bqueryd_controllers'
from rpc import RPC, RPCError
from controller import ControllerNode
from worker import WorkerNode
| <commit_before>__version__ = 0.6
import os
import logging
logger = logging.getLogger('bqueryd')
ch = logging.StreamHandler()
formatter = logging.Formatter('%(asctime)s %(name)s %(message)s', datefmt="%Y-%m-%d %H:%M:%S")
ch.setFormatter(formatter)
logger.addHandler(ch)
DEFAULT_DATA_DIR = '/srv/bcolz/'
INCOMING = os.path.join(DEFAULT_DATA_DIR, 'incoming')
if not os.path.exists(INCOMING):
os.mkdir(INCOMING)
REDIS_SET_KEY = 'bqueryd_controllers'
from rpc import RPC, RPCError
from controller import ControllerNode
from worker import WorkerNode
<commit_msg>Use os.makedirs in stead of os.mkdir to handle intermediate paths<commit_after> | __version__ = 0.8
import os
import logging
logger = logging.getLogger('bqueryd')
ch = logging.StreamHandler()
formatter = logging.Formatter('%(asctime)s %(name)s %(message)s', datefmt="%Y-%m-%d %H:%M:%S")
ch.setFormatter(formatter)
logger.addHandler(ch)
DEFAULT_DATA_DIR = '/srv/bcolz/'
INCOMING = os.path.join(DEFAULT_DATA_DIR, 'incoming')
if not os.path.exists(INCOMING):
os.makedirs(INCOMING)
REDIS_SET_KEY = 'bqueryd_controllers'
from rpc import RPC, RPCError
from controller import ControllerNode
from worker import WorkerNode
| __version__ = 0.6
import os
import logging
logger = logging.getLogger('bqueryd')
ch = logging.StreamHandler()
formatter = logging.Formatter('%(asctime)s %(name)s %(message)s', datefmt="%Y-%m-%d %H:%M:%S")
ch.setFormatter(formatter)
logger.addHandler(ch)
DEFAULT_DATA_DIR = '/srv/bcolz/'
INCOMING = os.path.join(DEFAULT_DATA_DIR, 'incoming')
if not os.path.exists(INCOMING):
os.mkdir(INCOMING)
REDIS_SET_KEY = 'bqueryd_controllers'
from rpc import RPC, RPCError
from controller import ControllerNode
from worker import WorkerNode
Use os.makedirs in stead of os.mkdir to handle intermediate paths__version__ = 0.8
import os
import logging
logger = logging.getLogger('bqueryd')
ch = logging.StreamHandler()
formatter = logging.Formatter('%(asctime)s %(name)s %(message)s', datefmt="%Y-%m-%d %H:%M:%S")
ch.setFormatter(formatter)
logger.addHandler(ch)
DEFAULT_DATA_DIR = '/srv/bcolz/'
INCOMING = os.path.join(DEFAULT_DATA_DIR, 'incoming')
if not os.path.exists(INCOMING):
os.makedirs(INCOMING)
REDIS_SET_KEY = 'bqueryd_controllers'
from rpc import RPC, RPCError
from controller import ControllerNode
from worker import WorkerNode
| <commit_before>__version__ = 0.6
import os
import logging
logger = logging.getLogger('bqueryd')
ch = logging.StreamHandler()
formatter = logging.Formatter('%(asctime)s %(name)s %(message)s', datefmt="%Y-%m-%d %H:%M:%S")
ch.setFormatter(formatter)
logger.addHandler(ch)
DEFAULT_DATA_DIR = '/srv/bcolz/'
INCOMING = os.path.join(DEFAULT_DATA_DIR, 'incoming')
if not os.path.exists(INCOMING):
os.mkdir(INCOMING)
REDIS_SET_KEY = 'bqueryd_controllers'
from rpc import RPC, RPCError
from controller import ControllerNode
from worker import WorkerNode
<commit_msg>Use os.makedirs in stead of os.mkdir to handle intermediate paths<commit_after>__version__ = 0.8
import os
import logging
logger = logging.getLogger('bqueryd')
ch = logging.StreamHandler()
formatter = logging.Formatter('%(asctime)s %(name)s %(message)s', datefmt="%Y-%m-%d %H:%M:%S")
ch.setFormatter(formatter)
logger.addHandler(ch)
DEFAULT_DATA_DIR = '/srv/bcolz/'
INCOMING = os.path.join(DEFAULT_DATA_DIR, 'incoming')
if not os.path.exists(INCOMING):
os.makedirs(INCOMING)
REDIS_SET_KEY = 'bqueryd_controllers'
from rpc import RPC, RPCError
from controller import ControllerNode
from worker import WorkerNode
|
d2197583c197745ad33fa9c59c6c7f9681b35078 | plugins/shorewall/__init__.py | plugins/shorewall/__init__.py | import logging, interfaces, os
from datetime import datetime
class Shorewall(interfaces.IOpenMesherConfigPlugin):
def activate(self):
self._register('shorewall/interfaces.conf')
self._register('shorewall/rules.conf')
def process(self, mesh, cliargs=None):
logging.debug('Generating Shorewall config...')
self._files = {}
for router in mesh.links:
self._files[router] = {}
self._files[router]['/shorewall/interfaces.mesh'] = self._templates['shorewall/interfaces.conf'].render(links = mesh.links[router], zone='vpn')
self._files[router]['/shorewall/rules.mesh'] = self._templates['shorewall/rules.conf'].render(links= mesh.links[router])
return True
def files(self):
""" Return a dictionary of routers containing a dictionary of filenames and contents """
return self._files
| import logging, interfaces, os
from datetime import datetime
class Shorewall(interfaces.IOpenMesherConfigPlugin):
def activate(self):
self._register('shorewall/interfaces.conf')
self._register('shorewall/rules.conf')
def setupargs(self, parser):
parser.add_argument('--wanzone', action='store', default='wan', help='Name of the WAN zone in Shorewall')
parser.add_argument('--fwzone', action='store', default='fw', help='Name of the Firewall zone in Shorewall')
parser.add_argument('--vpnzone', action='store', default='vpn', help='Name of the VPN zone in Shorewall')
def process(self, mesh, cliargs=None):
logging.debug('Generating Shorewall config...')
self._files = {}
for router in mesh.links:
self._files[router] = {}
self._files[router]['/shorewall/interfaces.mesh'] = self._templates['shorewall/interfaces.conf'].render(links = mesh.links[router], vpnzone=cliargs.vpnzone, fwzone=cliargs.fwzone, wanzone=cliargs.wanzone)
self._files[router]['/shorewall/rules.mesh'] = self._templates['shorewall/rules.conf'].render(links= mesh.links[router], vpnzone=cliargs.vpnzone, fwzone=cliargs.fwzone, wanzone=cliargs.wanzone)
return True
def files(self):
""" Return a dictionary of routers containing a dictionary of filenames and contents """
return self._files
| Add cliargs to shorewall for zone handling | Add cliargs to shorewall for zone handling
| Python | bsd-3-clause | heyaaron/openmesher,darkpixel/openmesher,heyaaron/openmesher,darkpixel/openmesher | import logging, interfaces, os
from datetime import datetime
class Shorewall(interfaces.IOpenMesherConfigPlugin):
def activate(self):
self._register('shorewall/interfaces.conf')
self._register('shorewall/rules.conf')
def process(self, mesh, cliargs=None):
logging.debug('Generating Shorewall config...')
self._files = {}
for router in mesh.links:
self._files[router] = {}
self._files[router]['/shorewall/interfaces.mesh'] = self._templates['shorewall/interfaces.conf'].render(links = mesh.links[router], zone='vpn')
self._files[router]['/shorewall/rules.mesh'] = self._templates['shorewall/rules.conf'].render(links= mesh.links[router])
return True
def files(self):
""" Return a dictionary of routers containing a dictionary of filenames and contents """
return self._files
Add cliargs to shorewall for zone handling | import logging, interfaces, os
from datetime import datetime
class Shorewall(interfaces.IOpenMesherConfigPlugin):
def activate(self):
self._register('shorewall/interfaces.conf')
self._register('shorewall/rules.conf')
def setupargs(self, parser):
parser.add_argument('--wanzone', action='store', default='wan', help='Name of the WAN zone in Shorewall')
parser.add_argument('--fwzone', action='store', default='fw', help='Name of the Firewall zone in Shorewall')
parser.add_argument('--vpnzone', action='store', default='vpn', help='Name of the VPN zone in Shorewall')
def process(self, mesh, cliargs=None):
logging.debug('Generating Shorewall config...')
self._files = {}
for router in mesh.links:
self._files[router] = {}
self._files[router]['/shorewall/interfaces.mesh'] = self._templates['shorewall/interfaces.conf'].render(links = mesh.links[router], vpnzone=cliargs.vpnzone, fwzone=cliargs.fwzone, wanzone=cliargs.wanzone)
self._files[router]['/shorewall/rules.mesh'] = self._templates['shorewall/rules.conf'].render(links= mesh.links[router], vpnzone=cliargs.vpnzone, fwzone=cliargs.fwzone, wanzone=cliargs.wanzone)
return True
def files(self):
""" Return a dictionary of routers containing a dictionary of filenames and contents """
return self._files
| <commit_before>import logging, interfaces, os
from datetime import datetime
class Shorewall(interfaces.IOpenMesherConfigPlugin):
def activate(self):
self._register('shorewall/interfaces.conf')
self._register('shorewall/rules.conf')
def process(self, mesh, cliargs=None):
logging.debug('Generating Shorewall config...')
self._files = {}
for router in mesh.links:
self._files[router] = {}
self._files[router]['/shorewall/interfaces.mesh'] = self._templates['shorewall/interfaces.conf'].render(links = mesh.links[router], zone='vpn')
self._files[router]['/shorewall/rules.mesh'] = self._templates['shorewall/rules.conf'].render(links= mesh.links[router])
return True
def files(self):
""" Return a dictionary of routers containing a dictionary of filenames and contents """
return self._files
<commit_msg>Add cliargs to shorewall for zone handling<commit_after> | import logging, interfaces, os
from datetime import datetime
class Shorewall(interfaces.IOpenMesherConfigPlugin):
def activate(self):
self._register('shorewall/interfaces.conf')
self._register('shorewall/rules.conf')
def setupargs(self, parser):
parser.add_argument('--wanzone', action='store', default='wan', help='Name of the WAN zone in Shorewall')
parser.add_argument('--fwzone', action='store', default='fw', help='Name of the Firewall zone in Shorewall')
parser.add_argument('--vpnzone', action='store', default='vpn', help='Name of the VPN zone in Shorewall')
def process(self, mesh, cliargs=None):
logging.debug('Generating Shorewall config...')
self._files = {}
for router in mesh.links:
self._files[router] = {}
self._files[router]['/shorewall/interfaces.mesh'] = self._templates['shorewall/interfaces.conf'].render(links = mesh.links[router], vpnzone=cliargs.vpnzone, fwzone=cliargs.fwzone, wanzone=cliargs.wanzone)
self._files[router]['/shorewall/rules.mesh'] = self._templates['shorewall/rules.conf'].render(links= mesh.links[router], vpnzone=cliargs.vpnzone, fwzone=cliargs.fwzone, wanzone=cliargs.wanzone)
return True
def files(self):
""" Return a dictionary of routers containing a dictionary of filenames and contents """
return self._files
| import logging, interfaces, os
from datetime import datetime
class Shorewall(interfaces.IOpenMesherConfigPlugin):
def activate(self):
self._register('shorewall/interfaces.conf')
self._register('shorewall/rules.conf')
def process(self, mesh, cliargs=None):
logging.debug('Generating Shorewall config...')
self._files = {}
for router in mesh.links:
self._files[router] = {}
self._files[router]['/shorewall/interfaces.mesh'] = self._templates['shorewall/interfaces.conf'].render(links = mesh.links[router], zone='vpn')
self._files[router]['/shorewall/rules.mesh'] = self._templates['shorewall/rules.conf'].render(links= mesh.links[router])
return True
def files(self):
""" Return a dictionary of routers containing a dictionary of filenames and contents """
return self._files
Add cliargs to shorewall for zone handlingimport logging, interfaces, os
from datetime import datetime
class Shorewall(interfaces.IOpenMesherConfigPlugin):
def activate(self):
self._register('shorewall/interfaces.conf')
self._register('shorewall/rules.conf')
def setupargs(self, parser):
parser.add_argument('--wanzone', action='store', default='wan', help='Name of the WAN zone in Shorewall')
parser.add_argument('--fwzone', action='store', default='fw', help='Name of the Firewall zone in Shorewall')
parser.add_argument('--vpnzone', action='store', default='vpn', help='Name of the VPN zone in Shorewall')
def process(self, mesh, cliargs=None):
logging.debug('Generating Shorewall config...')
self._files = {}
for router in mesh.links:
self._files[router] = {}
self._files[router]['/shorewall/interfaces.mesh'] = self._templates['shorewall/interfaces.conf'].render(links = mesh.links[router], vpnzone=cliargs.vpnzone, fwzone=cliargs.fwzone, wanzone=cliargs.wanzone)
self._files[router]['/shorewall/rules.mesh'] = self._templates['shorewall/rules.conf'].render(links= mesh.links[router], vpnzone=cliargs.vpnzone, fwzone=cliargs.fwzone, wanzone=cliargs.wanzone)
return True
def files(self):
""" Return a dictionary of routers containing a dictionary of filenames and contents """
return self._files
| <commit_before>import logging, interfaces, os
from datetime import datetime
class Shorewall(interfaces.IOpenMesherConfigPlugin):
def activate(self):
self._register('shorewall/interfaces.conf')
self._register('shorewall/rules.conf')
def process(self, mesh, cliargs=None):
logging.debug('Generating Shorewall config...')
self._files = {}
for router in mesh.links:
self._files[router] = {}
self._files[router]['/shorewall/interfaces.mesh'] = self._templates['shorewall/interfaces.conf'].render(links = mesh.links[router], zone='vpn')
self._files[router]['/shorewall/rules.mesh'] = self._templates['shorewall/rules.conf'].render(links= mesh.links[router])
return True
def files(self):
""" Return a dictionary of routers containing a dictionary of filenames and contents """
return self._files
<commit_msg>Add cliargs to shorewall for zone handling<commit_after>import logging, interfaces, os
from datetime import datetime
class Shorewall(interfaces.IOpenMesherConfigPlugin):
def activate(self):
self._register('shorewall/interfaces.conf')
self._register('shorewall/rules.conf')
def setupargs(self, parser):
parser.add_argument('--wanzone', action='store', default='wan', help='Name of the WAN zone in Shorewall')
parser.add_argument('--fwzone', action='store', default='fw', help='Name of the Firewall zone in Shorewall')
parser.add_argument('--vpnzone', action='store', default='vpn', help='Name of the VPN zone in Shorewall')
def process(self, mesh, cliargs=None):
logging.debug('Generating Shorewall config...')
self._files = {}
for router in mesh.links:
self._files[router] = {}
self._files[router]['/shorewall/interfaces.mesh'] = self._templates['shorewall/interfaces.conf'].render(links = mesh.links[router], vpnzone=cliargs.vpnzone, fwzone=cliargs.fwzone, wanzone=cliargs.wanzone)
self._files[router]['/shorewall/rules.mesh'] = self._templates['shorewall/rules.conf'].render(links= mesh.links[router], vpnzone=cliargs.vpnzone, fwzone=cliargs.fwzone, wanzone=cliargs.wanzone)
return True
def files(self):
""" Return a dictionary of routers containing a dictionary of filenames and contents """
return self._files
|
cdd79aa60f4ef707714a632188373a5c2c4b0af4 | mass_mailing_switzerland/models/crm_event_compassion.py | mass_mailing_switzerland/models/crm_event_compassion.py | ##############################################################################
#
# Copyright (C) 2018 Compassion CH (http://www.compassion.ch)
# Releasing children from poverty in Jesus' name
# @author: Nathan Fluckiger <nathan.fluckiger@hotmail.ch>
#
# The licence is in the file __manifest__.py
#
##############################################################################
from odoo import models, api
class EventCompassion(models.Model):
_inherit = "crm.event.compassion"
@api.model
def create(self, vals):
event = super().create(vals)
if event.campaign_id:
event.analytic_id.campaign_id = event.campaign_id
event.origin_id.campaign_id = event.campaign_id
return event
@api.multi
def write(self, vals):
res = super().write(vals)
for new_event in self:
if new_event.campaign_id:
new_event.analytic_id.campaign_id = new_event.campaign_id
new_event.origin_id.campaign_id = new_event.campaign_id
return res
| ##############################################################################
#
# Copyright (C) 2018 Compassion CH (http://www.compassion.ch)
# Releasing children from poverty in Jesus' name
# @author: Nathan Fluckiger <nathan.fluckiger@hotmail.ch>
#
# The licence is in the file __manifest__.py
#
##############################################################################
from odoo import models, api
class EventCompassion(models.Model):
_inherit = "crm.event.compassion"
@api.model
def create(self, vals):
event = super().create(vals)
if event.campaign_id:
event.analytic_id.campaign_id = event.campaign_id
event.origin_id.campaign_id = event.campaign_id
return event
@api.multi
def write(self, vals):
res = super().write(vals)
for new_event in self:
if new_event.campaign_id:
if new_event.analytic_id:
new_event.analytic_id.campaign_id = new_event.campaign_id
if new_event.origin_id:
new_event.origin_id.campaign_id = new_event.campaign_id
return res
| FIX bug in event creation | FIX bug in event creation
| Python | agpl-3.0 | CompassionCH/compassion-switzerland,eicher31/compassion-switzerland,CompassionCH/compassion-switzerland,eicher31/compassion-switzerland,CompassionCH/compassion-switzerland,eicher31/compassion-switzerland | ##############################################################################
#
# Copyright (C) 2018 Compassion CH (http://www.compassion.ch)
# Releasing children from poverty in Jesus' name
# @author: Nathan Fluckiger <nathan.fluckiger@hotmail.ch>
#
# The licence is in the file __manifest__.py
#
##############################################################################
from odoo import models, api
class EventCompassion(models.Model):
_inherit = "crm.event.compassion"
@api.model
def create(self, vals):
event = super().create(vals)
if event.campaign_id:
event.analytic_id.campaign_id = event.campaign_id
event.origin_id.campaign_id = event.campaign_id
return event
@api.multi
def write(self, vals):
res = super().write(vals)
for new_event in self:
if new_event.campaign_id:
new_event.analytic_id.campaign_id = new_event.campaign_id
new_event.origin_id.campaign_id = new_event.campaign_id
return res
FIX bug in event creation | ##############################################################################
#
# Copyright (C) 2018 Compassion CH (http://www.compassion.ch)
# Releasing children from poverty in Jesus' name
# @author: Nathan Fluckiger <nathan.fluckiger@hotmail.ch>
#
# The licence is in the file __manifest__.py
#
##############################################################################
from odoo import models, api
class EventCompassion(models.Model):
_inherit = "crm.event.compassion"
@api.model
def create(self, vals):
event = super().create(vals)
if event.campaign_id:
event.analytic_id.campaign_id = event.campaign_id
event.origin_id.campaign_id = event.campaign_id
return event
@api.multi
def write(self, vals):
res = super().write(vals)
for new_event in self:
if new_event.campaign_id:
if new_event.analytic_id:
new_event.analytic_id.campaign_id = new_event.campaign_id
if new_event.origin_id:
new_event.origin_id.campaign_id = new_event.campaign_id
return res
| <commit_before>##############################################################################
#
# Copyright (C) 2018 Compassion CH (http://www.compassion.ch)
# Releasing children from poverty in Jesus' name
# @author: Nathan Fluckiger <nathan.fluckiger@hotmail.ch>
#
# The licence is in the file __manifest__.py
#
##############################################################################
from odoo import models, api
class EventCompassion(models.Model):
_inherit = "crm.event.compassion"
@api.model
def create(self, vals):
event = super().create(vals)
if event.campaign_id:
event.analytic_id.campaign_id = event.campaign_id
event.origin_id.campaign_id = event.campaign_id
return event
@api.multi
def write(self, vals):
res = super().write(vals)
for new_event in self:
if new_event.campaign_id:
new_event.analytic_id.campaign_id = new_event.campaign_id
new_event.origin_id.campaign_id = new_event.campaign_id
return res
<commit_msg>FIX bug in event creation<commit_after> | ##############################################################################
#
# Copyright (C) 2018 Compassion CH (http://www.compassion.ch)
# Releasing children from poverty in Jesus' name
# @author: Nathan Fluckiger <nathan.fluckiger@hotmail.ch>
#
# The licence is in the file __manifest__.py
#
##############################################################################
from odoo import models, api
class EventCompassion(models.Model):
_inherit = "crm.event.compassion"
@api.model
def create(self, vals):
event = super().create(vals)
if event.campaign_id:
event.analytic_id.campaign_id = event.campaign_id
event.origin_id.campaign_id = event.campaign_id
return event
@api.multi
def write(self, vals):
res = super().write(vals)
for new_event in self:
if new_event.campaign_id:
if new_event.analytic_id:
new_event.analytic_id.campaign_id = new_event.campaign_id
if new_event.origin_id:
new_event.origin_id.campaign_id = new_event.campaign_id
return res
| ##############################################################################
#
# Copyright (C) 2018 Compassion CH (http://www.compassion.ch)
# Releasing children from poverty in Jesus' name
# @author: Nathan Fluckiger <nathan.fluckiger@hotmail.ch>
#
# The licence is in the file __manifest__.py
#
##############################################################################
from odoo import models, api
class EventCompassion(models.Model):
_inherit = "crm.event.compassion"
@api.model
def create(self, vals):
event = super().create(vals)
if event.campaign_id:
event.analytic_id.campaign_id = event.campaign_id
event.origin_id.campaign_id = event.campaign_id
return event
@api.multi
def write(self, vals):
res = super().write(vals)
for new_event in self:
if new_event.campaign_id:
new_event.analytic_id.campaign_id = new_event.campaign_id
new_event.origin_id.campaign_id = new_event.campaign_id
return res
FIX bug in event creation##############################################################################
#
# Copyright (C) 2018 Compassion CH (http://www.compassion.ch)
# Releasing children from poverty in Jesus' name
# @author: Nathan Fluckiger <nathan.fluckiger@hotmail.ch>
#
# The licence is in the file __manifest__.py
#
##############################################################################
from odoo import models, api
class EventCompassion(models.Model):
_inherit = "crm.event.compassion"
@api.model
def create(self, vals):
event = super().create(vals)
if event.campaign_id:
event.analytic_id.campaign_id = event.campaign_id
event.origin_id.campaign_id = event.campaign_id
return event
@api.multi
def write(self, vals):
res = super().write(vals)
for new_event in self:
if new_event.campaign_id:
if new_event.analytic_id:
new_event.analytic_id.campaign_id = new_event.campaign_id
if new_event.origin_id:
new_event.origin_id.campaign_id = new_event.campaign_id
return res
| <commit_before>##############################################################################
#
# Copyright (C) 2018 Compassion CH (http://www.compassion.ch)
# Releasing children from poverty in Jesus' name
# @author: Nathan Fluckiger <nathan.fluckiger@hotmail.ch>
#
# The licence is in the file __manifest__.py
#
##############################################################################
from odoo import models, api
class EventCompassion(models.Model):
_inherit = "crm.event.compassion"
@api.model
def create(self, vals):
event = super().create(vals)
if event.campaign_id:
event.analytic_id.campaign_id = event.campaign_id
event.origin_id.campaign_id = event.campaign_id
return event
@api.multi
def write(self, vals):
res = super().write(vals)
for new_event in self:
if new_event.campaign_id:
new_event.analytic_id.campaign_id = new_event.campaign_id
new_event.origin_id.campaign_id = new_event.campaign_id
return res
<commit_msg>FIX bug in event creation<commit_after>##############################################################################
#
# Copyright (C) 2018 Compassion CH (http://www.compassion.ch)
# Releasing children from poverty in Jesus' name
# @author: Nathan Fluckiger <nathan.fluckiger@hotmail.ch>
#
# The licence is in the file __manifest__.py
#
##############################################################################
from odoo import models, api
class EventCompassion(models.Model):
_inherit = "crm.event.compassion"
@api.model
def create(self, vals):
event = super().create(vals)
if event.campaign_id:
event.analytic_id.campaign_id = event.campaign_id
event.origin_id.campaign_id = event.campaign_id
return event
@api.multi
def write(self, vals):
res = super().write(vals)
for new_event in self:
if new_event.campaign_id:
if new_event.analytic_id:
new_event.analytic_id.campaign_id = new_event.campaign_id
if new_event.origin_id:
new_event.origin_id.campaign_id = new_event.campaign_id
return res
|
19e6c020bc7d640fe4c8ffbdf7825693d7e4a03a | scripts/missing-qq.py | scripts/missing-qq.py | #!/usr/bin/env python
import os
import xml.etree.ElementTree as ET
RES_FOLDER = os.path.abspath(os.path.join(os.path.dirname(__file__), "../app/src/main/res"))
EN_STRINGS = os.path.join(RES_FOLDER, "values/strings.xml")
QQ_STRINGS = os.path.join(RES_FOLDER, "values-qq/strings.xml")
# Get ElementTree containing all message names in English
enroot = ET.parse(EN_STRINGS).getroot()
# Get ElementTree containing all documented messages
qqroot = ET.parse(QQ_STRINGS).getroot()
# Create a set to store all documented messages
qqmsgs = set()
# Add all documented messages to that set
for child in qqroot:
qqmsgs.add(child.attrib['name'])
# Iterate through all messages and check that they're documented
for child in enroot:
if child.attrib['name'] not in qqmsgs:
print(child.attrib['name'] + " is undocumented!")
| #!/usr/bin/env python
import os
import sys
import xml.etree.ElementTree as ET
RES_FOLDER = os.path.abspath(os.path.join(os.path.dirname(__file__), "../app/src/main/res"))
EN_STRINGS = os.path.join(RES_FOLDER, "values/strings.xml")
QQ_STRINGS = os.path.join(RES_FOLDER, "values-qq/strings.xml")
# Get ElementTree containing all message names in English
enroot = ET.parse(EN_STRINGS).getroot()
# Get ElementTree containing all documented messages
qqroot = ET.parse(QQ_STRINGS).getroot()
# Create a set to store all documented messages
qqmsgs = set()
# Add all documented messages to that set
for child in qqroot:
qqmsgs.add(child.attrib['name'])
# Iterate through all messages and check that they're documented
missing = 0
for child in enroot:
if child.attrib['name'] not in qqmsgs:
print(child.attrib['name'] + " is undocumented!")
missing += 1
sys.exit(1 if missing else 0)
| Exit with nonzero when qq strings are missing | Exit with nonzero when qq strings are missing
Change-Id: Ife0f114dbe48faa445397aa7a94f74de2309d117
| Python | apache-2.0 | anirudh24seven/apps-android-wikipedia,wikimedia/apps-android-wikipedia,SAGROUP2/apps-android-wikipedia,SAGROUP2/apps-android-wikipedia,SAGROUP2/apps-android-wikipedia,SAGROUP2/apps-android-wikipedia,wikimedia/apps-android-wikipedia,wikimedia/apps-android-wikipedia,SAGROUP2/apps-android-wikipedia,dbrant/apps-android-wikipedia,anirudh24seven/apps-android-wikipedia,dbrant/apps-android-wikipedia,dbrant/apps-android-wikipedia,anirudh24seven/apps-android-wikipedia,wikimedia/apps-android-wikipedia,dbrant/apps-android-wikipedia,dbrant/apps-android-wikipedia,anirudh24seven/apps-android-wikipedia,anirudh24seven/apps-android-wikipedia | #!/usr/bin/env python
import os
import xml.etree.ElementTree as ET
RES_FOLDER = os.path.abspath(os.path.join(os.path.dirname(__file__), "../app/src/main/res"))
EN_STRINGS = os.path.join(RES_FOLDER, "values/strings.xml")
QQ_STRINGS = os.path.join(RES_FOLDER, "values-qq/strings.xml")
# Get ElementTree containing all message names in English
enroot = ET.parse(EN_STRINGS).getroot()
# Get ElementTree containing all documented messages
qqroot = ET.parse(QQ_STRINGS).getroot()
# Create a set to store all documented messages
qqmsgs = set()
# Add all documented messages to that set
for child in qqroot:
qqmsgs.add(child.attrib['name'])
# Iterate through all messages and check that they're documented
for child in enroot:
if child.attrib['name'] not in qqmsgs:
print(child.attrib['name'] + " is undocumented!")
Exit with nonzero when qq strings are missing
Change-Id: Ife0f114dbe48faa445397aa7a94f74de2309d117 | #!/usr/bin/env python
import os
import sys
import xml.etree.ElementTree as ET
RES_FOLDER = os.path.abspath(os.path.join(os.path.dirname(__file__), "../app/src/main/res"))
EN_STRINGS = os.path.join(RES_FOLDER, "values/strings.xml")
QQ_STRINGS = os.path.join(RES_FOLDER, "values-qq/strings.xml")
# Get ElementTree containing all message names in English
enroot = ET.parse(EN_STRINGS).getroot()
# Get ElementTree containing all documented messages
qqroot = ET.parse(QQ_STRINGS).getroot()
# Create a set to store all documented messages
qqmsgs = set()
# Add all documented messages to that set
for child in qqroot:
qqmsgs.add(child.attrib['name'])
# Iterate through all messages and check that they're documented
missing = 0
for child in enroot:
if child.attrib['name'] not in qqmsgs:
print(child.attrib['name'] + " is undocumented!")
missing += 1
sys.exit(1 if missing else 0)
| <commit_before>#!/usr/bin/env python
import os
import xml.etree.ElementTree as ET
RES_FOLDER = os.path.abspath(os.path.join(os.path.dirname(__file__), "../app/src/main/res"))
EN_STRINGS = os.path.join(RES_FOLDER, "values/strings.xml")
QQ_STRINGS = os.path.join(RES_FOLDER, "values-qq/strings.xml")
# Get ElementTree containing all message names in English
enroot = ET.parse(EN_STRINGS).getroot()
# Get ElementTree containing all documented messages
qqroot = ET.parse(QQ_STRINGS).getroot()
# Create a set to store all documented messages
qqmsgs = set()
# Add all documented messages to that set
for child in qqroot:
qqmsgs.add(child.attrib['name'])
# Iterate through all messages and check that they're documented
for child in enroot:
if child.attrib['name'] not in qqmsgs:
print(child.attrib['name'] + " is undocumented!")
<commit_msg>Exit with nonzero when qq strings are missing
Change-Id: Ife0f114dbe48faa445397aa7a94f74de2309d117<commit_after> | #!/usr/bin/env python
import os
import sys
import xml.etree.ElementTree as ET
RES_FOLDER = os.path.abspath(os.path.join(os.path.dirname(__file__), "../app/src/main/res"))
EN_STRINGS = os.path.join(RES_FOLDER, "values/strings.xml")
QQ_STRINGS = os.path.join(RES_FOLDER, "values-qq/strings.xml")
# Get ElementTree containing all message names in English
enroot = ET.parse(EN_STRINGS).getroot()
# Get ElementTree containing all documented messages
qqroot = ET.parse(QQ_STRINGS).getroot()
# Create a set to store all documented messages
qqmsgs = set()
# Add all documented messages to that set
for child in qqroot:
qqmsgs.add(child.attrib['name'])
# Iterate through all messages and check that they're documented
missing = 0
for child in enroot:
if child.attrib['name'] not in qqmsgs:
print(child.attrib['name'] + " is undocumented!")
missing += 1
sys.exit(1 if missing else 0)
| #!/usr/bin/env python
import os
import xml.etree.ElementTree as ET
RES_FOLDER = os.path.abspath(os.path.join(os.path.dirname(__file__), "../app/src/main/res"))
EN_STRINGS = os.path.join(RES_FOLDER, "values/strings.xml")
QQ_STRINGS = os.path.join(RES_FOLDER, "values-qq/strings.xml")
# Get ElementTree containing all message names in English
enroot = ET.parse(EN_STRINGS).getroot()
# Get ElementTree containing all documented messages
qqroot = ET.parse(QQ_STRINGS).getroot()
# Create a set to store all documented messages
qqmsgs = set()
# Add all documented messages to that set
for child in qqroot:
qqmsgs.add(child.attrib['name'])
# Iterate through all messages and check that they're documented
for child in enroot:
if child.attrib['name'] not in qqmsgs:
print(child.attrib['name'] + " is undocumented!")
Exit with nonzero when qq strings are missing
Change-Id: Ife0f114dbe48faa445397aa7a94f74de2309d117#!/usr/bin/env python
import os
import sys
import xml.etree.ElementTree as ET
RES_FOLDER = os.path.abspath(os.path.join(os.path.dirname(__file__), "../app/src/main/res"))
EN_STRINGS = os.path.join(RES_FOLDER, "values/strings.xml")
QQ_STRINGS = os.path.join(RES_FOLDER, "values-qq/strings.xml")
# Get ElementTree containing all message names in English
enroot = ET.parse(EN_STRINGS).getroot()
# Get ElementTree containing all documented messages
qqroot = ET.parse(QQ_STRINGS).getroot()
# Create a set to store all documented messages
qqmsgs = set()
# Add all documented messages to that set
for child in qqroot:
qqmsgs.add(child.attrib['name'])
# Iterate through all messages and check that they're documented
missing = 0
for child in enroot:
if child.attrib['name'] not in qqmsgs:
print(child.attrib['name'] + " is undocumented!")
missing += 1
sys.exit(1 if missing else 0)
| <commit_before>#!/usr/bin/env python
import os
import xml.etree.ElementTree as ET
RES_FOLDER = os.path.abspath(os.path.join(os.path.dirname(__file__), "../app/src/main/res"))
EN_STRINGS = os.path.join(RES_FOLDER, "values/strings.xml")
QQ_STRINGS = os.path.join(RES_FOLDER, "values-qq/strings.xml")
# Get ElementTree containing all message names in English
enroot = ET.parse(EN_STRINGS).getroot()
# Get ElementTree containing all documented messages
qqroot = ET.parse(QQ_STRINGS).getroot()
# Create a set to store all documented messages
qqmsgs = set()
# Add all documented messages to that set
for child in qqroot:
qqmsgs.add(child.attrib['name'])
# Iterate through all messages and check that they're documented
for child in enroot:
if child.attrib['name'] not in qqmsgs:
print(child.attrib['name'] + " is undocumented!")
<commit_msg>Exit with nonzero when qq strings are missing
Change-Id: Ife0f114dbe48faa445397aa7a94f74de2309d117<commit_after>#!/usr/bin/env python
import os
import sys
import xml.etree.ElementTree as ET
RES_FOLDER = os.path.abspath(os.path.join(os.path.dirname(__file__), "../app/src/main/res"))
EN_STRINGS = os.path.join(RES_FOLDER, "values/strings.xml")
QQ_STRINGS = os.path.join(RES_FOLDER, "values-qq/strings.xml")
# Get ElementTree containing all message names in English
enroot = ET.parse(EN_STRINGS).getroot()
# Get ElementTree containing all documented messages
qqroot = ET.parse(QQ_STRINGS).getroot()
# Create a set to store all documented messages
qqmsgs = set()
# Add all documented messages to that set
for child in qqroot:
qqmsgs.add(child.attrib['name'])
# Iterate through all messages and check that they're documented
missing = 0
for child in enroot:
if child.attrib['name'] not in qqmsgs:
print(child.attrib['name'] + " is undocumented!")
missing += 1
sys.exit(1 if missing else 0)
|
91c95c383803b6d25f8f68752ca7d507eaa9c7c8 | harvester/tests/test_cubeupload.py | harvester/tests/test_cubeupload.py | #!/usr/bin/env python3
#-*- coding: utf-8 -*-
import hashlib
import unittest
from harvester import harvester
class CupeuploadTest(unittest.TestCase):
def setUp(self):
self.nick = "test"
self.chan = '#brotherBot'
self.mask = "brotherBox!~brotherBo@unaffiliated/brotherbox"
self.h = harvester.HarvesterBot
def test_fetch_cubeupload_share(self):
msg = "http://cubeupload.com/im/YhUxlj.jpg"
test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca"
c = self.h._retrieve_content(self.h, self.mask, msg, self.chan)
md5 = hashlib.md5()
md5.update(c[0]['content'])
self.assertEqual(md5.hexdigest(), test_hash)
def test_fetch_cubeupload_raw(self):
msg = "http://i.cubeupload.com/YhUxlj.jpg"
test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca"
c = self.h._retrieve_content(self.h, self.mask, msg, self.chan)
md5 = hashlib.md5()
md5.update(c[0]['content'])
self.assertEqual(md5.hexdigest(), test_hash)
if __name__ == '__main__':
unittest.main()
| #!/usr/bin/env python3
#-*- coding: utf-8 -*-
import hashlib
import unittest
from harvester import harvester
class CubeuploadTest(unittest.TestCase):
def setUp(self):
self.nick = "test"
self.chan = '#brotherBot'
self.mask = "brotherBox!~brotherBo@unaffiliated/brotherbox"
self.h = harvester.HarvesterBot
def test_fetch_cubeupload_share(self):
msg = "http://cubeupload.com/im/YhUxlj.jpg"
test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca"
c = self.h._retrieve_content(self.h, self.mask, msg, self.chan)
md5 = hashlib.md5()
md5.update(c[0]['content'])
self.assertEqual(md5.hexdigest(), test_hash)
def test_fetch_cubeupload_raw(self):
msg = "http://i.cubeupload.com/YhUxlj.jpg"
test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca"
c = self.h._retrieve_content(self.h, self.mask, msg, self.chan)
md5 = hashlib.md5()
md5.update(c[0]['content'])
self.assertEqual(md5.hexdigest(), test_hash)
if __name__ == '__main__':
unittest.main()
| Fix minor spelling error in cubeupload test | Fix minor spelling error in cubeupload test
| Python | unlicense | luceatnobis/bravester,luceatnobis/harvester | #!/usr/bin/env python3
#-*- coding: utf-8 -*-
import hashlib
import unittest
from harvester import harvester
class CupeuploadTest(unittest.TestCase):
def setUp(self):
self.nick = "test"
self.chan = '#brotherBot'
self.mask = "brotherBox!~brotherBo@unaffiliated/brotherbox"
self.h = harvester.HarvesterBot
def test_fetch_cubeupload_share(self):
msg = "http://cubeupload.com/im/YhUxlj.jpg"
test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca"
c = self.h._retrieve_content(self.h, self.mask, msg, self.chan)
md5 = hashlib.md5()
md5.update(c[0]['content'])
self.assertEqual(md5.hexdigest(), test_hash)
def test_fetch_cubeupload_raw(self):
msg = "http://i.cubeupload.com/YhUxlj.jpg"
test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca"
c = self.h._retrieve_content(self.h, self.mask, msg, self.chan)
md5 = hashlib.md5()
md5.update(c[0]['content'])
self.assertEqual(md5.hexdigest(), test_hash)
if __name__ == '__main__':
unittest.main()
Fix minor spelling error in cubeupload test | #!/usr/bin/env python3
#-*- coding: utf-8 -*-
import hashlib
import unittest
from harvester import harvester
class CubeuploadTest(unittest.TestCase):
def setUp(self):
self.nick = "test"
self.chan = '#brotherBot'
self.mask = "brotherBox!~brotherBo@unaffiliated/brotherbox"
self.h = harvester.HarvesterBot
def test_fetch_cubeupload_share(self):
msg = "http://cubeupload.com/im/YhUxlj.jpg"
test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca"
c = self.h._retrieve_content(self.h, self.mask, msg, self.chan)
md5 = hashlib.md5()
md5.update(c[0]['content'])
self.assertEqual(md5.hexdigest(), test_hash)
def test_fetch_cubeupload_raw(self):
msg = "http://i.cubeupload.com/YhUxlj.jpg"
test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca"
c = self.h._retrieve_content(self.h, self.mask, msg, self.chan)
md5 = hashlib.md5()
md5.update(c[0]['content'])
self.assertEqual(md5.hexdigest(), test_hash)
if __name__ == '__main__':
unittest.main()
| <commit_before>#!/usr/bin/env python3
#-*- coding: utf-8 -*-
import hashlib
import unittest
from harvester import harvester
class CupeuploadTest(unittest.TestCase):
def setUp(self):
self.nick = "test"
self.chan = '#brotherBot'
self.mask = "brotherBox!~brotherBo@unaffiliated/brotherbox"
self.h = harvester.HarvesterBot
def test_fetch_cubeupload_share(self):
msg = "http://cubeupload.com/im/YhUxlj.jpg"
test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca"
c = self.h._retrieve_content(self.h, self.mask, msg, self.chan)
md5 = hashlib.md5()
md5.update(c[0]['content'])
self.assertEqual(md5.hexdigest(), test_hash)
def test_fetch_cubeupload_raw(self):
msg = "http://i.cubeupload.com/YhUxlj.jpg"
test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca"
c = self.h._retrieve_content(self.h, self.mask, msg, self.chan)
md5 = hashlib.md5()
md5.update(c[0]['content'])
self.assertEqual(md5.hexdigest(), test_hash)
if __name__ == '__main__':
unittest.main()
<commit_msg>Fix minor spelling error in cubeupload test<commit_after> | #!/usr/bin/env python3
#-*- coding: utf-8 -*-
import hashlib
import unittest
from harvester import harvester
class CubeuploadTest(unittest.TestCase):
def setUp(self):
self.nick = "test"
self.chan = '#brotherBot'
self.mask = "brotherBox!~brotherBo@unaffiliated/brotherbox"
self.h = harvester.HarvesterBot
def test_fetch_cubeupload_share(self):
msg = "http://cubeupload.com/im/YhUxlj.jpg"
test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca"
c = self.h._retrieve_content(self.h, self.mask, msg, self.chan)
md5 = hashlib.md5()
md5.update(c[0]['content'])
self.assertEqual(md5.hexdigest(), test_hash)
def test_fetch_cubeupload_raw(self):
msg = "http://i.cubeupload.com/YhUxlj.jpg"
test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca"
c = self.h._retrieve_content(self.h, self.mask, msg, self.chan)
md5 = hashlib.md5()
md5.update(c[0]['content'])
self.assertEqual(md5.hexdigest(), test_hash)
if __name__ == '__main__':
unittest.main()
| #!/usr/bin/env python3
#-*- coding: utf-8 -*-
import hashlib
import unittest
from harvester import harvester
class CupeuploadTest(unittest.TestCase):
def setUp(self):
self.nick = "test"
self.chan = '#brotherBot'
self.mask = "brotherBox!~brotherBo@unaffiliated/brotherbox"
self.h = harvester.HarvesterBot
def test_fetch_cubeupload_share(self):
msg = "http://cubeupload.com/im/YhUxlj.jpg"
test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca"
c = self.h._retrieve_content(self.h, self.mask, msg, self.chan)
md5 = hashlib.md5()
md5.update(c[0]['content'])
self.assertEqual(md5.hexdigest(), test_hash)
def test_fetch_cubeupload_raw(self):
msg = "http://i.cubeupload.com/YhUxlj.jpg"
test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca"
c = self.h._retrieve_content(self.h, self.mask, msg, self.chan)
md5 = hashlib.md5()
md5.update(c[0]['content'])
self.assertEqual(md5.hexdigest(), test_hash)
if __name__ == '__main__':
unittest.main()
Fix minor spelling error in cubeupload test#!/usr/bin/env python3
#-*- coding: utf-8 -*-
import hashlib
import unittest
from harvester import harvester
class CubeuploadTest(unittest.TestCase):
def setUp(self):
self.nick = "test"
self.chan = '#brotherBot'
self.mask = "brotherBox!~brotherBo@unaffiliated/brotherbox"
self.h = harvester.HarvesterBot
def test_fetch_cubeupload_share(self):
msg = "http://cubeupload.com/im/YhUxlj.jpg"
test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca"
c = self.h._retrieve_content(self.h, self.mask, msg, self.chan)
md5 = hashlib.md5()
md5.update(c[0]['content'])
self.assertEqual(md5.hexdigest(), test_hash)
def test_fetch_cubeupload_raw(self):
msg = "http://i.cubeupload.com/YhUxlj.jpg"
test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca"
c = self.h._retrieve_content(self.h, self.mask, msg, self.chan)
md5 = hashlib.md5()
md5.update(c[0]['content'])
self.assertEqual(md5.hexdigest(), test_hash)
if __name__ == '__main__':
unittest.main()
| <commit_before>#!/usr/bin/env python3
#-*- coding: utf-8 -*-
import hashlib
import unittest
from harvester import harvester
class CupeuploadTest(unittest.TestCase):
def setUp(self):
self.nick = "test"
self.chan = '#brotherBot'
self.mask = "brotherBox!~brotherBo@unaffiliated/brotherbox"
self.h = harvester.HarvesterBot
def test_fetch_cubeupload_share(self):
msg = "http://cubeupload.com/im/YhUxlj.jpg"
test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca"
c = self.h._retrieve_content(self.h, self.mask, msg, self.chan)
md5 = hashlib.md5()
md5.update(c[0]['content'])
self.assertEqual(md5.hexdigest(), test_hash)
def test_fetch_cubeupload_raw(self):
msg = "http://i.cubeupload.com/YhUxlj.jpg"
test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca"
c = self.h._retrieve_content(self.h, self.mask, msg, self.chan)
md5 = hashlib.md5()
md5.update(c[0]['content'])
self.assertEqual(md5.hexdigest(), test_hash)
if __name__ == '__main__':
unittest.main()
<commit_msg>Fix minor spelling error in cubeupload test<commit_after>#!/usr/bin/env python3
#-*- coding: utf-8 -*-
import hashlib
import unittest
from harvester import harvester
class CubeuploadTest(unittest.TestCase):
def setUp(self):
self.nick = "test"
self.chan = '#brotherBot'
self.mask = "brotherBox!~brotherBo@unaffiliated/brotherbox"
self.h = harvester.HarvesterBot
def test_fetch_cubeupload_share(self):
msg = "http://cubeupload.com/im/YhUxlj.jpg"
test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca"
c = self.h._retrieve_content(self.h, self.mask, msg, self.chan)
md5 = hashlib.md5()
md5.update(c[0]['content'])
self.assertEqual(md5.hexdigest(), test_hash)
def test_fetch_cubeupload_raw(self):
msg = "http://i.cubeupload.com/YhUxlj.jpg"
test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca"
c = self.h._retrieve_content(self.h, self.mask, msg, self.chan)
md5 = hashlib.md5()
md5.update(c[0]['content'])
self.assertEqual(md5.hexdigest(), test_hash)
if __name__ == '__main__':
unittest.main()
|
174c570d69d0958aa734794ffb7712ea37e70c6f | parse.py | parse.py | import sys
import configparser
def main():
config = configparser.ConfigParser(strict=False)
try:
section = sys.argv[1]
config_key = sys.argv[2]
config_value = sys.argv[3]
except IndexError:
print("Usage: cat test.ini | python parse.py <section> <option> <value>")
sys.exit(1)
try:
config.read_string(''.join(sys.stdin))
"""
Try to update or remove a config option from a section, or add the option to a new section.
"""
if section in config:
if config_key in config[section]:
if config_value == 'delete':
config.remove_option(section, config_key)
else:
config[section][config_key] = config_value
else:
config.add_section(section)
config.set(section, config_key, config_value)
config.write(sys.stdout)
except:
print("There was an error parsing the config.")
if __name__ == "__main__":
main()
| import sys
import configparser
def main():
config = configparser.ConfigParser(strict=False)
try:
section = sys.argv[1]
config_key = sys.argv[2]
config_value = sys.argv[3]
except IndexError:
print("Usage: cat test.ini | python parse.py <section> <option> <value>")
sys.exit(1)
try:
config.read_string(''.join(sys.stdin))
"""Try to update or remove a config option from a section, or add the option to a new section."""
if section in config:
if config_key in config[section]:
if config_value == 'delete':
config.remove_option(section, config_key)
else:
config[section][config_key] = config_value
else:
config.set(section, config_key, config_value)
else:
config.add_section(section)
config.set(section, config_key, config_value)
config.write(sys.stdout)
except:
print("There was an error parsing the config.")
if __name__ == "__main__":
main()
| Add new key to existing section. | Add new key to existing section.
| Python | mit | tonigrigoriu/ini-parser | import sys
import configparser
def main():
config = configparser.ConfigParser(strict=False)
try:
section = sys.argv[1]
config_key = sys.argv[2]
config_value = sys.argv[3]
except IndexError:
print("Usage: cat test.ini | python parse.py <section> <option> <value>")
sys.exit(1)
try:
config.read_string(''.join(sys.stdin))
"""
Try to update or remove a config option from a section, or add the option to a new section.
"""
if section in config:
if config_key in config[section]:
if config_value == 'delete':
config.remove_option(section, config_key)
else:
config[section][config_key] = config_value
else:
config.add_section(section)
config.set(section, config_key, config_value)
config.write(sys.stdout)
except:
print("There was an error parsing the config.")
if __name__ == "__main__":
main()
Add new key to existing section. | import sys
import configparser
def main():
config = configparser.ConfigParser(strict=False)
try:
section = sys.argv[1]
config_key = sys.argv[2]
config_value = sys.argv[3]
except IndexError:
print("Usage: cat test.ini | python parse.py <section> <option> <value>")
sys.exit(1)
try:
config.read_string(''.join(sys.stdin))
"""Try to update or remove a config option from a section, or add the option to a new section."""
if section in config:
if config_key in config[section]:
if config_value == 'delete':
config.remove_option(section, config_key)
else:
config[section][config_key] = config_value
else:
config.set(section, config_key, config_value)
else:
config.add_section(section)
config.set(section, config_key, config_value)
config.write(sys.stdout)
except:
print("There was an error parsing the config.")
if __name__ == "__main__":
main()
| <commit_before>import sys
import configparser
def main():
config = configparser.ConfigParser(strict=False)
try:
section = sys.argv[1]
config_key = sys.argv[2]
config_value = sys.argv[3]
except IndexError:
print("Usage: cat test.ini | python parse.py <section> <option> <value>")
sys.exit(1)
try:
config.read_string(''.join(sys.stdin))
"""
Try to update or remove a config option from a section, or add the option to a new section.
"""
if section in config:
if config_key in config[section]:
if config_value == 'delete':
config.remove_option(section, config_key)
else:
config[section][config_key] = config_value
else:
config.add_section(section)
config.set(section, config_key, config_value)
config.write(sys.stdout)
except:
print("There was an error parsing the config.")
if __name__ == "__main__":
main()
<commit_msg>Add new key to existing section.<commit_after> | import sys
import configparser
def main():
config = configparser.ConfigParser(strict=False)
try:
section = sys.argv[1]
config_key = sys.argv[2]
config_value = sys.argv[3]
except IndexError:
print("Usage: cat test.ini | python parse.py <section> <option> <value>")
sys.exit(1)
try:
config.read_string(''.join(sys.stdin))
"""Try to update or remove a config option from a section, or add the option to a new section."""
if section in config:
if config_key in config[section]:
if config_value == 'delete':
config.remove_option(section, config_key)
else:
config[section][config_key] = config_value
else:
config.set(section, config_key, config_value)
else:
config.add_section(section)
config.set(section, config_key, config_value)
config.write(sys.stdout)
except:
print("There was an error parsing the config.")
if __name__ == "__main__":
main()
| import sys
import configparser
def main():
config = configparser.ConfigParser(strict=False)
try:
section = sys.argv[1]
config_key = sys.argv[2]
config_value = sys.argv[3]
except IndexError:
print("Usage: cat test.ini | python parse.py <section> <option> <value>")
sys.exit(1)
try:
config.read_string(''.join(sys.stdin))
"""
Try to update or remove a config option from a section, or add the option to a new section.
"""
if section in config:
if config_key in config[section]:
if config_value == 'delete':
config.remove_option(section, config_key)
else:
config[section][config_key] = config_value
else:
config.add_section(section)
config.set(section, config_key, config_value)
config.write(sys.stdout)
except:
print("There was an error parsing the config.")
if __name__ == "__main__":
main()
Add new key to existing section.import sys
import configparser
def main():
config = configparser.ConfigParser(strict=False)
try:
section = sys.argv[1]
config_key = sys.argv[2]
config_value = sys.argv[3]
except IndexError:
print("Usage: cat test.ini | python parse.py <section> <option> <value>")
sys.exit(1)
try:
config.read_string(''.join(sys.stdin))
"""Try to update or remove a config option from a section, or add the option to a new section."""
if section in config:
if config_key in config[section]:
if config_value == 'delete':
config.remove_option(section, config_key)
else:
config[section][config_key] = config_value
else:
config.set(section, config_key, config_value)
else:
config.add_section(section)
config.set(section, config_key, config_value)
config.write(sys.stdout)
except:
print("There was an error parsing the config.")
if __name__ == "__main__":
main()
| <commit_before>import sys
import configparser
def main():
config = configparser.ConfigParser(strict=False)
try:
section = sys.argv[1]
config_key = sys.argv[2]
config_value = sys.argv[3]
except IndexError:
print("Usage: cat test.ini | python parse.py <section> <option> <value>")
sys.exit(1)
try:
config.read_string(''.join(sys.stdin))
"""
Try to update or remove a config option from a section, or add the option to a new section.
"""
if section in config:
if config_key in config[section]:
if config_value == 'delete':
config.remove_option(section, config_key)
else:
config[section][config_key] = config_value
else:
config.add_section(section)
config.set(section, config_key, config_value)
config.write(sys.stdout)
except:
print("There was an error parsing the config.")
if __name__ == "__main__":
main()
<commit_msg>Add new key to existing section.<commit_after>import sys
import configparser
def main():
config = configparser.ConfigParser(strict=False)
try:
section = sys.argv[1]
config_key = sys.argv[2]
config_value = sys.argv[3]
except IndexError:
print("Usage: cat test.ini | python parse.py <section> <option> <value>")
sys.exit(1)
try:
config.read_string(''.join(sys.stdin))
"""Try to update or remove a config option from a section, or add the option to a new section."""
if section in config:
if config_key in config[section]:
if config_value == 'delete':
config.remove_option(section, config_key)
else:
config[section][config_key] = config_value
else:
config.set(section, config_key, config_value)
else:
config.add_section(section)
config.set(section, config_key, config_value)
config.write(sys.stdout)
except:
print("There was an error parsing the config.")
if __name__ == "__main__":
main()
|
2e71f1a9deaf160ee666423e94ae526041cd32ff | engine.py | engine.py | # Use x, y coords for unit positions
# (97, 56) ... (104, 56)
# ... ...
# (97, 49) ... (104, 49)
#
# Algebraic notation for a position is:
# algebraic_pos = chr(x) + chr(y)
| # Use x, y coords for unit positions
# (97, 56) ... (104, 56)
# ... ...
# (97, 49) ... (104, 49)
#
# Algebraic notation for a position is:
# algebraic_pos = chr(x) + chr(y)
def _coord_to_algebraic(coord):
x, y = coord
return chr(x) + chr(y)
def _algebraic_to_coord(algebraic):
x, y = algebraic[0], algebraic[1]
return ord(x), ord(y)
| Add _coord_to_algebraic() and _algebraic_to_coord() to convert positions between (x, y) coords and 'a1' algebraic chess notation | Add _coord_to_algebraic() and _algebraic_to_coord() to convert positions between (x, y) coords and 'a1' algebraic chess notation
| Python | mit | EyuelAbebe/gamer,EyuelAbebe/gamer | # Use x, y coords for unit positions
# (97, 56) ... (104, 56)
# ... ...
# (97, 49) ... (104, 49)
#
# Algebraic notation for a position is:
# algebraic_pos = chr(x) + chr(y)
Add _coord_to_algebraic() and _algebraic_to_coord() to convert positions between (x, y) coords and 'a1' algebraic chess notation | # Use x, y coords for unit positions
# (97, 56) ... (104, 56)
# ... ...
# (97, 49) ... (104, 49)
#
# Algebraic notation for a position is:
# algebraic_pos = chr(x) + chr(y)
def _coord_to_algebraic(coord):
x, y = coord
return chr(x) + chr(y)
def _algebraic_to_coord(algebraic):
x, y = algebraic[0], algebraic[1]
return ord(x), ord(y)
| <commit_before># Use x, y coords for unit positions
# (97, 56) ... (104, 56)
# ... ...
# (97, 49) ... (104, 49)
#
# Algebraic notation for a position is:
# algebraic_pos = chr(x) + chr(y)
<commit_msg>Add _coord_to_algebraic() and _algebraic_to_coord() to convert positions between (x, y) coords and 'a1' algebraic chess notation<commit_after> | # Use x, y coords for unit positions
# (97, 56) ... (104, 56)
# ... ...
# (97, 49) ... (104, 49)
#
# Algebraic notation for a position is:
# algebraic_pos = chr(x) + chr(y)
def _coord_to_algebraic(coord):
x, y = coord
return chr(x) + chr(y)
def _algebraic_to_coord(algebraic):
x, y = algebraic[0], algebraic[1]
return ord(x), ord(y)
| # Use x, y coords for unit positions
# (97, 56) ... (104, 56)
# ... ...
# (97, 49) ... (104, 49)
#
# Algebraic notation for a position is:
# algebraic_pos = chr(x) + chr(y)
Add _coord_to_algebraic() and _algebraic_to_coord() to convert positions between (x, y) coords and 'a1' algebraic chess notation# Use x, y coords for unit positions
# (97, 56) ... (104, 56)
# ... ...
# (97, 49) ... (104, 49)
#
# Algebraic notation for a position is:
# algebraic_pos = chr(x) + chr(y)
def _coord_to_algebraic(coord):
x, y = coord
return chr(x) + chr(y)
def _algebraic_to_coord(algebraic):
x, y = algebraic[0], algebraic[1]
return ord(x), ord(y)
| <commit_before># Use x, y coords for unit positions
# (97, 56) ... (104, 56)
# ... ...
# (97, 49) ... (104, 49)
#
# Algebraic notation for a position is:
# algebraic_pos = chr(x) + chr(y)
<commit_msg>Add _coord_to_algebraic() and _algebraic_to_coord() to convert positions between (x, y) coords and 'a1' algebraic chess notation<commit_after># Use x, y coords for unit positions
# (97, 56) ... (104, 56)
# ... ...
# (97, 49) ... (104, 49)
#
# Algebraic notation for a position is:
# algebraic_pos = chr(x) + chr(y)
def _coord_to_algebraic(coord):
x, y = coord
return chr(x) + chr(y)
def _algebraic_to_coord(algebraic):
x, y = algebraic[0], algebraic[1]
return ord(x), ord(y)
|
215e37fce8b3fedf7bf31bf7c6393271c84141a2 | src/tapdisk/plugin.py | src/tapdisk/plugin.py | #!/usr/bin/env python
import os
import sys
import xapi
import xapi.plugin
from xapi.storage.datapath import log
class Implementation(xapi.plugin.Plugin_skeleton):
def query(self, dbg):
return {
"plugin": "tapdisk",
"name": "The tapdisk user-space datapath plugin",
"description": ("This plugin manages and configures tapdisk"
" instances backend by either raw or vhd"
" format files"),
"vendor": "Citrix",
"copyright": "(C) 2015 Citrix Inc",
"version": "3.0",
"required_api_version": "3.0",
"features": [
],
"configuration": {},
"required_cluster_stack": []}
if __name__ == "__main__":
log.log_call_argv()
cmd = xapi.plugin.Plugin_commandline(Implementation())
base = os.path.basename(sys.argv[0])
if base == "Plugin.Query":
cmd.query()
else:
raise xapi.plugin.Unimplemented(base)
| #!/usr/bin/env python
import os
import sys
import xapi
import xapi.storage.api.plugin
from xapi.storage import log
class Implementation(xapi.storage.api.plugin.Plugin_skeleton):
def query(self, dbg):
return {
"plugin": "tapdisk",
"name": "The tapdisk user-space datapath plugin",
"description": ("This plugin manages and configures tapdisk"
" instances backend by either raw or vhd"
" format files"),
"vendor": "Citrix",
"copyright": "(C) 2015 Citrix Inc",
"version": "3.0",
"required_api_version": "3.0",
"features": [
],
"configuration": {},
"required_cluster_stack": []}
if __name__ == "__main__":
log.log_call_argv()
cmd = xapi.storage.api.plugin.Plugin_commandline(Implementation())
base = os.path.basename(sys.argv[0])
if base == "Plugin.Query":
cmd.query()
else:
raise xapi.storage.api.plugin.Unimplemented(base)
| Use the new xapi.storage package structure | Use the new xapi.storage package structure
Signed-off-by: David Scott <63c9eb0ea83039690fefa11afe17873ba8278a56@eu.citrix.com>
| Python | lgpl-2.1 | xapi-project/xapi-storage-datapath-plugins,stefanopanella/xapi-storage-plugins,djs55/xapi-storage-datapath-plugins,stefanopanella/xapi-storage-plugins,stefanopanella/xapi-storage-plugins,jjd27/xapi-storage-datapath-plugins,robertbreker/xapi-storage-datapath-plugins | #!/usr/bin/env python
import os
import sys
import xapi
import xapi.plugin
from xapi.storage.datapath import log
class Implementation(xapi.plugin.Plugin_skeleton):
def query(self, dbg):
return {
"plugin": "tapdisk",
"name": "The tapdisk user-space datapath plugin",
"description": ("This plugin manages and configures tapdisk"
" instances backend by either raw or vhd"
" format files"),
"vendor": "Citrix",
"copyright": "(C) 2015 Citrix Inc",
"version": "3.0",
"required_api_version": "3.0",
"features": [
],
"configuration": {},
"required_cluster_stack": []}
if __name__ == "__main__":
log.log_call_argv()
cmd = xapi.plugin.Plugin_commandline(Implementation())
base = os.path.basename(sys.argv[0])
if base == "Plugin.Query":
cmd.query()
else:
raise xapi.plugin.Unimplemented(base)
Use the new xapi.storage package structure
Signed-off-by: David Scott <63c9eb0ea83039690fefa11afe17873ba8278a56@eu.citrix.com> | #!/usr/bin/env python
import os
import sys
import xapi
import xapi.storage.api.plugin
from xapi.storage import log
class Implementation(xapi.storage.api.plugin.Plugin_skeleton):
def query(self, dbg):
return {
"plugin": "tapdisk",
"name": "The tapdisk user-space datapath plugin",
"description": ("This plugin manages and configures tapdisk"
" instances backend by either raw or vhd"
" format files"),
"vendor": "Citrix",
"copyright": "(C) 2015 Citrix Inc",
"version": "3.0",
"required_api_version": "3.0",
"features": [
],
"configuration": {},
"required_cluster_stack": []}
if __name__ == "__main__":
log.log_call_argv()
cmd = xapi.storage.api.plugin.Plugin_commandline(Implementation())
base = os.path.basename(sys.argv[0])
if base == "Plugin.Query":
cmd.query()
else:
raise xapi.storage.api.plugin.Unimplemented(base)
| <commit_before>#!/usr/bin/env python
import os
import sys
import xapi
import xapi.plugin
from xapi.storage.datapath import log
class Implementation(xapi.plugin.Plugin_skeleton):
def query(self, dbg):
return {
"plugin": "tapdisk",
"name": "The tapdisk user-space datapath plugin",
"description": ("This plugin manages and configures tapdisk"
" instances backend by either raw or vhd"
" format files"),
"vendor": "Citrix",
"copyright": "(C) 2015 Citrix Inc",
"version": "3.0",
"required_api_version": "3.0",
"features": [
],
"configuration": {},
"required_cluster_stack": []}
if __name__ == "__main__":
log.log_call_argv()
cmd = xapi.plugin.Plugin_commandline(Implementation())
base = os.path.basename(sys.argv[0])
if base == "Plugin.Query":
cmd.query()
else:
raise xapi.plugin.Unimplemented(base)
<commit_msg>Use the new xapi.storage package structure
Signed-off-by: David Scott <63c9eb0ea83039690fefa11afe17873ba8278a56@eu.citrix.com><commit_after> | #!/usr/bin/env python
import os
import sys
import xapi
import xapi.storage.api.plugin
from xapi.storage import log
class Implementation(xapi.storage.api.plugin.Plugin_skeleton):
def query(self, dbg):
return {
"plugin": "tapdisk",
"name": "The tapdisk user-space datapath plugin",
"description": ("This plugin manages and configures tapdisk"
" instances backend by either raw or vhd"
" format files"),
"vendor": "Citrix",
"copyright": "(C) 2015 Citrix Inc",
"version": "3.0",
"required_api_version": "3.0",
"features": [
],
"configuration": {},
"required_cluster_stack": []}
if __name__ == "__main__":
log.log_call_argv()
cmd = xapi.storage.api.plugin.Plugin_commandline(Implementation())
base = os.path.basename(sys.argv[0])
if base == "Plugin.Query":
cmd.query()
else:
raise xapi.storage.api.plugin.Unimplemented(base)
| #!/usr/bin/env python
import os
import sys
import xapi
import xapi.plugin
from xapi.storage.datapath import log
class Implementation(xapi.plugin.Plugin_skeleton):
def query(self, dbg):
return {
"plugin": "tapdisk",
"name": "The tapdisk user-space datapath plugin",
"description": ("This plugin manages and configures tapdisk"
" instances backend by either raw or vhd"
" format files"),
"vendor": "Citrix",
"copyright": "(C) 2015 Citrix Inc",
"version": "3.0",
"required_api_version": "3.0",
"features": [
],
"configuration": {},
"required_cluster_stack": []}
if __name__ == "__main__":
log.log_call_argv()
cmd = xapi.plugin.Plugin_commandline(Implementation())
base = os.path.basename(sys.argv[0])
if base == "Plugin.Query":
cmd.query()
else:
raise xapi.plugin.Unimplemented(base)
Use the new xapi.storage package structure
Signed-off-by: David Scott <63c9eb0ea83039690fefa11afe17873ba8278a56@eu.citrix.com>#!/usr/bin/env python
import os
import sys
import xapi
import xapi.storage.api.plugin
from xapi.storage import log
class Implementation(xapi.storage.api.plugin.Plugin_skeleton):
def query(self, dbg):
return {
"plugin": "tapdisk",
"name": "The tapdisk user-space datapath plugin",
"description": ("This plugin manages and configures tapdisk"
" instances backend by either raw or vhd"
" format files"),
"vendor": "Citrix",
"copyright": "(C) 2015 Citrix Inc",
"version": "3.0",
"required_api_version": "3.0",
"features": [
],
"configuration": {},
"required_cluster_stack": []}
if __name__ == "__main__":
log.log_call_argv()
cmd = xapi.storage.api.plugin.Plugin_commandline(Implementation())
base = os.path.basename(sys.argv[0])
if base == "Plugin.Query":
cmd.query()
else:
raise xapi.storage.api.plugin.Unimplemented(base)
| <commit_before>#!/usr/bin/env python
import os
import sys
import xapi
import xapi.plugin
from xapi.storage.datapath import log
class Implementation(xapi.plugin.Plugin_skeleton):
def query(self, dbg):
return {
"plugin": "tapdisk",
"name": "The tapdisk user-space datapath plugin",
"description": ("This plugin manages and configures tapdisk"
" instances backend by either raw or vhd"
" format files"),
"vendor": "Citrix",
"copyright": "(C) 2015 Citrix Inc",
"version": "3.0",
"required_api_version": "3.0",
"features": [
],
"configuration": {},
"required_cluster_stack": []}
if __name__ == "__main__":
log.log_call_argv()
cmd = xapi.plugin.Plugin_commandline(Implementation())
base = os.path.basename(sys.argv[0])
if base == "Plugin.Query":
cmd.query()
else:
raise xapi.plugin.Unimplemented(base)
<commit_msg>Use the new xapi.storage package structure
Signed-off-by: David Scott <63c9eb0ea83039690fefa11afe17873ba8278a56@eu.citrix.com><commit_after>#!/usr/bin/env python
import os
import sys
import xapi
import xapi.storage.api.plugin
from xapi.storage import log
class Implementation(xapi.storage.api.plugin.Plugin_skeleton):
def query(self, dbg):
return {
"plugin": "tapdisk",
"name": "The tapdisk user-space datapath plugin",
"description": ("This plugin manages and configures tapdisk"
" instances backend by either raw or vhd"
" format files"),
"vendor": "Citrix",
"copyright": "(C) 2015 Citrix Inc",
"version": "3.0",
"required_api_version": "3.0",
"features": [
],
"configuration": {},
"required_cluster_stack": []}
if __name__ == "__main__":
log.log_call_argv()
cmd = xapi.storage.api.plugin.Plugin_commandline(Implementation())
base = os.path.basename(sys.argv[0])
if base == "Plugin.Query":
cmd.query()
else:
raise xapi.storage.api.plugin.Unimplemented(base)
|
273b463d8c86e60639b0c0ff54dad3e376a93060 | correctiv_justizgelder/views.py | correctiv_justizgelder/views.py | from django.views.generic import ListView, DetailView
from .forms import OrganisationSearchForm
from .models import Organisation
from .search_indexes import OrganisationIndex
from .search_utils import SearchPaginator
class OrganisationSearchView(ListView):
template_name = 'justizgelder/search.html'
paginate_by = 15
paginator_class = SearchPaginator
def get_queryset(self):
self.form = OrganisationSearchForm(self.request.GET)
self.result = self.form.search(size=self.paginate_by)
return self.result
def get_context_data(self, **kwargs):
context = super(OrganisationSearchView, self).get_context_data(**kwargs)
context['result'] = self.result
context['query'] = self.request.GET.get('q')
context['form'] = self.form
context['base_template'] = 'justizgelder/search_base.html'
if self.request.GET.get('embed'):
context['base_template'] = 'justizgelder/embed_base.html'
return context
class OrganisationDetail(DetailView):
template_name = 'justizgelder/organisation_detail.html'
model = Organisation
def get_context_data(self, **kwargs):
context = super(OrganisationDetail, self).get_context_data(**kwargs)
idx = OrganisationIndex()
context['mlt'] = idx.search(self.object.name,
size=15, sort=False,
aggregations=False
)
return context
| from django.views.generic import ListView, DetailView
from .forms import OrganisationSearchForm
from .models import Organisation
from .search_indexes import OrganisationIndex
from .search_utils import SearchPaginator
class OrganisationSearchView(ListView):
template_name = 'justizgelder/search.html'
paginate_by = 25
paginator_class = SearchPaginator
def get_queryset(self):
self.form = OrganisationSearchForm(self.request.GET)
self.result = self.form.search(size=self.paginate_by)
return self.result
def get_context_data(self, **kwargs):
context = super(OrganisationSearchView, self).get_context_data(**kwargs)
context['result'] = self.result
context['query'] = self.request.GET.get('q')
context['form'] = self.form
context['base_template'] = 'justizgelder/search_base.html'
if self.request.GET.get('embed'):
context['base_template'] = 'justizgelder/embed_base.html'
return context
class OrganisationDetail(DetailView):
template_name = 'justizgelder/organisation_detail.html'
model = Organisation
def get_context_data(self, **kwargs):
context = super(OrganisationDetail, self).get_context_data(**kwargs)
idx = OrganisationIndex()
context['mlt'] = idx.search(self.object.name,
size=15, sort=False,
aggregations=False
)
return context
| Increase results to 25 per page to make page look nicer | Increase results to 25 per page to make page look nicer | Python | mit | correctiv/correctiv-justizgelder,correctiv/correctiv-justizgelder | from django.views.generic import ListView, DetailView
from .forms import OrganisationSearchForm
from .models import Organisation
from .search_indexes import OrganisationIndex
from .search_utils import SearchPaginator
class OrganisationSearchView(ListView):
template_name = 'justizgelder/search.html'
paginate_by = 15
paginator_class = SearchPaginator
def get_queryset(self):
self.form = OrganisationSearchForm(self.request.GET)
self.result = self.form.search(size=self.paginate_by)
return self.result
def get_context_data(self, **kwargs):
context = super(OrganisationSearchView, self).get_context_data(**kwargs)
context['result'] = self.result
context['query'] = self.request.GET.get('q')
context['form'] = self.form
context['base_template'] = 'justizgelder/search_base.html'
if self.request.GET.get('embed'):
context['base_template'] = 'justizgelder/embed_base.html'
return context
class OrganisationDetail(DetailView):
template_name = 'justizgelder/organisation_detail.html'
model = Organisation
def get_context_data(self, **kwargs):
context = super(OrganisationDetail, self).get_context_data(**kwargs)
idx = OrganisationIndex()
context['mlt'] = idx.search(self.object.name,
size=15, sort=False,
aggregations=False
)
return context
Increase results to 25 per page to make page look nicer | from django.views.generic import ListView, DetailView
from .forms import OrganisationSearchForm
from .models import Organisation
from .search_indexes import OrganisationIndex
from .search_utils import SearchPaginator
class OrganisationSearchView(ListView):
template_name = 'justizgelder/search.html'
paginate_by = 25
paginator_class = SearchPaginator
def get_queryset(self):
self.form = OrganisationSearchForm(self.request.GET)
self.result = self.form.search(size=self.paginate_by)
return self.result
def get_context_data(self, **kwargs):
context = super(OrganisationSearchView, self).get_context_data(**kwargs)
context['result'] = self.result
context['query'] = self.request.GET.get('q')
context['form'] = self.form
context['base_template'] = 'justizgelder/search_base.html'
if self.request.GET.get('embed'):
context['base_template'] = 'justizgelder/embed_base.html'
return context
class OrganisationDetail(DetailView):
template_name = 'justizgelder/organisation_detail.html'
model = Organisation
def get_context_data(self, **kwargs):
context = super(OrganisationDetail, self).get_context_data(**kwargs)
idx = OrganisationIndex()
context['mlt'] = idx.search(self.object.name,
size=15, sort=False,
aggregations=False
)
return context
| <commit_before>from django.views.generic import ListView, DetailView
from .forms import OrganisationSearchForm
from .models import Organisation
from .search_indexes import OrganisationIndex
from .search_utils import SearchPaginator
class OrganisationSearchView(ListView):
template_name = 'justizgelder/search.html'
paginate_by = 15
paginator_class = SearchPaginator
def get_queryset(self):
self.form = OrganisationSearchForm(self.request.GET)
self.result = self.form.search(size=self.paginate_by)
return self.result
def get_context_data(self, **kwargs):
context = super(OrganisationSearchView, self).get_context_data(**kwargs)
context['result'] = self.result
context['query'] = self.request.GET.get('q')
context['form'] = self.form
context['base_template'] = 'justizgelder/search_base.html'
if self.request.GET.get('embed'):
context['base_template'] = 'justizgelder/embed_base.html'
return context
class OrganisationDetail(DetailView):
template_name = 'justizgelder/organisation_detail.html'
model = Organisation
def get_context_data(self, **kwargs):
context = super(OrganisationDetail, self).get_context_data(**kwargs)
idx = OrganisationIndex()
context['mlt'] = idx.search(self.object.name,
size=15, sort=False,
aggregations=False
)
return context
<commit_msg>Increase results to 25 per page to make page look nicer<commit_after> | from django.views.generic import ListView, DetailView
from .forms import OrganisationSearchForm
from .models import Organisation
from .search_indexes import OrganisationIndex
from .search_utils import SearchPaginator
class OrganisationSearchView(ListView):
template_name = 'justizgelder/search.html'
paginate_by = 25
paginator_class = SearchPaginator
def get_queryset(self):
self.form = OrganisationSearchForm(self.request.GET)
self.result = self.form.search(size=self.paginate_by)
return self.result
def get_context_data(self, **kwargs):
context = super(OrganisationSearchView, self).get_context_data(**kwargs)
context['result'] = self.result
context['query'] = self.request.GET.get('q')
context['form'] = self.form
context['base_template'] = 'justizgelder/search_base.html'
if self.request.GET.get('embed'):
context['base_template'] = 'justizgelder/embed_base.html'
return context
class OrganisationDetail(DetailView):
template_name = 'justizgelder/organisation_detail.html'
model = Organisation
def get_context_data(self, **kwargs):
context = super(OrganisationDetail, self).get_context_data(**kwargs)
idx = OrganisationIndex()
context['mlt'] = idx.search(self.object.name,
size=15, sort=False,
aggregations=False
)
return context
| from django.views.generic import ListView, DetailView
from .forms import OrganisationSearchForm
from .models import Organisation
from .search_indexes import OrganisationIndex
from .search_utils import SearchPaginator
class OrganisationSearchView(ListView):
template_name = 'justizgelder/search.html'
paginate_by = 15
paginator_class = SearchPaginator
def get_queryset(self):
self.form = OrganisationSearchForm(self.request.GET)
self.result = self.form.search(size=self.paginate_by)
return self.result
def get_context_data(self, **kwargs):
context = super(OrganisationSearchView, self).get_context_data(**kwargs)
context['result'] = self.result
context['query'] = self.request.GET.get('q')
context['form'] = self.form
context['base_template'] = 'justizgelder/search_base.html'
if self.request.GET.get('embed'):
context['base_template'] = 'justizgelder/embed_base.html'
return context
class OrganisationDetail(DetailView):
template_name = 'justizgelder/organisation_detail.html'
model = Organisation
def get_context_data(self, **kwargs):
context = super(OrganisationDetail, self).get_context_data(**kwargs)
idx = OrganisationIndex()
context['mlt'] = idx.search(self.object.name,
size=15, sort=False,
aggregations=False
)
return context
Increase results to 25 per page to make page look nicerfrom django.views.generic import ListView, DetailView
from .forms import OrganisationSearchForm
from .models import Organisation
from .search_indexes import OrganisationIndex
from .search_utils import SearchPaginator
class OrganisationSearchView(ListView):
template_name = 'justizgelder/search.html'
paginate_by = 25
paginator_class = SearchPaginator
def get_queryset(self):
self.form = OrganisationSearchForm(self.request.GET)
self.result = self.form.search(size=self.paginate_by)
return self.result
def get_context_data(self, **kwargs):
context = super(OrganisationSearchView, self).get_context_data(**kwargs)
context['result'] = self.result
context['query'] = self.request.GET.get('q')
context['form'] = self.form
context['base_template'] = 'justizgelder/search_base.html'
if self.request.GET.get('embed'):
context['base_template'] = 'justizgelder/embed_base.html'
return context
class OrganisationDetail(DetailView):
template_name = 'justizgelder/organisation_detail.html'
model = Organisation
def get_context_data(self, **kwargs):
context = super(OrganisationDetail, self).get_context_data(**kwargs)
idx = OrganisationIndex()
context['mlt'] = idx.search(self.object.name,
size=15, sort=False,
aggregations=False
)
return context
| <commit_before>from django.views.generic import ListView, DetailView
from .forms import OrganisationSearchForm
from .models import Organisation
from .search_indexes import OrganisationIndex
from .search_utils import SearchPaginator
class OrganisationSearchView(ListView):
template_name = 'justizgelder/search.html'
paginate_by = 15
paginator_class = SearchPaginator
def get_queryset(self):
self.form = OrganisationSearchForm(self.request.GET)
self.result = self.form.search(size=self.paginate_by)
return self.result
def get_context_data(self, **kwargs):
context = super(OrganisationSearchView, self).get_context_data(**kwargs)
context['result'] = self.result
context['query'] = self.request.GET.get('q')
context['form'] = self.form
context['base_template'] = 'justizgelder/search_base.html'
if self.request.GET.get('embed'):
context['base_template'] = 'justizgelder/embed_base.html'
return context
class OrganisationDetail(DetailView):
template_name = 'justizgelder/organisation_detail.html'
model = Organisation
def get_context_data(self, **kwargs):
context = super(OrganisationDetail, self).get_context_data(**kwargs)
idx = OrganisationIndex()
context['mlt'] = idx.search(self.object.name,
size=15, sort=False,
aggregations=False
)
return context
<commit_msg>Increase results to 25 per page to make page look nicer<commit_after>from django.views.generic import ListView, DetailView
from .forms import OrganisationSearchForm
from .models import Organisation
from .search_indexes import OrganisationIndex
from .search_utils import SearchPaginator
class OrganisationSearchView(ListView):
template_name = 'justizgelder/search.html'
paginate_by = 25
paginator_class = SearchPaginator
def get_queryset(self):
self.form = OrganisationSearchForm(self.request.GET)
self.result = self.form.search(size=self.paginate_by)
return self.result
def get_context_data(self, **kwargs):
context = super(OrganisationSearchView, self).get_context_data(**kwargs)
context['result'] = self.result
context['query'] = self.request.GET.get('q')
context['form'] = self.form
context['base_template'] = 'justizgelder/search_base.html'
if self.request.GET.get('embed'):
context['base_template'] = 'justizgelder/embed_base.html'
return context
class OrganisationDetail(DetailView):
template_name = 'justizgelder/organisation_detail.html'
model = Organisation
def get_context_data(self, **kwargs):
context = super(OrganisationDetail, self).get_context_data(**kwargs)
idx = OrganisationIndex()
context['mlt'] = idx.search(self.object.name,
size=15, sort=False,
aggregations=False
)
return context
|
90b1567ee8e1906b1d1724cf63cf8d383530da29 | nimp/commands/cis_tomat_mining.py | nimp/commands/cis_tomat_mining.py | # -*- coding: utf-8 -*-
from nimp.commands._cis_command import *
from nimp.utilities.ue3 import *
from nimp.utilities.deployment import *
from nimp.utilities.file_mapper import *
#-------------------------------------------------------------------------------
class CisTomatMining(CisCommand):
abstract = 0
def __init__(self):
CisCommand.__init__(self,
'cis-tomat-mining',
'Mines UE3 content into Tomat')
#---------------------------------------------------------------------------
def cis_configure_arguments(self, env, parser):
return True
#---------------------------------------------------------------------------
def _cis_run(self, env):
if call_process('.', ['pacman', '-S', '--noconfirm', 'repman']) != 0:
return False
if call_process('.', ['repman', 'add', 'dont-nod', 'http://pacman']) != 0:
return False
if call_process('.', ['pacman', '-S', '--noconfirm', '--needed', 'tomat-console']) != 0:
return False
return call_process('.',
[ 'TomatConsole',
'ImportFromUnreal',
'--RepositoryUri',
'sql://mining@console',
'--UnrealEnginePath',
'Binaries/Win64/ExampleGame.exe' ])
| # -*- coding: utf-8 -*-
from nimp.commands._cis_command import *
from nimp.utilities.ue3 import *
from nimp.utilities.deployment import *
from nimp.utilities.file_mapper import *
import tempfile
import shutil
#-------------------------------------------------------------------------------
class CisTomatMining(CisCommand):
abstract = 0
def __init__(self):
CisCommand.__init__(self,
'cis-tomat-mining',
'Mines UE3 content into Tomat')
#---------------------------------------------------------------------------
def cis_configure_arguments(self, env, parser):
return True
#---------------------------------------------------------------------------
def _cis_run(self, env):
if call_process('.', ['pacman', '-S', '--noconfirm', 'repman']) != 0:
return False
if call_process('.', ['repman', 'add', 'dont-nod', 'http://pacman/']) != 0:
return False
if call_process('.', ['pacman', '-Scc', '--noconfirm']) != 0:
return False
if call_process('.', ['pacman', '-S', '--noconfirm', '--force', 'tomat-console']) != 0:
return False
tmpdir = tempfile.mkdtemp()
success = True
if success and env.is_ue3:
success = ue3_commandlet(env.game, 'dnetomatminingcommandlet', [ tmpdir ])
if success:
success = call_process('.', [ 'TomatConsole',
'ImportFromUnreal',
'--RepositoryUri', 'sql://mining@console',
'--TmpDirectory', tmpdir ]) == 0
# Clean up after ourselves
shutil.rmtree(tmpdir)
return success
| Split the Tomat mining commandlet into an Unreal part and a TomatConsole part. | Split the Tomat mining commandlet into an Unreal part and a TomatConsole part.
| Python | mit | dontnod/nimp | # -*- coding: utf-8 -*-
from nimp.commands._cis_command import *
from nimp.utilities.ue3 import *
from nimp.utilities.deployment import *
from nimp.utilities.file_mapper import *
#-------------------------------------------------------------------------------
class CisTomatMining(CisCommand):
abstract = 0
def __init__(self):
CisCommand.__init__(self,
'cis-tomat-mining',
'Mines UE3 content into Tomat')
#---------------------------------------------------------------------------
def cis_configure_arguments(self, env, parser):
return True
#---------------------------------------------------------------------------
def _cis_run(self, env):
if call_process('.', ['pacman', '-S', '--noconfirm', 'repman']) != 0:
return False
if call_process('.', ['repman', 'add', 'dont-nod', 'http://pacman']) != 0:
return False
if call_process('.', ['pacman', '-S', '--noconfirm', '--needed', 'tomat-console']) != 0:
return False
return call_process('.',
[ 'TomatConsole',
'ImportFromUnreal',
'--RepositoryUri',
'sql://mining@console',
'--UnrealEnginePath',
'Binaries/Win64/ExampleGame.exe' ])
Split the Tomat mining commandlet into an Unreal part and a TomatConsole part. | # -*- coding: utf-8 -*-
from nimp.commands._cis_command import *
from nimp.utilities.ue3 import *
from nimp.utilities.deployment import *
from nimp.utilities.file_mapper import *
import tempfile
import shutil
#-------------------------------------------------------------------------------
class CisTomatMining(CisCommand):
abstract = 0
def __init__(self):
CisCommand.__init__(self,
'cis-tomat-mining',
'Mines UE3 content into Tomat')
#---------------------------------------------------------------------------
def cis_configure_arguments(self, env, parser):
return True
#---------------------------------------------------------------------------
def _cis_run(self, env):
if call_process('.', ['pacman', '-S', '--noconfirm', 'repman']) != 0:
return False
if call_process('.', ['repman', 'add', 'dont-nod', 'http://pacman/']) != 0:
return False
if call_process('.', ['pacman', '-Scc', '--noconfirm']) != 0:
return False
if call_process('.', ['pacman', '-S', '--noconfirm', '--force', 'tomat-console']) != 0:
return False
tmpdir = tempfile.mkdtemp()
success = True
if success and env.is_ue3:
success = ue3_commandlet(env.game, 'dnetomatminingcommandlet', [ tmpdir ])
if success:
success = call_process('.', [ 'TomatConsole',
'ImportFromUnreal',
'--RepositoryUri', 'sql://mining@console',
'--TmpDirectory', tmpdir ]) == 0
# Clean up after ourselves
shutil.rmtree(tmpdir)
return success
| <commit_before># -*- coding: utf-8 -*-
from nimp.commands._cis_command import *
from nimp.utilities.ue3 import *
from nimp.utilities.deployment import *
from nimp.utilities.file_mapper import *
#-------------------------------------------------------------------------------
class CisTomatMining(CisCommand):
abstract = 0
def __init__(self):
CisCommand.__init__(self,
'cis-tomat-mining',
'Mines UE3 content into Tomat')
#---------------------------------------------------------------------------
def cis_configure_arguments(self, env, parser):
return True
#---------------------------------------------------------------------------
def _cis_run(self, env):
if call_process('.', ['pacman', '-S', '--noconfirm', 'repman']) != 0:
return False
if call_process('.', ['repman', 'add', 'dont-nod', 'http://pacman']) != 0:
return False
if call_process('.', ['pacman', '-S', '--noconfirm', '--needed', 'tomat-console']) != 0:
return False
return call_process('.',
[ 'TomatConsole',
'ImportFromUnreal',
'--RepositoryUri',
'sql://mining@console',
'--UnrealEnginePath',
'Binaries/Win64/ExampleGame.exe' ])
<commit_msg>Split the Tomat mining commandlet into an Unreal part and a TomatConsole part.<commit_after> | # -*- coding: utf-8 -*-
from nimp.commands._cis_command import *
from nimp.utilities.ue3 import *
from nimp.utilities.deployment import *
from nimp.utilities.file_mapper import *
import tempfile
import shutil
#-------------------------------------------------------------------------------
class CisTomatMining(CisCommand):
abstract = 0
def __init__(self):
CisCommand.__init__(self,
'cis-tomat-mining',
'Mines UE3 content into Tomat')
#---------------------------------------------------------------------------
def cis_configure_arguments(self, env, parser):
return True
#---------------------------------------------------------------------------
def _cis_run(self, env):
if call_process('.', ['pacman', '-S', '--noconfirm', 'repman']) != 0:
return False
if call_process('.', ['repman', 'add', 'dont-nod', 'http://pacman/']) != 0:
return False
if call_process('.', ['pacman', '-Scc', '--noconfirm']) != 0:
return False
if call_process('.', ['pacman', '-S', '--noconfirm', '--force', 'tomat-console']) != 0:
return False
tmpdir = tempfile.mkdtemp()
success = True
if success and env.is_ue3:
success = ue3_commandlet(env.game, 'dnetomatminingcommandlet', [ tmpdir ])
if success:
success = call_process('.', [ 'TomatConsole',
'ImportFromUnreal',
'--RepositoryUri', 'sql://mining@console',
'--TmpDirectory', tmpdir ]) == 0
# Clean up after ourselves
shutil.rmtree(tmpdir)
return success
| # -*- coding: utf-8 -*-
from nimp.commands._cis_command import *
from nimp.utilities.ue3 import *
from nimp.utilities.deployment import *
from nimp.utilities.file_mapper import *
#-------------------------------------------------------------------------------
class CisTomatMining(CisCommand):
abstract = 0
def __init__(self):
CisCommand.__init__(self,
'cis-tomat-mining',
'Mines UE3 content into Tomat')
#---------------------------------------------------------------------------
def cis_configure_arguments(self, env, parser):
return True
#---------------------------------------------------------------------------
def _cis_run(self, env):
if call_process('.', ['pacman', '-S', '--noconfirm', 'repman']) != 0:
return False
if call_process('.', ['repman', 'add', 'dont-nod', 'http://pacman']) != 0:
return False
if call_process('.', ['pacman', '-S', '--noconfirm', '--needed', 'tomat-console']) != 0:
return False
return call_process('.',
[ 'TomatConsole',
'ImportFromUnreal',
'--RepositoryUri',
'sql://mining@console',
'--UnrealEnginePath',
'Binaries/Win64/ExampleGame.exe' ])
Split the Tomat mining commandlet into an Unreal part and a TomatConsole part.# -*- coding: utf-8 -*-
from nimp.commands._cis_command import *
from nimp.utilities.ue3 import *
from nimp.utilities.deployment import *
from nimp.utilities.file_mapper import *
import tempfile
import shutil
#-------------------------------------------------------------------------------
class CisTomatMining(CisCommand):
abstract = 0
def __init__(self):
CisCommand.__init__(self,
'cis-tomat-mining',
'Mines UE3 content into Tomat')
#---------------------------------------------------------------------------
def cis_configure_arguments(self, env, parser):
return True
#---------------------------------------------------------------------------
def _cis_run(self, env):
if call_process('.', ['pacman', '-S', '--noconfirm', 'repman']) != 0:
return False
if call_process('.', ['repman', 'add', 'dont-nod', 'http://pacman/']) != 0:
return False
if call_process('.', ['pacman', '-Scc', '--noconfirm']) != 0:
return False
if call_process('.', ['pacman', '-S', '--noconfirm', '--force', 'tomat-console']) != 0:
return False
tmpdir = tempfile.mkdtemp()
success = True
if success and env.is_ue3:
success = ue3_commandlet(env.game, 'dnetomatminingcommandlet', [ tmpdir ])
if success:
success = call_process('.', [ 'TomatConsole',
'ImportFromUnreal',
'--RepositoryUri', 'sql://mining@console',
'--TmpDirectory', tmpdir ]) == 0
# Clean up after ourselves
shutil.rmtree(tmpdir)
return success
| <commit_before># -*- coding: utf-8 -*-
from nimp.commands._cis_command import *
from nimp.utilities.ue3 import *
from nimp.utilities.deployment import *
from nimp.utilities.file_mapper import *
#-------------------------------------------------------------------------------
class CisTomatMining(CisCommand):
abstract = 0
def __init__(self):
CisCommand.__init__(self,
'cis-tomat-mining',
'Mines UE3 content into Tomat')
#---------------------------------------------------------------------------
def cis_configure_arguments(self, env, parser):
return True
#---------------------------------------------------------------------------
def _cis_run(self, env):
if call_process('.', ['pacman', '-S', '--noconfirm', 'repman']) != 0:
return False
if call_process('.', ['repman', 'add', 'dont-nod', 'http://pacman']) != 0:
return False
if call_process('.', ['pacman', '-S', '--noconfirm', '--needed', 'tomat-console']) != 0:
return False
return call_process('.',
[ 'TomatConsole',
'ImportFromUnreal',
'--RepositoryUri',
'sql://mining@console',
'--UnrealEnginePath',
'Binaries/Win64/ExampleGame.exe' ])
<commit_msg>Split the Tomat mining commandlet into an Unreal part and a TomatConsole part.<commit_after># -*- coding: utf-8 -*-
from nimp.commands._cis_command import *
from nimp.utilities.ue3 import *
from nimp.utilities.deployment import *
from nimp.utilities.file_mapper import *
import tempfile
import shutil
#-------------------------------------------------------------------------------
class CisTomatMining(CisCommand):
abstract = 0
def __init__(self):
CisCommand.__init__(self,
'cis-tomat-mining',
'Mines UE3 content into Tomat')
#---------------------------------------------------------------------------
def cis_configure_arguments(self, env, parser):
return True
#---------------------------------------------------------------------------
def _cis_run(self, env):
if call_process('.', ['pacman', '-S', '--noconfirm', 'repman']) != 0:
return False
if call_process('.', ['repman', 'add', 'dont-nod', 'http://pacman/']) != 0:
return False
if call_process('.', ['pacman', '-Scc', '--noconfirm']) != 0:
return False
if call_process('.', ['pacman', '-S', '--noconfirm', '--force', 'tomat-console']) != 0:
return False
tmpdir = tempfile.mkdtemp()
success = True
if success and env.is_ue3:
success = ue3_commandlet(env.game, 'dnetomatminingcommandlet', [ tmpdir ])
if success:
success = call_process('.', [ 'TomatConsole',
'ImportFromUnreal',
'--RepositoryUri', 'sql://mining@console',
'--TmpDirectory', tmpdir ]) == 0
# Clean up after ourselves
shutil.rmtree(tmpdir)
return success
|
e356ce2c6fc6a3383a4ab8f7eea1ecb3ef7aa978 | linter.py | linter.py | from SublimeLinter.lint import ComposerLinter
class Phpcs(ComposerLinter):
cmd = ('phpcs', '--report=emacs', '${args}', '-')
regex = r'^.*:(?P<line>[0-9]+):(?P<col>[0-9]+): (?:(?P<error>error)|(?P<warning>warning)) - (?P<message>(.(?!\(\S+\)$))+)( \((?P<code>\S+)\)$)?' # noqa: E501
defaults = {
'selector': 'source.php - text.blade, text.html.basic',
# we want auto-substitution of the filename,
# but `cmd` does not support that yet
'--stdin-path=': '${file}'
}
| from SublimeLinter.lint import ComposerLinter
class Phpcs(ComposerLinter):
cmd = ('phpcs', '--report=emacs', '${args}', '-')
regex = r'^.*:(?P<line>[0-9]+):(?P<col>[0-9]+): (?:(?P<error>error)|(?P<warning>warning)) - (?P<message>(.(?!\(\S+\)$))+)( \((?P<code>\S+)\)$)?' # noqa: E501
defaults = {
'selector': 'embedding.php, source.php - text.blade',
# we want auto-substitution of the filename,
# but `cmd` does not support that yet
'--stdin-path=': '${file}'
}
| Update selector for Sublime Text >= 4134 | Update selector for Sublime Text >= 4134
| Python | mit | SublimeLinter/SublimeLinter-phpcs | from SublimeLinter.lint import ComposerLinter
class Phpcs(ComposerLinter):
cmd = ('phpcs', '--report=emacs', '${args}', '-')
regex = r'^.*:(?P<line>[0-9]+):(?P<col>[0-9]+): (?:(?P<error>error)|(?P<warning>warning)) - (?P<message>(.(?!\(\S+\)$))+)( \((?P<code>\S+)\)$)?' # noqa: E501
defaults = {
'selector': 'source.php - text.blade, text.html.basic',
# we want auto-substitution of the filename,
# but `cmd` does not support that yet
'--stdin-path=': '${file}'
}
Update selector for Sublime Text >= 4134 | from SublimeLinter.lint import ComposerLinter
class Phpcs(ComposerLinter):
cmd = ('phpcs', '--report=emacs', '${args}', '-')
regex = r'^.*:(?P<line>[0-9]+):(?P<col>[0-9]+): (?:(?P<error>error)|(?P<warning>warning)) - (?P<message>(.(?!\(\S+\)$))+)( \((?P<code>\S+)\)$)?' # noqa: E501
defaults = {
'selector': 'embedding.php, source.php - text.blade',
# we want auto-substitution of the filename,
# but `cmd` does not support that yet
'--stdin-path=': '${file}'
}
| <commit_before>from SublimeLinter.lint import ComposerLinter
class Phpcs(ComposerLinter):
cmd = ('phpcs', '--report=emacs', '${args}', '-')
regex = r'^.*:(?P<line>[0-9]+):(?P<col>[0-9]+): (?:(?P<error>error)|(?P<warning>warning)) - (?P<message>(.(?!\(\S+\)$))+)( \((?P<code>\S+)\)$)?' # noqa: E501
defaults = {
'selector': 'source.php - text.blade, text.html.basic',
# we want auto-substitution of the filename,
# but `cmd` does not support that yet
'--stdin-path=': '${file}'
}
<commit_msg>Update selector for Sublime Text >= 4134<commit_after> | from SublimeLinter.lint import ComposerLinter
class Phpcs(ComposerLinter):
cmd = ('phpcs', '--report=emacs', '${args}', '-')
regex = r'^.*:(?P<line>[0-9]+):(?P<col>[0-9]+): (?:(?P<error>error)|(?P<warning>warning)) - (?P<message>(.(?!\(\S+\)$))+)( \((?P<code>\S+)\)$)?' # noqa: E501
defaults = {
'selector': 'embedding.php, source.php - text.blade',
# we want auto-substitution of the filename,
# but `cmd` does not support that yet
'--stdin-path=': '${file}'
}
| from SublimeLinter.lint import ComposerLinter
class Phpcs(ComposerLinter):
cmd = ('phpcs', '--report=emacs', '${args}', '-')
regex = r'^.*:(?P<line>[0-9]+):(?P<col>[0-9]+): (?:(?P<error>error)|(?P<warning>warning)) - (?P<message>(.(?!\(\S+\)$))+)( \((?P<code>\S+)\)$)?' # noqa: E501
defaults = {
'selector': 'source.php - text.blade, text.html.basic',
# we want auto-substitution of the filename,
# but `cmd` does not support that yet
'--stdin-path=': '${file}'
}
Update selector for Sublime Text >= 4134from SublimeLinter.lint import ComposerLinter
class Phpcs(ComposerLinter):
cmd = ('phpcs', '--report=emacs', '${args}', '-')
regex = r'^.*:(?P<line>[0-9]+):(?P<col>[0-9]+): (?:(?P<error>error)|(?P<warning>warning)) - (?P<message>(.(?!\(\S+\)$))+)( \((?P<code>\S+)\)$)?' # noqa: E501
defaults = {
'selector': 'embedding.php, source.php - text.blade',
# we want auto-substitution of the filename,
# but `cmd` does not support that yet
'--stdin-path=': '${file}'
}
| <commit_before>from SublimeLinter.lint import ComposerLinter
class Phpcs(ComposerLinter):
cmd = ('phpcs', '--report=emacs', '${args}', '-')
regex = r'^.*:(?P<line>[0-9]+):(?P<col>[0-9]+): (?:(?P<error>error)|(?P<warning>warning)) - (?P<message>(.(?!\(\S+\)$))+)( \((?P<code>\S+)\)$)?' # noqa: E501
defaults = {
'selector': 'source.php - text.blade, text.html.basic',
# we want auto-substitution of the filename,
# but `cmd` does not support that yet
'--stdin-path=': '${file}'
}
<commit_msg>Update selector for Sublime Text >= 4134<commit_after>from SublimeLinter.lint import ComposerLinter
class Phpcs(ComposerLinter):
cmd = ('phpcs', '--report=emacs', '${args}', '-')
regex = r'^.*:(?P<line>[0-9]+):(?P<col>[0-9]+): (?:(?P<error>error)|(?P<warning>warning)) - (?P<message>(.(?!\(\S+\)$))+)( \((?P<code>\S+)\)$)?' # noqa: E501
defaults = {
'selector': 'embedding.php, source.php - text.blade',
# we want auto-substitution of the filename,
# but `cmd` does not support that yet
'--stdin-path=': '${file}'
}
|
f61b81e968384859eb51a2ff14ca7709e8322ae8 | yunity/walls/models.py | yunity/walls/models.py | from django.db.models import ForeignKey, TextField
from config import settings
from yunity.base.models import BaseModel
class Wall(BaseModel):
pass
class WallPost(BaseModel):
wall = ForeignKey(Wall)
author = ForeignKey(settings.AUTH_USER_MODEL)
class WallPostContent(BaseModel):
post = ForeignKey(WallPost)
author = ForeignKey(settings.AUTH_USER_MODEL)
body = TextField()
| from django.db.models import ForeignKey, TextField
from config import settings
from yunity.base.models import BaseModel
class Wall(BaseModel):
def resolve_permissions(self, collector):
h = self.hub
if h.target_content_type.model == 'group':
g = h.target
""":type : Group"""
collector.add_hub(h, 'read')
if g.is_content_included_in_parent:
g = g.parent
while g:
collector.add_hub(g.hub, 'read')
g = g.parent
class WallPost(BaseModel):
wall = ForeignKey(Wall)
author = ForeignKey(settings.AUTH_USER_MODEL)
class WallPostContent(BaseModel):
post = ForeignKey(WallPost)
author = ForeignKey(settings.AUTH_USER_MODEL)
body = TextField()
| Implement basic permissions resolver for walls | Implement basic permissions resolver for walls
To be seen as a poc, collect all hub permissions for a basic permission
and settings/inheritance model for reading a wall.
with @nicksellen
| Python | agpl-3.0 | yunity/foodsaving-backend,yunity/yunity-core,yunity/foodsaving-backend,yunity/yunity-core,yunity/foodsaving-backend | from django.db.models import ForeignKey, TextField
from config import settings
from yunity.base.models import BaseModel
class Wall(BaseModel):
pass
class WallPost(BaseModel):
wall = ForeignKey(Wall)
author = ForeignKey(settings.AUTH_USER_MODEL)
class WallPostContent(BaseModel):
post = ForeignKey(WallPost)
author = ForeignKey(settings.AUTH_USER_MODEL)
body = TextField()
Implement basic permissions resolver for walls
To be seen as a poc, collect all hub permissions for a basic permission
and settings/inheritance model for reading a wall.
with @nicksellen | from django.db.models import ForeignKey, TextField
from config import settings
from yunity.base.models import BaseModel
class Wall(BaseModel):
def resolve_permissions(self, collector):
h = self.hub
if h.target_content_type.model == 'group':
g = h.target
""":type : Group"""
collector.add_hub(h, 'read')
if g.is_content_included_in_parent:
g = g.parent
while g:
collector.add_hub(g.hub, 'read')
g = g.parent
class WallPost(BaseModel):
wall = ForeignKey(Wall)
author = ForeignKey(settings.AUTH_USER_MODEL)
class WallPostContent(BaseModel):
post = ForeignKey(WallPost)
author = ForeignKey(settings.AUTH_USER_MODEL)
body = TextField()
| <commit_before>from django.db.models import ForeignKey, TextField
from config import settings
from yunity.base.models import BaseModel
class Wall(BaseModel):
pass
class WallPost(BaseModel):
wall = ForeignKey(Wall)
author = ForeignKey(settings.AUTH_USER_MODEL)
class WallPostContent(BaseModel):
post = ForeignKey(WallPost)
author = ForeignKey(settings.AUTH_USER_MODEL)
body = TextField()
<commit_msg>Implement basic permissions resolver for walls
To be seen as a poc, collect all hub permissions for a basic permission
and settings/inheritance model for reading a wall.
with @nicksellen<commit_after> | from django.db.models import ForeignKey, TextField
from config import settings
from yunity.base.models import BaseModel
class Wall(BaseModel):
def resolve_permissions(self, collector):
h = self.hub
if h.target_content_type.model == 'group':
g = h.target
""":type : Group"""
collector.add_hub(h, 'read')
if g.is_content_included_in_parent:
g = g.parent
while g:
collector.add_hub(g.hub, 'read')
g = g.parent
class WallPost(BaseModel):
wall = ForeignKey(Wall)
author = ForeignKey(settings.AUTH_USER_MODEL)
class WallPostContent(BaseModel):
post = ForeignKey(WallPost)
author = ForeignKey(settings.AUTH_USER_MODEL)
body = TextField()
| from django.db.models import ForeignKey, TextField
from config import settings
from yunity.base.models import BaseModel
class Wall(BaseModel):
pass
class WallPost(BaseModel):
wall = ForeignKey(Wall)
author = ForeignKey(settings.AUTH_USER_MODEL)
class WallPostContent(BaseModel):
post = ForeignKey(WallPost)
author = ForeignKey(settings.AUTH_USER_MODEL)
body = TextField()
Implement basic permissions resolver for walls
To be seen as a poc, collect all hub permissions for a basic permission
and settings/inheritance model for reading a wall.
with @nicksellenfrom django.db.models import ForeignKey, TextField
from config import settings
from yunity.base.models import BaseModel
class Wall(BaseModel):
def resolve_permissions(self, collector):
h = self.hub
if h.target_content_type.model == 'group':
g = h.target
""":type : Group"""
collector.add_hub(h, 'read')
if g.is_content_included_in_parent:
g = g.parent
while g:
collector.add_hub(g.hub, 'read')
g = g.parent
class WallPost(BaseModel):
wall = ForeignKey(Wall)
author = ForeignKey(settings.AUTH_USER_MODEL)
class WallPostContent(BaseModel):
post = ForeignKey(WallPost)
author = ForeignKey(settings.AUTH_USER_MODEL)
body = TextField()
| <commit_before>from django.db.models import ForeignKey, TextField
from config import settings
from yunity.base.models import BaseModel
class Wall(BaseModel):
pass
class WallPost(BaseModel):
wall = ForeignKey(Wall)
author = ForeignKey(settings.AUTH_USER_MODEL)
class WallPostContent(BaseModel):
post = ForeignKey(WallPost)
author = ForeignKey(settings.AUTH_USER_MODEL)
body = TextField()
<commit_msg>Implement basic permissions resolver for walls
To be seen as a poc, collect all hub permissions for a basic permission
and settings/inheritance model for reading a wall.
with @nicksellen<commit_after>from django.db.models import ForeignKey, TextField
from config import settings
from yunity.base.models import BaseModel
class Wall(BaseModel):
def resolve_permissions(self, collector):
h = self.hub
if h.target_content_type.model == 'group':
g = h.target
""":type : Group"""
collector.add_hub(h, 'read')
if g.is_content_included_in_parent:
g = g.parent
while g:
collector.add_hub(g.hub, 'read')
g = g.parent
class WallPost(BaseModel):
wall = ForeignKey(Wall)
author = ForeignKey(settings.AUTH_USER_MODEL)
class WallPostContent(BaseModel):
post = ForeignKey(WallPost)
author = ForeignKey(settings.AUTH_USER_MODEL)
body = TextField()
|
f83e8de41b4148f483a4de8ab624d89ee25fedb0 | soundbot.py | soundbot.py | from slacksocket import SlackSocket
import subprocess
import config
import os
def handle_cmd(cmd):
if cmd in ('ls','list'):
s.send_msg(list_files(), channel_name=config.slack_channel)
else:
playsound(cmd)
def playsound(sound):
subprocess.call([config.play_cmd,"{0}.mp3".format(sound)])
def list_files():
all_files = os.listdir('.')
mp3s = [parts[0] for file in all_files for parts in [file.split('.')] if parts[1]=='mp3']
mp3s.sort()
return ' '.join(mp3s)
s = SlackSocket(config.api_key,translate=True)
for event in s.events():
if event.event.get('channel')==config.slack_channel and event.event.get('type')=='message':
print(event.json)
handle_cmd(event.event.get('text'))
| from slacksocket import SlackSocket
import subprocess
import config
import os
def handle_cmd(cmd):
if cmd in ('ls','list'):
s.send_msg(list_files(), channel_name=config.slack_channel)
else:
playsound(cmd)
def playsound(sound):
subprocess.call([config.play_cmd,"{0}.mp3".format(sound)])
def list_files():
all_files = os.listdir('.')
mp3s = [parts[0] for file in all_files for parts in [file.split('.')] if parts[-1]=='mp3']
mp3s.sort()
return ' '.join(mp3s)
s = SlackSocket(config.api_key,translate=True)
for event in s.events():
if event.event.get('channel')==config.slack_channel and event.event.get('type')=='message':
print(event.json)
handle_cmd(event.event.get('text'))
| Make file listing more robust | Make file listing more robust
| Python | mit | sanderevers/slack-soundbot | from slacksocket import SlackSocket
import subprocess
import config
import os
def handle_cmd(cmd):
if cmd in ('ls','list'):
s.send_msg(list_files(), channel_name=config.slack_channel)
else:
playsound(cmd)
def playsound(sound):
subprocess.call([config.play_cmd,"{0}.mp3".format(sound)])
def list_files():
all_files = os.listdir('.')
mp3s = [parts[0] for file in all_files for parts in [file.split('.')] if parts[1]=='mp3']
mp3s.sort()
return ' '.join(mp3s)
s = SlackSocket(config.api_key,translate=True)
for event in s.events():
if event.event.get('channel')==config.slack_channel and event.event.get('type')=='message':
print(event.json)
handle_cmd(event.event.get('text'))
Make file listing more robust | from slacksocket import SlackSocket
import subprocess
import config
import os
def handle_cmd(cmd):
if cmd in ('ls','list'):
s.send_msg(list_files(), channel_name=config.slack_channel)
else:
playsound(cmd)
def playsound(sound):
subprocess.call([config.play_cmd,"{0}.mp3".format(sound)])
def list_files():
all_files = os.listdir('.')
mp3s = [parts[0] for file in all_files for parts in [file.split('.')] if parts[-1]=='mp3']
mp3s.sort()
return ' '.join(mp3s)
s = SlackSocket(config.api_key,translate=True)
for event in s.events():
if event.event.get('channel')==config.slack_channel and event.event.get('type')=='message':
print(event.json)
handle_cmd(event.event.get('text'))
| <commit_before>from slacksocket import SlackSocket
import subprocess
import config
import os
def handle_cmd(cmd):
if cmd in ('ls','list'):
s.send_msg(list_files(), channel_name=config.slack_channel)
else:
playsound(cmd)
def playsound(sound):
subprocess.call([config.play_cmd,"{0}.mp3".format(sound)])
def list_files():
all_files = os.listdir('.')
mp3s = [parts[0] for file in all_files for parts in [file.split('.')] if parts[1]=='mp3']
mp3s.sort()
return ' '.join(mp3s)
s = SlackSocket(config.api_key,translate=True)
for event in s.events():
if event.event.get('channel')==config.slack_channel and event.event.get('type')=='message':
print(event.json)
handle_cmd(event.event.get('text'))
<commit_msg>Make file listing more robust<commit_after> | from slacksocket import SlackSocket
import subprocess
import config
import os
def handle_cmd(cmd):
if cmd in ('ls','list'):
s.send_msg(list_files(), channel_name=config.slack_channel)
else:
playsound(cmd)
def playsound(sound):
subprocess.call([config.play_cmd,"{0}.mp3".format(sound)])
def list_files():
all_files = os.listdir('.')
mp3s = [parts[0] for file in all_files for parts in [file.split('.')] if parts[-1]=='mp3']
mp3s.sort()
return ' '.join(mp3s)
s = SlackSocket(config.api_key,translate=True)
for event in s.events():
if event.event.get('channel')==config.slack_channel and event.event.get('type')=='message':
print(event.json)
handle_cmd(event.event.get('text'))
| from slacksocket import SlackSocket
import subprocess
import config
import os
def handle_cmd(cmd):
if cmd in ('ls','list'):
s.send_msg(list_files(), channel_name=config.slack_channel)
else:
playsound(cmd)
def playsound(sound):
subprocess.call([config.play_cmd,"{0}.mp3".format(sound)])
def list_files():
all_files = os.listdir('.')
mp3s = [parts[0] for file in all_files for parts in [file.split('.')] if parts[1]=='mp3']
mp3s.sort()
return ' '.join(mp3s)
s = SlackSocket(config.api_key,translate=True)
for event in s.events():
if event.event.get('channel')==config.slack_channel and event.event.get('type')=='message':
print(event.json)
handle_cmd(event.event.get('text'))
Make file listing more robustfrom slacksocket import SlackSocket
import subprocess
import config
import os
def handle_cmd(cmd):
if cmd in ('ls','list'):
s.send_msg(list_files(), channel_name=config.slack_channel)
else:
playsound(cmd)
def playsound(sound):
subprocess.call([config.play_cmd,"{0}.mp3".format(sound)])
def list_files():
all_files = os.listdir('.')
mp3s = [parts[0] for file in all_files for parts in [file.split('.')] if parts[-1]=='mp3']
mp3s.sort()
return ' '.join(mp3s)
s = SlackSocket(config.api_key,translate=True)
for event in s.events():
if event.event.get('channel')==config.slack_channel and event.event.get('type')=='message':
print(event.json)
handle_cmd(event.event.get('text'))
| <commit_before>from slacksocket import SlackSocket
import subprocess
import config
import os
def handle_cmd(cmd):
if cmd in ('ls','list'):
s.send_msg(list_files(), channel_name=config.slack_channel)
else:
playsound(cmd)
def playsound(sound):
subprocess.call([config.play_cmd,"{0}.mp3".format(sound)])
def list_files():
all_files = os.listdir('.')
mp3s = [parts[0] for file in all_files for parts in [file.split('.')] if parts[1]=='mp3']
mp3s.sort()
return ' '.join(mp3s)
s = SlackSocket(config.api_key,translate=True)
for event in s.events():
if event.event.get('channel')==config.slack_channel and event.event.get('type')=='message':
print(event.json)
handle_cmd(event.event.get('text'))
<commit_msg>Make file listing more robust<commit_after>from slacksocket import SlackSocket
import subprocess
import config
import os
def handle_cmd(cmd):
if cmd in ('ls','list'):
s.send_msg(list_files(), channel_name=config.slack_channel)
else:
playsound(cmd)
def playsound(sound):
subprocess.call([config.play_cmd,"{0}.mp3".format(sound)])
def list_files():
all_files = os.listdir('.')
mp3s = [parts[0] for file in all_files for parts in [file.split('.')] if parts[-1]=='mp3']
mp3s.sort()
return ' '.join(mp3s)
s = SlackSocket(config.api_key,translate=True)
for event in s.events():
if event.event.get('channel')==config.slack_channel and event.event.get('type')=='message':
print(event.json)
handle_cmd(event.event.get('text'))
|
58970055d4905b50a3ed76a8ead39d9d0b572854 | hookit.py | hookit.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from datetime import datetime
from flask import Flask, request
app = Flask(__name__)
@app.route('/')
def index():
return 'Hello from Flask! (%s)' % datetime.now().strftime('%Y-%m-%d %H:%M:%S')
@app.route('/github')
def github():
if request.method != 'POST':
return 405
with open('test', 'w') as f:
f.write(request.get_json())
return ':)'
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
from datetime import datetime
from flask import Flask, request
app = Flask(__name__)
app.debug = True
@app.route('/')
def index():
return 'Hello from Flask! (%s)' % datetime.now().strftime('%Y-%m-%d %H:%M:%S')
@app.route('/github')
def github():
if request.method != 'POST':
return 'Not Allowed', 405
with open('test', 'w') as f:
f.write(request.get_json())
return ':)'
if __name__ == '__main__':
app.run()
| Add debugging options by default | Add debugging options by default
| Python | mit | pcostesi/hookit,pcostesi/hookit | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from datetime import datetime
from flask import Flask, request
app = Flask(__name__)
@app.route('/')
def index():
return 'Hello from Flask! (%s)' % datetime.now().strftime('%Y-%m-%d %H:%M:%S')
@app.route('/github')
def github():
if request.method != 'POST':
return 405
with open('test', 'w') as f:
f.write(request.get_json())
return ':)'
Add debugging options by default | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from datetime import datetime
from flask import Flask, request
app = Flask(__name__)
app.debug = True
@app.route('/')
def index():
return 'Hello from Flask! (%s)' % datetime.now().strftime('%Y-%m-%d %H:%M:%S')
@app.route('/github')
def github():
if request.method != 'POST':
return 'Not Allowed', 405
with open('test', 'w') as f:
f.write(request.get_json())
return ':)'
if __name__ == '__main__':
app.run()
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from datetime import datetime
from flask import Flask, request
app = Flask(__name__)
@app.route('/')
def index():
return 'Hello from Flask! (%s)' % datetime.now().strftime('%Y-%m-%d %H:%M:%S')
@app.route('/github')
def github():
if request.method != 'POST':
return 405
with open('test', 'w') as f:
f.write(request.get_json())
return ':)'
<commit_msg>Add debugging options by default<commit_after> | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from datetime import datetime
from flask import Flask, request
app = Flask(__name__)
app.debug = True
@app.route('/')
def index():
return 'Hello from Flask! (%s)' % datetime.now().strftime('%Y-%m-%d %H:%M:%S')
@app.route('/github')
def github():
if request.method != 'POST':
return 'Not Allowed', 405
with open('test', 'w') as f:
f.write(request.get_json())
return ':)'
if __name__ == '__main__':
app.run()
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
from datetime import datetime
from flask import Flask, request
app = Flask(__name__)
@app.route('/')
def index():
return 'Hello from Flask! (%s)' % datetime.now().strftime('%Y-%m-%d %H:%M:%S')
@app.route('/github')
def github():
if request.method != 'POST':
return 405
with open('test', 'w') as f:
f.write(request.get_json())
return ':)'
Add debugging options by default#!/usr/bin/env python
# -*- coding: utf-8 -*-
from datetime import datetime
from flask import Flask, request
app = Flask(__name__)
app.debug = True
@app.route('/')
def index():
return 'Hello from Flask! (%s)' % datetime.now().strftime('%Y-%m-%d %H:%M:%S')
@app.route('/github')
def github():
if request.method != 'POST':
return 'Not Allowed', 405
with open('test', 'w') as f:
f.write(request.get_json())
return ':)'
if __name__ == '__main__':
app.run()
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from datetime import datetime
from flask import Flask, request
app = Flask(__name__)
@app.route('/')
def index():
return 'Hello from Flask! (%s)' % datetime.now().strftime('%Y-%m-%d %H:%M:%S')
@app.route('/github')
def github():
if request.method != 'POST':
return 405
with open('test', 'w') as f:
f.write(request.get_json())
return ':)'
<commit_msg>Add debugging options by default<commit_after>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from datetime import datetime
from flask import Flask, request
app = Flask(__name__)
app.debug = True
@app.route('/')
def index():
return 'Hello from Flask! (%s)' % datetime.now().strftime('%Y-%m-%d %H:%M:%S')
@app.route('/github')
def github():
if request.method != 'POST':
return 'Not Allowed', 405
with open('test', 'w') as f:
f.write(request.get_json())
return ':)'
if __name__ == '__main__':
app.run()
|
3d48d62aca0455cf71a841fcfcddabaf770bd048 | plumeria/plugins/bible.py | plumeria/plugins/bible.py | from bs4 import BeautifulSoup
from bs4 import Comment
from bs4 import NavigableString
import plumeria.util.http as http
from plumeria.command import commands, CommandError
from plumeria.command.parse import Text
from plumeria.util.ratelimit import rate_limit
@commands.register("bible", "esv", category="Search", params=[Text('verse')])
@rate_limit()
async def search_esv(message, verse):
"""
Search for a bible passage from the English Standard Version.
Example::
bible Romans 12:16
"""
r = await http.get("http://www.esvapi.org/v2/rest/passageQuery", params={
"key": "IP",
"passage": verse,
"output-format": "crossway-xml-1.0"
})
doc = BeautifulSoup(r.text(), features="lxml")
if not doc.passage:
raise CommandError("Verse not found.")
lines = []
for verse_unit in doc.passage.content.find_all('verse-unit'):
num = int(verse_unit.find('verse-num').text)
woc = verse_unit.find('woc')
if woc:
text = woc.text
else:
text = "".join([str(node) for node in verse_unit.children
if isinstance(node, NavigableString) and not isinstance(node, Comment)])
lines.append("({}) {}".format(num, text.strip()))
return "\n".join(lines)
| from bs4 import BeautifulSoup
from bs4 import Comment
from bs4 import NavigableString
import plumeria.util.http as http
from plumeria.command import commands, CommandError
from plumeria.command.parse import Text
from plumeria.util.ratelimit import rate_limit
@commands.register("bible", "esv", category="Search", params=[Text('verse')])
@rate_limit()
async def search_esv(message, verse):
"""
Search for a bible passage from the English Standard Version.
Example::
bible Romans 12:16
"""
r = await http.get("http://www.esvapi.org/v2/rest/passageQuery", params={
"key": "IP",
"passage": verse,
"output-format": "crossway-xml-1.0"
})
doc = BeautifulSoup(r.text(), features="lxml")
if not doc.passage:
raise CommandError("Verse not found.")
lines = []
for verse_unit in doc.passage.content.find_all('verse-unit'):
num = int(verse_unit.find('verse-num').text)
woc = verse_unit.find('woc')
if woc:
text = woc.text
else:
text = "".join([str(node) for node in verse_unit.children
if isinstance(node, NavigableString) and not isinstance(node, Comment)])
lines.append("**{}** {}".format(num, text.strip()))
return "\n".join(lines)
| Use bold line number for verse. | Use bold line number for verse.
| Python | mit | sk89q/Plumeria,sk89q/Plumeria,sk89q/Plumeria | from bs4 import BeautifulSoup
from bs4 import Comment
from bs4 import NavigableString
import plumeria.util.http as http
from plumeria.command import commands, CommandError
from plumeria.command.parse import Text
from plumeria.util.ratelimit import rate_limit
@commands.register("bible", "esv", category="Search", params=[Text('verse')])
@rate_limit()
async def search_esv(message, verse):
"""
Search for a bible passage from the English Standard Version.
Example::
bible Romans 12:16
"""
r = await http.get("http://www.esvapi.org/v2/rest/passageQuery", params={
"key": "IP",
"passage": verse,
"output-format": "crossway-xml-1.0"
})
doc = BeautifulSoup(r.text(), features="lxml")
if not doc.passage:
raise CommandError("Verse not found.")
lines = []
for verse_unit in doc.passage.content.find_all('verse-unit'):
num = int(verse_unit.find('verse-num').text)
woc = verse_unit.find('woc')
if woc:
text = woc.text
else:
text = "".join([str(node) for node in verse_unit.children
if isinstance(node, NavigableString) and not isinstance(node, Comment)])
lines.append("({}) {}".format(num, text.strip()))
return "\n".join(lines)
Use bold line number for verse. | from bs4 import BeautifulSoup
from bs4 import Comment
from bs4 import NavigableString
import plumeria.util.http as http
from plumeria.command import commands, CommandError
from plumeria.command.parse import Text
from plumeria.util.ratelimit import rate_limit
@commands.register("bible", "esv", category="Search", params=[Text('verse')])
@rate_limit()
async def search_esv(message, verse):
"""
Search for a bible passage from the English Standard Version.
Example::
bible Romans 12:16
"""
r = await http.get("http://www.esvapi.org/v2/rest/passageQuery", params={
"key": "IP",
"passage": verse,
"output-format": "crossway-xml-1.0"
})
doc = BeautifulSoup(r.text(), features="lxml")
if not doc.passage:
raise CommandError("Verse not found.")
lines = []
for verse_unit in doc.passage.content.find_all('verse-unit'):
num = int(verse_unit.find('verse-num').text)
woc = verse_unit.find('woc')
if woc:
text = woc.text
else:
text = "".join([str(node) for node in verse_unit.children
if isinstance(node, NavigableString) and not isinstance(node, Comment)])
lines.append("**{}** {}".format(num, text.strip()))
return "\n".join(lines)
| <commit_before>from bs4 import BeautifulSoup
from bs4 import Comment
from bs4 import NavigableString
import plumeria.util.http as http
from plumeria.command import commands, CommandError
from plumeria.command.parse import Text
from plumeria.util.ratelimit import rate_limit
@commands.register("bible", "esv", category="Search", params=[Text('verse')])
@rate_limit()
async def search_esv(message, verse):
"""
Search for a bible passage from the English Standard Version.
Example::
bible Romans 12:16
"""
r = await http.get("http://www.esvapi.org/v2/rest/passageQuery", params={
"key": "IP",
"passage": verse,
"output-format": "crossway-xml-1.0"
})
doc = BeautifulSoup(r.text(), features="lxml")
if not doc.passage:
raise CommandError("Verse not found.")
lines = []
for verse_unit in doc.passage.content.find_all('verse-unit'):
num = int(verse_unit.find('verse-num').text)
woc = verse_unit.find('woc')
if woc:
text = woc.text
else:
text = "".join([str(node) for node in verse_unit.children
if isinstance(node, NavigableString) and not isinstance(node, Comment)])
lines.append("({}) {}".format(num, text.strip()))
return "\n".join(lines)
<commit_msg>Use bold line number for verse.<commit_after> | from bs4 import BeautifulSoup
from bs4 import Comment
from bs4 import NavigableString
import plumeria.util.http as http
from plumeria.command import commands, CommandError
from plumeria.command.parse import Text
from plumeria.util.ratelimit import rate_limit
@commands.register("bible", "esv", category="Search", params=[Text('verse')])
@rate_limit()
async def search_esv(message, verse):
"""
Search for a bible passage from the English Standard Version.
Example::
bible Romans 12:16
"""
r = await http.get("http://www.esvapi.org/v2/rest/passageQuery", params={
"key": "IP",
"passage": verse,
"output-format": "crossway-xml-1.0"
})
doc = BeautifulSoup(r.text(), features="lxml")
if not doc.passage:
raise CommandError("Verse not found.")
lines = []
for verse_unit in doc.passage.content.find_all('verse-unit'):
num = int(verse_unit.find('verse-num').text)
woc = verse_unit.find('woc')
if woc:
text = woc.text
else:
text = "".join([str(node) for node in verse_unit.children
if isinstance(node, NavigableString) and not isinstance(node, Comment)])
lines.append("**{}** {}".format(num, text.strip()))
return "\n".join(lines)
| from bs4 import BeautifulSoup
from bs4 import Comment
from bs4 import NavigableString
import plumeria.util.http as http
from plumeria.command import commands, CommandError
from plumeria.command.parse import Text
from plumeria.util.ratelimit import rate_limit
@commands.register("bible", "esv", category="Search", params=[Text('verse')])
@rate_limit()
async def search_esv(message, verse):
"""
Search for a bible passage from the English Standard Version.
Example::
bible Romans 12:16
"""
r = await http.get("http://www.esvapi.org/v2/rest/passageQuery", params={
"key": "IP",
"passage": verse,
"output-format": "crossway-xml-1.0"
})
doc = BeautifulSoup(r.text(), features="lxml")
if not doc.passage:
raise CommandError("Verse not found.")
lines = []
for verse_unit in doc.passage.content.find_all('verse-unit'):
num = int(verse_unit.find('verse-num').text)
woc = verse_unit.find('woc')
if woc:
text = woc.text
else:
text = "".join([str(node) for node in verse_unit.children
if isinstance(node, NavigableString) and not isinstance(node, Comment)])
lines.append("({}) {}".format(num, text.strip()))
return "\n".join(lines)
Use bold line number for verse.from bs4 import BeautifulSoup
from bs4 import Comment
from bs4 import NavigableString
import plumeria.util.http as http
from plumeria.command import commands, CommandError
from plumeria.command.parse import Text
from plumeria.util.ratelimit import rate_limit
@commands.register("bible", "esv", category="Search", params=[Text('verse')])
@rate_limit()
async def search_esv(message, verse):
"""
Search for a bible passage from the English Standard Version.
Example::
bible Romans 12:16
"""
r = await http.get("http://www.esvapi.org/v2/rest/passageQuery", params={
"key": "IP",
"passage": verse,
"output-format": "crossway-xml-1.0"
})
doc = BeautifulSoup(r.text(), features="lxml")
if not doc.passage:
raise CommandError("Verse not found.")
lines = []
for verse_unit in doc.passage.content.find_all('verse-unit'):
num = int(verse_unit.find('verse-num').text)
woc = verse_unit.find('woc')
if woc:
text = woc.text
else:
text = "".join([str(node) for node in verse_unit.children
if isinstance(node, NavigableString) and not isinstance(node, Comment)])
lines.append("**{}** {}".format(num, text.strip()))
return "\n".join(lines)
| <commit_before>from bs4 import BeautifulSoup
from bs4 import Comment
from bs4 import NavigableString
import plumeria.util.http as http
from plumeria.command import commands, CommandError
from plumeria.command.parse import Text
from plumeria.util.ratelimit import rate_limit
@commands.register("bible", "esv", category="Search", params=[Text('verse')])
@rate_limit()
async def search_esv(message, verse):
"""
Search for a bible passage from the English Standard Version.
Example::
bible Romans 12:16
"""
r = await http.get("http://www.esvapi.org/v2/rest/passageQuery", params={
"key": "IP",
"passage": verse,
"output-format": "crossway-xml-1.0"
})
doc = BeautifulSoup(r.text(), features="lxml")
if not doc.passage:
raise CommandError("Verse not found.")
lines = []
for verse_unit in doc.passage.content.find_all('verse-unit'):
num = int(verse_unit.find('verse-num').text)
woc = verse_unit.find('woc')
if woc:
text = woc.text
else:
text = "".join([str(node) for node in verse_unit.children
if isinstance(node, NavigableString) and not isinstance(node, Comment)])
lines.append("({}) {}".format(num, text.strip()))
return "\n".join(lines)
<commit_msg>Use bold line number for verse.<commit_after>from bs4 import BeautifulSoup
from bs4 import Comment
from bs4 import NavigableString
import plumeria.util.http as http
from plumeria.command import commands, CommandError
from plumeria.command.parse import Text
from plumeria.util.ratelimit import rate_limit
@commands.register("bible", "esv", category="Search", params=[Text('verse')])
@rate_limit()
async def search_esv(message, verse):
"""
Search for a bible passage from the English Standard Version.
Example::
bible Romans 12:16
"""
r = await http.get("http://www.esvapi.org/v2/rest/passageQuery", params={
"key": "IP",
"passage": verse,
"output-format": "crossway-xml-1.0"
})
doc = BeautifulSoup(r.text(), features="lxml")
if not doc.passage:
raise CommandError("Verse not found.")
lines = []
for verse_unit in doc.passage.content.find_all('verse-unit'):
num = int(verse_unit.find('verse-num').text)
woc = verse_unit.find('woc')
if woc:
text = woc.text
else:
text = "".join([str(node) for node in verse_unit.children
if isinstance(node, NavigableString) and not isinstance(node, Comment)])
lines.append("**{}** {}".format(num, text.strip()))
return "\n".join(lines)
|
f4f5852944d1fd1b9e96a70cb4496ee6e1e66dc0 | genome_designer/main/celery_util.py | genome_designer/main/celery_util.py | """
Methods for interfacing with the Celery task queue management library.
"""
from errno import errorcode
from celery.task.control import inspect
CELERY_ERROR_KEY = 'ERROR'
def get_celery_worker_status():
"""Checks whether celery is running and reports the error if not.
Source: http://stackoverflow.com/questions/8506914/detect-whether-celery-is-available-running
"""
try:
insp = inspect()
d = insp.stats()
if not d:
d = { CELERY_ERROR_KEY: 'No running Celery workers were found.' }
except IOError as e:
msg = "Error connecting to the backend: " + str(e)
if len(e.args) > 0 and errorcode.get(e.args[0]) == 'ECONNREFUSED':
msg += ' Check that the RabbitMQ server is running.'
d = { CELERY_ERROR_KEY: msg }
except ImportError as e:
d = { CELERY_ERROR_KEY: str(e)}
return d
| """
Methods for interfacing with the Celery task queue management library.
"""
from errno import errorcode
from celery.task.control import inspect
from django.conf import settings
CELERY_ERROR_KEY = 'ERROR'
def get_celery_worker_status():
"""Checks whether celery is running and reports the error if not.
Source: http://stackoverflow.com/questions/8506914/detect-whether-celery-is-available-running
"""
if settings.BROKER_BACKEND == 'memory':
# We are testing with in-memory celery. Celery is effectively running.
return {}
try:
insp = inspect()
d = insp.stats()
if not d:
d = { CELERY_ERROR_KEY: 'No running Celery workers were found.' }
except IOError as e:
msg = "Error connecting to the backend: " + str(e)
if len(e.args) > 0 and errorcode.get(e.args[0]) == 'ECONNREFUSED':
msg += ' Check that the RabbitMQ server is running.'
d = { CELERY_ERROR_KEY: msg }
except ImportError as e:
d = { CELERY_ERROR_KEY: str(e)}
return d
| Fix tests: Allow for celery not to be running when doing in-memory celery for tests. | Fix tests: Allow for celery not to be running when doing in-memory celery for tests.
| Python | mit | churchlab/millstone,churchlab/millstone,churchlab/millstone,churchlab/millstone,woodymit/millstone_accidental_source,woodymit/millstone_accidental_source,woodymit/millstone_accidental_source,woodymit/millstone,woodymit/millstone,woodymit/millstone_accidental_source,woodymit/millstone,woodymit/millstone | """
Methods for interfacing with the Celery task queue management library.
"""
from errno import errorcode
from celery.task.control import inspect
CELERY_ERROR_KEY = 'ERROR'
def get_celery_worker_status():
"""Checks whether celery is running and reports the error if not.
Source: http://stackoverflow.com/questions/8506914/detect-whether-celery-is-available-running
"""
try:
insp = inspect()
d = insp.stats()
if not d:
d = { CELERY_ERROR_KEY: 'No running Celery workers were found.' }
except IOError as e:
msg = "Error connecting to the backend: " + str(e)
if len(e.args) > 0 and errorcode.get(e.args[0]) == 'ECONNREFUSED':
msg += ' Check that the RabbitMQ server is running.'
d = { CELERY_ERROR_KEY: msg }
except ImportError as e:
d = { CELERY_ERROR_KEY: str(e)}
return d
Fix tests: Allow for celery not to be running when doing in-memory celery for tests. | """
Methods for interfacing with the Celery task queue management library.
"""
from errno import errorcode
from celery.task.control import inspect
from django.conf import settings
CELERY_ERROR_KEY = 'ERROR'
def get_celery_worker_status():
"""Checks whether celery is running and reports the error if not.
Source: http://stackoverflow.com/questions/8506914/detect-whether-celery-is-available-running
"""
if settings.BROKER_BACKEND == 'memory':
# We are testing with in-memory celery. Celery is effectively running.
return {}
try:
insp = inspect()
d = insp.stats()
if not d:
d = { CELERY_ERROR_KEY: 'No running Celery workers were found.' }
except IOError as e:
msg = "Error connecting to the backend: " + str(e)
if len(e.args) > 0 and errorcode.get(e.args[0]) == 'ECONNREFUSED':
msg += ' Check that the RabbitMQ server is running.'
d = { CELERY_ERROR_KEY: msg }
except ImportError as e:
d = { CELERY_ERROR_KEY: str(e)}
return d
| <commit_before>"""
Methods for interfacing with the Celery task queue management library.
"""
from errno import errorcode
from celery.task.control import inspect
CELERY_ERROR_KEY = 'ERROR'
def get_celery_worker_status():
"""Checks whether celery is running and reports the error if not.
Source: http://stackoverflow.com/questions/8506914/detect-whether-celery-is-available-running
"""
try:
insp = inspect()
d = insp.stats()
if not d:
d = { CELERY_ERROR_KEY: 'No running Celery workers were found.' }
except IOError as e:
msg = "Error connecting to the backend: " + str(e)
if len(e.args) > 0 and errorcode.get(e.args[0]) == 'ECONNREFUSED':
msg += ' Check that the RabbitMQ server is running.'
d = { CELERY_ERROR_KEY: msg }
except ImportError as e:
d = { CELERY_ERROR_KEY: str(e)}
return d
<commit_msg>Fix tests: Allow for celery not to be running when doing in-memory celery for tests.<commit_after> | """
Methods for interfacing with the Celery task queue management library.
"""
from errno import errorcode
from celery.task.control import inspect
from django.conf import settings
CELERY_ERROR_KEY = 'ERROR'
def get_celery_worker_status():
"""Checks whether celery is running and reports the error if not.
Source: http://stackoverflow.com/questions/8506914/detect-whether-celery-is-available-running
"""
if settings.BROKER_BACKEND == 'memory':
# We are testing with in-memory celery. Celery is effectively running.
return {}
try:
insp = inspect()
d = insp.stats()
if not d:
d = { CELERY_ERROR_KEY: 'No running Celery workers were found.' }
except IOError as e:
msg = "Error connecting to the backend: " + str(e)
if len(e.args) > 0 and errorcode.get(e.args[0]) == 'ECONNREFUSED':
msg += ' Check that the RabbitMQ server is running.'
d = { CELERY_ERROR_KEY: msg }
except ImportError as e:
d = { CELERY_ERROR_KEY: str(e)}
return d
| """
Methods for interfacing with the Celery task queue management library.
"""
from errno import errorcode
from celery.task.control import inspect
CELERY_ERROR_KEY = 'ERROR'
def get_celery_worker_status():
"""Checks whether celery is running and reports the error if not.
Source: http://stackoverflow.com/questions/8506914/detect-whether-celery-is-available-running
"""
try:
insp = inspect()
d = insp.stats()
if not d:
d = { CELERY_ERROR_KEY: 'No running Celery workers were found.' }
except IOError as e:
msg = "Error connecting to the backend: " + str(e)
if len(e.args) > 0 and errorcode.get(e.args[0]) == 'ECONNREFUSED':
msg += ' Check that the RabbitMQ server is running.'
d = { CELERY_ERROR_KEY: msg }
except ImportError as e:
d = { CELERY_ERROR_KEY: str(e)}
return d
Fix tests: Allow for celery not to be running when doing in-memory celery for tests."""
Methods for interfacing with the Celery task queue management library.
"""
from errno import errorcode
from celery.task.control import inspect
from django.conf import settings
CELERY_ERROR_KEY = 'ERROR'
def get_celery_worker_status():
"""Checks whether celery is running and reports the error if not.
Source: http://stackoverflow.com/questions/8506914/detect-whether-celery-is-available-running
"""
if settings.BROKER_BACKEND == 'memory':
# We are testing with in-memory celery. Celery is effectively running.
return {}
try:
insp = inspect()
d = insp.stats()
if not d:
d = { CELERY_ERROR_KEY: 'No running Celery workers were found.' }
except IOError as e:
msg = "Error connecting to the backend: " + str(e)
if len(e.args) > 0 and errorcode.get(e.args[0]) == 'ECONNREFUSED':
msg += ' Check that the RabbitMQ server is running.'
d = { CELERY_ERROR_KEY: msg }
except ImportError as e:
d = { CELERY_ERROR_KEY: str(e)}
return d
| <commit_before>"""
Methods for interfacing with the Celery task queue management library.
"""
from errno import errorcode
from celery.task.control import inspect
CELERY_ERROR_KEY = 'ERROR'
def get_celery_worker_status():
"""Checks whether celery is running and reports the error if not.
Source: http://stackoverflow.com/questions/8506914/detect-whether-celery-is-available-running
"""
try:
insp = inspect()
d = insp.stats()
if not d:
d = { CELERY_ERROR_KEY: 'No running Celery workers were found.' }
except IOError as e:
msg = "Error connecting to the backend: " + str(e)
if len(e.args) > 0 and errorcode.get(e.args[0]) == 'ECONNREFUSED':
msg += ' Check that the RabbitMQ server is running.'
d = { CELERY_ERROR_KEY: msg }
except ImportError as e:
d = { CELERY_ERROR_KEY: str(e)}
return d
<commit_msg>Fix tests: Allow for celery not to be running when doing in-memory celery for tests.<commit_after>"""
Methods for interfacing with the Celery task queue management library.
"""
from errno import errorcode
from celery.task.control import inspect
from django.conf import settings
CELERY_ERROR_KEY = 'ERROR'
def get_celery_worker_status():
"""Checks whether celery is running and reports the error if not.
Source: http://stackoverflow.com/questions/8506914/detect-whether-celery-is-available-running
"""
if settings.BROKER_BACKEND == 'memory':
# We are testing with in-memory celery. Celery is effectively running.
return {}
try:
insp = inspect()
d = insp.stats()
if not d:
d = { CELERY_ERROR_KEY: 'No running Celery workers were found.' }
except IOError as e:
msg = "Error connecting to the backend: " + str(e)
if len(e.args) > 0 and errorcode.get(e.args[0]) == 'ECONNREFUSED':
msg += ' Check that the RabbitMQ server is running.'
d = { CELERY_ERROR_KEY: msg }
except ImportError as e:
d = { CELERY_ERROR_KEY: str(e)}
return d
|
0b7f99bcb4e42c50263a7d8a42513876b02b445a | scikits/talkbox/tools/__init__.py | scikits/talkbox/tools/__init__.py | __all__ = []
import correlations
from correlations import *
__all__ += correlations.__all__
import cffilter
from cffilter import cslfilter as slfilter
__all__ += ['slfilter']
| __all__ = []
import correlations
from correlations import *
__all__ += correlations.__all__
import cffilter
from cffilter import cslfilter as slfilter
__all__ += ['slfilter']
from segmentaxis import segment_axis
__all__ += ['segment_axis']
| Put segment_axis in the main scikits.talkbox namespace. | Put segment_axis in the main scikits.talkbox namespace.
| Python | mit | cournape/talkbox,cournape/talkbox | __all__ = []
import correlations
from correlations import *
__all__ += correlations.__all__
import cffilter
from cffilter import cslfilter as slfilter
__all__ += ['slfilter']
Put segment_axis in the main scikits.talkbox namespace. | __all__ = []
import correlations
from correlations import *
__all__ += correlations.__all__
import cffilter
from cffilter import cslfilter as slfilter
__all__ += ['slfilter']
from segmentaxis import segment_axis
__all__ += ['segment_axis']
| <commit_before>__all__ = []
import correlations
from correlations import *
__all__ += correlations.__all__
import cffilter
from cffilter import cslfilter as slfilter
__all__ += ['slfilter']
<commit_msg>Put segment_axis in the main scikits.talkbox namespace.<commit_after> | __all__ = []
import correlations
from correlations import *
__all__ += correlations.__all__
import cffilter
from cffilter import cslfilter as slfilter
__all__ += ['slfilter']
from segmentaxis import segment_axis
__all__ += ['segment_axis']
| __all__ = []
import correlations
from correlations import *
__all__ += correlations.__all__
import cffilter
from cffilter import cslfilter as slfilter
__all__ += ['slfilter']
Put segment_axis in the main scikits.talkbox namespace.__all__ = []
import correlations
from correlations import *
__all__ += correlations.__all__
import cffilter
from cffilter import cslfilter as slfilter
__all__ += ['slfilter']
from segmentaxis import segment_axis
__all__ += ['segment_axis']
| <commit_before>__all__ = []
import correlations
from correlations import *
__all__ += correlations.__all__
import cffilter
from cffilter import cslfilter as slfilter
__all__ += ['slfilter']
<commit_msg>Put segment_axis in the main scikits.talkbox namespace.<commit_after>__all__ = []
import correlations
from correlations import *
__all__ += correlations.__all__
import cffilter
from cffilter import cslfilter as slfilter
__all__ += ['slfilter']
from segmentaxis import segment_axis
__all__ += ['segment_axis']
|
91fd97d7579673a0c310c734a1c1ef83a07b50d1 | phantasy/library/scan/datautil.py | phantasy/library/scan/datautil.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""utils for data crunching, saving.
"""
import numpy as np
class ScanDataFactory(object):
"""Post processor of data from scan server.
Parameters
----------
data : dict
Raw data retrieving from scan server regarding scan ID, after
completing certain scan task.
n_sample : int
Sample number for every scan device setup.
"""
def __init__(self, data, n_sample):
self._raw_data = data
self._n = n_sample
self._rebuild_data()
def _rebuild_data(self):
"""Rebuild raw_data
"""
self._data = {k:np.array(v.get('value')).reshape(-1, self._n)
for k,v in self._raw_data.iteritems()}
def get_average(self, name):
"""Get average.
Parameters
----------
name : str
Key name of raw_data.
"""
return self._data.get(name).mean(axis=1)
def get_errorbar(self, name):
"""Get errorbar
Parameters
----------
name : str
Key name of raw_data.
"""
return self._data.get(name).std(axis=1)
def get_all_names(self):
"""Get all key names of raw_data.
Returns
-------
ret : list
List of keys.
"""
return self._data.keys()
def save(self, ext='dat'):
"""
"""
pass
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""utils for data crunching, saving.
"""
import numpy as np
class ScanDataFactory(object):
"""Post processor of data from scan server.
Parameters
----------
data : dict
Raw data retrieving from scan server regarding scan ID, after
completing certain scan task.
n_sample : int
Sample number for every scan device setup.
"""
def __init__(self, data, n_sample):
self._raw_data = data
self._n = n_sample
self._rebuild_data()
@property
def raw_data(self):
"""dict: Dict of array, raw scan data."""
return self._raw_data
@property
def data(self):
"""dict: Dict of array, raw scan data after postprocessing."""
return self._data
def _rebuild_data(self):
"""Rebuild raw_data
"""
self._data = {k:np.array(v.get('value')).reshape(-1, self._n)
for k,v in self._raw_data.iteritems()}
def get_average(self, name):
"""Get average.
Parameters
----------
name : str
Key name of raw_data.
"""
return self._data.get(name).mean(axis=1)
def get_errorbar(self, name):
"""Get errorbar
Parameters
----------
name : str
Key name of raw_data.
"""
return self._data.get(name).std(axis=1)
def get_all_names(self):
"""Get all key names of raw_data.
Returns
-------
ret : list
List of keys.
"""
return self._data.keys()
def save(self, ext='dat'):
"""
"""
pass
| Make raw_data and data as properties | Make raw_data and data as properties
| Python | bsd-3-clause | archman/phantasy,archman/phantasy | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""utils for data crunching, saving.
"""
import numpy as np
class ScanDataFactory(object):
"""Post processor of data from scan server.
Parameters
----------
data : dict
Raw data retrieving from scan server regarding scan ID, after
completing certain scan task.
n_sample : int
Sample number for every scan device setup.
"""
def __init__(self, data, n_sample):
self._raw_data = data
self._n = n_sample
self._rebuild_data()
def _rebuild_data(self):
"""Rebuild raw_data
"""
self._data = {k:np.array(v.get('value')).reshape(-1, self._n)
for k,v in self._raw_data.iteritems()}
def get_average(self, name):
"""Get average.
Parameters
----------
name : str
Key name of raw_data.
"""
return self._data.get(name).mean(axis=1)
def get_errorbar(self, name):
"""Get errorbar
Parameters
----------
name : str
Key name of raw_data.
"""
return self._data.get(name).std(axis=1)
def get_all_names(self):
"""Get all key names of raw_data.
Returns
-------
ret : list
List of keys.
"""
return self._data.keys()
def save(self, ext='dat'):
"""
"""
pass
Make raw_data and data as properties | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""utils for data crunching, saving.
"""
import numpy as np
class ScanDataFactory(object):
"""Post processor of data from scan server.
Parameters
----------
data : dict
Raw data retrieving from scan server regarding scan ID, after
completing certain scan task.
n_sample : int
Sample number for every scan device setup.
"""
def __init__(self, data, n_sample):
self._raw_data = data
self._n = n_sample
self._rebuild_data()
@property
def raw_data(self):
"""dict: Dict of array, raw scan data."""
return self._raw_data
@property
def data(self):
"""dict: Dict of array, raw scan data after postprocessing."""
return self._data
def _rebuild_data(self):
"""Rebuild raw_data
"""
self._data = {k:np.array(v.get('value')).reshape(-1, self._n)
for k,v in self._raw_data.iteritems()}
def get_average(self, name):
"""Get average.
Parameters
----------
name : str
Key name of raw_data.
"""
return self._data.get(name).mean(axis=1)
def get_errorbar(self, name):
"""Get errorbar
Parameters
----------
name : str
Key name of raw_data.
"""
return self._data.get(name).std(axis=1)
def get_all_names(self):
"""Get all key names of raw_data.
Returns
-------
ret : list
List of keys.
"""
return self._data.keys()
def save(self, ext='dat'):
"""
"""
pass
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""utils for data crunching, saving.
"""
import numpy as np
class ScanDataFactory(object):
"""Post processor of data from scan server.
Parameters
----------
data : dict
Raw data retrieving from scan server regarding scan ID, after
completing certain scan task.
n_sample : int
Sample number for every scan device setup.
"""
def __init__(self, data, n_sample):
self._raw_data = data
self._n = n_sample
self._rebuild_data()
def _rebuild_data(self):
"""Rebuild raw_data
"""
self._data = {k:np.array(v.get('value')).reshape(-1, self._n)
for k,v in self._raw_data.iteritems()}
def get_average(self, name):
"""Get average.
Parameters
----------
name : str
Key name of raw_data.
"""
return self._data.get(name).mean(axis=1)
def get_errorbar(self, name):
"""Get errorbar
Parameters
----------
name : str
Key name of raw_data.
"""
return self._data.get(name).std(axis=1)
def get_all_names(self):
"""Get all key names of raw_data.
Returns
-------
ret : list
List of keys.
"""
return self._data.keys()
def save(self, ext='dat'):
"""
"""
pass
<commit_msg>Make raw_data and data as properties<commit_after> | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""utils for data crunching, saving.
"""
import numpy as np
class ScanDataFactory(object):
"""Post processor of data from scan server.
Parameters
----------
data : dict
Raw data retrieving from scan server regarding scan ID, after
completing certain scan task.
n_sample : int
Sample number for every scan device setup.
"""
def __init__(self, data, n_sample):
self._raw_data = data
self._n = n_sample
self._rebuild_data()
@property
def raw_data(self):
"""dict: Dict of array, raw scan data."""
return self._raw_data
@property
def data(self):
"""dict: Dict of array, raw scan data after postprocessing."""
return self._data
def _rebuild_data(self):
"""Rebuild raw_data
"""
self._data = {k:np.array(v.get('value')).reshape(-1, self._n)
for k,v in self._raw_data.iteritems()}
def get_average(self, name):
"""Get average.
Parameters
----------
name : str
Key name of raw_data.
"""
return self._data.get(name).mean(axis=1)
def get_errorbar(self, name):
"""Get errorbar
Parameters
----------
name : str
Key name of raw_data.
"""
return self._data.get(name).std(axis=1)
def get_all_names(self):
"""Get all key names of raw_data.
Returns
-------
ret : list
List of keys.
"""
return self._data.keys()
def save(self, ext='dat'):
"""
"""
pass
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""utils for data crunching, saving.
"""
import numpy as np
class ScanDataFactory(object):
"""Post processor of data from scan server.
Parameters
----------
data : dict
Raw data retrieving from scan server regarding scan ID, after
completing certain scan task.
n_sample : int
Sample number for every scan device setup.
"""
def __init__(self, data, n_sample):
self._raw_data = data
self._n = n_sample
self._rebuild_data()
def _rebuild_data(self):
"""Rebuild raw_data
"""
self._data = {k:np.array(v.get('value')).reshape(-1, self._n)
for k,v in self._raw_data.iteritems()}
def get_average(self, name):
"""Get average.
Parameters
----------
name : str
Key name of raw_data.
"""
return self._data.get(name).mean(axis=1)
def get_errorbar(self, name):
"""Get errorbar
Parameters
----------
name : str
Key name of raw_data.
"""
return self._data.get(name).std(axis=1)
def get_all_names(self):
"""Get all key names of raw_data.
Returns
-------
ret : list
List of keys.
"""
return self._data.keys()
def save(self, ext='dat'):
"""
"""
pass
Make raw_data and data as properties#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""utils for data crunching, saving.
"""
import numpy as np
class ScanDataFactory(object):
"""Post processor of data from scan server.
Parameters
----------
data : dict
Raw data retrieving from scan server regarding scan ID, after
completing certain scan task.
n_sample : int
Sample number for every scan device setup.
"""
def __init__(self, data, n_sample):
self._raw_data = data
self._n = n_sample
self._rebuild_data()
@property
def raw_data(self):
"""dict: Dict of array, raw scan data."""
return self._raw_data
@property
def data(self):
"""dict: Dict of array, raw scan data after postprocessing."""
return self._data
def _rebuild_data(self):
"""Rebuild raw_data
"""
self._data = {k:np.array(v.get('value')).reshape(-1, self._n)
for k,v in self._raw_data.iteritems()}
def get_average(self, name):
"""Get average.
Parameters
----------
name : str
Key name of raw_data.
"""
return self._data.get(name).mean(axis=1)
def get_errorbar(self, name):
"""Get errorbar
Parameters
----------
name : str
Key name of raw_data.
"""
return self._data.get(name).std(axis=1)
def get_all_names(self):
"""Get all key names of raw_data.
Returns
-------
ret : list
List of keys.
"""
return self._data.keys()
def save(self, ext='dat'):
"""
"""
pass
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""utils for data crunching, saving.
"""
import numpy as np
class ScanDataFactory(object):
"""Post processor of data from scan server.
Parameters
----------
data : dict
Raw data retrieving from scan server regarding scan ID, after
completing certain scan task.
n_sample : int
Sample number for every scan device setup.
"""
def __init__(self, data, n_sample):
self._raw_data = data
self._n = n_sample
self._rebuild_data()
def _rebuild_data(self):
"""Rebuild raw_data
"""
self._data = {k:np.array(v.get('value')).reshape(-1, self._n)
for k,v in self._raw_data.iteritems()}
def get_average(self, name):
"""Get average.
Parameters
----------
name : str
Key name of raw_data.
"""
return self._data.get(name).mean(axis=1)
def get_errorbar(self, name):
"""Get errorbar
Parameters
----------
name : str
Key name of raw_data.
"""
return self._data.get(name).std(axis=1)
def get_all_names(self):
"""Get all key names of raw_data.
Returns
-------
ret : list
List of keys.
"""
return self._data.keys()
def save(self, ext='dat'):
"""
"""
pass
<commit_msg>Make raw_data and data as properties<commit_after>#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""utils for data crunching, saving.
"""
import numpy as np
class ScanDataFactory(object):
"""Post processor of data from scan server.
Parameters
----------
data : dict
Raw data retrieving from scan server regarding scan ID, after
completing certain scan task.
n_sample : int
Sample number for every scan device setup.
"""
def __init__(self, data, n_sample):
self._raw_data = data
self._n = n_sample
self._rebuild_data()
@property
def raw_data(self):
"""dict: Dict of array, raw scan data."""
return self._raw_data
@property
def data(self):
"""dict: Dict of array, raw scan data after postprocessing."""
return self._data
def _rebuild_data(self):
"""Rebuild raw_data
"""
self._data = {k:np.array(v.get('value')).reshape(-1, self._n)
for k,v in self._raw_data.iteritems()}
def get_average(self, name):
"""Get average.
Parameters
----------
name : str
Key name of raw_data.
"""
return self._data.get(name).mean(axis=1)
def get_errorbar(self, name):
"""Get errorbar
Parameters
----------
name : str
Key name of raw_data.
"""
return self._data.get(name).std(axis=1)
def get_all_names(self):
"""Get all key names of raw_data.
Returns
-------
ret : list
List of keys.
"""
return self._data.keys()
def save(self, ext='dat'):
"""
"""
pass
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.