commit
stringlengths
40
40
old_file
stringlengths
4
118
new_file
stringlengths
4
118
old_contents
stringlengths
0
2.94k
new_contents
stringlengths
1
4.43k
subject
stringlengths
15
444
message
stringlengths
16
3.45k
lang
stringclasses
1 value
license
stringclasses
13 values
repos
stringlengths
5
43.2k
prompt
stringlengths
17
4.58k
response
stringlengths
1
4.43k
prompt_tagged
stringlengths
58
4.62k
response_tagged
stringlengths
1
4.43k
text
stringlengths
132
7.29k
text_tagged
stringlengths
173
7.33k
f1cc57acd774eace69da7ec0ae9a516207d8ff5c
pyrfc3339/__init__.py
pyrfc3339/__init__.py
""" pyRFC3339 parses and generates :RFC:`3339`-compliant timestamps using Python :class:`datetime.datetime` objects. >>> from pyrfc3339 import generate, parse >>> from datetime import datetime >>> import pytz >>> generate(datetime.utcnow().replace(tzinfo=pytz.utc)) #doctest:+ELLIPSIS '...T...Z' >>> parse('2009-01-01T10:01:02Z') datetime.datetime(2009, 1, 1, 10, 1, 2, tzinfo=<UTC>) >>> parse('2009-01-01T14:01:02-04:00') datetime.datetime(2009, 1, 1, 14, 1, 2, tzinfo=<UTC-04:00>) """ from generator import generate from parser import parse __all__ = ['generate', 'parse']
""" pyRFC3339 parses and generates :RFC:`3339`-compliant timestamps using Python :class:`datetime.datetime` objects. >>> from pyrfc3339 import generate, parse >>> from datetime import datetime >>> import pytz >>> generate(datetime.utcnow().replace(tzinfo=pytz.utc)) #doctest:+ELLIPSIS '...T...Z' >>> parse('2009-01-01T10:01:02Z') datetime.datetime(2009, 1, 1, 10, 1, 2, tzinfo=<UTC>) >>> parse('2009-01-01T14:01:02-04:00') datetime.datetime(2009, 1, 1, 14, 1, 2, tzinfo=<UTC-04:00>) """ from pyrfc3339.generator import generate from pyrfc3339.parser import parse __all__ = ['generate', 'parse']
Fix imports for Python 3
Fix imports for Python 3
Python
mit
kurtraschke/pyRFC3339
""" pyRFC3339 parses and generates :RFC:`3339`-compliant timestamps using Python :class:`datetime.datetime` objects. >>> from pyrfc3339 import generate, parse >>> from datetime import datetime >>> import pytz >>> generate(datetime.utcnow().replace(tzinfo=pytz.utc)) #doctest:+ELLIPSIS '...T...Z' >>> parse('2009-01-01T10:01:02Z') datetime.datetime(2009, 1, 1, 10, 1, 2, tzinfo=<UTC>) >>> parse('2009-01-01T14:01:02-04:00') datetime.datetime(2009, 1, 1, 14, 1, 2, tzinfo=<UTC-04:00>) """ from generator import generate from parser import parse __all__ = ['generate', 'parse'] Fix imports for Python 3
""" pyRFC3339 parses and generates :RFC:`3339`-compliant timestamps using Python :class:`datetime.datetime` objects. >>> from pyrfc3339 import generate, parse >>> from datetime import datetime >>> import pytz >>> generate(datetime.utcnow().replace(tzinfo=pytz.utc)) #doctest:+ELLIPSIS '...T...Z' >>> parse('2009-01-01T10:01:02Z') datetime.datetime(2009, 1, 1, 10, 1, 2, tzinfo=<UTC>) >>> parse('2009-01-01T14:01:02-04:00') datetime.datetime(2009, 1, 1, 14, 1, 2, tzinfo=<UTC-04:00>) """ from pyrfc3339.generator import generate from pyrfc3339.parser import parse __all__ = ['generate', 'parse']
<commit_before>""" pyRFC3339 parses and generates :RFC:`3339`-compliant timestamps using Python :class:`datetime.datetime` objects. >>> from pyrfc3339 import generate, parse >>> from datetime import datetime >>> import pytz >>> generate(datetime.utcnow().replace(tzinfo=pytz.utc)) #doctest:+ELLIPSIS '...T...Z' >>> parse('2009-01-01T10:01:02Z') datetime.datetime(2009, 1, 1, 10, 1, 2, tzinfo=<UTC>) >>> parse('2009-01-01T14:01:02-04:00') datetime.datetime(2009, 1, 1, 14, 1, 2, tzinfo=<UTC-04:00>) """ from generator import generate from parser import parse __all__ = ['generate', 'parse'] <commit_msg>Fix imports for Python 3<commit_after>
""" pyRFC3339 parses and generates :RFC:`3339`-compliant timestamps using Python :class:`datetime.datetime` objects. >>> from pyrfc3339 import generate, parse >>> from datetime import datetime >>> import pytz >>> generate(datetime.utcnow().replace(tzinfo=pytz.utc)) #doctest:+ELLIPSIS '...T...Z' >>> parse('2009-01-01T10:01:02Z') datetime.datetime(2009, 1, 1, 10, 1, 2, tzinfo=<UTC>) >>> parse('2009-01-01T14:01:02-04:00') datetime.datetime(2009, 1, 1, 14, 1, 2, tzinfo=<UTC-04:00>) """ from pyrfc3339.generator import generate from pyrfc3339.parser import parse __all__ = ['generate', 'parse']
""" pyRFC3339 parses and generates :RFC:`3339`-compliant timestamps using Python :class:`datetime.datetime` objects. >>> from pyrfc3339 import generate, parse >>> from datetime import datetime >>> import pytz >>> generate(datetime.utcnow().replace(tzinfo=pytz.utc)) #doctest:+ELLIPSIS '...T...Z' >>> parse('2009-01-01T10:01:02Z') datetime.datetime(2009, 1, 1, 10, 1, 2, tzinfo=<UTC>) >>> parse('2009-01-01T14:01:02-04:00') datetime.datetime(2009, 1, 1, 14, 1, 2, tzinfo=<UTC-04:00>) """ from generator import generate from parser import parse __all__ = ['generate', 'parse'] Fix imports for Python 3""" pyRFC3339 parses and generates :RFC:`3339`-compliant timestamps using Python :class:`datetime.datetime` objects. >>> from pyrfc3339 import generate, parse >>> from datetime import datetime >>> import pytz >>> generate(datetime.utcnow().replace(tzinfo=pytz.utc)) #doctest:+ELLIPSIS '...T...Z' >>> parse('2009-01-01T10:01:02Z') datetime.datetime(2009, 1, 1, 10, 1, 2, tzinfo=<UTC>) >>> parse('2009-01-01T14:01:02-04:00') datetime.datetime(2009, 1, 1, 14, 1, 2, tzinfo=<UTC-04:00>) """ from pyrfc3339.generator import generate from pyrfc3339.parser import parse __all__ = ['generate', 'parse']
<commit_before>""" pyRFC3339 parses and generates :RFC:`3339`-compliant timestamps using Python :class:`datetime.datetime` objects. >>> from pyrfc3339 import generate, parse >>> from datetime import datetime >>> import pytz >>> generate(datetime.utcnow().replace(tzinfo=pytz.utc)) #doctest:+ELLIPSIS '...T...Z' >>> parse('2009-01-01T10:01:02Z') datetime.datetime(2009, 1, 1, 10, 1, 2, tzinfo=<UTC>) >>> parse('2009-01-01T14:01:02-04:00') datetime.datetime(2009, 1, 1, 14, 1, 2, tzinfo=<UTC-04:00>) """ from generator import generate from parser import parse __all__ = ['generate', 'parse'] <commit_msg>Fix imports for Python 3<commit_after>""" pyRFC3339 parses and generates :RFC:`3339`-compliant timestamps using Python :class:`datetime.datetime` objects. >>> from pyrfc3339 import generate, parse >>> from datetime import datetime >>> import pytz >>> generate(datetime.utcnow().replace(tzinfo=pytz.utc)) #doctest:+ELLIPSIS '...T...Z' >>> parse('2009-01-01T10:01:02Z') datetime.datetime(2009, 1, 1, 10, 1, 2, tzinfo=<UTC>) >>> parse('2009-01-01T14:01:02-04:00') datetime.datetime(2009, 1, 1, 14, 1, 2, tzinfo=<UTC-04:00>) """ from pyrfc3339.generator import generate from pyrfc3339.parser import parse __all__ = ['generate', 'parse']
8337575314ae02e99eeded1ffb537a87a423b2c0
spam/ansiInventory.py
spam/ansiInventory.py
#!/usr/bin/env python # -*- coding: utf-8 -*- """ AnsibleInventory: INTRO: USAGE: """ import os import ansible.inventory class AnsibleInventory(object): ''' Ansible Inventory wrapper class. ''' def __init__(self, inventory_filename): ''' Initialize Inventory ''' if not os.path.exists(inventory_filename): print "Provide a valid inventory filename" return self.inventory = ansible.inventory.InventoryParser(inventory_filename) def get_hosts(self, group=None): ''' Get the hosts ''' if not group: return self.inventory.hosts.keys() groupobj = self.inventory.groups.get(group, None) if not groupobj: return None hostobjs = groupobj.get_hosts() hostlist = [] for host in hostobjs: hostlist.append(host.name) return hostlist
#!/usr/bin/env python # -*- coding: utf-8 -*- """ AnsibleInventory: INTRO: USAGE: """ import os import ansible.inventory class AnsibleInventory(object): ''' Ansible Inventory wrapper class. ''' def __init__(self, inventory_filename): ''' Initialize Inventory ''' if not os.path.exists(inventory_filename): print "Provide a valid inventory filename" return self.inventory = ansible.inventory.InventoryParser(inventory_filename) def get_hosts(self, group=None): ''' Get the hosts ''' hostlist = [] for group in self.inventory.groups: groupdict = {} groupdict['group'] = group groupdict['hostlist'] = [] groupobj = self.inventory.groups.get(group) for host in groupobj.get_hosts(): groupdict['hostlist'].append(host.name) hostlist.append(groupdict) return hostlist
Make changes to get_hosts() to return a list of dict
Make changes to get_hosts() to return a list of dict
Python
apache-2.0
bdastur/spam,bdastur/spam
#!/usr/bin/env python # -*- coding: utf-8 -*- """ AnsibleInventory: INTRO: USAGE: """ import os import ansible.inventory class AnsibleInventory(object): ''' Ansible Inventory wrapper class. ''' def __init__(self, inventory_filename): ''' Initialize Inventory ''' if not os.path.exists(inventory_filename): print "Provide a valid inventory filename" return self.inventory = ansible.inventory.InventoryParser(inventory_filename) def get_hosts(self, group=None): ''' Get the hosts ''' if not group: return self.inventory.hosts.keys() groupobj = self.inventory.groups.get(group, None) if not groupobj: return None hostobjs = groupobj.get_hosts() hostlist = [] for host in hostobjs: hostlist.append(host.name) return hostlist Make changes to get_hosts() to return a list of dict
#!/usr/bin/env python # -*- coding: utf-8 -*- """ AnsibleInventory: INTRO: USAGE: """ import os import ansible.inventory class AnsibleInventory(object): ''' Ansible Inventory wrapper class. ''' def __init__(self, inventory_filename): ''' Initialize Inventory ''' if not os.path.exists(inventory_filename): print "Provide a valid inventory filename" return self.inventory = ansible.inventory.InventoryParser(inventory_filename) def get_hosts(self, group=None): ''' Get the hosts ''' hostlist = [] for group in self.inventory.groups: groupdict = {} groupdict['group'] = group groupdict['hostlist'] = [] groupobj = self.inventory.groups.get(group) for host in groupobj.get_hosts(): groupdict['hostlist'].append(host.name) hostlist.append(groupdict) return hostlist
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- """ AnsibleInventory: INTRO: USAGE: """ import os import ansible.inventory class AnsibleInventory(object): ''' Ansible Inventory wrapper class. ''' def __init__(self, inventory_filename): ''' Initialize Inventory ''' if not os.path.exists(inventory_filename): print "Provide a valid inventory filename" return self.inventory = ansible.inventory.InventoryParser(inventory_filename) def get_hosts(self, group=None): ''' Get the hosts ''' if not group: return self.inventory.hosts.keys() groupobj = self.inventory.groups.get(group, None) if not groupobj: return None hostobjs = groupobj.get_hosts() hostlist = [] for host in hostobjs: hostlist.append(host.name) return hostlist <commit_msg>Make changes to get_hosts() to return a list of dict<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- """ AnsibleInventory: INTRO: USAGE: """ import os import ansible.inventory class AnsibleInventory(object): ''' Ansible Inventory wrapper class. ''' def __init__(self, inventory_filename): ''' Initialize Inventory ''' if not os.path.exists(inventory_filename): print "Provide a valid inventory filename" return self.inventory = ansible.inventory.InventoryParser(inventory_filename) def get_hosts(self, group=None): ''' Get the hosts ''' hostlist = [] for group in self.inventory.groups: groupdict = {} groupdict['group'] = group groupdict['hostlist'] = [] groupobj = self.inventory.groups.get(group) for host in groupobj.get_hosts(): groupdict['hostlist'].append(host.name) hostlist.append(groupdict) return hostlist
#!/usr/bin/env python # -*- coding: utf-8 -*- """ AnsibleInventory: INTRO: USAGE: """ import os import ansible.inventory class AnsibleInventory(object): ''' Ansible Inventory wrapper class. ''' def __init__(self, inventory_filename): ''' Initialize Inventory ''' if not os.path.exists(inventory_filename): print "Provide a valid inventory filename" return self.inventory = ansible.inventory.InventoryParser(inventory_filename) def get_hosts(self, group=None): ''' Get the hosts ''' if not group: return self.inventory.hosts.keys() groupobj = self.inventory.groups.get(group, None) if not groupobj: return None hostobjs = groupobj.get_hosts() hostlist = [] for host in hostobjs: hostlist.append(host.name) return hostlist Make changes to get_hosts() to return a list of dict#!/usr/bin/env python # -*- coding: utf-8 -*- """ AnsibleInventory: INTRO: USAGE: """ import os import ansible.inventory class AnsibleInventory(object): ''' Ansible Inventory wrapper class. ''' def __init__(self, inventory_filename): ''' Initialize Inventory ''' if not os.path.exists(inventory_filename): print "Provide a valid inventory filename" return self.inventory = ansible.inventory.InventoryParser(inventory_filename) def get_hosts(self, group=None): ''' Get the hosts ''' hostlist = [] for group in self.inventory.groups: groupdict = {} groupdict['group'] = group groupdict['hostlist'] = [] groupobj = self.inventory.groups.get(group) for host in groupobj.get_hosts(): groupdict['hostlist'].append(host.name) hostlist.append(groupdict) return hostlist
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- """ AnsibleInventory: INTRO: USAGE: """ import os import ansible.inventory class AnsibleInventory(object): ''' Ansible Inventory wrapper class. ''' def __init__(self, inventory_filename): ''' Initialize Inventory ''' if not os.path.exists(inventory_filename): print "Provide a valid inventory filename" return self.inventory = ansible.inventory.InventoryParser(inventory_filename) def get_hosts(self, group=None): ''' Get the hosts ''' if not group: return self.inventory.hosts.keys() groupobj = self.inventory.groups.get(group, None) if not groupobj: return None hostobjs = groupobj.get_hosts() hostlist = [] for host in hostobjs: hostlist.append(host.name) return hostlist <commit_msg>Make changes to get_hosts() to return a list of dict<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- """ AnsibleInventory: INTRO: USAGE: """ import os import ansible.inventory class AnsibleInventory(object): ''' Ansible Inventory wrapper class. ''' def __init__(self, inventory_filename): ''' Initialize Inventory ''' if not os.path.exists(inventory_filename): print "Provide a valid inventory filename" return self.inventory = ansible.inventory.InventoryParser(inventory_filename) def get_hosts(self, group=None): ''' Get the hosts ''' hostlist = [] for group in self.inventory.groups: groupdict = {} groupdict['group'] = group groupdict['hostlist'] = [] groupobj = self.inventory.groups.get(group) for host in groupobj.get_hosts(): groupdict['hostlist'].append(host.name) hostlist.append(groupdict) return hostlist
ffa3d12e5b45cad56367726bdce83de509bc33a7
state_tracker/state_defs.py
state_tracker/state_defs.py
# Copyright (c) 2001, Stanford University # All rights reserved. # # See the file LICENSE.txt for information on redistributing this software. import sys sys.path.append( "../glapi_parser" ) import apiutil apiutil.CopyrightDef() print """DESCRIPTION "" EXPORTS """ keys = apiutil.GetDispatchedFunctions("../glapi_parser/APIspec.txt") for func_name in apiutil.AllSpecials( 'state' ): print "crState%s" % func_name for func_name in apiutil.AllSpecials( 'state_feedback' ): print "crStateFeedback%s" % func_name for func_name in apiutil.AllSpecials( 'state_select' ): print "crStateSelect%s" % func_name print """crStateInit crStateCreateContext crStateDestroyContext crStateDiffContext crStateSwitchContext crStateMakeCurrent crStateSetCurrent crStateFlushFunc crStateFlushArg crStateDiffAPI crStateSetCurrentPointers crStateCurrentRecover crStateTransformUpdateTransform crStateColorMaterialRecover crStateError crStateUpdateColorBits crStateClientInit crStateGetCurrent crStateLimitsInit crStateMergeExtensions crStateRasterPosUpdate crStateTextureCheckDirtyImages crStateExtensionsInit crStateSetExtensionString crStateNativePixelPacking crStateUseServerArrays crStateComputeVersion __currentBits """
# Copyright (c) 2001, Stanford University # All rights reserved. # # See the file LICENSE.txt for information on redistributing this software. import sys sys.path.append( "../glapi_parser" ) import apiutil apiutil.CopyrightDef() print """DESCRIPTION "" EXPORTS """ keys = apiutil.GetDispatchedFunctions("../glapi_parser/APIspec.txt") for func_name in apiutil.AllSpecials( 'state' ): print "crState%s" % func_name for func_name in apiutil.AllSpecials( 'state_feedback' ): print "crStateFeedback%s" % func_name for func_name in apiutil.AllSpecials( 'state_select' ): print "crStateSelect%s" % func_name print """crStateInit crStateReadPixels crStateGetChromiumParametervCR crStateCreateContext crStateDestroyContext crStateDiffContext crStateSwitchContext crStateMakeCurrent crStateSetCurrent crStateFlushFunc crStateFlushArg crStateDiffAPI crStateSetCurrentPointers crStateCurrentRecover crStateTransformUpdateTransform crStateColorMaterialRecover crStateError crStateUpdateColorBits crStateClientInit crStateGetCurrent crStateLimitsInit crStateMergeExtensions crStateRasterPosUpdate crStateTextureCheckDirtyImages crStateExtensionsInit crStateSetExtensionString crStateNativePixelPacking crStateUseServerArrays crStateComputeVersion __currentBits """
Fix defs for ReadPixels and GetChromiumParametervCR
Fix defs for ReadPixels and GetChromiumParametervCR
Python
bsd-3-clause
rpavlik/chromium,rpavlik/chromium,rpavlik/chromium,rpavlik/chromium,rpavlik/chromium
# Copyright (c) 2001, Stanford University # All rights reserved. # # See the file LICENSE.txt for information on redistributing this software. import sys sys.path.append( "../glapi_parser" ) import apiutil apiutil.CopyrightDef() print """DESCRIPTION "" EXPORTS """ keys = apiutil.GetDispatchedFunctions("../glapi_parser/APIspec.txt") for func_name in apiutil.AllSpecials( 'state' ): print "crState%s" % func_name for func_name in apiutil.AllSpecials( 'state_feedback' ): print "crStateFeedback%s" % func_name for func_name in apiutil.AllSpecials( 'state_select' ): print "crStateSelect%s" % func_name print """crStateInit crStateCreateContext crStateDestroyContext crStateDiffContext crStateSwitchContext crStateMakeCurrent crStateSetCurrent crStateFlushFunc crStateFlushArg crStateDiffAPI crStateSetCurrentPointers crStateCurrentRecover crStateTransformUpdateTransform crStateColorMaterialRecover crStateError crStateUpdateColorBits crStateClientInit crStateGetCurrent crStateLimitsInit crStateMergeExtensions crStateRasterPosUpdate crStateTextureCheckDirtyImages crStateExtensionsInit crStateSetExtensionString crStateNativePixelPacking crStateUseServerArrays crStateComputeVersion __currentBits """ Fix defs for ReadPixels and GetChromiumParametervCR
# Copyright (c) 2001, Stanford University # All rights reserved. # # See the file LICENSE.txt for information on redistributing this software. import sys sys.path.append( "../glapi_parser" ) import apiutil apiutil.CopyrightDef() print """DESCRIPTION "" EXPORTS """ keys = apiutil.GetDispatchedFunctions("../glapi_parser/APIspec.txt") for func_name in apiutil.AllSpecials( 'state' ): print "crState%s" % func_name for func_name in apiutil.AllSpecials( 'state_feedback' ): print "crStateFeedback%s" % func_name for func_name in apiutil.AllSpecials( 'state_select' ): print "crStateSelect%s" % func_name print """crStateInit crStateReadPixels crStateGetChromiumParametervCR crStateCreateContext crStateDestroyContext crStateDiffContext crStateSwitchContext crStateMakeCurrent crStateSetCurrent crStateFlushFunc crStateFlushArg crStateDiffAPI crStateSetCurrentPointers crStateCurrentRecover crStateTransformUpdateTransform crStateColorMaterialRecover crStateError crStateUpdateColorBits crStateClientInit crStateGetCurrent crStateLimitsInit crStateMergeExtensions crStateRasterPosUpdate crStateTextureCheckDirtyImages crStateExtensionsInit crStateSetExtensionString crStateNativePixelPacking crStateUseServerArrays crStateComputeVersion __currentBits """
<commit_before># Copyright (c) 2001, Stanford University # All rights reserved. # # See the file LICENSE.txt for information on redistributing this software. import sys sys.path.append( "../glapi_parser" ) import apiutil apiutil.CopyrightDef() print """DESCRIPTION "" EXPORTS """ keys = apiutil.GetDispatchedFunctions("../glapi_parser/APIspec.txt") for func_name in apiutil.AllSpecials( 'state' ): print "crState%s" % func_name for func_name in apiutil.AllSpecials( 'state_feedback' ): print "crStateFeedback%s" % func_name for func_name in apiutil.AllSpecials( 'state_select' ): print "crStateSelect%s" % func_name print """crStateInit crStateCreateContext crStateDestroyContext crStateDiffContext crStateSwitchContext crStateMakeCurrent crStateSetCurrent crStateFlushFunc crStateFlushArg crStateDiffAPI crStateSetCurrentPointers crStateCurrentRecover crStateTransformUpdateTransform crStateColorMaterialRecover crStateError crStateUpdateColorBits crStateClientInit crStateGetCurrent crStateLimitsInit crStateMergeExtensions crStateRasterPosUpdate crStateTextureCheckDirtyImages crStateExtensionsInit crStateSetExtensionString crStateNativePixelPacking crStateUseServerArrays crStateComputeVersion __currentBits """ <commit_msg>Fix defs for ReadPixels and GetChromiumParametervCR<commit_after>
# Copyright (c) 2001, Stanford University # All rights reserved. # # See the file LICENSE.txt for information on redistributing this software. import sys sys.path.append( "../glapi_parser" ) import apiutil apiutil.CopyrightDef() print """DESCRIPTION "" EXPORTS """ keys = apiutil.GetDispatchedFunctions("../glapi_parser/APIspec.txt") for func_name in apiutil.AllSpecials( 'state' ): print "crState%s" % func_name for func_name in apiutil.AllSpecials( 'state_feedback' ): print "crStateFeedback%s" % func_name for func_name in apiutil.AllSpecials( 'state_select' ): print "crStateSelect%s" % func_name print """crStateInit crStateReadPixels crStateGetChromiumParametervCR crStateCreateContext crStateDestroyContext crStateDiffContext crStateSwitchContext crStateMakeCurrent crStateSetCurrent crStateFlushFunc crStateFlushArg crStateDiffAPI crStateSetCurrentPointers crStateCurrentRecover crStateTransformUpdateTransform crStateColorMaterialRecover crStateError crStateUpdateColorBits crStateClientInit crStateGetCurrent crStateLimitsInit crStateMergeExtensions crStateRasterPosUpdate crStateTextureCheckDirtyImages crStateExtensionsInit crStateSetExtensionString crStateNativePixelPacking crStateUseServerArrays crStateComputeVersion __currentBits """
# Copyright (c) 2001, Stanford University # All rights reserved. # # See the file LICENSE.txt for information on redistributing this software. import sys sys.path.append( "../glapi_parser" ) import apiutil apiutil.CopyrightDef() print """DESCRIPTION "" EXPORTS """ keys = apiutil.GetDispatchedFunctions("../glapi_parser/APIspec.txt") for func_name in apiutil.AllSpecials( 'state' ): print "crState%s" % func_name for func_name in apiutil.AllSpecials( 'state_feedback' ): print "crStateFeedback%s" % func_name for func_name in apiutil.AllSpecials( 'state_select' ): print "crStateSelect%s" % func_name print """crStateInit crStateCreateContext crStateDestroyContext crStateDiffContext crStateSwitchContext crStateMakeCurrent crStateSetCurrent crStateFlushFunc crStateFlushArg crStateDiffAPI crStateSetCurrentPointers crStateCurrentRecover crStateTransformUpdateTransform crStateColorMaterialRecover crStateError crStateUpdateColorBits crStateClientInit crStateGetCurrent crStateLimitsInit crStateMergeExtensions crStateRasterPosUpdate crStateTextureCheckDirtyImages crStateExtensionsInit crStateSetExtensionString crStateNativePixelPacking crStateUseServerArrays crStateComputeVersion __currentBits """ Fix defs for ReadPixels and GetChromiumParametervCR# Copyright (c) 2001, Stanford University # All rights reserved. # # See the file LICENSE.txt for information on redistributing this software. import sys sys.path.append( "../glapi_parser" ) import apiutil apiutil.CopyrightDef() print """DESCRIPTION "" EXPORTS """ keys = apiutil.GetDispatchedFunctions("../glapi_parser/APIspec.txt") for func_name in apiutil.AllSpecials( 'state' ): print "crState%s" % func_name for func_name in apiutil.AllSpecials( 'state_feedback' ): print "crStateFeedback%s" % func_name for func_name in apiutil.AllSpecials( 'state_select' ): print "crStateSelect%s" % func_name print """crStateInit crStateReadPixels crStateGetChromiumParametervCR crStateCreateContext crStateDestroyContext crStateDiffContext crStateSwitchContext crStateMakeCurrent crStateSetCurrent crStateFlushFunc crStateFlushArg crStateDiffAPI crStateSetCurrentPointers crStateCurrentRecover crStateTransformUpdateTransform crStateColorMaterialRecover crStateError crStateUpdateColorBits crStateClientInit crStateGetCurrent crStateLimitsInit crStateMergeExtensions crStateRasterPosUpdate crStateTextureCheckDirtyImages crStateExtensionsInit crStateSetExtensionString crStateNativePixelPacking crStateUseServerArrays crStateComputeVersion __currentBits """
<commit_before># Copyright (c) 2001, Stanford University # All rights reserved. # # See the file LICENSE.txt for information on redistributing this software. import sys sys.path.append( "../glapi_parser" ) import apiutil apiutil.CopyrightDef() print """DESCRIPTION "" EXPORTS """ keys = apiutil.GetDispatchedFunctions("../glapi_parser/APIspec.txt") for func_name in apiutil.AllSpecials( 'state' ): print "crState%s" % func_name for func_name in apiutil.AllSpecials( 'state_feedback' ): print "crStateFeedback%s" % func_name for func_name in apiutil.AllSpecials( 'state_select' ): print "crStateSelect%s" % func_name print """crStateInit crStateCreateContext crStateDestroyContext crStateDiffContext crStateSwitchContext crStateMakeCurrent crStateSetCurrent crStateFlushFunc crStateFlushArg crStateDiffAPI crStateSetCurrentPointers crStateCurrentRecover crStateTransformUpdateTransform crStateColorMaterialRecover crStateError crStateUpdateColorBits crStateClientInit crStateGetCurrent crStateLimitsInit crStateMergeExtensions crStateRasterPosUpdate crStateTextureCheckDirtyImages crStateExtensionsInit crStateSetExtensionString crStateNativePixelPacking crStateUseServerArrays crStateComputeVersion __currentBits """ <commit_msg>Fix defs for ReadPixels and GetChromiumParametervCR<commit_after># Copyright (c) 2001, Stanford University # All rights reserved. # # See the file LICENSE.txt for information on redistributing this software. import sys sys.path.append( "../glapi_parser" ) import apiutil apiutil.CopyrightDef() print """DESCRIPTION "" EXPORTS """ keys = apiutil.GetDispatchedFunctions("../glapi_parser/APIspec.txt") for func_name in apiutil.AllSpecials( 'state' ): print "crState%s" % func_name for func_name in apiutil.AllSpecials( 'state_feedback' ): print "crStateFeedback%s" % func_name for func_name in apiutil.AllSpecials( 'state_select' ): print "crStateSelect%s" % func_name print """crStateInit crStateReadPixels crStateGetChromiumParametervCR crStateCreateContext crStateDestroyContext crStateDiffContext crStateSwitchContext crStateMakeCurrent crStateSetCurrent crStateFlushFunc crStateFlushArg crStateDiffAPI crStateSetCurrentPointers crStateCurrentRecover crStateTransformUpdateTransform crStateColorMaterialRecover crStateError crStateUpdateColorBits crStateClientInit crStateGetCurrent crStateLimitsInit crStateMergeExtensions crStateRasterPosUpdate crStateTextureCheckDirtyImages crStateExtensionsInit crStateSetExtensionString crStateNativePixelPacking crStateUseServerArrays crStateComputeVersion __currentBits """
2b249d8a81c51d30d9175ac033c7a0b208684d59
tests/test_basic.py
tests/test_basic.py
import sys import pubrunner def test_countwords(): pubrunner.pubrun('examples/CountWords/',True,True)
import sys import pubrunner import pubrunner.command_line def test_countwords(): #pubrunner.pubrun('examples/CountWords/',True,True) sys.argv = ['--test','examples/CountWords/'] pubrunner.command_line.main()
Test case now runs main directly
Test case now runs main directly
Python
mit
jakelever/pubrunner,jakelever/pubrunner
import sys import pubrunner def test_countwords(): pubrunner.pubrun('examples/CountWords/',True,True) Test case now runs main directly
import sys import pubrunner import pubrunner.command_line def test_countwords(): #pubrunner.pubrun('examples/CountWords/',True,True) sys.argv = ['--test','examples/CountWords/'] pubrunner.command_line.main()
<commit_before>import sys import pubrunner def test_countwords(): pubrunner.pubrun('examples/CountWords/',True,True) <commit_msg>Test case now runs main directly<commit_after>
import sys import pubrunner import pubrunner.command_line def test_countwords(): #pubrunner.pubrun('examples/CountWords/',True,True) sys.argv = ['--test','examples/CountWords/'] pubrunner.command_line.main()
import sys import pubrunner def test_countwords(): pubrunner.pubrun('examples/CountWords/',True,True) Test case now runs main directlyimport sys import pubrunner import pubrunner.command_line def test_countwords(): #pubrunner.pubrun('examples/CountWords/',True,True) sys.argv = ['--test','examples/CountWords/'] pubrunner.command_line.main()
<commit_before>import sys import pubrunner def test_countwords(): pubrunner.pubrun('examples/CountWords/',True,True) <commit_msg>Test case now runs main directly<commit_after>import sys import pubrunner import pubrunner.command_line def test_countwords(): #pubrunner.pubrun('examples/CountWords/',True,True) sys.argv = ['--test','examples/CountWords/'] pubrunner.command_line.main()
806d3293ebbbd0f30f923e73def902e9c14a0879
tests/test_match.py
tests/test_match.py
import cv2 import numpy from nose.tools import raises import stbt from _stbt.core import _load_template def black(width=1280, height=720): return numpy.zeros((height, width, 3), dtype=numpy.uint8) def test_that_matchresult_image_matches_template_passed_to_match(): assert stbt.match("black.png", frame=black()).image == "black.png" def test_that_matchresult_str_image_matches_template_passed_to_match(): assert "image=\'black.png\'" in str(stbt.match("black.png", frame=black())) def test_that_matchresult_str_image_matches_template_passed_to_match_custom(): assert "image=<Custom Image>" in str( stbt.match(black(30, 30), frame=black())) @raises(ValueError) def test_that_match_rejects_greyscale_template(): grey = cv2.cvtColor(_load_template("black.png").image, cv2.cv.CV_BGR2GRAY) stbt.match(grey, frame=black())
import cv2 import numpy from nose.tools import raises import stbt from _stbt.core import _load_template def black(width=1280, height=720): return numpy.zeros((height, width, 3), dtype=numpy.uint8) def test_that_matchresult_image_matches_template_passed_to_match(): assert stbt.match("black.png", frame=black()).image == "black.png" def test_that_matchresult_str_image_matches_template_passed_to_match(): assert "image=\'black.png\'" in str(stbt.match("black.png", frame=black())) def test_that_matchresult_str_image_matches_template_passed_to_match_custom(): assert "image=<Custom Image>" in str( stbt.match(black(30, 30), frame=black())) def test_matchresult_region_when_first_pyramid_level_fails_to_match(): f = _imread("videotestsrc-full-frame.png") assert stbt.Region(184, 0, width=92, height=160) == stbt.match( "videotestsrc-redblue-flipped.png", frame=f).region @raises(ValueError) def test_that_match_rejects_greyscale_template(): grey = cv2.cvtColor(_load_template("black.png").image, cv2.cv.CV_BGR2GRAY) stbt.match(grey, frame=black())
Add test for the region reported for a failed match
tests: Add test for the region reported for a failed match If the match fails at the first level of the pyramid algorithm (when we're searching in a down-scaled version of the frame) then we have to up-scale the region we report for the best (but not good enough) match. This passes currently, but there was no test for it and I want to make sure I don't break it when I implement `stbt.match_all`.
Python
lgpl-2.1
martynjarvis/stb-tester,LewisHaley/stb-tester,LewisHaley/stb-tester,LewisHaley/stb-tester,martynjarvis/stb-tester,stb-tester/stb-tester,LewisHaley/stb-tester,martynjarvis/stb-tester,LewisHaley/stb-tester,stb-tester/stb-tester,martynjarvis/stb-tester,LewisHaley/stb-tester,martynjarvis/stb-tester,martynjarvis/stb-tester,stb-tester/stb-tester,LewisHaley/stb-tester,martynjarvis/stb-tester,stb-tester/stb-tester
import cv2 import numpy from nose.tools import raises import stbt from _stbt.core import _load_template def black(width=1280, height=720): return numpy.zeros((height, width, 3), dtype=numpy.uint8) def test_that_matchresult_image_matches_template_passed_to_match(): assert stbt.match("black.png", frame=black()).image == "black.png" def test_that_matchresult_str_image_matches_template_passed_to_match(): assert "image=\'black.png\'" in str(stbt.match("black.png", frame=black())) def test_that_matchresult_str_image_matches_template_passed_to_match_custom(): assert "image=<Custom Image>" in str( stbt.match(black(30, 30), frame=black())) @raises(ValueError) def test_that_match_rejects_greyscale_template(): grey = cv2.cvtColor(_load_template("black.png").image, cv2.cv.CV_BGR2GRAY) stbt.match(grey, frame=black()) tests: Add test for the region reported for a failed match If the match fails at the first level of the pyramid algorithm (when we're searching in a down-scaled version of the frame) then we have to up-scale the region we report for the best (but not good enough) match. This passes currently, but there was no test for it and I want to make sure I don't break it when I implement `stbt.match_all`.
import cv2 import numpy from nose.tools import raises import stbt from _stbt.core import _load_template def black(width=1280, height=720): return numpy.zeros((height, width, 3), dtype=numpy.uint8) def test_that_matchresult_image_matches_template_passed_to_match(): assert stbt.match("black.png", frame=black()).image == "black.png" def test_that_matchresult_str_image_matches_template_passed_to_match(): assert "image=\'black.png\'" in str(stbt.match("black.png", frame=black())) def test_that_matchresult_str_image_matches_template_passed_to_match_custom(): assert "image=<Custom Image>" in str( stbt.match(black(30, 30), frame=black())) def test_matchresult_region_when_first_pyramid_level_fails_to_match(): f = _imread("videotestsrc-full-frame.png") assert stbt.Region(184, 0, width=92, height=160) == stbt.match( "videotestsrc-redblue-flipped.png", frame=f).region @raises(ValueError) def test_that_match_rejects_greyscale_template(): grey = cv2.cvtColor(_load_template("black.png").image, cv2.cv.CV_BGR2GRAY) stbt.match(grey, frame=black())
<commit_before>import cv2 import numpy from nose.tools import raises import stbt from _stbt.core import _load_template def black(width=1280, height=720): return numpy.zeros((height, width, 3), dtype=numpy.uint8) def test_that_matchresult_image_matches_template_passed_to_match(): assert stbt.match("black.png", frame=black()).image == "black.png" def test_that_matchresult_str_image_matches_template_passed_to_match(): assert "image=\'black.png\'" in str(stbt.match("black.png", frame=black())) def test_that_matchresult_str_image_matches_template_passed_to_match_custom(): assert "image=<Custom Image>" in str( stbt.match(black(30, 30), frame=black())) @raises(ValueError) def test_that_match_rejects_greyscale_template(): grey = cv2.cvtColor(_load_template("black.png").image, cv2.cv.CV_BGR2GRAY) stbt.match(grey, frame=black()) <commit_msg>tests: Add test for the region reported for a failed match If the match fails at the first level of the pyramid algorithm (when we're searching in a down-scaled version of the frame) then we have to up-scale the region we report for the best (but not good enough) match. This passes currently, but there was no test for it and I want to make sure I don't break it when I implement `stbt.match_all`.<commit_after>
import cv2 import numpy from nose.tools import raises import stbt from _stbt.core import _load_template def black(width=1280, height=720): return numpy.zeros((height, width, 3), dtype=numpy.uint8) def test_that_matchresult_image_matches_template_passed_to_match(): assert stbt.match("black.png", frame=black()).image == "black.png" def test_that_matchresult_str_image_matches_template_passed_to_match(): assert "image=\'black.png\'" in str(stbt.match("black.png", frame=black())) def test_that_matchresult_str_image_matches_template_passed_to_match_custom(): assert "image=<Custom Image>" in str( stbt.match(black(30, 30), frame=black())) def test_matchresult_region_when_first_pyramid_level_fails_to_match(): f = _imread("videotestsrc-full-frame.png") assert stbt.Region(184, 0, width=92, height=160) == stbt.match( "videotestsrc-redblue-flipped.png", frame=f).region @raises(ValueError) def test_that_match_rejects_greyscale_template(): grey = cv2.cvtColor(_load_template("black.png").image, cv2.cv.CV_BGR2GRAY) stbt.match(grey, frame=black())
import cv2 import numpy from nose.tools import raises import stbt from _stbt.core import _load_template def black(width=1280, height=720): return numpy.zeros((height, width, 3), dtype=numpy.uint8) def test_that_matchresult_image_matches_template_passed_to_match(): assert stbt.match("black.png", frame=black()).image == "black.png" def test_that_matchresult_str_image_matches_template_passed_to_match(): assert "image=\'black.png\'" in str(stbt.match("black.png", frame=black())) def test_that_matchresult_str_image_matches_template_passed_to_match_custom(): assert "image=<Custom Image>" in str( stbt.match(black(30, 30), frame=black())) @raises(ValueError) def test_that_match_rejects_greyscale_template(): grey = cv2.cvtColor(_load_template("black.png").image, cv2.cv.CV_BGR2GRAY) stbt.match(grey, frame=black()) tests: Add test for the region reported for a failed match If the match fails at the first level of the pyramid algorithm (when we're searching in a down-scaled version of the frame) then we have to up-scale the region we report for the best (but not good enough) match. This passes currently, but there was no test for it and I want to make sure I don't break it when I implement `stbt.match_all`.import cv2 import numpy from nose.tools import raises import stbt from _stbt.core import _load_template def black(width=1280, height=720): return numpy.zeros((height, width, 3), dtype=numpy.uint8) def test_that_matchresult_image_matches_template_passed_to_match(): assert stbt.match("black.png", frame=black()).image == "black.png" def test_that_matchresult_str_image_matches_template_passed_to_match(): assert "image=\'black.png\'" in str(stbt.match("black.png", frame=black())) def test_that_matchresult_str_image_matches_template_passed_to_match_custom(): assert "image=<Custom Image>" in str( stbt.match(black(30, 30), frame=black())) def test_matchresult_region_when_first_pyramid_level_fails_to_match(): f = _imread("videotestsrc-full-frame.png") assert stbt.Region(184, 0, width=92, height=160) == stbt.match( "videotestsrc-redblue-flipped.png", frame=f).region @raises(ValueError) def test_that_match_rejects_greyscale_template(): grey = cv2.cvtColor(_load_template("black.png").image, cv2.cv.CV_BGR2GRAY) stbt.match(grey, frame=black())
<commit_before>import cv2 import numpy from nose.tools import raises import stbt from _stbt.core import _load_template def black(width=1280, height=720): return numpy.zeros((height, width, 3), dtype=numpy.uint8) def test_that_matchresult_image_matches_template_passed_to_match(): assert stbt.match("black.png", frame=black()).image == "black.png" def test_that_matchresult_str_image_matches_template_passed_to_match(): assert "image=\'black.png\'" in str(stbt.match("black.png", frame=black())) def test_that_matchresult_str_image_matches_template_passed_to_match_custom(): assert "image=<Custom Image>" in str( stbt.match(black(30, 30), frame=black())) @raises(ValueError) def test_that_match_rejects_greyscale_template(): grey = cv2.cvtColor(_load_template("black.png").image, cv2.cv.CV_BGR2GRAY) stbt.match(grey, frame=black()) <commit_msg>tests: Add test for the region reported for a failed match If the match fails at the first level of the pyramid algorithm (when we're searching in a down-scaled version of the frame) then we have to up-scale the region we report for the best (but not good enough) match. This passes currently, but there was no test for it and I want to make sure I don't break it when I implement `stbt.match_all`.<commit_after>import cv2 import numpy from nose.tools import raises import stbt from _stbt.core import _load_template def black(width=1280, height=720): return numpy.zeros((height, width, 3), dtype=numpy.uint8) def test_that_matchresult_image_matches_template_passed_to_match(): assert stbt.match("black.png", frame=black()).image == "black.png" def test_that_matchresult_str_image_matches_template_passed_to_match(): assert "image=\'black.png\'" in str(stbt.match("black.png", frame=black())) def test_that_matchresult_str_image_matches_template_passed_to_match_custom(): assert "image=<Custom Image>" in str( stbt.match(black(30, 30), frame=black())) def test_matchresult_region_when_first_pyramid_level_fails_to_match(): f = _imread("videotestsrc-full-frame.png") assert stbt.Region(184, 0, width=92, height=160) == stbt.match( "videotestsrc-redblue-flipped.png", frame=f).region @raises(ValueError) def test_that_match_rejects_greyscale_template(): grey = cv2.cvtColor(_load_template("black.png").image, cv2.cv.CV_BGR2GRAY) stbt.match(grey, frame=black())
b72f3f6e180bc332579e71b4abeba2b36319a55e
regscrape/settings.py
regscrape/settings.py
TARGET_SERVER = 'www.regulations.gov' DEBUG = True DB_NAME = 'regulations' ES_HOST = 'thrift://localhost:9500' DATA_DIR = '/data' EXTRACTORS = 2 DUMP_START = 0 DUMP_END = 3850000 DUMP_INCREMENT = 50000 MAX_WAIT = 600 CHUNK_SIZE = 10 FILTER = {} INSTANCES = 2 THREADS_PER_INSTANCE = 2 SITES = ['regsdotgov'] try: from local_settings import * except: pass
TARGET_SERVER = 'www.regulations.gov' DEBUG = True DB_NAME = 'regulations' ES_HOST = 'thrift://localhost:9500' DATA_DIR = '/data' EXTRACTORS = 2 DUMP_START = 0 DUMP_END = 3850000 DUMP_INCREMENT = 10000 MAX_WAIT = 600 CHUNK_SIZE = 10 FILTER = {} INSTANCES = 2 THREADS_PER_INSTANCE = 2 SITES = ['regsdotgov'] try: from local_settings import * except: pass
Reduce dump increment to 10k.
Reduce dump increment to 10k.
Python
bsd-3-clause
sunlightlabs/regulations-scraper,sunlightlabs/regulations-scraper,sunlightlabs/regulations-scraper
TARGET_SERVER = 'www.regulations.gov' DEBUG = True DB_NAME = 'regulations' ES_HOST = 'thrift://localhost:9500' DATA_DIR = '/data' EXTRACTORS = 2 DUMP_START = 0 DUMP_END = 3850000 DUMP_INCREMENT = 50000 MAX_WAIT = 600 CHUNK_SIZE = 10 FILTER = {} INSTANCES = 2 THREADS_PER_INSTANCE = 2 SITES = ['regsdotgov'] try: from local_settings import * except: pass Reduce dump increment to 10k.
TARGET_SERVER = 'www.regulations.gov' DEBUG = True DB_NAME = 'regulations' ES_HOST = 'thrift://localhost:9500' DATA_DIR = '/data' EXTRACTORS = 2 DUMP_START = 0 DUMP_END = 3850000 DUMP_INCREMENT = 10000 MAX_WAIT = 600 CHUNK_SIZE = 10 FILTER = {} INSTANCES = 2 THREADS_PER_INSTANCE = 2 SITES = ['regsdotgov'] try: from local_settings import * except: pass
<commit_before>TARGET_SERVER = 'www.regulations.gov' DEBUG = True DB_NAME = 'regulations' ES_HOST = 'thrift://localhost:9500' DATA_DIR = '/data' EXTRACTORS = 2 DUMP_START = 0 DUMP_END = 3850000 DUMP_INCREMENT = 50000 MAX_WAIT = 600 CHUNK_SIZE = 10 FILTER = {} INSTANCES = 2 THREADS_PER_INSTANCE = 2 SITES = ['regsdotgov'] try: from local_settings import * except: pass <commit_msg>Reduce dump increment to 10k.<commit_after>
TARGET_SERVER = 'www.regulations.gov' DEBUG = True DB_NAME = 'regulations' ES_HOST = 'thrift://localhost:9500' DATA_DIR = '/data' EXTRACTORS = 2 DUMP_START = 0 DUMP_END = 3850000 DUMP_INCREMENT = 10000 MAX_WAIT = 600 CHUNK_SIZE = 10 FILTER = {} INSTANCES = 2 THREADS_PER_INSTANCE = 2 SITES = ['regsdotgov'] try: from local_settings import * except: pass
TARGET_SERVER = 'www.regulations.gov' DEBUG = True DB_NAME = 'regulations' ES_HOST = 'thrift://localhost:9500' DATA_DIR = '/data' EXTRACTORS = 2 DUMP_START = 0 DUMP_END = 3850000 DUMP_INCREMENT = 50000 MAX_WAIT = 600 CHUNK_SIZE = 10 FILTER = {} INSTANCES = 2 THREADS_PER_INSTANCE = 2 SITES = ['regsdotgov'] try: from local_settings import * except: pass Reduce dump increment to 10k.TARGET_SERVER = 'www.regulations.gov' DEBUG = True DB_NAME = 'regulations' ES_HOST = 'thrift://localhost:9500' DATA_DIR = '/data' EXTRACTORS = 2 DUMP_START = 0 DUMP_END = 3850000 DUMP_INCREMENT = 10000 MAX_WAIT = 600 CHUNK_SIZE = 10 FILTER = {} INSTANCES = 2 THREADS_PER_INSTANCE = 2 SITES = ['regsdotgov'] try: from local_settings import * except: pass
<commit_before>TARGET_SERVER = 'www.regulations.gov' DEBUG = True DB_NAME = 'regulations' ES_HOST = 'thrift://localhost:9500' DATA_DIR = '/data' EXTRACTORS = 2 DUMP_START = 0 DUMP_END = 3850000 DUMP_INCREMENT = 50000 MAX_WAIT = 600 CHUNK_SIZE = 10 FILTER = {} INSTANCES = 2 THREADS_PER_INSTANCE = 2 SITES = ['regsdotgov'] try: from local_settings import * except: pass <commit_msg>Reduce dump increment to 10k.<commit_after>TARGET_SERVER = 'www.regulations.gov' DEBUG = True DB_NAME = 'regulations' ES_HOST = 'thrift://localhost:9500' DATA_DIR = '/data' EXTRACTORS = 2 DUMP_START = 0 DUMP_END = 3850000 DUMP_INCREMENT = 10000 MAX_WAIT = 600 CHUNK_SIZE = 10 FILTER = {} INSTANCES = 2 THREADS_PER_INSTANCE = 2 SITES = ['regsdotgov'] try: from local_settings import * except: pass
d254cf6960f2d04e02ed252c4461994483a9d0f5
launch_control/models/hw_device.py
launch_control/models/hw_device.py
""" Module with the HardwareDevice model. """ from launch_control.utils.json import PlainOldData class HardwareDevice(PlainOldData): """ Model representing any HardwareDevice A device is just a "device_type" attribute with a bag of properties and a human readable description. Individual device types can be freely added. For simplicity some common types of devices are provided as class properties DEVICE_xxx. Instances will come from a variety of factory classes, each capable of enumerating devices that it understands. The upside of having a common class like this is that it's easier to store it in the database _and_ not have to agree on a common set of properties for, say, all CPUs. If you want you can create instances manually, like this: >>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU) >>> cpu.desc = "800MHz OMAP3 Processor" >>> cpu.attributes['machine'] = 'arm' >>> cpu.attributes['mhz'] = 800 >>> cpu.attributes['vendor'] = 'Texas Instruments' """ DEVICE_CPU = "device.cpu" DEVICE_MEM = "device.mem" DEVICE_USB = "device.usb" DEVICE_PCI = "device.pci" DEVICE_BOARD = "device.board" __slots__ = ('device_type', 'desc', 'attributes') def __init__(self, device_type, description, attributes=None): self.device_type = device_type self.description = description self.attributes = attributes or {}
""" Module with the HardwareDevice model. """ from launch_control.utils.json import PlainOldData class HardwareDevice(PlainOldData): """ Model representing any HardwareDevice A device is just a "device_type" attribute with a bag of properties and a human readable description. Individual device types can be freely added. For simplicity some common types of devices are provided as class properties DEVICE_xxx. Instances will come from a variety of factory classes, each capable of enumerating devices that it understands. The upside of having a common class like this is that it's easier to store it in the database _and_ not have to agree on a common set of properties for, say, all CPUs. If you want you can create instances manually, like this: >>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU, ... u"800MHz OMAP3 Processor") >>> cpu.attributes[u'machine'] = u'arm' >>> cpu.attributes[u'mhz'] = '800' >>> cpu.attributes[u'vendor'] = u'Texas Instruments' """ DEVICE_CPU = "device.cpu" DEVICE_MEM = "device.mem" DEVICE_USB = "device.usb" DEVICE_PCI = "device.pci" DEVICE_BOARD = "device.board" __slots__ = ('device_type', 'desc', 'attributes') def __init__(self, device_type, description, attributes=None): self.device_type = device_type self.description = description self.attributes = attributes or {}
Fix HardwareDevice docstring to match implementation
Fix HardwareDevice docstring to match implementation
Python
agpl-3.0
Linaro/lava-server,OSSystems/lava-server,OSSystems/lava-server,Linaro/lava-server,OSSystems/lava-server,Linaro/lava-server,Linaro/lava-server
""" Module with the HardwareDevice model. """ from launch_control.utils.json import PlainOldData class HardwareDevice(PlainOldData): """ Model representing any HardwareDevice A device is just a "device_type" attribute with a bag of properties and a human readable description. Individual device types can be freely added. For simplicity some common types of devices are provided as class properties DEVICE_xxx. Instances will come from a variety of factory classes, each capable of enumerating devices that it understands. The upside of having a common class like this is that it's easier to store it in the database _and_ not have to agree on a common set of properties for, say, all CPUs. If you want you can create instances manually, like this: >>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU) >>> cpu.desc = "800MHz OMAP3 Processor" >>> cpu.attributes['machine'] = 'arm' >>> cpu.attributes['mhz'] = 800 >>> cpu.attributes['vendor'] = 'Texas Instruments' """ DEVICE_CPU = "device.cpu" DEVICE_MEM = "device.mem" DEVICE_USB = "device.usb" DEVICE_PCI = "device.pci" DEVICE_BOARD = "device.board" __slots__ = ('device_type', 'desc', 'attributes') def __init__(self, device_type, description, attributes=None): self.device_type = device_type self.description = description self.attributes = attributes or {} Fix HardwareDevice docstring to match implementation
""" Module with the HardwareDevice model. """ from launch_control.utils.json import PlainOldData class HardwareDevice(PlainOldData): """ Model representing any HardwareDevice A device is just a "device_type" attribute with a bag of properties and a human readable description. Individual device types can be freely added. For simplicity some common types of devices are provided as class properties DEVICE_xxx. Instances will come from a variety of factory classes, each capable of enumerating devices that it understands. The upside of having a common class like this is that it's easier to store it in the database _and_ not have to agree on a common set of properties for, say, all CPUs. If you want you can create instances manually, like this: >>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU, ... u"800MHz OMAP3 Processor") >>> cpu.attributes[u'machine'] = u'arm' >>> cpu.attributes[u'mhz'] = '800' >>> cpu.attributes[u'vendor'] = u'Texas Instruments' """ DEVICE_CPU = "device.cpu" DEVICE_MEM = "device.mem" DEVICE_USB = "device.usb" DEVICE_PCI = "device.pci" DEVICE_BOARD = "device.board" __slots__ = ('device_type', 'desc', 'attributes') def __init__(self, device_type, description, attributes=None): self.device_type = device_type self.description = description self.attributes = attributes or {}
<commit_before>""" Module with the HardwareDevice model. """ from launch_control.utils.json import PlainOldData class HardwareDevice(PlainOldData): """ Model representing any HardwareDevice A device is just a "device_type" attribute with a bag of properties and a human readable description. Individual device types can be freely added. For simplicity some common types of devices are provided as class properties DEVICE_xxx. Instances will come from a variety of factory classes, each capable of enumerating devices that it understands. The upside of having a common class like this is that it's easier to store it in the database _and_ not have to agree on a common set of properties for, say, all CPUs. If you want you can create instances manually, like this: >>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU) >>> cpu.desc = "800MHz OMAP3 Processor" >>> cpu.attributes['machine'] = 'arm' >>> cpu.attributes['mhz'] = 800 >>> cpu.attributes['vendor'] = 'Texas Instruments' """ DEVICE_CPU = "device.cpu" DEVICE_MEM = "device.mem" DEVICE_USB = "device.usb" DEVICE_PCI = "device.pci" DEVICE_BOARD = "device.board" __slots__ = ('device_type', 'desc', 'attributes') def __init__(self, device_type, description, attributes=None): self.device_type = device_type self.description = description self.attributes = attributes or {} <commit_msg>Fix HardwareDevice docstring to match implementation<commit_after>
""" Module with the HardwareDevice model. """ from launch_control.utils.json import PlainOldData class HardwareDevice(PlainOldData): """ Model representing any HardwareDevice A device is just a "device_type" attribute with a bag of properties and a human readable description. Individual device types can be freely added. For simplicity some common types of devices are provided as class properties DEVICE_xxx. Instances will come from a variety of factory classes, each capable of enumerating devices that it understands. The upside of having a common class like this is that it's easier to store it in the database _and_ not have to agree on a common set of properties for, say, all CPUs. If you want you can create instances manually, like this: >>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU, ... u"800MHz OMAP3 Processor") >>> cpu.attributes[u'machine'] = u'arm' >>> cpu.attributes[u'mhz'] = '800' >>> cpu.attributes[u'vendor'] = u'Texas Instruments' """ DEVICE_CPU = "device.cpu" DEVICE_MEM = "device.mem" DEVICE_USB = "device.usb" DEVICE_PCI = "device.pci" DEVICE_BOARD = "device.board" __slots__ = ('device_type', 'desc', 'attributes') def __init__(self, device_type, description, attributes=None): self.device_type = device_type self.description = description self.attributes = attributes or {}
""" Module with the HardwareDevice model. """ from launch_control.utils.json import PlainOldData class HardwareDevice(PlainOldData): """ Model representing any HardwareDevice A device is just a "device_type" attribute with a bag of properties and a human readable description. Individual device types can be freely added. For simplicity some common types of devices are provided as class properties DEVICE_xxx. Instances will come from a variety of factory classes, each capable of enumerating devices that it understands. The upside of having a common class like this is that it's easier to store it in the database _and_ not have to agree on a common set of properties for, say, all CPUs. If you want you can create instances manually, like this: >>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU) >>> cpu.desc = "800MHz OMAP3 Processor" >>> cpu.attributes['machine'] = 'arm' >>> cpu.attributes['mhz'] = 800 >>> cpu.attributes['vendor'] = 'Texas Instruments' """ DEVICE_CPU = "device.cpu" DEVICE_MEM = "device.mem" DEVICE_USB = "device.usb" DEVICE_PCI = "device.pci" DEVICE_BOARD = "device.board" __slots__ = ('device_type', 'desc', 'attributes') def __init__(self, device_type, description, attributes=None): self.device_type = device_type self.description = description self.attributes = attributes or {} Fix HardwareDevice docstring to match implementation""" Module with the HardwareDevice model. """ from launch_control.utils.json import PlainOldData class HardwareDevice(PlainOldData): """ Model representing any HardwareDevice A device is just a "device_type" attribute with a bag of properties and a human readable description. Individual device types can be freely added. For simplicity some common types of devices are provided as class properties DEVICE_xxx. Instances will come from a variety of factory classes, each capable of enumerating devices that it understands. The upside of having a common class like this is that it's easier to store it in the database _and_ not have to agree on a common set of properties for, say, all CPUs. If you want you can create instances manually, like this: >>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU, ... u"800MHz OMAP3 Processor") >>> cpu.attributes[u'machine'] = u'arm' >>> cpu.attributes[u'mhz'] = '800' >>> cpu.attributes[u'vendor'] = u'Texas Instruments' """ DEVICE_CPU = "device.cpu" DEVICE_MEM = "device.mem" DEVICE_USB = "device.usb" DEVICE_PCI = "device.pci" DEVICE_BOARD = "device.board" __slots__ = ('device_type', 'desc', 'attributes') def __init__(self, device_type, description, attributes=None): self.device_type = device_type self.description = description self.attributes = attributes or {}
<commit_before>""" Module with the HardwareDevice model. """ from launch_control.utils.json import PlainOldData class HardwareDevice(PlainOldData): """ Model representing any HardwareDevice A device is just a "device_type" attribute with a bag of properties and a human readable description. Individual device types can be freely added. For simplicity some common types of devices are provided as class properties DEVICE_xxx. Instances will come from a variety of factory classes, each capable of enumerating devices that it understands. The upside of having a common class like this is that it's easier to store it in the database _and_ not have to agree on a common set of properties for, say, all CPUs. If you want you can create instances manually, like this: >>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU) >>> cpu.desc = "800MHz OMAP3 Processor" >>> cpu.attributes['machine'] = 'arm' >>> cpu.attributes['mhz'] = 800 >>> cpu.attributes['vendor'] = 'Texas Instruments' """ DEVICE_CPU = "device.cpu" DEVICE_MEM = "device.mem" DEVICE_USB = "device.usb" DEVICE_PCI = "device.pci" DEVICE_BOARD = "device.board" __slots__ = ('device_type', 'desc', 'attributes') def __init__(self, device_type, description, attributes=None): self.device_type = device_type self.description = description self.attributes = attributes or {} <commit_msg>Fix HardwareDevice docstring to match implementation<commit_after>""" Module with the HardwareDevice model. """ from launch_control.utils.json import PlainOldData class HardwareDevice(PlainOldData): """ Model representing any HardwareDevice A device is just a "device_type" attribute with a bag of properties and a human readable description. Individual device types can be freely added. For simplicity some common types of devices are provided as class properties DEVICE_xxx. Instances will come from a variety of factory classes, each capable of enumerating devices that it understands. The upside of having a common class like this is that it's easier to store it in the database _and_ not have to agree on a common set of properties for, say, all CPUs. If you want you can create instances manually, like this: >>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU, ... u"800MHz OMAP3 Processor") >>> cpu.attributes[u'machine'] = u'arm' >>> cpu.attributes[u'mhz'] = '800' >>> cpu.attributes[u'vendor'] = u'Texas Instruments' """ DEVICE_CPU = "device.cpu" DEVICE_MEM = "device.mem" DEVICE_USB = "device.usb" DEVICE_PCI = "device.pci" DEVICE_BOARD = "device.board" __slots__ = ('device_type', 'desc', 'attributes') def __init__(self, device_type, description, attributes=None): self.device_type = device_type self.description = description self.attributes = attributes or {}
b97fd14bba5d45a6e4e3caa02bd947bddfd0ba8b
tools/sniper_stats_jobid.py
tools/sniper_stats_jobid.py
import sniper_stats, intelqueue, iqclient class SniperStatsJobid(sniper_stats.SniperStatsBase): def __init__(self, jobid): self.jobid = jobid self.ic = iqclient.IntelClient() def read_metricnames(self): return self.ic.graphite_dbresults(self.jobid, 'read_metricnames') def get_snapshots(self): return self.ic.graphite_dbresults(self.jobid, 'get_snapshots') def read_snapshot(self, prefix, metrics = None): return self.ic.graphite_dbresults(self.jobid, 'read_snapshot', {'prefix': prefix, 'metrics': metrics}) def get_topology(self): return self.ic.graphite_dbresults(self.jobid, 'get_topology') def get_markers(self): return self.ic.graphite_dbresults(self.jobid, 'get_markers') def get_events(self): return self.ic.graphite_dbresults(self.jobid, 'get_events')
import sniper_stats, intelqueue, iqclient class SniperStatsJobid(sniper_stats.SniperStatsBase): def __init__(self, jobid): self.jobid = jobid self.ic = iqclient.IntelClient() self.names = self.read_metricnames() def read_metricnames(self): return self.ic.graphite_dbresults(self.jobid, 'read_metricnames') def get_snapshots(self): return self.ic.graphite_dbresults(self.jobid, 'get_snapshots') def read_snapshot(self, prefix, metrics = None): return self.ic.graphite_dbresults(self.jobid, 'read_snapshot', {'prefix': prefix, 'metrics': metrics}) def get_topology(self): return self.ic.graphite_dbresults(self.jobid, 'get_topology') def get_markers(self): return self.ic.graphite_dbresults(self.jobid, 'get_markers') def get_events(self): return self.ic.graphite_dbresults(self.jobid, 'get_events')
Read metric names on startup for jobid-based stats so self.names is available as expected
[sniper_stats] Read metric names on startup for jobid-based stats so self.names is available as expected
Python
mit
abanaiyan/sniper,abanaiyan/sniper,abanaiyan/sniper,abanaiyan/sniper,abanaiyan/sniper
import sniper_stats, intelqueue, iqclient class SniperStatsJobid(sniper_stats.SniperStatsBase): def __init__(self, jobid): self.jobid = jobid self.ic = iqclient.IntelClient() def read_metricnames(self): return self.ic.graphite_dbresults(self.jobid, 'read_metricnames') def get_snapshots(self): return self.ic.graphite_dbresults(self.jobid, 'get_snapshots') def read_snapshot(self, prefix, metrics = None): return self.ic.graphite_dbresults(self.jobid, 'read_snapshot', {'prefix': prefix, 'metrics': metrics}) def get_topology(self): return self.ic.graphite_dbresults(self.jobid, 'get_topology') def get_markers(self): return self.ic.graphite_dbresults(self.jobid, 'get_markers') def get_events(self): return self.ic.graphite_dbresults(self.jobid, 'get_events') [sniper_stats] Read metric names on startup for jobid-based stats so self.names is available as expected
import sniper_stats, intelqueue, iqclient class SniperStatsJobid(sniper_stats.SniperStatsBase): def __init__(self, jobid): self.jobid = jobid self.ic = iqclient.IntelClient() self.names = self.read_metricnames() def read_metricnames(self): return self.ic.graphite_dbresults(self.jobid, 'read_metricnames') def get_snapshots(self): return self.ic.graphite_dbresults(self.jobid, 'get_snapshots') def read_snapshot(self, prefix, metrics = None): return self.ic.graphite_dbresults(self.jobid, 'read_snapshot', {'prefix': prefix, 'metrics': metrics}) def get_topology(self): return self.ic.graphite_dbresults(self.jobid, 'get_topology') def get_markers(self): return self.ic.graphite_dbresults(self.jobid, 'get_markers') def get_events(self): return self.ic.graphite_dbresults(self.jobid, 'get_events')
<commit_before>import sniper_stats, intelqueue, iqclient class SniperStatsJobid(sniper_stats.SniperStatsBase): def __init__(self, jobid): self.jobid = jobid self.ic = iqclient.IntelClient() def read_metricnames(self): return self.ic.graphite_dbresults(self.jobid, 'read_metricnames') def get_snapshots(self): return self.ic.graphite_dbresults(self.jobid, 'get_snapshots') def read_snapshot(self, prefix, metrics = None): return self.ic.graphite_dbresults(self.jobid, 'read_snapshot', {'prefix': prefix, 'metrics': metrics}) def get_topology(self): return self.ic.graphite_dbresults(self.jobid, 'get_topology') def get_markers(self): return self.ic.graphite_dbresults(self.jobid, 'get_markers') def get_events(self): return self.ic.graphite_dbresults(self.jobid, 'get_events') <commit_msg>[sniper_stats] Read metric names on startup for jobid-based stats so self.names is available as expected<commit_after>
import sniper_stats, intelqueue, iqclient class SniperStatsJobid(sniper_stats.SniperStatsBase): def __init__(self, jobid): self.jobid = jobid self.ic = iqclient.IntelClient() self.names = self.read_metricnames() def read_metricnames(self): return self.ic.graphite_dbresults(self.jobid, 'read_metricnames') def get_snapshots(self): return self.ic.graphite_dbresults(self.jobid, 'get_snapshots') def read_snapshot(self, prefix, metrics = None): return self.ic.graphite_dbresults(self.jobid, 'read_snapshot', {'prefix': prefix, 'metrics': metrics}) def get_topology(self): return self.ic.graphite_dbresults(self.jobid, 'get_topology') def get_markers(self): return self.ic.graphite_dbresults(self.jobid, 'get_markers') def get_events(self): return self.ic.graphite_dbresults(self.jobid, 'get_events')
import sniper_stats, intelqueue, iqclient class SniperStatsJobid(sniper_stats.SniperStatsBase): def __init__(self, jobid): self.jobid = jobid self.ic = iqclient.IntelClient() def read_metricnames(self): return self.ic.graphite_dbresults(self.jobid, 'read_metricnames') def get_snapshots(self): return self.ic.graphite_dbresults(self.jobid, 'get_snapshots') def read_snapshot(self, prefix, metrics = None): return self.ic.graphite_dbresults(self.jobid, 'read_snapshot', {'prefix': prefix, 'metrics': metrics}) def get_topology(self): return self.ic.graphite_dbresults(self.jobid, 'get_topology') def get_markers(self): return self.ic.graphite_dbresults(self.jobid, 'get_markers') def get_events(self): return self.ic.graphite_dbresults(self.jobid, 'get_events') [sniper_stats] Read metric names on startup for jobid-based stats so self.names is available as expectedimport sniper_stats, intelqueue, iqclient class SniperStatsJobid(sniper_stats.SniperStatsBase): def __init__(self, jobid): self.jobid = jobid self.ic = iqclient.IntelClient() self.names = self.read_metricnames() def read_metricnames(self): return self.ic.graphite_dbresults(self.jobid, 'read_metricnames') def get_snapshots(self): return self.ic.graphite_dbresults(self.jobid, 'get_snapshots') def read_snapshot(self, prefix, metrics = None): return self.ic.graphite_dbresults(self.jobid, 'read_snapshot', {'prefix': prefix, 'metrics': metrics}) def get_topology(self): return self.ic.graphite_dbresults(self.jobid, 'get_topology') def get_markers(self): return self.ic.graphite_dbresults(self.jobid, 'get_markers') def get_events(self): return self.ic.graphite_dbresults(self.jobid, 'get_events')
<commit_before>import sniper_stats, intelqueue, iqclient class SniperStatsJobid(sniper_stats.SniperStatsBase): def __init__(self, jobid): self.jobid = jobid self.ic = iqclient.IntelClient() def read_metricnames(self): return self.ic.graphite_dbresults(self.jobid, 'read_metricnames') def get_snapshots(self): return self.ic.graphite_dbresults(self.jobid, 'get_snapshots') def read_snapshot(self, prefix, metrics = None): return self.ic.graphite_dbresults(self.jobid, 'read_snapshot', {'prefix': prefix, 'metrics': metrics}) def get_topology(self): return self.ic.graphite_dbresults(self.jobid, 'get_topology') def get_markers(self): return self.ic.graphite_dbresults(self.jobid, 'get_markers') def get_events(self): return self.ic.graphite_dbresults(self.jobid, 'get_events') <commit_msg>[sniper_stats] Read metric names on startup for jobid-based stats so self.names is available as expected<commit_after>import sniper_stats, intelqueue, iqclient class SniperStatsJobid(sniper_stats.SniperStatsBase): def __init__(self, jobid): self.jobid = jobid self.ic = iqclient.IntelClient() self.names = self.read_metricnames() def read_metricnames(self): return self.ic.graphite_dbresults(self.jobid, 'read_metricnames') def get_snapshots(self): return self.ic.graphite_dbresults(self.jobid, 'get_snapshots') def read_snapshot(self, prefix, metrics = None): return self.ic.graphite_dbresults(self.jobid, 'read_snapshot', {'prefix': prefix, 'metrics': metrics}) def get_topology(self): return self.ic.graphite_dbresults(self.jobid, 'get_topology') def get_markers(self): return self.ic.graphite_dbresults(self.jobid, 'get_markers') def get_events(self): return self.ic.graphite_dbresults(self.jobid, 'get_events')
2a17b9fdb55806d6397f506066a2a7e8c480020b
pylinks/main/tests.py
pylinks/main/tests.py
from django.test import TestCase from analytics.models import GoogleAnalytics class GoogleAnalyticsTests(TestCase): def test_ga_tracking(self): GoogleAnalytics(site_id=1, web_property_id='12345').save() response = self.client.get('/') self.assertContains(response, "_gaq.push(['_setAccount', '12345']);")
from django.test import TestCase from analytics.models import GoogleAnalytics class GoogleAnalyticsTests(TestCase): def test_ga_tracking(self): GoogleAnalytics(site_id=1, web_property_id='12345').save() response = self.client.get('/') self.assertContains(response, "_gaq.push(['_setAccount', '12345']);") class AdminTests(TestCase): def test_admin_login_loads(self): self.assertEqual(self.client.get('/admin/login/').status_code, 200)
Add simple admin test just so we catch breakage early
Add simple admin test just so we catch breakage early
Python
mit
michaelmior/pylinks,michaelmior/pylinks,michaelmior/pylinks
from django.test import TestCase from analytics.models import GoogleAnalytics class GoogleAnalyticsTests(TestCase): def test_ga_tracking(self): GoogleAnalytics(site_id=1, web_property_id='12345').save() response = self.client.get('/') self.assertContains(response, "_gaq.push(['_setAccount', '12345']);") Add simple admin test just so we catch breakage early
from django.test import TestCase from analytics.models import GoogleAnalytics class GoogleAnalyticsTests(TestCase): def test_ga_tracking(self): GoogleAnalytics(site_id=1, web_property_id='12345').save() response = self.client.get('/') self.assertContains(response, "_gaq.push(['_setAccount', '12345']);") class AdminTests(TestCase): def test_admin_login_loads(self): self.assertEqual(self.client.get('/admin/login/').status_code, 200)
<commit_before>from django.test import TestCase from analytics.models import GoogleAnalytics class GoogleAnalyticsTests(TestCase): def test_ga_tracking(self): GoogleAnalytics(site_id=1, web_property_id='12345').save() response = self.client.get('/') self.assertContains(response, "_gaq.push(['_setAccount', '12345']);") <commit_msg>Add simple admin test just so we catch breakage early<commit_after>
from django.test import TestCase from analytics.models import GoogleAnalytics class GoogleAnalyticsTests(TestCase): def test_ga_tracking(self): GoogleAnalytics(site_id=1, web_property_id='12345').save() response = self.client.get('/') self.assertContains(response, "_gaq.push(['_setAccount', '12345']);") class AdminTests(TestCase): def test_admin_login_loads(self): self.assertEqual(self.client.get('/admin/login/').status_code, 200)
from django.test import TestCase from analytics.models import GoogleAnalytics class GoogleAnalyticsTests(TestCase): def test_ga_tracking(self): GoogleAnalytics(site_id=1, web_property_id='12345').save() response = self.client.get('/') self.assertContains(response, "_gaq.push(['_setAccount', '12345']);") Add simple admin test just so we catch breakage earlyfrom django.test import TestCase from analytics.models import GoogleAnalytics class GoogleAnalyticsTests(TestCase): def test_ga_tracking(self): GoogleAnalytics(site_id=1, web_property_id='12345').save() response = self.client.get('/') self.assertContains(response, "_gaq.push(['_setAccount', '12345']);") class AdminTests(TestCase): def test_admin_login_loads(self): self.assertEqual(self.client.get('/admin/login/').status_code, 200)
<commit_before>from django.test import TestCase from analytics.models import GoogleAnalytics class GoogleAnalyticsTests(TestCase): def test_ga_tracking(self): GoogleAnalytics(site_id=1, web_property_id='12345').save() response = self.client.get('/') self.assertContains(response, "_gaq.push(['_setAccount', '12345']);") <commit_msg>Add simple admin test just so we catch breakage early<commit_after>from django.test import TestCase from analytics.models import GoogleAnalytics class GoogleAnalyticsTests(TestCase): def test_ga_tracking(self): GoogleAnalytics(site_id=1, web_property_id='12345').save() response = self.client.get('/') self.assertContains(response, "_gaq.push(['_setAccount', '12345']);") class AdminTests(TestCase): def test_admin_login_loads(self): self.assertEqual(self.client.get('/admin/login/').status_code, 200)
36e8b7f7dd4de93c61f49d65106f2a0410945e2d
pyoracc/model/line.py
pyoracc/model/line.py
from mako.template import Template class Line(object): template = Template("""${label}. \\ % for word in words: ${word} \\ % endfor % if lemmas: \n#lem: \\ % for lemma in lemmas: ${lemma}; \\ % endfor \n %endif """, output_encoding='utf-8') def __init__(self, label): self.label = label self.words = [] self.lemmas = [] self.witnesses = [] self.translation = None self.notes = [] self.references = [] self.links = [] def __str__(self): return self.template.render_unicode(**vars(self)) def serialize(self): return self.template.render_unicode(**vars(self))
from mako.template import Template class Line(object): template = Template("""\n${label}.\t\\ ${' '.join(words)}\\ % if references: % for reference in references: ^${reference}^ % endfor % endif % if lemmas: \n#lem:\\ ${'; '.join(lemmas)}\\ % endif % if notes: \n % for note in notes: ${note.serialize()} % endfor % endif % if links: \n#link: \\ % for link in links: ${link}; % endfor % endif """, output_encoding='utf-8') def __init__(self, label): self.label = label self.words = [] self.lemmas = [] self.witnesses = [] self.translation = None self.notes = [] self.references = [] self.links = [] def __str__(self): return self.template.render_unicode(**vars(self)) def serialize(self): return self.template.render_unicode(**vars(self))
Use join for serializing words and lemmas to avoid printing last ; as required by ATF format. Print also references, notes and links.
Use join for serializing words and lemmas to avoid printing last ; as required by ATF format. Print also references, notes and links.
Python
mit
UCL/pyoracc
from mako.template import Template class Line(object): template = Template("""${label}. \\ % for word in words: ${word} \\ % endfor % if lemmas: \n#lem: \\ % for lemma in lemmas: ${lemma}; \\ % endfor \n %endif """, output_encoding='utf-8') def __init__(self, label): self.label = label self.words = [] self.lemmas = [] self.witnesses = [] self.translation = None self.notes = [] self.references = [] self.links = [] def __str__(self): return self.template.render_unicode(**vars(self)) def serialize(self): return self.template.render_unicode(**vars(self)) Use join for serializing words and lemmas to avoid printing last ; as required by ATF format. Print also references, notes and links.
from mako.template import Template class Line(object): template = Template("""\n${label}.\t\\ ${' '.join(words)}\\ % if references: % for reference in references: ^${reference}^ % endfor % endif % if lemmas: \n#lem:\\ ${'; '.join(lemmas)}\\ % endif % if notes: \n % for note in notes: ${note.serialize()} % endfor % endif % if links: \n#link: \\ % for link in links: ${link}; % endfor % endif """, output_encoding='utf-8') def __init__(self, label): self.label = label self.words = [] self.lemmas = [] self.witnesses = [] self.translation = None self.notes = [] self.references = [] self.links = [] def __str__(self): return self.template.render_unicode(**vars(self)) def serialize(self): return self.template.render_unicode(**vars(self))
<commit_before>from mako.template import Template class Line(object): template = Template("""${label}. \\ % for word in words: ${word} \\ % endfor % if lemmas: \n#lem: \\ % for lemma in lemmas: ${lemma}; \\ % endfor \n %endif """, output_encoding='utf-8') def __init__(self, label): self.label = label self.words = [] self.lemmas = [] self.witnesses = [] self.translation = None self.notes = [] self.references = [] self.links = [] def __str__(self): return self.template.render_unicode(**vars(self)) def serialize(self): return self.template.render_unicode(**vars(self)) <commit_msg>Use join for serializing words and lemmas to avoid printing last ; as required by ATF format. Print also references, notes and links.<commit_after>
from mako.template import Template class Line(object): template = Template("""\n${label}.\t\\ ${' '.join(words)}\\ % if references: % for reference in references: ^${reference}^ % endfor % endif % if lemmas: \n#lem:\\ ${'; '.join(lemmas)}\\ % endif % if notes: \n % for note in notes: ${note.serialize()} % endfor % endif % if links: \n#link: \\ % for link in links: ${link}; % endfor % endif """, output_encoding='utf-8') def __init__(self, label): self.label = label self.words = [] self.lemmas = [] self.witnesses = [] self.translation = None self.notes = [] self.references = [] self.links = [] def __str__(self): return self.template.render_unicode(**vars(self)) def serialize(self): return self.template.render_unicode(**vars(self))
from mako.template import Template class Line(object): template = Template("""${label}. \\ % for word in words: ${word} \\ % endfor % if lemmas: \n#lem: \\ % for lemma in lemmas: ${lemma}; \\ % endfor \n %endif """, output_encoding='utf-8') def __init__(self, label): self.label = label self.words = [] self.lemmas = [] self.witnesses = [] self.translation = None self.notes = [] self.references = [] self.links = [] def __str__(self): return self.template.render_unicode(**vars(self)) def serialize(self): return self.template.render_unicode(**vars(self)) Use join for serializing words and lemmas to avoid printing last ; as required by ATF format. Print also references, notes and links.from mako.template import Template class Line(object): template = Template("""\n${label}.\t\\ ${' '.join(words)}\\ % if references: % for reference in references: ^${reference}^ % endfor % endif % if lemmas: \n#lem:\\ ${'; '.join(lemmas)}\\ % endif % if notes: \n % for note in notes: ${note.serialize()} % endfor % endif % if links: \n#link: \\ % for link in links: ${link}; % endfor % endif """, output_encoding='utf-8') def __init__(self, label): self.label = label self.words = [] self.lemmas = [] self.witnesses = [] self.translation = None self.notes = [] self.references = [] self.links = [] def __str__(self): return self.template.render_unicode(**vars(self)) def serialize(self): return self.template.render_unicode(**vars(self))
<commit_before>from mako.template import Template class Line(object): template = Template("""${label}. \\ % for word in words: ${word} \\ % endfor % if lemmas: \n#lem: \\ % for lemma in lemmas: ${lemma}; \\ % endfor \n %endif """, output_encoding='utf-8') def __init__(self, label): self.label = label self.words = [] self.lemmas = [] self.witnesses = [] self.translation = None self.notes = [] self.references = [] self.links = [] def __str__(self): return self.template.render_unicode(**vars(self)) def serialize(self): return self.template.render_unicode(**vars(self)) <commit_msg>Use join for serializing words and lemmas to avoid printing last ; as required by ATF format. Print also references, notes and links.<commit_after>from mako.template import Template class Line(object): template = Template("""\n${label}.\t\\ ${' '.join(words)}\\ % if references: % for reference in references: ^${reference}^ % endfor % endif % if lemmas: \n#lem:\\ ${'; '.join(lemmas)}\\ % endif % if notes: \n % for note in notes: ${note.serialize()} % endfor % endif % if links: \n#link: \\ % for link in links: ${link}; % endfor % endif """, output_encoding='utf-8') def __init__(self, label): self.label = label self.words = [] self.lemmas = [] self.witnesses = [] self.translation = None self.notes = [] self.references = [] self.links = [] def __str__(self): return self.template.render_unicode(**vars(self)) def serialize(self): return self.template.render_unicode(**vars(self))
87b1d823f09a20547b08f769636bfc7bcc7f0766
setup.py
setup.py
from setuptools import setup setup( name='rhino', version='1.0', packages=['rhino'], include_package_data=True, #requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup(<3.1)', 'jinja2'], #install_requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup<3.1', 'jinja2'], requires=['Django', 'south', 'jinja2', 'oauth2'], )
from setuptools import setup setup( name='rhino', version='1.0', packages=['rhino'], include_package_data=True, #requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup(<3.1)', 'jinja2'], #install_requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup<3.1', 'jinja2'], requires=['Django', 'south', 'jinja2', 'oauth2', 'pytidylib', 'remoteobjects'], )
Call out these requirements too
Call out these requirements too
Python
mit
markpasc/leapfrog,markpasc/leapfrog
from setuptools import setup setup( name='rhino', version='1.0', packages=['rhino'], include_package_data=True, #requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup(<3.1)', 'jinja2'], #install_requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup<3.1', 'jinja2'], requires=['Django', 'south', 'jinja2', 'oauth2'], ) Call out these requirements too
from setuptools import setup setup( name='rhino', version='1.0', packages=['rhino'], include_package_data=True, #requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup(<3.1)', 'jinja2'], #install_requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup<3.1', 'jinja2'], requires=['Django', 'south', 'jinja2', 'oauth2', 'pytidylib', 'remoteobjects'], )
<commit_before>from setuptools import setup setup( name='rhino', version='1.0', packages=['rhino'], include_package_data=True, #requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup(<3.1)', 'jinja2'], #install_requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup<3.1', 'jinja2'], requires=['Django', 'south', 'jinja2', 'oauth2'], ) <commit_msg>Call out these requirements too<commit_after>
from setuptools import setup setup( name='rhino', version='1.0', packages=['rhino'], include_package_data=True, #requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup(<3.1)', 'jinja2'], #install_requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup<3.1', 'jinja2'], requires=['Django', 'south', 'jinja2', 'oauth2', 'pytidylib', 'remoteobjects'], )
from setuptools import setup setup( name='rhino', version='1.0', packages=['rhino'], include_package_data=True, #requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup(<3.1)', 'jinja2'], #install_requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup<3.1', 'jinja2'], requires=['Django', 'south', 'jinja2', 'oauth2'], ) Call out these requirements toofrom setuptools import setup setup( name='rhino', version='1.0', packages=['rhino'], include_package_data=True, #requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup(<3.1)', 'jinja2'], #install_requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup<3.1', 'jinja2'], requires=['Django', 'south', 'jinja2', 'oauth2', 'pytidylib', 'remoteobjects'], )
<commit_before>from setuptools import setup setup( name='rhino', version='1.0', packages=['rhino'], include_package_data=True, #requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup(<3.1)', 'jinja2'], #install_requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup<3.1', 'jinja2'], requires=['Django', 'south', 'jinja2', 'oauth2'], ) <commit_msg>Call out these requirements too<commit_after>from setuptools import setup setup( name='rhino', version='1.0', packages=['rhino'], include_package_data=True, #requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup(<3.1)', 'jinja2'], #install_requires=['Django', 'south', 'celery', 'django_celery', 'httplib2', 'passogva', 'BeautifulSoup<3.1', 'jinja2'], requires=['Django', 'south', 'jinja2', 'oauth2', 'pytidylib', 'remoteobjects'], )
58701c0d750714f8ded53627b0f8c22f256376c6
setup.py
setup.py
#!/usr/bin/env python # Copyright (c) 2013 Soren Hansen # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. from setuptools import setup, find_packages setup( name='basicdb', version='0.1', description='Basic database service', long_description=open('README.rst', 'r').read(), author='Soren Hansen', author_email='soren@linux2go.dk', url='http://github.com/sorenh/basicdb', packages=find_packages(), include_package_data=True, license='Apache 2.0', keywords='basicdb simpledb')
#!/usr/bin/env python # Copyright (c) 2013 Soren Hansen # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. from setuptools import setup, find_packages import pkg_resources setup( name='basicdb', version='0.1', description='Basic database service', long_description=pkg_resources.resource_string(__name__, "README.rst"), author='Soren Hansen', author_email='soren@linux2go.dk', url='http://github.com/sorenh/basicdb', packages=find_packages(), include_package_data=True, license='Apache 2.0', keywords='basicdb simpledb')
Use pkg_resources to read README.rst
Use pkg_resources to read README.rst
Python
apache-2.0
JioCloud/basicdb,varunarya10/basicdb,sorenh/basicdb
#!/usr/bin/env python # Copyright (c) 2013 Soren Hansen # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. from setuptools import setup, find_packages setup( name='basicdb', version='0.1', description='Basic database service', long_description=open('README.rst', 'r').read(), author='Soren Hansen', author_email='soren@linux2go.dk', url='http://github.com/sorenh/basicdb', packages=find_packages(), include_package_data=True, license='Apache 2.0', keywords='basicdb simpledb') Use pkg_resources to read README.rst
#!/usr/bin/env python # Copyright (c) 2013 Soren Hansen # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. from setuptools import setup, find_packages import pkg_resources setup( name='basicdb', version='0.1', description='Basic database service', long_description=pkg_resources.resource_string(__name__, "README.rst"), author='Soren Hansen', author_email='soren@linux2go.dk', url='http://github.com/sorenh/basicdb', packages=find_packages(), include_package_data=True, license='Apache 2.0', keywords='basicdb simpledb')
<commit_before>#!/usr/bin/env python # Copyright (c) 2013 Soren Hansen # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. from setuptools import setup, find_packages setup( name='basicdb', version='0.1', description='Basic database service', long_description=open('README.rst', 'r').read(), author='Soren Hansen', author_email='soren@linux2go.dk', url='http://github.com/sorenh/basicdb', packages=find_packages(), include_package_data=True, license='Apache 2.0', keywords='basicdb simpledb') <commit_msg>Use pkg_resources to read README.rst<commit_after>
#!/usr/bin/env python # Copyright (c) 2013 Soren Hansen # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. from setuptools import setup, find_packages import pkg_resources setup( name='basicdb', version='0.1', description='Basic database service', long_description=pkg_resources.resource_string(__name__, "README.rst"), author='Soren Hansen', author_email='soren@linux2go.dk', url='http://github.com/sorenh/basicdb', packages=find_packages(), include_package_data=True, license='Apache 2.0', keywords='basicdb simpledb')
#!/usr/bin/env python # Copyright (c) 2013 Soren Hansen # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. from setuptools import setup, find_packages setup( name='basicdb', version='0.1', description='Basic database service', long_description=open('README.rst', 'r').read(), author='Soren Hansen', author_email='soren@linux2go.dk', url='http://github.com/sorenh/basicdb', packages=find_packages(), include_package_data=True, license='Apache 2.0', keywords='basicdb simpledb') Use pkg_resources to read README.rst#!/usr/bin/env python # Copyright (c) 2013 Soren Hansen # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. from setuptools import setup, find_packages import pkg_resources setup( name='basicdb', version='0.1', description='Basic database service', long_description=pkg_resources.resource_string(__name__, "README.rst"), author='Soren Hansen', author_email='soren@linux2go.dk', url='http://github.com/sorenh/basicdb', packages=find_packages(), include_package_data=True, license='Apache 2.0', keywords='basicdb simpledb')
<commit_before>#!/usr/bin/env python # Copyright (c) 2013 Soren Hansen # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. from setuptools import setup, find_packages setup( name='basicdb', version='0.1', description='Basic database service', long_description=open('README.rst', 'r').read(), author='Soren Hansen', author_email='soren@linux2go.dk', url='http://github.com/sorenh/basicdb', packages=find_packages(), include_package_data=True, license='Apache 2.0', keywords='basicdb simpledb') <commit_msg>Use pkg_resources to read README.rst<commit_after>#!/usr/bin/env python # Copyright (c) 2013 Soren Hansen # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. from setuptools import setup, find_packages import pkg_resources setup( name='basicdb', version='0.1', description='Basic database service', long_description=pkg_resources.resource_string(__name__, "README.rst"), author='Soren Hansen', author_email='soren@linux2go.dk', url='http://github.com/sorenh/basicdb', packages=find_packages(), include_package_data=True, license='Apache 2.0', keywords='basicdb simpledb')
7d8e91ce410bf1add9a21777afc0517198c11ced
setup.py
setup.py
from setuptools import setup, find_packages setup( name = "OvhApi", version = "1.0", description='Python module to manage Ovh API', author='Guilhem Lettron', author_email='guilhem.lettron@optiflows.com', url='https://github.com/optiflows/OvhApi', packages = find_packages(), )
from setuptools import setup, find_packages setup( name = "ovhapi", version = "1.0", description='Python module to manage Ovh API', maintainer='Guilhem Lettron', maintainer_email='guilhem.lettron@optiflows.com', url='https://github.com/optiflows/OvhApi', packages = find_packages(), )
Change Name to ovhapi Set me as maintainer (not author)
Change Name to ovhapi Set me as maintainer (not author)
Python
bsd-3-clause
optiflows/ovhapi
from setuptools import setup, find_packages setup( name = "OvhApi", version = "1.0", description='Python module to manage Ovh API', author='Guilhem Lettron', author_email='guilhem.lettron@optiflows.com', url='https://github.com/optiflows/OvhApi', packages = find_packages(), ) Change Name to ovhapi Set me as maintainer (not author)
from setuptools import setup, find_packages setup( name = "ovhapi", version = "1.0", description='Python module to manage Ovh API', maintainer='Guilhem Lettron', maintainer_email='guilhem.lettron@optiflows.com', url='https://github.com/optiflows/OvhApi', packages = find_packages(), )
<commit_before>from setuptools import setup, find_packages setup( name = "OvhApi", version = "1.0", description='Python module to manage Ovh API', author='Guilhem Lettron', author_email='guilhem.lettron@optiflows.com', url='https://github.com/optiflows/OvhApi', packages = find_packages(), ) <commit_msg>Change Name to ovhapi Set me as maintainer (not author)<commit_after>
from setuptools import setup, find_packages setup( name = "ovhapi", version = "1.0", description='Python module to manage Ovh API', maintainer='Guilhem Lettron', maintainer_email='guilhem.lettron@optiflows.com', url='https://github.com/optiflows/OvhApi', packages = find_packages(), )
from setuptools import setup, find_packages setup( name = "OvhApi", version = "1.0", description='Python module to manage Ovh API', author='Guilhem Lettron', author_email='guilhem.lettron@optiflows.com', url='https://github.com/optiflows/OvhApi', packages = find_packages(), ) Change Name to ovhapi Set me as maintainer (not author)from setuptools import setup, find_packages setup( name = "ovhapi", version = "1.0", description='Python module to manage Ovh API', maintainer='Guilhem Lettron', maintainer_email='guilhem.lettron@optiflows.com', url='https://github.com/optiflows/OvhApi', packages = find_packages(), )
<commit_before>from setuptools import setup, find_packages setup( name = "OvhApi", version = "1.0", description='Python module to manage Ovh API', author='Guilhem Lettron', author_email='guilhem.lettron@optiflows.com', url='https://github.com/optiflows/OvhApi', packages = find_packages(), ) <commit_msg>Change Name to ovhapi Set me as maintainer (not author)<commit_after>from setuptools import setup, find_packages setup( name = "ovhapi", version = "1.0", description='Python module to manage Ovh API', maintainer='Guilhem Lettron', maintainer_email='guilhem.lettron@optiflows.com', url='https://github.com/optiflows/OvhApi', packages = find_packages(), )
360a8c395373bcab0b725b1ac8f8dfd581d6e2b9
setup.py
setup.py
from setuptools import setup, find_packages setup( name='stop-words', version=__import__("stop_words").get_version(), description='Get list of common stop words in various languages in Python', long_description=open('README.rst').read(), license=open('LICENSE').read(), author='Alireza Savand', author_email='alireza.savand@gmail.com', url='https://github.com/Alir3z4/python-stop-words', packages=find_packages(), zip_safe=False, package_data={ 'stop_words': [ 'stop-words/*.txt', 'stop-words/languages.json', ] }, classifiers=[ 'Programming Language :: Python', 'Intended Audience :: Developers', 'Operating System :: OS Independent', 'Topic :: Software Development', 'Development Status :: 6 - Mature', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Topic :: Text Processing', 'Topic :: Text Processing :: Filters', 'License :: OSI Approved :: BSD License', ], )
from setuptools import setup, find_packages setup( name='stop-words', version=__import__("stop_words").get_version(), description='Get list of common stop words in various languages in Python', long_description=open('README.rst').read(), license=open('LICENSE').read(), author='Alireza Savand', author_email='alireza.savand@gmail.com', url='https://github.com/Alir3z4/python-stop-words', packages=find_packages(), zip_safe=False, package_data={ 'stop_words': [ 'stop-words/*.txt', 'stop-words/languages.json', ] }, classifiers=[ 'Programming Language :: Python', 'Intended Audience :: Developers', 'Operating System :: OS Independent', 'Topic :: Software Development', 'Development Status :: 6 - Mature', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Topic :: Text Processing', 'Topic :: Text Processing :: Filters', 'License :: OSI Approved :: BSD License', ], )
Add Python34 and Python35 to classifiers
Add Python34 and Python35 to classifiers
Python
bsd-3-clause
Alir3z4/python-stop-words
from setuptools import setup, find_packages setup( name='stop-words', version=__import__("stop_words").get_version(), description='Get list of common stop words in various languages in Python', long_description=open('README.rst').read(), license=open('LICENSE').read(), author='Alireza Savand', author_email='alireza.savand@gmail.com', url='https://github.com/Alir3z4/python-stop-words', packages=find_packages(), zip_safe=False, package_data={ 'stop_words': [ 'stop-words/*.txt', 'stop-words/languages.json', ] }, classifiers=[ 'Programming Language :: Python', 'Intended Audience :: Developers', 'Operating System :: OS Independent', 'Topic :: Software Development', 'Development Status :: 6 - Mature', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Topic :: Text Processing', 'Topic :: Text Processing :: Filters', 'License :: OSI Approved :: BSD License', ], ) Add Python34 and Python35 to classifiers
from setuptools import setup, find_packages setup( name='stop-words', version=__import__("stop_words").get_version(), description='Get list of common stop words in various languages in Python', long_description=open('README.rst').read(), license=open('LICENSE').read(), author='Alireza Savand', author_email='alireza.savand@gmail.com', url='https://github.com/Alir3z4/python-stop-words', packages=find_packages(), zip_safe=False, package_data={ 'stop_words': [ 'stop-words/*.txt', 'stop-words/languages.json', ] }, classifiers=[ 'Programming Language :: Python', 'Intended Audience :: Developers', 'Operating System :: OS Independent', 'Topic :: Software Development', 'Development Status :: 6 - Mature', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Topic :: Text Processing', 'Topic :: Text Processing :: Filters', 'License :: OSI Approved :: BSD License', ], )
<commit_before>from setuptools import setup, find_packages setup( name='stop-words', version=__import__("stop_words").get_version(), description='Get list of common stop words in various languages in Python', long_description=open('README.rst').read(), license=open('LICENSE').read(), author='Alireza Savand', author_email='alireza.savand@gmail.com', url='https://github.com/Alir3z4/python-stop-words', packages=find_packages(), zip_safe=False, package_data={ 'stop_words': [ 'stop-words/*.txt', 'stop-words/languages.json', ] }, classifiers=[ 'Programming Language :: Python', 'Intended Audience :: Developers', 'Operating System :: OS Independent', 'Topic :: Software Development', 'Development Status :: 6 - Mature', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Topic :: Text Processing', 'Topic :: Text Processing :: Filters', 'License :: OSI Approved :: BSD License', ], ) <commit_msg>Add Python34 and Python35 to classifiers<commit_after>
from setuptools import setup, find_packages setup( name='stop-words', version=__import__("stop_words").get_version(), description='Get list of common stop words in various languages in Python', long_description=open('README.rst').read(), license=open('LICENSE').read(), author='Alireza Savand', author_email='alireza.savand@gmail.com', url='https://github.com/Alir3z4/python-stop-words', packages=find_packages(), zip_safe=False, package_data={ 'stop_words': [ 'stop-words/*.txt', 'stop-words/languages.json', ] }, classifiers=[ 'Programming Language :: Python', 'Intended Audience :: Developers', 'Operating System :: OS Independent', 'Topic :: Software Development', 'Development Status :: 6 - Mature', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Topic :: Text Processing', 'Topic :: Text Processing :: Filters', 'License :: OSI Approved :: BSD License', ], )
from setuptools import setup, find_packages setup( name='stop-words', version=__import__("stop_words").get_version(), description='Get list of common stop words in various languages in Python', long_description=open('README.rst').read(), license=open('LICENSE').read(), author='Alireza Savand', author_email='alireza.savand@gmail.com', url='https://github.com/Alir3z4/python-stop-words', packages=find_packages(), zip_safe=False, package_data={ 'stop_words': [ 'stop-words/*.txt', 'stop-words/languages.json', ] }, classifiers=[ 'Programming Language :: Python', 'Intended Audience :: Developers', 'Operating System :: OS Independent', 'Topic :: Software Development', 'Development Status :: 6 - Mature', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Topic :: Text Processing', 'Topic :: Text Processing :: Filters', 'License :: OSI Approved :: BSD License', ], ) Add Python34 and Python35 to classifiersfrom setuptools import setup, find_packages setup( name='stop-words', version=__import__("stop_words").get_version(), description='Get list of common stop words in various languages in Python', long_description=open('README.rst').read(), license=open('LICENSE').read(), author='Alireza Savand', author_email='alireza.savand@gmail.com', url='https://github.com/Alir3z4/python-stop-words', packages=find_packages(), zip_safe=False, package_data={ 'stop_words': [ 'stop-words/*.txt', 'stop-words/languages.json', ] }, classifiers=[ 'Programming Language :: Python', 'Intended Audience :: Developers', 'Operating System :: OS Independent', 'Topic :: Software Development', 'Development Status :: 6 - Mature', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Topic :: Text Processing', 'Topic :: Text Processing :: Filters', 'License :: OSI Approved :: BSD License', ], )
<commit_before>from setuptools import setup, find_packages setup( name='stop-words', version=__import__("stop_words").get_version(), description='Get list of common stop words in various languages in Python', long_description=open('README.rst').read(), license=open('LICENSE').read(), author='Alireza Savand', author_email='alireza.savand@gmail.com', url='https://github.com/Alir3z4/python-stop-words', packages=find_packages(), zip_safe=False, package_data={ 'stop_words': [ 'stop-words/*.txt', 'stop-words/languages.json', ] }, classifiers=[ 'Programming Language :: Python', 'Intended Audience :: Developers', 'Operating System :: OS Independent', 'Topic :: Software Development', 'Development Status :: 6 - Mature', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Topic :: Text Processing', 'Topic :: Text Processing :: Filters', 'License :: OSI Approved :: BSD License', ], ) <commit_msg>Add Python34 and Python35 to classifiers<commit_after>from setuptools import setup, find_packages setup( name='stop-words', version=__import__("stop_words").get_version(), description='Get list of common stop words in various languages in Python', long_description=open('README.rst').read(), license=open('LICENSE').read(), author='Alireza Savand', author_email='alireza.savand@gmail.com', url='https://github.com/Alir3z4/python-stop-words', packages=find_packages(), zip_safe=False, package_data={ 'stop_words': [ 'stop-words/*.txt', 'stop-words/languages.json', ] }, classifiers=[ 'Programming Language :: Python', 'Intended Audience :: Developers', 'Operating System :: OS Independent', 'Topic :: Software Development', 'Development Status :: 6 - Mature', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Topic :: Text Processing', 'Topic :: Text Processing :: Filters', 'License :: OSI Approved :: BSD License', ], )
27434395a599d7e42d2295056396937d89bb53a6
setup.py
setup.py
import sys from setuptools import find_packages, setup VERSION = '2.0.dev0' install_requires = [ 'django-local-settings>=1.0a10', 'stashward', ] if sys.version_info[:2] < (3, 4): install_requires.append('enum34') setup( name='django-arcutils', version=VERSION, url='https://github.com/PSU-OIT-ARC/django-arcutils', author='PSU - OIT - ARC', author_email='consultants@pdx.edu', description='Common utilities used in ARC Django projects', packages=find_packages(), include_package_data=True, zip_safe=False, install_requires=install_requires, extras_require={ 'cas': [ 'django-cas-client>=1.2.0', ], 'ldap': [ 'certifi>=2015.11.20.1', 'ldap3>=1.0.3', ], 'dev': [ 'django>=1.7,<1.9', 'flake8', 'ldap3', 'mock', 'model_mommy', ], }, entry_points=""" [console_scripts] arcutils = arcutils.__main__:main """, classifiers=[ 'Development Status :: 3 - Alpha', 'Framework :: Django', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], )
import sys from setuptools import find_packages, setup VERSION = '2.0.dev0' install_requires = [ 'django-local-settings>=1.0a10', 'stashward', ] if sys.version_info[:2] < (3, 4): install_requires.append('enum34') setup( name='django-arcutils', version=VERSION, url='https://github.com/PSU-OIT-ARC/django-arcutils', author='PSU - OIT - ARC', author_email='consultants@pdx.edu', description='Common utilities used in ARC Django projects', packages=find_packages(), include_package_data=True, zip_safe=False, install_requires=install_requires, extras_require={ 'cas': [ 'django-cas-client>=1.2.0', ], 'ldap': [ 'certifi>=2015.11.20.1', 'ldap3>=1.0.3', ], 'dev': [ 'django>=1.7,<1.9', 'djangorestframework>3.3', 'flake8', 'ldap3', 'mock', 'model_mommy', ], }, entry_points=""" [console_scripts] arcutils = arcutils.__main__:main """, classifiers=[ 'Development Status :: 3 - Alpha', 'Framework :: Django', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], )
Add DRF as a dev dependency
Add DRF as a dev dependency
Python
mit
PSU-OIT-ARC/django-arcutils,wylee/django-arcutils,wylee/django-arcutils,PSU-OIT-ARC/django-arcutils
import sys from setuptools import find_packages, setup VERSION = '2.0.dev0' install_requires = [ 'django-local-settings>=1.0a10', 'stashward', ] if sys.version_info[:2] < (3, 4): install_requires.append('enum34') setup( name='django-arcutils', version=VERSION, url='https://github.com/PSU-OIT-ARC/django-arcutils', author='PSU - OIT - ARC', author_email='consultants@pdx.edu', description='Common utilities used in ARC Django projects', packages=find_packages(), include_package_data=True, zip_safe=False, install_requires=install_requires, extras_require={ 'cas': [ 'django-cas-client>=1.2.0', ], 'ldap': [ 'certifi>=2015.11.20.1', 'ldap3>=1.0.3', ], 'dev': [ 'django>=1.7,<1.9', 'flake8', 'ldap3', 'mock', 'model_mommy', ], }, entry_points=""" [console_scripts] arcutils = arcutils.__main__:main """, classifiers=[ 'Development Status :: 3 - Alpha', 'Framework :: Django', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], ) Add DRF as a dev dependency
import sys from setuptools import find_packages, setup VERSION = '2.0.dev0' install_requires = [ 'django-local-settings>=1.0a10', 'stashward', ] if sys.version_info[:2] < (3, 4): install_requires.append('enum34') setup( name='django-arcutils', version=VERSION, url='https://github.com/PSU-OIT-ARC/django-arcutils', author='PSU - OIT - ARC', author_email='consultants@pdx.edu', description='Common utilities used in ARC Django projects', packages=find_packages(), include_package_data=True, zip_safe=False, install_requires=install_requires, extras_require={ 'cas': [ 'django-cas-client>=1.2.0', ], 'ldap': [ 'certifi>=2015.11.20.1', 'ldap3>=1.0.3', ], 'dev': [ 'django>=1.7,<1.9', 'djangorestframework>3.3', 'flake8', 'ldap3', 'mock', 'model_mommy', ], }, entry_points=""" [console_scripts] arcutils = arcutils.__main__:main """, classifiers=[ 'Development Status :: 3 - Alpha', 'Framework :: Django', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], )
<commit_before>import sys from setuptools import find_packages, setup VERSION = '2.0.dev0' install_requires = [ 'django-local-settings>=1.0a10', 'stashward', ] if sys.version_info[:2] < (3, 4): install_requires.append('enum34') setup( name='django-arcutils', version=VERSION, url='https://github.com/PSU-OIT-ARC/django-arcutils', author='PSU - OIT - ARC', author_email='consultants@pdx.edu', description='Common utilities used in ARC Django projects', packages=find_packages(), include_package_data=True, zip_safe=False, install_requires=install_requires, extras_require={ 'cas': [ 'django-cas-client>=1.2.0', ], 'ldap': [ 'certifi>=2015.11.20.1', 'ldap3>=1.0.3', ], 'dev': [ 'django>=1.7,<1.9', 'flake8', 'ldap3', 'mock', 'model_mommy', ], }, entry_points=""" [console_scripts] arcutils = arcutils.__main__:main """, classifiers=[ 'Development Status :: 3 - Alpha', 'Framework :: Django', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], ) <commit_msg>Add DRF as a dev dependency<commit_after>
import sys from setuptools import find_packages, setup VERSION = '2.0.dev0' install_requires = [ 'django-local-settings>=1.0a10', 'stashward', ] if sys.version_info[:2] < (3, 4): install_requires.append('enum34') setup( name='django-arcutils', version=VERSION, url='https://github.com/PSU-OIT-ARC/django-arcutils', author='PSU - OIT - ARC', author_email='consultants@pdx.edu', description='Common utilities used in ARC Django projects', packages=find_packages(), include_package_data=True, zip_safe=False, install_requires=install_requires, extras_require={ 'cas': [ 'django-cas-client>=1.2.0', ], 'ldap': [ 'certifi>=2015.11.20.1', 'ldap3>=1.0.3', ], 'dev': [ 'django>=1.7,<1.9', 'djangorestframework>3.3', 'flake8', 'ldap3', 'mock', 'model_mommy', ], }, entry_points=""" [console_scripts] arcutils = arcutils.__main__:main """, classifiers=[ 'Development Status :: 3 - Alpha', 'Framework :: Django', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], )
import sys from setuptools import find_packages, setup VERSION = '2.0.dev0' install_requires = [ 'django-local-settings>=1.0a10', 'stashward', ] if sys.version_info[:2] < (3, 4): install_requires.append('enum34') setup( name='django-arcutils', version=VERSION, url='https://github.com/PSU-OIT-ARC/django-arcutils', author='PSU - OIT - ARC', author_email='consultants@pdx.edu', description='Common utilities used in ARC Django projects', packages=find_packages(), include_package_data=True, zip_safe=False, install_requires=install_requires, extras_require={ 'cas': [ 'django-cas-client>=1.2.0', ], 'ldap': [ 'certifi>=2015.11.20.1', 'ldap3>=1.0.3', ], 'dev': [ 'django>=1.7,<1.9', 'flake8', 'ldap3', 'mock', 'model_mommy', ], }, entry_points=""" [console_scripts] arcutils = arcutils.__main__:main """, classifiers=[ 'Development Status :: 3 - Alpha', 'Framework :: Django', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], ) Add DRF as a dev dependencyimport sys from setuptools import find_packages, setup VERSION = '2.0.dev0' install_requires = [ 'django-local-settings>=1.0a10', 'stashward', ] if sys.version_info[:2] < (3, 4): install_requires.append('enum34') setup( name='django-arcutils', version=VERSION, url='https://github.com/PSU-OIT-ARC/django-arcutils', author='PSU - OIT - ARC', author_email='consultants@pdx.edu', description='Common utilities used in ARC Django projects', packages=find_packages(), include_package_data=True, zip_safe=False, install_requires=install_requires, extras_require={ 'cas': [ 'django-cas-client>=1.2.0', ], 'ldap': [ 'certifi>=2015.11.20.1', 'ldap3>=1.0.3', ], 'dev': [ 'django>=1.7,<1.9', 'djangorestframework>3.3', 'flake8', 'ldap3', 'mock', 'model_mommy', ], }, entry_points=""" [console_scripts] arcutils = arcutils.__main__:main """, classifiers=[ 'Development Status :: 3 - Alpha', 'Framework :: Django', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], )
<commit_before>import sys from setuptools import find_packages, setup VERSION = '2.0.dev0' install_requires = [ 'django-local-settings>=1.0a10', 'stashward', ] if sys.version_info[:2] < (3, 4): install_requires.append('enum34') setup( name='django-arcutils', version=VERSION, url='https://github.com/PSU-OIT-ARC/django-arcutils', author='PSU - OIT - ARC', author_email='consultants@pdx.edu', description='Common utilities used in ARC Django projects', packages=find_packages(), include_package_data=True, zip_safe=False, install_requires=install_requires, extras_require={ 'cas': [ 'django-cas-client>=1.2.0', ], 'ldap': [ 'certifi>=2015.11.20.1', 'ldap3>=1.0.3', ], 'dev': [ 'django>=1.7,<1.9', 'flake8', 'ldap3', 'mock', 'model_mommy', ], }, entry_points=""" [console_scripts] arcutils = arcutils.__main__:main """, classifiers=[ 'Development Status :: 3 - Alpha', 'Framework :: Django', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], ) <commit_msg>Add DRF as a dev dependency<commit_after>import sys from setuptools import find_packages, setup VERSION = '2.0.dev0' install_requires = [ 'django-local-settings>=1.0a10', 'stashward', ] if sys.version_info[:2] < (3, 4): install_requires.append('enum34') setup( name='django-arcutils', version=VERSION, url='https://github.com/PSU-OIT-ARC/django-arcutils', author='PSU - OIT - ARC', author_email='consultants@pdx.edu', description='Common utilities used in ARC Django projects', packages=find_packages(), include_package_data=True, zip_safe=False, install_requires=install_requires, extras_require={ 'cas': [ 'django-cas-client>=1.2.0', ], 'ldap': [ 'certifi>=2015.11.20.1', 'ldap3>=1.0.3', ], 'dev': [ 'django>=1.7,<1.9', 'djangorestframework>3.3', 'flake8', 'ldap3', 'mock', 'model_mommy', ], }, entry_points=""" [console_scripts] arcutils = arcutils.__main__:main """, classifiers=[ 'Development Status :: 3 - Alpha', 'Framework :: Django', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', ], )
5392af2781d6a6a7c71c15ab7699feb3e3b8f2f2
setup.py
setup.py
__author__ = 'katharine' import os import sys from setuptools import setup, find_packages requirements_path = os.path.join(os.path.dirname(__file__), 'requirements.txt') with open(requirements_path) as requirements_file: requirements = [line.strip() for line in requirements_file.readlines()] setup(name='pypkjs', version='3.6', description='PebbleKit JS in Python!', url='https://github.com/pebble/pypkjs', author='Pebble Technology Corporation', author_email='katharine@pebble.com', license='MIT', packages=find_packages(), install_requires=requirements, entry_points={ 'console_scripts': [ 'pypkjs=runner.websocket:run_tool' ], }, zip_safe=False)
__author__ = 'katharine' import os import sys from setuptools import setup, find_packages requirements_path = os.path.join(os.path.dirname(__file__), 'requirements.txt') with open(requirements_path) as requirements_file: requirements = [line.strip() for line in requirements_file.readlines()] setup(name='pypkjs', version='3.6', description='PebbleKit JS in Python!', url='https://github.com/pebble/pypkjs', author='Pebble Technology Corporation', author_email='katharine@pebble.com', license='MIT', packages=find_packages(), install_requires=requirements, package_data={ 'javascript.navigator': 'GeoLiteCity.dat' }, entry_points={ 'console_scripts': [ 'pypkjs=runner.websocket:run_tool' ], }, zip_safe=False)
Add package data for GeoIP DB
Add package data for GeoIP DB
Python
mit
pebble/pypkjs
__author__ = 'katharine' import os import sys from setuptools import setup, find_packages requirements_path = os.path.join(os.path.dirname(__file__), 'requirements.txt') with open(requirements_path) as requirements_file: requirements = [line.strip() for line in requirements_file.readlines()] setup(name='pypkjs', version='3.6', description='PebbleKit JS in Python!', url='https://github.com/pebble/pypkjs', author='Pebble Technology Corporation', author_email='katharine@pebble.com', license='MIT', packages=find_packages(), install_requires=requirements, entry_points={ 'console_scripts': [ 'pypkjs=runner.websocket:run_tool' ], }, zip_safe=False) Add package data for GeoIP DB
__author__ = 'katharine' import os import sys from setuptools import setup, find_packages requirements_path = os.path.join(os.path.dirname(__file__), 'requirements.txt') with open(requirements_path) as requirements_file: requirements = [line.strip() for line in requirements_file.readlines()] setup(name='pypkjs', version='3.6', description='PebbleKit JS in Python!', url='https://github.com/pebble/pypkjs', author='Pebble Technology Corporation', author_email='katharine@pebble.com', license='MIT', packages=find_packages(), install_requires=requirements, package_data={ 'javascript.navigator': 'GeoLiteCity.dat' }, entry_points={ 'console_scripts': [ 'pypkjs=runner.websocket:run_tool' ], }, zip_safe=False)
<commit_before>__author__ = 'katharine' import os import sys from setuptools import setup, find_packages requirements_path = os.path.join(os.path.dirname(__file__), 'requirements.txt') with open(requirements_path) as requirements_file: requirements = [line.strip() for line in requirements_file.readlines()] setup(name='pypkjs', version='3.6', description='PebbleKit JS in Python!', url='https://github.com/pebble/pypkjs', author='Pebble Technology Corporation', author_email='katharine@pebble.com', license='MIT', packages=find_packages(), install_requires=requirements, entry_points={ 'console_scripts': [ 'pypkjs=runner.websocket:run_tool' ], }, zip_safe=False) <commit_msg>Add package data for GeoIP DB<commit_after>
__author__ = 'katharine' import os import sys from setuptools import setup, find_packages requirements_path = os.path.join(os.path.dirname(__file__), 'requirements.txt') with open(requirements_path) as requirements_file: requirements = [line.strip() for line in requirements_file.readlines()] setup(name='pypkjs', version='3.6', description='PebbleKit JS in Python!', url='https://github.com/pebble/pypkjs', author='Pebble Technology Corporation', author_email='katharine@pebble.com', license='MIT', packages=find_packages(), install_requires=requirements, package_data={ 'javascript.navigator': 'GeoLiteCity.dat' }, entry_points={ 'console_scripts': [ 'pypkjs=runner.websocket:run_tool' ], }, zip_safe=False)
__author__ = 'katharine' import os import sys from setuptools import setup, find_packages requirements_path = os.path.join(os.path.dirname(__file__), 'requirements.txt') with open(requirements_path) as requirements_file: requirements = [line.strip() for line in requirements_file.readlines()] setup(name='pypkjs', version='3.6', description='PebbleKit JS in Python!', url='https://github.com/pebble/pypkjs', author='Pebble Technology Corporation', author_email='katharine@pebble.com', license='MIT', packages=find_packages(), install_requires=requirements, entry_points={ 'console_scripts': [ 'pypkjs=runner.websocket:run_tool' ], }, zip_safe=False) Add package data for GeoIP DB__author__ = 'katharine' import os import sys from setuptools import setup, find_packages requirements_path = os.path.join(os.path.dirname(__file__), 'requirements.txt') with open(requirements_path) as requirements_file: requirements = [line.strip() for line in requirements_file.readlines()] setup(name='pypkjs', version='3.6', description='PebbleKit JS in Python!', url='https://github.com/pebble/pypkjs', author='Pebble Technology Corporation', author_email='katharine@pebble.com', license='MIT', packages=find_packages(), install_requires=requirements, package_data={ 'javascript.navigator': 'GeoLiteCity.dat' }, entry_points={ 'console_scripts': [ 'pypkjs=runner.websocket:run_tool' ], }, zip_safe=False)
<commit_before>__author__ = 'katharine' import os import sys from setuptools import setup, find_packages requirements_path = os.path.join(os.path.dirname(__file__), 'requirements.txt') with open(requirements_path) as requirements_file: requirements = [line.strip() for line in requirements_file.readlines()] setup(name='pypkjs', version='3.6', description='PebbleKit JS in Python!', url='https://github.com/pebble/pypkjs', author='Pebble Technology Corporation', author_email='katharine@pebble.com', license='MIT', packages=find_packages(), install_requires=requirements, entry_points={ 'console_scripts': [ 'pypkjs=runner.websocket:run_tool' ], }, zip_safe=False) <commit_msg>Add package data for GeoIP DB<commit_after>__author__ = 'katharine' import os import sys from setuptools import setup, find_packages requirements_path = os.path.join(os.path.dirname(__file__), 'requirements.txt') with open(requirements_path) as requirements_file: requirements = [line.strip() for line in requirements_file.readlines()] setup(name='pypkjs', version='3.6', description='PebbleKit JS in Python!', url='https://github.com/pebble/pypkjs', author='Pebble Technology Corporation', author_email='katharine@pebble.com', license='MIT', packages=find_packages(), install_requires=requirements, package_data={ 'javascript.navigator': 'GeoLiteCity.dat' }, entry_points={ 'console_scripts': [ 'pypkjs=runner.websocket:run_tool' ], }, zip_safe=False)
e8770775250371766f47317a8aa40e034a5d75de
setup.py
setup.py
#!/usr/bin/env python from setuptools import setup, find_packages requires = [ ] setup(name='mirrors', version='0.1.1', description='rsync mirror manager', author='Ethan House', author_email='ehouse@csh.rit.edu', packages=find_packages(), install_requires=requires, zip_safe=False, entry_points=""" [console_scripts] mirrors=mirrors:main """)
#!/usr/bin/env python from setuptools import setup, find_packages requires = [ ] setup(name='mirrors', version='0.2.0', description='rsync mirror manager', author='Ethan House', author_email='ehouse@csh.rit.edu', packages=find_packages(), install_requires=requires, zip_safe=False, entry_points=""" [console_scripts] mirrors=mirrors:main """)
Move release version up to 0.2.0
Move release version up to 0.2.0
Python
bsd-2-clause
ehouse/mirrors
#!/usr/bin/env python from setuptools import setup, find_packages requires = [ ] setup(name='mirrors', version='0.1.1', description='rsync mirror manager', author='Ethan House', author_email='ehouse@csh.rit.edu', packages=find_packages(), install_requires=requires, zip_safe=False, entry_points=""" [console_scripts] mirrors=mirrors:main """) Move release version up to 0.2.0
#!/usr/bin/env python from setuptools import setup, find_packages requires = [ ] setup(name='mirrors', version='0.2.0', description='rsync mirror manager', author='Ethan House', author_email='ehouse@csh.rit.edu', packages=find_packages(), install_requires=requires, zip_safe=False, entry_points=""" [console_scripts] mirrors=mirrors:main """)
<commit_before>#!/usr/bin/env python from setuptools import setup, find_packages requires = [ ] setup(name='mirrors', version='0.1.1', description='rsync mirror manager', author='Ethan House', author_email='ehouse@csh.rit.edu', packages=find_packages(), install_requires=requires, zip_safe=False, entry_points=""" [console_scripts] mirrors=mirrors:main """) <commit_msg>Move release version up to 0.2.0<commit_after>
#!/usr/bin/env python from setuptools import setup, find_packages requires = [ ] setup(name='mirrors', version='0.2.0', description='rsync mirror manager', author='Ethan House', author_email='ehouse@csh.rit.edu', packages=find_packages(), install_requires=requires, zip_safe=False, entry_points=""" [console_scripts] mirrors=mirrors:main """)
#!/usr/bin/env python from setuptools import setup, find_packages requires = [ ] setup(name='mirrors', version='0.1.1', description='rsync mirror manager', author='Ethan House', author_email='ehouse@csh.rit.edu', packages=find_packages(), install_requires=requires, zip_safe=False, entry_points=""" [console_scripts] mirrors=mirrors:main """) Move release version up to 0.2.0#!/usr/bin/env python from setuptools import setup, find_packages requires = [ ] setup(name='mirrors', version='0.2.0', description='rsync mirror manager', author='Ethan House', author_email='ehouse@csh.rit.edu', packages=find_packages(), install_requires=requires, zip_safe=False, entry_points=""" [console_scripts] mirrors=mirrors:main """)
<commit_before>#!/usr/bin/env python from setuptools import setup, find_packages requires = [ ] setup(name='mirrors', version='0.1.1', description='rsync mirror manager', author='Ethan House', author_email='ehouse@csh.rit.edu', packages=find_packages(), install_requires=requires, zip_safe=False, entry_points=""" [console_scripts] mirrors=mirrors:main """) <commit_msg>Move release version up to 0.2.0<commit_after>#!/usr/bin/env python from setuptools import setup, find_packages requires = [ ] setup(name='mirrors', version='0.2.0', description='rsync mirror manager', author='Ethan House', author_email='ehouse@csh.rit.edu', packages=find_packages(), install_requires=requires, zip_safe=False, entry_points=""" [console_scripts] mirrors=mirrors:main """)
a463ac8ae112dd19bfc1c8e2df170023114ded07
setup.py
setup.py
#!/usr/bin/env python from setuptools import setup, find_packages from setuputils import find_version, read setup( name='astor', version=find_version('astor/__init__.py'), description='Read/rewrite/write Python ASTs', long_description=read('README.rst'), author='Patrick Maupin', author_email='pmaupin@gmail.com', platforms='Independent', url='https://github.com/berkerpeksag/astor', packages=find_packages(exclude=['tests']), py_modules=['setuputils'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: Implementation', 'Programming Language :: Python :: Implementation :: CPython', 'Programming Language :: Python :: Implementation :: PyPy', 'Topic :: Software Development :: Code Generators', 'Topic :: Software Development :: Compilers', ], keywords='ast, codegen', )
#!/usr/bin/env python from setuptools import setup, find_packages from setuputils import find_version, read setup( name='astor', version=find_version('astor/__init__.py'), description='Read/rewrite/write Python ASTs', long_description=read('README.rst'), author='Patrick Maupin', author_email='pmaupin@gmail.com', platforms='Independent', url='https://github.com/berkerpeksag/astor', packages=find_packages(exclude=['tests']), py_modules=['setuputils'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: Implementation', 'Programming Language :: Python :: Implementation :: CPython', 'Programming Language :: Python :: Implementation :: PyPy', 'Topic :: Software Development :: Code Generators', 'Topic :: Software Development :: Compilers', ], keywords='ast, codegen', )
Remove 3.5 for now. It's not added to PyPI yet.
Remove 3.5 for now. It's not added to PyPI yet.
Python
bsd-3-clause
berkerpeksag/astor
#!/usr/bin/env python from setuptools import setup, find_packages from setuputils import find_version, read setup( name='astor', version=find_version('astor/__init__.py'), description='Read/rewrite/write Python ASTs', long_description=read('README.rst'), author='Patrick Maupin', author_email='pmaupin@gmail.com', platforms='Independent', url='https://github.com/berkerpeksag/astor', packages=find_packages(exclude=['tests']), py_modules=['setuputils'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: Implementation', 'Programming Language :: Python :: Implementation :: CPython', 'Programming Language :: Python :: Implementation :: PyPy', 'Topic :: Software Development :: Code Generators', 'Topic :: Software Development :: Compilers', ], keywords='ast, codegen', ) Remove 3.5 for now. It's not added to PyPI yet.
#!/usr/bin/env python from setuptools import setup, find_packages from setuputils import find_version, read setup( name='astor', version=find_version('astor/__init__.py'), description='Read/rewrite/write Python ASTs', long_description=read('README.rst'), author='Patrick Maupin', author_email='pmaupin@gmail.com', platforms='Independent', url='https://github.com/berkerpeksag/astor', packages=find_packages(exclude=['tests']), py_modules=['setuputils'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: Implementation', 'Programming Language :: Python :: Implementation :: CPython', 'Programming Language :: Python :: Implementation :: PyPy', 'Topic :: Software Development :: Code Generators', 'Topic :: Software Development :: Compilers', ], keywords='ast, codegen', )
<commit_before>#!/usr/bin/env python from setuptools import setup, find_packages from setuputils import find_version, read setup( name='astor', version=find_version('astor/__init__.py'), description='Read/rewrite/write Python ASTs', long_description=read('README.rst'), author='Patrick Maupin', author_email='pmaupin@gmail.com', platforms='Independent', url='https://github.com/berkerpeksag/astor', packages=find_packages(exclude=['tests']), py_modules=['setuputils'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: Implementation', 'Programming Language :: Python :: Implementation :: CPython', 'Programming Language :: Python :: Implementation :: PyPy', 'Topic :: Software Development :: Code Generators', 'Topic :: Software Development :: Compilers', ], keywords='ast, codegen', ) <commit_msg>Remove 3.5 for now. It's not added to PyPI yet.<commit_after>
#!/usr/bin/env python from setuptools import setup, find_packages from setuputils import find_version, read setup( name='astor', version=find_version('astor/__init__.py'), description='Read/rewrite/write Python ASTs', long_description=read('README.rst'), author='Patrick Maupin', author_email='pmaupin@gmail.com', platforms='Independent', url='https://github.com/berkerpeksag/astor', packages=find_packages(exclude=['tests']), py_modules=['setuputils'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: Implementation', 'Programming Language :: Python :: Implementation :: CPython', 'Programming Language :: Python :: Implementation :: PyPy', 'Topic :: Software Development :: Code Generators', 'Topic :: Software Development :: Compilers', ], keywords='ast, codegen', )
#!/usr/bin/env python from setuptools import setup, find_packages from setuputils import find_version, read setup( name='astor', version=find_version('astor/__init__.py'), description='Read/rewrite/write Python ASTs', long_description=read('README.rst'), author='Patrick Maupin', author_email='pmaupin@gmail.com', platforms='Independent', url='https://github.com/berkerpeksag/astor', packages=find_packages(exclude=['tests']), py_modules=['setuputils'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: Implementation', 'Programming Language :: Python :: Implementation :: CPython', 'Programming Language :: Python :: Implementation :: PyPy', 'Topic :: Software Development :: Code Generators', 'Topic :: Software Development :: Compilers', ], keywords='ast, codegen', ) Remove 3.5 for now. It's not added to PyPI yet.#!/usr/bin/env python from setuptools import setup, find_packages from setuputils import find_version, read setup( name='astor', version=find_version('astor/__init__.py'), description='Read/rewrite/write Python ASTs', long_description=read('README.rst'), author='Patrick Maupin', author_email='pmaupin@gmail.com', platforms='Independent', url='https://github.com/berkerpeksag/astor', packages=find_packages(exclude=['tests']), py_modules=['setuputils'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: Implementation', 'Programming Language :: Python :: Implementation :: CPython', 'Programming Language :: Python :: Implementation :: PyPy', 'Topic :: Software Development :: Code Generators', 'Topic :: Software Development :: Compilers', ], keywords='ast, codegen', )
<commit_before>#!/usr/bin/env python from setuptools import setup, find_packages from setuputils import find_version, read setup( name='astor', version=find_version('astor/__init__.py'), description='Read/rewrite/write Python ASTs', long_description=read('README.rst'), author='Patrick Maupin', author_email='pmaupin@gmail.com', platforms='Independent', url='https://github.com/berkerpeksag/astor', packages=find_packages(exclude=['tests']), py_modules=['setuputils'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: Implementation', 'Programming Language :: Python :: Implementation :: CPython', 'Programming Language :: Python :: Implementation :: PyPy', 'Topic :: Software Development :: Code Generators', 'Topic :: Software Development :: Compilers', ], keywords='ast, codegen', ) <commit_msg>Remove 3.5 for now. It's not added to PyPI yet.<commit_after>#!/usr/bin/env python from setuptools import setup, find_packages from setuputils import find_version, read setup( name='astor', version=find_version('astor/__init__.py'), description='Read/rewrite/write Python ASTs', long_description=read('README.rst'), author='Patrick Maupin', author_email='pmaupin@gmail.com', platforms='Independent', url='https://github.com/berkerpeksag/astor', packages=find_packages(exclude=['tests']), py_modules=['setuputils'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: Implementation', 'Programming Language :: Python :: Implementation :: CPython', 'Programming Language :: Python :: Implementation :: PyPy', 'Topic :: Software Development :: Code Generators', 'Topic :: Software Development :: Compilers', ], keywords='ast, codegen', )
9275bc3a9373e453e06f0aebc883a773dfb97627
setup.py
setup.py
from setuptools import setup, find_packages from tiddlywebwiki import __version__ as VERSION setup( name = 'tiddlywebwiki', version = VERSION, description = 'A TiddlyWeb plugin to provide a multi-user TiddlyWiki environment.', author = 'FND', author_email = 'FNDo@gmx.net', packages = find_packages(exclude=['test']), scripts = ['twinstance'], platforms = 'Posix; MacOS X; Windows', install_requires = [ 'tiddlyweb>=0.9.96', 'tiddlywebplugins.wikklytextrender', 'tiddlywebplugins.status>=0.5', 'tiddlywebplugins.differ', 'tiddlywebplugins.atom', 'tiddlywebplugins.twimport', 'tiddlywebplugins.utils', 'tiddlywebplugins.instancer>=0.5.5', 'BeautifulSoup', 'wikklytext'], include_package_data = True, zip_safe = False )
from setuptools import setup, find_packages from tiddlywebwiki import __version__ as VERSION setup( name = 'tiddlywebwiki', version = VERSION, description = 'A TiddlyWeb plugin to provide a multi-user TiddlyWiki environment.', author = 'FND', author_email = 'FNDo@gmx.net', packages = find_packages(exclude=['test']), scripts = ['twinstance'], platforms = 'Posix; MacOS X; Windows', install_requires = [ 'tiddlyweb>=0.9.96', 'tiddlywebplugins.wikklytextrender', 'tiddlywebplugins.status>=0.5', 'tiddlywebplugins.differ', 'tiddlywebplugins.atom', 'tiddlywebplugins.twimport', 'tiddlywebplugins.utils', 'tiddlywebplugins.instancer>=0.5.5', 'wikklytext'], include_package_data = True, zip_safe = False )
Remove BeautifulSoup from direct dependency list.
Remove BeautifulSoup from direct dependency list.
Python
bsd-3-clause
tiddlyweb/tiddlywebwiki,tiddlyweb/tiddlywebwiki,tiddlyweb/tiddlywebwiki
from setuptools import setup, find_packages from tiddlywebwiki import __version__ as VERSION setup( name = 'tiddlywebwiki', version = VERSION, description = 'A TiddlyWeb plugin to provide a multi-user TiddlyWiki environment.', author = 'FND', author_email = 'FNDo@gmx.net', packages = find_packages(exclude=['test']), scripts = ['twinstance'], platforms = 'Posix; MacOS X; Windows', install_requires = [ 'tiddlyweb>=0.9.96', 'tiddlywebplugins.wikklytextrender', 'tiddlywebplugins.status>=0.5', 'tiddlywebplugins.differ', 'tiddlywebplugins.atom', 'tiddlywebplugins.twimport', 'tiddlywebplugins.utils', 'tiddlywebplugins.instancer>=0.5.5', 'BeautifulSoup', 'wikklytext'], include_package_data = True, zip_safe = False ) Remove BeautifulSoup from direct dependency list.
from setuptools import setup, find_packages from tiddlywebwiki import __version__ as VERSION setup( name = 'tiddlywebwiki', version = VERSION, description = 'A TiddlyWeb plugin to provide a multi-user TiddlyWiki environment.', author = 'FND', author_email = 'FNDo@gmx.net', packages = find_packages(exclude=['test']), scripts = ['twinstance'], platforms = 'Posix; MacOS X; Windows', install_requires = [ 'tiddlyweb>=0.9.96', 'tiddlywebplugins.wikklytextrender', 'tiddlywebplugins.status>=0.5', 'tiddlywebplugins.differ', 'tiddlywebplugins.atom', 'tiddlywebplugins.twimport', 'tiddlywebplugins.utils', 'tiddlywebplugins.instancer>=0.5.5', 'wikklytext'], include_package_data = True, zip_safe = False )
<commit_before>from setuptools import setup, find_packages from tiddlywebwiki import __version__ as VERSION setup( name = 'tiddlywebwiki', version = VERSION, description = 'A TiddlyWeb plugin to provide a multi-user TiddlyWiki environment.', author = 'FND', author_email = 'FNDo@gmx.net', packages = find_packages(exclude=['test']), scripts = ['twinstance'], platforms = 'Posix; MacOS X; Windows', install_requires = [ 'tiddlyweb>=0.9.96', 'tiddlywebplugins.wikklytextrender', 'tiddlywebplugins.status>=0.5', 'tiddlywebplugins.differ', 'tiddlywebplugins.atom', 'tiddlywebplugins.twimport', 'tiddlywebplugins.utils', 'tiddlywebplugins.instancer>=0.5.5', 'BeautifulSoup', 'wikklytext'], include_package_data = True, zip_safe = False ) <commit_msg>Remove BeautifulSoup from direct dependency list.<commit_after>
from setuptools import setup, find_packages from tiddlywebwiki import __version__ as VERSION setup( name = 'tiddlywebwiki', version = VERSION, description = 'A TiddlyWeb plugin to provide a multi-user TiddlyWiki environment.', author = 'FND', author_email = 'FNDo@gmx.net', packages = find_packages(exclude=['test']), scripts = ['twinstance'], platforms = 'Posix; MacOS X; Windows', install_requires = [ 'tiddlyweb>=0.9.96', 'tiddlywebplugins.wikklytextrender', 'tiddlywebplugins.status>=0.5', 'tiddlywebplugins.differ', 'tiddlywebplugins.atom', 'tiddlywebplugins.twimport', 'tiddlywebplugins.utils', 'tiddlywebplugins.instancer>=0.5.5', 'wikklytext'], include_package_data = True, zip_safe = False )
from setuptools import setup, find_packages from tiddlywebwiki import __version__ as VERSION setup( name = 'tiddlywebwiki', version = VERSION, description = 'A TiddlyWeb plugin to provide a multi-user TiddlyWiki environment.', author = 'FND', author_email = 'FNDo@gmx.net', packages = find_packages(exclude=['test']), scripts = ['twinstance'], platforms = 'Posix; MacOS X; Windows', install_requires = [ 'tiddlyweb>=0.9.96', 'tiddlywebplugins.wikklytextrender', 'tiddlywebplugins.status>=0.5', 'tiddlywebplugins.differ', 'tiddlywebplugins.atom', 'tiddlywebplugins.twimport', 'tiddlywebplugins.utils', 'tiddlywebplugins.instancer>=0.5.5', 'BeautifulSoup', 'wikklytext'], include_package_data = True, zip_safe = False ) Remove BeautifulSoup from direct dependency list.from setuptools import setup, find_packages from tiddlywebwiki import __version__ as VERSION setup( name = 'tiddlywebwiki', version = VERSION, description = 'A TiddlyWeb plugin to provide a multi-user TiddlyWiki environment.', author = 'FND', author_email = 'FNDo@gmx.net', packages = find_packages(exclude=['test']), scripts = ['twinstance'], platforms = 'Posix; MacOS X; Windows', install_requires = [ 'tiddlyweb>=0.9.96', 'tiddlywebplugins.wikklytextrender', 'tiddlywebplugins.status>=0.5', 'tiddlywebplugins.differ', 'tiddlywebplugins.atom', 'tiddlywebplugins.twimport', 'tiddlywebplugins.utils', 'tiddlywebplugins.instancer>=0.5.5', 'wikklytext'], include_package_data = True, zip_safe = False )
<commit_before>from setuptools import setup, find_packages from tiddlywebwiki import __version__ as VERSION setup( name = 'tiddlywebwiki', version = VERSION, description = 'A TiddlyWeb plugin to provide a multi-user TiddlyWiki environment.', author = 'FND', author_email = 'FNDo@gmx.net', packages = find_packages(exclude=['test']), scripts = ['twinstance'], platforms = 'Posix; MacOS X; Windows', install_requires = [ 'tiddlyweb>=0.9.96', 'tiddlywebplugins.wikklytextrender', 'tiddlywebplugins.status>=0.5', 'tiddlywebplugins.differ', 'tiddlywebplugins.atom', 'tiddlywebplugins.twimport', 'tiddlywebplugins.utils', 'tiddlywebplugins.instancer>=0.5.5', 'BeautifulSoup', 'wikklytext'], include_package_data = True, zip_safe = False ) <commit_msg>Remove BeautifulSoup from direct dependency list.<commit_after>from setuptools import setup, find_packages from tiddlywebwiki import __version__ as VERSION setup( name = 'tiddlywebwiki', version = VERSION, description = 'A TiddlyWeb plugin to provide a multi-user TiddlyWiki environment.', author = 'FND', author_email = 'FNDo@gmx.net', packages = find_packages(exclude=['test']), scripts = ['twinstance'], platforms = 'Posix; MacOS X; Windows', install_requires = [ 'tiddlyweb>=0.9.96', 'tiddlywebplugins.wikklytextrender', 'tiddlywebplugins.status>=0.5', 'tiddlywebplugins.differ', 'tiddlywebplugins.atom', 'tiddlywebplugins.twimport', 'tiddlywebplugins.utils', 'tiddlywebplugins.instancer>=0.5.5', 'wikklytext'], include_package_data = True, zip_safe = False )
7d547301e047556a7f95a76e80c0cf1fde5aa960
setup.py
setup.py
import os import re from setuptools import setup HERE = os.path.dirname(os.path.abspath(__file__)) def get_version(): filename = os.path.join(HERE, 'transfluent.py') contents = open(filename).read() pattern = r"^__version__ = '(.*?)'$" return re.search(pattern, contents, re.MULTILINE).group(1) setup( name='transfluent', version=get_version(), description='A Python wrapper for Transfluent API', long_description=( open('README.rst').read() + '\n' + open('CHANGES.rst').read() ), author='Janne Vanhala', author_email='janne.vanhala@gmail.com', url='http://github.com/jpvanhal/python-transfluent', py_modules=['transfluent'], license=open('LICENSE').read(), platforms='any', install_requires=[ 'requests>=1.0', ], classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: Implementation :: PyPy', 'Topic :: Software Development :: Libraries :: Python Modules' ] )
import os import re from setuptools import setup HERE = os.path.dirname(os.path.abspath(__file__)) def get_version(): filename = os.path.join(HERE, 'transfluent.py') contents = open(filename).read() pattern = r"^__version__ = '(.*?)'$" return re.search(pattern, contents, re.MULTILINE).group(1) setup( name='transfluent', version=get_version(), description='A Python wrapper for Transfluent API', long_description=( open('README.rst').read() + '\n' + open('CHANGES.rst').read() ), author='Janne Vanhala', author_email='janne.vanhala@gmail.com', url='http://github.com/jpvanhal/python-transfluent', py_modules=['transfluent'], license=open('LICENSE').read(), platforms='any', install_requires=[ 'requests>=1.0', ], classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: Implementation :: PyPy', 'Topic :: Software Development :: Libraries :: Python Modules' ] )
Add Python 3.5 to trove classifiers
Add Python 3.5 to trove classifiers
Python
bsd-3-clause
jpvanhal/python-transfluent
import os import re from setuptools import setup HERE = os.path.dirname(os.path.abspath(__file__)) def get_version(): filename = os.path.join(HERE, 'transfluent.py') contents = open(filename).read() pattern = r"^__version__ = '(.*?)'$" return re.search(pattern, contents, re.MULTILINE).group(1) setup( name='transfluent', version=get_version(), description='A Python wrapper for Transfluent API', long_description=( open('README.rst').read() + '\n' + open('CHANGES.rst').read() ), author='Janne Vanhala', author_email='janne.vanhala@gmail.com', url='http://github.com/jpvanhal/python-transfluent', py_modules=['transfluent'], license=open('LICENSE').read(), platforms='any', install_requires=[ 'requests>=1.0', ], classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: Implementation :: PyPy', 'Topic :: Software Development :: Libraries :: Python Modules' ] ) Add Python 3.5 to trove classifiers
import os import re from setuptools import setup HERE = os.path.dirname(os.path.abspath(__file__)) def get_version(): filename = os.path.join(HERE, 'transfluent.py') contents = open(filename).read() pattern = r"^__version__ = '(.*?)'$" return re.search(pattern, contents, re.MULTILINE).group(1) setup( name='transfluent', version=get_version(), description='A Python wrapper for Transfluent API', long_description=( open('README.rst').read() + '\n' + open('CHANGES.rst').read() ), author='Janne Vanhala', author_email='janne.vanhala@gmail.com', url='http://github.com/jpvanhal/python-transfluent', py_modules=['transfluent'], license=open('LICENSE').read(), platforms='any', install_requires=[ 'requests>=1.0', ], classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: Implementation :: PyPy', 'Topic :: Software Development :: Libraries :: Python Modules' ] )
<commit_before>import os import re from setuptools import setup HERE = os.path.dirname(os.path.abspath(__file__)) def get_version(): filename = os.path.join(HERE, 'transfluent.py') contents = open(filename).read() pattern = r"^__version__ = '(.*?)'$" return re.search(pattern, contents, re.MULTILINE).group(1) setup( name='transfluent', version=get_version(), description='A Python wrapper for Transfluent API', long_description=( open('README.rst').read() + '\n' + open('CHANGES.rst').read() ), author='Janne Vanhala', author_email='janne.vanhala@gmail.com', url='http://github.com/jpvanhal/python-transfluent', py_modules=['transfluent'], license=open('LICENSE').read(), platforms='any', install_requires=[ 'requests>=1.0', ], classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: Implementation :: PyPy', 'Topic :: Software Development :: Libraries :: Python Modules' ] ) <commit_msg>Add Python 3.5 to trove classifiers<commit_after>
import os import re from setuptools import setup HERE = os.path.dirname(os.path.abspath(__file__)) def get_version(): filename = os.path.join(HERE, 'transfluent.py') contents = open(filename).read() pattern = r"^__version__ = '(.*?)'$" return re.search(pattern, contents, re.MULTILINE).group(1) setup( name='transfluent', version=get_version(), description='A Python wrapper for Transfluent API', long_description=( open('README.rst').read() + '\n' + open('CHANGES.rst').read() ), author='Janne Vanhala', author_email='janne.vanhala@gmail.com', url='http://github.com/jpvanhal/python-transfluent', py_modules=['transfluent'], license=open('LICENSE').read(), platforms='any', install_requires=[ 'requests>=1.0', ], classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: Implementation :: PyPy', 'Topic :: Software Development :: Libraries :: Python Modules' ] )
import os import re from setuptools import setup HERE = os.path.dirname(os.path.abspath(__file__)) def get_version(): filename = os.path.join(HERE, 'transfluent.py') contents = open(filename).read() pattern = r"^__version__ = '(.*?)'$" return re.search(pattern, contents, re.MULTILINE).group(1) setup( name='transfluent', version=get_version(), description='A Python wrapper for Transfluent API', long_description=( open('README.rst').read() + '\n' + open('CHANGES.rst').read() ), author='Janne Vanhala', author_email='janne.vanhala@gmail.com', url='http://github.com/jpvanhal/python-transfluent', py_modules=['transfluent'], license=open('LICENSE').read(), platforms='any', install_requires=[ 'requests>=1.0', ], classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: Implementation :: PyPy', 'Topic :: Software Development :: Libraries :: Python Modules' ] ) Add Python 3.5 to trove classifiersimport os import re from setuptools import setup HERE = os.path.dirname(os.path.abspath(__file__)) def get_version(): filename = os.path.join(HERE, 'transfluent.py') contents = open(filename).read() pattern = r"^__version__ = '(.*?)'$" return re.search(pattern, contents, re.MULTILINE).group(1) setup( name='transfluent', version=get_version(), description='A Python wrapper for Transfluent API', long_description=( open('README.rst').read() + '\n' + open('CHANGES.rst').read() ), author='Janne Vanhala', author_email='janne.vanhala@gmail.com', url='http://github.com/jpvanhal/python-transfluent', py_modules=['transfluent'], license=open('LICENSE').read(), platforms='any', install_requires=[ 'requests>=1.0', ], classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: Implementation :: PyPy', 'Topic :: Software Development :: Libraries :: Python Modules' ] )
<commit_before>import os import re from setuptools import setup HERE = os.path.dirname(os.path.abspath(__file__)) def get_version(): filename = os.path.join(HERE, 'transfluent.py') contents = open(filename).read() pattern = r"^__version__ = '(.*?)'$" return re.search(pattern, contents, re.MULTILINE).group(1) setup( name='transfluent', version=get_version(), description='A Python wrapper for Transfluent API', long_description=( open('README.rst').read() + '\n' + open('CHANGES.rst').read() ), author='Janne Vanhala', author_email='janne.vanhala@gmail.com', url='http://github.com/jpvanhal/python-transfluent', py_modules=['transfluent'], license=open('LICENSE').read(), platforms='any', install_requires=[ 'requests>=1.0', ], classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: Implementation :: PyPy', 'Topic :: Software Development :: Libraries :: Python Modules' ] ) <commit_msg>Add Python 3.5 to trove classifiers<commit_after>import os import re from setuptools import setup HERE = os.path.dirname(os.path.abspath(__file__)) def get_version(): filename = os.path.join(HERE, 'transfluent.py') contents = open(filename).read() pattern = r"^__version__ = '(.*?)'$" return re.search(pattern, contents, re.MULTILINE).group(1) setup( name='transfluent', version=get_version(), description='A Python wrapper for Transfluent API', long_description=( open('README.rst').read() + '\n' + open('CHANGES.rst').read() ), author='Janne Vanhala', author_email='janne.vanhala@gmail.com', url='http://github.com/jpvanhal/python-transfluent', py_modules=['transfluent'], license=open('LICENSE').read(), platforms='any', install_requires=[ 'requests>=1.0', ], classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: Implementation :: PyPy', 'Topic :: Software Development :: Libraries :: Python Modules' ] )
aa647c7e9a35b2293739c55e353319d2626f1f33
setup.py
setup.py
#! /usr/bin/env python3 # -*- coding:Utf8 -*- #-------------------------------------------------------------------------------------------------------------- # All necessary import: #-------------------------------------------------------------------------------------------------------------- import os, sys, glob import setuptools as st from distutils.core import setup from distutils.command.install_data import install_data packages = st.find_packages() #-------------------------------------------------------------------------------------------------------------- # Call the setup function: #-------------------------------------------------------------------------------------------------------------- setup( name = 'kmsi', version = '0.1', description = 'Python Module for analysis gadget simulation.', author = 'Guillaume Plum', packages = packages, cmdclass = {'install_data': install_data}, # data_files = [ # ('share/LibThese/animation-plugins', ["share/LibThese/animation-plugins/__init__.py"]), #glob.glob("share/LibThese/animation-plugins/*.py")), # ], # scripts = [ # 'scripts/animationv2.py', # 'scripts/models_plot.py', # ], ) #vim:spelllang=
#! /usr/bin/env python3 # -*- coding:Utf8 -*- #-------------------------------------------------------------------------------------------------------------- # All necessary import: #-------------------------------------------------------------------------------------------------------------- import os, sys, glob import setuptools as st from distutils.core import setup from distutils.command.install_data import install_data packages = st.find_packages() #-------------------------------------------------------------------------------------------------------------- # Call the setup function: #-------------------------------------------------------------------------------------------------------------- setup( name = 'kmsi', version = '0.1', description = 'Python Module for analysis gadget simulation.', author = 'Guillaume Plum', packages = packages, cmdclass = {'install_data': install_data}, data_files = [ ('share/kmsi', ["share/kmsi/config.yaml"]), #glob.glob("share/LibThese/animation-plugins/*.py")), ], scripts = [ 'scripts/kmsi', ], ) #vim:spelllang=
Add the kmsi script to the install files.
Add the kmsi script to the install files.
Python
mit
ElricleNecro/kmsi
#! /usr/bin/env python3 # -*- coding:Utf8 -*- #-------------------------------------------------------------------------------------------------------------- # All necessary import: #-------------------------------------------------------------------------------------------------------------- import os, sys, glob import setuptools as st from distutils.core import setup from distutils.command.install_data import install_data packages = st.find_packages() #-------------------------------------------------------------------------------------------------------------- # Call the setup function: #-------------------------------------------------------------------------------------------------------------- setup( name = 'kmsi', version = '0.1', description = 'Python Module for analysis gadget simulation.', author = 'Guillaume Plum', packages = packages, cmdclass = {'install_data': install_data}, # data_files = [ # ('share/LibThese/animation-plugins', ["share/LibThese/animation-plugins/__init__.py"]), #glob.glob("share/LibThese/animation-plugins/*.py")), # ], # scripts = [ # 'scripts/animationv2.py', # 'scripts/models_plot.py', # ], ) #vim:spelllang= Add the kmsi script to the install files.
#! /usr/bin/env python3 # -*- coding:Utf8 -*- #-------------------------------------------------------------------------------------------------------------- # All necessary import: #-------------------------------------------------------------------------------------------------------------- import os, sys, glob import setuptools as st from distutils.core import setup from distutils.command.install_data import install_data packages = st.find_packages() #-------------------------------------------------------------------------------------------------------------- # Call the setup function: #-------------------------------------------------------------------------------------------------------------- setup( name = 'kmsi', version = '0.1', description = 'Python Module for analysis gadget simulation.', author = 'Guillaume Plum', packages = packages, cmdclass = {'install_data': install_data}, data_files = [ ('share/kmsi', ["share/kmsi/config.yaml"]), #glob.glob("share/LibThese/animation-plugins/*.py")), ], scripts = [ 'scripts/kmsi', ], ) #vim:spelllang=
<commit_before>#! /usr/bin/env python3 # -*- coding:Utf8 -*- #-------------------------------------------------------------------------------------------------------------- # All necessary import: #-------------------------------------------------------------------------------------------------------------- import os, sys, glob import setuptools as st from distutils.core import setup from distutils.command.install_data import install_data packages = st.find_packages() #-------------------------------------------------------------------------------------------------------------- # Call the setup function: #-------------------------------------------------------------------------------------------------------------- setup( name = 'kmsi', version = '0.1', description = 'Python Module for analysis gadget simulation.', author = 'Guillaume Plum', packages = packages, cmdclass = {'install_data': install_data}, # data_files = [ # ('share/LibThese/animation-plugins', ["share/LibThese/animation-plugins/__init__.py"]), #glob.glob("share/LibThese/animation-plugins/*.py")), # ], # scripts = [ # 'scripts/animationv2.py', # 'scripts/models_plot.py', # ], ) #vim:spelllang= <commit_msg>Add the kmsi script to the install files.<commit_after>
#! /usr/bin/env python3 # -*- coding:Utf8 -*- #-------------------------------------------------------------------------------------------------------------- # All necessary import: #-------------------------------------------------------------------------------------------------------------- import os, sys, glob import setuptools as st from distutils.core import setup from distutils.command.install_data import install_data packages = st.find_packages() #-------------------------------------------------------------------------------------------------------------- # Call the setup function: #-------------------------------------------------------------------------------------------------------------- setup( name = 'kmsi', version = '0.1', description = 'Python Module for analysis gadget simulation.', author = 'Guillaume Plum', packages = packages, cmdclass = {'install_data': install_data}, data_files = [ ('share/kmsi', ["share/kmsi/config.yaml"]), #glob.glob("share/LibThese/animation-plugins/*.py")), ], scripts = [ 'scripts/kmsi', ], ) #vim:spelllang=
#! /usr/bin/env python3 # -*- coding:Utf8 -*- #-------------------------------------------------------------------------------------------------------------- # All necessary import: #-------------------------------------------------------------------------------------------------------------- import os, sys, glob import setuptools as st from distutils.core import setup from distutils.command.install_data import install_data packages = st.find_packages() #-------------------------------------------------------------------------------------------------------------- # Call the setup function: #-------------------------------------------------------------------------------------------------------------- setup( name = 'kmsi', version = '0.1', description = 'Python Module for analysis gadget simulation.', author = 'Guillaume Plum', packages = packages, cmdclass = {'install_data': install_data}, # data_files = [ # ('share/LibThese/animation-plugins', ["share/LibThese/animation-plugins/__init__.py"]), #glob.glob("share/LibThese/animation-plugins/*.py")), # ], # scripts = [ # 'scripts/animationv2.py', # 'scripts/models_plot.py', # ], ) #vim:spelllang= Add the kmsi script to the install files.#! /usr/bin/env python3 # -*- coding:Utf8 -*- #-------------------------------------------------------------------------------------------------------------- # All necessary import: #-------------------------------------------------------------------------------------------------------------- import os, sys, glob import setuptools as st from distutils.core import setup from distutils.command.install_data import install_data packages = st.find_packages() #-------------------------------------------------------------------------------------------------------------- # Call the setup function: #-------------------------------------------------------------------------------------------------------------- setup( name = 'kmsi', version = '0.1', description = 'Python Module for analysis gadget simulation.', author = 'Guillaume Plum', packages = packages, cmdclass = {'install_data': install_data}, data_files = [ ('share/kmsi', ["share/kmsi/config.yaml"]), #glob.glob("share/LibThese/animation-plugins/*.py")), ], scripts = [ 'scripts/kmsi', ], ) #vim:spelllang=
<commit_before>#! /usr/bin/env python3 # -*- coding:Utf8 -*- #-------------------------------------------------------------------------------------------------------------- # All necessary import: #-------------------------------------------------------------------------------------------------------------- import os, sys, glob import setuptools as st from distutils.core import setup from distutils.command.install_data import install_data packages = st.find_packages() #-------------------------------------------------------------------------------------------------------------- # Call the setup function: #-------------------------------------------------------------------------------------------------------------- setup( name = 'kmsi', version = '0.1', description = 'Python Module for analysis gadget simulation.', author = 'Guillaume Plum', packages = packages, cmdclass = {'install_data': install_data}, # data_files = [ # ('share/LibThese/animation-plugins', ["share/LibThese/animation-plugins/__init__.py"]), #glob.glob("share/LibThese/animation-plugins/*.py")), # ], # scripts = [ # 'scripts/animationv2.py', # 'scripts/models_plot.py', # ], ) #vim:spelllang= <commit_msg>Add the kmsi script to the install files.<commit_after>#! /usr/bin/env python3 # -*- coding:Utf8 -*- #-------------------------------------------------------------------------------------------------------------- # All necessary import: #-------------------------------------------------------------------------------------------------------------- import os, sys, glob import setuptools as st from distutils.core import setup from distutils.command.install_data import install_data packages = st.find_packages() #-------------------------------------------------------------------------------------------------------------- # Call the setup function: #-------------------------------------------------------------------------------------------------------------- setup( name = 'kmsi', version = '0.1', description = 'Python Module for analysis gadget simulation.', author = 'Guillaume Plum', packages = packages, cmdclass = {'install_data': install_data}, data_files = [ ('share/kmsi', ["share/kmsi/config.yaml"]), #glob.glob("share/LibThese/animation-plugins/*.py")), ], scripts = [ 'scripts/kmsi', ], ) #vim:spelllang=
f92c37200bb889188af21f0280c908f1bc2bcbff
setup.py
setup.py
from setuptools import setup, find_packages setup( name='incuna-pigeon', version='0.0.0', description='Notification management', url='https://github.com/incuna/incuna-pigeon', author='Incuna', author_email='admin@incuna.com', license='BSD', classifiers=[ 'Development Status :: 1 - Planning', 'Intended Audience :: Developers', 'Topic :: Communications', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', ], keywords='notifications', packages=find_packages(), )
from setuptools import setup, find_packages setup( name='incuna-pigeon', version='0.0.0', description='Notification management', url='https://github.com/incuna/incuna-pigeon', author='Incuna', author_email='admin@incuna.com', license='BSD', classifiers=[ 'Development Status :: 1 - Planning', 'Intended Audience :: Developers', 'Topic :: Communications', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3 :: Only', 'Programming Language :: Python :: 3.4', ], keywords='notifications', packages=find_packages(), )
Add Python 3 Only trove classifier
Add Python 3 Only trove classifier
Python
bsd-2-clause
incuna/incuna-pigeon
from setuptools import setup, find_packages setup( name='incuna-pigeon', version='0.0.0', description='Notification management', url='https://github.com/incuna/incuna-pigeon', author='Incuna', author_email='admin@incuna.com', license='BSD', classifiers=[ 'Development Status :: 1 - Planning', 'Intended Audience :: Developers', 'Topic :: Communications', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', ], keywords='notifications', packages=find_packages(), ) Add Python 3 Only trove classifier
from setuptools import setup, find_packages setup( name='incuna-pigeon', version='0.0.0', description='Notification management', url='https://github.com/incuna/incuna-pigeon', author='Incuna', author_email='admin@incuna.com', license='BSD', classifiers=[ 'Development Status :: 1 - Planning', 'Intended Audience :: Developers', 'Topic :: Communications', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3 :: Only', 'Programming Language :: Python :: 3.4', ], keywords='notifications', packages=find_packages(), )
<commit_before>from setuptools import setup, find_packages setup( name='incuna-pigeon', version='0.0.0', description='Notification management', url='https://github.com/incuna/incuna-pigeon', author='Incuna', author_email='admin@incuna.com', license='BSD', classifiers=[ 'Development Status :: 1 - Planning', 'Intended Audience :: Developers', 'Topic :: Communications', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', ], keywords='notifications', packages=find_packages(), ) <commit_msg>Add Python 3 Only trove classifier<commit_after>
from setuptools import setup, find_packages setup( name='incuna-pigeon', version='0.0.0', description='Notification management', url='https://github.com/incuna/incuna-pigeon', author='Incuna', author_email='admin@incuna.com', license='BSD', classifiers=[ 'Development Status :: 1 - Planning', 'Intended Audience :: Developers', 'Topic :: Communications', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3 :: Only', 'Programming Language :: Python :: 3.4', ], keywords='notifications', packages=find_packages(), )
from setuptools import setup, find_packages setup( name='incuna-pigeon', version='0.0.0', description='Notification management', url='https://github.com/incuna/incuna-pigeon', author='Incuna', author_email='admin@incuna.com', license='BSD', classifiers=[ 'Development Status :: 1 - Planning', 'Intended Audience :: Developers', 'Topic :: Communications', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', ], keywords='notifications', packages=find_packages(), ) Add Python 3 Only trove classifierfrom setuptools import setup, find_packages setup( name='incuna-pigeon', version='0.0.0', description='Notification management', url='https://github.com/incuna/incuna-pigeon', author='Incuna', author_email='admin@incuna.com', license='BSD', classifiers=[ 'Development Status :: 1 - Planning', 'Intended Audience :: Developers', 'Topic :: Communications', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3 :: Only', 'Programming Language :: Python :: 3.4', ], keywords='notifications', packages=find_packages(), )
<commit_before>from setuptools import setup, find_packages setup( name='incuna-pigeon', version='0.0.0', description='Notification management', url='https://github.com/incuna/incuna-pigeon', author='Incuna', author_email='admin@incuna.com', license='BSD', classifiers=[ 'Development Status :: 1 - Planning', 'Intended Audience :: Developers', 'Topic :: Communications', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', ], keywords='notifications', packages=find_packages(), ) <commit_msg>Add Python 3 Only trove classifier<commit_after>from setuptools import setup, find_packages setup( name='incuna-pigeon', version='0.0.0', description='Notification management', url='https://github.com/incuna/incuna-pigeon', author='Incuna', author_email='admin@incuna.com', license='BSD', classifiers=[ 'Development Status :: 1 - Planning', 'Intended Audience :: Developers', 'Topic :: Communications', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3 :: Only', 'Programming Language :: Python :: 3.4', ], keywords='notifications', packages=find_packages(), )
51c65f37ea5f0d2cd98de8e63f541d533e1f8a65
setup.py
setup.py
#!/usr/bin/env python from codecs import open from setuptools import find_packages, setup with open('README.rst', 'r', 'utf-8') as f: readme = f.read() setup( name='django-paginationlinks', version='0.1.1', description='Django Pagination Links', long_description=readme, url='https://github.com/blancltd/django-paginationlinks', maintainer='Blanc Ltd', maintainer_email='studio@blanc.ltd.uk', platforms=['any'], packages=find_packages(), classifiers=[ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], license='BSD', )
#!/usr/bin/env python from codecs import open from setuptools import find_packages, setup with open('README.rst', 'r', 'utf-8') as f: readme = f.read() setup( name='django-paginationlinks', version='0.1.1', description='Django Pagination Links', long_description=readme, url='https://github.com/developersociety/django-paginationlinks', maintainer='Blanc Ltd', maintainer_email='studio@blanc.ltd.uk', platforms=['any'], packages=find_packages(), classifiers=[ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], license='BSD', )
Update GitHub repos from blancltd to developersociety
Update GitHub repos from blancltd to developersociety
Python
bsd-3-clause
blancltd/django-paginationlinks
#!/usr/bin/env python from codecs import open from setuptools import find_packages, setup with open('README.rst', 'r', 'utf-8') as f: readme = f.read() setup( name='django-paginationlinks', version='0.1.1', description='Django Pagination Links', long_description=readme, url='https://github.com/blancltd/django-paginationlinks', maintainer='Blanc Ltd', maintainer_email='studio@blanc.ltd.uk', platforms=['any'], packages=find_packages(), classifiers=[ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], license='BSD', ) Update GitHub repos from blancltd to developersociety
#!/usr/bin/env python from codecs import open from setuptools import find_packages, setup with open('README.rst', 'r', 'utf-8') as f: readme = f.read() setup( name='django-paginationlinks', version='0.1.1', description='Django Pagination Links', long_description=readme, url='https://github.com/developersociety/django-paginationlinks', maintainer='Blanc Ltd', maintainer_email='studio@blanc.ltd.uk', platforms=['any'], packages=find_packages(), classifiers=[ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], license='BSD', )
<commit_before>#!/usr/bin/env python from codecs import open from setuptools import find_packages, setup with open('README.rst', 'r', 'utf-8') as f: readme = f.read() setup( name='django-paginationlinks', version='0.1.1', description='Django Pagination Links', long_description=readme, url='https://github.com/blancltd/django-paginationlinks', maintainer='Blanc Ltd', maintainer_email='studio@blanc.ltd.uk', platforms=['any'], packages=find_packages(), classifiers=[ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], license='BSD', ) <commit_msg>Update GitHub repos from blancltd to developersociety<commit_after>
#!/usr/bin/env python from codecs import open from setuptools import find_packages, setup with open('README.rst', 'r', 'utf-8') as f: readme = f.read() setup( name='django-paginationlinks', version='0.1.1', description='Django Pagination Links', long_description=readme, url='https://github.com/developersociety/django-paginationlinks', maintainer='Blanc Ltd', maintainer_email='studio@blanc.ltd.uk', platforms=['any'], packages=find_packages(), classifiers=[ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], license='BSD', )
#!/usr/bin/env python from codecs import open from setuptools import find_packages, setup with open('README.rst', 'r', 'utf-8') as f: readme = f.read() setup( name='django-paginationlinks', version='0.1.1', description='Django Pagination Links', long_description=readme, url='https://github.com/blancltd/django-paginationlinks', maintainer='Blanc Ltd', maintainer_email='studio@blanc.ltd.uk', platforms=['any'], packages=find_packages(), classifiers=[ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], license='BSD', ) Update GitHub repos from blancltd to developersociety#!/usr/bin/env python from codecs import open from setuptools import find_packages, setup with open('README.rst', 'r', 'utf-8') as f: readme = f.read() setup( name='django-paginationlinks', version='0.1.1', description='Django Pagination Links', long_description=readme, url='https://github.com/developersociety/django-paginationlinks', maintainer='Blanc Ltd', maintainer_email='studio@blanc.ltd.uk', platforms=['any'], packages=find_packages(), classifiers=[ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], license='BSD', )
<commit_before>#!/usr/bin/env python from codecs import open from setuptools import find_packages, setup with open('README.rst', 'r', 'utf-8') as f: readme = f.read() setup( name='django-paginationlinks', version='0.1.1', description='Django Pagination Links', long_description=readme, url='https://github.com/blancltd/django-paginationlinks', maintainer='Blanc Ltd', maintainer_email='studio@blanc.ltd.uk', platforms=['any'], packages=find_packages(), classifiers=[ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], license='BSD', ) <commit_msg>Update GitHub repos from blancltd to developersociety<commit_after>#!/usr/bin/env python from codecs import open from setuptools import find_packages, setup with open('README.rst', 'r', 'utf-8') as f: readme = f.read() setup( name='django-paginationlinks', version='0.1.1', description='Django Pagination Links', long_description=readme, url='https://github.com/developersociety/django-paginationlinks', maintainer='Blanc Ltd', maintainer_email='studio@blanc.ltd.uk', platforms=['any'], packages=find_packages(), classifiers=[ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], license='BSD', )
e8efd3b22d7b301771d72b2d7b9ca4cca474e302
setup.py
setup.py
try: from setuptools import setup except ImportError: from distutils.core import setup config = { 'description': 'pelican-do', 'author': 'Commands to automate common pelican tasks', 'url': '', 'download_url': '', 'author_email': 'gustavoajz@gmail.com', 'version': '0.1', 'install_requires': [ 'click==6.2', 'Jinja2==2.8', 'awesome-slugify==1.6.5', ], 'extras_require': { 'development': [ ], }, 'setup_requires': [ 'pytest-runner', ], 'tests_require': [ 'pytest==2.8.5', 'pytest-cov==2.2.0' ], 'packages': ['pelican_do'], 'scripts': [], 'name': 'pelican-do', 'entry_points': { 'console_scripts': ['pelican-do=pelican_do.main:main'] } } setup(**config)
try: from setuptools import setup except ImportError: from distutils.core import setup config = { 'description': 'pelican-do', 'author': 'Commands to automate common pelican tasks', 'url': '', 'download_url': '', 'author_email': 'gustavoajz@gmail.com', 'version': '0.1', 'install_requires': [ 'click==6.2', 'Jinja2==2.8', 'awesome-slugify==1.6.5', ], 'extras_require': { 'development': [ ], }, 'setup_requires': [ 'pytest-runner', ], 'tests_require': [ 'pytest>=2.6.4', 'pytest-cov==2.2.0' ], 'packages': ['pelican_do'], 'scripts': [], 'name': 'pelican-do', 'entry_points': { 'console_scripts': ['pelican-do=pelican_do.main:main'] } } setup(**config)
Downgrade pytest version to be able to use default shippable minion
Downgrade pytest version to be able to use default shippable minion
Python
mit
gusajz/pelican-do,gusajz/pelican-do
try: from setuptools import setup except ImportError: from distutils.core import setup config = { 'description': 'pelican-do', 'author': 'Commands to automate common pelican tasks', 'url': '', 'download_url': '', 'author_email': 'gustavoajz@gmail.com', 'version': '0.1', 'install_requires': [ 'click==6.2', 'Jinja2==2.8', 'awesome-slugify==1.6.5', ], 'extras_require': { 'development': [ ], }, 'setup_requires': [ 'pytest-runner', ], 'tests_require': [ 'pytest==2.8.5', 'pytest-cov==2.2.0' ], 'packages': ['pelican_do'], 'scripts': [], 'name': 'pelican-do', 'entry_points': { 'console_scripts': ['pelican-do=pelican_do.main:main'] } } setup(**config) Downgrade pytest version to be able to use default shippable minion
try: from setuptools import setup except ImportError: from distutils.core import setup config = { 'description': 'pelican-do', 'author': 'Commands to automate common pelican tasks', 'url': '', 'download_url': '', 'author_email': 'gustavoajz@gmail.com', 'version': '0.1', 'install_requires': [ 'click==6.2', 'Jinja2==2.8', 'awesome-slugify==1.6.5', ], 'extras_require': { 'development': [ ], }, 'setup_requires': [ 'pytest-runner', ], 'tests_require': [ 'pytest>=2.6.4', 'pytest-cov==2.2.0' ], 'packages': ['pelican_do'], 'scripts': [], 'name': 'pelican-do', 'entry_points': { 'console_scripts': ['pelican-do=pelican_do.main:main'] } } setup(**config)
<commit_before> try: from setuptools import setup except ImportError: from distutils.core import setup config = { 'description': 'pelican-do', 'author': 'Commands to automate common pelican tasks', 'url': '', 'download_url': '', 'author_email': 'gustavoajz@gmail.com', 'version': '0.1', 'install_requires': [ 'click==6.2', 'Jinja2==2.8', 'awesome-slugify==1.6.5', ], 'extras_require': { 'development': [ ], }, 'setup_requires': [ 'pytest-runner', ], 'tests_require': [ 'pytest==2.8.5', 'pytest-cov==2.2.0' ], 'packages': ['pelican_do'], 'scripts': [], 'name': 'pelican-do', 'entry_points': { 'console_scripts': ['pelican-do=pelican_do.main:main'] } } setup(**config) <commit_msg>Downgrade pytest version to be able to use default shippable minion<commit_after>
try: from setuptools import setup except ImportError: from distutils.core import setup config = { 'description': 'pelican-do', 'author': 'Commands to automate common pelican tasks', 'url': '', 'download_url': '', 'author_email': 'gustavoajz@gmail.com', 'version': '0.1', 'install_requires': [ 'click==6.2', 'Jinja2==2.8', 'awesome-slugify==1.6.5', ], 'extras_require': { 'development': [ ], }, 'setup_requires': [ 'pytest-runner', ], 'tests_require': [ 'pytest>=2.6.4', 'pytest-cov==2.2.0' ], 'packages': ['pelican_do'], 'scripts': [], 'name': 'pelican-do', 'entry_points': { 'console_scripts': ['pelican-do=pelican_do.main:main'] } } setup(**config)
try: from setuptools import setup except ImportError: from distutils.core import setup config = { 'description': 'pelican-do', 'author': 'Commands to automate common pelican tasks', 'url': '', 'download_url': '', 'author_email': 'gustavoajz@gmail.com', 'version': '0.1', 'install_requires': [ 'click==6.2', 'Jinja2==2.8', 'awesome-slugify==1.6.5', ], 'extras_require': { 'development': [ ], }, 'setup_requires': [ 'pytest-runner', ], 'tests_require': [ 'pytest==2.8.5', 'pytest-cov==2.2.0' ], 'packages': ['pelican_do'], 'scripts': [], 'name': 'pelican-do', 'entry_points': { 'console_scripts': ['pelican-do=pelican_do.main:main'] } } setup(**config) Downgrade pytest version to be able to use default shippable minion try: from setuptools import setup except ImportError: from distutils.core import setup config = { 'description': 'pelican-do', 'author': 'Commands to automate common pelican tasks', 'url': '', 'download_url': '', 'author_email': 'gustavoajz@gmail.com', 'version': '0.1', 'install_requires': [ 'click==6.2', 'Jinja2==2.8', 'awesome-slugify==1.6.5', ], 'extras_require': { 'development': [ ], }, 'setup_requires': [ 'pytest-runner', ], 'tests_require': [ 'pytest>=2.6.4', 'pytest-cov==2.2.0' ], 'packages': ['pelican_do'], 'scripts': [], 'name': 'pelican-do', 'entry_points': { 'console_scripts': ['pelican-do=pelican_do.main:main'] } } setup(**config)
<commit_before> try: from setuptools import setup except ImportError: from distutils.core import setup config = { 'description': 'pelican-do', 'author': 'Commands to automate common pelican tasks', 'url': '', 'download_url': '', 'author_email': 'gustavoajz@gmail.com', 'version': '0.1', 'install_requires': [ 'click==6.2', 'Jinja2==2.8', 'awesome-slugify==1.6.5', ], 'extras_require': { 'development': [ ], }, 'setup_requires': [ 'pytest-runner', ], 'tests_require': [ 'pytest==2.8.5', 'pytest-cov==2.2.0' ], 'packages': ['pelican_do'], 'scripts': [], 'name': 'pelican-do', 'entry_points': { 'console_scripts': ['pelican-do=pelican_do.main:main'] } } setup(**config) <commit_msg>Downgrade pytest version to be able to use default shippable minion<commit_after> try: from setuptools import setup except ImportError: from distutils.core import setup config = { 'description': 'pelican-do', 'author': 'Commands to automate common pelican tasks', 'url': '', 'download_url': '', 'author_email': 'gustavoajz@gmail.com', 'version': '0.1', 'install_requires': [ 'click==6.2', 'Jinja2==2.8', 'awesome-slugify==1.6.5', ], 'extras_require': { 'development': [ ], }, 'setup_requires': [ 'pytest-runner', ], 'tests_require': [ 'pytest>=2.6.4', 'pytest-cov==2.2.0' ], 'packages': ['pelican_do'], 'scripts': [], 'name': 'pelican-do', 'entry_points': { 'console_scripts': ['pelican-do=pelican_do.main:main'] } } setup(**config)
b7acc8ca9c6c41aff7ffb419125f54d21da09652
setup.py
setup.py
#!/usr/bin/env python from distutils.core import setup setup(name='plyprotobuf', version='1.0', description='Protobuf Parsing Library that uses ply', author='Dusan Klinec', url='https://github.com/sb98052/plyprotobuf', packages=['plyproto'], )
#!/usr/bin/env python from distutils.core import setup setup(name='plyprotobuf', version='1.0', description='Protobuf Parsing Library that uses ply', author='Dusan Klinec', url='https://github.com/sb98052/plyprotobuf', packages=['plyproto'], install_requires=['ply'] )
Add dependency to ply package
Add dependency to ply package
Python
apache-2.0
sb98052/plyprotobuf
#!/usr/bin/env python from distutils.core import setup setup(name='plyprotobuf', version='1.0', description='Protobuf Parsing Library that uses ply', author='Dusan Klinec', url='https://github.com/sb98052/plyprotobuf', packages=['plyproto'], ) Add dependency to ply package
#!/usr/bin/env python from distutils.core import setup setup(name='plyprotobuf', version='1.0', description='Protobuf Parsing Library that uses ply', author='Dusan Klinec', url='https://github.com/sb98052/plyprotobuf', packages=['plyproto'], install_requires=['ply'] )
<commit_before>#!/usr/bin/env python from distutils.core import setup setup(name='plyprotobuf', version='1.0', description='Protobuf Parsing Library that uses ply', author='Dusan Klinec', url='https://github.com/sb98052/plyprotobuf', packages=['plyproto'], ) <commit_msg>Add dependency to ply package<commit_after>
#!/usr/bin/env python from distutils.core import setup setup(name='plyprotobuf', version='1.0', description='Protobuf Parsing Library that uses ply', author='Dusan Klinec', url='https://github.com/sb98052/plyprotobuf', packages=['plyproto'], install_requires=['ply'] )
#!/usr/bin/env python from distutils.core import setup setup(name='plyprotobuf', version='1.0', description='Protobuf Parsing Library that uses ply', author='Dusan Klinec', url='https://github.com/sb98052/plyprotobuf', packages=['plyproto'], ) Add dependency to ply package#!/usr/bin/env python from distutils.core import setup setup(name='plyprotobuf', version='1.0', description='Protobuf Parsing Library that uses ply', author='Dusan Klinec', url='https://github.com/sb98052/plyprotobuf', packages=['plyproto'], install_requires=['ply'] )
<commit_before>#!/usr/bin/env python from distutils.core import setup setup(name='plyprotobuf', version='1.0', description='Protobuf Parsing Library that uses ply', author='Dusan Klinec', url='https://github.com/sb98052/plyprotobuf', packages=['plyproto'], ) <commit_msg>Add dependency to ply package<commit_after>#!/usr/bin/env python from distutils.core import setup setup(name='plyprotobuf', version='1.0', description='Protobuf Parsing Library that uses ply', author='Dusan Klinec', url='https://github.com/sb98052/plyprotobuf', packages=['plyproto'], install_requires=['ply'] )
08b4cc4e065e63eef522756888fa8a75d9bf6ddb
setup.py
setup.py
from setuptools import setup, find_packages setup( name='django-nap', version='0.3', description='A light REST tool for Django', author='Curtis Maloney', author_email='curtis@tinbrain.net', url='http://github.com/funkybob/django-nap', keywords=['django', 'json', 'rest'], packages = find_packages(), zip_safe=False, classifiers = [ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], requires = [ 'Django (>=1.4)', ], )
from setuptools import setup, find_packages setup( name='django-nap', version='0.4', description='A light REST tool for Django', author='Curtis Maloney', author_email='curtis@tinbrain.net', url='http://github.com/funkybob/django-nap', keywords=['django', 'json', 'rest'], packages = find_packages(excludes=['test.*']), zip_safe=False, classifiers = [ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], requires = [ 'Django (>=1.4)', ], )
Fix find packages Bump version
Fix find packages Bump version
Python
bsd-3-clause
limbera/django-nap,MarkusH/django-nap
from setuptools import setup, find_packages setup( name='django-nap', version='0.3', description='A light REST tool for Django', author='Curtis Maloney', author_email='curtis@tinbrain.net', url='http://github.com/funkybob/django-nap', keywords=['django', 'json', 'rest'], packages = find_packages(), zip_safe=False, classifiers = [ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], requires = [ 'Django (>=1.4)', ], ) Fix find packages Bump version
from setuptools import setup, find_packages setup( name='django-nap', version='0.4', description='A light REST tool for Django', author='Curtis Maloney', author_email='curtis@tinbrain.net', url='http://github.com/funkybob/django-nap', keywords=['django', 'json', 'rest'], packages = find_packages(excludes=['test.*']), zip_safe=False, classifiers = [ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], requires = [ 'Django (>=1.4)', ], )
<commit_before>from setuptools import setup, find_packages setup( name='django-nap', version='0.3', description='A light REST tool for Django', author='Curtis Maloney', author_email='curtis@tinbrain.net', url='http://github.com/funkybob/django-nap', keywords=['django', 'json', 'rest'], packages = find_packages(), zip_safe=False, classifiers = [ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], requires = [ 'Django (>=1.4)', ], ) <commit_msg>Fix find packages Bump version<commit_after>
from setuptools import setup, find_packages setup( name='django-nap', version='0.4', description='A light REST tool for Django', author='Curtis Maloney', author_email='curtis@tinbrain.net', url='http://github.com/funkybob/django-nap', keywords=['django', 'json', 'rest'], packages = find_packages(excludes=['test.*']), zip_safe=False, classifiers = [ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], requires = [ 'Django (>=1.4)', ], )
from setuptools import setup, find_packages setup( name='django-nap', version='0.3', description='A light REST tool for Django', author='Curtis Maloney', author_email='curtis@tinbrain.net', url='http://github.com/funkybob/django-nap', keywords=['django', 'json', 'rest'], packages = find_packages(), zip_safe=False, classifiers = [ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], requires = [ 'Django (>=1.4)', ], ) Fix find packages Bump versionfrom setuptools import setup, find_packages setup( name='django-nap', version='0.4', description='A light REST tool for Django', author='Curtis Maloney', author_email='curtis@tinbrain.net', url='http://github.com/funkybob/django-nap', keywords=['django', 'json', 'rest'], packages = find_packages(excludes=['test.*']), zip_safe=False, classifiers = [ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], requires = [ 'Django (>=1.4)', ], )
<commit_before>from setuptools import setup, find_packages setup( name='django-nap', version='0.3', description='A light REST tool for Django', author='Curtis Maloney', author_email='curtis@tinbrain.net', url='http://github.com/funkybob/django-nap', keywords=['django', 'json', 'rest'], packages = find_packages(), zip_safe=False, classifiers = [ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], requires = [ 'Django (>=1.4)', ], ) <commit_msg>Fix find packages Bump version<commit_after>from setuptools import setup, find_packages setup( name='django-nap', version='0.4', description='A light REST tool for Django', author='Curtis Maloney', author_email='curtis@tinbrain.net', url='http://github.com/funkybob/django-nap', keywords=['django', 'json', 'rest'], packages = find_packages(excludes=['test.*']), zip_safe=False, classifiers = [ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], requires = [ 'Django (>=1.4)', ], )
1bb67543b4187ee05c616afb731c229aaa94fdd3
setup.py
setup.py
import os import sys from setuptools import setup, find_packages def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() requirements = ['httplib2', 'argparse', 'prettytable'] if sys.version_info < (2, 6): requirements.append('simplejson') setup( name = "python-keystoneclient", version = "2.7", description = "Client library for OpenStack Keystone API", long_description = read('README.rst'), url = 'https://github.com/4P/python-keystoneclient', license = 'Apache', author = 'Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss', author_email = 'gabriel.hurley@nebula.com', packages = find_packages(exclude=['tests', 'tests.*']), classifiers = [ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], install_requires = requirements, tests_require = ["nose", "mock", "mox"], test_suite = "nose.collector", entry_points = { 'console_scripts': ['keystone = keystoneclient.shell:main'] } )
import os import sys from setuptools import setup, find_packages def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() requirements = ['httplib2', 'argparse', 'prettytable'] if sys.version_info < (2, 6): requirements.append('simplejson') setup( name = "python-keystoneclient", version = "2012.1", description = "Client library for OpenStack Keystone API", long_description = read('README.rst'), url = 'https://github.com/openstack/python-keystoneclient', license = 'Apache', author = 'Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss', author_email = 'gabriel.hurley@nebula.com', packages = find_packages(exclude=['tests', 'tests.*']), classifiers = [ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], install_requires = requirements, tests_require = ["nose", "mock", "mox"], test_suite = "nose.collector", entry_points = { 'console_scripts': ['keystone = keystoneclient.shell:main'] } )
Adjust version number to match other deliveries
Adjust version number to match other deliveries Set version from 2.7 to 2012.1 to match the other OpenStack Keystone deliveries (python-keystoneclient will be released as part of Keystone 2012.1~e3). Also adjusted the location of the git repository to match new location. Fixes bug 917656. Change-Id: I4d8d071e3cdc5665e29a89067958f5f1e8964221
Python
apache-2.0
citrix-openstack-build/keystoneauth,jamielennox/keystoneauth,sileht/keystoneauth
import os import sys from setuptools import setup, find_packages def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() requirements = ['httplib2', 'argparse', 'prettytable'] if sys.version_info < (2, 6): requirements.append('simplejson') setup( name = "python-keystoneclient", version = "2.7", description = "Client library for OpenStack Keystone API", long_description = read('README.rst'), url = 'https://github.com/4P/python-keystoneclient', license = 'Apache', author = 'Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss', author_email = 'gabriel.hurley@nebula.com', packages = find_packages(exclude=['tests', 'tests.*']), classifiers = [ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], install_requires = requirements, tests_require = ["nose", "mock", "mox"], test_suite = "nose.collector", entry_points = { 'console_scripts': ['keystone = keystoneclient.shell:main'] } ) Adjust version number to match other deliveries Set version from 2.7 to 2012.1 to match the other OpenStack Keystone deliveries (python-keystoneclient will be released as part of Keystone 2012.1~e3). Also adjusted the location of the git repository to match new location. Fixes bug 917656. Change-Id: I4d8d071e3cdc5665e29a89067958f5f1e8964221
import os import sys from setuptools import setup, find_packages def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() requirements = ['httplib2', 'argparse', 'prettytable'] if sys.version_info < (2, 6): requirements.append('simplejson') setup( name = "python-keystoneclient", version = "2012.1", description = "Client library for OpenStack Keystone API", long_description = read('README.rst'), url = 'https://github.com/openstack/python-keystoneclient', license = 'Apache', author = 'Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss', author_email = 'gabriel.hurley@nebula.com', packages = find_packages(exclude=['tests', 'tests.*']), classifiers = [ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], install_requires = requirements, tests_require = ["nose", "mock", "mox"], test_suite = "nose.collector", entry_points = { 'console_scripts': ['keystone = keystoneclient.shell:main'] } )
<commit_before>import os import sys from setuptools import setup, find_packages def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() requirements = ['httplib2', 'argparse', 'prettytable'] if sys.version_info < (2, 6): requirements.append('simplejson') setup( name = "python-keystoneclient", version = "2.7", description = "Client library for OpenStack Keystone API", long_description = read('README.rst'), url = 'https://github.com/4P/python-keystoneclient', license = 'Apache', author = 'Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss', author_email = 'gabriel.hurley@nebula.com', packages = find_packages(exclude=['tests', 'tests.*']), classifiers = [ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], install_requires = requirements, tests_require = ["nose", "mock", "mox"], test_suite = "nose.collector", entry_points = { 'console_scripts': ['keystone = keystoneclient.shell:main'] } ) <commit_msg>Adjust version number to match other deliveries Set version from 2.7 to 2012.1 to match the other OpenStack Keystone deliveries (python-keystoneclient will be released as part of Keystone 2012.1~e3). Also adjusted the location of the git repository to match new location. Fixes bug 917656. Change-Id: I4d8d071e3cdc5665e29a89067958f5f1e8964221<commit_after>
import os import sys from setuptools import setup, find_packages def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() requirements = ['httplib2', 'argparse', 'prettytable'] if sys.version_info < (2, 6): requirements.append('simplejson') setup( name = "python-keystoneclient", version = "2012.1", description = "Client library for OpenStack Keystone API", long_description = read('README.rst'), url = 'https://github.com/openstack/python-keystoneclient', license = 'Apache', author = 'Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss', author_email = 'gabriel.hurley@nebula.com', packages = find_packages(exclude=['tests', 'tests.*']), classifiers = [ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], install_requires = requirements, tests_require = ["nose", "mock", "mox"], test_suite = "nose.collector", entry_points = { 'console_scripts': ['keystone = keystoneclient.shell:main'] } )
import os import sys from setuptools import setup, find_packages def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() requirements = ['httplib2', 'argparse', 'prettytable'] if sys.version_info < (2, 6): requirements.append('simplejson') setup( name = "python-keystoneclient", version = "2.7", description = "Client library for OpenStack Keystone API", long_description = read('README.rst'), url = 'https://github.com/4P/python-keystoneclient', license = 'Apache', author = 'Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss', author_email = 'gabriel.hurley@nebula.com', packages = find_packages(exclude=['tests', 'tests.*']), classifiers = [ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], install_requires = requirements, tests_require = ["nose", "mock", "mox"], test_suite = "nose.collector", entry_points = { 'console_scripts': ['keystone = keystoneclient.shell:main'] } ) Adjust version number to match other deliveries Set version from 2.7 to 2012.1 to match the other OpenStack Keystone deliveries (python-keystoneclient will be released as part of Keystone 2012.1~e3). Also adjusted the location of the git repository to match new location. Fixes bug 917656. Change-Id: I4d8d071e3cdc5665e29a89067958f5f1e8964221import os import sys from setuptools import setup, find_packages def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() requirements = ['httplib2', 'argparse', 'prettytable'] if sys.version_info < (2, 6): requirements.append('simplejson') setup( name = "python-keystoneclient", version = "2012.1", description = "Client library for OpenStack Keystone API", long_description = read('README.rst'), url = 'https://github.com/openstack/python-keystoneclient', license = 'Apache', author = 'Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss', author_email = 'gabriel.hurley@nebula.com', packages = find_packages(exclude=['tests', 'tests.*']), classifiers = [ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], install_requires = requirements, tests_require = ["nose", "mock", "mox"], test_suite = "nose.collector", entry_points = { 'console_scripts': ['keystone = keystoneclient.shell:main'] } )
<commit_before>import os import sys from setuptools import setup, find_packages def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() requirements = ['httplib2', 'argparse', 'prettytable'] if sys.version_info < (2, 6): requirements.append('simplejson') setup( name = "python-keystoneclient", version = "2.7", description = "Client library for OpenStack Keystone API", long_description = read('README.rst'), url = 'https://github.com/4P/python-keystoneclient', license = 'Apache', author = 'Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss', author_email = 'gabriel.hurley@nebula.com', packages = find_packages(exclude=['tests', 'tests.*']), classifiers = [ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], install_requires = requirements, tests_require = ["nose", "mock", "mox"], test_suite = "nose.collector", entry_points = { 'console_scripts': ['keystone = keystoneclient.shell:main'] } ) <commit_msg>Adjust version number to match other deliveries Set version from 2.7 to 2012.1 to match the other OpenStack Keystone deliveries (python-keystoneclient will be released as part of Keystone 2012.1~e3). Also adjusted the location of the git repository to match new location. Fixes bug 917656. Change-Id: I4d8d071e3cdc5665e29a89067958f5f1e8964221<commit_after>import os import sys from setuptools import setup, find_packages def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() requirements = ['httplib2', 'argparse', 'prettytable'] if sys.version_info < (2, 6): requirements.append('simplejson') setup( name = "python-keystoneclient", version = "2012.1", description = "Client library for OpenStack Keystone API", long_description = read('README.rst'), url = 'https://github.com/openstack/python-keystoneclient', license = 'Apache', author = 'Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss', author_email = 'gabriel.hurley@nebula.com', packages = find_packages(exclude=['tests', 'tests.*']), classifiers = [ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], install_requires = requirements, tests_require = ["nose", "mock", "mox"], test_suite = "nose.collector", entry_points = { 'console_scripts': ['keystone = keystoneclient.shell:main'] } )
710d41a1c1b328793224975ee5afc4aebe462f28
setup.py
setup.py
from setuptools import setup, find_packages import os version = '0.4' here = os.path.abspath(os.path.dirname(__file__)) try: README = open(os.path.join(here, 'README.txt')).read() CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read() except IOError: README = CHANGES = '' setup(name='tgext.admin', version=version, description="Admin Controller add-on for basic TG identity model.", long_description=README + "\n" + CHANGES, # Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers classifiers=[ "Programming Language :: Python", "Topic :: Software Development :: Libraries :: Python Modules", ], keywords='TG2, TG, sprox, Rest, internet, adminn', author='Christopher Perkins', author_email='chris@percious.com', url='tgtools.googlecode.com', license='MIT', packages=find_packages(exclude=['ez_setup']), namespace_packages=['tgext'], include_package_data=True, zip_safe=True, install_requires=[ 'setuptools', 'tgext.crud>=0.4', # -*- Extra requirements: -*- ], entry_points=""" # -*- Entry points: -*- """, )
from setuptools import setup, find_packages import os version = '0.5' here = os.path.abspath(os.path.dirname(__file__)) try: README = open(os.path.join(here, 'README.txt')).read() CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read() except IOError: README = CHANGES = '' setup(name='tgext.admin', version=version, description="Admin Controller add-on for basic TG identity model.", long_description=README + "\n" + CHANGES, # Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers classifiers=[ "Programming Language :: Python", "Topic :: Software Development :: Libraries :: Python Modules", ], keywords='turbogears2.extension, TG2, TG, sprox, Rest, internet, admin', author='Christopher Perkins', author_email='chris@percious.com', url='tgtools.googlecode.com', license='MIT', packages=find_packages(exclude=['ez_setup']), namespace_packages=['tgext'], include_package_data=True, zip_safe=True, install_requires=[ 'setuptools', 'tgext.crud>=0.4', # -*- Extra requirements: -*- ], entry_points=""" # -*- Entry points: -*- """, )
Raise version and change keywords for upcoming release
Raise version and change keywords for upcoming release
Python
mit
TurboGears/tgext.admin,TurboGears/tgext.admin
from setuptools import setup, find_packages import os version = '0.4' here = os.path.abspath(os.path.dirname(__file__)) try: README = open(os.path.join(here, 'README.txt')).read() CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read() except IOError: README = CHANGES = '' setup(name='tgext.admin', version=version, description="Admin Controller add-on for basic TG identity model.", long_description=README + "\n" + CHANGES, # Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers classifiers=[ "Programming Language :: Python", "Topic :: Software Development :: Libraries :: Python Modules", ], keywords='TG2, TG, sprox, Rest, internet, adminn', author='Christopher Perkins', author_email='chris@percious.com', url='tgtools.googlecode.com', license='MIT', packages=find_packages(exclude=['ez_setup']), namespace_packages=['tgext'], include_package_data=True, zip_safe=True, install_requires=[ 'setuptools', 'tgext.crud>=0.4', # -*- Extra requirements: -*- ], entry_points=""" # -*- Entry points: -*- """, ) Raise version and change keywords for upcoming release
from setuptools import setup, find_packages import os version = '0.5' here = os.path.abspath(os.path.dirname(__file__)) try: README = open(os.path.join(here, 'README.txt')).read() CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read() except IOError: README = CHANGES = '' setup(name='tgext.admin', version=version, description="Admin Controller add-on for basic TG identity model.", long_description=README + "\n" + CHANGES, # Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers classifiers=[ "Programming Language :: Python", "Topic :: Software Development :: Libraries :: Python Modules", ], keywords='turbogears2.extension, TG2, TG, sprox, Rest, internet, admin', author='Christopher Perkins', author_email='chris@percious.com', url='tgtools.googlecode.com', license='MIT', packages=find_packages(exclude=['ez_setup']), namespace_packages=['tgext'], include_package_data=True, zip_safe=True, install_requires=[ 'setuptools', 'tgext.crud>=0.4', # -*- Extra requirements: -*- ], entry_points=""" # -*- Entry points: -*- """, )
<commit_before>from setuptools import setup, find_packages import os version = '0.4' here = os.path.abspath(os.path.dirname(__file__)) try: README = open(os.path.join(here, 'README.txt')).read() CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read() except IOError: README = CHANGES = '' setup(name='tgext.admin', version=version, description="Admin Controller add-on for basic TG identity model.", long_description=README + "\n" + CHANGES, # Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers classifiers=[ "Programming Language :: Python", "Topic :: Software Development :: Libraries :: Python Modules", ], keywords='TG2, TG, sprox, Rest, internet, adminn', author='Christopher Perkins', author_email='chris@percious.com', url='tgtools.googlecode.com', license='MIT', packages=find_packages(exclude=['ez_setup']), namespace_packages=['tgext'], include_package_data=True, zip_safe=True, install_requires=[ 'setuptools', 'tgext.crud>=0.4', # -*- Extra requirements: -*- ], entry_points=""" # -*- Entry points: -*- """, ) <commit_msg>Raise version and change keywords for upcoming release<commit_after>
from setuptools import setup, find_packages import os version = '0.5' here = os.path.abspath(os.path.dirname(__file__)) try: README = open(os.path.join(here, 'README.txt')).read() CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read() except IOError: README = CHANGES = '' setup(name='tgext.admin', version=version, description="Admin Controller add-on for basic TG identity model.", long_description=README + "\n" + CHANGES, # Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers classifiers=[ "Programming Language :: Python", "Topic :: Software Development :: Libraries :: Python Modules", ], keywords='turbogears2.extension, TG2, TG, sprox, Rest, internet, admin', author='Christopher Perkins', author_email='chris@percious.com', url='tgtools.googlecode.com', license='MIT', packages=find_packages(exclude=['ez_setup']), namespace_packages=['tgext'], include_package_data=True, zip_safe=True, install_requires=[ 'setuptools', 'tgext.crud>=0.4', # -*- Extra requirements: -*- ], entry_points=""" # -*- Entry points: -*- """, )
from setuptools import setup, find_packages import os version = '0.4' here = os.path.abspath(os.path.dirname(__file__)) try: README = open(os.path.join(here, 'README.txt')).read() CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read() except IOError: README = CHANGES = '' setup(name='tgext.admin', version=version, description="Admin Controller add-on for basic TG identity model.", long_description=README + "\n" + CHANGES, # Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers classifiers=[ "Programming Language :: Python", "Topic :: Software Development :: Libraries :: Python Modules", ], keywords='TG2, TG, sprox, Rest, internet, adminn', author='Christopher Perkins', author_email='chris@percious.com', url='tgtools.googlecode.com', license='MIT', packages=find_packages(exclude=['ez_setup']), namespace_packages=['tgext'], include_package_data=True, zip_safe=True, install_requires=[ 'setuptools', 'tgext.crud>=0.4', # -*- Extra requirements: -*- ], entry_points=""" # -*- Entry points: -*- """, ) Raise version and change keywords for upcoming releasefrom setuptools import setup, find_packages import os version = '0.5' here = os.path.abspath(os.path.dirname(__file__)) try: README = open(os.path.join(here, 'README.txt')).read() CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read() except IOError: README = CHANGES = '' setup(name='tgext.admin', version=version, description="Admin Controller add-on for basic TG identity model.", long_description=README + "\n" + CHANGES, # Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers classifiers=[ "Programming Language :: Python", "Topic :: Software Development :: Libraries :: Python Modules", ], keywords='turbogears2.extension, TG2, TG, sprox, Rest, internet, admin', author='Christopher Perkins', author_email='chris@percious.com', url='tgtools.googlecode.com', license='MIT', packages=find_packages(exclude=['ez_setup']), namespace_packages=['tgext'], include_package_data=True, zip_safe=True, install_requires=[ 'setuptools', 'tgext.crud>=0.4', # -*- Extra requirements: -*- ], entry_points=""" # -*- Entry points: -*- """, )
<commit_before>from setuptools import setup, find_packages import os version = '0.4' here = os.path.abspath(os.path.dirname(__file__)) try: README = open(os.path.join(here, 'README.txt')).read() CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read() except IOError: README = CHANGES = '' setup(name='tgext.admin', version=version, description="Admin Controller add-on for basic TG identity model.", long_description=README + "\n" + CHANGES, # Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers classifiers=[ "Programming Language :: Python", "Topic :: Software Development :: Libraries :: Python Modules", ], keywords='TG2, TG, sprox, Rest, internet, adminn', author='Christopher Perkins', author_email='chris@percious.com', url='tgtools.googlecode.com', license='MIT', packages=find_packages(exclude=['ez_setup']), namespace_packages=['tgext'], include_package_data=True, zip_safe=True, install_requires=[ 'setuptools', 'tgext.crud>=0.4', # -*- Extra requirements: -*- ], entry_points=""" # -*- Entry points: -*- """, ) <commit_msg>Raise version and change keywords for upcoming release<commit_after>from setuptools import setup, find_packages import os version = '0.5' here = os.path.abspath(os.path.dirname(__file__)) try: README = open(os.path.join(here, 'README.txt')).read() CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read() except IOError: README = CHANGES = '' setup(name='tgext.admin', version=version, description="Admin Controller add-on for basic TG identity model.", long_description=README + "\n" + CHANGES, # Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers classifiers=[ "Programming Language :: Python", "Topic :: Software Development :: Libraries :: Python Modules", ], keywords='turbogears2.extension, TG2, TG, sprox, Rest, internet, admin', author='Christopher Perkins', author_email='chris@percious.com', url='tgtools.googlecode.com', license='MIT', packages=find_packages(exclude=['ez_setup']), namespace_packages=['tgext'], include_package_data=True, zip_safe=True, install_requires=[ 'setuptools', 'tgext.crud>=0.4', # -*- Extra requirements: -*- ], entry_points=""" # -*- Entry points: -*- """, )
c983c4dc27c547dab0c5f4afd5438231d26ce840
setup.py
setup.py
from setuptools import setup setup( name='ncdjango', description='A map server for NetCDF data', keywords='netcdf,django,map server', version='0.4.0', packages=[ 'ncdjango', 'ncdjango.geoprocessing', 'ncdjango.migrations', 'ncdjango.interfaces', 'ncdjango.interfaces.arcgis', 'ncdjango.interfaces.arcgis_extended', 'ncdjango.interfaces.data' ], install_requires=[ 'six', 'requests', 'Django>=1.7.0,<1.9.0', 'Pillow>=2.9.0', 'Shapely>=1.3.2', 'GDAL>=1.11.0', 'django-tastypie>=0.11.1', 'netCDF4>=1.1.6', 'numpy>=1.8.1', 'pyproj>=1.9.4', 'fiona', 'rasterio>=0.28.0', 'clover', 'ply>=3.8', 'celery>=3.1.19', 'djangorestframework' ], dependency_links=[ 'git+https://github.com/consbio/clover.git' ], url='https://github.com/consbio/ncdjango', license='BSD', )
from setuptools import setup setup( name='ncdjango', description='A map server for NetCDF data', keywords='netcdf,django,map server', version='0.4.0', packages=[ 'ncdjango', 'ncdjango.geoprocessing', 'ncdjango.migrations', 'ncdjango.interfaces', 'ncdjango.interfaces.arcgis', 'ncdjango.interfaces.arcgis_extended', 'ncdjango.interfaces.data' ], install_requires=[ 'six', 'requests', 'Django>=1.7.0,<1.9.0', 'Pillow>=2.9.0', 'Shapely>=1.3.2', 'django-tastypie>=0.11.1', 'netCDF4>=1.1.6', 'numpy>=1.8.1', 'pyproj>=1.9.4', 'fiona', 'rasterio>=0.28.0', 'clover', 'ply>=3.8', 'celery>=3.1.19', 'djangorestframework' ], dependency_links=[ 'git+https://github.com/consbio/clover.git' ], url='https://github.com/consbio/ncdjango', license='BSD', )
Remove GDAL requirement (it's not a direct dependency)
Remove GDAL requirement (it's not a direct dependency)
Python
bsd-3-clause
consbio/ncdjango,consbio/ncdjango
from setuptools import setup setup( name='ncdjango', description='A map server for NetCDF data', keywords='netcdf,django,map server', version='0.4.0', packages=[ 'ncdjango', 'ncdjango.geoprocessing', 'ncdjango.migrations', 'ncdjango.interfaces', 'ncdjango.interfaces.arcgis', 'ncdjango.interfaces.arcgis_extended', 'ncdjango.interfaces.data' ], install_requires=[ 'six', 'requests', 'Django>=1.7.0,<1.9.0', 'Pillow>=2.9.0', 'Shapely>=1.3.2', 'GDAL>=1.11.0', 'django-tastypie>=0.11.1', 'netCDF4>=1.1.6', 'numpy>=1.8.1', 'pyproj>=1.9.4', 'fiona', 'rasterio>=0.28.0', 'clover', 'ply>=3.8', 'celery>=3.1.19', 'djangorestframework' ], dependency_links=[ 'git+https://github.com/consbio/clover.git' ], url='https://github.com/consbio/ncdjango', license='BSD', ) Remove GDAL requirement (it's not a direct dependency)
from setuptools import setup setup( name='ncdjango', description='A map server for NetCDF data', keywords='netcdf,django,map server', version='0.4.0', packages=[ 'ncdjango', 'ncdjango.geoprocessing', 'ncdjango.migrations', 'ncdjango.interfaces', 'ncdjango.interfaces.arcgis', 'ncdjango.interfaces.arcgis_extended', 'ncdjango.interfaces.data' ], install_requires=[ 'six', 'requests', 'Django>=1.7.0,<1.9.0', 'Pillow>=2.9.0', 'Shapely>=1.3.2', 'django-tastypie>=0.11.1', 'netCDF4>=1.1.6', 'numpy>=1.8.1', 'pyproj>=1.9.4', 'fiona', 'rasterio>=0.28.0', 'clover', 'ply>=3.8', 'celery>=3.1.19', 'djangorestframework' ], dependency_links=[ 'git+https://github.com/consbio/clover.git' ], url='https://github.com/consbio/ncdjango', license='BSD', )
<commit_before>from setuptools import setup setup( name='ncdjango', description='A map server for NetCDF data', keywords='netcdf,django,map server', version='0.4.0', packages=[ 'ncdjango', 'ncdjango.geoprocessing', 'ncdjango.migrations', 'ncdjango.interfaces', 'ncdjango.interfaces.arcgis', 'ncdjango.interfaces.arcgis_extended', 'ncdjango.interfaces.data' ], install_requires=[ 'six', 'requests', 'Django>=1.7.0,<1.9.0', 'Pillow>=2.9.0', 'Shapely>=1.3.2', 'GDAL>=1.11.0', 'django-tastypie>=0.11.1', 'netCDF4>=1.1.6', 'numpy>=1.8.1', 'pyproj>=1.9.4', 'fiona', 'rasterio>=0.28.0', 'clover', 'ply>=3.8', 'celery>=3.1.19', 'djangorestframework' ], dependency_links=[ 'git+https://github.com/consbio/clover.git' ], url='https://github.com/consbio/ncdjango', license='BSD', ) <commit_msg>Remove GDAL requirement (it's not a direct dependency)<commit_after>
from setuptools import setup setup( name='ncdjango', description='A map server for NetCDF data', keywords='netcdf,django,map server', version='0.4.0', packages=[ 'ncdjango', 'ncdjango.geoprocessing', 'ncdjango.migrations', 'ncdjango.interfaces', 'ncdjango.interfaces.arcgis', 'ncdjango.interfaces.arcgis_extended', 'ncdjango.interfaces.data' ], install_requires=[ 'six', 'requests', 'Django>=1.7.0,<1.9.0', 'Pillow>=2.9.0', 'Shapely>=1.3.2', 'django-tastypie>=0.11.1', 'netCDF4>=1.1.6', 'numpy>=1.8.1', 'pyproj>=1.9.4', 'fiona', 'rasterio>=0.28.0', 'clover', 'ply>=3.8', 'celery>=3.1.19', 'djangorestframework' ], dependency_links=[ 'git+https://github.com/consbio/clover.git' ], url='https://github.com/consbio/ncdjango', license='BSD', )
from setuptools import setup setup( name='ncdjango', description='A map server for NetCDF data', keywords='netcdf,django,map server', version='0.4.0', packages=[ 'ncdjango', 'ncdjango.geoprocessing', 'ncdjango.migrations', 'ncdjango.interfaces', 'ncdjango.interfaces.arcgis', 'ncdjango.interfaces.arcgis_extended', 'ncdjango.interfaces.data' ], install_requires=[ 'six', 'requests', 'Django>=1.7.0,<1.9.0', 'Pillow>=2.9.0', 'Shapely>=1.3.2', 'GDAL>=1.11.0', 'django-tastypie>=0.11.1', 'netCDF4>=1.1.6', 'numpy>=1.8.1', 'pyproj>=1.9.4', 'fiona', 'rasterio>=0.28.0', 'clover', 'ply>=3.8', 'celery>=3.1.19', 'djangorestframework' ], dependency_links=[ 'git+https://github.com/consbio/clover.git' ], url='https://github.com/consbio/ncdjango', license='BSD', ) Remove GDAL requirement (it's not a direct dependency)from setuptools import setup setup( name='ncdjango', description='A map server for NetCDF data', keywords='netcdf,django,map server', version='0.4.0', packages=[ 'ncdjango', 'ncdjango.geoprocessing', 'ncdjango.migrations', 'ncdjango.interfaces', 'ncdjango.interfaces.arcgis', 'ncdjango.interfaces.arcgis_extended', 'ncdjango.interfaces.data' ], install_requires=[ 'six', 'requests', 'Django>=1.7.0,<1.9.0', 'Pillow>=2.9.0', 'Shapely>=1.3.2', 'django-tastypie>=0.11.1', 'netCDF4>=1.1.6', 'numpy>=1.8.1', 'pyproj>=1.9.4', 'fiona', 'rasterio>=0.28.0', 'clover', 'ply>=3.8', 'celery>=3.1.19', 'djangorestframework' ], dependency_links=[ 'git+https://github.com/consbio/clover.git' ], url='https://github.com/consbio/ncdjango', license='BSD', )
<commit_before>from setuptools import setup setup( name='ncdjango', description='A map server for NetCDF data', keywords='netcdf,django,map server', version='0.4.0', packages=[ 'ncdjango', 'ncdjango.geoprocessing', 'ncdjango.migrations', 'ncdjango.interfaces', 'ncdjango.interfaces.arcgis', 'ncdjango.interfaces.arcgis_extended', 'ncdjango.interfaces.data' ], install_requires=[ 'six', 'requests', 'Django>=1.7.0,<1.9.0', 'Pillow>=2.9.0', 'Shapely>=1.3.2', 'GDAL>=1.11.0', 'django-tastypie>=0.11.1', 'netCDF4>=1.1.6', 'numpy>=1.8.1', 'pyproj>=1.9.4', 'fiona', 'rasterio>=0.28.0', 'clover', 'ply>=3.8', 'celery>=3.1.19', 'djangorestframework' ], dependency_links=[ 'git+https://github.com/consbio/clover.git' ], url='https://github.com/consbio/ncdjango', license='BSD', ) <commit_msg>Remove GDAL requirement (it's not a direct dependency)<commit_after>from setuptools import setup setup( name='ncdjango', description='A map server for NetCDF data', keywords='netcdf,django,map server', version='0.4.0', packages=[ 'ncdjango', 'ncdjango.geoprocessing', 'ncdjango.migrations', 'ncdjango.interfaces', 'ncdjango.interfaces.arcgis', 'ncdjango.interfaces.arcgis_extended', 'ncdjango.interfaces.data' ], install_requires=[ 'six', 'requests', 'Django>=1.7.0,<1.9.0', 'Pillow>=2.9.0', 'Shapely>=1.3.2', 'django-tastypie>=0.11.1', 'netCDF4>=1.1.6', 'numpy>=1.8.1', 'pyproj>=1.9.4', 'fiona', 'rasterio>=0.28.0', 'clover', 'ply>=3.8', 'celery>=3.1.19', 'djangorestframework' ], dependency_links=[ 'git+https://github.com/consbio/clover.git' ], url='https://github.com/consbio/ncdjango', license='BSD', )
0a33b7d8df544226df711db33a27f45421c19290
setup.py
setup.py
from setuptools import setup version = '2.0.0' setup(name='pyactiveresource', version=version, description='ActiveResource for Python', author='Shopify', author_email='developers@shopify.com', url='https://github.com/Shopify/pyactiveresource/', packages=['pyactiveresource', 'pyactiveresource/testing'], license='MIT License', test_suite='test', tests_require=[ 'python-dateutil<2.0', # >= 2.0 is for python>=3.0 'PyYAML', ], platforms=['any'], classifiers=['Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Software Development', 'Topic :: Software Development :: Libraries', 'Topic :: Software Development :: Libraries :: Python Modules'] )
from setuptools import setup import sys version = '2.0.0' if sys.version_info >= (3,): python_dateutils_version = 'python-dateutil>=2.0' else: python_dateutils_version = 'python-dateutil<2.0' setup(name='pyactiveresource', version=version, description='ActiveResource for Python', author='Shopify', author_email='developers@shopify.com', url='https://github.com/Shopify/pyactiveresource/', packages=['pyactiveresource', 'pyactiveresource/testing'], license='MIT License', test_suite='test', tests_require=[ python_dateutils_version, 'PyYAML', ], platforms=['any'], classifiers=['Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Software Development', 'Topic :: Software Development :: Libraries', 'Topic :: Software Development :: Libraries :: Python Modules'] )
Use the right version of python-dateutils when using python 3.
Use the right version of python-dateutils when using python 3.
Python
mit
metric-collective/pyactiveresource,piran/pyactiveresource,varesa/pyactiveresource,hockeybuggy/pyactiveresource
from setuptools import setup version = '2.0.0' setup(name='pyactiveresource', version=version, description='ActiveResource for Python', author='Shopify', author_email='developers@shopify.com', url='https://github.com/Shopify/pyactiveresource/', packages=['pyactiveresource', 'pyactiveresource/testing'], license='MIT License', test_suite='test', tests_require=[ 'python-dateutil<2.0', # >= 2.0 is for python>=3.0 'PyYAML', ], platforms=['any'], classifiers=['Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Software Development', 'Topic :: Software Development :: Libraries', 'Topic :: Software Development :: Libraries :: Python Modules'] ) Use the right version of python-dateutils when using python 3.
from setuptools import setup import sys version = '2.0.0' if sys.version_info >= (3,): python_dateutils_version = 'python-dateutil>=2.0' else: python_dateutils_version = 'python-dateutil<2.0' setup(name='pyactiveresource', version=version, description='ActiveResource for Python', author='Shopify', author_email='developers@shopify.com', url='https://github.com/Shopify/pyactiveresource/', packages=['pyactiveresource', 'pyactiveresource/testing'], license='MIT License', test_suite='test', tests_require=[ python_dateutils_version, 'PyYAML', ], platforms=['any'], classifiers=['Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Software Development', 'Topic :: Software Development :: Libraries', 'Topic :: Software Development :: Libraries :: Python Modules'] )
<commit_before>from setuptools import setup version = '2.0.0' setup(name='pyactiveresource', version=version, description='ActiveResource for Python', author='Shopify', author_email='developers@shopify.com', url='https://github.com/Shopify/pyactiveresource/', packages=['pyactiveresource', 'pyactiveresource/testing'], license='MIT License', test_suite='test', tests_require=[ 'python-dateutil<2.0', # >= 2.0 is for python>=3.0 'PyYAML', ], platforms=['any'], classifiers=['Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Software Development', 'Topic :: Software Development :: Libraries', 'Topic :: Software Development :: Libraries :: Python Modules'] ) <commit_msg>Use the right version of python-dateutils when using python 3.<commit_after>
from setuptools import setup import sys version = '2.0.0' if sys.version_info >= (3,): python_dateutils_version = 'python-dateutil>=2.0' else: python_dateutils_version = 'python-dateutil<2.0' setup(name='pyactiveresource', version=version, description='ActiveResource for Python', author='Shopify', author_email='developers@shopify.com', url='https://github.com/Shopify/pyactiveresource/', packages=['pyactiveresource', 'pyactiveresource/testing'], license='MIT License', test_suite='test', tests_require=[ python_dateutils_version, 'PyYAML', ], platforms=['any'], classifiers=['Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Software Development', 'Topic :: Software Development :: Libraries', 'Topic :: Software Development :: Libraries :: Python Modules'] )
from setuptools import setup version = '2.0.0' setup(name='pyactiveresource', version=version, description='ActiveResource for Python', author='Shopify', author_email='developers@shopify.com', url='https://github.com/Shopify/pyactiveresource/', packages=['pyactiveresource', 'pyactiveresource/testing'], license='MIT License', test_suite='test', tests_require=[ 'python-dateutil<2.0', # >= 2.0 is for python>=3.0 'PyYAML', ], platforms=['any'], classifiers=['Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Software Development', 'Topic :: Software Development :: Libraries', 'Topic :: Software Development :: Libraries :: Python Modules'] ) Use the right version of python-dateutils when using python 3.from setuptools import setup import sys version = '2.0.0' if sys.version_info >= (3,): python_dateutils_version = 'python-dateutil>=2.0' else: python_dateutils_version = 'python-dateutil<2.0' setup(name='pyactiveresource', version=version, description='ActiveResource for Python', author='Shopify', author_email='developers@shopify.com', url='https://github.com/Shopify/pyactiveresource/', packages=['pyactiveresource', 'pyactiveresource/testing'], license='MIT License', test_suite='test', tests_require=[ python_dateutils_version, 'PyYAML', ], platforms=['any'], classifiers=['Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Software Development', 'Topic :: Software Development :: Libraries', 'Topic :: Software Development :: Libraries :: Python Modules'] )
<commit_before>from setuptools import setup version = '2.0.0' setup(name='pyactiveresource', version=version, description='ActiveResource for Python', author='Shopify', author_email='developers@shopify.com', url='https://github.com/Shopify/pyactiveresource/', packages=['pyactiveresource', 'pyactiveresource/testing'], license='MIT License', test_suite='test', tests_require=[ 'python-dateutil<2.0', # >= 2.0 is for python>=3.0 'PyYAML', ], platforms=['any'], classifiers=['Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Software Development', 'Topic :: Software Development :: Libraries', 'Topic :: Software Development :: Libraries :: Python Modules'] ) <commit_msg>Use the right version of python-dateutils when using python 3.<commit_after>from setuptools import setup import sys version = '2.0.0' if sys.version_info >= (3,): python_dateutils_version = 'python-dateutil>=2.0' else: python_dateutils_version = 'python-dateutil<2.0' setup(name='pyactiveresource', version=version, description='ActiveResource for Python', author='Shopify', author_email='developers@shopify.com', url='https://github.com/Shopify/pyactiveresource/', packages=['pyactiveresource', 'pyactiveresource/testing'], license='MIT License', test_suite='test', tests_require=[ python_dateutils_version, 'PyYAML', ], platforms=['any'], classifiers=['Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Software Development', 'Topic :: Software Development :: Libraries', 'Topic :: Software Development :: Libraries :: Python Modules'] )
973473bd7ba0e19e8e86492d4e15b272c441b278
setup.py
setup.py
"""Setuptools configuration for engine.""" from setuptools import setup from setuptools import find_packages with open('README.rst', 'r') as readmefile: README = readmefile.read() setup( name='asyncdef.engine', version='0.1.0', url='https://github.com/asyncdef/engine', description='Core event loop implementation.', author="Kevin Conway", author_email="kevinjacobconway@gmail.com", long_description=README, license='Apache 2.0', packages=find_packages(exclude=['build', 'dist', 'docs']), install_requires=[ 'iface<2.0.0', 'asyncdef.interfaces<2.0.0', ], extras_require={ 'testing': [ 'pep257', 'pep8', 'pyenchant', 'pyflakes', 'pylint', 'pytest', 'pytest-cov', ], }, entry_points={ 'console_scripts': [ ], }, include_package_data=True, zip_safe=False, )
"""Setuptools configuration for engine.""" from setuptools import setup from setuptools import find_packages with open('README.rst', 'r') as readmefile: README = readmefile.read() setup( name='asyncdef.engine', version='0.1.0', url='https://github.com/asyncdef/engine', description='Core event loop implementation.', author="Kevin Conway", author_email="kevinjacobconway@gmail.com", long_description=README, license='Apache 2.0', packages=[ 'asyncdef', 'asyncdef.engine', 'asyncdef.engine.processors', ], install_requires=[ 'iface<2.0.0', 'asyncdef.interfaces<2.0.0', ], extras_require={ 'testing': [ 'pep257', 'pep8', 'pyenchant', 'pyflakes', 'pylint', 'pytest', 'pytest-cov', ], }, entry_points={ 'console_scripts': [ ], }, include_package_data=True, zip_safe=False, )
Swap find_packages for a manual list
Swap find_packages for a manual list This is required to support PEP420 namespace packages.
Python
apache-2.0
asyncdef/engine
"""Setuptools configuration for engine.""" from setuptools import setup from setuptools import find_packages with open('README.rst', 'r') as readmefile: README = readmefile.read() setup( name='asyncdef.engine', version='0.1.0', url='https://github.com/asyncdef/engine', description='Core event loop implementation.', author="Kevin Conway", author_email="kevinjacobconway@gmail.com", long_description=README, license='Apache 2.0', packages=find_packages(exclude=['build', 'dist', 'docs']), install_requires=[ 'iface<2.0.0', 'asyncdef.interfaces<2.0.0', ], extras_require={ 'testing': [ 'pep257', 'pep8', 'pyenchant', 'pyflakes', 'pylint', 'pytest', 'pytest-cov', ], }, entry_points={ 'console_scripts': [ ], }, include_package_data=True, zip_safe=False, ) Swap find_packages for a manual list This is required to support PEP420 namespace packages.
"""Setuptools configuration for engine.""" from setuptools import setup from setuptools import find_packages with open('README.rst', 'r') as readmefile: README = readmefile.read() setup( name='asyncdef.engine', version='0.1.0', url='https://github.com/asyncdef/engine', description='Core event loop implementation.', author="Kevin Conway", author_email="kevinjacobconway@gmail.com", long_description=README, license='Apache 2.0', packages=[ 'asyncdef', 'asyncdef.engine', 'asyncdef.engine.processors', ], install_requires=[ 'iface<2.0.0', 'asyncdef.interfaces<2.0.0', ], extras_require={ 'testing': [ 'pep257', 'pep8', 'pyenchant', 'pyflakes', 'pylint', 'pytest', 'pytest-cov', ], }, entry_points={ 'console_scripts': [ ], }, include_package_data=True, zip_safe=False, )
<commit_before>"""Setuptools configuration for engine.""" from setuptools import setup from setuptools import find_packages with open('README.rst', 'r') as readmefile: README = readmefile.read() setup( name='asyncdef.engine', version='0.1.0', url='https://github.com/asyncdef/engine', description='Core event loop implementation.', author="Kevin Conway", author_email="kevinjacobconway@gmail.com", long_description=README, license='Apache 2.0', packages=find_packages(exclude=['build', 'dist', 'docs']), install_requires=[ 'iface<2.0.0', 'asyncdef.interfaces<2.0.0', ], extras_require={ 'testing': [ 'pep257', 'pep8', 'pyenchant', 'pyflakes', 'pylint', 'pytest', 'pytest-cov', ], }, entry_points={ 'console_scripts': [ ], }, include_package_data=True, zip_safe=False, ) <commit_msg>Swap find_packages for a manual list This is required to support PEP420 namespace packages.<commit_after>
"""Setuptools configuration for engine.""" from setuptools import setup from setuptools import find_packages with open('README.rst', 'r') as readmefile: README = readmefile.read() setup( name='asyncdef.engine', version='0.1.0', url='https://github.com/asyncdef/engine', description='Core event loop implementation.', author="Kevin Conway", author_email="kevinjacobconway@gmail.com", long_description=README, license='Apache 2.0', packages=[ 'asyncdef', 'asyncdef.engine', 'asyncdef.engine.processors', ], install_requires=[ 'iface<2.0.0', 'asyncdef.interfaces<2.0.0', ], extras_require={ 'testing': [ 'pep257', 'pep8', 'pyenchant', 'pyflakes', 'pylint', 'pytest', 'pytest-cov', ], }, entry_points={ 'console_scripts': [ ], }, include_package_data=True, zip_safe=False, )
"""Setuptools configuration for engine.""" from setuptools import setup from setuptools import find_packages with open('README.rst', 'r') as readmefile: README = readmefile.read() setup( name='asyncdef.engine', version='0.1.0', url='https://github.com/asyncdef/engine', description='Core event loop implementation.', author="Kevin Conway", author_email="kevinjacobconway@gmail.com", long_description=README, license='Apache 2.0', packages=find_packages(exclude=['build', 'dist', 'docs']), install_requires=[ 'iface<2.0.0', 'asyncdef.interfaces<2.0.0', ], extras_require={ 'testing': [ 'pep257', 'pep8', 'pyenchant', 'pyflakes', 'pylint', 'pytest', 'pytest-cov', ], }, entry_points={ 'console_scripts': [ ], }, include_package_data=True, zip_safe=False, ) Swap find_packages for a manual list This is required to support PEP420 namespace packages."""Setuptools configuration for engine.""" from setuptools import setup from setuptools import find_packages with open('README.rst', 'r') as readmefile: README = readmefile.read() setup( name='asyncdef.engine', version='0.1.0', url='https://github.com/asyncdef/engine', description='Core event loop implementation.', author="Kevin Conway", author_email="kevinjacobconway@gmail.com", long_description=README, license='Apache 2.0', packages=[ 'asyncdef', 'asyncdef.engine', 'asyncdef.engine.processors', ], install_requires=[ 'iface<2.0.0', 'asyncdef.interfaces<2.0.0', ], extras_require={ 'testing': [ 'pep257', 'pep8', 'pyenchant', 'pyflakes', 'pylint', 'pytest', 'pytest-cov', ], }, entry_points={ 'console_scripts': [ ], }, include_package_data=True, zip_safe=False, )
<commit_before>"""Setuptools configuration for engine.""" from setuptools import setup from setuptools import find_packages with open('README.rst', 'r') as readmefile: README = readmefile.read() setup( name='asyncdef.engine', version='0.1.0', url='https://github.com/asyncdef/engine', description='Core event loop implementation.', author="Kevin Conway", author_email="kevinjacobconway@gmail.com", long_description=README, license='Apache 2.0', packages=find_packages(exclude=['build', 'dist', 'docs']), install_requires=[ 'iface<2.0.0', 'asyncdef.interfaces<2.0.0', ], extras_require={ 'testing': [ 'pep257', 'pep8', 'pyenchant', 'pyflakes', 'pylint', 'pytest', 'pytest-cov', ], }, entry_points={ 'console_scripts': [ ], }, include_package_data=True, zip_safe=False, ) <commit_msg>Swap find_packages for a manual list This is required to support PEP420 namespace packages.<commit_after>"""Setuptools configuration for engine.""" from setuptools import setup from setuptools import find_packages with open('README.rst', 'r') as readmefile: README = readmefile.read() setup( name='asyncdef.engine', version='0.1.0', url='https://github.com/asyncdef/engine', description='Core event loop implementation.', author="Kevin Conway", author_email="kevinjacobconway@gmail.com", long_description=README, license='Apache 2.0', packages=[ 'asyncdef', 'asyncdef.engine', 'asyncdef.engine.processors', ], install_requires=[ 'iface<2.0.0', 'asyncdef.interfaces<2.0.0', ], extras_require={ 'testing': [ 'pep257', 'pep8', 'pyenchant', 'pyflakes', 'pylint', 'pytest', 'pytest-cov', ], }, entry_points={ 'console_scripts': [ ], }, include_package_data=True, zip_safe=False, )
5acb2258164ce5f8d7c8c297b32534547e861450
lms/djangoapps/debug/views.py
lms/djangoapps/debug/views.py
"""Views for debugging and diagnostics""" import pprint from django.http import Http404 from django.contrib.auth.decorators import login_required from django_future.csrf import ensure_csrf_cookie, csrf_exempt from mitxmako.shortcuts import render_to_response from codejail.safe_exec import safe_exec @login_required @ensure_csrf_cookie def run_python(request): if not request.user.is_staff: raise Http404 c = {} c['code'] = '' c['results'] = None if request.method == 'POST': py_code = c['code'] = request.POST.get('code') g = {} try: safe_exec(py_code, g) except Exception as e: c['results'] = str(e) else: c['results'] = pprint.pformat(g) return render_to_response("debug/run_python_form.html", c)
"""Views for debugging and diagnostics""" import pprint import traceback from django.http import Http404 from django.contrib.auth.decorators import login_required from django_future.csrf import ensure_csrf_cookie, csrf_exempt from mitxmako.shortcuts import render_to_response from codejail.safe_exec import safe_exec @login_required @ensure_csrf_cookie def run_python(request): """A page to allow testing the Python sandbox on a production server.""" if not request.user.is_staff: raise Http404 c = {} c['code'] = '' c['results'] = None if request.method == 'POST': py_code = c['code'] = request.POST.get('code') g = {} try: safe_exec(py_code, g) except Exception as e: c['results'] = traceback.format_exc() else: c['results'] = pprint.pformat(g) return render_to_response("debug/run_python_form.html", c)
Print the full traceback when execution fails.
Print the full traceback when execution fails.
Python
agpl-3.0
cpennington/edx-platform,Shrhawk/edx-platform,chrisndodge/edx-platform,xinjiguaike/edx-platform,leansoft/edx-platform,jamiefolsom/edx-platform,chand3040/cloud_that,nttks/jenkins-test,PepperPD/edx-pepper-platform,edry/edx-platform,naresh21/synergetics-edx-platform,doganov/edx-platform,adoosii/edx-platform,eduNEXT/edx-platform,analyseuc3m/ANALYSE-v1,shubhdev/edxOnBaadal,vikas1885/test1,motion2015/edx-platform,ubc/edx-platform,benpatterson/edx-platform,JioEducation/edx-platform,praveen-pal/edx-platform,olexiim/edx-platform,xuxiao19910803/edx,carsongee/edx-platform,shurihell/testasia,lduarte1991/edx-platform,waheedahmed/edx-platform,kxliugang/edx-platform,dsajkl/123,Lektorium-LLC/edx-platform,ferabra/edx-platform,nanolearningllc/edx-platform-cypress-2,nikolas/edx-platform,amir-qayyum-khan/edx-platform,torchingloom/edx-platform,TeachAtTUM/edx-platform,Semi-global/edx-platform,mushtaqak/edx-platform,mcgachey/edx-platform,ampax/edx-platform,OmarIthawi/edx-platform,10clouds/edx-platform,nanolearningllc/edx-platform-cypress,jamiefolsom/edx-platform,procangroup/edx-platform,Edraak/edx-platform,shashank971/edx-platform,yokose-ks/edx-platform,hkawasaki/kawasaki-aio8-1,simbs/edx-platform,jbzdak/edx-platform,nttks/edx-platform,utecuy/edx-platform,jazztpt/edx-platform,msegado/edx-platform,philanthropy-u/edx-platform,ahmedaljazzar/edx-platform,DefyVentures/edx-platform,ZLLab-Mooc/edx-platform,hkawasaki/kawasaki-aio8-2,hamzehd/edx-platform,jswope00/griffinx,cecep-edu/edx-platform,chand3040/cloud_that,bigdatauniversity/edx-platform,shubhdev/edx-platform,OmarIthawi/edx-platform,zerobatu/edx-platform,auferack08/edx-platform,Shrhawk/edx-platform,alexthered/kienhoc-platform,jazkarta/edx-platform,LICEF/edx-platform,waheedahmed/edx-platform,amir-qayyum-khan/edx-platform,wwj718/ANALYSE,adoosii/edx-platform,sudheerchintala/LearnEraPlatForm,etzhou/edx-platform,eestay/edx-platform,zerobatu/edx-platform,Edraak/edraak-platform,ferabra/edx-platform,tiagochiavericosta/edx-platform,OmarIthawi/edx-platform,vismartltd/edx-platform,EduPepperPDTesting/pepper2013-testing,utecuy/edx-platform,nanolearningllc/edx-platform-cypress,4eek/edx-platform,zadgroup/edx-platform,stvstnfrd/edx-platform,solashirai/edx-platform,caesar2164/edx-platform,CourseTalk/edx-platform,Livit/Livit.Learn.EdX,EduPepperPD/pepper2013,eduNEXT/edunext-platform,rue89-tech/edx-platform,pepeportela/edx-platform,antoviaque/edx-platform,J861449197/edx-platform,raccoongang/edx-platform,cselis86/edx-platform,MSOpenTech/edx-platform,zerobatu/edx-platform,eduNEXT/edx-platform,eduNEXT/edunext-platform,synergeticsedx/deployment-wipro,nttks/edx-platform,olexiim/edx-platform,nagyistoce/edx-platform,morpheby/levelup-by,tanmaykm/edx-platform,torchingloom/edx-platform,gymnasium/edx-platform,gsehub/edx-platform,chand3040/cloud_that,cognitiveclass/edx-platform,DefyVentures/edx-platform,fly19890211/edx-platform,jruiperezv/ANALYSE,jazkarta/edx-platform,mjirayu/sit_academy,hmcmooc/muddx-platform,franosincic/edx-platform,shashank971/edx-platform,rhndg/openedx,tanmaykm/edx-platform,mbareta/edx-platform-ft,auferack08/edx-platform,eestay/edx-platform,cognitiveclass/edx-platform,carsongee/edx-platform,martynovp/edx-platform,zadgroup/edx-platform,bdero/edx-platform,Unow/edx-platform,Livit/Livit.Learn.EdX,hkawasaki/kawasaki-aio8-1,beni55/edx-platform,B-MOOC/edx-platform,rationalAgent/edx-platform-custom,playm2mboy/edx-platform,fintech-circle/edx-platform,edry/edx-platform,Edraak/edx-platform,devs1991/test_edx_docmode,atsolakid/edx-platform,morenopc/edx-platform,Lektorium-LLC/edx-platform,proversity-org/edx-platform,rue89-tech/edx-platform,knehez/edx-platform,zerobatu/edx-platform,ahmadio/edx-platform,mahendra-r/edx-platform,don-github/edx-platform,Edraak/circleci-edx-platform,LearnEra/LearnEraPlaftform,ampax/edx-platform-backup,zubair-arbi/edx-platform,UXE/local-edx,olexiim/edx-platform,defance/edx-platform,prarthitm/edxplatform,franosincic/edx-platform,morpheby/levelup-by,defance/edx-platform,shashank971/edx-platform,miptliot/edx-platform,eemirtekin/edx-platform,martynovp/edx-platform,shabab12/edx-platform,Kalyzee/edx-platform,eduNEXT/edx-platform,10clouds/edx-platform,sameetb-cuelogic/edx-platform-test,arifsetiawan/edx-platform,caesar2164/edx-platform,zofuthan/edx-platform,beacloudgenius/edx-platform,IndonesiaX/edx-platform,defance/edx-platform,vasyarv/edx-platform,LICEF/edx-platform,nttks/edx-platform,RPI-OPENEDX/edx-platform,edx-solutions/edx-platform,eemirtekin/edx-platform,xuxiao19910803/edx,pepeportela/edx-platform,Endika/edx-platform,arifsetiawan/edx-platform,leansoft/edx-platform,antonve/s4-project-mooc,yokose-ks/edx-platform,nanolearning/edx-platform,EduPepperPD/pepper2013,DNFcode/edx-platform,alexthered/kienhoc-platform,nikolas/edx-platform,chauhanhardik/populo_2,itsjeyd/edx-platform,procangroup/edx-platform,devs1991/test_edx_docmode,dkarakats/edx-platform,PepperPD/edx-pepper-platform,yokose-ks/edx-platform,ESOedX/edx-platform,solashirai/edx-platform,pabloborrego93/edx-platform,playm2mboy/edx-platform,peterm-itr/edx-platform,doismellburning/edx-platform,valtech-mooc/edx-platform,DNFcode/edx-platform,lduarte1991/edx-platform,mjirayu/sit_academy,ahmadiga/min_edx,BehavioralInsightsTeam/edx-platform,nanolearningllc/edx-platform-cypress-2,motion2015/a3,ampax/edx-platform-backup,ovnicraft/edx-platform,jswope00/GAI,atsolakid/edx-platform,zofuthan/edx-platform,tiagochiavericosta/edx-platform,jbzdak/edx-platform,xuxiao19910803/edx-platform,jolyonb/edx-platform,fintech-circle/edx-platform,RPI-OPENEDX/edx-platform,IONISx/edx-platform,J861449197/edx-platform,antonve/s4-project-mooc,ovnicraft/edx-platform,chand3040/cloud_that,mahendra-r/edx-platform,marcore/edx-platform,mitocw/edx-platform,hkawasaki/kawasaki-aio8-0,Unow/edx-platform,jruiperezv/ANALYSE,ubc/edx-platform,longmen21/edx-platform,shurihell/testasia,wwj718/ANALYSE,appsembler/edx-platform,morenopc/edx-platform,bigdatauniversity/edx-platform,syjeon/new_edx,eemirtekin/edx-platform,ahmadio/edx-platform,chauhanhardik/populo,fintech-circle/edx-platform,IITBinterns13/edx-platform-dev,kalebhartje/schoolboost,jamesblunt/edx-platform,nttks/edx-platform,bdero/edx-platform,B-MOOC/edx-platform,don-github/edx-platform,MakeHer/edx-platform,arbrandes/edx-platform,IndonesiaX/edx-platform,simbs/edx-platform,apigee/edx-platform,teltek/edx-platform,shubhdev/openedx,zofuthan/edx-platform,lduarte1991/edx-platform,AkA84/edx-platform,martynovp/edx-platform,Softmotions/edx-platform,defance/edx-platform,miptliot/edx-platform,SravanthiSinha/edx-platform,nikolas/edx-platform,kursitet/edx-platform,nanolearningllc/edx-platform-cypress-2,cyanna/edx-platform,cecep-edu/edx-platform,mahendra-r/edx-platform,abdoosh00/edx-rtl-final,ferabra/edx-platform,kxliugang/edx-platform,dkarakats/edx-platform,benpatterson/edx-platform,raccoongang/edx-platform,dcosentino/edx-platform,jbassen/edx-platform,mtlchun/edx,iivic/BoiseStateX,bitifirefly/edx-platform,ahmadio/edx-platform,mbareta/edx-platform-ft,chrisndodge/edx-platform,kursitet/edx-platform,bitifirefly/edx-platform,mtlchun/edx,chudaol/edx-platform,rue89-tech/edx-platform,mjg2203/edx-platform-seas,ubc/edx-platform,LearnEra/LearnEraPlaftform,10clouds/edx-platform,longmen21/edx-platform,marcore/edx-platform,carsongee/edx-platform,pdehaye/theming-edx-platform,Unow/edx-platform,chand3040/cloud_that,TsinghuaX/edx-platform,kamalx/edx-platform,WatanabeYasumasa/edx-platform,benpatterson/edx-platform,pdehaye/theming-edx-platform,beacloudgenius/edx-platform,gsehub/edx-platform,hkawasaki/kawasaki-aio8-1,syjeon/new_edx,xuxiao19910803/edx-platform,jswope00/GAI,proversity-org/edx-platform,PepperPD/edx-pepper-platform,EduPepperPD/pepper2013,playm2mboy/edx-platform,torchingloom/edx-platform,fly19890211/edx-platform,chauhanhardik/populo_2,ak2703/edx-platform,dsajkl/reqiop,ampax/edx-platform,mitocw/edx-platform,shubhdev/openedx,hmcmooc/muddx-platform,RPI-OPENEDX/edx-platform,mahendra-r/edx-platform,dsajkl/123,chauhanhardik/populo,rhndg/openedx,Softmotions/edx-platform,halvertoluke/edx-platform,miptliot/edx-platform,Edraak/circleci-edx-platform,rhndg/openedx,mcgachey/edx-platform,jazztpt/edx-platform,Ayub-Khan/edx-platform,pku9104038/edx-platform,vasyarv/edx-platform,ampax/edx-platform,jelugbo/tundex,y12uc231/edx-platform,chauhanhardik/populo,jamesblunt/edx-platform,pku9104038/edx-platform,kalebhartje/schoolboost,hamzehd/edx-platform,pelikanchik/edx-platform,WatanabeYasumasa/edx-platform,UOMx/edx-platform,DNFcode/edx-platform,ZLLab-Mooc/edx-platform,appsembler/edx-platform,jelugbo/tundex,longmen21/edx-platform,ovnicraft/edx-platform,Shrhawk/edx-platform,pomegranited/edx-platform,romain-li/edx-platform,simbs/edx-platform,philanthropy-u/edx-platform,doismellburning/edx-platform,wwj718/ANALYSE,IONISx/edx-platform,praveen-pal/edx-platform,franosincic/edx-platform,MakeHer/edx-platform,unicri/edx-platform,Softmotions/edx-platform,xuxiao19910803/edx-platform,IONISx/edx-platform,SivilTaram/edx-platform,mushtaqak/edx-platform,eestay/edx-platform,zhenzhai/edx-platform,waheedahmed/edx-platform,apigee/edx-platform,JCBarahona/edX,stvstnfrd/edx-platform,dsajkl/123,deepsrijit1105/edx-platform,polimediaupv/edx-platform,jzoldak/edx-platform,Semi-global/edx-platform,zubair-arbi/edx-platform,ampax/edx-platform-backup,pabloborrego93/edx-platform,ZLLab-Mooc/edx-platform,ovnicraft/edx-platform,chudaol/edx-platform,arbrandes/edx-platform,chauhanhardik/populo,Edraak/edx-platform,cecep-edu/edx-platform,ahmadiga/min_edx,Kalyzee/edx-platform,mtlchun/edx,etzhou/edx-platform,mjirayu/sit_academy,JCBarahona/edX,UXE/local-edx,antoviaque/edx-platform,jazztpt/edx-platform,hkawasaki/kawasaki-aio8-0,shubhdev/openedx,EduPepperPDTesting/pepper2013-testing,Edraak/edraak-platform,fly19890211/edx-platform,bdero/edx-platform,UOMx/edx-platform,AkA84/edx-platform,DefyVentures/edx-platform,fly19890211/edx-platform,RPI-OPENEDX/edx-platform,beni55/edx-platform,analyseuc3m/ANALYSE-v1,kmoocdev2/edx-platform,J861449197/edx-platform,cselis86/edx-platform,abdoosh00/edraak,antoviaque/edx-platform,PepperPD/edx-pepper-platform,shubhdev/openedx,mbareta/edx-platform-ft,TsinghuaX/edx-platform,morenopc/edx-platform,edx-solutions/edx-platform,kmoocdev/edx-platform,pepeportela/edx-platform,peterm-itr/edx-platform,halvertoluke/edx-platform,valtech-mooc/edx-platform,mjg2203/edx-platform-seas,xinjiguaike/edx-platform,angelapper/edx-platform,openfun/edx-platform,jruiperezv/ANALYSE,olexiim/edx-platform,CredoReference/edx-platform,auferack08/edx-platform,edx/edx-platform,IndonesiaX/edx-platform,shurihell/testasia,SravanthiSinha/edx-platform,msegado/edx-platform,jjmiranda/edx-platform,bitifirefly/edx-platform,JioEducation/edx-platform,mtlchun/edx,4eek/edx-platform,kxliugang/edx-platform,kmoocdev/edx-platform,Endika/edx-platform,y12uc231/edx-platform,vikas1885/test1,jazkarta/edx-platform,dsajkl/123,yokose-ks/edx-platform,appsembler/edx-platform,cognitiveclass/edx-platform,analyseuc3m/ANALYSE-v1,arbrandes/edx-platform,cecep-edu/edx-platform,caesar2164/edx-platform,jolyonb/edx-platform,Endika/edx-platform,dsajkl/reqiop,wwj718/edx-platform,dkarakats/edx-platform,WatanabeYasumasa/edx-platform,cselis86/edx-platform,chauhanhardik/populo,nagyistoce/edx-platform,torchingloom/edx-platform,shashank971/edx-platform,jonathan-beard/edx-platform,shurihell/testasia,Stanford-Online/edx-platform,rismalrv/edx-platform,auferack08/edx-platform,jswope00/griffinx,valtech-mooc/edx-platform,motion2015/edx-platform,stvstnfrd/edx-platform,stvstnfrd/edx-platform,doismellburning/edx-platform,xingyepei/edx-platform,longmen21/edx-platform,openfun/edx-platform,xinjiguaike/edx-platform,rationalAgent/edx-platform-custom,cselis86/edx-platform,JCBarahona/edX,kxliugang/edx-platform,10clouds/edx-platform,amir-qayyum-khan/edx-platform,syjeon/new_edx,zadgroup/edx-platform,pelikanchik/edx-platform,ahmedaljazzar/edx-platform,ferabra/edx-platform,prarthitm/edxplatform,sudheerchintala/LearnEraPlatForm,hmcmooc/muddx-platform,dcosentino/edx-platform,wwj718/edx-platform,CourseTalk/edx-platform,romain-li/edx-platform,louyihua/edx-platform,jazkarta/edx-platform,nanolearning/edx-platform,AkA84/edx-platform,yokose-ks/edx-platform,ahmadio/edx-platform,nanolearning/edx-platform,pelikanchik/edx-platform,gymnasium/edx-platform,solashirai/edx-platform,Ayub-Khan/edx-platform,polimediaupv/edx-platform,xinjiguaike/edx-platform,BehavioralInsightsTeam/edx-platform,EduPepperPDTesting/pepper2013-testing,cselis86/edx-platform,jbzdak/edx-platform,romain-li/edx-platform,motion2015/a3,hkawasaki/kawasaki-aio8-0,jelugbo/tundex,EduPepperPD/pepper2013,doganov/edx-platform,vikas1885/test1,kmoocdev2/edx-platform,wwj718/ANALYSE,knehez/edx-platform,IITBinterns13/edx-platform-dev,IndonesiaX/edx-platform,don-github/edx-platform,xuxiao19910803/edx-platform,y12uc231/edx-platform,JCBarahona/edX,ubc/edx-platform,J861449197/edx-platform,abdoosh00/edraak,Semi-global/edx-platform,tiagochiavericosta/edx-platform,SivilTaram/edx-platform,DNFcode/edx-platform,edry/edx-platform,franosincic/edx-platform,edx-solutions/edx-platform,jbassen/edx-platform,bigdatauniversity/edx-platform,EDUlib/edx-platform,abdoosh00/edraak,pomegranited/edx-platform,jswope00/griffinx,Lektorium-LLC/edx-platform,raccoongang/edx-platform,MSOpenTech/edx-platform,sameetb-cuelogic/edx-platform-test,iivic/BoiseStateX,xingyepei/edx-platform,beacloudgenius/edx-platform,etzhou/edx-platform,tiagochiavericosta/edx-platform,jonathan-beard/edx-platform,apigee/edx-platform,MSOpenTech/edx-platform,simbs/edx-platform,ampax/edx-platform-backup,arbrandes/edx-platform,devs1991/test_edx_docmode,MakeHer/edx-platform,deepsrijit1105/edx-platform,UXE/local-edx,Livit/Livit.Learn.EdX,ZLLab-Mooc/edx-platform,ferabra/edx-platform,eduNEXT/edunext-platform,proversity-org/edx-platform,angelapper/edx-platform,CourseTalk/edx-platform,leansoft/edx-platform,alexthered/kienhoc-platform,chudaol/edx-platform,angelapper/edx-platform,mitocw/edx-platform,wwj718/ANALYSE,morenopc/edx-platform,iivic/BoiseStateX,kxliugang/edx-platform,prarthitm/edxplatform,pabloborrego93/edx-platform,jazkarta/edx-platform-for-isc,romain-li/edx-platform,xingyepei/edx-platform,miptliot/edx-platform,morenopc/edx-platform,kmoocdev/edx-platform,devs1991/test_edx_docmode,peterm-itr/edx-platform,TeachAtTUM/edx-platform,shashank971/edx-platform,jonathan-beard/edx-platform,DNFcode/edx-platform,jbassen/edx-platform,knehez/edx-platform,jazztpt/edx-platform,solashirai/edx-platform,Unow/edx-platform,ovnicraft/edx-platform,cyanna/edx-platform,zofuthan/edx-platform,doganov/edx-platform,ak2703/edx-platform,antonve/s4-project-mooc,kalebhartje/schoolboost,edx-solutions/edx-platform,B-MOOC/edx-platform,rue89-tech/edx-platform,nttks/jenkins-test,Kalyzee/edx-platform,motion2015/a3,hastexo/edx-platform,EDUlib/edx-platform,DefyVentures/edx-platform,don-github/edx-platform,jazkarta/edx-platform-for-isc,unicri/edx-platform,mushtaqak/edx-platform,eduNEXT/edx-platform,Stanford-Online/edx-platform,cpennington/edx-platform,RPI-OPENEDX/edx-platform,IONISx/edx-platform,zerobatu/edx-platform,chauhanhardik/populo_2,synergeticsedx/deployment-wipro,alexthered/kienhoc-platform,shabab12/edx-platform,arifsetiawan/edx-platform,jjmiranda/edx-platform,synergeticsedx/deployment-wipro,jonathan-beard/edx-platform,zadgroup/edx-platform,unicri/edx-platform,deepsrijit1105/edx-platform,beacloudgenius/edx-platform,adoosii/edx-platform,appliedx/edx-platform,J861449197/edx-platform,zhenzhai/edx-platform,mjg2203/edx-platform-seas,nagyistoce/edx-platform,zofuthan/edx-platform,bigdatauniversity/edx-platform,dkarakats/edx-platform,kamalx/edx-platform,devs1991/test_edx_docmode,abdoosh00/edx-rtl-final,Lektorium-LLC/edx-platform,Edraak/circleci-edx-platform,pdehaye/theming-edx-platform,jolyonb/edx-platform,halvertoluke/edx-platform,jzoldak/edx-platform,ampax/edx-platform-backup,shubhdev/edxOnBaadal,cognitiveclass/edx-platform,chudaol/edx-platform,kmoocdev2/edx-platform,mitocw/edx-platform,Semi-global/edx-platform,utecuy/edx-platform,Kalyzee/edx-platform,gymnasium/edx-platform,edry/edx-platform,a-parhom/edx-platform,nttks/edx-platform,polimediaupv/edx-platform,fly19890211/edx-platform,procangroup/edx-platform,jswope00/griffinx,devs1991/test_edx_docmode,vasyarv/edx-platform,msegado/edx-platform,knehez/edx-platform,jazztpt/edx-platform,vismartltd/edx-platform,DefyVentures/edx-platform,alu042/edx-platform,atsolakid/edx-platform,naresh21/synergetics-edx-platform,Shrhawk/edx-platform,rationalAgent/edx-platform-custom,hastexo/edx-platform,marcore/edx-platform,chrisndodge/edx-platform,cyanna/edx-platform,MakeHer/edx-platform,nikolas/edx-platform,pdehaye/theming-edx-platform,jazkarta/edx-platform-for-isc,rismalrv/edx-platform,beni55/edx-platform,JCBarahona/edX,BehavioralInsightsTeam/edx-platform,LearnEra/LearnEraPlaftform,angelapper/edx-platform,syjeon/new_edx,doganov/edx-platform,torchingloom/edx-platform,valtech-mooc/edx-platform,unicri/edx-platform,a-parhom/edx-platform,motion2015/edx-platform,cyanna/edx-platform,philanthropy-u/edx-platform,ESOedX/edx-platform,morpheby/levelup-by,alexthered/kienhoc-platform,EduPepperPD/pepper2013,rhndg/openedx,dcosentino/edx-platform,analyseuc3m/ANALYSE-v1,dkarakats/edx-platform,EduPepperPDTesting/pepper2013-testing,longmen21/edx-platform,LICEF/edx-platform,zadgroup/edx-platform,jazkarta/edx-platform-for-isc,kamalx/edx-platform,wwj718/edx-platform,motion2015/edx-platform,bdero/edx-platform,kursitet/edx-platform,kmoocdev/edx-platform,chrisndodge/edx-platform,inares/edx-platform,prarthitm/edxplatform,sudheerchintala/LearnEraPlatForm,Kalyzee/edx-platform,beni55/edx-platform,kursitet/edx-platform,SravanthiSinha/edx-platform,IONISx/edx-platform,antonve/s4-project-mooc,mushtaqak/edx-platform,alu042/edx-platform,mcgachey/edx-platform,appliedx/edx-platform,procangroup/edx-platform,xingyepei/edx-platform,shubhdev/edxOnBaadal,WatanabeYasumasa/edx-platform,shubhdev/edx-platform,jruiperezv/ANALYSE,mjg2203/edx-platform-seas,amir-qayyum-khan/edx-platform,Softmotions/edx-platform,praveen-pal/edx-platform,jamiefolsom/edx-platform,Ayub-Khan/edx-platform,jamesblunt/edx-platform,halvertoluke/edx-platform,MSOpenTech/edx-platform,waheedahmed/edx-platform,caesar2164/edx-platform,alu042/edx-platform,iivic/BoiseStateX,louyihua/edx-platform,zhenzhai/edx-platform,antoviaque/edx-platform,jbassen/edx-platform,Edraak/circleci-edx-platform,BehavioralInsightsTeam/edx-platform,itsjeyd/edx-platform,leansoft/edx-platform,martynovp/edx-platform,AkA84/edx-platform,beni55/edx-platform,gsehub/edx-platform,abdoosh00/edx-rtl-final,louyihua/edx-platform,doismellburning/edx-platform,jelugbo/tundex,rhndg/openedx,LearnEra/LearnEraPlaftform,vikas1885/test1,ZLLab-Mooc/edx-platform,IITBinterns13/edx-platform-dev,EduPepperPDTesting/pepper2013-testing,PepperPD/edx-pepper-platform,simbs/edx-platform,peterm-itr/edx-platform,hmcmooc/muddx-platform,shabab12/edx-platform,appsembler/edx-platform,ahmedaljazzar/edx-platform,gsehub/edx-platform,jjmiranda/edx-platform,Semi-global/edx-platform,xinjiguaike/edx-platform,devs1991/test_edx_docmode,shubhdev/edx-platform,vikas1885/test1,TsinghuaX/edx-platform,ahmadiga/min_edx,bitifirefly/edx-platform,JioEducation/edx-platform,arifsetiawan/edx-platform,abdoosh00/edx-rtl-final,nanolearning/edx-platform,jruiperezv/ANALYSE,synergeticsedx/deployment-wipro,vismartltd/edx-platform,LICEF/edx-platform,naresh21/synergetics-edx-platform,valtech-mooc/edx-platform,ESOedX/edx-platform,utecuy/edx-platform,Ayub-Khan/edx-platform,motion2015/a3,sudheerchintala/LearnEraPlatForm,hamzehd/edx-platform,rismalrv/edx-platform,edx/edx-platform,SivilTaram/edx-platform,kmoocdev2/edx-platform,shubhdev/edxOnBaadal,vismartltd/edx-platform,philanthropy-u/edx-platform,Livit/Livit.Learn.EdX,EDUlib/edx-platform,motion2015/a3,benpatterson/edx-platform,iivic/BoiseStateX,adoosii/edx-platform,tiagochiavericosta/edx-platform,ahmadiga/min_edx,xuxiao19910803/edx,inares/edx-platform,ak2703/edx-platform,polimediaupv/edx-platform,waheedahmed/edx-platform,teltek/edx-platform,mjirayu/sit_academy,etzhou/edx-platform,UXE/local-edx,shabab12/edx-platform,leansoft/edx-platform,ahmadiga/min_edx,pomegranited/edx-platform,jelugbo/tundex,zhenzhai/edx-platform,y12uc231/edx-platform,ahmedaljazzar/edx-platform,fintech-circle/edx-platform,jazkarta/edx-platform,hkawasaki/kawasaki-aio8-2,SravanthiSinha/edx-platform,lduarte1991/edx-platform,franosincic/edx-platform,UOMx/edx-platform,kalebhartje/schoolboost,chudaol/edx-platform,pepeportela/edx-platform,rationalAgent/edx-platform-custom,hkawasaki/kawasaki-aio8-2,Endika/edx-platform,hkawasaki/kawasaki-aio8-0,mtlchun/edx,itsjeyd/edx-platform,jswope00/GAI,apigee/edx-platform,shubhdev/edxOnBaadal,msegado/edx-platform,jswope00/griffinx,shubhdev/edx-platform,UOMx/edx-platform,tanmaykm/edx-platform,openfun/edx-platform,polimediaupv/edx-platform,B-MOOC/edx-platform,wwj718/edx-platform,rationalAgent/edx-platform-custom,vasyarv/edx-platform,shubhdev/edx-platform,edry/edx-platform,martynovp/edx-platform,TsinghuaX/edx-platform,LICEF/edx-platform,pabloborrego93/edx-platform,hkawasaki/kawasaki-aio8-1,TeachAtTUM/edx-platform,xingyepei/edx-platform,Shrhawk/edx-platform,pomegranited/edx-platform,carsongee/edx-platform,Edraak/edx-platform,jjmiranda/edx-platform,sameetb-cuelogic/edx-platform-test,xuxiao19910803/edx,ampax/edx-platform,devs1991/test_edx_docmode,y12uc231/edx-platform,CredoReference/edx-platform,dsajkl/123,alu042/edx-platform,andyzsf/edx,inares/edx-platform,ubc/edx-platform,openfun/edx-platform,nanolearning/edx-platform,praveen-pal/edx-platform,pku9104038/edx-platform,cyanna/edx-platform,deepsrijit1105/edx-platform,Edraak/edraak-platform,rue89-tech/edx-platform,benpatterson/edx-platform,appliedx/edx-platform,itsjeyd/edx-platform,cpennington/edx-platform,ak2703/edx-platform,a-parhom/edx-platform,B-MOOC/edx-platform,jamesblunt/edx-platform,jamesblunt/edx-platform,eestay/edx-platform,arifsetiawan/edx-platform,edx/edx-platform,nttks/jenkins-test,chauhanhardik/populo_2,eduNEXT/edunext-platform,kmoocdev2/edx-platform,naresh21/synergetics-edx-platform,andyzsf/edx,dsajkl/reqiop,edx/edx-platform,CourseTalk/edx-platform,andyzsf/edx,nttks/jenkins-test,vismartltd/edx-platform,dcosentino/edx-platform,jbzdak/edx-platform,AkA84/edx-platform,cpennington/edx-platform,teltek/edx-platform,Stanford-Online/edx-platform,doganov/edx-platform,antonve/s4-project-mooc,abdoosh00/edraak,mjirayu/sit_academy,4eek/edx-platform,chauhanhardik/populo_2,cognitiveclass/edx-platform,kmoocdev/edx-platform,nanolearningllc/edx-platform-cypress,inares/edx-platform,kamalx/edx-platform,jbzdak/edx-platform,EduPepperPDTesting/pepper2013-testing,doismellburning/edx-platform,jbassen/edx-platform,jamiefolsom/edx-platform,kursitet/edx-platform,knehez/edx-platform,teltek/edx-platform,CredoReference/edx-platform,vasyarv/edx-platform,louyihua/edx-platform,EDUlib/edx-platform,hamzehd/edx-platform,hamzehd/edx-platform,atsolakid/edx-platform,shurihell/testasia,mushtaqak/edx-platform,nanolearningllc/edx-platform-cypress-2,pomegranited/edx-platform,MSOpenTech/edx-platform,sameetb-cuelogic/edx-platform-test,Stanford-Online/edx-platform,beacloudgenius/edx-platform,jolyonb/edx-platform,Edraak/edx-platform,solashirai/edx-platform,tanmaykm/edx-platform,CredoReference/edx-platform,4eek/edx-platform,utecuy/edx-platform,jonathan-beard/edx-platform,jzoldak/edx-platform,SivilTaram/edx-platform,Ayub-Khan/edx-platform,jamiefolsom/edx-platform,andyzsf/edx,hastexo/edx-platform,Edraak/edraak-platform,halvertoluke/edx-platform,nanolearningllc/edx-platform-cypress-2,adoosii/edx-platform,don-github/edx-platform,dsajkl/reqiop,marcore/edx-platform,hkawasaki/kawasaki-aio8-2,appliedx/edx-platform,nagyistoce/edx-platform,romain-li/edx-platform,playm2mboy/edx-platform,mcgachey/edx-platform,wwj718/edx-platform,mahendra-r/edx-platform,cecep-edu/edx-platform,proversity-org/edx-platform,jswope00/GAI,rismalrv/edx-platform,nttks/jenkins-test,bitifirefly/edx-platform,raccoongang/edx-platform,ahmadio/edx-platform,IndonesiaX/edx-platform,appliedx/edx-platform,inares/edx-platform,bigdatauniversity/edx-platform,morpheby/levelup-by,IITBinterns13/edx-platform-dev,sameetb-cuelogic/edx-platform-test,kamalx/edx-platform,MakeHer/edx-platform,xuxiao19910803/edx,zubair-arbi/edx-platform,xuxiao19910803/edx-platform,dcosentino/edx-platform,rismalrv/edx-platform,gymnasium/edx-platform,playm2mboy/edx-platform,Edraak/circleci-edx-platform,a-parhom/edx-platform,SravanthiSinha/edx-platform,zubair-arbi/edx-platform,JioEducation/edx-platform,TeachAtTUM/edx-platform,pelikanchik/edx-platform,openfun/edx-platform,nanolearningllc/edx-platform-cypress,nanolearningllc/edx-platform-cypress,etzhou/edx-platform,jazkarta/edx-platform-for-isc,4eek/edx-platform,kalebhartje/schoolboost,mcgachey/edx-platform,pku9104038/edx-platform,atsolakid/edx-platform,mbareta/edx-platform-ft,msegado/edx-platform,Softmotions/edx-platform,eemirtekin/edx-platform,nikolas/edx-platform,hastexo/edx-platform,eestay/edx-platform,ESOedX/edx-platform,OmarIthawi/edx-platform,zhenzhai/edx-platform,nagyistoce/edx-platform,shubhdev/openedx,olexiim/edx-platform,eemirtekin/edx-platform,SivilTaram/edx-platform,motion2015/edx-platform,unicri/edx-platform,zubair-arbi/edx-platform,jzoldak/edx-platform,ak2703/edx-platform
"""Views for debugging and diagnostics""" import pprint from django.http import Http404 from django.contrib.auth.decorators import login_required from django_future.csrf import ensure_csrf_cookie, csrf_exempt from mitxmako.shortcuts import render_to_response from codejail.safe_exec import safe_exec @login_required @ensure_csrf_cookie def run_python(request): if not request.user.is_staff: raise Http404 c = {} c['code'] = '' c['results'] = None if request.method == 'POST': py_code = c['code'] = request.POST.get('code') g = {} try: safe_exec(py_code, g) except Exception as e: c['results'] = str(e) else: c['results'] = pprint.pformat(g) return render_to_response("debug/run_python_form.html", c) Print the full traceback when execution fails.
"""Views for debugging and diagnostics""" import pprint import traceback from django.http import Http404 from django.contrib.auth.decorators import login_required from django_future.csrf import ensure_csrf_cookie, csrf_exempt from mitxmako.shortcuts import render_to_response from codejail.safe_exec import safe_exec @login_required @ensure_csrf_cookie def run_python(request): """A page to allow testing the Python sandbox on a production server.""" if not request.user.is_staff: raise Http404 c = {} c['code'] = '' c['results'] = None if request.method == 'POST': py_code = c['code'] = request.POST.get('code') g = {} try: safe_exec(py_code, g) except Exception as e: c['results'] = traceback.format_exc() else: c['results'] = pprint.pformat(g) return render_to_response("debug/run_python_form.html", c)
<commit_before>"""Views for debugging and diagnostics""" import pprint from django.http import Http404 from django.contrib.auth.decorators import login_required from django_future.csrf import ensure_csrf_cookie, csrf_exempt from mitxmako.shortcuts import render_to_response from codejail.safe_exec import safe_exec @login_required @ensure_csrf_cookie def run_python(request): if not request.user.is_staff: raise Http404 c = {} c['code'] = '' c['results'] = None if request.method == 'POST': py_code = c['code'] = request.POST.get('code') g = {} try: safe_exec(py_code, g) except Exception as e: c['results'] = str(e) else: c['results'] = pprint.pformat(g) return render_to_response("debug/run_python_form.html", c) <commit_msg>Print the full traceback when execution fails.<commit_after>
"""Views for debugging and diagnostics""" import pprint import traceback from django.http import Http404 from django.contrib.auth.decorators import login_required from django_future.csrf import ensure_csrf_cookie, csrf_exempt from mitxmako.shortcuts import render_to_response from codejail.safe_exec import safe_exec @login_required @ensure_csrf_cookie def run_python(request): """A page to allow testing the Python sandbox on a production server.""" if not request.user.is_staff: raise Http404 c = {} c['code'] = '' c['results'] = None if request.method == 'POST': py_code = c['code'] = request.POST.get('code') g = {} try: safe_exec(py_code, g) except Exception as e: c['results'] = traceback.format_exc() else: c['results'] = pprint.pformat(g) return render_to_response("debug/run_python_form.html", c)
"""Views for debugging and diagnostics""" import pprint from django.http import Http404 from django.contrib.auth.decorators import login_required from django_future.csrf import ensure_csrf_cookie, csrf_exempt from mitxmako.shortcuts import render_to_response from codejail.safe_exec import safe_exec @login_required @ensure_csrf_cookie def run_python(request): if not request.user.is_staff: raise Http404 c = {} c['code'] = '' c['results'] = None if request.method == 'POST': py_code = c['code'] = request.POST.get('code') g = {} try: safe_exec(py_code, g) except Exception as e: c['results'] = str(e) else: c['results'] = pprint.pformat(g) return render_to_response("debug/run_python_form.html", c) Print the full traceback when execution fails."""Views for debugging and diagnostics""" import pprint import traceback from django.http import Http404 from django.contrib.auth.decorators import login_required from django_future.csrf import ensure_csrf_cookie, csrf_exempt from mitxmako.shortcuts import render_to_response from codejail.safe_exec import safe_exec @login_required @ensure_csrf_cookie def run_python(request): """A page to allow testing the Python sandbox on a production server.""" if not request.user.is_staff: raise Http404 c = {} c['code'] = '' c['results'] = None if request.method == 'POST': py_code = c['code'] = request.POST.get('code') g = {} try: safe_exec(py_code, g) except Exception as e: c['results'] = traceback.format_exc() else: c['results'] = pprint.pformat(g) return render_to_response("debug/run_python_form.html", c)
<commit_before>"""Views for debugging and diagnostics""" import pprint from django.http import Http404 from django.contrib.auth.decorators import login_required from django_future.csrf import ensure_csrf_cookie, csrf_exempt from mitxmako.shortcuts import render_to_response from codejail.safe_exec import safe_exec @login_required @ensure_csrf_cookie def run_python(request): if not request.user.is_staff: raise Http404 c = {} c['code'] = '' c['results'] = None if request.method == 'POST': py_code = c['code'] = request.POST.get('code') g = {} try: safe_exec(py_code, g) except Exception as e: c['results'] = str(e) else: c['results'] = pprint.pformat(g) return render_to_response("debug/run_python_form.html", c) <commit_msg>Print the full traceback when execution fails.<commit_after>"""Views for debugging and diagnostics""" import pprint import traceback from django.http import Http404 from django.contrib.auth.decorators import login_required from django_future.csrf import ensure_csrf_cookie, csrf_exempt from mitxmako.shortcuts import render_to_response from codejail.safe_exec import safe_exec @login_required @ensure_csrf_cookie def run_python(request): """A page to allow testing the Python sandbox on a production server.""" if not request.user.is_staff: raise Http404 c = {} c['code'] = '' c['results'] = None if request.method == 'POST': py_code = c['code'] = request.POST.get('code') g = {} try: safe_exec(py_code, g) except Exception as e: c['results'] = traceback.format_exc() else: c['results'] = pprint.pformat(g) return render_to_response("debug/run_python_form.html", c)
4b03c2c39c90bd1563954df16f353348f20d7280
setup.py
setup.py
#!/usr/bin/env python # -*- coding: utf-8 -*- """The Screen class lets you to do positioned writes to the dos terminal. The Screen class also allows you to specify the colors for foreground and background, to the extent the dos terminal allows. """ classifiers = """\ Development Status :: 3 - Alpha Environment :: Win32 (MS Windows) Intended Audience :: Developers License :: OSI Approved :: BSD License Operating System :: Microsoft :: MS-DOS Programming Language :: Python Topic :: Software Development :: Libraries Topic :: System :: Shells Topic :: Terminals """ from distutils.core import setup doclines = __doc__.split('\n') setup( name='dosbox-screen', version='0.0.1', requires=['ctypes', 'colorama'], description=doclines[0], classifiers=[line for line in classifiers.split('\n') if line], long_description=' '.join(doclines), license="BSD", #platform='win32', author='Bjorn Pettersen', author_email='bjorn@tkbe.org', url='https://github.com/thebjorn/doscmd-screen', download_url='https://github.com/thebjorn/doscmd-screen', py_modules=['screen'] )
#!/usr/bin/env python # -*- coding: utf-8 -*- """The Screen class lets you to do positioned writes to the dos terminal. The Screen class also allows you to specify the colors for foreground and background, to the extent the dos terminal allows. """ classifiers = """\ Development Status :: 3 - Alpha Environment :: Win32 (MS Windows) Intended Audience :: Developers License :: OSI Approved :: BSD License Operating System :: Microsoft :: MS-DOS Programming Language :: Python Topic :: Software Development :: Libraries Topic :: System :: Shells Topic :: Terminals """ from distutils.core import setup doclines = __doc__.split('\n') setup( name='dosbox-screen', version='0.0.3', requires=['colorama'], install_requires=['colorama'], description=doclines[0], classifiers=[line for line in classifiers.split('\n') if line], long_description=' '.join(doclines), license="BSD", #platform='win32', author='Bjorn Pettersen', author_email='bjorn@tkbe.org', url='https://github.com/thebjorn/doscmd-screen', download_url='https://github.com/thebjorn/doscmd-screen', py_modules=['screen'] )
Add install_requires and remove ctypes from requirements.
Add install_requires and remove ctypes from requirements.
Python
mit
thebjorn/doscmd-screen
#!/usr/bin/env python # -*- coding: utf-8 -*- """The Screen class lets you to do positioned writes to the dos terminal. The Screen class also allows you to specify the colors for foreground and background, to the extent the dos terminal allows. """ classifiers = """\ Development Status :: 3 - Alpha Environment :: Win32 (MS Windows) Intended Audience :: Developers License :: OSI Approved :: BSD License Operating System :: Microsoft :: MS-DOS Programming Language :: Python Topic :: Software Development :: Libraries Topic :: System :: Shells Topic :: Terminals """ from distutils.core import setup doclines = __doc__.split('\n') setup( name='dosbox-screen', version='0.0.1', requires=['ctypes', 'colorama'], description=doclines[0], classifiers=[line for line in classifiers.split('\n') if line], long_description=' '.join(doclines), license="BSD", #platform='win32', author='Bjorn Pettersen', author_email='bjorn@tkbe.org', url='https://github.com/thebjorn/doscmd-screen', download_url='https://github.com/thebjorn/doscmd-screen', py_modules=['screen'] ) Add install_requires and remove ctypes from requirements.
#!/usr/bin/env python # -*- coding: utf-8 -*- """The Screen class lets you to do positioned writes to the dos terminal. The Screen class also allows you to specify the colors for foreground and background, to the extent the dos terminal allows. """ classifiers = """\ Development Status :: 3 - Alpha Environment :: Win32 (MS Windows) Intended Audience :: Developers License :: OSI Approved :: BSD License Operating System :: Microsoft :: MS-DOS Programming Language :: Python Topic :: Software Development :: Libraries Topic :: System :: Shells Topic :: Terminals """ from distutils.core import setup doclines = __doc__.split('\n') setup( name='dosbox-screen', version='0.0.3', requires=['colorama'], install_requires=['colorama'], description=doclines[0], classifiers=[line for line in classifiers.split('\n') if line], long_description=' '.join(doclines), license="BSD", #platform='win32', author='Bjorn Pettersen', author_email='bjorn@tkbe.org', url='https://github.com/thebjorn/doscmd-screen', download_url='https://github.com/thebjorn/doscmd-screen', py_modules=['screen'] )
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- """The Screen class lets you to do positioned writes to the dos terminal. The Screen class also allows you to specify the colors for foreground and background, to the extent the dos terminal allows. """ classifiers = """\ Development Status :: 3 - Alpha Environment :: Win32 (MS Windows) Intended Audience :: Developers License :: OSI Approved :: BSD License Operating System :: Microsoft :: MS-DOS Programming Language :: Python Topic :: Software Development :: Libraries Topic :: System :: Shells Topic :: Terminals """ from distutils.core import setup doclines = __doc__.split('\n') setup( name='dosbox-screen', version='0.0.1', requires=['ctypes', 'colorama'], description=doclines[0], classifiers=[line for line in classifiers.split('\n') if line], long_description=' '.join(doclines), license="BSD", #platform='win32', author='Bjorn Pettersen', author_email='bjorn@tkbe.org', url='https://github.com/thebjorn/doscmd-screen', download_url='https://github.com/thebjorn/doscmd-screen', py_modules=['screen'] ) <commit_msg>Add install_requires and remove ctypes from requirements.<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- """The Screen class lets you to do positioned writes to the dos terminal. The Screen class also allows you to specify the colors for foreground and background, to the extent the dos terminal allows. """ classifiers = """\ Development Status :: 3 - Alpha Environment :: Win32 (MS Windows) Intended Audience :: Developers License :: OSI Approved :: BSD License Operating System :: Microsoft :: MS-DOS Programming Language :: Python Topic :: Software Development :: Libraries Topic :: System :: Shells Topic :: Terminals """ from distutils.core import setup doclines = __doc__.split('\n') setup( name='dosbox-screen', version='0.0.3', requires=['colorama'], install_requires=['colorama'], description=doclines[0], classifiers=[line for line in classifiers.split('\n') if line], long_description=' '.join(doclines), license="BSD", #platform='win32', author='Bjorn Pettersen', author_email='bjorn@tkbe.org', url='https://github.com/thebjorn/doscmd-screen', download_url='https://github.com/thebjorn/doscmd-screen', py_modules=['screen'] )
#!/usr/bin/env python # -*- coding: utf-8 -*- """The Screen class lets you to do positioned writes to the dos terminal. The Screen class also allows you to specify the colors for foreground and background, to the extent the dos terminal allows. """ classifiers = """\ Development Status :: 3 - Alpha Environment :: Win32 (MS Windows) Intended Audience :: Developers License :: OSI Approved :: BSD License Operating System :: Microsoft :: MS-DOS Programming Language :: Python Topic :: Software Development :: Libraries Topic :: System :: Shells Topic :: Terminals """ from distutils.core import setup doclines = __doc__.split('\n') setup( name='dosbox-screen', version='0.0.1', requires=['ctypes', 'colorama'], description=doclines[0], classifiers=[line for line in classifiers.split('\n') if line], long_description=' '.join(doclines), license="BSD", #platform='win32', author='Bjorn Pettersen', author_email='bjorn@tkbe.org', url='https://github.com/thebjorn/doscmd-screen', download_url='https://github.com/thebjorn/doscmd-screen', py_modules=['screen'] ) Add install_requires and remove ctypes from requirements.#!/usr/bin/env python # -*- coding: utf-8 -*- """The Screen class lets you to do positioned writes to the dos terminal. The Screen class also allows you to specify the colors for foreground and background, to the extent the dos terminal allows. """ classifiers = """\ Development Status :: 3 - Alpha Environment :: Win32 (MS Windows) Intended Audience :: Developers License :: OSI Approved :: BSD License Operating System :: Microsoft :: MS-DOS Programming Language :: Python Topic :: Software Development :: Libraries Topic :: System :: Shells Topic :: Terminals """ from distutils.core import setup doclines = __doc__.split('\n') setup( name='dosbox-screen', version='0.0.3', requires=['colorama'], install_requires=['colorama'], description=doclines[0], classifiers=[line for line in classifiers.split('\n') if line], long_description=' '.join(doclines), license="BSD", #platform='win32', author='Bjorn Pettersen', author_email='bjorn@tkbe.org', url='https://github.com/thebjorn/doscmd-screen', download_url='https://github.com/thebjorn/doscmd-screen', py_modules=['screen'] )
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- """The Screen class lets you to do positioned writes to the dos terminal. The Screen class also allows you to specify the colors for foreground and background, to the extent the dos terminal allows. """ classifiers = """\ Development Status :: 3 - Alpha Environment :: Win32 (MS Windows) Intended Audience :: Developers License :: OSI Approved :: BSD License Operating System :: Microsoft :: MS-DOS Programming Language :: Python Topic :: Software Development :: Libraries Topic :: System :: Shells Topic :: Terminals """ from distutils.core import setup doclines = __doc__.split('\n') setup( name='dosbox-screen', version='0.0.1', requires=['ctypes', 'colorama'], description=doclines[0], classifiers=[line for line in classifiers.split('\n') if line], long_description=' '.join(doclines), license="BSD", #platform='win32', author='Bjorn Pettersen', author_email='bjorn@tkbe.org', url='https://github.com/thebjorn/doscmd-screen', download_url='https://github.com/thebjorn/doscmd-screen', py_modules=['screen'] ) <commit_msg>Add install_requires and remove ctypes from requirements.<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- """The Screen class lets you to do positioned writes to the dos terminal. The Screen class also allows you to specify the colors for foreground and background, to the extent the dos terminal allows. """ classifiers = """\ Development Status :: 3 - Alpha Environment :: Win32 (MS Windows) Intended Audience :: Developers License :: OSI Approved :: BSD License Operating System :: Microsoft :: MS-DOS Programming Language :: Python Topic :: Software Development :: Libraries Topic :: System :: Shells Topic :: Terminals """ from distutils.core import setup doclines = __doc__.split('\n') setup( name='dosbox-screen', version='0.0.3', requires=['colorama'], install_requires=['colorama'], description=doclines[0], classifiers=[line for line in classifiers.split('\n') if line], long_description=' '.join(doclines), license="BSD", #platform='win32', author='Bjorn Pettersen', author_email='bjorn@tkbe.org', url='https://github.com/thebjorn/doscmd-screen', download_url='https://github.com/thebjorn/doscmd-screen', py_modules=['screen'] )
7182f52f495174dc7a9689100f5298e848b8229c
setup.py
setup.py
"""Cloud browser package.""" from setuptools import setup, find_packages from cloud_browser import __version__ # Base packages. MOD_NAME = "cloud_browser" PKGS = [x for x in find_packages() if x.split('.')[0] == MOD_NAME] setup( name="django-cloud-browser", version=__version__, description="Django Cloud Browser application.", long_description="Browser for cloud datastores (Rackspace, AWS, etc.).", url="https://github.com/ryan-roemer/django-cloud-browser", author="Ryan Roemer", author_email="ryan@loose-bits.com", classifiers=[ "Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP :: Site Management", ], install_requires=[ "distribute", ], packages=PKGS, include_package_data=True, )
"""Cloud browser package.""" from __future__ import with_statement import os from setuptools import setup, find_packages from cloud_browser import __version__ ############################################################################### # Base packages. ############################################################################### MOD_NAME = "cloud_browser" PKGS = [x for x in find_packages() if x.split('.')[0] == MOD_NAME] ############################################################################### # Helpers. ############################################################################### def read_file(name): """Read file name (without extension) to string.""" cur_path = os.path.dirname(__file__) exts = ('txt', 'rst') for ext in exts: path = os.path.join(cur_path, '.'.join((name, ext))) if os.path.exists(path): with open(path, 'rb') as file_obj: return file_obj.read() return '' ############################################################################### # Setup. ############################################################################### setup( name="django-cloud-browser", version=__version__, description="Django Cloud Browser application.", long_description=read_file("README"), url="https://github.com/ryan-roemer/django-cloud-browser", author="Ryan Roemer", author_email="ryan@loose-bits.com", classifiers=[ "Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP :: Site Management", ], install_requires=[ "distribute", ], packages=PKGS, include_package_data=True, )
Add long description read straight from README.
Setup: Add long description read straight from README.
Python
mit
ryan-roemer/django-cloud-browser,ryan-roemer/django-cloud-browser,UrbanDaddy/django-cloud-browser,UrbanDaddy/django-cloud-browser,ryan-roemer/django-cloud-browser
"""Cloud browser package.""" from setuptools import setup, find_packages from cloud_browser import __version__ # Base packages. MOD_NAME = "cloud_browser" PKGS = [x for x in find_packages() if x.split('.')[0] == MOD_NAME] setup( name="django-cloud-browser", version=__version__, description="Django Cloud Browser application.", long_description="Browser for cloud datastores (Rackspace, AWS, etc.).", url="https://github.com/ryan-roemer/django-cloud-browser", author="Ryan Roemer", author_email="ryan@loose-bits.com", classifiers=[ "Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP :: Site Management", ], install_requires=[ "distribute", ], packages=PKGS, include_package_data=True, ) Setup: Add long description read straight from README.
"""Cloud browser package.""" from __future__ import with_statement import os from setuptools import setup, find_packages from cloud_browser import __version__ ############################################################################### # Base packages. ############################################################################### MOD_NAME = "cloud_browser" PKGS = [x for x in find_packages() if x.split('.')[0] == MOD_NAME] ############################################################################### # Helpers. ############################################################################### def read_file(name): """Read file name (without extension) to string.""" cur_path = os.path.dirname(__file__) exts = ('txt', 'rst') for ext in exts: path = os.path.join(cur_path, '.'.join((name, ext))) if os.path.exists(path): with open(path, 'rb') as file_obj: return file_obj.read() return '' ############################################################################### # Setup. ############################################################################### setup( name="django-cloud-browser", version=__version__, description="Django Cloud Browser application.", long_description=read_file("README"), url="https://github.com/ryan-roemer/django-cloud-browser", author="Ryan Roemer", author_email="ryan@loose-bits.com", classifiers=[ "Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP :: Site Management", ], install_requires=[ "distribute", ], packages=PKGS, include_package_data=True, )
<commit_before>"""Cloud browser package.""" from setuptools import setup, find_packages from cloud_browser import __version__ # Base packages. MOD_NAME = "cloud_browser" PKGS = [x for x in find_packages() if x.split('.')[0] == MOD_NAME] setup( name="django-cloud-browser", version=__version__, description="Django Cloud Browser application.", long_description="Browser for cloud datastores (Rackspace, AWS, etc.).", url="https://github.com/ryan-roemer/django-cloud-browser", author="Ryan Roemer", author_email="ryan@loose-bits.com", classifiers=[ "Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP :: Site Management", ], install_requires=[ "distribute", ], packages=PKGS, include_package_data=True, ) <commit_msg>Setup: Add long description read straight from README.<commit_after>
"""Cloud browser package.""" from __future__ import with_statement import os from setuptools import setup, find_packages from cloud_browser import __version__ ############################################################################### # Base packages. ############################################################################### MOD_NAME = "cloud_browser" PKGS = [x for x in find_packages() if x.split('.')[0] == MOD_NAME] ############################################################################### # Helpers. ############################################################################### def read_file(name): """Read file name (without extension) to string.""" cur_path = os.path.dirname(__file__) exts = ('txt', 'rst') for ext in exts: path = os.path.join(cur_path, '.'.join((name, ext))) if os.path.exists(path): with open(path, 'rb') as file_obj: return file_obj.read() return '' ############################################################################### # Setup. ############################################################################### setup( name="django-cloud-browser", version=__version__, description="Django Cloud Browser application.", long_description=read_file("README"), url="https://github.com/ryan-roemer/django-cloud-browser", author="Ryan Roemer", author_email="ryan@loose-bits.com", classifiers=[ "Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP :: Site Management", ], install_requires=[ "distribute", ], packages=PKGS, include_package_data=True, )
"""Cloud browser package.""" from setuptools import setup, find_packages from cloud_browser import __version__ # Base packages. MOD_NAME = "cloud_browser" PKGS = [x for x in find_packages() if x.split('.')[0] == MOD_NAME] setup( name="django-cloud-browser", version=__version__, description="Django Cloud Browser application.", long_description="Browser for cloud datastores (Rackspace, AWS, etc.).", url="https://github.com/ryan-roemer/django-cloud-browser", author="Ryan Roemer", author_email="ryan@loose-bits.com", classifiers=[ "Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP :: Site Management", ], install_requires=[ "distribute", ], packages=PKGS, include_package_data=True, ) Setup: Add long description read straight from README."""Cloud browser package.""" from __future__ import with_statement import os from setuptools import setup, find_packages from cloud_browser import __version__ ############################################################################### # Base packages. ############################################################################### MOD_NAME = "cloud_browser" PKGS = [x for x in find_packages() if x.split('.')[0] == MOD_NAME] ############################################################################### # Helpers. ############################################################################### def read_file(name): """Read file name (without extension) to string.""" cur_path = os.path.dirname(__file__) exts = ('txt', 'rst') for ext in exts: path = os.path.join(cur_path, '.'.join((name, ext))) if os.path.exists(path): with open(path, 'rb') as file_obj: return file_obj.read() return '' ############################################################################### # Setup. ############################################################################### setup( name="django-cloud-browser", version=__version__, description="Django Cloud Browser application.", long_description=read_file("README"), url="https://github.com/ryan-roemer/django-cloud-browser", author="Ryan Roemer", author_email="ryan@loose-bits.com", classifiers=[ "Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP :: Site Management", ], install_requires=[ "distribute", ], packages=PKGS, include_package_data=True, )
<commit_before>"""Cloud browser package.""" from setuptools import setup, find_packages from cloud_browser import __version__ # Base packages. MOD_NAME = "cloud_browser" PKGS = [x for x in find_packages() if x.split('.')[0] == MOD_NAME] setup( name="django-cloud-browser", version=__version__, description="Django Cloud Browser application.", long_description="Browser for cloud datastores (Rackspace, AWS, etc.).", url="https://github.com/ryan-roemer/django-cloud-browser", author="Ryan Roemer", author_email="ryan@loose-bits.com", classifiers=[ "Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP :: Site Management", ], install_requires=[ "distribute", ], packages=PKGS, include_package_data=True, ) <commit_msg>Setup: Add long description read straight from README.<commit_after>"""Cloud browser package.""" from __future__ import with_statement import os from setuptools import setup, find_packages from cloud_browser import __version__ ############################################################################### # Base packages. ############################################################################### MOD_NAME = "cloud_browser" PKGS = [x for x in find_packages() if x.split('.')[0] == MOD_NAME] ############################################################################### # Helpers. ############################################################################### def read_file(name): """Read file name (without extension) to string.""" cur_path = os.path.dirname(__file__) exts = ('txt', 'rst') for ext in exts: path = os.path.join(cur_path, '.'.join((name, ext))) if os.path.exists(path): with open(path, 'rb') as file_obj: return file_obj.read() return '' ############################################################################### # Setup. ############################################################################### setup( name="django-cloud-browser", version=__version__, description="Django Cloud Browser application.", long_description=read_file("README"), url="https://github.com/ryan-roemer/django-cloud-browser", author="Ryan Roemer", author_email="ryan@loose-bits.com", classifiers=[ "Development Status :: 3 - Alpha", "Environment :: Web Environment", "Framework :: Django", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Internet", "Topic :: Internet :: WWW/HTTP :: Site Management", ], install_requires=[ "distribute", ], packages=PKGS, include_package_data=True, )
18e07203967dc2ee53a992e61ab709fc0a58d882
setup.py
setup.py
import io import os import re from setuptools import setup, find_packages def find_version(): file_dir = os.path.dirname(__file__) with io.open(os.path.join(file_dir, 'auth0', '__init__.py')) as f: version = re.search(r'^__version__ = [\'"]([^\'"]*)[\'"]', f.read()) if version: return version.group(1) else: raise RuntimeError("Unable to find version string.") with open('README.rst') as f: long_description = f.read() setup( name='auth0-python', version=find_version(), description='Auth0 Python SDK', long_description=long_description, author='Auth0', author_email='support@auth0.com', license='MIT', packages=find_packages(), install_requires=['requests'], extras_require={'test': ['mock']}, classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Operating System :: OS Independent', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], url='https://github.com/auth0/auth0-python', )
import io import os import re from setuptools import setup, find_packages def find_version(): file_dir = os.path.dirname(__file__) with io.open(os.path.join(file_dir, 'auth0', '__init__.py')) as f: version = re.search(r'^__version__ = [\'"]([^\'"]*)[\'"]', f.read()) if version: return version.group(1) else: raise RuntimeError("Unable to find version string.") with io.open('README.rst', encoding='utf-8') as f: long_description = f.read() setup( name='auth0-python', version=find_version(), description='Auth0 Python SDK', long_description=long_description, author='Auth0', author_email='support@auth0.com', license='MIT', packages=find_packages(), install_requires=['requests'], extras_require={'test': ['mock']}, classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Operating System :: OS Independent', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], url='https://github.com/auth0/auth0-python', )
Use io.open with UTF-8 encoding to avoid UnicodeDecodeError
Use io.open with UTF-8 encoding to avoid UnicodeDecodeError
Python
mit
auth0/auth0-python,auth0/auth0-python
import io import os import re from setuptools import setup, find_packages def find_version(): file_dir = os.path.dirname(__file__) with io.open(os.path.join(file_dir, 'auth0', '__init__.py')) as f: version = re.search(r'^__version__ = [\'"]([^\'"]*)[\'"]', f.read()) if version: return version.group(1) else: raise RuntimeError("Unable to find version string.") with open('README.rst') as f: long_description = f.read() setup( name='auth0-python', version=find_version(), description='Auth0 Python SDK', long_description=long_description, author='Auth0', author_email='support@auth0.com', license='MIT', packages=find_packages(), install_requires=['requests'], extras_require={'test': ['mock']}, classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Operating System :: OS Independent', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], url='https://github.com/auth0/auth0-python', ) Use io.open with UTF-8 encoding to avoid UnicodeDecodeError
import io import os import re from setuptools import setup, find_packages def find_version(): file_dir = os.path.dirname(__file__) with io.open(os.path.join(file_dir, 'auth0', '__init__.py')) as f: version = re.search(r'^__version__ = [\'"]([^\'"]*)[\'"]', f.read()) if version: return version.group(1) else: raise RuntimeError("Unable to find version string.") with io.open('README.rst', encoding='utf-8') as f: long_description = f.read() setup( name='auth0-python', version=find_version(), description='Auth0 Python SDK', long_description=long_description, author='Auth0', author_email='support@auth0.com', license='MIT', packages=find_packages(), install_requires=['requests'], extras_require={'test': ['mock']}, classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Operating System :: OS Independent', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], url='https://github.com/auth0/auth0-python', )
<commit_before>import io import os import re from setuptools import setup, find_packages def find_version(): file_dir = os.path.dirname(__file__) with io.open(os.path.join(file_dir, 'auth0', '__init__.py')) as f: version = re.search(r'^__version__ = [\'"]([^\'"]*)[\'"]', f.read()) if version: return version.group(1) else: raise RuntimeError("Unable to find version string.") with open('README.rst') as f: long_description = f.read() setup( name='auth0-python', version=find_version(), description='Auth0 Python SDK', long_description=long_description, author='Auth0', author_email='support@auth0.com', license='MIT', packages=find_packages(), install_requires=['requests'], extras_require={'test': ['mock']}, classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Operating System :: OS Independent', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], url='https://github.com/auth0/auth0-python', ) <commit_msg>Use io.open with UTF-8 encoding to avoid UnicodeDecodeError<commit_after>
import io import os import re from setuptools import setup, find_packages def find_version(): file_dir = os.path.dirname(__file__) with io.open(os.path.join(file_dir, 'auth0', '__init__.py')) as f: version = re.search(r'^__version__ = [\'"]([^\'"]*)[\'"]', f.read()) if version: return version.group(1) else: raise RuntimeError("Unable to find version string.") with io.open('README.rst', encoding='utf-8') as f: long_description = f.read() setup( name='auth0-python', version=find_version(), description='Auth0 Python SDK', long_description=long_description, author='Auth0', author_email='support@auth0.com', license='MIT', packages=find_packages(), install_requires=['requests'], extras_require={'test': ['mock']}, classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Operating System :: OS Independent', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], url='https://github.com/auth0/auth0-python', )
import io import os import re from setuptools import setup, find_packages def find_version(): file_dir = os.path.dirname(__file__) with io.open(os.path.join(file_dir, 'auth0', '__init__.py')) as f: version = re.search(r'^__version__ = [\'"]([^\'"]*)[\'"]', f.read()) if version: return version.group(1) else: raise RuntimeError("Unable to find version string.") with open('README.rst') as f: long_description = f.read() setup( name='auth0-python', version=find_version(), description='Auth0 Python SDK', long_description=long_description, author='Auth0', author_email='support@auth0.com', license='MIT', packages=find_packages(), install_requires=['requests'], extras_require={'test': ['mock']}, classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Operating System :: OS Independent', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], url='https://github.com/auth0/auth0-python', ) Use io.open with UTF-8 encoding to avoid UnicodeDecodeErrorimport io import os import re from setuptools import setup, find_packages def find_version(): file_dir = os.path.dirname(__file__) with io.open(os.path.join(file_dir, 'auth0', '__init__.py')) as f: version = re.search(r'^__version__ = [\'"]([^\'"]*)[\'"]', f.read()) if version: return version.group(1) else: raise RuntimeError("Unable to find version string.") with io.open('README.rst', encoding='utf-8') as f: long_description = f.read() setup( name='auth0-python', version=find_version(), description='Auth0 Python SDK', long_description=long_description, author='Auth0', author_email='support@auth0.com', license='MIT', packages=find_packages(), install_requires=['requests'], extras_require={'test': ['mock']}, classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Operating System :: OS Independent', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], url='https://github.com/auth0/auth0-python', )
<commit_before>import io import os import re from setuptools import setup, find_packages def find_version(): file_dir = os.path.dirname(__file__) with io.open(os.path.join(file_dir, 'auth0', '__init__.py')) as f: version = re.search(r'^__version__ = [\'"]([^\'"]*)[\'"]', f.read()) if version: return version.group(1) else: raise RuntimeError("Unable to find version string.") with open('README.rst') as f: long_description = f.read() setup( name='auth0-python', version=find_version(), description='Auth0 Python SDK', long_description=long_description, author='Auth0', author_email='support@auth0.com', license='MIT', packages=find_packages(), install_requires=['requests'], extras_require={'test': ['mock']}, classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Operating System :: OS Independent', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], url='https://github.com/auth0/auth0-python', ) <commit_msg>Use io.open with UTF-8 encoding to avoid UnicodeDecodeError<commit_after>import io import os import re from setuptools import setup, find_packages def find_version(): file_dir = os.path.dirname(__file__) with io.open(os.path.join(file_dir, 'auth0', '__init__.py')) as f: version = re.search(r'^__version__ = [\'"]([^\'"]*)[\'"]', f.read()) if version: return version.group(1) else: raise RuntimeError("Unable to find version string.") with io.open('README.rst', encoding='utf-8') as f: long_description = f.read() setup( name='auth0-python', version=find_version(), description='Auth0 Python SDK', long_description=long_description, author='Auth0', author_email='support@auth0.com', license='MIT', packages=find_packages(), install_requires=['requests'], extras_require={'test': ['mock']}, classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Operating System :: OS Independent', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], url='https://github.com/auth0/auth0-python', )
8a3d4015136d6aebccb092208f7c6d02b5c93e13
setup.py
setup.py
from setuptools import setup, find_packages setup( name='ssbio', version='0.1', author='Nathan Mih', author_email='nmih@ucsd.edu', license='MIT', url='http://github.com/nmih/ssbio', description='Various tools and functions to enable structural systems biology', packages=find_packages(), package_dir={'ssbio': 'ssbio'}, package_data={'ssbio': ['ssbio/etc/*']}, scripts = ['ssbio/scripts/cleanpdb', 'ssbio/scripts/aggprop', 'ssbio/scripts/thermostability', 'ssbio/structure/mutatepdb.py', 'ssbio/structure/tleap.py', 'ssbio/databases/drugbank.py', 'ssbio/structure/properties/msmsprops.py', 'ssbio/dock/dock.py'], long_description=open('README.md').read(), install_requires=['biopython', 'numpy', 'tqdm', 'pandas', 'requests', 'cachetools', 'bioservices', 'prody', 'xmltodict'] )
from setuptools import setup, find_packages setup( name='ssbio', version='0.1', author='Nathan Mih', author_email='nmih@ucsd.edu', license='MIT', url='http://github.com/nmih/ssbio', description='Various tools and functions to enable structural systems biology', packages=find_packages(), package_dir={'ssbio': 'ssbio'}, package_data={'ssbio': ['ssbio/etc/*']}, scripts = ['ssbio/scripts/cleanpdb', 'ssbio/scripts/aggprop', 'ssbio/scripts/thermostability', 'ssbio/structure/mutatepdb.py', 'ssbio/structure/tleap.py', 'ssbio/databases/drugbank.py', 'ssbio/structure/properties/msmsprops.py', 'ssbio/dock/dock.py'], long_description=open('README.md').read(), install_requires=['biopython', 'numpy', 'tqdm', 'pandas', 'requests', 'cachetools', 'bioservices', 'prody', 'xmltodict', 'bidict'] )
Add bidict to required packages
Add bidict to required packages
Python
mit
SBRG/ssbio,nmih/ssbio,nmih/ssbio,SBRG/ssbio
from setuptools import setup, find_packages setup( name='ssbio', version='0.1', author='Nathan Mih', author_email='nmih@ucsd.edu', license='MIT', url='http://github.com/nmih/ssbio', description='Various tools and functions to enable structural systems biology', packages=find_packages(), package_dir={'ssbio': 'ssbio'}, package_data={'ssbio': ['ssbio/etc/*']}, scripts = ['ssbio/scripts/cleanpdb', 'ssbio/scripts/aggprop', 'ssbio/scripts/thermostability', 'ssbio/structure/mutatepdb.py', 'ssbio/structure/tleap.py', 'ssbio/databases/drugbank.py', 'ssbio/structure/properties/msmsprops.py', 'ssbio/dock/dock.py'], long_description=open('README.md').read(), install_requires=['biopython', 'numpy', 'tqdm', 'pandas', 'requests', 'cachetools', 'bioservices', 'prody', 'xmltodict'] )Add bidict to required packages
from setuptools import setup, find_packages setup( name='ssbio', version='0.1', author='Nathan Mih', author_email='nmih@ucsd.edu', license='MIT', url='http://github.com/nmih/ssbio', description='Various tools and functions to enable structural systems biology', packages=find_packages(), package_dir={'ssbio': 'ssbio'}, package_data={'ssbio': ['ssbio/etc/*']}, scripts = ['ssbio/scripts/cleanpdb', 'ssbio/scripts/aggprop', 'ssbio/scripts/thermostability', 'ssbio/structure/mutatepdb.py', 'ssbio/structure/tleap.py', 'ssbio/databases/drugbank.py', 'ssbio/structure/properties/msmsprops.py', 'ssbio/dock/dock.py'], long_description=open('README.md').read(), install_requires=['biopython', 'numpy', 'tqdm', 'pandas', 'requests', 'cachetools', 'bioservices', 'prody', 'xmltodict', 'bidict'] )
<commit_before>from setuptools import setup, find_packages setup( name='ssbio', version='0.1', author='Nathan Mih', author_email='nmih@ucsd.edu', license='MIT', url='http://github.com/nmih/ssbio', description='Various tools and functions to enable structural systems biology', packages=find_packages(), package_dir={'ssbio': 'ssbio'}, package_data={'ssbio': ['ssbio/etc/*']}, scripts = ['ssbio/scripts/cleanpdb', 'ssbio/scripts/aggprop', 'ssbio/scripts/thermostability', 'ssbio/structure/mutatepdb.py', 'ssbio/structure/tleap.py', 'ssbio/databases/drugbank.py', 'ssbio/structure/properties/msmsprops.py', 'ssbio/dock/dock.py'], long_description=open('README.md').read(), install_requires=['biopython', 'numpy', 'tqdm', 'pandas', 'requests', 'cachetools', 'bioservices', 'prody', 'xmltodict'] )<commit_msg>Add bidict to required packages<commit_after>
from setuptools import setup, find_packages setup( name='ssbio', version='0.1', author='Nathan Mih', author_email='nmih@ucsd.edu', license='MIT', url='http://github.com/nmih/ssbio', description='Various tools and functions to enable structural systems biology', packages=find_packages(), package_dir={'ssbio': 'ssbio'}, package_data={'ssbio': ['ssbio/etc/*']}, scripts = ['ssbio/scripts/cleanpdb', 'ssbio/scripts/aggprop', 'ssbio/scripts/thermostability', 'ssbio/structure/mutatepdb.py', 'ssbio/structure/tleap.py', 'ssbio/databases/drugbank.py', 'ssbio/structure/properties/msmsprops.py', 'ssbio/dock/dock.py'], long_description=open('README.md').read(), install_requires=['biopython', 'numpy', 'tqdm', 'pandas', 'requests', 'cachetools', 'bioservices', 'prody', 'xmltodict', 'bidict'] )
from setuptools import setup, find_packages setup( name='ssbio', version='0.1', author='Nathan Mih', author_email='nmih@ucsd.edu', license='MIT', url='http://github.com/nmih/ssbio', description='Various tools and functions to enable structural systems biology', packages=find_packages(), package_dir={'ssbio': 'ssbio'}, package_data={'ssbio': ['ssbio/etc/*']}, scripts = ['ssbio/scripts/cleanpdb', 'ssbio/scripts/aggprop', 'ssbio/scripts/thermostability', 'ssbio/structure/mutatepdb.py', 'ssbio/structure/tleap.py', 'ssbio/databases/drugbank.py', 'ssbio/structure/properties/msmsprops.py', 'ssbio/dock/dock.py'], long_description=open('README.md').read(), install_requires=['biopython', 'numpy', 'tqdm', 'pandas', 'requests', 'cachetools', 'bioservices', 'prody', 'xmltodict'] )Add bidict to required packagesfrom setuptools import setup, find_packages setup( name='ssbio', version='0.1', author='Nathan Mih', author_email='nmih@ucsd.edu', license='MIT', url='http://github.com/nmih/ssbio', description='Various tools and functions to enable structural systems biology', packages=find_packages(), package_dir={'ssbio': 'ssbio'}, package_data={'ssbio': ['ssbio/etc/*']}, scripts = ['ssbio/scripts/cleanpdb', 'ssbio/scripts/aggprop', 'ssbio/scripts/thermostability', 'ssbio/structure/mutatepdb.py', 'ssbio/structure/tleap.py', 'ssbio/databases/drugbank.py', 'ssbio/structure/properties/msmsprops.py', 'ssbio/dock/dock.py'], long_description=open('README.md').read(), install_requires=['biopython', 'numpy', 'tqdm', 'pandas', 'requests', 'cachetools', 'bioservices', 'prody', 'xmltodict', 'bidict'] )
<commit_before>from setuptools import setup, find_packages setup( name='ssbio', version='0.1', author='Nathan Mih', author_email='nmih@ucsd.edu', license='MIT', url='http://github.com/nmih/ssbio', description='Various tools and functions to enable structural systems biology', packages=find_packages(), package_dir={'ssbio': 'ssbio'}, package_data={'ssbio': ['ssbio/etc/*']}, scripts = ['ssbio/scripts/cleanpdb', 'ssbio/scripts/aggprop', 'ssbio/scripts/thermostability', 'ssbio/structure/mutatepdb.py', 'ssbio/structure/tleap.py', 'ssbio/databases/drugbank.py', 'ssbio/structure/properties/msmsprops.py', 'ssbio/dock/dock.py'], long_description=open('README.md').read(), install_requires=['biopython', 'numpy', 'tqdm', 'pandas', 'requests', 'cachetools', 'bioservices', 'prody', 'xmltodict'] )<commit_msg>Add bidict to required packages<commit_after>from setuptools import setup, find_packages setup( name='ssbio', version='0.1', author='Nathan Mih', author_email='nmih@ucsd.edu', license='MIT', url='http://github.com/nmih/ssbio', description='Various tools and functions to enable structural systems biology', packages=find_packages(), package_dir={'ssbio': 'ssbio'}, package_data={'ssbio': ['ssbio/etc/*']}, scripts = ['ssbio/scripts/cleanpdb', 'ssbio/scripts/aggprop', 'ssbio/scripts/thermostability', 'ssbio/structure/mutatepdb.py', 'ssbio/structure/tleap.py', 'ssbio/databases/drugbank.py', 'ssbio/structure/properties/msmsprops.py', 'ssbio/dock/dock.py'], long_description=open('README.md').read(), install_requires=['biopython', 'numpy', 'tqdm', 'pandas', 'requests', 'cachetools', 'bioservices', 'prody', 'xmltodict', 'bidict'] )
695743c3a887224c212b434de7d6e2ccf08d4620
setup.py
setup.py
#!/usr/bin/env python from distutils.core import setup setup( name='wagtail_embed_videos', version='0.0.5', description='Embed Videos for Wagtail CMS.', long_description=README, author='Diogo Marques', author_email='doriva.marques.29@gmail.com', maintainer='Diogo Marques', maintainer_email='doriva.marques.29@gmail.com', url='https://github.com/infoportugal/wagtail-embedvideos', packages=['wagtail_embed_videos', 'wagtail_embed_videos.views'], package_data={'wagtail_embed_videos': ['static/wagtail_embed_videos/js/*.js']}, requires=['django(>=1.7)', 'wagtail(>=1.0)', 'django-embed-video(>=1.0)'], install_requires=['wagtail', 'django-embed-video'] classifiers=[ 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Operating System :: OS Independent', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'Framework :: Django', 'Framework :: Wagtail CMS', 'License :: OSI Approved :: BSD License'], license='New BSD', )
#!/usr/bin/env python from distutils.core import setup setup( name='wagtail_embed_videos', version='0.0.5', description='Embed Videos for Wagtail CMS.', long_description=README, author='Diogo Marques', author_email='doriva.marques.29@gmail.com', maintainer='Diogo Marques', maintainer_email='doriva.marques.29@gmail.com', url='https://github.com/infoportugal/wagtail-embedvideos', packages=['wagtail_embed_videos', 'wagtail_embed_videos.views'], package_data={'wagtail_embed_videos': ['static/wagtail_embed_videos/js/*.js']}, requires=['django(>=1.7)', 'wagtail(>=1.0)', 'django-embed-video(>=1.0)'], install_requires=['wagtail', 'django-embed-video'], classifiers=[ 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Operating System :: OS Independent', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'Framework :: Django', 'Framework :: Wagtail CMS', 'License :: OSI Approved :: BSD License'], license='New BSD', )
Fix other SyntaxError: invalid syntax
Fix other SyntaxError: invalid syntax Fix Traceback (most recent call last): File "<string>", line 20, in <module> File "/tmp/pip-7dj0aup6-build/setup.py", line 19 classifiers=[ ^ SyntaxError: invalid syntax
Python
bsd-3-clause
SalahAdDin/wagtail-embedvideos,SalahAdDin/wagtail-embedvideos,infoportugal/wagtail-embedvideos,infoportugal/wagtail-embedvideos,SalahAdDin/wagtail-embedvideos,infoportugal/wagtail-embedvideos
#!/usr/bin/env python from distutils.core import setup setup( name='wagtail_embed_videos', version='0.0.5', description='Embed Videos for Wagtail CMS.', long_description=README, author='Diogo Marques', author_email='doriva.marques.29@gmail.com', maintainer='Diogo Marques', maintainer_email='doriva.marques.29@gmail.com', url='https://github.com/infoportugal/wagtail-embedvideos', packages=['wagtail_embed_videos', 'wagtail_embed_videos.views'], package_data={'wagtail_embed_videos': ['static/wagtail_embed_videos/js/*.js']}, requires=['django(>=1.7)', 'wagtail(>=1.0)', 'django-embed-video(>=1.0)'], install_requires=['wagtail', 'django-embed-video'] classifiers=[ 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Operating System :: OS Independent', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'Framework :: Django', 'Framework :: Wagtail CMS', 'License :: OSI Approved :: BSD License'], license='New BSD', ) Fix other SyntaxError: invalid syntax Fix Traceback (most recent call last): File "<string>", line 20, in <module> File "/tmp/pip-7dj0aup6-build/setup.py", line 19 classifiers=[ ^ SyntaxError: invalid syntax
#!/usr/bin/env python from distutils.core import setup setup( name='wagtail_embed_videos', version='0.0.5', description='Embed Videos for Wagtail CMS.', long_description=README, author='Diogo Marques', author_email='doriva.marques.29@gmail.com', maintainer='Diogo Marques', maintainer_email='doriva.marques.29@gmail.com', url='https://github.com/infoportugal/wagtail-embedvideos', packages=['wagtail_embed_videos', 'wagtail_embed_videos.views'], package_data={'wagtail_embed_videos': ['static/wagtail_embed_videos/js/*.js']}, requires=['django(>=1.7)', 'wagtail(>=1.0)', 'django-embed-video(>=1.0)'], install_requires=['wagtail', 'django-embed-video'], classifiers=[ 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Operating System :: OS Independent', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'Framework :: Django', 'Framework :: Wagtail CMS', 'License :: OSI Approved :: BSD License'], license='New BSD', )
<commit_before>#!/usr/bin/env python from distutils.core import setup setup( name='wagtail_embed_videos', version='0.0.5', description='Embed Videos for Wagtail CMS.', long_description=README, author='Diogo Marques', author_email='doriva.marques.29@gmail.com', maintainer='Diogo Marques', maintainer_email='doriva.marques.29@gmail.com', url='https://github.com/infoportugal/wagtail-embedvideos', packages=['wagtail_embed_videos', 'wagtail_embed_videos.views'], package_data={'wagtail_embed_videos': ['static/wagtail_embed_videos/js/*.js']}, requires=['django(>=1.7)', 'wagtail(>=1.0)', 'django-embed-video(>=1.0)'], install_requires=['wagtail', 'django-embed-video'] classifiers=[ 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Operating System :: OS Independent', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'Framework :: Django', 'Framework :: Wagtail CMS', 'License :: OSI Approved :: BSD License'], license='New BSD', ) <commit_msg>Fix other SyntaxError: invalid syntax Fix Traceback (most recent call last): File "<string>", line 20, in <module> File "/tmp/pip-7dj0aup6-build/setup.py", line 19 classifiers=[ ^ SyntaxError: invalid syntax<commit_after>
#!/usr/bin/env python from distutils.core import setup setup( name='wagtail_embed_videos', version='0.0.5', description='Embed Videos for Wagtail CMS.', long_description=README, author='Diogo Marques', author_email='doriva.marques.29@gmail.com', maintainer='Diogo Marques', maintainer_email='doriva.marques.29@gmail.com', url='https://github.com/infoportugal/wagtail-embedvideos', packages=['wagtail_embed_videos', 'wagtail_embed_videos.views'], package_data={'wagtail_embed_videos': ['static/wagtail_embed_videos/js/*.js']}, requires=['django(>=1.7)', 'wagtail(>=1.0)', 'django-embed-video(>=1.0)'], install_requires=['wagtail', 'django-embed-video'], classifiers=[ 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Operating System :: OS Independent', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'Framework :: Django', 'Framework :: Wagtail CMS', 'License :: OSI Approved :: BSD License'], license='New BSD', )
#!/usr/bin/env python from distutils.core import setup setup( name='wagtail_embed_videos', version='0.0.5', description='Embed Videos for Wagtail CMS.', long_description=README, author='Diogo Marques', author_email='doriva.marques.29@gmail.com', maintainer='Diogo Marques', maintainer_email='doriva.marques.29@gmail.com', url='https://github.com/infoportugal/wagtail-embedvideos', packages=['wagtail_embed_videos', 'wagtail_embed_videos.views'], package_data={'wagtail_embed_videos': ['static/wagtail_embed_videos/js/*.js']}, requires=['django(>=1.7)', 'wagtail(>=1.0)', 'django-embed-video(>=1.0)'], install_requires=['wagtail', 'django-embed-video'] classifiers=[ 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Operating System :: OS Independent', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'Framework :: Django', 'Framework :: Wagtail CMS', 'License :: OSI Approved :: BSD License'], license='New BSD', ) Fix other SyntaxError: invalid syntax Fix Traceback (most recent call last): File "<string>", line 20, in <module> File "/tmp/pip-7dj0aup6-build/setup.py", line 19 classifiers=[ ^ SyntaxError: invalid syntax#!/usr/bin/env python from distutils.core import setup setup( name='wagtail_embed_videos', version='0.0.5', description='Embed Videos for Wagtail CMS.', long_description=README, author='Diogo Marques', author_email='doriva.marques.29@gmail.com', maintainer='Diogo Marques', maintainer_email='doriva.marques.29@gmail.com', url='https://github.com/infoportugal/wagtail-embedvideos', packages=['wagtail_embed_videos', 'wagtail_embed_videos.views'], package_data={'wagtail_embed_videos': ['static/wagtail_embed_videos/js/*.js']}, requires=['django(>=1.7)', 'wagtail(>=1.0)', 'django-embed-video(>=1.0)'], install_requires=['wagtail', 'django-embed-video'], classifiers=[ 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Operating System :: OS Independent', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'Framework :: Django', 'Framework :: Wagtail CMS', 'License :: OSI Approved :: BSD License'], license='New BSD', )
<commit_before>#!/usr/bin/env python from distutils.core import setup setup( name='wagtail_embed_videos', version='0.0.5', description='Embed Videos for Wagtail CMS.', long_description=README, author='Diogo Marques', author_email='doriva.marques.29@gmail.com', maintainer='Diogo Marques', maintainer_email='doriva.marques.29@gmail.com', url='https://github.com/infoportugal/wagtail-embedvideos', packages=['wagtail_embed_videos', 'wagtail_embed_videos.views'], package_data={'wagtail_embed_videos': ['static/wagtail_embed_videos/js/*.js']}, requires=['django(>=1.7)', 'wagtail(>=1.0)', 'django-embed-video(>=1.0)'], install_requires=['wagtail', 'django-embed-video'] classifiers=[ 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Operating System :: OS Independent', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'Framework :: Django', 'Framework :: Wagtail CMS', 'License :: OSI Approved :: BSD License'], license='New BSD', ) <commit_msg>Fix other SyntaxError: invalid syntax Fix Traceback (most recent call last): File "<string>", line 20, in <module> File "/tmp/pip-7dj0aup6-build/setup.py", line 19 classifiers=[ ^ SyntaxError: invalid syntax<commit_after>#!/usr/bin/env python from distutils.core import setup setup( name='wagtail_embed_videos', version='0.0.5', description='Embed Videos for Wagtail CMS.', long_description=README, author='Diogo Marques', author_email='doriva.marques.29@gmail.com', maintainer='Diogo Marques', maintainer_email='doriva.marques.29@gmail.com', url='https://github.com/infoportugal/wagtail-embedvideos', packages=['wagtail_embed_videos', 'wagtail_embed_videos.views'], package_data={'wagtail_embed_videos': ['static/wagtail_embed_videos/js/*.js']}, requires=['django(>=1.7)', 'wagtail(>=1.0)', 'django-embed-video(>=1.0)'], install_requires=['wagtail', 'django-embed-video'], classifiers=[ 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Operating System :: OS Independent', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'Framework :: Django', 'Framework :: Wagtail CMS', 'License :: OSI Approved :: BSD License'], license='New BSD', )
55745f668715c294cd5662712b2d1ccb7726f125
setup.py
setup.py
from distutils.core import setup import multi_schema setup( name = "django-multi-schema", version = multi_schema.__version__, description = "Postgres schema support in django.", url = "http://hg.schinckel.net/django-multi-schema", author = "Matthew Schinckel", author_email = "matt@schinckel.net", packages = [ "multi_schema", ], classifiers = [ 'Programming Language :: Python', 'Operating System :: OS Independent', 'Framework :: Django', ], )
from distutils.core import setup import multi_schema setup( name = "django-multi-schema", version = multi_schema.__version__, description = "Postgres schema support in django.", url = "http://hg.schinckel.net/django-multi-schema", author = "Matthew Schinckel", author_email = "matt@schinckel.net", packages = [ "multi_schema", ], install_requires = [ 'south==0.7.4', ], classifiers = [ 'Programming Language :: Python', 'Operating System :: OS Independent', 'Framework :: Django', ], )
Add south as a dependency, so we can apply a version. Does not need to be installed in INSTALLED_APPS.
Add south as a dependency, so we can apply a version. Does not need to be installed in INSTALLED_APPS.
Python
bsd-3-clause
schinckel/django-boardinghouse,schinckel/django-boardinghouse,schinckel/django-boardinghouse
from distutils.core import setup import multi_schema setup( name = "django-multi-schema", version = multi_schema.__version__, description = "Postgres schema support in django.", url = "http://hg.schinckel.net/django-multi-schema", author = "Matthew Schinckel", author_email = "matt@schinckel.net", packages = [ "multi_schema", ], classifiers = [ 'Programming Language :: Python', 'Operating System :: OS Independent', 'Framework :: Django', ], ) Add south as a dependency, so we can apply a version. Does not need to be installed in INSTALLED_APPS.
from distutils.core import setup import multi_schema setup( name = "django-multi-schema", version = multi_schema.__version__, description = "Postgres schema support in django.", url = "http://hg.schinckel.net/django-multi-schema", author = "Matthew Schinckel", author_email = "matt@schinckel.net", packages = [ "multi_schema", ], install_requires = [ 'south==0.7.4', ], classifiers = [ 'Programming Language :: Python', 'Operating System :: OS Independent', 'Framework :: Django', ], )
<commit_before>from distutils.core import setup import multi_schema setup( name = "django-multi-schema", version = multi_schema.__version__, description = "Postgres schema support in django.", url = "http://hg.schinckel.net/django-multi-schema", author = "Matthew Schinckel", author_email = "matt@schinckel.net", packages = [ "multi_schema", ], classifiers = [ 'Programming Language :: Python', 'Operating System :: OS Independent', 'Framework :: Django', ], ) <commit_msg>Add south as a dependency, so we can apply a version. Does not need to be installed in INSTALLED_APPS.<commit_after>
from distutils.core import setup import multi_schema setup( name = "django-multi-schema", version = multi_schema.__version__, description = "Postgres schema support in django.", url = "http://hg.schinckel.net/django-multi-schema", author = "Matthew Schinckel", author_email = "matt@schinckel.net", packages = [ "multi_schema", ], install_requires = [ 'south==0.7.4', ], classifiers = [ 'Programming Language :: Python', 'Operating System :: OS Independent', 'Framework :: Django', ], )
from distutils.core import setup import multi_schema setup( name = "django-multi-schema", version = multi_schema.__version__, description = "Postgres schema support in django.", url = "http://hg.schinckel.net/django-multi-schema", author = "Matthew Schinckel", author_email = "matt@schinckel.net", packages = [ "multi_schema", ], classifiers = [ 'Programming Language :: Python', 'Operating System :: OS Independent', 'Framework :: Django', ], ) Add south as a dependency, so we can apply a version. Does not need to be installed in INSTALLED_APPS.from distutils.core import setup import multi_schema setup( name = "django-multi-schema", version = multi_schema.__version__, description = "Postgres schema support in django.", url = "http://hg.schinckel.net/django-multi-schema", author = "Matthew Schinckel", author_email = "matt@schinckel.net", packages = [ "multi_schema", ], install_requires = [ 'south==0.7.4', ], classifiers = [ 'Programming Language :: Python', 'Operating System :: OS Independent', 'Framework :: Django', ], )
<commit_before>from distutils.core import setup import multi_schema setup( name = "django-multi-schema", version = multi_schema.__version__, description = "Postgres schema support in django.", url = "http://hg.schinckel.net/django-multi-schema", author = "Matthew Schinckel", author_email = "matt@schinckel.net", packages = [ "multi_schema", ], classifiers = [ 'Programming Language :: Python', 'Operating System :: OS Independent', 'Framework :: Django', ], ) <commit_msg>Add south as a dependency, so we can apply a version. Does not need to be installed in INSTALLED_APPS.<commit_after>from distutils.core import setup import multi_schema setup( name = "django-multi-schema", version = multi_schema.__version__, description = "Postgres schema support in django.", url = "http://hg.schinckel.net/django-multi-schema", author = "Matthew Schinckel", author_email = "matt@schinckel.net", packages = [ "multi_schema", ], install_requires = [ 'south==0.7.4', ], classifiers = [ 'Programming Language :: Python', 'Operating System :: OS Independent', 'Framework :: Django', ], )
a84eca4bffd30c068194cc6c6d8176178fe26e78
setup.py
setup.py
#!/usr/bin/env python from setuptools import setup, find_packages from os.path import dirname, join import sys, os # When creating the sdist, make sure the django.mo file also exists: if 'sdist' in sys.argv: try: os.chdir('fluent_comments') from django.core.management.commands.compilemessages import compile_messages compile_messages(sys.stderr) finally: os.chdir('..') setup( name='django-fluent-comments', version='0.8.0', license='Apache License, Version 2.0', install_requires=[ 'Django>=1.2.0', 'django-crispy-forms>=1.1.1', 'akismet>=0.2', ], description='A modern, ajax-based appearance for django.contrib.comments', long_description=open('README.rst').read(), author='Diederik van der Boor', author_email='opensource@edoburu.nl', url='https://github.com/edoburu/django-fluent-comments', download_url='https://github.com/edoburu/django-fluent-comments/zipball/master', packages=find_packages(exclude=('example*',)), include_package_data=True, zip_safe=False, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ] )
#!/usr/bin/env python from setuptools import setup, find_packages from os.path import dirname, join import sys, os # When creating the sdist, make sure the django.mo file also exists: if 'sdist' in sys.argv: try: os.chdir('fluent_comments') from django.core.management.commands.compilemessages import compile_messages compile_messages(sys.stderr) finally: os.chdir('..') setup( name='django-fluent-comments', version='0.8.0', license='Apache License, Version 2.0', install_requires=[ 'django-crispy-forms>=1.1.1', 'akismet>=0.2', ], requires=[ 'Django (>=1.2)', ], description='A modern, ajax-based appearance for django.contrib.comments', long_description=open('README.rst').read(), author='Diederik van der Boor', author_email='opensource@edoburu.nl', url='https://github.com/edoburu/django-fluent-comments', download_url='https://github.com/edoburu/django-fluent-comments/zipball/master', packages=find_packages(exclude=('example*',)), include_package_data=True, zip_safe=False, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ] )
Move Django out of `install_requires` to to `requires` block.
Move Django out of `install_requires` to to `requires` block.
Python
apache-2.0
Afnarel/django-fluent-comments,BangorUniversity/django-fluent-comments,mgpyh/django-fluent-comments,akszydelko/django-fluent-comments,PetrDlouhy/django-fluent-comments,django-fluent/django-fluent-comments,mgpyh/django-fluent-comments,PetrDlouhy/django-fluent-comments,PetrDlouhy/django-fluent-comments,Afnarel/django-fluent-comments,edoburu/django-fluent-comments,edoburu/django-fluent-comments,django-fluent/django-fluent-comments,BangorUniversity/django-fluent-comments,akszydelko/django-fluent-comments,akszydelko/django-fluent-comments,django-fluent/django-fluent-comments,mgpyh/django-fluent-comments,Afnarel/django-fluent-comments,edoburu/django-fluent-comments,django-fluent/django-fluent-comments,BangorUniversity/django-fluent-comments
#!/usr/bin/env python from setuptools import setup, find_packages from os.path import dirname, join import sys, os # When creating the sdist, make sure the django.mo file also exists: if 'sdist' in sys.argv: try: os.chdir('fluent_comments') from django.core.management.commands.compilemessages import compile_messages compile_messages(sys.stderr) finally: os.chdir('..') setup( name='django-fluent-comments', version='0.8.0', license='Apache License, Version 2.0', install_requires=[ 'Django>=1.2.0', 'django-crispy-forms>=1.1.1', 'akismet>=0.2', ], description='A modern, ajax-based appearance for django.contrib.comments', long_description=open('README.rst').read(), author='Diederik van der Boor', author_email='opensource@edoburu.nl', url='https://github.com/edoburu/django-fluent-comments', download_url='https://github.com/edoburu/django-fluent-comments/zipball/master', packages=find_packages(exclude=('example*',)), include_package_data=True, zip_safe=False, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ] ) Move Django out of `install_requires` to to `requires` block.
#!/usr/bin/env python from setuptools import setup, find_packages from os.path import dirname, join import sys, os # When creating the sdist, make sure the django.mo file also exists: if 'sdist' in sys.argv: try: os.chdir('fluent_comments') from django.core.management.commands.compilemessages import compile_messages compile_messages(sys.stderr) finally: os.chdir('..') setup( name='django-fluent-comments', version='0.8.0', license='Apache License, Version 2.0', install_requires=[ 'django-crispy-forms>=1.1.1', 'akismet>=0.2', ], requires=[ 'Django (>=1.2)', ], description='A modern, ajax-based appearance for django.contrib.comments', long_description=open('README.rst').read(), author='Diederik van der Boor', author_email='opensource@edoburu.nl', url='https://github.com/edoburu/django-fluent-comments', download_url='https://github.com/edoburu/django-fluent-comments/zipball/master', packages=find_packages(exclude=('example*',)), include_package_data=True, zip_safe=False, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ] )
<commit_before>#!/usr/bin/env python from setuptools import setup, find_packages from os.path import dirname, join import sys, os # When creating the sdist, make sure the django.mo file also exists: if 'sdist' in sys.argv: try: os.chdir('fluent_comments') from django.core.management.commands.compilemessages import compile_messages compile_messages(sys.stderr) finally: os.chdir('..') setup( name='django-fluent-comments', version='0.8.0', license='Apache License, Version 2.0', install_requires=[ 'Django>=1.2.0', 'django-crispy-forms>=1.1.1', 'akismet>=0.2', ], description='A modern, ajax-based appearance for django.contrib.comments', long_description=open('README.rst').read(), author='Diederik van der Boor', author_email='opensource@edoburu.nl', url='https://github.com/edoburu/django-fluent-comments', download_url='https://github.com/edoburu/django-fluent-comments/zipball/master', packages=find_packages(exclude=('example*',)), include_package_data=True, zip_safe=False, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ] ) <commit_msg>Move Django out of `install_requires` to to `requires` block.<commit_after>
#!/usr/bin/env python from setuptools import setup, find_packages from os.path import dirname, join import sys, os # When creating the sdist, make sure the django.mo file also exists: if 'sdist' in sys.argv: try: os.chdir('fluent_comments') from django.core.management.commands.compilemessages import compile_messages compile_messages(sys.stderr) finally: os.chdir('..') setup( name='django-fluent-comments', version='0.8.0', license='Apache License, Version 2.0', install_requires=[ 'django-crispy-forms>=1.1.1', 'akismet>=0.2', ], requires=[ 'Django (>=1.2)', ], description='A modern, ajax-based appearance for django.contrib.comments', long_description=open('README.rst').read(), author='Diederik van der Boor', author_email='opensource@edoburu.nl', url='https://github.com/edoburu/django-fluent-comments', download_url='https://github.com/edoburu/django-fluent-comments/zipball/master', packages=find_packages(exclude=('example*',)), include_package_data=True, zip_safe=False, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ] )
#!/usr/bin/env python from setuptools import setup, find_packages from os.path import dirname, join import sys, os # When creating the sdist, make sure the django.mo file also exists: if 'sdist' in sys.argv: try: os.chdir('fluent_comments') from django.core.management.commands.compilemessages import compile_messages compile_messages(sys.stderr) finally: os.chdir('..') setup( name='django-fluent-comments', version='0.8.0', license='Apache License, Version 2.0', install_requires=[ 'Django>=1.2.0', 'django-crispy-forms>=1.1.1', 'akismet>=0.2', ], description='A modern, ajax-based appearance for django.contrib.comments', long_description=open('README.rst').read(), author='Diederik van der Boor', author_email='opensource@edoburu.nl', url='https://github.com/edoburu/django-fluent-comments', download_url='https://github.com/edoburu/django-fluent-comments/zipball/master', packages=find_packages(exclude=('example*',)), include_package_data=True, zip_safe=False, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ] ) Move Django out of `install_requires` to to `requires` block.#!/usr/bin/env python from setuptools import setup, find_packages from os.path import dirname, join import sys, os # When creating the sdist, make sure the django.mo file also exists: if 'sdist' in sys.argv: try: os.chdir('fluent_comments') from django.core.management.commands.compilemessages import compile_messages compile_messages(sys.stderr) finally: os.chdir('..') setup( name='django-fluent-comments', version='0.8.0', license='Apache License, Version 2.0', install_requires=[ 'django-crispy-forms>=1.1.1', 'akismet>=0.2', ], requires=[ 'Django (>=1.2)', ], description='A modern, ajax-based appearance for django.contrib.comments', long_description=open('README.rst').read(), author='Diederik van der Boor', author_email='opensource@edoburu.nl', url='https://github.com/edoburu/django-fluent-comments', download_url='https://github.com/edoburu/django-fluent-comments/zipball/master', packages=find_packages(exclude=('example*',)), include_package_data=True, zip_safe=False, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ] )
<commit_before>#!/usr/bin/env python from setuptools import setup, find_packages from os.path import dirname, join import sys, os # When creating the sdist, make sure the django.mo file also exists: if 'sdist' in sys.argv: try: os.chdir('fluent_comments') from django.core.management.commands.compilemessages import compile_messages compile_messages(sys.stderr) finally: os.chdir('..') setup( name='django-fluent-comments', version='0.8.0', license='Apache License, Version 2.0', install_requires=[ 'Django>=1.2.0', 'django-crispy-forms>=1.1.1', 'akismet>=0.2', ], description='A modern, ajax-based appearance for django.contrib.comments', long_description=open('README.rst').read(), author='Diederik van der Boor', author_email='opensource@edoburu.nl', url='https://github.com/edoburu/django-fluent-comments', download_url='https://github.com/edoburu/django-fluent-comments/zipball/master', packages=find_packages(exclude=('example*',)), include_package_data=True, zip_safe=False, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ] ) <commit_msg>Move Django out of `install_requires` to to `requires` block.<commit_after>#!/usr/bin/env python from setuptools import setup, find_packages from os.path import dirname, join import sys, os # When creating the sdist, make sure the django.mo file also exists: if 'sdist' in sys.argv: try: os.chdir('fluent_comments') from django.core.management.commands.compilemessages import compile_messages compile_messages(sys.stderr) finally: os.chdir('..') setup( name='django-fluent-comments', version='0.8.0', license='Apache License, Version 2.0', install_requires=[ 'django-crispy-forms>=1.1.1', 'akismet>=0.2', ], requires=[ 'Django (>=1.2)', ], description='A modern, ajax-based appearance for django.contrib.comments', long_description=open('README.rst').read(), author='Diederik van der Boor', author_email='opensource@edoburu.nl', url='https://github.com/edoburu/django-fluent-comments', download_url='https://github.com/edoburu/django-fluent-comments/zipball/master', packages=find_packages(exclude=('example*',)), include_package_data=True, zip_safe=False, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ] )
b588213ccf61a45a1e0e96f626c528483b3aea7e
setup.py
setup.py
""" Flask-Static --------------- Generates a static website from a Flask application. """ from setuptools import setup setup( name='Flask-Static', version='0.1', url='http://exyr.org/Flask-Static/', license='BSD', author='Simon Sapin', author_email='simon.sapin@exyr.org', description='Generates a static website from a Flask application', long_description=__doc__, packages=['flaskext'], namespace_packages=['flaskext'], test_suite='test_flaskstatic', zip_safe=False, platforms='any', install_requires=[ 'Flask', ], classifiers=[ 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules' ] )
""" Flask-Static --------------- Generates a static website from a Flask application. """ from setuptools import setup setup( name='Flask-Static', version='0.1dev', url='http://exyr.org/Flask-Static/', license='BSD', author='Simon Sapin', author_email='simon.sapin@exyr.org', description='Generates a static website from a Flask application', long_description=__doc__, packages=['flaskext'], namespace_packages=['flaskext'], test_suite='test_flaskstatic', zip_safe=False, platforms='any', install_requires=[ 'Flask', ], classifiers=[ 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules' ] )
Change version number to 0.1dev
Change version number to 0.1dev
Python
bsd-3-clause
SimonSapin/Frozen-Flask,SimonSapin/Frozen-Flask,SimonSapin/Frozen-Flask
""" Flask-Static --------------- Generates a static website from a Flask application. """ from setuptools import setup setup( name='Flask-Static', version='0.1', url='http://exyr.org/Flask-Static/', license='BSD', author='Simon Sapin', author_email='simon.sapin@exyr.org', description='Generates a static website from a Flask application', long_description=__doc__, packages=['flaskext'], namespace_packages=['flaskext'], test_suite='test_flaskstatic', zip_safe=False, platforms='any', install_requires=[ 'Flask', ], classifiers=[ 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules' ] ) Change version number to 0.1dev
""" Flask-Static --------------- Generates a static website from a Flask application. """ from setuptools import setup setup( name='Flask-Static', version='0.1dev', url='http://exyr.org/Flask-Static/', license='BSD', author='Simon Sapin', author_email='simon.sapin@exyr.org', description='Generates a static website from a Flask application', long_description=__doc__, packages=['flaskext'], namespace_packages=['flaskext'], test_suite='test_flaskstatic', zip_safe=False, platforms='any', install_requires=[ 'Flask', ], classifiers=[ 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules' ] )
<commit_before>""" Flask-Static --------------- Generates a static website from a Flask application. """ from setuptools import setup setup( name='Flask-Static', version='0.1', url='http://exyr.org/Flask-Static/', license='BSD', author='Simon Sapin', author_email='simon.sapin@exyr.org', description='Generates a static website from a Flask application', long_description=__doc__, packages=['flaskext'], namespace_packages=['flaskext'], test_suite='test_flaskstatic', zip_safe=False, platforms='any', install_requires=[ 'Flask', ], classifiers=[ 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules' ] ) <commit_msg>Change version number to 0.1dev<commit_after>
""" Flask-Static --------------- Generates a static website from a Flask application. """ from setuptools import setup setup( name='Flask-Static', version='0.1dev', url='http://exyr.org/Flask-Static/', license='BSD', author='Simon Sapin', author_email='simon.sapin@exyr.org', description='Generates a static website from a Flask application', long_description=__doc__, packages=['flaskext'], namespace_packages=['flaskext'], test_suite='test_flaskstatic', zip_safe=False, platforms='any', install_requires=[ 'Flask', ], classifiers=[ 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules' ] )
""" Flask-Static --------------- Generates a static website from a Flask application. """ from setuptools import setup setup( name='Flask-Static', version='0.1', url='http://exyr.org/Flask-Static/', license='BSD', author='Simon Sapin', author_email='simon.sapin@exyr.org', description='Generates a static website from a Flask application', long_description=__doc__, packages=['flaskext'], namespace_packages=['flaskext'], test_suite='test_flaskstatic', zip_safe=False, platforms='any', install_requires=[ 'Flask', ], classifiers=[ 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules' ] ) Change version number to 0.1dev""" Flask-Static --------------- Generates a static website from a Flask application. """ from setuptools import setup setup( name='Flask-Static', version='0.1dev', url='http://exyr.org/Flask-Static/', license='BSD', author='Simon Sapin', author_email='simon.sapin@exyr.org', description='Generates a static website from a Flask application', long_description=__doc__, packages=['flaskext'], namespace_packages=['flaskext'], test_suite='test_flaskstatic', zip_safe=False, platforms='any', install_requires=[ 'Flask', ], classifiers=[ 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules' ] )
<commit_before>""" Flask-Static --------------- Generates a static website from a Flask application. """ from setuptools import setup setup( name='Flask-Static', version='0.1', url='http://exyr.org/Flask-Static/', license='BSD', author='Simon Sapin', author_email='simon.sapin@exyr.org', description='Generates a static website from a Flask application', long_description=__doc__, packages=['flaskext'], namespace_packages=['flaskext'], test_suite='test_flaskstatic', zip_safe=False, platforms='any', install_requires=[ 'Flask', ], classifiers=[ 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules' ] ) <commit_msg>Change version number to 0.1dev<commit_after>""" Flask-Static --------------- Generates a static website from a Flask application. """ from setuptools import setup setup( name='Flask-Static', version='0.1dev', url='http://exyr.org/Flask-Static/', license='BSD', author='Simon Sapin', author_email='simon.sapin@exyr.org', description='Generates a static website from a Flask application', long_description=__doc__, packages=['flaskext'], namespace_packages=['flaskext'], test_suite='test_flaskstatic', zip_safe=False, platforms='any', install_requires=[ 'Flask', ], classifiers=[ 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules' ] )
53440a46cb38194e12c383a718f392b9da2175b6
setup.py
setup.py
#!/usr/bin/python import setuptools import numpy from setuptools import Extension setuptools.setup( name='mrpt', version='0.1', url='http://github.com/teemupitkanen/mrpt', install_requires=[], packages={ '.': 'mrpt' }, zip_safe=False, test_suite='py.test', entry_points='', ext_modules = [ Extension('mrptlib', sources = [ 'cpp/mrptmodule.cpp', ], extra_compile_args=['-std=c++11', '-O3', '-march=native', '-ffast-math', '-s', '-fno-rtti', '-fopenmp', '-DNDEBUG'], extra_link_args=['-lgomp'], libraries = ['stdc++'], include_dirs = ['cpp/lib', numpy.get_include()] ) ] )
#!/usr/bin/python import setuptools import numpy from setuptools import Extension # Not all CPUs have march as a tuning parameter import platform cputune = ['-march=native',] if platform.machine() == "ppc64le": cputune = ['-mcpu=native',] setuptools.setup( name='mrpt', version='0.1', url='http://github.com/teemupitkanen/mrpt', install_requires=[], packages={ '.': 'mrpt' }, zip_safe=False, test_suite='py.test', entry_points='', ext_modules = [ Extension('mrptlib', sources = [ 'cpp/mrptmodule.cpp', ], extra_compile_args=['-std=c++11', '-O3', '-ffast-math', '-s', '-fno-rtti', '-fopenmp', '-DNDEBUG'] + cputune, extra_link_args=['-lgomp'], libraries = ['stdc++'], include_dirs = ['cpp/lib', numpy.get_include()] ) ] )
Solve ppc64le build due to unexisting -march
Solve ppc64le build due to unexisting -march On ppc64le (e.g: POWER8 machines), the -mcpu should be used instead.
Python
mit
teemupitkanen/mrpt,teemupitkanen/mrpt,teemupitkanen/mrpt
#!/usr/bin/python import setuptools import numpy from setuptools import Extension setuptools.setup( name='mrpt', version='0.1', url='http://github.com/teemupitkanen/mrpt', install_requires=[], packages={ '.': 'mrpt' }, zip_safe=False, test_suite='py.test', entry_points='', ext_modules = [ Extension('mrptlib', sources = [ 'cpp/mrptmodule.cpp', ], extra_compile_args=['-std=c++11', '-O3', '-march=native', '-ffast-math', '-s', '-fno-rtti', '-fopenmp', '-DNDEBUG'], extra_link_args=['-lgomp'], libraries = ['stdc++'], include_dirs = ['cpp/lib', numpy.get_include()] ) ] ) Solve ppc64le build due to unexisting -march On ppc64le (e.g: POWER8 machines), the -mcpu should be used instead.
#!/usr/bin/python import setuptools import numpy from setuptools import Extension # Not all CPUs have march as a tuning parameter import platform cputune = ['-march=native',] if platform.machine() == "ppc64le": cputune = ['-mcpu=native',] setuptools.setup( name='mrpt', version='0.1', url='http://github.com/teemupitkanen/mrpt', install_requires=[], packages={ '.': 'mrpt' }, zip_safe=False, test_suite='py.test', entry_points='', ext_modules = [ Extension('mrptlib', sources = [ 'cpp/mrptmodule.cpp', ], extra_compile_args=['-std=c++11', '-O3', '-ffast-math', '-s', '-fno-rtti', '-fopenmp', '-DNDEBUG'] + cputune, extra_link_args=['-lgomp'], libraries = ['stdc++'], include_dirs = ['cpp/lib', numpy.get_include()] ) ] )
<commit_before>#!/usr/bin/python import setuptools import numpy from setuptools import Extension setuptools.setup( name='mrpt', version='0.1', url='http://github.com/teemupitkanen/mrpt', install_requires=[], packages={ '.': 'mrpt' }, zip_safe=False, test_suite='py.test', entry_points='', ext_modules = [ Extension('mrptlib', sources = [ 'cpp/mrptmodule.cpp', ], extra_compile_args=['-std=c++11', '-O3', '-march=native', '-ffast-math', '-s', '-fno-rtti', '-fopenmp', '-DNDEBUG'], extra_link_args=['-lgomp'], libraries = ['stdc++'], include_dirs = ['cpp/lib', numpy.get_include()] ) ] ) <commit_msg>Solve ppc64le build due to unexisting -march On ppc64le (e.g: POWER8 machines), the -mcpu should be used instead.<commit_after>
#!/usr/bin/python import setuptools import numpy from setuptools import Extension # Not all CPUs have march as a tuning parameter import platform cputune = ['-march=native',] if platform.machine() == "ppc64le": cputune = ['-mcpu=native',] setuptools.setup( name='mrpt', version='0.1', url='http://github.com/teemupitkanen/mrpt', install_requires=[], packages={ '.': 'mrpt' }, zip_safe=False, test_suite='py.test', entry_points='', ext_modules = [ Extension('mrptlib', sources = [ 'cpp/mrptmodule.cpp', ], extra_compile_args=['-std=c++11', '-O3', '-ffast-math', '-s', '-fno-rtti', '-fopenmp', '-DNDEBUG'] + cputune, extra_link_args=['-lgomp'], libraries = ['stdc++'], include_dirs = ['cpp/lib', numpy.get_include()] ) ] )
#!/usr/bin/python import setuptools import numpy from setuptools import Extension setuptools.setup( name='mrpt', version='0.1', url='http://github.com/teemupitkanen/mrpt', install_requires=[], packages={ '.': 'mrpt' }, zip_safe=False, test_suite='py.test', entry_points='', ext_modules = [ Extension('mrptlib', sources = [ 'cpp/mrptmodule.cpp', ], extra_compile_args=['-std=c++11', '-O3', '-march=native', '-ffast-math', '-s', '-fno-rtti', '-fopenmp', '-DNDEBUG'], extra_link_args=['-lgomp'], libraries = ['stdc++'], include_dirs = ['cpp/lib', numpy.get_include()] ) ] ) Solve ppc64le build due to unexisting -march On ppc64le (e.g: POWER8 machines), the -mcpu should be used instead.#!/usr/bin/python import setuptools import numpy from setuptools import Extension # Not all CPUs have march as a tuning parameter import platform cputune = ['-march=native',] if platform.machine() == "ppc64le": cputune = ['-mcpu=native',] setuptools.setup( name='mrpt', version='0.1', url='http://github.com/teemupitkanen/mrpt', install_requires=[], packages={ '.': 'mrpt' }, zip_safe=False, test_suite='py.test', entry_points='', ext_modules = [ Extension('mrptlib', sources = [ 'cpp/mrptmodule.cpp', ], extra_compile_args=['-std=c++11', '-O3', '-ffast-math', '-s', '-fno-rtti', '-fopenmp', '-DNDEBUG'] + cputune, extra_link_args=['-lgomp'], libraries = ['stdc++'], include_dirs = ['cpp/lib', numpy.get_include()] ) ] )
<commit_before>#!/usr/bin/python import setuptools import numpy from setuptools import Extension setuptools.setup( name='mrpt', version='0.1', url='http://github.com/teemupitkanen/mrpt', install_requires=[], packages={ '.': 'mrpt' }, zip_safe=False, test_suite='py.test', entry_points='', ext_modules = [ Extension('mrptlib', sources = [ 'cpp/mrptmodule.cpp', ], extra_compile_args=['-std=c++11', '-O3', '-march=native', '-ffast-math', '-s', '-fno-rtti', '-fopenmp', '-DNDEBUG'], extra_link_args=['-lgomp'], libraries = ['stdc++'], include_dirs = ['cpp/lib', numpy.get_include()] ) ] ) <commit_msg>Solve ppc64le build due to unexisting -march On ppc64le (e.g: POWER8 machines), the -mcpu should be used instead.<commit_after>#!/usr/bin/python import setuptools import numpy from setuptools import Extension # Not all CPUs have march as a tuning parameter import platform cputune = ['-march=native',] if platform.machine() == "ppc64le": cputune = ['-mcpu=native',] setuptools.setup( name='mrpt', version='0.1', url='http://github.com/teemupitkanen/mrpt', install_requires=[], packages={ '.': 'mrpt' }, zip_safe=False, test_suite='py.test', entry_points='', ext_modules = [ Extension('mrptlib', sources = [ 'cpp/mrptmodule.cpp', ], extra_compile_args=['-std=c++11', '-O3', '-ffast-math', '-s', '-fno-rtti', '-fopenmp', '-DNDEBUG'] + cputune, extra_link_args=['-lgomp'], libraries = ['stdc++'], include_dirs = ['cpp/lib', numpy.get_include()] ) ] )
a714511115bfee0fbdc6c70bd0abfceaa08384f6
idlk/__init__.py
idlk/__init__.py
from __future__ import absolute_import from __future__ import division from __future__ import unicode_literals import os import sys import unicodedata import idlk.base41 if sys.version_info[0] == 3: _get_byte = lambda c: c else: _get_byte = ord def hash_macroman(data): h = 0 for c in data: h = ((h << 8) + h) + _get_byte(c) return h % 0xFFFEECED def idlk(filename): # Normalize to NFC. filename = unicodedata.normalize('NFC', filename) # Convert to lowercase first. filename = filename.lower() # The original algorithm seems to prefer Mac Roman encoding as long as # there are no non-mappable characters in the file name. try: macroman_name = filename.encode("macroman") except UnicodeEncodeError: pass else: hashed = base41.encode(hash_macroman(macroman_name)) base, ext = os.path.splitext(macroman_name) return "~{:s}~{:s}.idlk".format(base[0:18].decode("macroman"), hashed) # Regrettably the encoding / hashing algorithm for unicode filenames is # not currently known. Please file a feature request/patch if you # discover a working implementation. return False
""" A lock filename generator for idlk files used by a well known DTP suite. """ from __future__ import absolute_import from __future__ import division from __future__ import unicode_literals import os import sys import unicodedata from idlk import base41 if sys.version_info[0] == 3: _get_byte = lambda c: c else: _get_byte = ord def hash_macroman(data): """ Compute the hash for the given byte string. """ result = 0 for char in data: result = ((result << 8) + result) + _get_byte(char) return result % 0xFFFEECED def idlk(filename): """ Generate the lock file name for the given file. """ # Normalize to NFC. filename = unicodedata.normalize('NFC', filename) # Convert to lowercase first. filename = filename.lower() # The original algorithm seems to prefer Mac Roman encoding as long as # there are no non-mappable characters in the file name. try: macroman_name = filename.encode("macroman") except UnicodeEncodeError: pass else: hashed = base41.encode(hash_macroman(macroman_name)) base = os.path.splitext(macroman_name)[0] return "~{:s}~{:s}.idlk".format(base[0:18].decode("macroman"), hashed) # Regrettably the encoding / hashing algorithm for unicode filenames is # not currently known. Please file a feature request/patch if you # discover a working implementation. return False
Fix issues reported by pylint
Fix issues reported by pylint
Python
mit
znerol/py-idlk
from __future__ import absolute_import from __future__ import division from __future__ import unicode_literals import os import sys import unicodedata import idlk.base41 if sys.version_info[0] == 3: _get_byte = lambda c: c else: _get_byte = ord def hash_macroman(data): h = 0 for c in data: h = ((h << 8) + h) + _get_byte(c) return h % 0xFFFEECED def idlk(filename): # Normalize to NFC. filename = unicodedata.normalize('NFC', filename) # Convert to lowercase first. filename = filename.lower() # The original algorithm seems to prefer Mac Roman encoding as long as # there are no non-mappable characters in the file name. try: macroman_name = filename.encode("macroman") except UnicodeEncodeError: pass else: hashed = base41.encode(hash_macroman(macroman_name)) base, ext = os.path.splitext(macroman_name) return "~{:s}~{:s}.idlk".format(base[0:18].decode("macroman"), hashed) # Regrettably the encoding / hashing algorithm for unicode filenames is # not currently known. Please file a feature request/patch if you # discover a working implementation. return False Fix issues reported by pylint
""" A lock filename generator for idlk files used by a well known DTP suite. """ from __future__ import absolute_import from __future__ import division from __future__ import unicode_literals import os import sys import unicodedata from idlk import base41 if sys.version_info[0] == 3: _get_byte = lambda c: c else: _get_byte = ord def hash_macroman(data): """ Compute the hash for the given byte string. """ result = 0 for char in data: result = ((result << 8) + result) + _get_byte(char) return result % 0xFFFEECED def idlk(filename): """ Generate the lock file name for the given file. """ # Normalize to NFC. filename = unicodedata.normalize('NFC', filename) # Convert to lowercase first. filename = filename.lower() # The original algorithm seems to prefer Mac Roman encoding as long as # there are no non-mappable characters in the file name. try: macroman_name = filename.encode("macroman") except UnicodeEncodeError: pass else: hashed = base41.encode(hash_macroman(macroman_name)) base = os.path.splitext(macroman_name)[0] return "~{:s}~{:s}.idlk".format(base[0:18].decode("macroman"), hashed) # Regrettably the encoding / hashing algorithm for unicode filenames is # not currently known. Please file a feature request/patch if you # discover a working implementation. return False
<commit_before>from __future__ import absolute_import from __future__ import division from __future__ import unicode_literals import os import sys import unicodedata import idlk.base41 if sys.version_info[0] == 3: _get_byte = lambda c: c else: _get_byte = ord def hash_macroman(data): h = 0 for c in data: h = ((h << 8) + h) + _get_byte(c) return h % 0xFFFEECED def idlk(filename): # Normalize to NFC. filename = unicodedata.normalize('NFC', filename) # Convert to lowercase first. filename = filename.lower() # The original algorithm seems to prefer Mac Roman encoding as long as # there are no non-mappable characters in the file name. try: macroman_name = filename.encode("macroman") except UnicodeEncodeError: pass else: hashed = base41.encode(hash_macroman(macroman_name)) base, ext = os.path.splitext(macroman_name) return "~{:s}~{:s}.idlk".format(base[0:18].decode("macroman"), hashed) # Regrettably the encoding / hashing algorithm for unicode filenames is # not currently known. Please file a feature request/patch if you # discover a working implementation. return False <commit_msg>Fix issues reported by pylint<commit_after>
""" A lock filename generator for idlk files used by a well known DTP suite. """ from __future__ import absolute_import from __future__ import division from __future__ import unicode_literals import os import sys import unicodedata from idlk import base41 if sys.version_info[0] == 3: _get_byte = lambda c: c else: _get_byte = ord def hash_macroman(data): """ Compute the hash for the given byte string. """ result = 0 for char in data: result = ((result << 8) + result) + _get_byte(char) return result % 0xFFFEECED def idlk(filename): """ Generate the lock file name for the given file. """ # Normalize to NFC. filename = unicodedata.normalize('NFC', filename) # Convert to lowercase first. filename = filename.lower() # The original algorithm seems to prefer Mac Roman encoding as long as # there are no non-mappable characters in the file name. try: macroman_name = filename.encode("macroman") except UnicodeEncodeError: pass else: hashed = base41.encode(hash_macroman(macroman_name)) base = os.path.splitext(macroman_name)[0] return "~{:s}~{:s}.idlk".format(base[0:18].decode("macroman"), hashed) # Regrettably the encoding / hashing algorithm for unicode filenames is # not currently known. Please file a feature request/patch if you # discover a working implementation. return False
from __future__ import absolute_import from __future__ import division from __future__ import unicode_literals import os import sys import unicodedata import idlk.base41 if sys.version_info[0] == 3: _get_byte = lambda c: c else: _get_byte = ord def hash_macroman(data): h = 0 for c in data: h = ((h << 8) + h) + _get_byte(c) return h % 0xFFFEECED def idlk(filename): # Normalize to NFC. filename = unicodedata.normalize('NFC', filename) # Convert to lowercase first. filename = filename.lower() # The original algorithm seems to prefer Mac Roman encoding as long as # there are no non-mappable characters in the file name. try: macroman_name = filename.encode("macroman") except UnicodeEncodeError: pass else: hashed = base41.encode(hash_macroman(macroman_name)) base, ext = os.path.splitext(macroman_name) return "~{:s}~{:s}.idlk".format(base[0:18].decode("macroman"), hashed) # Regrettably the encoding / hashing algorithm for unicode filenames is # not currently known. Please file a feature request/patch if you # discover a working implementation. return False Fix issues reported by pylint""" A lock filename generator for idlk files used by a well known DTP suite. """ from __future__ import absolute_import from __future__ import division from __future__ import unicode_literals import os import sys import unicodedata from idlk import base41 if sys.version_info[0] == 3: _get_byte = lambda c: c else: _get_byte = ord def hash_macroman(data): """ Compute the hash for the given byte string. """ result = 0 for char in data: result = ((result << 8) + result) + _get_byte(char) return result % 0xFFFEECED def idlk(filename): """ Generate the lock file name for the given file. """ # Normalize to NFC. filename = unicodedata.normalize('NFC', filename) # Convert to lowercase first. filename = filename.lower() # The original algorithm seems to prefer Mac Roman encoding as long as # there are no non-mappable characters in the file name. try: macroman_name = filename.encode("macroman") except UnicodeEncodeError: pass else: hashed = base41.encode(hash_macroman(macroman_name)) base = os.path.splitext(macroman_name)[0] return "~{:s}~{:s}.idlk".format(base[0:18].decode("macroman"), hashed) # Regrettably the encoding / hashing algorithm for unicode filenames is # not currently known. Please file a feature request/patch if you # discover a working implementation. return False
<commit_before>from __future__ import absolute_import from __future__ import division from __future__ import unicode_literals import os import sys import unicodedata import idlk.base41 if sys.version_info[0] == 3: _get_byte = lambda c: c else: _get_byte = ord def hash_macroman(data): h = 0 for c in data: h = ((h << 8) + h) + _get_byte(c) return h % 0xFFFEECED def idlk(filename): # Normalize to NFC. filename = unicodedata.normalize('NFC', filename) # Convert to lowercase first. filename = filename.lower() # The original algorithm seems to prefer Mac Roman encoding as long as # there are no non-mappable characters in the file name. try: macroman_name = filename.encode("macroman") except UnicodeEncodeError: pass else: hashed = base41.encode(hash_macroman(macroman_name)) base, ext = os.path.splitext(macroman_name) return "~{:s}~{:s}.idlk".format(base[0:18].decode("macroman"), hashed) # Regrettably the encoding / hashing algorithm for unicode filenames is # not currently known. Please file a feature request/patch if you # discover a working implementation. return False <commit_msg>Fix issues reported by pylint<commit_after>""" A lock filename generator for idlk files used by a well known DTP suite. """ from __future__ import absolute_import from __future__ import division from __future__ import unicode_literals import os import sys import unicodedata from idlk import base41 if sys.version_info[0] == 3: _get_byte = lambda c: c else: _get_byte = ord def hash_macroman(data): """ Compute the hash for the given byte string. """ result = 0 for char in data: result = ((result << 8) + result) + _get_byte(char) return result % 0xFFFEECED def idlk(filename): """ Generate the lock file name for the given file. """ # Normalize to NFC. filename = unicodedata.normalize('NFC', filename) # Convert to lowercase first. filename = filename.lower() # The original algorithm seems to prefer Mac Roman encoding as long as # there are no non-mappable characters in the file name. try: macroman_name = filename.encode("macroman") except UnicodeEncodeError: pass else: hashed = base41.encode(hash_macroman(macroman_name)) base = os.path.splitext(macroman_name)[0] return "~{:s}~{:s}.idlk".format(base[0:18].decode("macroman"), hashed) # Regrettably the encoding / hashing algorithm for unicode filenames is # not currently known. Please file a feature request/patch if you # discover a working implementation. return False
3bc0876e7bae2cfb62724f1e5dce1a93f71b7252
docstring_parser/parser/__init__.py
docstring_parser/parser/__init__.py
"""Docstring parsing.""" from . import rest from .common import ParseError _styles = {"rest": rest.parse} def parse(text: str, style: str = "auto"): """ Parse the docstring into its components. :param str text: docstring text to parse :param text style: docstring style, choose from: 'rest', 'auto' :returns: parsed docstring """ if style == "auto": rets = [] for _parse in _styles.values(): try: rets.append(_parse(text)) except ParseError as e: exc = e if not rets: raise exc return sorted(rets, key=lambda ret: len(ret.meta), reverse=True)[0] else: return _styles[style]
"""Docstring parsing.""" from . import rest from .common import ParseError, Docstring _styles = {"rest": rest.parse} def _parse_score(docstring: Docstring) -> int: """ Produce a score for the parsing. :param Docstring docstring: parsed docstring representation :returns int: parse score, higher is better """ score = 0 if docstring.short_description: score += 1 if docstring.long_description: score += docstring.long_description.count('\n') score += len(docstring.params) score += len(docstring.raises) if docstring.returns: score += 2 return score def parse(text: str, style: str = 'auto') -> Docstring: """ Parse the docstring into its components. :param str text: docstring text to parse :param str style: docstring style, choose from: 'rest', 'auto' :returns Docstring: parsed docstring representation """ if style == 'auto': rets = [] for _parse in _styles.values(): try: rets.append(_parse(text)) except ParseError as e: exc = e if not rets: raise exc return sorted(rets, key=_parse_score, reverse=True)[0] else: return _styles[style](text)
Fix parsing when style specified, add 'auto' score
Fix parsing when style specified, add 'auto' score
Python
mit
rr-/docstring_parser
"""Docstring parsing.""" from . import rest from .common import ParseError _styles = {"rest": rest.parse} def parse(text: str, style: str = "auto"): """ Parse the docstring into its components. :param str text: docstring text to parse :param text style: docstring style, choose from: 'rest', 'auto' :returns: parsed docstring """ if style == "auto": rets = [] for _parse in _styles.values(): try: rets.append(_parse(text)) except ParseError as e: exc = e if not rets: raise exc return sorted(rets, key=lambda ret: len(ret.meta), reverse=True)[0] else: return _styles[style] Fix parsing when style specified, add 'auto' score
"""Docstring parsing.""" from . import rest from .common import ParseError, Docstring _styles = {"rest": rest.parse} def _parse_score(docstring: Docstring) -> int: """ Produce a score for the parsing. :param Docstring docstring: parsed docstring representation :returns int: parse score, higher is better """ score = 0 if docstring.short_description: score += 1 if docstring.long_description: score += docstring.long_description.count('\n') score += len(docstring.params) score += len(docstring.raises) if docstring.returns: score += 2 return score def parse(text: str, style: str = 'auto') -> Docstring: """ Parse the docstring into its components. :param str text: docstring text to parse :param str style: docstring style, choose from: 'rest', 'auto' :returns Docstring: parsed docstring representation """ if style == 'auto': rets = [] for _parse in _styles.values(): try: rets.append(_parse(text)) except ParseError as e: exc = e if not rets: raise exc return sorted(rets, key=_parse_score, reverse=True)[0] else: return _styles[style](text)
<commit_before>"""Docstring parsing.""" from . import rest from .common import ParseError _styles = {"rest": rest.parse} def parse(text: str, style: str = "auto"): """ Parse the docstring into its components. :param str text: docstring text to parse :param text style: docstring style, choose from: 'rest', 'auto' :returns: parsed docstring """ if style == "auto": rets = [] for _parse in _styles.values(): try: rets.append(_parse(text)) except ParseError as e: exc = e if not rets: raise exc return sorted(rets, key=lambda ret: len(ret.meta), reverse=True)[0] else: return _styles[style] <commit_msg>Fix parsing when style specified, add 'auto' score<commit_after>
"""Docstring parsing.""" from . import rest from .common import ParseError, Docstring _styles = {"rest": rest.parse} def _parse_score(docstring: Docstring) -> int: """ Produce a score for the parsing. :param Docstring docstring: parsed docstring representation :returns int: parse score, higher is better """ score = 0 if docstring.short_description: score += 1 if docstring.long_description: score += docstring.long_description.count('\n') score += len(docstring.params) score += len(docstring.raises) if docstring.returns: score += 2 return score def parse(text: str, style: str = 'auto') -> Docstring: """ Parse the docstring into its components. :param str text: docstring text to parse :param str style: docstring style, choose from: 'rest', 'auto' :returns Docstring: parsed docstring representation """ if style == 'auto': rets = [] for _parse in _styles.values(): try: rets.append(_parse(text)) except ParseError as e: exc = e if not rets: raise exc return sorted(rets, key=_parse_score, reverse=True)[0] else: return _styles[style](text)
"""Docstring parsing.""" from . import rest from .common import ParseError _styles = {"rest": rest.parse} def parse(text: str, style: str = "auto"): """ Parse the docstring into its components. :param str text: docstring text to parse :param text style: docstring style, choose from: 'rest', 'auto' :returns: parsed docstring """ if style == "auto": rets = [] for _parse in _styles.values(): try: rets.append(_parse(text)) except ParseError as e: exc = e if not rets: raise exc return sorted(rets, key=lambda ret: len(ret.meta), reverse=True)[0] else: return _styles[style] Fix parsing when style specified, add 'auto' score"""Docstring parsing.""" from . import rest from .common import ParseError, Docstring _styles = {"rest": rest.parse} def _parse_score(docstring: Docstring) -> int: """ Produce a score for the parsing. :param Docstring docstring: parsed docstring representation :returns int: parse score, higher is better """ score = 0 if docstring.short_description: score += 1 if docstring.long_description: score += docstring.long_description.count('\n') score += len(docstring.params) score += len(docstring.raises) if docstring.returns: score += 2 return score def parse(text: str, style: str = 'auto') -> Docstring: """ Parse the docstring into its components. :param str text: docstring text to parse :param str style: docstring style, choose from: 'rest', 'auto' :returns Docstring: parsed docstring representation """ if style == 'auto': rets = [] for _parse in _styles.values(): try: rets.append(_parse(text)) except ParseError as e: exc = e if not rets: raise exc return sorted(rets, key=_parse_score, reverse=True)[0] else: return _styles[style](text)
<commit_before>"""Docstring parsing.""" from . import rest from .common import ParseError _styles = {"rest": rest.parse} def parse(text: str, style: str = "auto"): """ Parse the docstring into its components. :param str text: docstring text to parse :param text style: docstring style, choose from: 'rest', 'auto' :returns: parsed docstring """ if style == "auto": rets = [] for _parse in _styles.values(): try: rets.append(_parse(text)) except ParseError as e: exc = e if not rets: raise exc return sorted(rets, key=lambda ret: len(ret.meta), reverse=True)[0] else: return _styles[style] <commit_msg>Fix parsing when style specified, add 'auto' score<commit_after>"""Docstring parsing.""" from . import rest from .common import ParseError, Docstring _styles = {"rest": rest.parse} def _parse_score(docstring: Docstring) -> int: """ Produce a score for the parsing. :param Docstring docstring: parsed docstring representation :returns int: parse score, higher is better """ score = 0 if docstring.short_description: score += 1 if docstring.long_description: score += docstring.long_description.count('\n') score += len(docstring.params) score += len(docstring.raises) if docstring.returns: score += 2 return score def parse(text: str, style: str = 'auto') -> Docstring: """ Parse the docstring into its components. :param str text: docstring text to parse :param str style: docstring style, choose from: 'rest', 'auto' :returns Docstring: parsed docstring representation """ if style == 'auto': rets = [] for _parse in _styles.values(): try: rets.append(_parse(text)) except ParseError as e: exc = e if not rets: raise exc return sorted(rets, key=_parse_score, reverse=True)[0] else: return _styles[style](text)
0e1bdcb4e6d2404bb832ab86ec7bf526c1c90bbb
teami18n/teami18n/models.py
teami18n/teami18n/models.py
from django.db import models from django_countries import countries class Country(models.Model): code = models.CharField(max_length=2, choices=tuple(countries), unique=True) class Podcast(models.Model): story_id = models.CharField(max_length=16, unique=True) link = models.URLField() title = models.TextField() teaser = models.TextField() program_name = models.TextField(blank=True) show_date = models.DateTimeField(null=True, blank=True) image_link = models.URLField(null=True, blank=True) countries = models.ManyToManyField(Country, related_name="podcasts")
from django.db import models from django_countries import countries class Country(models.Model): code = models.CharField(max_length=2, choices=tuple(countries), unique=True) def __unicode__(self): return self.code class Podcast(models.Model): story_id = models.CharField(max_length=16, unique=True) link = models.URLField() title = models.TextField() teaser = models.TextField() program_name = models.TextField(blank=True) show_date = models.DateTimeField(null=True, blank=True) image_link = models.URLField(null=True, blank=True) countries = models.ManyToManyField(Country, related_name="podcasts") def __unicode__(self): return self.title
Add nice name for working in the shell
Add nice name for working in the shell
Python
mit
team-i18n/hackaway,team-i18n/hackaway,team-i18n/hackaway
from django.db import models from django_countries import countries class Country(models.Model): code = models.CharField(max_length=2, choices=tuple(countries), unique=True) class Podcast(models.Model): story_id = models.CharField(max_length=16, unique=True) link = models.URLField() title = models.TextField() teaser = models.TextField() program_name = models.TextField(blank=True) show_date = models.DateTimeField(null=True, blank=True) image_link = models.URLField(null=True, blank=True) countries = models.ManyToManyField(Country, related_name="podcasts") Add nice name for working in the shell
from django.db import models from django_countries import countries class Country(models.Model): code = models.CharField(max_length=2, choices=tuple(countries), unique=True) def __unicode__(self): return self.code class Podcast(models.Model): story_id = models.CharField(max_length=16, unique=True) link = models.URLField() title = models.TextField() teaser = models.TextField() program_name = models.TextField(blank=True) show_date = models.DateTimeField(null=True, blank=True) image_link = models.URLField(null=True, blank=True) countries = models.ManyToManyField(Country, related_name="podcasts") def __unicode__(self): return self.title
<commit_before>from django.db import models from django_countries import countries class Country(models.Model): code = models.CharField(max_length=2, choices=tuple(countries), unique=True) class Podcast(models.Model): story_id = models.CharField(max_length=16, unique=True) link = models.URLField() title = models.TextField() teaser = models.TextField() program_name = models.TextField(blank=True) show_date = models.DateTimeField(null=True, blank=True) image_link = models.URLField(null=True, blank=True) countries = models.ManyToManyField(Country, related_name="podcasts") <commit_msg>Add nice name for working in the shell<commit_after>
from django.db import models from django_countries import countries class Country(models.Model): code = models.CharField(max_length=2, choices=tuple(countries), unique=True) def __unicode__(self): return self.code class Podcast(models.Model): story_id = models.CharField(max_length=16, unique=True) link = models.URLField() title = models.TextField() teaser = models.TextField() program_name = models.TextField(blank=True) show_date = models.DateTimeField(null=True, blank=True) image_link = models.URLField(null=True, blank=True) countries = models.ManyToManyField(Country, related_name="podcasts") def __unicode__(self): return self.title
from django.db import models from django_countries import countries class Country(models.Model): code = models.CharField(max_length=2, choices=tuple(countries), unique=True) class Podcast(models.Model): story_id = models.CharField(max_length=16, unique=True) link = models.URLField() title = models.TextField() teaser = models.TextField() program_name = models.TextField(blank=True) show_date = models.DateTimeField(null=True, blank=True) image_link = models.URLField(null=True, blank=True) countries = models.ManyToManyField(Country, related_name="podcasts") Add nice name for working in the shellfrom django.db import models from django_countries import countries class Country(models.Model): code = models.CharField(max_length=2, choices=tuple(countries), unique=True) def __unicode__(self): return self.code class Podcast(models.Model): story_id = models.CharField(max_length=16, unique=True) link = models.URLField() title = models.TextField() teaser = models.TextField() program_name = models.TextField(blank=True) show_date = models.DateTimeField(null=True, blank=True) image_link = models.URLField(null=True, blank=True) countries = models.ManyToManyField(Country, related_name="podcasts") def __unicode__(self): return self.title
<commit_before>from django.db import models from django_countries import countries class Country(models.Model): code = models.CharField(max_length=2, choices=tuple(countries), unique=True) class Podcast(models.Model): story_id = models.CharField(max_length=16, unique=True) link = models.URLField() title = models.TextField() teaser = models.TextField() program_name = models.TextField(blank=True) show_date = models.DateTimeField(null=True, blank=True) image_link = models.URLField(null=True, blank=True) countries = models.ManyToManyField(Country, related_name="podcasts") <commit_msg>Add nice name for working in the shell<commit_after>from django.db import models from django_countries import countries class Country(models.Model): code = models.CharField(max_length=2, choices=tuple(countries), unique=True) def __unicode__(self): return self.code class Podcast(models.Model): story_id = models.CharField(max_length=16, unique=True) link = models.URLField() title = models.TextField() teaser = models.TextField() program_name = models.TextField(blank=True) show_date = models.DateTimeField(null=True, blank=True) image_link = models.URLField(null=True, blank=True) countries = models.ManyToManyField(Country, related_name="podcasts") def __unicode__(self): return self.title
e1a0e3e6895ce14822b111ee17b182a79b7b28c9
miniraf/calc.py
miniraf/calc.py
def create_parser(subparsers): pass
import argparse from astropy.io import fits import sys OP_MAP = {"+": lambda x, y: x + y, "-": lambda x, y: x - y, "*": lambda x, y: x * y, "/": lambda x, y: x / y} def create_parser(subparsers): parser_calc = subparsers.add_parser("calc", help="calc help") parser_calc.add_argument("-o", "--output", metavar="OUTFILE", default=sys.stdout.buffer) parser_calc.add_argument("file1") parser_calc.add_argument("op", choices=["+", "-", "*", "/"]) parser_calc.add_argument("file2") parser_calc.set_defaults(func=main) def load_fits_data(filename): with fits.open(filename) as f: data = f[0].data return data def main(args): a, b = load_fits_data(args.file1), load_fits_data(args.file2) result = OP_MAP[args.op](a, b) hdu = fits.PrimaryHDU(result) hdu.writeto(args.output)
Add simple four-function output option
Add simple four-function output option Signed-off-by: Lizhou Sha <d6acb26e253550574bc1141efa0eb5e6de15daeb@mit.edu>
Python
mit
vulpicastor/miniraf
def create_parser(subparsers): pass Add simple four-function output option Signed-off-by: Lizhou Sha <d6acb26e253550574bc1141efa0eb5e6de15daeb@mit.edu>
import argparse from astropy.io import fits import sys OP_MAP = {"+": lambda x, y: x + y, "-": lambda x, y: x - y, "*": lambda x, y: x * y, "/": lambda x, y: x / y} def create_parser(subparsers): parser_calc = subparsers.add_parser("calc", help="calc help") parser_calc.add_argument("-o", "--output", metavar="OUTFILE", default=sys.stdout.buffer) parser_calc.add_argument("file1") parser_calc.add_argument("op", choices=["+", "-", "*", "/"]) parser_calc.add_argument("file2") parser_calc.set_defaults(func=main) def load_fits_data(filename): with fits.open(filename) as f: data = f[0].data return data def main(args): a, b = load_fits_data(args.file1), load_fits_data(args.file2) result = OP_MAP[args.op](a, b) hdu = fits.PrimaryHDU(result) hdu.writeto(args.output)
<commit_before>def create_parser(subparsers): pass <commit_msg>Add simple four-function output option Signed-off-by: Lizhou Sha <d6acb26e253550574bc1141efa0eb5e6de15daeb@mit.edu><commit_after>
import argparse from astropy.io import fits import sys OP_MAP = {"+": lambda x, y: x + y, "-": lambda x, y: x - y, "*": lambda x, y: x * y, "/": lambda x, y: x / y} def create_parser(subparsers): parser_calc = subparsers.add_parser("calc", help="calc help") parser_calc.add_argument("-o", "--output", metavar="OUTFILE", default=sys.stdout.buffer) parser_calc.add_argument("file1") parser_calc.add_argument("op", choices=["+", "-", "*", "/"]) parser_calc.add_argument("file2") parser_calc.set_defaults(func=main) def load_fits_data(filename): with fits.open(filename) as f: data = f[0].data return data def main(args): a, b = load_fits_data(args.file1), load_fits_data(args.file2) result = OP_MAP[args.op](a, b) hdu = fits.PrimaryHDU(result) hdu.writeto(args.output)
def create_parser(subparsers): pass Add simple four-function output option Signed-off-by: Lizhou Sha <d6acb26e253550574bc1141efa0eb5e6de15daeb@mit.edu>import argparse from astropy.io import fits import sys OP_MAP = {"+": lambda x, y: x + y, "-": lambda x, y: x - y, "*": lambda x, y: x * y, "/": lambda x, y: x / y} def create_parser(subparsers): parser_calc = subparsers.add_parser("calc", help="calc help") parser_calc.add_argument("-o", "--output", metavar="OUTFILE", default=sys.stdout.buffer) parser_calc.add_argument("file1") parser_calc.add_argument("op", choices=["+", "-", "*", "/"]) parser_calc.add_argument("file2") parser_calc.set_defaults(func=main) def load_fits_data(filename): with fits.open(filename) as f: data = f[0].data return data def main(args): a, b = load_fits_data(args.file1), load_fits_data(args.file2) result = OP_MAP[args.op](a, b) hdu = fits.PrimaryHDU(result) hdu.writeto(args.output)
<commit_before>def create_parser(subparsers): pass <commit_msg>Add simple four-function output option Signed-off-by: Lizhou Sha <d6acb26e253550574bc1141efa0eb5e6de15daeb@mit.edu><commit_after>import argparse from astropy.io import fits import sys OP_MAP = {"+": lambda x, y: x + y, "-": lambda x, y: x - y, "*": lambda x, y: x * y, "/": lambda x, y: x / y} def create_parser(subparsers): parser_calc = subparsers.add_parser("calc", help="calc help") parser_calc.add_argument("-o", "--output", metavar="OUTFILE", default=sys.stdout.buffer) parser_calc.add_argument("file1") parser_calc.add_argument("op", choices=["+", "-", "*", "/"]) parser_calc.add_argument("file2") parser_calc.set_defaults(func=main) def load_fits_data(filename): with fits.open(filename) as f: data = f[0].data return data def main(args): a, b = load_fits_data(args.file1), load_fits_data(args.file2) result = OP_MAP[args.op](a, b) hdu = fits.PrimaryHDU(result) hdu.writeto(args.output)
8ca16832b54c887e6e3a84d7018181bf7e55fba0
comrade/core/context_processors.py
comrade/core/context_processors.py
from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def default(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT return context
from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def default(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT return context def ssl_media(request): if request.is_secure(): ssl_media_url = settings.MEDIA_URL.replace('http://','https://') else: ssl_media_url = settings.MEDIA_URL return {'MEDIA_URL': ssl_media_url}
Add SSL media context processor.
Add SSL media context processor.
Python
mit
bueda/django-comrade
from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def default(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT return context Add SSL media context processor.
from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def default(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT return context def ssl_media(request): if request.is_secure(): ssl_media_url = settings.MEDIA_URL.replace('http://','https://') else: ssl_media_url = settings.MEDIA_URL return {'MEDIA_URL': ssl_media_url}
<commit_before>from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def default(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT return context <commit_msg>Add SSL media context processor.<commit_after>
from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def default(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT return context def ssl_media(request): if request.is_secure(): ssl_media_url = settings.MEDIA_URL.replace('http://','https://') else: ssl_media_url = settings.MEDIA_URL return {'MEDIA_URL': ssl_media_url}
from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def default(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT return context Add SSL media context processor.from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def default(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT return context def ssl_media(request): if request.is_secure(): ssl_media_url = settings.MEDIA_URL.replace('http://','https://') else: ssl_media_url = settings.MEDIA_URL return {'MEDIA_URL': ssl_media_url}
<commit_before>from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def default(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT return context <commit_msg>Add SSL media context processor.<commit_after>from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def default(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT return context def ssl_media(request): if request.is_secure(): ssl_media_url = settings.MEDIA_URL.replace('http://','https://') else: ssl_media_url = settings.MEDIA_URL return {'MEDIA_URL': ssl_media_url}
7a85762ead43d8ba75547488eecda120417e8c2a
lib/python/opendiamond/helpers.py
lib/python/opendiamond/helpers.py
# # The OpenDiamond Platform for Interactive Search # Version 4 # # Copyright (c) 2009 Carnegie Mellon University # All rights reserved. # # This software is distributed under the terms of the Eclipse Public # License, Version 1.0 which can be found in the file named LICENSE. # ANY USE, REPRODUCTION OR DISTRIBUTION OF THIS SOFTWARE CONSTITUTES # RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT # from subprocess import Popen, PIPE def _GenerateCookie(scopelist, servers): cmd = ["cookiecutter"] for server in servers: cmd.extend(['-s', server]) for url in scopelist: cmd.extend(['-u', url]) return Popen(cmd, stdout=PIPE).stdout.read() def GenerateCookie(scopelist, servers, proxies=None): if not proxies: return _GenerateCookie(scopelist, servers) cookie = [] n = len(proxies) for i in range(n): scope = [ '/proxy/%dof%d/%s:5873%s' % (i+1, n, server, scope) for scope in scopelist for server in servers ] cookie.append(_GenerateCookie(scope, (proxies[i],))) return ''.join(cookie)
# # The OpenDiamond Platform for Interactive Search # Version 4 # # Copyright (c) 2009 Carnegie Mellon University # All rights reserved. # # This software is distributed under the terms of the Eclipse Public # License, Version 1.0 which can be found in the file named LICENSE. # ANY USE, REPRODUCTION OR DISTRIBUTION OF THIS SOFTWARE CONSTITUTES # RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT #
Remove Python wrapper function for executing cookiecutter program
Remove Python wrapper function for executing cookiecutter program
Python
epl-1.0
cmusatyalab/opendiamond,cmusatyalab/opendiamond,cmusatyalab/opendiamond,cmusatyalab/opendiamond,cmusatyalab/opendiamond
# # The OpenDiamond Platform for Interactive Search # Version 4 # # Copyright (c) 2009 Carnegie Mellon University # All rights reserved. # # This software is distributed under the terms of the Eclipse Public # License, Version 1.0 which can be found in the file named LICENSE. # ANY USE, REPRODUCTION OR DISTRIBUTION OF THIS SOFTWARE CONSTITUTES # RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT # from subprocess import Popen, PIPE def _GenerateCookie(scopelist, servers): cmd = ["cookiecutter"] for server in servers: cmd.extend(['-s', server]) for url in scopelist: cmd.extend(['-u', url]) return Popen(cmd, stdout=PIPE).stdout.read() def GenerateCookie(scopelist, servers, proxies=None): if not proxies: return _GenerateCookie(scopelist, servers) cookie = [] n = len(proxies) for i in range(n): scope = [ '/proxy/%dof%d/%s:5873%s' % (i+1, n, server, scope) for scope in scopelist for server in servers ] cookie.append(_GenerateCookie(scope, (proxies[i],))) return ''.join(cookie) Remove Python wrapper function for executing cookiecutter program
# # The OpenDiamond Platform for Interactive Search # Version 4 # # Copyright (c) 2009 Carnegie Mellon University # All rights reserved. # # This software is distributed under the terms of the Eclipse Public # License, Version 1.0 which can be found in the file named LICENSE. # ANY USE, REPRODUCTION OR DISTRIBUTION OF THIS SOFTWARE CONSTITUTES # RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT #
<commit_before># # The OpenDiamond Platform for Interactive Search # Version 4 # # Copyright (c) 2009 Carnegie Mellon University # All rights reserved. # # This software is distributed under the terms of the Eclipse Public # License, Version 1.0 which can be found in the file named LICENSE. # ANY USE, REPRODUCTION OR DISTRIBUTION OF THIS SOFTWARE CONSTITUTES # RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT # from subprocess import Popen, PIPE def _GenerateCookie(scopelist, servers): cmd = ["cookiecutter"] for server in servers: cmd.extend(['-s', server]) for url in scopelist: cmd.extend(['-u', url]) return Popen(cmd, stdout=PIPE).stdout.read() def GenerateCookie(scopelist, servers, proxies=None): if not proxies: return _GenerateCookie(scopelist, servers) cookie = [] n = len(proxies) for i in range(n): scope = [ '/proxy/%dof%d/%s:5873%s' % (i+1, n, server, scope) for scope in scopelist for server in servers ] cookie.append(_GenerateCookie(scope, (proxies[i],))) return ''.join(cookie) <commit_msg>Remove Python wrapper function for executing cookiecutter program<commit_after>
# # The OpenDiamond Platform for Interactive Search # Version 4 # # Copyright (c) 2009 Carnegie Mellon University # All rights reserved. # # This software is distributed under the terms of the Eclipse Public # License, Version 1.0 which can be found in the file named LICENSE. # ANY USE, REPRODUCTION OR DISTRIBUTION OF THIS SOFTWARE CONSTITUTES # RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT #
# # The OpenDiamond Platform for Interactive Search # Version 4 # # Copyright (c) 2009 Carnegie Mellon University # All rights reserved. # # This software is distributed under the terms of the Eclipse Public # License, Version 1.0 which can be found in the file named LICENSE. # ANY USE, REPRODUCTION OR DISTRIBUTION OF THIS SOFTWARE CONSTITUTES # RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT # from subprocess import Popen, PIPE def _GenerateCookie(scopelist, servers): cmd = ["cookiecutter"] for server in servers: cmd.extend(['-s', server]) for url in scopelist: cmd.extend(['-u', url]) return Popen(cmd, stdout=PIPE).stdout.read() def GenerateCookie(scopelist, servers, proxies=None): if not proxies: return _GenerateCookie(scopelist, servers) cookie = [] n = len(proxies) for i in range(n): scope = [ '/proxy/%dof%d/%s:5873%s' % (i+1, n, server, scope) for scope in scopelist for server in servers ] cookie.append(_GenerateCookie(scope, (proxies[i],))) return ''.join(cookie) Remove Python wrapper function for executing cookiecutter program# # The OpenDiamond Platform for Interactive Search # Version 4 # # Copyright (c) 2009 Carnegie Mellon University # All rights reserved. # # This software is distributed under the terms of the Eclipse Public # License, Version 1.0 which can be found in the file named LICENSE. # ANY USE, REPRODUCTION OR DISTRIBUTION OF THIS SOFTWARE CONSTITUTES # RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT #
<commit_before># # The OpenDiamond Platform for Interactive Search # Version 4 # # Copyright (c) 2009 Carnegie Mellon University # All rights reserved. # # This software is distributed under the terms of the Eclipse Public # License, Version 1.0 which can be found in the file named LICENSE. # ANY USE, REPRODUCTION OR DISTRIBUTION OF THIS SOFTWARE CONSTITUTES # RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT # from subprocess import Popen, PIPE def _GenerateCookie(scopelist, servers): cmd = ["cookiecutter"] for server in servers: cmd.extend(['-s', server]) for url in scopelist: cmd.extend(['-u', url]) return Popen(cmd, stdout=PIPE).stdout.read() def GenerateCookie(scopelist, servers, proxies=None): if not proxies: return _GenerateCookie(scopelist, servers) cookie = [] n = len(proxies) for i in range(n): scope = [ '/proxy/%dof%d/%s:5873%s' % (i+1, n, server, scope) for scope in scopelist for server in servers ] cookie.append(_GenerateCookie(scope, (proxies[i],))) return ''.join(cookie) <commit_msg>Remove Python wrapper function for executing cookiecutter program<commit_after># # The OpenDiamond Platform for Interactive Search # Version 4 # # Copyright (c) 2009 Carnegie Mellon University # All rights reserved. # # This software is distributed under the terms of the Eclipse Public # License, Version 1.0 which can be found in the file named LICENSE. # ANY USE, REPRODUCTION OR DISTRIBUTION OF THIS SOFTWARE CONSTITUTES # RECIPIENT'S ACCEPTANCE OF THIS AGREEMENT #
41b241de6f2afa94b442007518d481526bfb66ae
linked-list/remove-k-from-list.py
linked-list/remove-k-from-list.py
# Given a singly linked list of integers l and an integer k, remove all elements from list l that have a value equal to k class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None
# Given a singly linked list of integers l and an integer k, remove all elements from list l that have a value equal to k class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None class LinkedList(object): def __init__(self, head=None): self.head = head
Add initialization to linked list class
Add initialization to linked list class
Python
mit
derekmpham/interview-prep,derekmpham/interview-prep
# Given a singly linked list of integers l and an integer k, remove all elements from list l that have a value equal to k class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None Add initialization to linked list class
# Given a singly linked list of integers l and an integer k, remove all elements from list l that have a value equal to k class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None class LinkedList(object): def __init__(self, head=None): self.head = head
<commit_before># Given a singly linked list of integers l and an integer k, remove all elements from list l that have a value equal to k class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None <commit_msg>Add initialization to linked list class<commit_after>
# Given a singly linked list of integers l and an integer k, remove all elements from list l that have a value equal to k class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None class LinkedList(object): def __init__(self, head=None): self.head = head
# Given a singly linked list of integers l and an integer k, remove all elements from list l that have a value equal to k class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None Add initialization to linked list class# Given a singly linked list of integers l and an integer k, remove all elements from list l that have a value equal to k class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None class LinkedList(object): def __init__(self, head=None): self.head = head
<commit_before># Given a singly linked list of integers l and an integer k, remove all elements from list l that have a value equal to k class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None <commit_msg>Add initialization to linked list class<commit_after># Given a singly linked list of integers l and an integer k, remove all elements from list l that have a value equal to k class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None class LinkedList(object): def __init__(self, head=None): self.head = head
da50d1b66f662f5e3e1b89fd88632f7076c32084
apps/careers/models.py
apps/careers/models.py
from cms import sitemaps from cms.apps.pages.models import ContentBase from cms.models import HtmlField, PageBase from django.db import models from historylinks import shortcuts as historylinks class Careers(ContentBase): classifier = 'apps' urlconf = '{{ project_name }}.apps.careers.urls' per_page = models.PositiveIntegerField( 'careers per page', default=10, blank=True, null=True ) def __str__(self): return self.page.title class Career(PageBase): page = models.ForeignKey( Careers ) title = models.CharField( max_length=256, ) slug = models.CharField( max_length=256, unique=True ) location = models.CharField( max_length=256, blank=True, null=True ) summary = models.TextField( blank=True, null=True ) description = HtmlField() email_address = models.EmailField() order = models.PositiveIntegerField( default=0 ) class Meta: ordering = ['order'] def __str__(self): return self.title def get_absolute_url(self): return self.page.page.reverse('career_detail', kwargs={ 'slug': self.slug, }) historylinks.register(Career) sitemaps.register(Career)
from cms import sitemaps from cms.apps.pages.models import ContentBase from cms.models import HtmlField, PageBase from django.db import models from historylinks import shortcuts as historylinks class Careers(ContentBase): classifier = 'apps' urlconf = '{{ project_name }}.apps.careers.urls' per_page = models.PositiveIntegerField( 'careers per page', default=10, blank=True, null=True ) def __str__(self): return self.page.title class Career(PageBase): page = models.ForeignKey( Careers ) location = models.CharField( max_length=256, blank=True, null=True ) summary = models.TextField( blank=True, null=True ) description = HtmlField() email_address = models.EmailField() order = models.PositiveIntegerField( default=0 ) class Meta: ordering = ['order'] def __str__(self): return self.title def get_absolute_url(self): return self.page.page.reverse('career_detail', kwargs={ 'slug': self.slug, }) historylinks.register(Career) sitemaps.register(Career)
Remove duplicate fields from Career
Remove duplicate fields from Career
Python
mit
onespacemedia/cms-jobs,onespacemedia/cms-jobs
from cms import sitemaps from cms.apps.pages.models import ContentBase from cms.models import HtmlField, PageBase from django.db import models from historylinks import shortcuts as historylinks class Careers(ContentBase): classifier = 'apps' urlconf = '{{ project_name }}.apps.careers.urls' per_page = models.PositiveIntegerField( 'careers per page', default=10, blank=True, null=True ) def __str__(self): return self.page.title class Career(PageBase): page = models.ForeignKey( Careers ) title = models.CharField( max_length=256, ) slug = models.CharField( max_length=256, unique=True ) location = models.CharField( max_length=256, blank=True, null=True ) summary = models.TextField( blank=True, null=True ) description = HtmlField() email_address = models.EmailField() order = models.PositiveIntegerField( default=0 ) class Meta: ordering = ['order'] def __str__(self): return self.title def get_absolute_url(self): return self.page.page.reverse('career_detail', kwargs={ 'slug': self.slug, }) historylinks.register(Career) sitemaps.register(Career) Remove duplicate fields from Career
from cms import sitemaps from cms.apps.pages.models import ContentBase from cms.models import HtmlField, PageBase from django.db import models from historylinks import shortcuts as historylinks class Careers(ContentBase): classifier = 'apps' urlconf = '{{ project_name }}.apps.careers.urls' per_page = models.PositiveIntegerField( 'careers per page', default=10, blank=True, null=True ) def __str__(self): return self.page.title class Career(PageBase): page = models.ForeignKey( Careers ) location = models.CharField( max_length=256, blank=True, null=True ) summary = models.TextField( blank=True, null=True ) description = HtmlField() email_address = models.EmailField() order = models.PositiveIntegerField( default=0 ) class Meta: ordering = ['order'] def __str__(self): return self.title def get_absolute_url(self): return self.page.page.reverse('career_detail', kwargs={ 'slug': self.slug, }) historylinks.register(Career) sitemaps.register(Career)
<commit_before>from cms import sitemaps from cms.apps.pages.models import ContentBase from cms.models import HtmlField, PageBase from django.db import models from historylinks import shortcuts as historylinks class Careers(ContentBase): classifier = 'apps' urlconf = '{{ project_name }}.apps.careers.urls' per_page = models.PositiveIntegerField( 'careers per page', default=10, blank=True, null=True ) def __str__(self): return self.page.title class Career(PageBase): page = models.ForeignKey( Careers ) title = models.CharField( max_length=256, ) slug = models.CharField( max_length=256, unique=True ) location = models.CharField( max_length=256, blank=True, null=True ) summary = models.TextField( blank=True, null=True ) description = HtmlField() email_address = models.EmailField() order = models.PositiveIntegerField( default=0 ) class Meta: ordering = ['order'] def __str__(self): return self.title def get_absolute_url(self): return self.page.page.reverse('career_detail', kwargs={ 'slug': self.slug, }) historylinks.register(Career) sitemaps.register(Career) <commit_msg>Remove duplicate fields from Career<commit_after>
from cms import sitemaps from cms.apps.pages.models import ContentBase from cms.models import HtmlField, PageBase from django.db import models from historylinks import shortcuts as historylinks class Careers(ContentBase): classifier = 'apps' urlconf = '{{ project_name }}.apps.careers.urls' per_page = models.PositiveIntegerField( 'careers per page', default=10, blank=True, null=True ) def __str__(self): return self.page.title class Career(PageBase): page = models.ForeignKey( Careers ) location = models.CharField( max_length=256, blank=True, null=True ) summary = models.TextField( blank=True, null=True ) description = HtmlField() email_address = models.EmailField() order = models.PositiveIntegerField( default=0 ) class Meta: ordering = ['order'] def __str__(self): return self.title def get_absolute_url(self): return self.page.page.reverse('career_detail', kwargs={ 'slug': self.slug, }) historylinks.register(Career) sitemaps.register(Career)
from cms import sitemaps from cms.apps.pages.models import ContentBase from cms.models import HtmlField, PageBase from django.db import models from historylinks import shortcuts as historylinks class Careers(ContentBase): classifier = 'apps' urlconf = '{{ project_name }}.apps.careers.urls' per_page = models.PositiveIntegerField( 'careers per page', default=10, blank=True, null=True ) def __str__(self): return self.page.title class Career(PageBase): page = models.ForeignKey( Careers ) title = models.CharField( max_length=256, ) slug = models.CharField( max_length=256, unique=True ) location = models.CharField( max_length=256, blank=True, null=True ) summary = models.TextField( blank=True, null=True ) description = HtmlField() email_address = models.EmailField() order = models.PositiveIntegerField( default=0 ) class Meta: ordering = ['order'] def __str__(self): return self.title def get_absolute_url(self): return self.page.page.reverse('career_detail', kwargs={ 'slug': self.slug, }) historylinks.register(Career) sitemaps.register(Career) Remove duplicate fields from Careerfrom cms import sitemaps from cms.apps.pages.models import ContentBase from cms.models import HtmlField, PageBase from django.db import models from historylinks import shortcuts as historylinks class Careers(ContentBase): classifier = 'apps' urlconf = '{{ project_name }}.apps.careers.urls' per_page = models.PositiveIntegerField( 'careers per page', default=10, blank=True, null=True ) def __str__(self): return self.page.title class Career(PageBase): page = models.ForeignKey( Careers ) location = models.CharField( max_length=256, blank=True, null=True ) summary = models.TextField( blank=True, null=True ) description = HtmlField() email_address = models.EmailField() order = models.PositiveIntegerField( default=0 ) class Meta: ordering = ['order'] def __str__(self): return self.title def get_absolute_url(self): return self.page.page.reverse('career_detail', kwargs={ 'slug': self.slug, }) historylinks.register(Career) sitemaps.register(Career)
<commit_before>from cms import sitemaps from cms.apps.pages.models import ContentBase from cms.models import HtmlField, PageBase from django.db import models from historylinks import shortcuts as historylinks class Careers(ContentBase): classifier = 'apps' urlconf = '{{ project_name }}.apps.careers.urls' per_page = models.PositiveIntegerField( 'careers per page', default=10, blank=True, null=True ) def __str__(self): return self.page.title class Career(PageBase): page = models.ForeignKey( Careers ) title = models.CharField( max_length=256, ) slug = models.CharField( max_length=256, unique=True ) location = models.CharField( max_length=256, blank=True, null=True ) summary = models.TextField( blank=True, null=True ) description = HtmlField() email_address = models.EmailField() order = models.PositiveIntegerField( default=0 ) class Meta: ordering = ['order'] def __str__(self): return self.title def get_absolute_url(self): return self.page.page.reverse('career_detail', kwargs={ 'slug': self.slug, }) historylinks.register(Career) sitemaps.register(Career) <commit_msg>Remove duplicate fields from Career<commit_after>from cms import sitemaps from cms.apps.pages.models import ContentBase from cms.models import HtmlField, PageBase from django.db import models from historylinks import shortcuts as historylinks class Careers(ContentBase): classifier = 'apps' urlconf = '{{ project_name }}.apps.careers.urls' per_page = models.PositiveIntegerField( 'careers per page', default=10, blank=True, null=True ) def __str__(self): return self.page.title class Career(PageBase): page = models.ForeignKey( Careers ) location = models.CharField( max_length=256, blank=True, null=True ) summary = models.TextField( blank=True, null=True ) description = HtmlField() email_address = models.EmailField() order = models.PositiveIntegerField( default=0 ) class Meta: ordering = ['order'] def __str__(self): return self.title def get_absolute_url(self): return self.page.page.reverse('career_detail', kwargs={ 'slug': self.slug, }) historylinks.register(Career) sitemaps.register(Career)
8416a3ed1a6af2d0037f77744d809441591086cd
mrp_bom_location/models/mrp_bom.py
mrp_bom_location/models/mrp_bom.py
# Copyright 2017 Eficent Business and IT Consulting Services S.L. # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from odoo import fields, models class MrpBom(models.Model): _inherit = "mrp.bom" location_id = fields.Many2one( related='picking_type_id.default_location_dest_id', store=True, ) class MrpBomLine(models.Model): _inherit = "mrp.bom.line" location_id = fields.Many2one( related='bom_id.picking_type_id.default_location_src_id', store=True, )
# Copyright 2017 Eficent Business and IT Consulting Services S.L. # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from odoo import fields, models class MrpBom(models.Model): _inherit = "mrp.bom" location_id = fields.Many2one( related='picking_type_id.default_location_dest_id', readonly=True, store=True, ) class MrpBomLine(models.Model): _inherit = "mrp.bom.line" location_id = fields.Many2one( related='bom_id.picking_type_id.default_location_src_id', readonly=True, store=True, )
Make the related location readonly
[IMP] Make the related location readonly
Python
agpl-3.0
OCA/manufacture,OCA/manufacture
# Copyright 2017 Eficent Business and IT Consulting Services S.L. # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from odoo import fields, models class MrpBom(models.Model): _inherit = "mrp.bom" location_id = fields.Many2one( related='picking_type_id.default_location_dest_id', store=True, ) class MrpBomLine(models.Model): _inherit = "mrp.bom.line" location_id = fields.Many2one( related='bom_id.picking_type_id.default_location_src_id', store=True, ) [IMP] Make the related location readonly
# Copyright 2017 Eficent Business and IT Consulting Services S.L. # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from odoo import fields, models class MrpBom(models.Model): _inherit = "mrp.bom" location_id = fields.Many2one( related='picking_type_id.default_location_dest_id', readonly=True, store=True, ) class MrpBomLine(models.Model): _inherit = "mrp.bom.line" location_id = fields.Many2one( related='bom_id.picking_type_id.default_location_src_id', readonly=True, store=True, )
<commit_before># Copyright 2017 Eficent Business and IT Consulting Services S.L. # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from odoo import fields, models class MrpBom(models.Model): _inherit = "mrp.bom" location_id = fields.Many2one( related='picking_type_id.default_location_dest_id', store=True, ) class MrpBomLine(models.Model): _inherit = "mrp.bom.line" location_id = fields.Many2one( related='bom_id.picking_type_id.default_location_src_id', store=True, ) <commit_msg>[IMP] Make the related location readonly<commit_after>
# Copyright 2017 Eficent Business and IT Consulting Services S.L. # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from odoo import fields, models class MrpBom(models.Model): _inherit = "mrp.bom" location_id = fields.Many2one( related='picking_type_id.default_location_dest_id', readonly=True, store=True, ) class MrpBomLine(models.Model): _inherit = "mrp.bom.line" location_id = fields.Many2one( related='bom_id.picking_type_id.default_location_src_id', readonly=True, store=True, )
# Copyright 2017 Eficent Business and IT Consulting Services S.L. # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from odoo import fields, models class MrpBom(models.Model): _inherit = "mrp.bom" location_id = fields.Many2one( related='picking_type_id.default_location_dest_id', store=True, ) class MrpBomLine(models.Model): _inherit = "mrp.bom.line" location_id = fields.Many2one( related='bom_id.picking_type_id.default_location_src_id', store=True, ) [IMP] Make the related location readonly# Copyright 2017 Eficent Business and IT Consulting Services S.L. # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from odoo import fields, models class MrpBom(models.Model): _inherit = "mrp.bom" location_id = fields.Many2one( related='picking_type_id.default_location_dest_id', readonly=True, store=True, ) class MrpBomLine(models.Model): _inherit = "mrp.bom.line" location_id = fields.Many2one( related='bom_id.picking_type_id.default_location_src_id', readonly=True, store=True, )
<commit_before># Copyright 2017 Eficent Business and IT Consulting Services S.L. # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from odoo import fields, models class MrpBom(models.Model): _inherit = "mrp.bom" location_id = fields.Many2one( related='picking_type_id.default_location_dest_id', store=True, ) class MrpBomLine(models.Model): _inherit = "mrp.bom.line" location_id = fields.Many2one( related='bom_id.picking_type_id.default_location_src_id', store=True, ) <commit_msg>[IMP] Make the related location readonly<commit_after># Copyright 2017 Eficent Business and IT Consulting Services S.L. # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from odoo import fields, models class MrpBom(models.Model): _inherit = "mrp.bom" location_id = fields.Many2one( related='picking_type_id.default_location_dest_id', readonly=True, store=True, ) class MrpBomLine(models.Model): _inherit = "mrp.bom.line" location_id = fields.Many2one( related='bom_id.picking_type_id.default_location_src_id', readonly=True, store=True, )
57a7651ba9583830ab32fae0bb8d790bb2bdb6a8
setup.py
setup.py
#!/usr/bin/env python from setuptools import setup setup(name='programmabletuple', version='0.2.0', description='Python metaclass for making named tuples with programmability', long_description=open('README.rst').read(), author='Tschijnmo TSCHAU', author_email='tschijnmotschau@gmail.com', url='https://github.com/tschijnmo/programmabletuple', license='MIT', packages=['programmabletuple', ], classifiers=[ 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries :: Python Modules', ], )
#!/usr/bin/env python from setuptools import setup setup(name='programmabletuple', version='0.3.0', description='Python metaclass for making named tuples with programmability', long_description=open('README.rst').read(), author='Tschijnmo TSCHAU', author_email='tschijnmotschau@gmail.com', url='https://github.com/tschijnmo/programmabletuple', license='MIT', packages=['programmabletuple', ], classifiers=[ 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries :: Python Modules', ], )
Change the version number to 3.0
Change the version number to 3.0 The most notable change from 2.0 is the new initializer.
Python
mit
tschijnmo/programmabletuple
#!/usr/bin/env python from setuptools import setup setup(name='programmabletuple', version='0.2.0', description='Python metaclass for making named tuples with programmability', long_description=open('README.rst').read(), author='Tschijnmo TSCHAU', author_email='tschijnmotschau@gmail.com', url='https://github.com/tschijnmo/programmabletuple', license='MIT', packages=['programmabletuple', ], classifiers=[ 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries :: Python Modules', ], ) Change the version number to 3.0 The most notable change from 2.0 is the new initializer.
#!/usr/bin/env python from setuptools import setup setup(name='programmabletuple', version='0.3.0', description='Python metaclass for making named tuples with programmability', long_description=open('README.rst').read(), author='Tschijnmo TSCHAU', author_email='tschijnmotschau@gmail.com', url='https://github.com/tschijnmo/programmabletuple', license='MIT', packages=['programmabletuple', ], classifiers=[ 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries :: Python Modules', ], )
<commit_before>#!/usr/bin/env python from setuptools import setup setup(name='programmabletuple', version='0.2.0', description='Python metaclass for making named tuples with programmability', long_description=open('README.rst').read(), author='Tschijnmo TSCHAU', author_email='tschijnmotschau@gmail.com', url='https://github.com/tschijnmo/programmabletuple', license='MIT', packages=['programmabletuple', ], classifiers=[ 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries :: Python Modules', ], ) <commit_msg>Change the version number to 3.0 The most notable change from 2.0 is the new initializer.<commit_after>
#!/usr/bin/env python from setuptools import setup setup(name='programmabletuple', version='0.3.0', description='Python metaclass for making named tuples with programmability', long_description=open('README.rst').read(), author='Tschijnmo TSCHAU', author_email='tschijnmotschau@gmail.com', url='https://github.com/tschijnmo/programmabletuple', license='MIT', packages=['programmabletuple', ], classifiers=[ 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries :: Python Modules', ], )
#!/usr/bin/env python from setuptools import setup setup(name='programmabletuple', version='0.2.0', description='Python metaclass for making named tuples with programmability', long_description=open('README.rst').read(), author='Tschijnmo TSCHAU', author_email='tschijnmotschau@gmail.com', url='https://github.com/tschijnmo/programmabletuple', license='MIT', packages=['programmabletuple', ], classifiers=[ 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries :: Python Modules', ], ) Change the version number to 3.0 The most notable change from 2.0 is the new initializer.#!/usr/bin/env python from setuptools import setup setup(name='programmabletuple', version='0.3.0', description='Python metaclass for making named tuples with programmability', long_description=open('README.rst').read(), author='Tschijnmo TSCHAU', author_email='tschijnmotschau@gmail.com', url='https://github.com/tschijnmo/programmabletuple', license='MIT', packages=['programmabletuple', ], classifiers=[ 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries :: Python Modules', ], )
<commit_before>#!/usr/bin/env python from setuptools import setup setup(name='programmabletuple', version='0.2.0', description='Python metaclass for making named tuples with programmability', long_description=open('README.rst').read(), author='Tschijnmo TSCHAU', author_email='tschijnmotschau@gmail.com', url='https://github.com/tschijnmo/programmabletuple', license='MIT', packages=['programmabletuple', ], classifiers=[ 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries :: Python Modules', ], ) <commit_msg>Change the version number to 3.0 The most notable change from 2.0 is the new initializer.<commit_after>#!/usr/bin/env python from setuptools import setup setup(name='programmabletuple', version='0.3.0', description='Python metaclass for making named tuples with programmability', long_description=open('README.rst').read(), author='Tschijnmo TSCHAU', author_email='tschijnmotschau@gmail.com', url='https://github.com/tschijnmo/programmabletuple', license='MIT', packages=['programmabletuple', ], classifiers=[ 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries :: Python Modules', ], )
21afbaab7deb874703f4968ea1337b59120f0ad0
music-stream.py
music-stream.py
import urllib.request import subprocess LIMIT = 10 PLAYER = 'vlc' url = 'http://streams.twitch.tv/kraken/streams?limit='+str(LIMIT)+'&offset=0&game=Music&broadcaster_language=&on_site=1' with urllib.request.urlopen(url) as response: html = response.read().decode('utf8') i = 0 urls = [] for line in html.split(','): if 'status' in line: status = line.split('"')[-2] status = ''.join(i for i in status if ord(i)<128) #filter non ascii characters if 'display_name' in line: name = line.split('"')[-2] print(str(i) + ') ' + name + ' : ' + status) i += 1 if 'url' in line: url = line.split('"')[-2] urls.append(url) choice = LIMIT while (choice >= LIMIT): choice = int(input('Choose a stream\n')) cmd = ['livestreamer', urls[choice], 'audio'] if PLAYER != 'vlc': cmd.append('-p') cmd.append(PLAYER) subprocess.Popen(cmd, shell=False)
import urllib.request import subprocess LIMIT = 10 PLAYER = 'vlc' STREAMS_URL = 'http://streams.twitch.tv/kraken/streams?limit='+str(LIMIT)+'&offset=0&game=Music&broadcaster_language=&on_site=1' while True: with urllib.request.urlopen(STREAMS_URL) as response: html = response.read().decode('utf8') i = 0 urls = [] for line in html.split(','): if 'status' in line: status = line.split('"')[-2] status = ''.join(i for i in status if ord(i)<128) #filter non ascii characters if 'display_name' in line: name = line.split('"')[-2] print(str(i) + ') ' + name + ' : ' + status) i += 1 if 'url' in line: url = line.split('"')[-2] urls.append(url) choice = LIMIT while (choice >= LIMIT): choice = int(input('Choose a stream\n')) cmd = ['livestreamer', urls[choice], 'audio'] if PLAYER != 'vlc': cmd.append('-p') cmd.append(PLAYER) subprocess.call(cmd, shell=False) print('\n\n\n')
Refresh streams list when player is closed
Refresh streams list when player is closed
Python
mit
GaudyZircon/music-stream
import urllib.request import subprocess LIMIT = 10 PLAYER = 'vlc' url = 'http://streams.twitch.tv/kraken/streams?limit='+str(LIMIT)+'&offset=0&game=Music&broadcaster_language=&on_site=1' with urllib.request.urlopen(url) as response: html = response.read().decode('utf8') i = 0 urls = [] for line in html.split(','): if 'status' in line: status = line.split('"')[-2] status = ''.join(i for i in status if ord(i)<128) #filter non ascii characters if 'display_name' in line: name = line.split('"')[-2] print(str(i) + ') ' + name + ' : ' + status) i += 1 if 'url' in line: url = line.split('"')[-2] urls.append(url) choice = LIMIT while (choice >= LIMIT): choice = int(input('Choose a stream\n')) cmd = ['livestreamer', urls[choice], 'audio'] if PLAYER != 'vlc': cmd.append('-p') cmd.append(PLAYER) subprocess.Popen(cmd, shell=False) Refresh streams list when player is closed
import urllib.request import subprocess LIMIT = 10 PLAYER = 'vlc' STREAMS_URL = 'http://streams.twitch.tv/kraken/streams?limit='+str(LIMIT)+'&offset=0&game=Music&broadcaster_language=&on_site=1' while True: with urllib.request.urlopen(STREAMS_URL) as response: html = response.read().decode('utf8') i = 0 urls = [] for line in html.split(','): if 'status' in line: status = line.split('"')[-2] status = ''.join(i for i in status if ord(i)<128) #filter non ascii characters if 'display_name' in line: name = line.split('"')[-2] print(str(i) + ') ' + name + ' : ' + status) i += 1 if 'url' in line: url = line.split('"')[-2] urls.append(url) choice = LIMIT while (choice >= LIMIT): choice = int(input('Choose a stream\n')) cmd = ['livestreamer', urls[choice], 'audio'] if PLAYER != 'vlc': cmd.append('-p') cmd.append(PLAYER) subprocess.call(cmd, shell=False) print('\n\n\n')
<commit_before>import urllib.request import subprocess LIMIT = 10 PLAYER = 'vlc' url = 'http://streams.twitch.tv/kraken/streams?limit='+str(LIMIT)+'&offset=0&game=Music&broadcaster_language=&on_site=1' with urllib.request.urlopen(url) as response: html = response.read().decode('utf8') i = 0 urls = [] for line in html.split(','): if 'status' in line: status = line.split('"')[-2] status = ''.join(i for i in status if ord(i)<128) #filter non ascii characters if 'display_name' in line: name = line.split('"')[-2] print(str(i) + ') ' + name + ' : ' + status) i += 1 if 'url' in line: url = line.split('"')[-2] urls.append(url) choice = LIMIT while (choice >= LIMIT): choice = int(input('Choose a stream\n')) cmd = ['livestreamer', urls[choice], 'audio'] if PLAYER != 'vlc': cmd.append('-p') cmd.append(PLAYER) subprocess.Popen(cmd, shell=False) <commit_msg>Refresh streams list when player is closed<commit_after>
import urllib.request import subprocess LIMIT = 10 PLAYER = 'vlc' STREAMS_URL = 'http://streams.twitch.tv/kraken/streams?limit='+str(LIMIT)+'&offset=0&game=Music&broadcaster_language=&on_site=1' while True: with urllib.request.urlopen(STREAMS_URL) as response: html = response.read().decode('utf8') i = 0 urls = [] for line in html.split(','): if 'status' in line: status = line.split('"')[-2] status = ''.join(i for i in status if ord(i)<128) #filter non ascii characters if 'display_name' in line: name = line.split('"')[-2] print(str(i) + ') ' + name + ' : ' + status) i += 1 if 'url' in line: url = line.split('"')[-2] urls.append(url) choice = LIMIT while (choice >= LIMIT): choice = int(input('Choose a stream\n')) cmd = ['livestreamer', urls[choice], 'audio'] if PLAYER != 'vlc': cmd.append('-p') cmd.append(PLAYER) subprocess.call(cmd, shell=False) print('\n\n\n')
import urllib.request import subprocess LIMIT = 10 PLAYER = 'vlc' url = 'http://streams.twitch.tv/kraken/streams?limit='+str(LIMIT)+'&offset=0&game=Music&broadcaster_language=&on_site=1' with urllib.request.urlopen(url) as response: html = response.read().decode('utf8') i = 0 urls = [] for line in html.split(','): if 'status' in line: status = line.split('"')[-2] status = ''.join(i for i in status if ord(i)<128) #filter non ascii characters if 'display_name' in line: name = line.split('"')[-2] print(str(i) + ') ' + name + ' : ' + status) i += 1 if 'url' in line: url = line.split('"')[-2] urls.append(url) choice = LIMIT while (choice >= LIMIT): choice = int(input('Choose a stream\n')) cmd = ['livestreamer', urls[choice], 'audio'] if PLAYER != 'vlc': cmd.append('-p') cmd.append(PLAYER) subprocess.Popen(cmd, shell=False) Refresh streams list when player is closedimport urllib.request import subprocess LIMIT = 10 PLAYER = 'vlc' STREAMS_URL = 'http://streams.twitch.tv/kraken/streams?limit='+str(LIMIT)+'&offset=0&game=Music&broadcaster_language=&on_site=1' while True: with urllib.request.urlopen(STREAMS_URL) as response: html = response.read().decode('utf8') i = 0 urls = [] for line in html.split(','): if 'status' in line: status = line.split('"')[-2] status = ''.join(i for i in status if ord(i)<128) #filter non ascii characters if 'display_name' in line: name = line.split('"')[-2] print(str(i) + ') ' + name + ' : ' + status) i += 1 if 'url' in line: url = line.split('"')[-2] urls.append(url) choice = LIMIT while (choice >= LIMIT): choice = int(input('Choose a stream\n')) cmd = ['livestreamer', urls[choice], 'audio'] if PLAYER != 'vlc': cmd.append('-p') cmd.append(PLAYER) subprocess.call(cmd, shell=False) print('\n\n\n')
<commit_before>import urllib.request import subprocess LIMIT = 10 PLAYER = 'vlc' url = 'http://streams.twitch.tv/kraken/streams?limit='+str(LIMIT)+'&offset=0&game=Music&broadcaster_language=&on_site=1' with urllib.request.urlopen(url) as response: html = response.read().decode('utf8') i = 0 urls = [] for line in html.split(','): if 'status' in line: status = line.split('"')[-2] status = ''.join(i for i in status if ord(i)<128) #filter non ascii characters if 'display_name' in line: name = line.split('"')[-2] print(str(i) + ') ' + name + ' : ' + status) i += 1 if 'url' in line: url = line.split('"')[-2] urls.append(url) choice = LIMIT while (choice >= LIMIT): choice = int(input('Choose a stream\n')) cmd = ['livestreamer', urls[choice], 'audio'] if PLAYER != 'vlc': cmd.append('-p') cmd.append(PLAYER) subprocess.Popen(cmd, shell=False) <commit_msg>Refresh streams list when player is closed<commit_after>import urllib.request import subprocess LIMIT = 10 PLAYER = 'vlc' STREAMS_URL = 'http://streams.twitch.tv/kraken/streams?limit='+str(LIMIT)+'&offset=0&game=Music&broadcaster_language=&on_site=1' while True: with urllib.request.urlopen(STREAMS_URL) as response: html = response.read().decode('utf8') i = 0 urls = [] for line in html.split(','): if 'status' in line: status = line.split('"')[-2] status = ''.join(i for i in status if ord(i)<128) #filter non ascii characters if 'display_name' in line: name = line.split('"')[-2] print(str(i) + ') ' + name + ' : ' + status) i += 1 if 'url' in line: url = line.split('"')[-2] urls.append(url) choice = LIMIT while (choice >= LIMIT): choice = int(input('Choose a stream\n')) cmd = ['livestreamer', urls[choice], 'audio'] if PLAYER != 'vlc': cmd.append('-p') cmd.append(PLAYER) subprocess.call(cmd, shell=False) print('\n\n\n')
9d9827721e3d4c45f8917662d2f59759fb4ecd66
muv/__init__.py
muv/__init__.py
""" Miscellaneous utilities. """ import numpy as np def kennard_stone(d, k): """ Use the Kennard-Stone algorithm to select k maximally separated examples from a dataset. See R. W. Kennard and L. A. Stone (1969): Computer Aided Design of Experiments, Technometrics, 11:1, 137-148. Algorithm --------- 1. Choose the two examples separated by the largest distance. In the case of a tie, use the first examples returned by np.where. 2. For the remaining k - 2 selections, choose the example with the greatest distance to the closest example among all previously chosen points. Parameters ---------- d : ndarray Pairwise distance matrix between dataset examples. k : int Number of examples to select. """ assert 1 < k < d.shape[0] chosen = [] # choose initial points first = np.where(d == np.amax(d)) chosen.append(first[0][0]) chosen.append(first[1][0]) d = np.ma.array(d, mask=np.ones_like(d, dtype=bool)) # choose remaining points while len(chosen) < k: d.mask[:, chosen] = False d.mask[chosen] = True print d p = np.ma.argmax(np.ma.amin(d, axis=1)) chosen.append(p) return chosen
""" Miscellaneous utilities. """ import numpy as np class MUV(object): """ Generate maximum unbiased validation (MUV) datasets for virtual screening as described in Rohrer and Baumann, J. Chem. Inf. Model. 2009, 49, 169-184. """ def kennard_stone(d, k): """ Use the Kennard-Stone algorithm to select k maximally separated examples from a dataset. See Kennard and Stone, Technometrics 1969, 11, 137-148. Algorithm --------- 1. Choose the two examples separated by the largest distance. In the case of a tie, use the first examples returned by np.where. 2. For the remaining k - 2 selections, choose the example with the greatest distance to the closest example among all previously chosen points. Parameters ---------- d : ndarray Pairwise distance matrix between dataset examples. k : int Number of examples to select. """ assert 1 < k < d.shape[0] chosen = [] # choose initial points first = np.where(d == np.amax(d)) chosen.append(first[0][0]) chosen.append(first[1][0]) d = np.ma.array(d, mask=np.ones_like(d, dtype=bool)) # choose remaining points while len(chosen) < k: d.mask[:, chosen] = False d.mask[chosen] = True print d p = np.ma.argmax(np.ma.amin(d, axis=1)) chosen.append(p) return chosen
Fix reference and add MUV class
Fix reference and add MUV class
Python
bsd-3-clause
skearnes/muv
""" Miscellaneous utilities. """ import numpy as np def kennard_stone(d, k): """ Use the Kennard-Stone algorithm to select k maximally separated examples from a dataset. See R. W. Kennard and L. A. Stone (1969): Computer Aided Design of Experiments, Technometrics, 11:1, 137-148. Algorithm --------- 1. Choose the two examples separated by the largest distance. In the case of a tie, use the first examples returned by np.where. 2. For the remaining k - 2 selections, choose the example with the greatest distance to the closest example among all previously chosen points. Parameters ---------- d : ndarray Pairwise distance matrix between dataset examples. k : int Number of examples to select. """ assert 1 < k < d.shape[0] chosen = [] # choose initial points first = np.where(d == np.amax(d)) chosen.append(first[0][0]) chosen.append(first[1][0]) d = np.ma.array(d, mask=np.ones_like(d, dtype=bool)) # choose remaining points while len(chosen) < k: d.mask[:, chosen] = False d.mask[chosen] = True print d p = np.ma.argmax(np.ma.amin(d, axis=1)) chosen.append(p) return chosen Fix reference and add MUV class
""" Miscellaneous utilities. """ import numpy as np class MUV(object): """ Generate maximum unbiased validation (MUV) datasets for virtual screening as described in Rohrer and Baumann, J. Chem. Inf. Model. 2009, 49, 169-184. """ def kennard_stone(d, k): """ Use the Kennard-Stone algorithm to select k maximally separated examples from a dataset. See Kennard and Stone, Technometrics 1969, 11, 137-148. Algorithm --------- 1. Choose the two examples separated by the largest distance. In the case of a tie, use the first examples returned by np.where. 2. For the remaining k - 2 selections, choose the example with the greatest distance to the closest example among all previously chosen points. Parameters ---------- d : ndarray Pairwise distance matrix between dataset examples. k : int Number of examples to select. """ assert 1 < k < d.shape[0] chosen = [] # choose initial points first = np.where(d == np.amax(d)) chosen.append(first[0][0]) chosen.append(first[1][0]) d = np.ma.array(d, mask=np.ones_like(d, dtype=bool)) # choose remaining points while len(chosen) < k: d.mask[:, chosen] = False d.mask[chosen] = True print d p = np.ma.argmax(np.ma.amin(d, axis=1)) chosen.append(p) return chosen
<commit_before>""" Miscellaneous utilities. """ import numpy as np def kennard_stone(d, k): """ Use the Kennard-Stone algorithm to select k maximally separated examples from a dataset. See R. W. Kennard and L. A. Stone (1969): Computer Aided Design of Experiments, Technometrics, 11:1, 137-148. Algorithm --------- 1. Choose the two examples separated by the largest distance. In the case of a tie, use the first examples returned by np.where. 2. For the remaining k - 2 selections, choose the example with the greatest distance to the closest example among all previously chosen points. Parameters ---------- d : ndarray Pairwise distance matrix between dataset examples. k : int Number of examples to select. """ assert 1 < k < d.shape[0] chosen = [] # choose initial points first = np.where(d == np.amax(d)) chosen.append(first[0][0]) chosen.append(first[1][0]) d = np.ma.array(d, mask=np.ones_like(d, dtype=bool)) # choose remaining points while len(chosen) < k: d.mask[:, chosen] = False d.mask[chosen] = True print d p = np.ma.argmax(np.ma.amin(d, axis=1)) chosen.append(p) return chosen <commit_msg>Fix reference and add MUV class<commit_after>
""" Miscellaneous utilities. """ import numpy as np class MUV(object): """ Generate maximum unbiased validation (MUV) datasets for virtual screening as described in Rohrer and Baumann, J. Chem. Inf. Model. 2009, 49, 169-184. """ def kennard_stone(d, k): """ Use the Kennard-Stone algorithm to select k maximally separated examples from a dataset. See Kennard and Stone, Technometrics 1969, 11, 137-148. Algorithm --------- 1. Choose the two examples separated by the largest distance. In the case of a tie, use the first examples returned by np.where. 2. For the remaining k - 2 selections, choose the example with the greatest distance to the closest example among all previously chosen points. Parameters ---------- d : ndarray Pairwise distance matrix between dataset examples. k : int Number of examples to select. """ assert 1 < k < d.shape[0] chosen = [] # choose initial points first = np.where(d == np.amax(d)) chosen.append(first[0][0]) chosen.append(first[1][0]) d = np.ma.array(d, mask=np.ones_like(d, dtype=bool)) # choose remaining points while len(chosen) < k: d.mask[:, chosen] = False d.mask[chosen] = True print d p = np.ma.argmax(np.ma.amin(d, axis=1)) chosen.append(p) return chosen
""" Miscellaneous utilities. """ import numpy as np def kennard_stone(d, k): """ Use the Kennard-Stone algorithm to select k maximally separated examples from a dataset. See R. W. Kennard and L. A. Stone (1969): Computer Aided Design of Experiments, Technometrics, 11:1, 137-148. Algorithm --------- 1. Choose the two examples separated by the largest distance. In the case of a tie, use the first examples returned by np.where. 2. For the remaining k - 2 selections, choose the example with the greatest distance to the closest example among all previously chosen points. Parameters ---------- d : ndarray Pairwise distance matrix between dataset examples. k : int Number of examples to select. """ assert 1 < k < d.shape[0] chosen = [] # choose initial points first = np.where(d == np.amax(d)) chosen.append(first[0][0]) chosen.append(first[1][0]) d = np.ma.array(d, mask=np.ones_like(d, dtype=bool)) # choose remaining points while len(chosen) < k: d.mask[:, chosen] = False d.mask[chosen] = True print d p = np.ma.argmax(np.ma.amin(d, axis=1)) chosen.append(p) return chosen Fix reference and add MUV class""" Miscellaneous utilities. """ import numpy as np class MUV(object): """ Generate maximum unbiased validation (MUV) datasets for virtual screening as described in Rohrer and Baumann, J. Chem. Inf. Model. 2009, 49, 169-184. """ def kennard_stone(d, k): """ Use the Kennard-Stone algorithm to select k maximally separated examples from a dataset. See Kennard and Stone, Technometrics 1969, 11, 137-148. Algorithm --------- 1. Choose the two examples separated by the largest distance. In the case of a tie, use the first examples returned by np.where. 2. For the remaining k - 2 selections, choose the example with the greatest distance to the closest example among all previously chosen points. Parameters ---------- d : ndarray Pairwise distance matrix between dataset examples. k : int Number of examples to select. """ assert 1 < k < d.shape[0] chosen = [] # choose initial points first = np.where(d == np.amax(d)) chosen.append(first[0][0]) chosen.append(first[1][0]) d = np.ma.array(d, mask=np.ones_like(d, dtype=bool)) # choose remaining points while len(chosen) < k: d.mask[:, chosen] = False d.mask[chosen] = True print d p = np.ma.argmax(np.ma.amin(d, axis=1)) chosen.append(p) return chosen
<commit_before>""" Miscellaneous utilities. """ import numpy as np def kennard_stone(d, k): """ Use the Kennard-Stone algorithm to select k maximally separated examples from a dataset. See R. W. Kennard and L. A. Stone (1969): Computer Aided Design of Experiments, Technometrics, 11:1, 137-148. Algorithm --------- 1. Choose the two examples separated by the largest distance. In the case of a tie, use the first examples returned by np.where. 2. For the remaining k - 2 selections, choose the example with the greatest distance to the closest example among all previously chosen points. Parameters ---------- d : ndarray Pairwise distance matrix between dataset examples. k : int Number of examples to select. """ assert 1 < k < d.shape[0] chosen = [] # choose initial points first = np.where(d == np.amax(d)) chosen.append(first[0][0]) chosen.append(first[1][0]) d = np.ma.array(d, mask=np.ones_like(d, dtype=bool)) # choose remaining points while len(chosen) < k: d.mask[:, chosen] = False d.mask[chosen] = True print d p = np.ma.argmax(np.ma.amin(d, axis=1)) chosen.append(p) return chosen <commit_msg>Fix reference and add MUV class<commit_after>""" Miscellaneous utilities. """ import numpy as np class MUV(object): """ Generate maximum unbiased validation (MUV) datasets for virtual screening as described in Rohrer and Baumann, J. Chem. Inf. Model. 2009, 49, 169-184. """ def kennard_stone(d, k): """ Use the Kennard-Stone algorithm to select k maximally separated examples from a dataset. See Kennard and Stone, Technometrics 1969, 11, 137-148. Algorithm --------- 1. Choose the two examples separated by the largest distance. In the case of a tie, use the first examples returned by np.where. 2. For the remaining k - 2 selections, choose the example with the greatest distance to the closest example among all previously chosen points. Parameters ---------- d : ndarray Pairwise distance matrix between dataset examples. k : int Number of examples to select. """ assert 1 < k < d.shape[0] chosen = [] # choose initial points first = np.where(d == np.amax(d)) chosen.append(first[0][0]) chosen.append(first[1][0]) d = np.ma.array(d, mask=np.ones_like(d, dtype=bool)) # choose remaining points while len(chosen) < k: d.mask[:, chosen] = False d.mask[chosen] = True print d p = np.ma.argmax(np.ma.amin(d, axis=1)) chosen.append(p) return chosen
50a1f9cbd5e9ab3a279e195bba06708aae58d3c2
setup.py
setup.py
from setuptools import setup, find_packages setup(name='notation', description='Better human-readable numbers.', long_description=open('README.rst').read(), author='Stijn Debrouwere', author_email='stijn@debrouwere.org', url='https://github.com/debrouwere/python-notation/', download_url='https://www.github.com/debrouwere/python-notation/tarball/master', version='1.0.0', license='ISC', packages=find_packages(), keywords='human numbers format notation scientific engineering', install_requires=[], test_suite='notation.tests', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: ISC License', 'Operating System :: OS Independent', 'Topic :: Scientific/Engineering :: Information Analysis', 'Topic :: Utilities', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', ], )
from setuptools import setup, find_packages setup(name='notation', description='Better human-readable numbers.', long_description=open('README.rst').read(), author='Stijn Debrouwere', author_email='stijn@debrouwere.org', url='https://github.com/debrouwere/python-notation/', download_url='https://www.github.com/debrouwere/python-notation/tarball/master', version='1.0.0', license='ISC', packages=find_packages(), keywords='human numbers format notation scientific engineering', install_requires=[], test_suite='notation.tests', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: ISC License (ISCL)', 'Operating System :: OS Independent', 'Topic :: Scientific/Engineering :: Information Analysis', 'Topic :: Utilities', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', ], )
Fix reference to ISC license.
Fix reference to ISC license.
Python
isc
debrouwere/python-ballpark
from setuptools import setup, find_packages setup(name='notation', description='Better human-readable numbers.', long_description=open('README.rst').read(), author='Stijn Debrouwere', author_email='stijn@debrouwere.org', url='https://github.com/debrouwere/python-notation/', download_url='https://www.github.com/debrouwere/python-notation/tarball/master', version='1.0.0', license='ISC', packages=find_packages(), keywords='human numbers format notation scientific engineering', install_requires=[], test_suite='notation.tests', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: ISC License', 'Operating System :: OS Independent', 'Topic :: Scientific/Engineering :: Information Analysis', 'Topic :: Utilities', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', ], ) Fix reference to ISC license.
from setuptools import setup, find_packages setup(name='notation', description='Better human-readable numbers.', long_description=open('README.rst').read(), author='Stijn Debrouwere', author_email='stijn@debrouwere.org', url='https://github.com/debrouwere/python-notation/', download_url='https://www.github.com/debrouwere/python-notation/tarball/master', version='1.0.0', license='ISC', packages=find_packages(), keywords='human numbers format notation scientific engineering', install_requires=[], test_suite='notation.tests', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: ISC License (ISCL)', 'Operating System :: OS Independent', 'Topic :: Scientific/Engineering :: Information Analysis', 'Topic :: Utilities', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', ], )
<commit_before>from setuptools import setup, find_packages setup(name='notation', description='Better human-readable numbers.', long_description=open('README.rst').read(), author='Stijn Debrouwere', author_email='stijn@debrouwere.org', url='https://github.com/debrouwere/python-notation/', download_url='https://www.github.com/debrouwere/python-notation/tarball/master', version='1.0.0', license='ISC', packages=find_packages(), keywords='human numbers format notation scientific engineering', install_requires=[], test_suite='notation.tests', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: ISC License', 'Operating System :: OS Independent', 'Topic :: Scientific/Engineering :: Information Analysis', 'Topic :: Utilities', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', ], ) <commit_msg>Fix reference to ISC license.<commit_after>
from setuptools import setup, find_packages setup(name='notation', description='Better human-readable numbers.', long_description=open('README.rst').read(), author='Stijn Debrouwere', author_email='stijn@debrouwere.org', url='https://github.com/debrouwere/python-notation/', download_url='https://www.github.com/debrouwere/python-notation/tarball/master', version='1.0.0', license='ISC', packages=find_packages(), keywords='human numbers format notation scientific engineering', install_requires=[], test_suite='notation.tests', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: ISC License (ISCL)', 'Operating System :: OS Independent', 'Topic :: Scientific/Engineering :: Information Analysis', 'Topic :: Utilities', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', ], )
from setuptools import setup, find_packages setup(name='notation', description='Better human-readable numbers.', long_description=open('README.rst').read(), author='Stijn Debrouwere', author_email='stijn@debrouwere.org', url='https://github.com/debrouwere/python-notation/', download_url='https://www.github.com/debrouwere/python-notation/tarball/master', version='1.0.0', license='ISC', packages=find_packages(), keywords='human numbers format notation scientific engineering', install_requires=[], test_suite='notation.tests', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: ISC License', 'Operating System :: OS Independent', 'Topic :: Scientific/Engineering :: Information Analysis', 'Topic :: Utilities', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', ], ) Fix reference to ISC license.from setuptools import setup, find_packages setup(name='notation', description='Better human-readable numbers.', long_description=open('README.rst').read(), author='Stijn Debrouwere', author_email='stijn@debrouwere.org', url='https://github.com/debrouwere/python-notation/', download_url='https://www.github.com/debrouwere/python-notation/tarball/master', version='1.0.0', license='ISC', packages=find_packages(), keywords='human numbers format notation scientific engineering', install_requires=[], test_suite='notation.tests', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: ISC License (ISCL)', 'Operating System :: OS Independent', 'Topic :: Scientific/Engineering :: Information Analysis', 'Topic :: Utilities', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', ], )
<commit_before>from setuptools import setup, find_packages setup(name='notation', description='Better human-readable numbers.', long_description=open('README.rst').read(), author='Stijn Debrouwere', author_email='stijn@debrouwere.org', url='https://github.com/debrouwere/python-notation/', download_url='https://www.github.com/debrouwere/python-notation/tarball/master', version='1.0.0', license='ISC', packages=find_packages(), keywords='human numbers format notation scientific engineering', install_requires=[], test_suite='notation.tests', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: ISC License', 'Operating System :: OS Independent', 'Topic :: Scientific/Engineering :: Information Analysis', 'Topic :: Utilities', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', ], ) <commit_msg>Fix reference to ISC license.<commit_after>from setuptools import setup, find_packages setup(name='notation', description='Better human-readable numbers.', long_description=open('README.rst').read(), author='Stijn Debrouwere', author_email='stijn@debrouwere.org', url='https://github.com/debrouwere/python-notation/', download_url='https://www.github.com/debrouwere/python-notation/tarball/master', version='1.0.0', license='ISC', packages=find_packages(), keywords='human numbers format notation scientific engineering', install_requires=[], test_suite='notation.tests', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: ISC License (ISCL)', 'Operating System :: OS Independent', 'Topic :: Scientific/Engineering :: Information Analysis', 'Topic :: Utilities', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.4', ], )
06e2d872bda93ed3b520e6ab7ee646aa2f7a0fcd
setup.py
setup.py
##################################################### # Copyright (c) Xuanyi Dong [GitHub D-X-Y], 2019.08 # ##################################################### # [2020.03.09] Upgrade to v1.2 import os from setuptools import setup def read(fname='README.md'): with open(os.path.join(os.path.dirname(__file__), fname), encoding='utf-8') as cfile: return cfile.read() setup( name = "nas_bench_201", version = "1.2", author = "Xuanyi Dong", author_email = "dongxuanyi888@gmail.com", description = "API for NAS-Bench-201 (a benchmark for neural architecture search).", license = "MIT", keywords = "NAS Dataset API DeepLearning", url = "https://github.com/D-X-Y/NAS-Bench-201", packages=['nas_201_api'], long_description=read('README.md'), long_description_content_type='text/markdown', classifiers=[ "Programming Language :: Python", "Topic :: Database", "Topic :: Scientific/Engineering :: Artificial Intelligence", "License :: OSI Approved :: MIT License", ], )
##################################################### # Copyright (c) Xuanyi Dong [GitHub D-X-Y], 2019.08 # ##################################################### # [2020.02.25] Initialize the API as v1.1 # [2020.03.09] Upgrade the API to v1.2 # [2020.03.16] Upgrade the API to v1.3 import os from setuptools import setup def read(fname='README.md'): with open(os.path.join(os.path.dirname(__file__), fname), encoding='utf-8') as cfile: return cfile.read() setup( name = "nas_bench_201", version = "1.3", author = "Xuanyi Dong", author_email = "dongxuanyi888@gmail.com", description = "API for NAS-Bench-201 (a benchmark for neural architecture search).", license = "MIT", keywords = "NAS Dataset API DeepLearning", url = "https://github.com/D-X-Y/NAS-Bench-201", packages=['nas_201_api'], long_description=read('README.md'), long_description_content_type='text/markdown', classifiers=[ "Programming Language :: Python", "Topic :: Database", "Topic :: Scientific/Engineering :: Artificial Intelligence", "License :: OSI Approved :: MIT License", ], )
Upgrade API to v1.3 and benchmark file to v1.1
Upgrade API to v1.3 and benchmark file to v1.1
Python
mit
D-X-Y/ResNeXt,D-X-Y/ResNeXt
##################################################### # Copyright (c) Xuanyi Dong [GitHub D-X-Y], 2019.08 # ##################################################### # [2020.03.09] Upgrade to v1.2 import os from setuptools import setup def read(fname='README.md'): with open(os.path.join(os.path.dirname(__file__), fname), encoding='utf-8') as cfile: return cfile.read() setup( name = "nas_bench_201", version = "1.2", author = "Xuanyi Dong", author_email = "dongxuanyi888@gmail.com", description = "API for NAS-Bench-201 (a benchmark for neural architecture search).", license = "MIT", keywords = "NAS Dataset API DeepLearning", url = "https://github.com/D-X-Y/NAS-Bench-201", packages=['nas_201_api'], long_description=read('README.md'), long_description_content_type='text/markdown', classifiers=[ "Programming Language :: Python", "Topic :: Database", "Topic :: Scientific/Engineering :: Artificial Intelligence", "License :: OSI Approved :: MIT License", ], ) Upgrade API to v1.3 and benchmark file to v1.1
##################################################### # Copyright (c) Xuanyi Dong [GitHub D-X-Y], 2019.08 # ##################################################### # [2020.02.25] Initialize the API as v1.1 # [2020.03.09] Upgrade the API to v1.2 # [2020.03.16] Upgrade the API to v1.3 import os from setuptools import setup def read(fname='README.md'): with open(os.path.join(os.path.dirname(__file__), fname), encoding='utf-8') as cfile: return cfile.read() setup( name = "nas_bench_201", version = "1.3", author = "Xuanyi Dong", author_email = "dongxuanyi888@gmail.com", description = "API for NAS-Bench-201 (a benchmark for neural architecture search).", license = "MIT", keywords = "NAS Dataset API DeepLearning", url = "https://github.com/D-X-Y/NAS-Bench-201", packages=['nas_201_api'], long_description=read('README.md'), long_description_content_type='text/markdown', classifiers=[ "Programming Language :: Python", "Topic :: Database", "Topic :: Scientific/Engineering :: Artificial Intelligence", "License :: OSI Approved :: MIT License", ], )
<commit_before>##################################################### # Copyright (c) Xuanyi Dong [GitHub D-X-Y], 2019.08 # ##################################################### # [2020.03.09] Upgrade to v1.2 import os from setuptools import setup def read(fname='README.md'): with open(os.path.join(os.path.dirname(__file__), fname), encoding='utf-8') as cfile: return cfile.read() setup( name = "nas_bench_201", version = "1.2", author = "Xuanyi Dong", author_email = "dongxuanyi888@gmail.com", description = "API for NAS-Bench-201 (a benchmark for neural architecture search).", license = "MIT", keywords = "NAS Dataset API DeepLearning", url = "https://github.com/D-X-Y/NAS-Bench-201", packages=['nas_201_api'], long_description=read('README.md'), long_description_content_type='text/markdown', classifiers=[ "Programming Language :: Python", "Topic :: Database", "Topic :: Scientific/Engineering :: Artificial Intelligence", "License :: OSI Approved :: MIT License", ], ) <commit_msg>Upgrade API to v1.3 and benchmark file to v1.1<commit_after>
##################################################### # Copyright (c) Xuanyi Dong [GitHub D-X-Y], 2019.08 # ##################################################### # [2020.02.25] Initialize the API as v1.1 # [2020.03.09] Upgrade the API to v1.2 # [2020.03.16] Upgrade the API to v1.3 import os from setuptools import setup def read(fname='README.md'): with open(os.path.join(os.path.dirname(__file__), fname), encoding='utf-8') as cfile: return cfile.read() setup( name = "nas_bench_201", version = "1.3", author = "Xuanyi Dong", author_email = "dongxuanyi888@gmail.com", description = "API for NAS-Bench-201 (a benchmark for neural architecture search).", license = "MIT", keywords = "NAS Dataset API DeepLearning", url = "https://github.com/D-X-Y/NAS-Bench-201", packages=['nas_201_api'], long_description=read('README.md'), long_description_content_type='text/markdown', classifiers=[ "Programming Language :: Python", "Topic :: Database", "Topic :: Scientific/Engineering :: Artificial Intelligence", "License :: OSI Approved :: MIT License", ], )
##################################################### # Copyright (c) Xuanyi Dong [GitHub D-X-Y], 2019.08 # ##################################################### # [2020.03.09] Upgrade to v1.2 import os from setuptools import setup def read(fname='README.md'): with open(os.path.join(os.path.dirname(__file__), fname), encoding='utf-8') as cfile: return cfile.read() setup( name = "nas_bench_201", version = "1.2", author = "Xuanyi Dong", author_email = "dongxuanyi888@gmail.com", description = "API for NAS-Bench-201 (a benchmark for neural architecture search).", license = "MIT", keywords = "NAS Dataset API DeepLearning", url = "https://github.com/D-X-Y/NAS-Bench-201", packages=['nas_201_api'], long_description=read('README.md'), long_description_content_type='text/markdown', classifiers=[ "Programming Language :: Python", "Topic :: Database", "Topic :: Scientific/Engineering :: Artificial Intelligence", "License :: OSI Approved :: MIT License", ], ) Upgrade API to v1.3 and benchmark file to v1.1##################################################### # Copyright (c) Xuanyi Dong [GitHub D-X-Y], 2019.08 # ##################################################### # [2020.02.25] Initialize the API as v1.1 # [2020.03.09] Upgrade the API to v1.2 # [2020.03.16] Upgrade the API to v1.3 import os from setuptools import setup def read(fname='README.md'): with open(os.path.join(os.path.dirname(__file__), fname), encoding='utf-8') as cfile: return cfile.read() setup( name = "nas_bench_201", version = "1.3", author = "Xuanyi Dong", author_email = "dongxuanyi888@gmail.com", description = "API for NAS-Bench-201 (a benchmark for neural architecture search).", license = "MIT", keywords = "NAS Dataset API DeepLearning", url = "https://github.com/D-X-Y/NAS-Bench-201", packages=['nas_201_api'], long_description=read('README.md'), long_description_content_type='text/markdown', classifiers=[ "Programming Language :: Python", "Topic :: Database", "Topic :: Scientific/Engineering :: Artificial Intelligence", "License :: OSI Approved :: MIT License", ], )
<commit_before>##################################################### # Copyright (c) Xuanyi Dong [GitHub D-X-Y], 2019.08 # ##################################################### # [2020.03.09] Upgrade to v1.2 import os from setuptools import setup def read(fname='README.md'): with open(os.path.join(os.path.dirname(__file__), fname), encoding='utf-8') as cfile: return cfile.read() setup( name = "nas_bench_201", version = "1.2", author = "Xuanyi Dong", author_email = "dongxuanyi888@gmail.com", description = "API for NAS-Bench-201 (a benchmark for neural architecture search).", license = "MIT", keywords = "NAS Dataset API DeepLearning", url = "https://github.com/D-X-Y/NAS-Bench-201", packages=['nas_201_api'], long_description=read('README.md'), long_description_content_type='text/markdown', classifiers=[ "Programming Language :: Python", "Topic :: Database", "Topic :: Scientific/Engineering :: Artificial Intelligence", "License :: OSI Approved :: MIT License", ], ) <commit_msg>Upgrade API to v1.3 and benchmark file to v1.1<commit_after>##################################################### # Copyright (c) Xuanyi Dong [GitHub D-X-Y], 2019.08 # ##################################################### # [2020.02.25] Initialize the API as v1.1 # [2020.03.09] Upgrade the API to v1.2 # [2020.03.16] Upgrade the API to v1.3 import os from setuptools import setup def read(fname='README.md'): with open(os.path.join(os.path.dirname(__file__), fname), encoding='utf-8') as cfile: return cfile.read() setup( name = "nas_bench_201", version = "1.3", author = "Xuanyi Dong", author_email = "dongxuanyi888@gmail.com", description = "API for NAS-Bench-201 (a benchmark for neural architecture search).", license = "MIT", keywords = "NAS Dataset API DeepLearning", url = "https://github.com/D-X-Y/NAS-Bench-201", packages=['nas_201_api'], long_description=read('README.md'), long_description_content_type='text/markdown', classifiers=[ "Programming Language :: Python", "Topic :: Database", "Topic :: Scientific/Engineering :: Artificial Intelligence", "License :: OSI Approved :: MIT License", ], )
2bcff1f482fe4209dcf1bd53f9b535fefdd82aa3
setup.py
setup.py
from setuptools import setup, find_packages setup(name='googleanalytics', description='A wrapper for the Google Analytics API.', long_description=open('README.rst').read(), author='Stijn Debrouwere', author_email='stijn@debrouwere.org', url='https://github.com/debrouwere/google-analytics/', download_url='http://www.github.com/debrouwere/google-analytics/tarball/master', version='0.8.4', license='ISC', packages=find_packages(), keywords='data analytics api wrapper google', scripts=[ 'bin/googleanalytics' ], install_requires=[ 'oauth2client==1.3', 'google-api-python-client==1.3', 'python-dateutil==1.5', 'addressable>=1', 'inspector>=0.2', 'flask==0.10', 'keyring==4', 'click==3.3', 'pyyaml>=3' ], test_suite='googleanalytics.tests', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Scientific/Engineering :: Information Analysis', ], )
from setuptools import setup, find_packages setup(name='googleanalytics', description='A wrapper for the Google Analytics API.', long_description=open('README.rst').read(), author='Stijn Debrouwere', author_email='stijn@debrouwere.org', url='https://github.com/debrouwere/google-analytics/', download_url='http://www.github.com/debrouwere/google-analytics/tarball/master', version='0.8.5', license='ISC', packages=find_packages(), keywords='data analytics api wrapper google', scripts=[ 'bin/googleanalytics' ], install_requires=[ 'oauth2client==1.3', 'google-api-python-client==1.3', 'python-dateutil==1.5', 'addressable>=1', 'inspect-it>=0.2', 'flask==0.10', 'keyring==4', 'click==3.3', 'pyyaml>=3' ], test_suite='googleanalytics.tests', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Scientific/Engineering :: Information Analysis', ], )
Deal with dependency rename: inspector to inspect-it.
Deal with dependency rename: inspector to inspect-it.
Python
isc
debrouwere/google-analytics
from setuptools import setup, find_packages setup(name='googleanalytics', description='A wrapper for the Google Analytics API.', long_description=open('README.rst').read(), author='Stijn Debrouwere', author_email='stijn@debrouwere.org', url='https://github.com/debrouwere/google-analytics/', download_url='http://www.github.com/debrouwere/google-analytics/tarball/master', version='0.8.4', license='ISC', packages=find_packages(), keywords='data analytics api wrapper google', scripts=[ 'bin/googleanalytics' ], install_requires=[ 'oauth2client==1.3', 'google-api-python-client==1.3', 'python-dateutil==1.5', 'addressable>=1', 'inspector>=0.2', 'flask==0.10', 'keyring==4', 'click==3.3', 'pyyaml>=3' ], test_suite='googleanalytics.tests', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Scientific/Engineering :: Information Analysis', ], )Deal with dependency rename: inspector to inspect-it.
from setuptools import setup, find_packages setup(name='googleanalytics', description='A wrapper for the Google Analytics API.', long_description=open('README.rst').read(), author='Stijn Debrouwere', author_email='stijn@debrouwere.org', url='https://github.com/debrouwere/google-analytics/', download_url='http://www.github.com/debrouwere/google-analytics/tarball/master', version='0.8.5', license='ISC', packages=find_packages(), keywords='data analytics api wrapper google', scripts=[ 'bin/googleanalytics' ], install_requires=[ 'oauth2client==1.3', 'google-api-python-client==1.3', 'python-dateutil==1.5', 'addressable>=1', 'inspect-it>=0.2', 'flask==0.10', 'keyring==4', 'click==3.3', 'pyyaml>=3' ], test_suite='googleanalytics.tests', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Scientific/Engineering :: Information Analysis', ], )
<commit_before>from setuptools import setup, find_packages setup(name='googleanalytics', description='A wrapper for the Google Analytics API.', long_description=open('README.rst').read(), author='Stijn Debrouwere', author_email='stijn@debrouwere.org', url='https://github.com/debrouwere/google-analytics/', download_url='http://www.github.com/debrouwere/google-analytics/tarball/master', version='0.8.4', license='ISC', packages=find_packages(), keywords='data analytics api wrapper google', scripts=[ 'bin/googleanalytics' ], install_requires=[ 'oauth2client==1.3', 'google-api-python-client==1.3', 'python-dateutil==1.5', 'addressable>=1', 'inspector>=0.2', 'flask==0.10', 'keyring==4', 'click==3.3', 'pyyaml>=3' ], test_suite='googleanalytics.tests', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Scientific/Engineering :: Information Analysis', ], )<commit_msg>Deal with dependency rename: inspector to inspect-it.<commit_after>
from setuptools import setup, find_packages setup(name='googleanalytics', description='A wrapper for the Google Analytics API.', long_description=open('README.rst').read(), author='Stijn Debrouwere', author_email='stijn@debrouwere.org', url='https://github.com/debrouwere/google-analytics/', download_url='http://www.github.com/debrouwere/google-analytics/tarball/master', version='0.8.5', license='ISC', packages=find_packages(), keywords='data analytics api wrapper google', scripts=[ 'bin/googleanalytics' ], install_requires=[ 'oauth2client==1.3', 'google-api-python-client==1.3', 'python-dateutil==1.5', 'addressable>=1', 'inspect-it>=0.2', 'flask==0.10', 'keyring==4', 'click==3.3', 'pyyaml>=3' ], test_suite='googleanalytics.tests', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Scientific/Engineering :: Information Analysis', ], )
from setuptools import setup, find_packages setup(name='googleanalytics', description='A wrapper for the Google Analytics API.', long_description=open('README.rst').read(), author='Stijn Debrouwere', author_email='stijn@debrouwere.org', url='https://github.com/debrouwere/google-analytics/', download_url='http://www.github.com/debrouwere/google-analytics/tarball/master', version='0.8.4', license='ISC', packages=find_packages(), keywords='data analytics api wrapper google', scripts=[ 'bin/googleanalytics' ], install_requires=[ 'oauth2client==1.3', 'google-api-python-client==1.3', 'python-dateutil==1.5', 'addressable>=1', 'inspector>=0.2', 'flask==0.10', 'keyring==4', 'click==3.3', 'pyyaml>=3' ], test_suite='googleanalytics.tests', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Scientific/Engineering :: Information Analysis', ], )Deal with dependency rename: inspector to inspect-it.from setuptools import setup, find_packages setup(name='googleanalytics', description='A wrapper for the Google Analytics API.', long_description=open('README.rst').read(), author='Stijn Debrouwere', author_email='stijn@debrouwere.org', url='https://github.com/debrouwere/google-analytics/', download_url='http://www.github.com/debrouwere/google-analytics/tarball/master', version='0.8.5', license='ISC', packages=find_packages(), keywords='data analytics api wrapper google', scripts=[ 'bin/googleanalytics' ], install_requires=[ 'oauth2client==1.3', 'google-api-python-client==1.3', 'python-dateutil==1.5', 'addressable>=1', 'inspect-it>=0.2', 'flask==0.10', 'keyring==4', 'click==3.3', 'pyyaml>=3' ], test_suite='googleanalytics.tests', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Scientific/Engineering :: Information Analysis', ], )
<commit_before>from setuptools import setup, find_packages setup(name='googleanalytics', description='A wrapper for the Google Analytics API.', long_description=open('README.rst').read(), author='Stijn Debrouwere', author_email='stijn@debrouwere.org', url='https://github.com/debrouwere/google-analytics/', download_url='http://www.github.com/debrouwere/google-analytics/tarball/master', version='0.8.4', license='ISC', packages=find_packages(), keywords='data analytics api wrapper google', scripts=[ 'bin/googleanalytics' ], install_requires=[ 'oauth2client==1.3', 'google-api-python-client==1.3', 'python-dateutil==1.5', 'addressable>=1', 'inspector>=0.2', 'flask==0.10', 'keyring==4', 'click==3.3', 'pyyaml>=3' ], test_suite='googleanalytics.tests', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Scientific/Engineering :: Information Analysis', ], )<commit_msg>Deal with dependency rename: inspector to inspect-it.<commit_after>from setuptools import setup, find_packages setup(name='googleanalytics', description='A wrapper for the Google Analytics API.', long_description=open('README.rst').read(), author='Stijn Debrouwere', author_email='stijn@debrouwere.org', url='https://github.com/debrouwere/google-analytics/', download_url='http://www.github.com/debrouwere/google-analytics/tarball/master', version='0.8.5', license='ISC', packages=find_packages(), keywords='data analytics api wrapper google', scripts=[ 'bin/googleanalytics' ], install_requires=[ 'oauth2client==1.3', 'google-api-python-client==1.3', 'python-dateutil==1.5', 'addressable>=1', 'inspect-it>=0.2', 'flask==0.10', 'keyring==4', 'click==3.3', 'pyyaml>=3' ], test_suite='googleanalytics.tests', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Scientific/Engineering :: Information Analysis', ], )
f7459258c0d19de00c042768c68aeeb699c4f04c
setup.py
setup.py
from setuptools import setup, find_packages import pygout readme = open('README.rst', 'r').read() DESCRIPTION = readme.split('\n')[0] LONG_DESCRIPTION = readme INSTALL_REQUIRES = [ 'Pygments == 1.5', ] setup( name = 'PygOut', version = pygout.__version__, url = 'http://github.com/alanbriolat/PygOut', license = 'BSD License', author = 'Alan Briolat, Helen Gray', description = DESCRIPTION, long_descrption = LONG_DESCRIPTION, packages = find_packages(), platforms = 'any', install_requires = INSTALL_REQUIRES, entry_points = { 'console_scripts': ['pygout = pygout.cmdline:main'], }, classifiers = [ 'Development Status :: 2 - Pre-Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: End Users/Desktop', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', ], )
import sys from setuptools import setup, find_packages import pygout readme = open('README.rst', 'r').read() DESCRIPTION = readme.split('\n')[0] LONG_DESCRIPTION = readme INSTALL_REQUIRES = [ 'Pygments == 1.5', ] if sys.version_info < (2, 7): INSTALL_REQUIRES.append('argparse >= 1.1') if sys.version_info < (3, 0): INSTALL_REQUIRES.append('configparser >= 3.0') setup( name = 'PygOut', version = pygout.__version__, url = 'http://github.com/alanbriolat/PygOut', license = 'BSD License', author = 'Alan Briolat, Helen M. Gray', description = DESCRIPTION, long_descrption = LONG_DESCRIPTION, packages = find_packages(), platforms = 'any', install_requires = INSTALL_REQUIRES, entry_points = { 'console_scripts': ['pygout = pygout.cmdline:main'], }, classifiers = [ 'Development Status :: 2 - Pre-Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: End Users/Desktop', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', ], )
Add dependency on argparse and new configparser
Add dependency on argparse and new configparser
Python
bsd-3-clause
alanbriolat/PygOut
from setuptools import setup, find_packages import pygout readme = open('README.rst', 'r').read() DESCRIPTION = readme.split('\n')[0] LONG_DESCRIPTION = readme INSTALL_REQUIRES = [ 'Pygments == 1.5', ] setup( name = 'PygOut', version = pygout.__version__, url = 'http://github.com/alanbriolat/PygOut', license = 'BSD License', author = 'Alan Briolat, Helen Gray', description = DESCRIPTION, long_descrption = LONG_DESCRIPTION, packages = find_packages(), platforms = 'any', install_requires = INSTALL_REQUIRES, entry_points = { 'console_scripts': ['pygout = pygout.cmdline:main'], }, classifiers = [ 'Development Status :: 2 - Pre-Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: End Users/Desktop', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', ], ) Add dependency on argparse and new configparser
import sys from setuptools import setup, find_packages import pygout readme = open('README.rst', 'r').read() DESCRIPTION = readme.split('\n')[0] LONG_DESCRIPTION = readme INSTALL_REQUIRES = [ 'Pygments == 1.5', ] if sys.version_info < (2, 7): INSTALL_REQUIRES.append('argparse >= 1.1') if sys.version_info < (3, 0): INSTALL_REQUIRES.append('configparser >= 3.0') setup( name = 'PygOut', version = pygout.__version__, url = 'http://github.com/alanbriolat/PygOut', license = 'BSD License', author = 'Alan Briolat, Helen M. Gray', description = DESCRIPTION, long_descrption = LONG_DESCRIPTION, packages = find_packages(), platforms = 'any', install_requires = INSTALL_REQUIRES, entry_points = { 'console_scripts': ['pygout = pygout.cmdline:main'], }, classifiers = [ 'Development Status :: 2 - Pre-Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: End Users/Desktop', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', ], )
<commit_before>from setuptools import setup, find_packages import pygout readme = open('README.rst', 'r').read() DESCRIPTION = readme.split('\n')[0] LONG_DESCRIPTION = readme INSTALL_REQUIRES = [ 'Pygments == 1.5', ] setup( name = 'PygOut', version = pygout.__version__, url = 'http://github.com/alanbriolat/PygOut', license = 'BSD License', author = 'Alan Briolat, Helen Gray', description = DESCRIPTION, long_descrption = LONG_DESCRIPTION, packages = find_packages(), platforms = 'any', install_requires = INSTALL_REQUIRES, entry_points = { 'console_scripts': ['pygout = pygout.cmdline:main'], }, classifiers = [ 'Development Status :: 2 - Pre-Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: End Users/Desktop', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', ], ) <commit_msg>Add dependency on argparse and new configparser<commit_after>
import sys from setuptools import setup, find_packages import pygout readme = open('README.rst', 'r').read() DESCRIPTION = readme.split('\n')[0] LONG_DESCRIPTION = readme INSTALL_REQUIRES = [ 'Pygments == 1.5', ] if sys.version_info < (2, 7): INSTALL_REQUIRES.append('argparse >= 1.1') if sys.version_info < (3, 0): INSTALL_REQUIRES.append('configparser >= 3.0') setup( name = 'PygOut', version = pygout.__version__, url = 'http://github.com/alanbriolat/PygOut', license = 'BSD License', author = 'Alan Briolat, Helen M. Gray', description = DESCRIPTION, long_descrption = LONG_DESCRIPTION, packages = find_packages(), platforms = 'any', install_requires = INSTALL_REQUIRES, entry_points = { 'console_scripts': ['pygout = pygout.cmdline:main'], }, classifiers = [ 'Development Status :: 2 - Pre-Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: End Users/Desktop', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', ], )
from setuptools import setup, find_packages import pygout readme = open('README.rst', 'r').read() DESCRIPTION = readme.split('\n')[0] LONG_DESCRIPTION = readme INSTALL_REQUIRES = [ 'Pygments == 1.5', ] setup( name = 'PygOut', version = pygout.__version__, url = 'http://github.com/alanbriolat/PygOut', license = 'BSD License', author = 'Alan Briolat, Helen Gray', description = DESCRIPTION, long_descrption = LONG_DESCRIPTION, packages = find_packages(), platforms = 'any', install_requires = INSTALL_REQUIRES, entry_points = { 'console_scripts': ['pygout = pygout.cmdline:main'], }, classifiers = [ 'Development Status :: 2 - Pre-Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: End Users/Desktop', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', ], ) Add dependency on argparse and new configparserimport sys from setuptools import setup, find_packages import pygout readme = open('README.rst', 'r').read() DESCRIPTION = readme.split('\n')[0] LONG_DESCRIPTION = readme INSTALL_REQUIRES = [ 'Pygments == 1.5', ] if sys.version_info < (2, 7): INSTALL_REQUIRES.append('argparse >= 1.1') if sys.version_info < (3, 0): INSTALL_REQUIRES.append('configparser >= 3.0') setup( name = 'PygOut', version = pygout.__version__, url = 'http://github.com/alanbriolat/PygOut', license = 'BSD License', author = 'Alan Briolat, Helen M. Gray', description = DESCRIPTION, long_descrption = LONG_DESCRIPTION, packages = find_packages(), platforms = 'any', install_requires = INSTALL_REQUIRES, entry_points = { 'console_scripts': ['pygout = pygout.cmdline:main'], }, classifiers = [ 'Development Status :: 2 - Pre-Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: End Users/Desktop', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', ], )
<commit_before>from setuptools import setup, find_packages import pygout readme = open('README.rst', 'r').read() DESCRIPTION = readme.split('\n')[0] LONG_DESCRIPTION = readme INSTALL_REQUIRES = [ 'Pygments == 1.5', ] setup( name = 'PygOut', version = pygout.__version__, url = 'http://github.com/alanbriolat/PygOut', license = 'BSD License', author = 'Alan Briolat, Helen Gray', description = DESCRIPTION, long_descrption = LONG_DESCRIPTION, packages = find_packages(), platforms = 'any', install_requires = INSTALL_REQUIRES, entry_points = { 'console_scripts': ['pygout = pygout.cmdline:main'], }, classifiers = [ 'Development Status :: 2 - Pre-Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: End Users/Desktop', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', ], ) <commit_msg>Add dependency on argparse and new configparser<commit_after>import sys from setuptools import setup, find_packages import pygout readme = open('README.rst', 'r').read() DESCRIPTION = readme.split('\n')[0] LONG_DESCRIPTION = readme INSTALL_REQUIRES = [ 'Pygments == 1.5', ] if sys.version_info < (2, 7): INSTALL_REQUIRES.append('argparse >= 1.1') if sys.version_info < (3, 0): INSTALL_REQUIRES.append('configparser >= 3.0') setup( name = 'PygOut', version = pygout.__version__, url = 'http://github.com/alanbriolat/PygOut', license = 'BSD License', author = 'Alan Briolat, Helen M. Gray', description = DESCRIPTION, long_descrption = LONG_DESCRIPTION, packages = find_packages(), platforms = 'any', install_requires = INSTALL_REQUIRES, entry_points = { 'console_scripts': ['pygout = pygout.cmdline:main'], }, classifiers = [ 'Development Status :: 2 - Pre-Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: End Users/Desktop', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', ], )
98a58ee53e918b0587e1a792c5d08c3c005a7770
setup.py
setup.py
import sys from setuptools import setup meta = dict( name="stacklogger", version="0.1.0", description="A stack-aware logging extension", author="Will Maier", author_email="willmaier@ml1.net", py_modules=["stacklogger"], test_suite="tests.py", install_requires=["setuptools"], keywords="logging stack frame", url="http://packages.python.org/stacklogger", classifiers=[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Programming Language :: Python" "Topic :: System :: Logging", ], ) # Automatic conversion for Python 3 requires distribute. if False and sys.version_info >= (3,): meta.update(dict( use_2to3=True, )) setup(**meta)
import sys from setuptools import setup meta = dict( name="stacklogger", version="0.1.0", description="A stack-aware logging extension", author="Will Maier", author_email="willmaier@ml1.net", py_modules=["stacklogger"], test_suite="tests", install_requires=["setuptools"], keywords="logging stack frame", url="http://packages.python.org/stacklogger", classifiers=[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Programming Language :: Python" "Topic :: System :: Logging", ], ) # Automatic conversion for Python 3 requires distribute. if False and sys.version_info >= (3,): meta.update(dict( use_2to3=True, )) setup(**meta)
Fix test_suite (uses module, not file, name).
Fix test_suite (uses module, not file, name).
Python
isc
whilp/stacklogger
import sys from setuptools import setup meta = dict( name="stacklogger", version="0.1.0", description="A stack-aware logging extension", author="Will Maier", author_email="willmaier@ml1.net", py_modules=["stacklogger"], test_suite="tests.py", install_requires=["setuptools"], keywords="logging stack frame", url="http://packages.python.org/stacklogger", classifiers=[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Programming Language :: Python" "Topic :: System :: Logging", ], ) # Automatic conversion for Python 3 requires distribute. if False and sys.version_info >= (3,): meta.update(dict( use_2to3=True, )) setup(**meta) Fix test_suite (uses module, not file, name).
import sys from setuptools import setup meta = dict( name="stacklogger", version="0.1.0", description="A stack-aware logging extension", author="Will Maier", author_email="willmaier@ml1.net", py_modules=["stacklogger"], test_suite="tests", install_requires=["setuptools"], keywords="logging stack frame", url="http://packages.python.org/stacklogger", classifiers=[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Programming Language :: Python" "Topic :: System :: Logging", ], ) # Automatic conversion for Python 3 requires distribute. if False and sys.version_info >= (3,): meta.update(dict( use_2to3=True, )) setup(**meta)
<commit_before>import sys from setuptools import setup meta = dict( name="stacklogger", version="0.1.0", description="A stack-aware logging extension", author="Will Maier", author_email="willmaier@ml1.net", py_modules=["stacklogger"], test_suite="tests.py", install_requires=["setuptools"], keywords="logging stack frame", url="http://packages.python.org/stacklogger", classifiers=[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Programming Language :: Python" "Topic :: System :: Logging", ], ) # Automatic conversion for Python 3 requires distribute. if False and sys.version_info >= (3,): meta.update(dict( use_2to3=True, )) setup(**meta) <commit_msg>Fix test_suite (uses module, not file, name).<commit_after>
import sys from setuptools import setup meta = dict( name="stacklogger", version="0.1.0", description="A stack-aware logging extension", author="Will Maier", author_email="willmaier@ml1.net", py_modules=["stacklogger"], test_suite="tests", install_requires=["setuptools"], keywords="logging stack frame", url="http://packages.python.org/stacklogger", classifiers=[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Programming Language :: Python" "Topic :: System :: Logging", ], ) # Automatic conversion for Python 3 requires distribute. if False and sys.version_info >= (3,): meta.update(dict( use_2to3=True, )) setup(**meta)
import sys from setuptools import setup meta = dict( name="stacklogger", version="0.1.0", description="A stack-aware logging extension", author="Will Maier", author_email="willmaier@ml1.net", py_modules=["stacklogger"], test_suite="tests.py", install_requires=["setuptools"], keywords="logging stack frame", url="http://packages.python.org/stacklogger", classifiers=[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Programming Language :: Python" "Topic :: System :: Logging", ], ) # Automatic conversion for Python 3 requires distribute. if False and sys.version_info >= (3,): meta.update(dict( use_2to3=True, )) setup(**meta) Fix test_suite (uses module, not file, name).import sys from setuptools import setup meta = dict( name="stacklogger", version="0.1.0", description="A stack-aware logging extension", author="Will Maier", author_email="willmaier@ml1.net", py_modules=["stacklogger"], test_suite="tests", install_requires=["setuptools"], keywords="logging stack frame", url="http://packages.python.org/stacklogger", classifiers=[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Programming Language :: Python" "Topic :: System :: Logging", ], ) # Automatic conversion for Python 3 requires distribute. if False and sys.version_info >= (3,): meta.update(dict( use_2to3=True, )) setup(**meta)
<commit_before>import sys from setuptools import setup meta = dict( name="stacklogger", version="0.1.0", description="A stack-aware logging extension", author="Will Maier", author_email="willmaier@ml1.net", py_modules=["stacklogger"], test_suite="tests.py", install_requires=["setuptools"], keywords="logging stack frame", url="http://packages.python.org/stacklogger", classifiers=[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Programming Language :: Python" "Topic :: System :: Logging", ], ) # Automatic conversion for Python 3 requires distribute. if False and sys.version_info >= (3,): meta.update(dict( use_2to3=True, )) setup(**meta) <commit_msg>Fix test_suite (uses module, not file, name).<commit_after>import sys from setuptools import setup meta = dict( name="stacklogger", version="0.1.0", description="A stack-aware logging extension", author="Will Maier", author_email="willmaier@ml1.net", py_modules=["stacklogger"], test_suite="tests", install_requires=["setuptools"], keywords="logging stack frame", url="http://packages.python.org/stacklogger", classifiers=[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: BSD License", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Programming Language :: Python" "Topic :: System :: Logging", ], ) # Automatic conversion for Python 3 requires distribute. if False and sys.version_info >= (3,): meta.update(dict( use_2to3=True, )) setup(**meta)
dc76cf3a5fcd60223ac903900a56945bb6ee607a
setup.py
setup.py
import sys import os from setuptools import setup long_description = open('README.rst').read() classifiers = [ 'Development Status :: 4 - Beta', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', ] setup_kwargs = dict( name='powershift-cli', version='1.1.0', description='Pluggable command line client for OpenShift.', long_description=long_description, url='https://github.com/getwarped/powershift-cli', author='Graham Dumpleton', author_email='Graham.Dumpleton@gmail.com', license='BSD', classifiers=classifiers, keywords='openshift kubernetes', packages=['powershift', 'powershift.cli'], package_dir={'powershift': 'src/powershift'}, package_data={'powershift.cli': ['completion-bash.sh']}, entry_points = {'console_scripts':['powershift = powershift.cli:main']}, install_requires=['click'], ) setup(**setup_kwargs)
import sys import os from setuptools import setup long_description = open('README.rst').read() classifiers = [ 'Development Status :: 4 - Beta', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', ] setup_kwargs = dict( name='powershift-cli', version='1.1.0', description='Pluggable command line client for OpenShift.', long_description=long_description, url='https://github.com/getwarped/powershift-cli', author='Graham Dumpleton', author_email='Graham.Dumpleton@gmail.com', license='BSD', classifiers=classifiers, keywords='openshift kubernetes', packages=['powershift', 'powershift.cli'], package_dir={'powershift': 'src/powershift'}, package_data={'powershift.cli': ['completion-bash.sh']}, entry_points = {'console_scripts':['powershift = powershift.cli:main']}, install_requires=['click'], extra_requires={'cluster': ['powershift-cluster>=1.1.0']}, ) setup(**setup_kwargs)
Allow pulling in of cluster plugin using powershift-cli[cluster] as install target.
Allow pulling in of cluster plugin using powershift-cli[cluster] as install target.
Python
bsd-2-clause
getwarped/powershift-cli,getwarped/powershift-cli
import sys import os from setuptools import setup long_description = open('README.rst').read() classifiers = [ 'Development Status :: 4 - Beta', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', ] setup_kwargs = dict( name='powershift-cli', version='1.1.0', description='Pluggable command line client for OpenShift.', long_description=long_description, url='https://github.com/getwarped/powershift-cli', author='Graham Dumpleton', author_email='Graham.Dumpleton@gmail.com', license='BSD', classifiers=classifiers, keywords='openshift kubernetes', packages=['powershift', 'powershift.cli'], package_dir={'powershift': 'src/powershift'}, package_data={'powershift.cli': ['completion-bash.sh']}, entry_points = {'console_scripts':['powershift = powershift.cli:main']}, install_requires=['click'], ) setup(**setup_kwargs) Allow pulling in of cluster plugin using powershift-cli[cluster] as install target.
import sys import os from setuptools import setup long_description = open('README.rst').read() classifiers = [ 'Development Status :: 4 - Beta', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', ] setup_kwargs = dict( name='powershift-cli', version='1.1.0', description='Pluggable command line client for OpenShift.', long_description=long_description, url='https://github.com/getwarped/powershift-cli', author='Graham Dumpleton', author_email='Graham.Dumpleton@gmail.com', license='BSD', classifiers=classifiers, keywords='openshift kubernetes', packages=['powershift', 'powershift.cli'], package_dir={'powershift': 'src/powershift'}, package_data={'powershift.cli': ['completion-bash.sh']}, entry_points = {'console_scripts':['powershift = powershift.cli:main']}, install_requires=['click'], extra_requires={'cluster': ['powershift-cluster>=1.1.0']}, ) setup(**setup_kwargs)
<commit_before>import sys import os from setuptools import setup long_description = open('README.rst').read() classifiers = [ 'Development Status :: 4 - Beta', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', ] setup_kwargs = dict( name='powershift-cli', version='1.1.0', description='Pluggable command line client for OpenShift.', long_description=long_description, url='https://github.com/getwarped/powershift-cli', author='Graham Dumpleton', author_email='Graham.Dumpleton@gmail.com', license='BSD', classifiers=classifiers, keywords='openshift kubernetes', packages=['powershift', 'powershift.cli'], package_dir={'powershift': 'src/powershift'}, package_data={'powershift.cli': ['completion-bash.sh']}, entry_points = {'console_scripts':['powershift = powershift.cli:main']}, install_requires=['click'], ) setup(**setup_kwargs) <commit_msg>Allow pulling in of cluster plugin using powershift-cli[cluster] as install target.<commit_after>
import sys import os from setuptools import setup long_description = open('README.rst').read() classifiers = [ 'Development Status :: 4 - Beta', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', ] setup_kwargs = dict( name='powershift-cli', version='1.1.0', description='Pluggable command line client for OpenShift.', long_description=long_description, url='https://github.com/getwarped/powershift-cli', author='Graham Dumpleton', author_email='Graham.Dumpleton@gmail.com', license='BSD', classifiers=classifiers, keywords='openshift kubernetes', packages=['powershift', 'powershift.cli'], package_dir={'powershift': 'src/powershift'}, package_data={'powershift.cli': ['completion-bash.sh']}, entry_points = {'console_scripts':['powershift = powershift.cli:main']}, install_requires=['click'], extra_requires={'cluster': ['powershift-cluster>=1.1.0']}, ) setup(**setup_kwargs)
import sys import os from setuptools import setup long_description = open('README.rst').read() classifiers = [ 'Development Status :: 4 - Beta', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', ] setup_kwargs = dict( name='powershift-cli', version='1.1.0', description='Pluggable command line client for OpenShift.', long_description=long_description, url='https://github.com/getwarped/powershift-cli', author='Graham Dumpleton', author_email='Graham.Dumpleton@gmail.com', license='BSD', classifiers=classifiers, keywords='openshift kubernetes', packages=['powershift', 'powershift.cli'], package_dir={'powershift': 'src/powershift'}, package_data={'powershift.cli': ['completion-bash.sh']}, entry_points = {'console_scripts':['powershift = powershift.cli:main']}, install_requires=['click'], ) setup(**setup_kwargs) Allow pulling in of cluster plugin using powershift-cli[cluster] as install target.import sys import os from setuptools import setup long_description = open('README.rst').read() classifiers = [ 'Development Status :: 4 - Beta', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', ] setup_kwargs = dict( name='powershift-cli', version='1.1.0', description='Pluggable command line client for OpenShift.', long_description=long_description, url='https://github.com/getwarped/powershift-cli', author='Graham Dumpleton', author_email='Graham.Dumpleton@gmail.com', license='BSD', classifiers=classifiers, keywords='openshift kubernetes', packages=['powershift', 'powershift.cli'], package_dir={'powershift': 'src/powershift'}, package_data={'powershift.cli': ['completion-bash.sh']}, entry_points = {'console_scripts':['powershift = powershift.cli:main']}, install_requires=['click'], extra_requires={'cluster': ['powershift-cluster>=1.1.0']}, ) setup(**setup_kwargs)
<commit_before>import sys import os from setuptools import setup long_description = open('README.rst').read() classifiers = [ 'Development Status :: 4 - Beta', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', ] setup_kwargs = dict( name='powershift-cli', version='1.1.0', description='Pluggable command line client for OpenShift.', long_description=long_description, url='https://github.com/getwarped/powershift-cli', author='Graham Dumpleton', author_email='Graham.Dumpleton@gmail.com', license='BSD', classifiers=classifiers, keywords='openshift kubernetes', packages=['powershift', 'powershift.cli'], package_dir={'powershift': 'src/powershift'}, package_data={'powershift.cli': ['completion-bash.sh']}, entry_points = {'console_scripts':['powershift = powershift.cli:main']}, install_requires=['click'], ) setup(**setup_kwargs) <commit_msg>Allow pulling in of cluster plugin using powershift-cli[cluster] as install target.<commit_after>import sys import os from setuptools import setup long_description = open('README.rst').read() classifiers = [ 'Development Status :: 4 - Beta', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', ] setup_kwargs = dict( name='powershift-cli', version='1.1.0', description='Pluggable command line client for OpenShift.', long_description=long_description, url='https://github.com/getwarped/powershift-cli', author='Graham Dumpleton', author_email='Graham.Dumpleton@gmail.com', license='BSD', classifiers=classifiers, keywords='openshift kubernetes', packages=['powershift', 'powershift.cli'], package_dir={'powershift': 'src/powershift'}, package_data={'powershift.cli': ['completion-bash.sh']}, entry_points = {'console_scripts':['powershift = powershift.cli:main']}, install_requires=['click'], extra_requires={'cluster': ['powershift-cluster>=1.1.0']}, ) setup(**setup_kwargs)
b7eccddf76f484da10340f56215c0d07237d34ce
setup.py
setup.py
from distutils.core import setup setup( name='Pykka', version='0.1', author='Stein Magnus Jodal', author_email='stein.magnus@jodal.no', packages=['pykka'], url='http://github.com/jodal/pykka', license='Apache License, Version 2.0', description='Pykka makes actors look like regular objects', long_description=open('README.rst').read(), classifiers=[ 'Development Status :: 1 - Planning', 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache Software License', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries', ], )
from distutils.core import setup setup( name='Pykka', version='0.1', author='Stein Magnus Jodal', author_email='stein.magnus@jodal.no', packages=['pykka'], url='http://github.com/jodal/pykka', license='Apache License, Version 2.0', description='Pykka makes actors look like regular objects', long_description=open('README.rst').read(), classifiers=[ 'Development Status :: 2 - Pre-Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache Software License', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries', ], )
Change dev status to pre-alpha
Change dev status to pre-alpha
Python
apache-2.0
tamland/pykka,jodal/pykka,tempbottle/pykka
from distutils.core import setup setup( name='Pykka', version='0.1', author='Stein Magnus Jodal', author_email='stein.magnus@jodal.no', packages=['pykka'], url='http://github.com/jodal/pykka', license='Apache License, Version 2.0', description='Pykka makes actors look like regular objects', long_description=open('README.rst').read(), classifiers=[ 'Development Status :: 1 - Planning', 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache Software License', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries', ], ) Change dev status to pre-alpha
from distutils.core import setup setup( name='Pykka', version='0.1', author='Stein Magnus Jodal', author_email='stein.magnus@jodal.no', packages=['pykka'], url='http://github.com/jodal/pykka', license='Apache License, Version 2.0', description='Pykka makes actors look like regular objects', long_description=open('README.rst').read(), classifiers=[ 'Development Status :: 2 - Pre-Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache Software License', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries', ], )
<commit_before>from distutils.core import setup setup( name='Pykka', version='0.1', author='Stein Magnus Jodal', author_email='stein.magnus@jodal.no', packages=['pykka'], url='http://github.com/jodal/pykka', license='Apache License, Version 2.0', description='Pykka makes actors look like regular objects', long_description=open('README.rst').read(), classifiers=[ 'Development Status :: 1 - Planning', 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache Software License', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries', ], ) <commit_msg>Change dev status to pre-alpha<commit_after>
from distutils.core import setup setup( name='Pykka', version='0.1', author='Stein Magnus Jodal', author_email='stein.magnus@jodal.no', packages=['pykka'], url='http://github.com/jodal/pykka', license='Apache License, Version 2.0', description='Pykka makes actors look like regular objects', long_description=open('README.rst').read(), classifiers=[ 'Development Status :: 2 - Pre-Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache Software License', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries', ], )
from distutils.core import setup setup( name='Pykka', version='0.1', author='Stein Magnus Jodal', author_email='stein.magnus@jodal.no', packages=['pykka'], url='http://github.com/jodal/pykka', license='Apache License, Version 2.0', description='Pykka makes actors look like regular objects', long_description=open('README.rst').read(), classifiers=[ 'Development Status :: 1 - Planning', 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache Software License', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries', ], ) Change dev status to pre-alphafrom distutils.core import setup setup( name='Pykka', version='0.1', author='Stein Magnus Jodal', author_email='stein.magnus@jodal.no', packages=['pykka'], url='http://github.com/jodal/pykka', license='Apache License, Version 2.0', description='Pykka makes actors look like regular objects', long_description=open('README.rst').read(), classifiers=[ 'Development Status :: 2 - Pre-Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache Software License', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries', ], )
<commit_before>from distutils.core import setup setup( name='Pykka', version='0.1', author='Stein Magnus Jodal', author_email='stein.magnus@jodal.no', packages=['pykka'], url='http://github.com/jodal/pykka', license='Apache License, Version 2.0', description='Pykka makes actors look like regular objects', long_description=open('README.rst').read(), classifiers=[ 'Development Status :: 1 - Planning', 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache Software License', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries', ], ) <commit_msg>Change dev status to pre-alpha<commit_after>from distutils.core import setup setup( name='Pykka', version='0.1', author='Stein Magnus Jodal', author_email='stein.magnus@jodal.no', packages=['pykka'], url='http://github.com/jodal/pykka', license='Apache License, Version 2.0', description='Pykka makes actors look like regular objects', long_description=open('README.rst').read(), classifiers=[ 'Development Status :: 2 - Pre-Alpha', 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache Software License', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries', ], )
83aa3020a585d41ae2d7f2a7759e42389c22bc65
setup.py
setup.py
from setuptools import find_packages, setup setup(name='satnogsclient', packages=find_packages(), version='0.2.4', author='SatNOGS team', author_email='client-dev@satnogs.org', url='https://github.com/satnogs/satnogs-client/', description='SatNOGS Client', install_requires=['APScheduler', 'SQLAlchemy', 'requests', 'validators', 'python-dateutil', 'ephem', 'pytz'], dependency_links=[ 'git+https://github.com/brandon-rhodes/pyephem.git@47d0ba3616ee6c308f2eed319af3901592d00f70#egg=ephem' ], scripts=['satnogsclient/bin/satnogs-poller', 'satnogsclient/bin/satnogs-task'])
from setuptools import find_packages, setup setup(name='satnogsclient', packages=find_packages(), version='0.2.4', author='SatNOGS team', author_email='client-dev@satnogs.org', url='https://github.com/satnogs/satnogs-client/', description='SatNOGS Client', install_requires=['APScheduler', 'SQLAlchemy', 'requests', 'validators', 'python-dateutil', 'ephem', 'pytz'], scripts=['satnogsclient/bin/satnogs-poller', 'satnogsclient/bin/satnogs-task'])
Upgrade pyephem to latest version.
Upgrade pyephem to latest version. * Pyephem v3.7.6.0 now includes this required patch: https://github.com/brandon-rhodes/pyephem/commit/47d0ba3616ee6c308f2eed319af3901592d00f70 * No need to specify a custom dependency in our setup.py.
Python
agpl-3.0
adamkalis/satnogs-client,cshields/satnogs-client,cshields/satnogs-client,adamkalis/satnogs-client
from setuptools import find_packages, setup setup(name='satnogsclient', packages=find_packages(), version='0.2.4', author='SatNOGS team', author_email='client-dev@satnogs.org', url='https://github.com/satnogs/satnogs-client/', description='SatNOGS Client', install_requires=['APScheduler', 'SQLAlchemy', 'requests', 'validators', 'python-dateutil', 'ephem', 'pytz'], dependency_links=[ 'git+https://github.com/brandon-rhodes/pyephem.git@47d0ba3616ee6c308f2eed319af3901592d00f70#egg=ephem' ], scripts=['satnogsclient/bin/satnogs-poller', 'satnogsclient/bin/satnogs-task']) Upgrade pyephem to latest version. * Pyephem v3.7.6.0 now includes this required patch: https://github.com/brandon-rhodes/pyephem/commit/47d0ba3616ee6c308f2eed319af3901592d00f70 * No need to specify a custom dependency in our setup.py.
from setuptools import find_packages, setup setup(name='satnogsclient', packages=find_packages(), version='0.2.4', author='SatNOGS team', author_email='client-dev@satnogs.org', url='https://github.com/satnogs/satnogs-client/', description='SatNOGS Client', install_requires=['APScheduler', 'SQLAlchemy', 'requests', 'validators', 'python-dateutil', 'ephem', 'pytz'], scripts=['satnogsclient/bin/satnogs-poller', 'satnogsclient/bin/satnogs-task'])
<commit_before>from setuptools import find_packages, setup setup(name='satnogsclient', packages=find_packages(), version='0.2.4', author='SatNOGS team', author_email='client-dev@satnogs.org', url='https://github.com/satnogs/satnogs-client/', description='SatNOGS Client', install_requires=['APScheduler', 'SQLAlchemy', 'requests', 'validators', 'python-dateutil', 'ephem', 'pytz'], dependency_links=[ 'git+https://github.com/brandon-rhodes/pyephem.git@47d0ba3616ee6c308f2eed319af3901592d00f70#egg=ephem' ], scripts=['satnogsclient/bin/satnogs-poller', 'satnogsclient/bin/satnogs-task']) <commit_msg>Upgrade pyephem to latest version. * Pyephem v3.7.6.0 now includes this required patch: https://github.com/brandon-rhodes/pyephem/commit/47d0ba3616ee6c308f2eed319af3901592d00f70 * No need to specify a custom dependency in our setup.py.<commit_after>
from setuptools import find_packages, setup setup(name='satnogsclient', packages=find_packages(), version='0.2.4', author='SatNOGS team', author_email='client-dev@satnogs.org', url='https://github.com/satnogs/satnogs-client/', description='SatNOGS Client', install_requires=['APScheduler', 'SQLAlchemy', 'requests', 'validators', 'python-dateutil', 'ephem', 'pytz'], scripts=['satnogsclient/bin/satnogs-poller', 'satnogsclient/bin/satnogs-task'])
from setuptools import find_packages, setup setup(name='satnogsclient', packages=find_packages(), version='0.2.4', author='SatNOGS team', author_email='client-dev@satnogs.org', url='https://github.com/satnogs/satnogs-client/', description='SatNOGS Client', install_requires=['APScheduler', 'SQLAlchemy', 'requests', 'validators', 'python-dateutil', 'ephem', 'pytz'], dependency_links=[ 'git+https://github.com/brandon-rhodes/pyephem.git@47d0ba3616ee6c308f2eed319af3901592d00f70#egg=ephem' ], scripts=['satnogsclient/bin/satnogs-poller', 'satnogsclient/bin/satnogs-task']) Upgrade pyephem to latest version. * Pyephem v3.7.6.0 now includes this required patch: https://github.com/brandon-rhodes/pyephem/commit/47d0ba3616ee6c308f2eed319af3901592d00f70 * No need to specify a custom dependency in our setup.py.from setuptools import find_packages, setup setup(name='satnogsclient', packages=find_packages(), version='0.2.4', author='SatNOGS team', author_email='client-dev@satnogs.org', url='https://github.com/satnogs/satnogs-client/', description='SatNOGS Client', install_requires=['APScheduler', 'SQLAlchemy', 'requests', 'validators', 'python-dateutil', 'ephem', 'pytz'], scripts=['satnogsclient/bin/satnogs-poller', 'satnogsclient/bin/satnogs-task'])
<commit_before>from setuptools import find_packages, setup setup(name='satnogsclient', packages=find_packages(), version='0.2.4', author='SatNOGS team', author_email='client-dev@satnogs.org', url='https://github.com/satnogs/satnogs-client/', description='SatNOGS Client', install_requires=['APScheduler', 'SQLAlchemy', 'requests', 'validators', 'python-dateutil', 'ephem', 'pytz'], dependency_links=[ 'git+https://github.com/brandon-rhodes/pyephem.git@47d0ba3616ee6c308f2eed319af3901592d00f70#egg=ephem' ], scripts=['satnogsclient/bin/satnogs-poller', 'satnogsclient/bin/satnogs-task']) <commit_msg>Upgrade pyephem to latest version. * Pyephem v3.7.6.0 now includes this required patch: https://github.com/brandon-rhodes/pyephem/commit/47d0ba3616ee6c308f2eed319af3901592d00f70 * No need to specify a custom dependency in our setup.py.<commit_after>from setuptools import find_packages, setup setup(name='satnogsclient', packages=find_packages(), version='0.2.4', author='SatNOGS team', author_email='client-dev@satnogs.org', url='https://github.com/satnogs/satnogs-client/', description='SatNOGS Client', install_requires=['APScheduler', 'SQLAlchemy', 'requests', 'validators', 'python-dateutil', 'ephem', 'pytz'], scripts=['satnogsclient/bin/satnogs-poller', 'satnogsclient/bin/satnogs-task'])
b86612bdefb58dd7f1f930cfb756dcd16f77c770
setup.py
setup.py
from setuptools import setup def readme(): with open("README.rst", 'r') as f: return f.read() setup( name="asciigraf", version="0.3.0", packages=["asciigraf"], description="A python library for making ascii-art into network graphs.", classifiers=[ "Development Status :: 3 - Alpha", "License :: OSI Approved :: MIT License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Topic :: Software Development :: Libraries :: Python Modules", ], long_description=readme(), author="Opus One Solutions", author_email="rnd@opusonesolutions.com", url="https://github.com/AnjoMan/asciigraf", keywords=["graph", "network", "testing", "parser"], license="MIT", install_requires=[ 'networkx==1.11', ], extras_require={ "test": ["pytest", "pytest-cov"], }, )
from setuptools import setup def readme(): with open("README.rst", 'r') as f: return f.read() setup( name="asciigraf", version="0.3.0", packages=["asciigraf"], description="A python library for making ascii-art into network graphs.", classifiers=[ "Development Status :: 3 - Alpha", "License :: OSI Approved :: MIT License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Topic :: Software Development :: Libraries :: Python Modules", ], long_description=readme(), author="Opus One Solutions", author_email="rnd@opusonesolutions.com", url="https://github.com/opusonesolutions/asciigraf", keywords=["graph", "network", "testing", "parser"], license="MIT", install_requires=[ 'networkx==1.11', ], extras_require={ "test": ["pytest", "pytest-cov"], }, )
Change package url to point to opusonesolutions group
Change package url to point to opusonesolutions group
Python
mit
AnjoMan/asciigraf
from setuptools import setup def readme(): with open("README.rst", 'r') as f: return f.read() setup( name="asciigraf", version="0.3.0", packages=["asciigraf"], description="A python library for making ascii-art into network graphs.", classifiers=[ "Development Status :: 3 - Alpha", "License :: OSI Approved :: MIT License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Topic :: Software Development :: Libraries :: Python Modules", ], long_description=readme(), author="Opus One Solutions", author_email="rnd@opusonesolutions.com", url="https://github.com/AnjoMan/asciigraf", keywords=["graph", "network", "testing", "parser"], license="MIT", install_requires=[ 'networkx==1.11', ], extras_require={ "test": ["pytest", "pytest-cov"], }, ) Change package url to point to opusonesolutions group
from setuptools import setup def readme(): with open("README.rst", 'r') as f: return f.read() setup( name="asciigraf", version="0.3.0", packages=["asciigraf"], description="A python library for making ascii-art into network graphs.", classifiers=[ "Development Status :: 3 - Alpha", "License :: OSI Approved :: MIT License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Topic :: Software Development :: Libraries :: Python Modules", ], long_description=readme(), author="Opus One Solutions", author_email="rnd@opusonesolutions.com", url="https://github.com/opusonesolutions/asciigraf", keywords=["graph", "network", "testing", "parser"], license="MIT", install_requires=[ 'networkx==1.11', ], extras_require={ "test": ["pytest", "pytest-cov"], }, )
<commit_before>from setuptools import setup def readme(): with open("README.rst", 'r') as f: return f.read() setup( name="asciigraf", version="0.3.0", packages=["asciigraf"], description="A python library for making ascii-art into network graphs.", classifiers=[ "Development Status :: 3 - Alpha", "License :: OSI Approved :: MIT License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Topic :: Software Development :: Libraries :: Python Modules", ], long_description=readme(), author="Opus One Solutions", author_email="rnd@opusonesolutions.com", url="https://github.com/AnjoMan/asciigraf", keywords=["graph", "network", "testing", "parser"], license="MIT", install_requires=[ 'networkx==1.11', ], extras_require={ "test": ["pytest", "pytest-cov"], }, ) <commit_msg>Change package url to point to opusonesolutions group<commit_after>
from setuptools import setup def readme(): with open("README.rst", 'r') as f: return f.read() setup( name="asciigraf", version="0.3.0", packages=["asciigraf"], description="A python library for making ascii-art into network graphs.", classifiers=[ "Development Status :: 3 - Alpha", "License :: OSI Approved :: MIT License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Topic :: Software Development :: Libraries :: Python Modules", ], long_description=readme(), author="Opus One Solutions", author_email="rnd@opusonesolutions.com", url="https://github.com/opusonesolutions/asciigraf", keywords=["graph", "network", "testing", "parser"], license="MIT", install_requires=[ 'networkx==1.11', ], extras_require={ "test": ["pytest", "pytest-cov"], }, )
from setuptools import setup def readme(): with open("README.rst", 'r') as f: return f.read() setup( name="asciigraf", version="0.3.0", packages=["asciigraf"], description="A python library for making ascii-art into network graphs.", classifiers=[ "Development Status :: 3 - Alpha", "License :: OSI Approved :: MIT License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Topic :: Software Development :: Libraries :: Python Modules", ], long_description=readme(), author="Opus One Solutions", author_email="rnd@opusonesolutions.com", url="https://github.com/AnjoMan/asciigraf", keywords=["graph", "network", "testing", "parser"], license="MIT", install_requires=[ 'networkx==1.11', ], extras_require={ "test": ["pytest", "pytest-cov"], }, ) Change package url to point to opusonesolutions groupfrom setuptools import setup def readme(): with open("README.rst", 'r') as f: return f.read() setup( name="asciigraf", version="0.3.0", packages=["asciigraf"], description="A python library for making ascii-art into network graphs.", classifiers=[ "Development Status :: 3 - Alpha", "License :: OSI Approved :: MIT License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Topic :: Software Development :: Libraries :: Python Modules", ], long_description=readme(), author="Opus One Solutions", author_email="rnd@opusonesolutions.com", url="https://github.com/opusonesolutions/asciigraf", keywords=["graph", "network", "testing", "parser"], license="MIT", install_requires=[ 'networkx==1.11', ], extras_require={ "test": ["pytest", "pytest-cov"], }, )
<commit_before>from setuptools import setup def readme(): with open("README.rst", 'r') as f: return f.read() setup( name="asciigraf", version="0.3.0", packages=["asciigraf"], description="A python library for making ascii-art into network graphs.", classifiers=[ "Development Status :: 3 - Alpha", "License :: OSI Approved :: MIT License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Topic :: Software Development :: Libraries :: Python Modules", ], long_description=readme(), author="Opus One Solutions", author_email="rnd@opusonesolutions.com", url="https://github.com/AnjoMan/asciigraf", keywords=["graph", "network", "testing", "parser"], license="MIT", install_requires=[ 'networkx==1.11', ], extras_require={ "test": ["pytest", "pytest-cov"], }, ) <commit_msg>Change package url to point to opusonesolutions group<commit_after>from setuptools import setup def readme(): with open("README.rst", 'r') as f: return f.read() setup( name="asciigraf", version="0.3.0", packages=["asciigraf"], description="A python library for making ascii-art into network graphs.", classifiers=[ "Development Status :: 3 - Alpha", "License :: OSI Approved :: MIT License", "Programming Language :: Python", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Topic :: Software Development :: Libraries :: Python Modules", ], long_description=readme(), author="Opus One Solutions", author_email="rnd@opusonesolutions.com", url="https://github.com/opusonesolutions/asciigraf", keywords=["graph", "network", "testing", "parser"], license="MIT", install_requires=[ 'networkx==1.11', ], extras_require={ "test": ["pytest", "pytest-cov"], }, )
d616642e11c0151f44cdae6038d8cdae07abdf8c
setup.py
setup.py
from distutils.core import setup setup( name='Getty Art', version='0.0.1', author='Clemens Wolff', author_email='clemens.wolff+pypi@gmail.com', packages=['getty_art'], url='https://github.com/c-w/GettyArt', download_url='http://pypi.python.org/pypi/GettyArt', license='LICENSE.txt', description='Scraper for art available from getty.edu', long_description=open('README.rst').read(), )
from distutils.core import setup setup( name='Getty Art', version='0.0.1', author='Clemens Wolff', author_email='clemens.wolff+pypi@gmail.com', packages=['getty_art'], url='https://github.com/c-w/GettyArt', download_url='http://pypi.python.org/pypi/GettyArt', license='LICENSE.txt', description='Scraper for the fine art on getty.edu', long_description=open('README.rst').read(), )
Make tag-line consistent with GitHub
Make tag-line consistent with GitHub
Python
mit
c-w/GettyArt
from distutils.core import setup setup( name='Getty Art', version='0.0.1', author='Clemens Wolff', author_email='clemens.wolff+pypi@gmail.com', packages=['getty_art'], url='https://github.com/c-w/GettyArt', download_url='http://pypi.python.org/pypi/GettyArt', license='LICENSE.txt', description='Scraper for art available from getty.edu', long_description=open('README.rst').read(), ) Make tag-line consistent with GitHub
from distutils.core import setup setup( name='Getty Art', version='0.0.1', author='Clemens Wolff', author_email='clemens.wolff+pypi@gmail.com', packages=['getty_art'], url='https://github.com/c-w/GettyArt', download_url='http://pypi.python.org/pypi/GettyArt', license='LICENSE.txt', description='Scraper for the fine art on getty.edu', long_description=open('README.rst').read(), )
<commit_before>from distutils.core import setup setup( name='Getty Art', version='0.0.1', author='Clemens Wolff', author_email='clemens.wolff+pypi@gmail.com', packages=['getty_art'], url='https://github.com/c-w/GettyArt', download_url='http://pypi.python.org/pypi/GettyArt', license='LICENSE.txt', description='Scraper for art available from getty.edu', long_description=open('README.rst').read(), ) <commit_msg>Make tag-line consistent with GitHub<commit_after>
from distutils.core import setup setup( name='Getty Art', version='0.0.1', author='Clemens Wolff', author_email='clemens.wolff+pypi@gmail.com', packages=['getty_art'], url='https://github.com/c-w/GettyArt', download_url='http://pypi.python.org/pypi/GettyArt', license='LICENSE.txt', description='Scraper for the fine art on getty.edu', long_description=open('README.rst').read(), )
from distutils.core import setup setup( name='Getty Art', version='0.0.1', author='Clemens Wolff', author_email='clemens.wolff+pypi@gmail.com', packages=['getty_art'], url='https://github.com/c-w/GettyArt', download_url='http://pypi.python.org/pypi/GettyArt', license='LICENSE.txt', description='Scraper for art available from getty.edu', long_description=open('README.rst').read(), ) Make tag-line consistent with GitHubfrom distutils.core import setup setup( name='Getty Art', version='0.0.1', author='Clemens Wolff', author_email='clemens.wolff+pypi@gmail.com', packages=['getty_art'], url='https://github.com/c-w/GettyArt', download_url='http://pypi.python.org/pypi/GettyArt', license='LICENSE.txt', description='Scraper for the fine art on getty.edu', long_description=open('README.rst').read(), )
<commit_before>from distutils.core import setup setup( name='Getty Art', version='0.0.1', author='Clemens Wolff', author_email='clemens.wolff+pypi@gmail.com', packages=['getty_art'], url='https://github.com/c-w/GettyArt', download_url='http://pypi.python.org/pypi/GettyArt', license='LICENSE.txt', description='Scraper for art available from getty.edu', long_description=open('README.rst').read(), ) <commit_msg>Make tag-line consistent with GitHub<commit_after>from distutils.core import setup setup( name='Getty Art', version='0.0.1', author='Clemens Wolff', author_email='clemens.wolff+pypi@gmail.com', packages=['getty_art'], url='https://github.com/c-w/GettyArt', download_url='http://pypi.python.org/pypi/GettyArt', license='LICENSE.txt', description='Scraper for the fine art on getty.edu', long_description=open('README.rst').read(), )
ec25b097520930b19a8ae63a1bab23efcc3ba752
setup.py
setup.py
# # Copyright 2017 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import setuptools def readme(): with open('docs/README.md') as f: return f.read() setuptools.setup( name='pstar', description='pstar: better python collections', long_description=readme(), version='0.1.0', url='https://github.com/iansf/pstar', download_url='https://github.com/iansf/pstar/archive/0.1.0.tar.gz', author='Ian Fischer, Google', author_email='iansf@google.com', packages=['pstar'], license='Apache 2.0', install_requires=[], test_suite='nose.collector', tests_require=['nose'], )
# # Copyright 2017 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import setuptools def readme(): return '`pstar` documentation and source code can be found at https://github.com/iansf/pstar.' def version(): return '0.1.1' setuptools.setup( name='pstar', description='pstar: numpy for arbitrary data', long_description=readme(), version=version(), url='https://github.com/iansf/pstar', download_url='https://github.com/iansf/pstar/archive/%s.tar.gz' % version(), author='Ian Fischer, Google', author_email='iansf@google.com', packages=['pstar'], license='Apache 2.0', install_requires=[], test_suite='nose.collector', tests_require=['nose'], )
Fix bug with pip install. Update version to 0.1.1.
Fix bug with pip install. Update version to 0.1.1.
Python
apache-2.0
iansf/pstar
# # Copyright 2017 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import setuptools def readme(): with open('docs/README.md') as f: return f.read() setuptools.setup( name='pstar', description='pstar: better python collections', long_description=readme(), version='0.1.0', url='https://github.com/iansf/pstar', download_url='https://github.com/iansf/pstar/archive/0.1.0.tar.gz', author='Ian Fischer, Google', author_email='iansf@google.com', packages=['pstar'], license='Apache 2.0', install_requires=[], test_suite='nose.collector', tests_require=['nose'], ) Fix bug with pip install. Update version to 0.1.1.
# # Copyright 2017 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import setuptools def readme(): return '`pstar` documentation and source code can be found at https://github.com/iansf/pstar.' def version(): return '0.1.1' setuptools.setup( name='pstar', description='pstar: numpy for arbitrary data', long_description=readme(), version=version(), url='https://github.com/iansf/pstar', download_url='https://github.com/iansf/pstar/archive/%s.tar.gz' % version(), author='Ian Fischer, Google', author_email='iansf@google.com', packages=['pstar'], license='Apache 2.0', install_requires=[], test_suite='nose.collector', tests_require=['nose'], )
<commit_before># # Copyright 2017 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import setuptools def readme(): with open('docs/README.md') as f: return f.read() setuptools.setup( name='pstar', description='pstar: better python collections', long_description=readme(), version='0.1.0', url='https://github.com/iansf/pstar', download_url='https://github.com/iansf/pstar/archive/0.1.0.tar.gz', author='Ian Fischer, Google', author_email='iansf@google.com', packages=['pstar'], license='Apache 2.0', install_requires=[], test_suite='nose.collector', tests_require=['nose'], ) <commit_msg>Fix bug with pip install. Update version to 0.1.1.<commit_after>
# # Copyright 2017 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import setuptools def readme(): return '`pstar` documentation and source code can be found at https://github.com/iansf/pstar.' def version(): return '0.1.1' setuptools.setup( name='pstar', description='pstar: numpy for arbitrary data', long_description=readme(), version=version(), url='https://github.com/iansf/pstar', download_url='https://github.com/iansf/pstar/archive/%s.tar.gz' % version(), author='Ian Fischer, Google', author_email='iansf@google.com', packages=['pstar'], license='Apache 2.0', install_requires=[], test_suite='nose.collector', tests_require=['nose'], )
# # Copyright 2017 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import setuptools def readme(): with open('docs/README.md') as f: return f.read() setuptools.setup( name='pstar', description='pstar: better python collections', long_description=readme(), version='0.1.0', url='https://github.com/iansf/pstar', download_url='https://github.com/iansf/pstar/archive/0.1.0.tar.gz', author='Ian Fischer, Google', author_email='iansf@google.com', packages=['pstar'], license='Apache 2.0', install_requires=[], test_suite='nose.collector', tests_require=['nose'], ) Fix bug with pip install. Update version to 0.1.1.# # Copyright 2017 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import setuptools def readme(): return '`pstar` documentation and source code can be found at https://github.com/iansf/pstar.' def version(): return '0.1.1' setuptools.setup( name='pstar', description='pstar: numpy for arbitrary data', long_description=readme(), version=version(), url='https://github.com/iansf/pstar', download_url='https://github.com/iansf/pstar/archive/%s.tar.gz' % version(), author='Ian Fischer, Google', author_email='iansf@google.com', packages=['pstar'], license='Apache 2.0', install_requires=[], test_suite='nose.collector', tests_require=['nose'], )
<commit_before># # Copyright 2017 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import setuptools def readme(): with open('docs/README.md') as f: return f.read() setuptools.setup( name='pstar', description='pstar: better python collections', long_description=readme(), version='0.1.0', url='https://github.com/iansf/pstar', download_url='https://github.com/iansf/pstar/archive/0.1.0.tar.gz', author='Ian Fischer, Google', author_email='iansf@google.com', packages=['pstar'], license='Apache 2.0', install_requires=[], test_suite='nose.collector', tests_require=['nose'], ) <commit_msg>Fix bug with pip install. Update version to 0.1.1.<commit_after># # Copyright 2017 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import setuptools def readme(): return '`pstar` documentation and source code can be found at https://github.com/iansf/pstar.' def version(): return '0.1.1' setuptools.setup( name='pstar', description='pstar: numpy for arbitrary data', long_description=readme(), version=version(), url='https://github.com/iansf/pstar', download_url='https://github.com/iansf/pstar/archive/%s.tar.gz' % version(), author='Ian Fischer, Google', author_email='iansf@google.com', packages=['pstar'], license='Apache 2.0', install_requires=[], test_suite='nose.collector', tests_require=['nose'], )
63349f4a73c8cb99f495cc15621ecf946e9e652e
setup.py
setup.py
# -*- coding: utf-8 -*- VERSION = '0.2' from setuptools import setup setup( name='nutshell', packages=["nutshell"], version=VERSION, description='A minimal python library to access Nutshell CRM:s JSON-RPC API.', author=u'Emil Stenström', author_email='em@kth.se', url='https://github.com/EmilStenstrom/python-nutshell', download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/' + VERSION, install_requires=["requests>=2.9.1", "six>=1.10.0"], tests_require=["mock>=1.0.1", "nose>=1.3.7"], test_suite="nose.collector", keywords=['nutshell', 'nutshell-crm', 'json-rpc'], classifiers=[], )
# -*- coding: utf-8 -*- VERSION = '0.2' from setuptools import setup setup( name='nutshell', packages=["nutshell"], version=VERSION, description='A minimal python library to access Nutshell CRM:s JSON-RPC API.', author=u'Emil Stenström', author_email='em@kth.se', url='https://github.com/EmilStenstrom/python-nutshell', download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/' + VERSION, install_requires=["requests>=2.9.1", "six>=1.10.0"], tests_require=["mock>=1.0.1", "nose>=1.3.7"], test_suite="nose.collector", keywords=['nutshell', 'nutshell-crm', 'json-rpc'], classifiers=[ "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Operating System :: OS Independent", ])
Add proper Python version classifiers.
Add proper Python version classifiers.
Python
mit
EmilStenstrom/python-nutshell
# -*- coding: utf-8 -*- VERSION = '0.2' from setuptools import setup setup( name='nutshell', packages=["nutshell"], version=VERSION, description='A minimal python library to access Nutshell CRM:s JSON-RPC API.', author=u'Emil Stenström', author_email='em@kth.se', url='https://github.com/EmilStenstrom/python-nutshell', download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/' + VERSION, install_requires=["requests>=2.9.1", "six>=1.10.0"], tests_require=["mock>=1.0.1", "nose>=1.3.7"], test_suite="nose.collector", keywords=['nutshell', 'nutshell-crm', 'json-rpc'], classifiers=[], ) Add proper Python version classifiers.
# -*- coding: utf-8 -*- VERSION = '0.2' from setuptools import setup setup( name='nutshell', packages=["nutshell"], version=VERSION, description='A minimal python library to access Nutshell CRM:s JSON-RPC API.', author=u'Emil Stenström', author_email='em@kth.se', url='https://github.com/EmilStenstrom/python-nutshell', download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/' + VERSION, install_requires=["requests>=2.9.1", "six>=1.10.0"], tests_require=["mock>=1.0.1", "nose>=1.3.7"], test_suite="nose.collector", keywords=['nutshell', 'nutshell-crm', 'json-rpc'], classifiers=[ "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Operating System :: OS Independent", ])
<commit_before># -*- coding: utf-8 -*- VERSION = '0.2' from setuptools import setup setup( name='nutshell', packages=["nutshell"], version=VERSION, description='A minimal python library to access Nutshell CRM:s JSON-RPC API.', author=u'Emil Stenström', author_email='em@kth.se', url='https://github.com/EmilStenstrom/python-nutshell', download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/' + VERSION, install_requires=["requests>=2.9.1", "six>=1.10.0"], tests_require=["mock>=1.0.1", "nose>=1.3.7"], test_suite="nose.collector", keywords=['nutshell', 'nutshell-crm', 'json-rpc'], classifiers=[], ) <commit_msg>Add proper Python version classifiers.<commit_after>
# -*- coding: utf-8 -*- VERSION = '0.2' from setuptools import setup setup( name='nutshell', packages=["nutshell"], version=VERSION, description='A minimal python library to access Nutshell CRM:s JSON-RPC API.', author=u'Emil Stenström', author_email='em@kth.se', url='https://github.com/EmilStenstrom/python-nutshell', download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/' + VERSION, install_requires=["requests>=2.9.1", "six>=1.10.0"], tests_require=["mock>=1.0.1", "nose>=1.3.7"], test_suite="nose.collector", keywords=['nutshell', 'nutshell-crm', 'json-rpc'], classifiers=[ "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Operating System :: OS Independent", ])
# -*- coding: utf-8 -*- VERSION = '0.2' from setuptools import setup setup( name='nutshell', packages=["nutshell"], version=VERSION, description='A minimal python library to access Nutshell CRM:s JSON-RPC API.', author=u'Emil Stenström', author_email='em@kth.se', url='https://github.com/EmilStenstrom/python-nutshell', download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/' + VERSION, install_requires=["requests>=2.9.1", "six>=1.10.0"], tests_require=["mock>=1.0.1", "nose>=1.3.7"], test_suite="nose.collector", keywords=['nutshell', 'nutshell-crm', 'json-rpc'], classifiers=[], ) Add proper Python version classifiers.# -*- coding: utf-8 -*- VERSION = '0.2' from setuptools import setup setup( name='nutshell', packages=["nutshell"], version=VERSION, description='A minimal python library to access Nutshell CRM:s JSON-RPC API.', author=u'Emil Stenström', author_email='em@kth.se', url='https://github.com/EmilStenstrom/python-nutshell', download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/' + VERSION, install_requires=["requests>=2.9.1", "six>=1.10.0"], tests_require=["mock>=1.0.1", "nose>=1.3.7"], test_suite="nose.collector", keywords=['nutshell', 'nutshell-crm', 'json-rpc'], classifiers=[ "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Operating System :: OS Independent", ])
<commit_before># -*- coding: utf-8 -*- VERSION = '0.2' from setuptools import setup setup( name='nutshell', packages=["nutshell"], version=VERSION, description='A minimal python library to access Nutshell CRM:s JSON-RPC API.', author=u'Emil Stenström', author_email='em@kth.se', url='https://github.com/EmilStenstrom/python-nutshell', download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/' + VERSION, install_requires=["requests>=2.9.1", "six>=1.10.0"], tests_require=["mock>=1.0.1", "nose>=1.3.7"], test_suite="nose.collector", keywords=['nutshell', 'nutshell-crm', 'json-rpc'], classifiers=[], ) <commit_msg>Add proper Python version classifiers.<commit_after># -*- coding: utf-8 -*- VERSION = '0.2' from setuptools import setup setup( name='nutshell', packages=["nutshell"], version=VERSION, description='A minimal python library to access Nutshell CRM:s JSON-RPC API.', author=u'Emil Stenström', author_email='em@kth.se', url='https://github.com/EmilStenstrom/python-nutshell', download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/' + VERSION, install_requires=["requests>=2.9.1", "six>=1.10.0"], tests_require=["mock>=1.0.1", "nose>=1.3.7"], test_suite="nose.collector", keywords=['nutshell', 'nutshell-crm', 'json-rpc'], classifiers=[ "Programming Language :: Python", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Operating System :: OS Independent", ])
d363631106c95365e138bea08ca0f8811c1dba67
setup.py
setup.py
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-glitter-news', version='0.1', description='Django Glitter News for Django', long_description=open('README.rst').read(), url='https://github.com/blancltd/django-glitter-news', maintainer='Blanc Ltd', maintainer_email='studio@blanc.ltd.uk', platforms=['any'], packages=find_packages(), install_requires=[ 'django-glitter', ], classifiers=[ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', ], license='BSD 3-Clause', )
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-glitter-news', version='0.1', description='Django Glitter News for Django', long_description=open('README.rst').read(), url='https://github.com/blancltd/django-glitter-news', maintainer='Blanc Ltd', maintainer_email='studio@blanc.ltd.uk', platforms=['any'], packages=find_packages(), install_requires=[ 'django-glitter', ], classifiers=[ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], license='BSD 3-Clause', )
Update python versions we care about
Update python versions we care about
Python
bsd-2-clause
blancltd/glitter-news
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-glitter-news', version='0.1', description='Django Glitter News for Django', long_description=open('README.rst').read(), url='https://github.com/blancltd/django-glitter-news', maintainer='Blanc Ltd', maintainer_email='studio@blanc.ltd.uk', platforms=['any'], packages=find_packages(), install_requires=[ 'django-glitter', ], classifiers=[ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', ], license='BSD 3-Clause', ) Update python versions we care about
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-glitter-news', version='0.1', description='Django Glitter News for Django', long_description=open('README.rst').read(), url='https://github.com/blancltd/django-glitter-news', maintainer='Blanc Ltd', maintainer_email='studio@blanc.ltd.uk', platforms=['any'], packages=find_packages(), install_requires=[ 'django-glitter', ], classifiers=[ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], license='BSD 3-Clause', )
<commit_before>#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-glitter-news', version='0.1', description='Django Glitter News for Django', long_description=open('README.rst').read(), url='https://github.com/blancltd/django-glitter-news', maintainer='Blanc Ltd', maintainer_email='studio@blanc.ltd.uk', platforms=['any'], packages=find_packages(), install_requires=[ 'django-glitter', ], classifiers=[ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', ], license='BSD 3-Clause', ) <commit_msg>Update python versions we care about<commit_after>
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-glitter-news', version='0.1', description='Django Glitter News for Django', long_description=open('README.rst').read(), url='https://github.com/blancltd/django-glitter-news', maintainer='Blanc Ltd', maintainer_email='studio@blanc.ltd.uk', platforms=['any'], packages=find_packages(), install_requires=[ 'django-glitter', ], classifiers=[ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], license='BSD 3-Clause', )
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-glitter-news', version='0.1', description='Django Glitter News for Django', long_description=open('README.rst').read(), url='https://github.com/blancltd/django-glitter-news', maintainer='Blanc Ltd', maintainer_email='studio@blanc.ltd.uk', platforms=['any'], packages=find_packages(), install_requires=[ 'django-glitter', ], classifiers=[ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', ], license='BSD 3-Clause', ) Update python versions we care about#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-glitter-news', version='0.1', description='Django Glitter News for Django', long_description=open('README.rst').read(), url='https://github.com/blancltd/django-glitter-news', maintainer='Blanc Ltd', maintainer_email='studio@blanc.ltd.uk', platforms=['any'], packages=find_packages(), install_requires=[ 'django-glitter', ], classifiers=[ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], license='BSD 3-Clause', )
<commit_before>#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-glitter-news', version='0.1', description='Django Glitter News for Django', long_description=open('README.rst').read(), url='https://github.com/blancltd/django-glitter-news', maintainer='Blanc Ltd', maintainer_email='studio@blanc.ltd.uk', platforms=['any'], packages=find_packages(), install_requires=[ 'django-glitter', ], classifiers=[ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.2', 'Programming Language :: Python :: 3.3', ], license='BSD 3-Clause', ) <commit_msg>Update python versions we care about<commit_after>#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-glitter-news', version='0.1', description='Django Glitter News for Django', long_description=open('README.rst').read(), url='https://github.com/blancltd/django-glitter-news', maintainer='Blanc Ltd', maintainer_email='studio@blanc.ltd.uk', platforms=['any'], packages=find_packages(), install_requires=[ 'django-glitter', ], classifiers=[ 'Environment :: Web Environment', 'Framework :: Django', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], license='BSD 3-Clause', )
34e04331fae60e63d0bc0daf47161bc8507835b8
setup.py
setup.py
#!/usr/bin/env python import os import sys from serfclient import __version__ try: from setuptools import setup from setuptools.command.test import test as TestCommand class PyTest(TestCommand): def finalize_options(self): TestCommand.finalize_options(self) self.test_args = [] self.test_suite = True def run_tests(self): import pytest errno = pytest.main(self.test_args) sys.exit(errno) except ImportError: from distutils.core import setup PyTest = lambda x: x try: long_description = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read() except: long_description = None setup( name='serfclient', version=__version__, description='Python client for the Serf orchestration tool', long_description=long_description, url='https://github.com/KushalP/serfclient-py', author='Kushal Pisavadia', author_email='kushal@violentlymild.com', maintainer='Kushal Pisavadia', maintainer_email='kushal@violentlymild.com', keywords=['Serf', 'orchestration', 'service discovery'], license='MIT', packages=['serfclient'], install_requires=['msgpack-python >= 0.4.0'], tests_require=['pytest >= 2.5.2', 'pytest-cov >= 1.6'], cmdclass={'test': PyTest} )
#!/usr/bin/env python import os import sys from serfclient import __version__ try: from setuptools import setup from setuptools.command.test import test as TestCommand class PyTest(TestCommand): def finalize_options(self): TestCommand.finalize_options(self) self.test_args = [] self.test_suite = True def run_tests(self): import pytest errno = pytest.main(self.test_args) sys.exit(errno) except ImportError: from distutils.core import setup PyTest = lambda x: x try: long_description = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read() except: long_description = None setup( name='serfclient', version=__version__, description='Python client for the Serf orchestration tool', long_description=long_description, url='https://github.com/KushalP/serfclient-py', author='Kushal Pisavadia', author_email='kushal@violentlymild.com', maintainer='Kushal Pisavadia', maintainer_email='kushal@violentlymild.com', keywords=['Serf', 'orchestration', 'service discovery'], license='MIT', packages=['serfclient'], install_requires=['msgpack-python >= 0.4.0'], tests_require=['pytest >= 2.5.2', 'pytest-cov >= 1.6', 'python-coveralls >= 2.4.2'], cmdclass={'test': PyTest} )
Add python-coveralls as a test dependency
Add python-coveralls as a test dependency This is so that we can push our coverage stats to coverage.io.
Python
mit
charleswhchan/serfclient-py,KushalP/serfclient-py
#!/usr/bin/env python import os import sys from serfclient import __version__ try: from setuptools import setup from setuptools.command.test import test as TestCommand class PyTest(TestCommand): def finalize_options(self): TestCommand.finalize_options(self) self.test_args = [] self.test_suite = True def run_tests(self): import pytest errno = pytest.main(self.test_args) sys.exit(errno) except ImportError: from distutils.core import setup PyTest = lambda x: x try: long_description = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read() except: long_description = None setup( name='serfclient', version=__version__, description='Python client for the Serf orchestration tool', long_description=long_description, url='https://github.com/KushalP/serfclient-py', author='Kushal Pisavadia', author_email='kushal@violentlymild.com', maintainer='Kushal Pisavadia', maintainer_email='kushal@violentlymild.com', keywords=['Serf', 'orchestration', 'service discovery'], license='MIT', packages=['serfclient'], install_requires=['msgpack-python >= 0.4.0'], tests_require=['pytest >= 2.5.2', 'pytest-cov >= 1.6'], cmdclass={'test': PyTest} ) Add python-coveralls as a test dependency This is so that we can push our coverage stats to coverage.io.
#!/usr/bin/env python import os import sys from serfclient import __version__ try: from setuptools import setup from setuptools.command.test import test as TestCommand class PyTest(TestCommand): def finalize_options(self): TestCommand.finalize_options(self) self.test_args = [] self.test_suite = True def run_tests(self): import pytest errno = pytest.main(self.test_args) sys.exit(errno) except ImportError: from distutils.core import setup PyTest = lambda x: x try: long_description = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read() except: long_description = None setup( name='serfclient', version=__version__, description='Python client for the Serf orchestration tool', long_description=long_description, url='https://github.com/KushalP/serfclient-py', author='Kushal Pisavadia', author_email='kushal@violentlymild.com', maintainer='Kushal Pisavadia', maintainer_email='kushal@violentlymild.com', keywords=['Serf', 'orchestration', 'service discovery'], license='MIT', packages=['serfclient'], install_requires=['msgpack-python >= 0.4.0'], tests_require=['pytest >= 2.5.2', 'pytest-cov >= 1.6', 'python-coveralls >= 2.4.2'], cmdclass={'test': PyTest} )
<commit_before>#!/usr/bin/env python import os import sys from serfclient import __version__ try: from setuptools import setup from setuptools.command.test import test as TestCommand class PyTest(TestCommand): def finalize_options(self): TestCommand.finalize_options(self) self.test_args = [] self.test_suite = True def run_tests(self): import pytest errno = pytest.main(self.test_args) sys.exit(errno) except ImportError: from distutils.core import setup PyTest = lambda x: x try: long_description = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read() except: long_description = None setup( name='serfclient', version=__version__, description='Python client for the Serf orchestration tool', long_description=long_description, url='https://github.com/KushalP/serfclient-py', author='Kushal Pisavadia', author_email='kushal@violentlymild.com', maintainer='Kushal Pisavadia', maintainer_email='kushal@violentlymild.com', keywords=['Serf', 'orchestration', 'service discovery'], license='MIT', packages=['serfclient'], install_requires=['msgpack-python >= 0.4.0'], tests_require=['pytest >= 2.5.2', 'pytest-cov >= 1.6'], cmdclass={'test': PyTest} ) <commit_msg>Add python-coveralls as a test dependency This is so that we can push our coverage stats to coverage.io.<commit_after>
#!/usr/bin/env python import os import sys from serfclient import __version__ try: from setuptools import setup from setuptools.command.test import test as TestCommand class PyTest(TestCommand): def finalize_options(self): TestCommand.finalize_options(self) self.test_args = [] self.test_suite = True def run_tests(self): import pytest errno = pytest.main(self.test_args) sys.exit(errno) except ImportError: from distutils.core import setup PyTest = lambda x: x try: long_description = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read() except: long_description = None setup( name='serfclient', version=__version__, description='Python client for the Serf orchestration tool', long_description=long_description, url='https://github.com/KushalP/serfclient-py', author='Kushal Pisavadia', author_email='kushal@violentlymild.com', maintainer='Kushal Pisavadia', maintainer_email='kushal@violentlymild.com', keywords=['Serf', 'orchestration', 'service discovery'], license='MIT', packages=['serfclient'], install_requires=['msgpack-python >= 0.4.0'], tests_require=['pytest >= 2.5.2', 'pytest-cov >= 1.6', 'python-coveralls >= 2.4.2'], cmdclass={'test': PyTest} )
#!/usr/bin/env python import os import sys from serfclient import __version__ try: from setuptools import setup from setuptools.command.test import test as TestCommand class PyTest(TestCommand): def finalize_options(self): TestCommand.finalize_options(self) self.test_args = [] self.test_suite = True def run_tests(self): import pytest errno = pytest.main(self.test_args) sys.exit(errno) except ImportError: from distutils.core import setup PyTest = lambda x: x try: long_description = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read() except: long_description = None setup( name='serfclient', version=__version__, description='Python client for the Serf orchestration tool', long_description=long_description, url='https://github.com/KushalP/serfclient-py', author='Kushal Pisavadia', author_email='kushal@violentlymild.com', maintainer='Kushal Pisavadia', maintainer_email='kushal@violentlymild.com', keywords=['Serf', 'orchestration', 'service discovery'], license='MIT', packages=['serfclient'], install_requires=['msgpack-python >= 0.4.0'], tests_require=['pytest >= 2.5.2', 'pytest-cov >= 1.6'], cmdclass={'test': PyTest} ) Add python-coveralls as a test dependency This is so that we can push our coverage stats to coverage.io.#!/usr/bin/env python import os import sys from serfclient import __version__ try: from setuptools import setup from setuptools.command.test import test as TestCommand class PyTest(TestCommand): def finalize_options(self): TestCommand.finalize_options(self) self.test_args = [] self.test_suite = True def run_tests(self): import pytest errno = pytest.main(self.test_args) sys.exit(errno) except ImportError: from distutils.core import setup PyTest = lambda x: x try: long_description = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read() except: long_description = None setup( name='serfclient', version=__version__, description='Python client for the Serf orchestration tool', long_description=long_description, url='https://github.com/KushalP/serfclient-py', author='Kushal Pisavadia', author_email='kushal@violentlymild.com', maintainer='Kushal Pisavadia', maintainer_email='kushal@violentlymild.com', keywords=['Serf', 'orchestration', 'service discovery'], license='MIT', packages=['serfclient'], install_requires=['msgpack-python >= 0.4.0'], tests_require=['pytest >= 2.5.2', 'pytest-cov >= 1.6', 'python-coveralls >= 2.4.2'], cmdclass={'test': PyTest} )
<commit_before>#!/usr/bin/env python import os import sys from serfclient import __version__ try: from setuptools import setup from setuptools.command.test import test as TestCommand class PyTest(TestCommand): def finalize_options(self): TestCommand.finalize_options(self) self.test_args = [] self.test_suite = True def run_tests(self): import pytest errno = pytest.main(self.test_args) sys.exit(errno) except ImportError: from distutils.core import setup PyTest = lambda x: x try: long_description = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read() except: long_description = None setup( name='serfclient', version=__version__, description='Python client for the Serf orchestration tool', long_description=long_description, url='https://github.com/KushalP/serfclient-py', author='Kushal Pisavadia', author_email='kushal@violentlymild.com', maintainer='Kushal Pisavadia', maintainer_email='kushal@violentlymild.com', keywords=['Serf', 'orchestration', 'service discovery'], license='MIT', packages=['serfclient'], install_requires=['msgpack-python >= 0.4.0'], tests_require=['pytest >= 2.5.2', 'pytest-cov >= 1.6'], cmdclass={'test': PyTest} ) <commit_msg>Add python-coveralls as a test dependency This is so that we can push our coverage stats to coverage.io.<commit_after>#!/usr/bin/env python import os import sys from serfclient import __version__ try: from setuptools import setup from setuptools.command.test import test as TestCommand class PyTest(TestCommand): def finalize_options(self): TestCommand.finalize_options(self) self.test_args = [] self.test_suite = True def run_tests(self): import pytest errno = pytest.main(self.test_args) sys.exit(errno) except ImportError: from distutils.core import setup PyTest = lambda x: x try: long_description = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read() except: long_description = None setup( name='serfclient', version=__version__, description='Python client for the Serf orchestration tool', long_description=long_description, url='https://github.com/KushalP/serfclient-py', author='Kushal Pisavadia', author_email='kushal@violentlymild.com', maintainer='Kushal Pisavadia', maintainer_email='kushal@violentlymild.com', keywords=['Serf', 'orchestration', 'service discovery'], license='MIT', packages=['serfclient'], install_requires=['msgpack-python >= 0.4.0'], tests_require=['pytest >= 2.5.2', 'pytest-cov >= 1.6', 'python-coveralls >= 2.4.2'], cmdclass={'test': PyTest} )
302086ac111798bb3a5a977ed443ab274ee28dec
setup.py
setup.py
import os import sys import setuptools from keystoneclient.openstack.common import setup def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() requires = setup.parse_requirements() depend_links = setup.parse_dependency_links() tests_require = setup.parse_requirements(['tools/test-requires']) setuptools.setup( name="python-keystoneclient", version=setup.get_post_version('keystoneclient'), description="Client library for OpenStack Keystone API", long_description=read('README.rst'), url='https://github.com/openstack/python-keystoneclient', license='Apache', author='Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss', author_email='gabriel.hurley@nebula.com', packages=setuptools.find_packages(exclude=['tests', 'tests.*']), classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], install_requires=requires, dependency_links=depend_links, cmdclass=setup.get_cmdclass(), tests_require=tests_require, test_suite="nose.collector", entry_points={ 'console_scripts': ['keystone = keystoneclient.shell:main'] } )
import os import sys import setuptools from keystoneclient.openstack.common import setup def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() requires = setup.parse_requirements() depend_links = setup.parse_dependency_links() tests_require = setup.parse_requirements(['tools/test-requires']) setuptools.setup( name="python-keystoneclient", version=setup.get_post_version('keystoneclient'), description="Client library for OpenStack Identity API (Keystone)", long_description=read('README.rst'), url='https://github.com/openstack/python-keystoneclient', license='Apache', author='Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss', author_email='gabriel.hurley@nebula.com', packages=setuptools.find_packages(exclude=['tests', 'tests.*']), classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], install_requires=requires, dependency_links=depend_links, cmdclass=setup.get_cmdclass(), tests_require=tests_require, test_suite="nose.collector", entry_points={ 'console_scripts': ['keystone = keystoneclient.shell:main'] } )
Replace refs to 'Keystone API' with 'Identity API'
Replace refs to 'Keystone API' with 'Identity API' Formally, OpenStack Keystone implements the OpenStack Identity API, and this is a client to the API, not to Keystone itself. Change-Id: If568866221a29ba041f0f2cd56dc81deeb9ebc00
Python
apache-2.0
sileht/keystoneauth,jamielennox/keystoneauth,citrix-openstack-build/keystoneauth
import os import sys import setuptools from keystoneclient.openstack.common import setup def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() requires = setup.parse_requirements() depend_links = setup.parse_dependency_links() tests_require = setup.parse_requirements(['tools/test-requires']) setuptools.setup( name="python-keystoneclient", version=setup.get_post_version('keystoneclient'), description="Client library for OpenStack Keystone API", long_description=read('README.rst'), url='https://github.com/openstack/python-keystoneclient', license='Apache', author='Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss', author_email='gabriel.hurley@nebula.com', packages=setuptools.find_packages(exclude=['tests', 'tests.*']), classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], install_requires=requires, dependency_links=depend_links, cmdclass=setup.get_cmdclass(), tests_require=tests_require, test_suite="nose.collector", entry_points={ 'console_scripts': ['keystone = keystoneclient.shell:main'] } ) Replace refs to 'Keystone API' with 'Identity API' Formally, OpenStack Keystone implements the OpenStack Identity API, and this is a client to the API, not to Keystone itself. Change-Id: If568866221a29ba041f0f2cd56dc81deeb9ebc00
import os import sys import setuptools from keystoneclient.openstack.common import setup def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() requires = setup.parse_requirements() depend_links = setup.parse_dependency_links() tests_require = setup.parse_requirements(['tools/test-requires']) setuptools.setup( name="python-keystoneclient", version=setup.get_post_version('keystoneclient'), description="Client library for OpenStack Identity API (Keystone)", long_description=read('README.rst'), url='https://github.com/openstack/python-keystoneclient', license='Apache', author='Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss', author_email='gabriel.hurley@nebula.com', packages=setuptools.find_packages(exclude=['tests', 'tests.*']), classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], install_requires=requires, dependency_links=depend_links, cmdclass=setup.get_cmdclass(), tests_require=tests_require, test_suite="nose.collector", entry_points={ 'console_scripts': ['keystone = keystoneclient.shell:main'] } )
<commit_before>import os import sys import setuptools from keystoneclient.openstack.common import setup def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() requires = setup.parse_requirements() depend_links = setup.parse_dependency_links() tests_require = setup.parse_requirements(['tools/test-requires']) setuptools.setup( name="python-keystoneclient", version=setup.get_post_version('keystoneclient'), description="Client library for OpenStack Keystone API", long_description=read('README.rst'), url='https://github.com/openstack/python-keystoneclient', license='Apache', author='Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss', author_email='gabriel.hurley@nebula.com', packages=setuptools.find_packages(exclude=['tests', 'tests.*']), classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], install_requires=requires, dependency_links=depend_links, cmdclass=setup.get_cmdclass(), tests_require=tests_require, test_suite="nose.collector", entry_points={ 'console_scripts': ['keystone = keystoneclient.shell:main'] } ) <commit_msg>Replace refs to 'Keystone API' with 'Identity API' Formally, OpenStack Keystone implements the OpenStack Identity API, and this is a client to the API, not to Keystone itself. Change-Id: If568866221a29ba041f0f2cd56dc81deeb9ebc00<commit_after>
import os import sys import setuptools from keystoneclient.openstack.common import setup def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() requires = setup.parse_requirements() depend_links = setup.parse_dependency_links() tests_require = setup.parse_requirements(['tools/test-requires']) setuptools.setup( name="python-keystoneclient", version=setup.get_post_version('keystoneclient'), description="Client library for OpenStack Identity API (Keystone)", long_description=read('README.rst'), url='https://github.com/openstack/python-keystoneclient', license='Apache', author='Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss', author_email='gabriel.hurley@nebula.com', packages=setuptools.find_packages(exclude=['tests', 'tests.*']), classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], install_requires=requires, dependency_links=depend_links, cmdclass=setup.get_cmdclass(), tests_require=tests_require, test_suite="nose.collector", entry_points={ 'console_scripts': ['keystone = keystoneclient.shell:main'] } )
import os import sys import setuptools from keystoneclient.openstack.common import setup def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() requires = setup.parse_requirements() depend_links = setup.parse_dependency_links() tests_require = setup.parse_requirements(['tools/test-requires']) setuptools.setup( name="python-keystoneclient", version=setup.get_post_version('keystoneclient'), description="Client library for OpenStack Keystone API", long_description=read('README.rst'), url='https://github.com/openstack/python-keystoneclient', license='Apache', author='Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss', author_email='gabriel.hurley@nebula.com', packages=setuptools.find_packages(exclude=['tests', 'tests.*']), classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], install_requires=requires, dependency_links=depend_links, cmdclass=setup.get_cmdclass(), tests_require=tests_require, test_suite="nose.collector", entry_points={ 'console_scripts': ['keystone = keystoneclient.shell:main'] } ) Replace refs to 'Keystone API' with 'Identity API' Formally, OpenStack Keystone implements the OpenStack Identity API, and this is a client to the API, not to Keystone itself. Change-Id: If568866221a29ba041f0f2cd56dc81deeb9ebc00import os import sys import setuptools from keystoneclient.openstack.common import setup def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() requires = setup.parse_requirements() depend_links = setup.parse_dependency_links() tests_require = setup.parse_requirements(['tools/test-requires']) setuptools.setup( name="python-keystoneclient", version=setup.get_post_version('keystoneclient'), description="Client library for OpenStack Identity API (Keystone)", long_description=read('README.rst'), url='https://github.com/openstack/python-keystoneclient', license='Apache', author='Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss', author_email='gabriel.hurley@nebula.com', packages=setuptools.find_packages(exclude=['tests', 'tests.*']), classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], install_requires=requires, dependency_links=depend_links, cmdclass=setup.get_cmdclass(), tests_require=tests_require, test_suite="nose.collector", entry_points={ 'console_scripts': ['keystone = keystoneclient.shell:main'] } )
<commit_before>import os import sys import setuptools from keystoneclient.openstack.common import setup def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() requires = setup.parse_requirements() depend_links = setup.parse_dependency_links() tests_require = setup.parse_requirements(['tools/test-requires']) setuptools.setup( name="python-keystoneclient", version=setup.get_post_version('keystoneclient'), description="Client library for OpenStack Keystone API", long_description=read('README.rst'), url='https://github.com/openstack/python-keystoneclient', license='Apache', author='Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss', author_email='gabriel.hurley@nebula.com', packages=setuptools.find_packages(exclude=['tests', 'tests.*']), classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], install_requires=requires, dependency_links=depend_links, cmdclass=setup.get_cmdclass(), tests_require=tests_require, test_suite="nose.collector", entry_points={ 'console_scripts': ['keystone = keystoneclient.shell:main'] } ) <commit_msg>Replace refs to 'Keystone API' with 'Identity API' Formally, OpenStack Keystone implements the OpenStack Identity API, and this is a client to the API, not to Keystone itself. Change-Id: If568866221a29ba041f0f2cd56dc81deeb9ebc00<commit_after>import os import sys import setuptools from keystoneclient.openstack.common import setup def read(fname): return open(os.path.join(os.path.dirname(__file__), fname)).read() requires = setup.parse_requirements() depend_links = setup.parse_dependency_links() tests_require = setup.parse_requirements(['tools/test-requires']) setuptools.setup( name="python-keystoneclient", version=setup.get_post_version('keystoneclient'), description="Client library for OpenStack Identity API (Keystone)", long_description=read('README.rst'), url='https://github.com/openstack/python-keystoneclient', license='Apache', author='Nebula Inc, based on work by Rackspace and Jacob Kaplan-Moss', author_email='gabriel.hurley@nebula.com', packages=setuptools.find_packages(exclude=['tests', 'tests.*']), classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'Intended Audience :: Developers', 'Intended Audience :: Information Technology', 'License :: OSI Approved :: Apache Software License', 'Operating System :: OS Independent', 'Programming Language :: Python', ], install_requires=requires, dependency_links=depend_links, cmdclass=setup.get_cmdclass(), tests_require=tests_require, test_suite="nose.collector", entry_points={ 'console_scripts': ['keystone = keystoneclient.shell:main'] } )
4f2a4ad90c85fb50eb02127676dec3e257e2b874
setup.py
setup.py
# -*- coding: utf-8 -*- """Package setup for heroku-tools CLI application.""" import os from setuptools import setup dependencies = ['click', 'sarge', 'pyyaml', 'requests', 'dateutils'] README = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read() setup( name='heroku-tools', version='0.2.1', url='https://github.com/yunojuno/heroku-tools', license='MIT', author='Hugo Rodger-Brown', author_email='hugo@yunojuno.com', description=( "Command line application for managing Heroku applications." ), long_description=README, include_package_data=True, packages=[ 'heroku_tools', ], install_requires=dependencies, entry_points={ 'console_scripts': [ 'heroku-tools = heroku_tools:entry_point', ], }, classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 2', ] )
# -*- coding: utf-8 -*- """Package setup for heroku-tools CLI application.""" import os from setuptools import setup dependencies = ['click', 'sarge', 'pyyaml', 'requests', 'dateutils'] README = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read() setup( name='heroku-tools', version='0.2.2', url='https://github.com/yunojuno/heroku-tools', license='MIT', author='Hugo Rodger-Brown', author_email='hugo@yunojuno.com', description=( "Command line application for managing Heroku applications." ), long_description=README, include_package_data=True, packages=[ 'heroku_tools', 'heroku_tools.settings', ], install_requires=dependencies, entry_points={ 'console_scripts': [ 'heroku-tools = heroku_tools:entry_point', ], }, classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 2', ] )
Add missing settings package to release (0.2.2)
Add missing settings package to release (0.2.2)
Python
mit
yunojuno/heroku-tools
# -*- coding: utf-8 -*- """Package setup for heroku-tools CLI application.""" import os from setuptools import setup dependencies = ['click', 'sarge', 'pyyaml', 'requests', 'dateutils'] README = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read() setup( name='heroku-tools', version='0.2.1', url='https://github.com/yunojuno/heroku-tools', license='MIT', author='Hugo Rodger-Brown', author_email='hugo@yunojuno.com', description=( "Command line application for managing Heroku applications." ), long_description=README, include_package_data=True, packages=[ 'heroku_tools', ], install_requires=dependencies, entry_points={ 'console_scripts': [ 'heroku-tools = heroku_tools:entry_point', ], }, classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 2', ] ) Add missing settings package to release (0.2.2)
# -*- coding: utf-8 -*- """Package setup for heroku-tools CLI application.""" import os from setuptools import setup dependencies = ['click', 'sarge', 'pyyaml', 'requests', 'dateutils'] README = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read() setup( name='heroku-tools', version='0.2.2', url='https://github.com/yunojuno/heroku-tools', license='MIT', author='Hugo Rodger-Brown', author_email='hugo@yunojuno.com', description=( "Command line application for managing Heroku applications." ), long_description=README, include_package_data=True, packages=[ 'heroku_tools', 'heroku_tools.settings', ], install_requires=dependencies, entry_points={ 'console_scripts': [ 'heroku-tools = heroku_tools:entry_point', ], }, classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 2', ] )
<commit_before># -*- coding: utf-8 -*- """Package setup for heroku-tools CLI application.""" import os from setuptools import setup dependencies = ['click', 'sarge', 'pyyaml', 'requests', 'dateutils'] README = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read() setup( name='heroku-tools', version='0.2.1', url='https://github.com/yunojuno/heroku-tools', license='MIT', author='Hugo Rodger-Brown', author_email='hugo@yunojuno.com', description=( "Command line application for managing Heroku applications." ), long_description=README, include_package_data=True, packages=[ 'heroku_tools', ], install_requires=dependencies, entry_points={ 'console_scripts': [ 'heroku-tools = heroku_tools:entry_point', ], }, classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 2', ] ) <commit_msg>Add missing settings package to release (0.2.2)<commit_after>
# -*- coding: utf-8 -*- """Package setup for heroku-tools CLI application.""" import os from setuptools import setup dependencies = ['click', 'sarge', 'pyyaml', 'requests', 'dateutils'] README = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read() setup( name='heroku-tools', version='0.2.2', url='https://github.com/yunojuno/heroku-tools', license='MIT', author='Hugo Rodger-Brown', author_email='hugo@yunojuno.com', description=( "Command line application for managing Heroku applications." ), long_description=README, include_package_data=True, packages=[ 'heroku_tools', 'heroku_tools.settings', ], install_requires=dependencies, entry_points={ 'console_scripts': [ 'heroku-tools = heroku_tools:entry_point', ], }, classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 2', ] )
# -*- coding: utf-8 -*- """Package setup for heroku-tools CLI application.""" import os from setuptools import setup dependencies = ['click', 'sarge', 'pyyaml', 'requests', 'dateutils'] README = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read() setup( name='heroku-tools', version='0.2.1', url='https://github.com/yunojuno/heroku-tools', license='MIT', author='Hugo Rodger-Brown', author_email='hugo@yunojuno.com', description=( "Command line application for managing Heroku applications." ), long_description=README, include_package_data=True, packages=[ 'heroku_tools', ], install_requires=dependencies, entry_points={ 'console_scripts': [ 'heroku-tools = heroku_tools:entry_point', ], }, classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 2', ] ) Add missing settings package to release (0.2.2)# -*- coding: utf-8 -*- """Package setup for heroku-tools CLI application.""" import os from setuptools import setup dependencies = ['click', 'sarge', 'pyyaml', 'requests', 'dateutils'] README = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read() setup( name='heroku-tools', version='0.2.2', url='https://github.com/yunojuno/heroku-tools', license='MIT', author='Hugo Rodger-Brown', author_email='hugo@yunojuno.com', description=( "Command line application for managing Heroku applications." ), long_description=README, include_package_data=True, packages=[ 'heroku_tools', 'heroku_tools.settings', ], install_requires=dependencies, entry_points={ 'console_scripts': [ 'heroku-tools = heroku_tools:entry_point', ], }, classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 2', ] )
<commit_before># -*- coding: utf-8 -*- """Package setup for heroku-tools CLI application.""" import os from setuptools import setup dependencies = ['click', 'sarge', 'pyyaml', 'requests', 'dateutils'] README = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read() setup( name='heroku-tools', version='0.2.1', url='https://github.com/yunojuno/heroku-tools', license='MIT', author='Hugo Rodger-Brown', author_email='hugo@yunojuno.com', description=( "Command line application for managing Heroku applications." ), long_description=README, include_package_data=True, packages=[ 'heroku_tools', ], install_requires=dependencies, entry_points={ 'console_scripts': [ 'heroku-tools = heroku_tools:entry_point', ], }, classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 2', ] ) <commit_msg>Add missing settings package to release (0.2.2)<commit_after># -*- coding: utf-8 -*- """Package setup for heroku-tools CLI application.""" import os from setuptools import setup dependencies = ['click', 'sarge', 'pyyaml', 'requests', 'dateutils'] README = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read() setup( name='heroku-tools', version='0.2.2', url='https://github.com/yunojuno/heroku-tools', license='MIT', author='Hugo Rodger-Brown', author_email='hugo@yunojuno.com', description=( "Command line application for managing Heroku applications." ), long_description=README, include_package_data=True, packages=[ 'heroku_tools', 'heroku_tools.settings', ], install_requires=dependencies, entry_points={ 'console_scripts': [ 'heroku-tools = heroku_tools:entry_point', ], }, classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python', 'Programming Language :: Python :: 2', ] )
21e99ff4617d4307d2223b79c25d3642132e27b3
setup.py
setup.py
from setuptools import setup setup(name='journal_dates', packages=[''], version='0.1', description='Prints a monthly journal template', url='http://github.com/bzamecnik/journal_dates', author='Bohumir Zamecnik', author_email='bohumir.zamecnik@gmail.com', license='MIT', install_requires=['arrow'], zip_safe=False, entry_points={ 'console_scripts': [ 'journal_dates=journal_dates:main', ], })
from setuptools import setup setup(name='journal_dates', packages=[''], version='0.1', description='Prints a monthly journal template', url='http://github.com/bzamecnik/journal_dates.py', author='Bohumir Zamecnik', author_email='bohumir.zamecnik@gmail.com', license='MIT', install_requires=['arrow'], zip_safe=False, entry_points={ 'console_scripts': [ 'journal_dates=journal_dates:main', ], })
Fix the github repo name.
Fix the github repo name.
Python
mit
bzamecnik/journal_dates
from setuptools import setup setup(name='journal_dates', packages=[''], version='0.1', description='Prints a monthly journal template', url='http://github.com/bzamecnik/journal_dates', author='Bohumir Zamecnik', author_email='bohumir.zamecnik@gmail.com', license='MIT', install_requires=['arrow'], zip_safe=False, entry_points={ 'console_scripts': [ 'journal_dates=journal_dates:main', ], }) Fix the github repo name.
from setuptools import setup setup(name='journal_dates', packages=[''], version='0.1', description='Prints a monthly journal template', url='http://github.com/bzamecnik/journal_dates.py', author='Bohumir Zamecnik', author_email='bohumir.zamecnik@gmail.com', license='MIT', install_requires=['arrow'], zip_safe=False, entry_points={ 'console_scripts': [ 'journal_dates=journal_dates:main', ], })
<commit_before>from setuptools import setup setup(name='journal_dates', packages=[''], version='0.1', description='Prints a monthly journal template', url='http://github.com/bzamecnik/journal_dates', author='Bohumir Zamecnik', author_email='bohumir.zamecnik@gmail.com', license='MIT', install_requires=['arrow'], zip_safe=False, entry_points={ 'console_scripts': [ 'journal_dates=journal_dates:main', ], }) <commit_msg>Fix the github repo name.<commit_after>
from setuptools import setup setup(name='journal_dates', packages=[''], version='0.1', description='Prints a monthly journal template', url='http://github.com/bzamecnik/journal_dates.py', author='Bohumir Zamecnik', author_email='bohumir.zamecnik@gmail.com', license='MIT', install_requires=['arrow'], zip_safe=False, entry_points={ 'console_scripts': [ 'journal_dates=journal_dates:main', ], })
from setuptools import setup setup(name='journal_dates', packages=[''], version='0.1', description='Prints a monthly journal template', url='http://github.com/bzamecnik/journal_dates', author='Bohumir Zamecnik', author_email='bohumir.zamecnik@gmail.com', license='MIT', install_requires=['arrow'], zip_safe=False, entry_points={ 'console_scripts': [ 'journal_dates=journal_dates:main', ], }) Fix the github repo name.from setuptools import setup setup(name='journal_dates', packages=[''], version='0.1', description='Prints a monthly journal template', url='http://github.com/bzamecnik/journal_dates.py', author='Bohumir Zamecnik', author_email='bohumir.zamecnik@gmail.com', license='MIT', install_requires=['arrow'], zip_safe=False, entry_points={ 'console_scripts': [ 'journal_dates=journal_dates:main', ], })
<commit_before>from setuptools import setup setup(name='journal_dates', packages=[''], version='0.1', description='Prints a monthly journal template', url='http://github.com/bzamecnik/journal_dates', author='Bohumir Zamecnik', author_email='bohumir.zamecnik@gmail.com', license='MIT', install_requires=['arrow'], zip_safe=False, entry_points={ 'console_scripts': [ 'journal_dates=journal_dates:main', ], }) <commit_msg>Fix the github repo name.<commit_after>from setuptools import setup setup(name='journal_dates', packages=[''], version='0.1', description='Prints a monthly journal template', url='http://github.com/bzamecnik/journal_dates.py', author='Bohumir Zamecnik', author_email='bohumir.zamecnik@gmail.com', license='MIT', install_requires=['arrow'], zip_safe=False, entry_points={ 'console_scripts': [ 'journal_dates=journal_dates:main', ], })
86fc1b91aeb9dc17b5776ef372050c707c64fc30
setup.py
setup.py
#!/usr/bin/env python """ properties: Fancy properties for Python. """ from distutils.core import setup from setuptools import find_packages CLASSIFIERS = [ 'Development Status :: 4 - Beta', 'Programming Language :: Python', 'Topic :: Scientific/Engineering', 'Topic :: Scientific/Engineering :: Mathematics', 'Topic :: Scientific/Engineering :: Physics', 'Operating System :: Microsoft :: Windows', 'Operating System :: POSIX', 'Operating System :: Unix', 'Operating System :: MacOS', 'Natural Language :: English', ] with open("README.rst") as f: LONG_DESCRIPTION = ''.join(f.readlines()) setup( name="properties", version="0.2.3", packages=find_packages(exclude=('tests',)), install_requires=[ 'future', 'numpy>=1.7', 'six', 'vectormath>=0.1.0', ], author="3point Science", author_email="info@3ptscience.com", description="properties", long_description=LONG_DESCRIPTION, keywords="property", url="http://steno3d.com/", download_url="http://github.com/3ptscience/properties", classifiers=CLASSIFIERS, platforms=["Windows", "Linux", "Solaris", "Mac OS-X", "Unix"], use_2to3=False, )
#!/usr/bin/env python """ properties: Fancy properties for Python. """ from distutils.core import setup from setuptools import find_packages CLASSIFIERS = [ 'Development Status :: 4 - Beta', 'Programming Language :: Python', 'Topic :: Scientific/Engineering', 'Topic :: Scientific/Engineering :: Mathematics', 'Topic :: Scientific/Engineering :: Physics', 'Operating System :: Microsoft :: Windows', 'Operating System :: POSIX', 'Operating System :: Unix', 'Operating System :: MacOS', 'Natural Language :: English', ] with open("README.rst") as f: LONG_DESCRIPTION = ''.join(f.readlines()) setup( name="properties", version="0.2.3", packages=find_packages(exclude=('tests',)), install_requires=[ 'numpy>=1.7', 'six', 'vectormath>=0.1.1', ], author="3point Science", author_email="info@3ptscience.com", description="properties", long_description=LONG_DESCRIPTION, keywords="property", url="http://steno3d.com/", download_url="http://github.com/3ptscience/properties", classifiers=CLASSIFIERS, platforms=["Windows", "Linux", "Solaris", "Mac OS-X", "Unix"], use_2to3=False, )
Remove future, bump vectormath dependencies
Remove future, bump vectormath dependencies
Python
mit
3ptscience/properties,aranzgeo/properties
#!/usr/bin/env python """ properties: Fancy properties for Python. """ from distutils.core import setup from setuptools import find_packages CLASSIFIERS = [ 'Development Status :: 4 - Beta', 'Programming Language :: Python', 'Topic :: Scientific/Engineering', 'Topic :: Scientific/Engineering :: Mathematics', 'Topic :: Scientific/Engineering :: Physics', 'Operating System :: Microsoft :: Windows', 'Operating System :: POSIX', 'Operating System :: Unix', 'Operating System :: MacOS', 'Natural Language :: English', ] with open("README.rst") as f: LONG_DESCRIPTION = ''.join(f.readlines()) setup( name="properties", version="0.2.3", packages=find_packages(exclude=('tests',)), install_requires=[ 'future', 'numpy>=1.7', 'six', 'vectormath>=0.1.0', ], author="3point Science", author_email="info@3ptscience.com", description="properties", long_description=LONG_DESCRIPTION, keywords="property", url="http://steno3d.com/", download_url="http://github.com/3ptscience/properties", classifiers=CLASSIFIERS, platforms=["Windows", "Linux", "Solaris", "Mac OS-X", "Unix"], use_2to3=False, ) Remove future, bump vectormath dependencies
#!/usr/bin/env python """ properties: Fancy properties for Python. """ from distutils.core import setup from setuptools import find_packages CLASSIFIERS = [ 'Development Status :: 4 - Beta', 'Programming Language :: Python', 'Topic :: Scientific/Engineering', 'Topic :: Scientific/Engineering :: Mathematics', 'Topic :: Scientific/Engineering :: Physics', 'Operating System :: Microsoft :: Windows', 'Operating System :: POSIX', 'Operating System :: Unix', 'Operating System :: MacOS', 'Natural Language :: English', ] with open("README.rst") as f: LONG_DESCRIPTION = ''.join(f.readlines()) setup( name="properties", version="0.2.3", packages=find_packages(exclude=('tests',)), install_requires=[ 'numpy>=1.7', 'six', 'vectormath>=0.1.1', ], author="3point Science", author_email="info@3ptscience.com", description="properties", long_description=LONG_DESCRIPTION, keywords="property", url="http://steno3d.com/", download_url="http://github.com/3ptscience/properties", classifiers=CLASSIFIERS, platforms=["Windows", "Linux", "Solaris", "Mac OS-X", "Unix"], use_2to3=False, )
<commit_before>#!/usr/bin/env python """ properties: Fancy properties for Python. """ from distutils.core import setup from setuptools import find_packages CLASSIFIERS = [ 'Development Status :: 4 - Beta', 'Programming Language :: Python', 'Topic :: Scientific/Engineering', 'Topic :: Scientific/Engineering :: Mathematics', 'Topic :: Scientific/Engineering :: Physics', 'Operating System :: Microsoft :: Windows', 'Operating System :: POSIX', 'Operating System :: Unix', 'Operating System :: MacOS', 'Natural Language :: English', ] with open("README.rst") as f: LONG_DESCRIPTION = ''.join(f.readlines()) setup( name="properties", version="0.2.3", packages=find_packages(exclude=('tests',)), install_requires=[ 'future', 'numpy>=1.7', 'six', 'vectormath>=0.1.0', ], author="3point Science", author_email="info@3ptscience.com", description="properties", long_description=LONG_DESCRIPTION, keywords="property", url="http://steno3d.com/", download_url="http://github.com/3ptscience/properties", classifiers=CLASSIFIERS, platforms=["Windows", "Linux", "Solaris", "Mac OS-X", "Unix"], use_2to3=False, ) <commit_msg>Remove future, bump vectormath dependencies<commit_after>
#!/usr/bin/env python """ properties: Fancy properties for Python. """ from distutils.core import setup from setuptools import find_packages CLASSIFIERS = [ 'Development Status :: 4 - Beta', 'Programming Language :: Python', 'Topic :: Scientific/Engineering', 'Topic :: Scientific/Engineering :: Mathematics', 'Topic :: Scientific/Engineering :: Physics', 'Operating System :: Microsoft :: Windows', 'Operating System :: POSIX', 'Operating System :: Unix', 'Operating System :: MacOS', 'Natural Language :: English', ] with open("README.rst") as f: LONG_DESCRIPTION = ''.join(f.readlines()) setup( name="properties", version="0.2.3", packages=find_packages(exclude=('tests',)), install_requires=[ 'numpy>=1.7', 'six', 'vectormath>=0.1.1', ], author="3point Science", author_email="info@3ptscience.com", description="properties", long_description=LONG_DESCRIPTION, keywords="property", url="http://steno3d.com/", download_url="http://github.com/3ptscience/properties", classifiers=CLASSIFIERS, platforms=["Windows", "Linux", "Solaris", "Mac OS-X", "Unix"], use_2to3=False, )
#!/usr/bin/env python """ properties: Fancy properties for Python. """ from distutils.core import setup from setuptools import find_packages CLASSIFIERS = [ 'Development Status :: 4 - Beta', 'Programming Language :: Python', 'Topic :: Scientific/Engineering', 'Topic :: Scientific/Engineering :: Mathematics', 'Topic :: Scientific/Engineering :: Physics', 'Operating System :: Microsoft :: Windows', 'Operating System :: POSIX', 'Operating System :: Unix', 'Operating System :: MacOS', 'Natural Language :: English', ] with open("README.rst") as f: LONG_DESCRIPTION = ''.join(f.readlines()) setup( name="properties", version="0.2.3", packages=find_packages(exclude=('tests',)), install_requires=[ 'future', 'numpy>=1.7', 'six', 'vectormath>=0.1.0', ], author="3point Science", author_email="info@3ptscience.com", description="properties", long_description=LONG_DESCRIPTION, keywords="property", url="http://steno3d.com/", download_url="http://github.com/3ptscience/properties", classifiers=CLASSIFIERS, platforms=["Windows", "Linux", "Solaris", "Mac OS-X", "Unix"], use_2to3=False, ) Remove future, bump vectormath dependencies#!/usr/bin/env python """ properties: Fancy properties for Python. """ from distutils.core import setup from setuptools import find_packages CLASSIFIERS = [ 'Development Status :: 4 - Beta', 'Programming Language :: Python', 'Topic :: Scientific/Engineering', 'Topic :: Scientific/Engineering :: Mathematics', 'Topic :: Scientific/Engineering :: Physics', 'Operating System :: Microsoft :: Windows', 'Operating System :: POSIX', 'Operating System :: Unix', 'Operating System :: MacOS', 'Natural Language :: English', ] with open("README.rst") as f: LONG_DESCRIPTION = ''.join(f.readlines()) setup( name="properties", version="0.2.3", packages=find_packages(exclude=('tests',)), install_requires=[ 'numpy>=1.7', 'six', 'vectormath>=0.1.1', ], author="3point Science", author_email="info@3ptscience.com", description="properties", long_description=LONG_DESCRIPTION, keywords="property", url="http://steno3d.com/", download_url="http://github.com/3ptscience/properties", classifiers=CLASSIFIERS, platforms=["Windows", "Linux", "Solaris", "Mac OS-X", "Unix"], use_2to3=False, )
<commit_before>#!/usr/bin/env python """ properties: Fancy properties for Python. """ from distutils.core import setup from setuptools import find_packages CLASSIFIERS = [ 'Development Status :: 4 - Beta', 'Programming Language :: Python', 'Topic :: Scientific/Engineering', 'Topic :: Scientific/Engineering :: Mathematics', 'Topic :: Scientific/Engineering :: Physics', 'Operating System :: Microsoft :: Windows', 'Operating System :: POSIX', 'Operating System :: Unix', 'Operating System :: MacOS', 'Natural Language :: English', ] with open("README.rst") as f: LONG_DESCRIPTION = ''.join(f.readlines()) setup( name="properties", version="0.2.3", packages=find_packages(exclude=('tests',)), install_requires=[ 'future', 'numpy>=1.7', 'six', 'vectormath>=0.1.0', ], author="3point Science", author_email="info@3ptscience.com", description="properties", long_description=LONG_DESCRIPTION, keywords="property", url="http://steno3d.com/", download_url="http://github.com/3ptscience/properties", classifiers=CLASSIFIERS, platforms=["Windows", "Linux", "Solaris", "Mac OS-X", "Unix"], use_2to3=False, ) <commit_msg>Remove future, bump vectormath dependencies<commit_after>#!/usr/bin/env python """ properties: Fancy properties for Python. """ from distutils.core import setup from setuptools import find_packages CLASSIFIERS = [ 'Development Status :: 4 - Beta', 'Programming Language :: Python', 'Topic :: Scientific/Engineering', 'Topic :: Scientific/Engineering :: Mathematics', 'Topic :: Scientific/Engineering :: Physics', 'Operating System :: Microsoft :: Windows', 'Operating System :: POSIX', 'Operating System :: Unix', 'Operating System :: MacOS', 'Natural Language :: English', ] with open("README.rst") as f: LONG_DESCRIPTION = ''.join(f.readlines()) setup( name="properties", version="0.2.3", packages=find_packages(exclude=('tests',)), install_requires=[ 'numpy>=1.7', 'six', 'vectormath>=0.1.1', ], author="3point Science", author_email="info@3ptscience.com", description="properties", long_description=LONG_DESCRIPTION, keywords="property", url="http://steno3d.com/", download_url="http://github.com/3ptscience/properties", classifiers=CLASSIFIERS, platforms=["Windows", "Linux", "Solaris", "Mac OS-X", "Unix"], use_2to3=False, )
c4b4a7ab03728b907d0c64e3322fd08a56f5c9ce
suave.py
suave.py
#!/usr/bin/env python import curses import os import time from box import Box from utils import load_yaml def main(screen): """ Draws and redraws the screen. """ # Hide the cursor. curses.curs_set(0) # Load config from file. config = load_yaml(os.path.expanduser('~/.suave/config.yml')) # Create boxes from config. boxes = [] for box in config: boxes.append( Box( screen=screen, rows=box['rows'], columns=box['columns'], rows_offset=box['rows-offset'], columns_offset=box['columns-offset'], command=box['command'], interval=box['interval'], ) ) while True: # Re/draw the screen and boxes. screen.erase() screen.refresh() [box.refresh() for box in boxes] # Wait before redrawing again. time.sleep(1) curses.wrapper(main)
#!/usr/bin/env python import curses import os import time from box import Box from utils import load_yaml def main(screen): """ Draws and redraws the screen. """ # Hide the cursor. curses.curs_set(0) # Load config from file. config = load_yaml(os.path.expanduser('~/.suave/config.yml')) # Create boxes from config. boxes = [] for box in config: boxes.append( Box( screen=screen, rows=box['rows'], columns=box['columns'], rows_offset=box['rows-offset'], columns_offset=box['columns-offset'], command=box['command'], interval=box['interval'], ) ) while True: # Re/draw the screen and boxes. screen.clear() screen.refresh() [box.refresh() for box in boxes] # Wait before redrawing again. time.sleep(1) curses.wrapper(main)
Clear the screen rather than erasing it before refreshing
Clear the screen rather than erasing it before refreshing
Python
mit
countermeasure/suave
#!/usr/bin/env python import curses import os import time from box import Box from utils import load_yaml def main(screen): """ Draws and redraws the screen. """ # Hide the cursor. curses.curs_set(0) # Load config from file. config = load_yaml(os.path.expanduser('~/.suave/config.yml')) # Create boxes from config. boxes = [] for box in config: boxes.append( Box( screen=screen, rows=box['rows'], columns=box['columns'], rows_offset=box['rows-offset'], columns_offset=box['columns-offset'], command=box['command'], interval=box['interval'], ) ) while True: # Re/draw the screen and boxes. screen.erase() screen.refresh() [box.refresh() for box in boxes] # Wait before redrawing again. time.sleep(1) curses.wrapper(main) Clear the screen rather than erasing it before refreshing
#!/usr/bin/env python import curses import os import time from box import Box from utils import load_yaml def main(screen): """ Draws and redraws the screen. """ # Hide the cursor. curses.curs_set(0) # Load config from file. config = load_yaml(os.path.expanduser('~/.suave/config.yml')) # Create boxes from config. boxes = [] for box in config: boxes.append( Box( screen=screen, rows=box['rows'], columns=box['columns'], rows_offset=box['rows-offset'], columns_offset=box['columns-offset'], command=box['command'], interval=box['interval'], ) ) while True: # Re/draw the screen and boxes. screen.clear() screen.refresh() [box.refresh() for box in boxes] # Wait before redrawing again. time.sleep(1) curses.wrapper(main)
<commit_before>#!/usr/bin/env python import curses import os import time from box import Box from utils import load_yaml def main(screen): """ Draws and redraws the screen. """ # Hide the cursor. curses.curs_set(0) # Load config from file. config = load_yaml(os.path.expanduser('~/.suave/config.yml')) # Create boxes from config. boxes = [] for box in config: boxes.append( Box( screen=screen, rows=box['rows'], columns=box['columns'], rows_offset=box['rows-offset'], columns_offset=box['columns-offset'], command=box['command'], interval=box['interval'], ) ) while True: # Re/draw the screen and boxes. screen.erase() screen.refresh() [box.refresh() for box in boxes] # Wait before redrawing again. time.sleep(1) curses.wrapper(main) <commit_msg>Clear the screen rather than erasing it before refreshing<commit_after>
#!/usr/bin/env python import curses import os import time from box import Box from utils import load_yaml def main(screen): """ Draws and redraws the screen. """ # Hide the cursor. curses.curs_set(0) # Load config from file. config = load_yaml(os.path.expanduser('~/.suave/config.yml')) # Create boxes from config. boxes = [] for box in config: boxes.append( Box( screen=screen, rows=box['rows'], columns=box['columns'], rows_offset=box['rows-offset'], columns_offset=box['columns-offset'], command=box['command'], interval=box['interval'], ) ) while True: # Re/draw the screen and boxes. screen.clear() screen.refresh() [box.refresh() for box in boxes] # Wait before redrawing again. time.sleep(1) curses.wrapper(main)
#!/usr/bin/env python import curses import os import time from box import Box from utils import load_yaml def main(screen): """ Draws and redraws the screen. """ # Hide the cursor. curses.curs_set(0) # Load config from file. config = load_yaml(os.path.expanduser('~/.suave/config.yml')) # Create boxes from config. boxes = [] for box in config: boxes.append( Box( screen=screen, rows=box['rows'], columns=box['columns'], rows_offset=box['rows-offset'], columns_offset=box['columns-offset'], command=box['command'], interval=box['interval'], ) ) while True: # Re/draw the screen and boxes. screen.erase() screen.refresh() [box.refresh() for box in boxes] # Wait before redrawing again. time.sleep(1) curses.wrapper(main) Clear the screen rather than erasing it before refreshing#!/usr/bin/env python import curses import os import time from box import Box from utils import load_yaml def main(screen): """ Draws and redraws the screen. """ # Hide the cursor. curses.curs_set(0) # Load config from file. config = load_yaml(os.path.expanduser('~/.suave/config.yml')) # Create boxes from config. boxes = [] for box in config: boxes.append( Box( screen=screen, rows=box['rows'], columns=box['columns'], rows_offset=box['rows-offset'], columns_offset=box['columns-offset'], command=box['command'], interval=box['interval'], ) ) while True: # Re/draw the screen and boxes. screen.clear() screen.refresh() [box.refresh() for box in boxes] # Wait before redrawing again. time.sleep(1) curses.wrapper(main)
<commit_before>#!/usr/bin/env python import curses import os import time from box import Box from utils import load_yaml def main(screen): """ Draws and redraws the screen. """ # Hide the cursor. curses.curs_set(0) # Load config from file. config = load_yaml(os.path.expanduser('~/.suave/config.yml')) # Create boxes from config. boxes = [] for box in config: boxes.append( Box( screen=screen, rows=box['rows'], columns=box['columns'], rows_offset=box['rows-offset'], columns_offset=box['columns-offset'], command=box['command'], interval=box['interval'], ) ) while True: # Re/draw the screen and boxes. screen.erase() screen.refresh() [box.refresh() for box in boxes] # Wait before redrawing again. time.sleep(1) curses.wrapper(main) <commit_msg>Clear the screen rather than erasing it before refreshing<commit_after>#!/usr/bin/env python import curses import os import time from box import Box from utils import load_yaml def main(screen): """ Draws and redraws the screen. """ # Hide the cursor. curses.curs_set(0) # Load config from file. config = load_yaml(os.path.expanduser('~/.suave/config.yml')) # Create boxes from config. boxes = [] for box in config: boxes.append( Box( screen=screen, rows=box['rows'], columns=box['columns'], rows_offset=box['rows-offset'], columns_offset=box['columns-offset'], command=box['command'], interval=box['interval'], ) ) while True: # Re/draw the screen and boxes. screen.clear() screen.refresh() [box.refresh() for box in boxes] # Wait before redrawing again. time.sleep(1) curses.wrapper(main)
4ee409a5635b1d027f5d3c68fb2a62f554c9c801
ib_insync/__init__.py
ib_insync/__init__.py
import sys import importlib from .version import __version__, __version_info__ # noqa from . import util if sys.version_info < (3, 6, 0): raise RuntimeError('ib_insync requires Python 3.6 or higher') try: import ibapi except ImportError: raise RuntimeError( 'IB API from http://interactivebrokers.github.io is required') if util.ibapiVersionInfo() < (9, 73, 6): raise RuntimeError( f'Old version ({ibapi.__version__}) of ibapi package detected. ' 'The newest version from http://interactivebrokers.github.io ' 'is required') from .version import __version__, __version_info__ # noqa from .objects import * # noqa from .event import * # noqa from .contract import * # noqa from .order import * # noqa from .ticker import * # noqa from .ib import * # noqa from .client import * # noqa from .wrapper import * # noqa from .flexreport import * # noqa from .ibcontroller import * # noqa from . import util # noqa __all__ = ['util'] for _m in ( objects, event, contract, order, ticker, ib, # noqa client, wrapper, flexreport, ibcontroller): # noqa __all__ += _m.__all__ del sys del importlib del ibapi
import sys import importlib if sys.version_info < (3, 6, 0): raise RuntimeError('ib_insync requires Python 3.6 or higher') try: import ibapi except ImportError: raise RuntimeError( 'IB API from http://interactivebrokers.github.io is required') from . import util # noqa if util.ibapiVersionInfo() < (9, 73, 6): raise RuntimeError( f'Old version ({ibapi.__version__}) of ibapi package detected. ' 'The newest version from http://interactivebrokers.github.io ' 'is required') from .version import __version__, __version_info__ # noqa from .objects import * # noqa from .event import * # noqa from .contract import * # noqa from .order import * # noqa from .ticker import * # noqa from .ib import * # noqa from .client import * # noqa from .wrapper import * # noqa from .flexreport import * # noqa from .ibcontroller import * # noqa __all__ = ['util'] for _m in ( objects, event, contract, order, ticker, ib, # noqa client, wrapper, flexreport, ibcontroller): # noqa __all__ += _m.__all__ del sys del importlib del ibapi
Fix explicit check for presence of ibapi package
Fix explicit check for presence of ibapi package
Python
bsd-2-clause
erdewit/ib_insync,erdewit/ib_insync
import sys import importlib from .version import __version__, __version_info__ # noqa from . import util if sys.version_info < (3, 6, 0): raise RuntimeError('ib_insync requires Python 3.6 or higher') try: import ibapi except ImportError: raise RuntimeError( 'IB API from http://interactivebrokers.github.io is required') if util.ibapiVersionInfo() < (9, 73, 6): raise RuntimeError( f'Old version ({ibapi.__version__}) of ibapi package detected. ' 'The newest version from http://interactivebrokers.github.io ' 'is required') from .version import __version__, __version_info__ # noqa from .objects import * # noqa from .event import * # noqa from .contract import * # noqa from .order import * # noqa from .ticker import * # noqa from .ib import * # noqa from .client import * # noqa from .wrapper import * # noqa from .flexreport import * # noqa from .ibcontroller import * # noqa from . import util # noqa __all__ = ['util'] for _m in ( objects, event, contract, order, ticker, ib, # noqa client, wrapper, flexreport, ibcontroller): # noqa __all__ += _m.__all__ del sys del importlib del ibapi Fix explicit check for presence of ibapi package
import sys import importlib if sys.version_info < (3, 6, 0): raise RuntimeError('ib_insync requires Python 3.6 or higher') try: import ibapi except ImportError: raise RuntimeError( 'IB API from http://interactivebrokers.github.io is required') from . import util # noqa if util.ibapiVersionInfo() < (9, 73, 6): raise RuntimeError( f'Old version ({ibapi.__version__}) of ibapi package detected. ' 'The newest version from http://interactivebrokers.github.io ' 'is required') from .version import __version__, __version_info__ # noqa from .objects import * # noqa from .event import * # noqa from .contract import * # noqa from .order import * # noqa from .ticker import * # noqa from .ib import * # noqa from .client import * # noqa from .wrapper import * # noqa from .flexreport import * # noqa from .ibcontroller import * # noqa __all__ = ['util'] for _m in ( objects, event, contract, order, ticker, ib, # noqa client, wrapper, flexreport, ibcontroller): # noqa __all__ += _m.__all__ del sys del importlib del ibapi
<commit_before>import sys import importlib from .version import __version__, __version_info__ # noqa from . import util if sys.version_info < (3, 6, 0): raise RuntimeError('ib_insync requires Python 3.6 or higher') try: import ibapi except ImportError: raise RuntimeError( 'IB API from http://interactivebrokers.github.io is required') if util.ibapiVersionInfo() < (9, 73, 6): raise RuntimeError( f'Old version ({ibapi.__version__}) of ibapi package detected. ' 'The newest version from http://interactivebrokers.github.io ' 'is required') from .version import __version__, __version_info__ # noqa from .objects import * # noqa from .event import * # noqa from .contract import * # noqa from .order import * # noqa from .ticker import * # noqa from .ib import * # noqa from .client import * # noqa from .wrapper import * # noqa from .flexreport import * # noqa from .ibcontroller import * # noqa from . import util # noqa __all__ = ['util'] for _m in ( objects, event, contract, order, ticker, ib, # noqa client, wrapper, flexreport, ibcontroller): # noqa __all__ += _m.__all__ del sys del importlib del ibapi <commit_msg>Fix explicit check for presence of ibapi package<commit_after>
import sys import importlib if sys.version_info < (3, 6, 0): raise RuntimeError('ib_insync requires Python 3.6 or higher') try: import ibapi except ImportError: raise RuntimeError( 'IB API from http://interactivebrokers.github.io is required') from . import util # noqa if util.ibapiVersionInfo() < (9, 73, 6): raise RuntimeError( f'Old version ({ibapi.__version__}) of ibapi package detected. ' 'The newest version from http://interactivebrokers.github.io ' 'is required') from .version import __version__, __version_info__ # noqa from .objects import * # noqa from .event import * # noqa from .contract import * # noqa from .order import * # noqa from .ticker import * # noqa from .ib import * # noqa from .client import * # noqa from .wrapper import * # noqa from .flexreport import * # noqa from .ibcontroller import * # noqa __all__ = ['util'] for _m in ( objects, event, contract, order, ticker, ib, # noqa client, wrapper, flexreport, ibcontroller): # noqa __all__ += _m.__all__ del sys del importlib del ibapi
import sys import importlib from .version import __version__, __version_info__ # noqa from . import util if sys.version_info < (3, 6, 0): raise RuntimeError('ib_insync requires Python 3.6 or higher') try: import ibapi except ImportError: raise RuntimeError( 'IB API from http://interactivebrokers.github.io is required') if util.ibapiVersionInfo() < (9, 73, 6): raise RuntimeError( f'Old version ({ibapi.__version__}) of ibapi package detected. ' 'The newest version from http://interactivebrokers.github.io ' 'is required') from .version import __version__, __version_info__ # noqa from .objects import * # noqa from .event import * # noqa from .contract import * # noqa from .order import * # noqa from .ticker import * # noqa from .ib import * # noqa from .client import * # noqa from .wrapper import * # noqa from .flexreport import * # noqa from .ibcontroller import * # noqa from . import util # noqa __all__ = ['util'] for _m in ( objects, event, contract, order, ticker, ib, # noqa client, wrapper, flexreport, ibcontroller): # noqa __all__ += _m.__all__ del sys del importlib del ibapi Fix explicit check for presence of ibapi packageimport sys import importlib if sys.version_info < (3, 6, 0): raise RuntimeError('ib_insync requires Python 3.6 or higher') try: import ibapi except ImportError: raise RuntimeError( 'IB API from http://interactivebrokers.github.io is required') from . import util # noqa if util.ibapiVersionInfo() < (9, 73, 6): raise RuntimeError( f'Old version ({ibapi.__version__}) of ibapi package detected. ' 'The newest version from http://interactivebrokers.github.io ' 'is required') from .version import __version__, __version_info__ # noqa from .objects import * # noqa from .event import * # noqa from .contract import * # noqa from .order import * # noqa from .ticker import * # noqa from .ib import * # noqa from .client import * # noqa from .wrapper import * # noqa from .flexreport import * # noqa from .ibcontroller import * # noqa __all__ = ['util'] for _m in ( objects, event, contract, order, ticker, ib, # noqa client, wrapper, flexreport, ibcontroller): # noqa __all__ += _m.__all__ del sys del importlib del ibapi
<commit_before>import sys import importlib from .version import __version__, __version_info__ # noqa from . import util if sys.version_info < (3, 6, 0): raise RuntimeError('ib_insync requires Python 3.6 or higher') try: import ibapi except ImportError: raise RuntimeError( 'IB API from http://interactivebrokers.github.io is required') if util.ibapiVersionInfo() < (9, 73, 6): raise RuntimeError( f'Old version ({ibapi.__version__}) of ibapi package detected. ' 'The newest version from http://interactivebrokers.github.io ' 'is required') from .version import __version__, __version_info__ # noqa from .objects import * # noqa from .event import * # noqa from .contract import * # noqa from .order import * # noqa from .ticker import * # noqa from .ib import * # noqa from .client import * # noqa from .wrapper import * # noqa from .flexreport import * # noqa from .ibcontroller import * # noqa from . import util # noqa __all__ = ['util'] for _m in ( objects, event, contract, order, ticker, ib, # noqa client, wrapper, flexreport, ibcontroller): # noqa __all__ += _m.__all__ del sys del importlib del ibapi <commit_msg>Fix explicit check for presence of ibapi package<commit_after>import sys import importlib if sys.version_info < (3, 6, 0): raise RuntimeError('ib_insync requires Python 3.6 or higher') try: import ibapi except ImportError: raise RuntimeError( 'IB API from http://interactivebrokers.github.io is required') from . import util # noqa if util.ibapiVersionInfo() < (9, 73, 6): raise RuntimeError( f'Old version ({ibapi.__version__}) of ibapi package detected. ' 'The newest version from http://interactivebrokers.github.io ' 'is required') from .version import __version__, __version_info__ # noqa from .objects import * # noqa from .event import * # noqa from .contract import * # noqa from .order import * # noqa from .ticker import * # noqa from .ib import * # noqa from .client import * # noqa from .wrapper import * # noqa from .flexreport import * # noqa from .ibcontroller import * # noqa __all__ = ['util'] for _m in ( objects, event, contract, order, ticker, ib, # noqa client, wrapper, flexreport, ibcontroller): # noqa __all__ += _m.__all__ del sys del importlib del ibapi
50395649bf65b2ee7ef31fc8fedf62b9a66e96bf
metaci/cumulusci/utils.py
metaci/cumulusci/utils.py
from cumulusci.core.config import ConnectedAppOAuthConfig from django.conf import settings def get_connected_app(): return ConnectedAppOAuthConfig({ 'callback_url': settings.CONNECTED_APP_CALLBACK_URL, 'client_id': settings.CONNECTED_APP_CLIENT_ID, 'client_secret': settings.CONNECTED_APP_CLIENT_SECRET, })
from cumulusci.core.config import ConnectedAppOAuthConfig from django.conf import settings def get_connected_app(): return ConnectedAppOAuthConfig({ 'callback_url': settings.CONNECTED_APP_CALLBACK_URL, 'client_id': settings.CONNECTED_APP_CLIENT_ID, 'client_secret': settings.CONNECTED_APP_CLIENT_SECRET, })
Revert file to original state
Revert file to original state
Python
bsd-3-clause
SalesforceFoundation/mrbelvedereci,SalesforceFoundation/mrbelvedereci,SalesforceFoundation/mrbelvedereci,SalesforceFoundation/mrbelvedereci
from cumulusci.core.config import ConnectedAppOAuthConfig from django.conf import settings def get_connected_app(): return ConnectedAppOAuthConfig({ 'callback_url': settings.CONNECTED_APP_CALLBACK_URL, 'client_id': settings.CONNECTED_APP_CLIENT_ID, 'client_secret': settings.CONNECTED_APP_CLIENT_SECRET, }) Revert file to original state
from cumulusci.core.config import ConnectedAppOAuthConfig from django.conf import settings def get_connected_app(): return ConnectedAppOAuthConfig({ 'callback_url': settings.CONNECTED_APP_CALLBACK_URL, 'client_id': settings.CONNECTED_APP_CLIENT_ID, 'client_secret': settings.CONNECTED_APP_CLIENT_SECRET, })
<commit_before>from cumulusci.core.config import ConnectedAppOAuthConfig from django.conf import settings def get_connected_app(): return ConnectedAppOAuthConfig({ 'callback_url': settings.CONNECTED_APP_CALLBACK_URL, 'client_id': settings.CONNECTED_APP_CLIENT_ID, 'client_secret': settings.CONNECTED_APP_CLIENT_SECRET, }) <commit_msg>Revert file to original state<commit_after>
from cumulusci.core.config import ConnectedAppOAuthConfig from django.conf import settings def get_connected_app(): return ConnectedAppOAuthConfig({ 'callback_url': settings.CONNECTED_APP_CALLBACK_URL, 'client_id': settings.CONNECTED_APP_CLIENT_ID, 'client_secret': settings.CONNECTED_APP_CLIENT_SECRET, })
from cumulusci.core.config import ConnectedAppOAuthConfig from django.conf import settings def get_connected_app(): return ConnectedAppOAuthConfig({ 'callback_url': settings.CONNECTED_APP_CALLBACK_URL, 'client_id': settings.CONNECTED_APP_CLIENT_ID, 'client_secret': settings.CONNECTED_APP_CLIENT_SECRET, }) Revert file to original statefrom cumulusci.core.config import ConnectedAppOAuthConfig from django.conf import settings def get_connected_app(): return ConnectedAppOAuthConfig({ 'callback_url': settings.CONNECTED_APP_CALLBACK_URL, 'client_id': settings.CONNECTED_APP_CLIENT_ID, 'client_secret': settings.CONNECTED_APP_CLIENT_SECRET, })
<commit_before>from cumulusci.core.config import ConnectedAppOAuthConfig from django.conf import settings def get_connected_app(): return ConnectedAppOAuthConfig({ 'callback_url': settings.CONNECTED_APP_CALLBACK_URL, 'client_id': settings.CONNECTED_APP_CLIENT_ID, 'client_secret': settings.CONNECTED_APP_CLIENT_SECRET, }) <commit_msg>Revert file to original state<commit_after>from cumulusci.core.config import ConnectedAppOAuthConfig from django.conf import settings def get_connected_app(): return ConnectedAppOAuthConfig({ 'callback_url': settings.CONNECTED_APP_CALLBACK_URL, 'client_id': settings.CONNECTED_APP_CLIENT_ID, 'client_secret': settings.CONNECTED_APP_CLIENT_SECRET, })
525c224080b3ac13864fbd3b5b9db2e884691edf
polyaxon/sidecar/sidecar/sidecar/monitor.py
polyaxon/sidecar/sidecar/sidecar/monitor.py
import ocular from polyaxon_schemas.pod import PodLifeCycle def is_container_terminated(event, container_id): statuses_by_name = ocular.processor.get_container_statuses_by_name( event.status.to_dict().get('container_statuses', [])) statuses = ocular.processor.get_container_status(statuses_by_name, (container_id,)) return statuses.get('state', {}).get('terminated') def is_pod_running(k8s_manager, pod_id, container_id): event = k8s_manager.k8s_api.read_namespaced_pod_status(pod_id, k8s_manager.namespace) is_terminated = is_container_terminated(event=event, container_id=container_id) return ( event.status.phase in {PodLifeCycle.RUNNING, PodLifeCycle.PENDING, PodLifeCycle.CONTAINER_CREATING}and not is_terminated )
import ocular from polyaxon_schemas.pod import PodLifeCycle def is_container_terminated(event, container_id): container_statuses = event.status.to_dict().get('container_statuses') or [] statuses_by_name = ocular.processor.get_container_statuses_by_name(container_statuses) statuses = ocular.processor.get_container_status(statuses_by_name, (container_id,)) statuses = statuses or {} return statuses.get('state', {}).get('terminated') def is_pod_running(k8s_manager, pod_id, container_id): event = k8s_manager.k8s_api.read_namespaced_pod_status(pod_id, k8s_manager.namespace) is_terminated = is_container_terminated(event=event, container_id=container_id) return ( event.status.phase in {PodLifeCycle.RUNNING, PodLifeCycle.PENDING, PodLifeCycle.CONTAINER_CREATING} and not is_terminated )
Fix sidecar check for terminated containers
Fix sidecar check for terminated containers
Python
apache-2.0
polyaxon/polyaxon,polyaxon/polyaxon,polyaxon/polyaxon
import ocular from polyaxon_schemas.pod import PodLifeCycle def is_container_terminated(event, container_id): statuses_by_name = ocular.processor.get_container_statuses_by_name( event.status.to_dict().get('container_statuses', [])) statuses = ocular.processor.get_container_status(statuses_by_name, (container_id,)) return statuses.get('state', {}).get('terminated') def is_pod_running(k8s_manager, pod_id, container_id): event = k8s_manager.k8s_api.read_namespaced_pod_status(pod_id, k8s_manager.namespace) is_terminated = is_container_terminated(event=event, container_id=container_id) return ( event.status.phase in {PodLifeCycle.RUNNING, PodLifeCycle.PENDING, PodLifeCycle.CONTAINER_CREATING}and not is_terminated ) Fix sidecar check for terminated containers
import ocular from polyaxon_schemas.pod import PodLifeCycle def is_container_terminated(event, container_id): container_statuses = event.status.to_dict().get('container_statuses') or [] statuses_by_name = ocular.processor.get_container_statuses_by_name(container_statuses) statuses = ocular.processor.get_container_status(statuses_by_name, (container_id,)) statuses = statuses or {} return statuses.get('state', {}).get('terminated') def is_pod_running(k8s_manager, pod_id, container_id): event = k8s_manager.k8s_api.read_namespaced_pod_status(pod_id, k8s_manager.namespace) is_terminated = is_container_terminated(event=event, container_id=container_id) return ( event.status.phase in {PodLifeCycle.RUNNING, PodLifeCycle.PENDING, PodLifeCycle.CONTAINER_CREATING} and not is_terminated )
<commit_before>import ocular from polyaxon_schemas.pod import PodLifeCycle def is_container_terminated(event, container_id): statuses_by_name = ocular.processor.get_container_statuses_by_name( event.status.to_dict().get('container_statuses', [])) statuses = ocular.processor.get_container_status(statuses_by_name, (container_id,)) return statuses.get('state', {}).get('terminated') def is_pod_running(k8s_manager, pod_id, container_id): event = k8s_manager.k8s_api.read_namespaced_pod_status(pod_id, k8s_manager.namespace) is_terminated = is_container_terminated(event=event, container_id=container_id) return ( event.status.phase in {PodLifeCycle.RUNNING, PodLifeCycle.PENDING, PodLifeCycle.CONTAINER_CREATING}and not is_terminated ) <commit_msg>Fix sidecar check for terminated containers<commit_after>
import ocular from polyaxon_schemas.pod import PodLifeCycle def is_container_terminated(event, container_id): container_statuses = event.status.to_dict().get('container_statuses') or [] statuses_by_name = ocular.processor.get_container_statuses_by_name(container_statuses) statuses = ocular.processor.get_container_status(statuses_by_name, (container_id,)) statuses = statuses or {} return statuses.get('state', {}).get('terminated') def is_pod_running(k8s_manager, pod_id, container_id): event = k8s_manager.k8s_api.read_namespaced_pod_status(pod_id, k8s_manager.namespace) is_terminated = is_container_terminated(event=event, container_id=container_id) return ( event.status.phase in {PodLifeCycle.RUNNING, PodLifeCycle.PENDING, PodLifeCycle.CONTAINER_CREATING} and not is_terminated )
import ocular from polyaxon_schemas.pod import PodLifeCycle def is_container_terminated(event, container_id): statuses_by_name = ocular.processor.get_container_statuses_by_name( event.status.to_dict().get('container_statuses', [])) statuses = ocular.processor.get_container_status(statuses_by_name, (container_id,)) return statuses.get('state', {}).get('terminated') def is_pod_running(k8s_manager, pod_id, container_id): event = k8s_manager.k8s_api.read_namespaced_pod_status(pod_id, k8s_manager.namespace) is_terminated = is_container_terminated(event=event, container_id=container_id) return ( event.status.phase in {PodLifeCycle.RUNNING, PodLifeCycle.PENDING, PodLifeCycle.CONTAINER_CREATING}and not is_terminated ) Fix sidecar check for terminated containersimport ocular from polyaxon_schemas.pod import PodLifeCycle def is_container_terminated(event, container_id): container_statuses = event.status.to_dict().get('container_statuses') or [] statuses_by_name = ocular.processor.get_container_statuses_by_name(container_statuses) statuses = ocular.processor.get_container_status(statuses_by_name, (container_id,)) statuses = statuses or {} return statuses.get('state', {}).get('terminated') def is_pod_running(k8s_manager, pod_id, container_id): event = k8s_manager.k8s_api.read_namespaced_pod_status(pod_id, k8s_manager.namespace) is_terminated = is_container_terminated(event=event, container_id=container_id) return ( event.status.phase in {PodLifeCycle.RUNNING, PodLifeCycle.PENDING, PodLifeCycle.CONTAINER_CREATING} and not is_terminated )
<commit_before>import ocular from polyaxon_schemas.pod import PodLifeCycle def is_container_terminated(event, container_id): statuses_by_name = ocular.processor.get_container_statuses_by_name( event.status.to_dict().get('container_statuses', [])) statuses = ocular.processor.get_container_status(statuses_by_name, (container_id,)) return statuses.get('state', {}).get('terminated') def is_pod_running(k8s_manager, pod_id, container_id): event = k8s_manager.k8s_api.read_namespaced_pod_status(pod_id, k8s_manager.namespace) is_terminated = is_container_terminated(event=event, container_id=container_id) return ( event.status.phase in {PodLifeCycle.RUNNING, PodLifeCycle.PENDING, PodLifeCycle.CONTAINER_CREATING}and not is_terminated ) <commit_msg>Fix sidecar check for terminated containers<commit_after>import ocular from polyaxon_schemas.pod import PodLifeCycle def is_container_terminated(event, container_id): container_statuses = event.status.to_dict().get('container_statuses') or [] statuses_by_name = ocular.processor.get_container_statuses_by_name(container_statuses) statuses = ocular.processor.get_container_status(statuses_by_name, (container_id,)) statuses = statuses or {} return statuses.get('state', {}).get('terminated') def is_pod_running(k8s_manager, pod_id, container_id): event = k8s_manager.k8s_api.read_namespaced_pod_status(pod_id, k8s_manager.namespace) is_terminated = is_container_terminated(event=event, container_id=container_id) return ( event.status.phase in {PodLifeCycle.RUNNING, PodLifeCycle.PENDING, PodLifeCycle.CONTAINER_CREATING} and not is_terminated )
d387ab6335ba73a0ecbc1ffa55e9b35ff119bd58
journal/views.py
journal/views.py
import json from django.utils.decorators import method_decorator from django.views.decorators.csrf import csrf_exempt from django.views.generic import View from django.http import JsonResponse from .models import Entry @method_decorator(csrf_exempt, name='dispatch') class RestView(View): def get(self, request): last = request.GET.get('last', None) if last is None: entries = Entry.objects.all() else: entries = Entry.objects.filter(id__gt=last) ret = map(lambda x: {'id': x.id, 'content': x.content.decode()}, entries) return JsonResponse({'entries': list(ret)}) @csrf_exempt def put(self, request): entries = json.loads(request.body.decode()) for entry in entries: Entry(content=entry.encode()).save() res = JsonResponse({'ok': 1}) res.status_code = 201 return res
import json from django.utils.decorators import method_decorator from django.views.decorators.csrf import csrf_exempt from django.views.generic import View from django.http import JsonResponse from .models import Entry @method_decorator(csrf_exempt, name='dispatch') class RestView(View): def get(self, request): last = request.GET.get('last', None) if last is None: entries = Entry.objects.all() else: entries = Entry.objects.filter(id__gt=last) ret = map(lambda x: {'id': x.id, 'content': x.content.decode()}, entries) return JsonResponse({'entries': list(ret)}) @csrf_exempt def put(self, request): entries = json.loads(request.body.decode()) for entry in entries['entries']: Entry(content=entry['content'].encode()).save() res = JsonResponse({'ok': 1}) res.status_code = 201 return res
Update the protocol to mirror the return result of get.
Update the protocol to mirror the return result of get.
Python
agpl-3.0
etesync/journal-manager
import json from django.utils.decorators import method_decorator from django.views.decorators.csrf import csrf_exempt from django.views.generic import View from django.http import JsonResponse from .models import Entry @method_decorator(csrf_exempt, name='dispatch') class RestView(View): def get(self, request): last = request.GET.get('last', None) if last is None: entries = Entry.objects.all() else: entries = Entry.objects.filter(id__gt=last) ret = map(lambda x: {'id': x.id, 'content': x.content.decode()}, entries) return JsonResponse({'entries': list(ret)}) @csrf_exempt def put(self, request): entries = json.loads(request.body.decode()) for entry in entries: Entry(content=entry.encode()).save() res = JsonResponse({'ok': 1}) res.status_code = 201 return res Update the protocol to mirror the return result of get.
import json from django.utils.decorators import method_decorator from django.views.decorators.csrf import csrf_exempt from django.views.generic import View from django.http import JsonResponse from .models import Entry @method_decorator(csrf_exempt, name='dispatch') class RestView(View): def get(self, request): last = request.GET.get('last', None) if last is None: entries = Entry.objects.all() else: entries = Entry.objects.filter(id__gt=last) ret = map(lambda x: {'id': x.id, 'content': x.content.decode()}, entries) return JsonResponse({'entries': list(ret)}) @csrf_exempt def put(self, request): entries = json.loads(request.body.decode()) for entry in entries['entries']: Entry(content=entry['content'].encode()).save() res = JsonResponse({'ok': 1}) res.status_code = 201 return res
<commit_before>import json from django.utils.decorators import method_decorator from django.views.decorators.csrf import csrf_exempt from django.views.generic import View from django.http import JsonResponse from .models import Entry @method_decorator(csrf_exempt, name='dispatch') class RestView(View): def get(self, request): last = request.GET.get('last', None) if last is None: entries = Entry.objects.all() else: entries = Entry.objects.filter(id__gt=last) ret = map(lambda x: {'id': x.id, 'content': x.content.decode()}, entries) return JsonResponse({'entries': list(ret)}) @csrf_exempt def put(self, request): entries = json.loads(request.body.decode()) for entry in entries: Entry(content=entry.encode()).save() res = JsonResponse({'ok': 1}) res.status_code = 201 return res <commit_msg>Update the protocol to mirror the return result of get.<commit_after>
import json from django.utils.decorators import method_decorator from django.views.decorators.csrf import csrf_exempt from django.views.generic import View from django.http import JsonResponse from .models import Entry @method_decorator(csrf_exempt, name='dispatch') class RestView(View): def get(self, request): last = request.GET.get('last', None) if last is None: entries = Entry.objects.all() else: entries = Entry.objects.filter(id__gt=last) ret = map(lambda x: {'id': x.id, 'content': x.content.decode()}, entries) return JsonResponse({'entries': list(ret)}) @csrf_exempt def put(self, request): entries = json.loads(request.body.decode()) for entry in entries['entries']: Entry(content=entry['content'].encode()).save() res = JsonResponse({'ok': 1}) res.status_code = 201 return res
import json from django.utils.decorators import method_decorator from django.views.decorators.csrf import csrf_exempt from django.views.generic import View from django.http import JsonResponse from .models import Entry @method_decorator(csrf_exempt, name='dispatch') class RestView(View): def get(self, request): last = request.GET.get('last', None) if last is None: entries = Entry.objects.all() else: entries = Entry.objects.filter(id__gt=last) ret = map(lambda x: {'id': x.id, 'content': x.content.decode()}, entries) return JsonResponse({'entries': list(ret)}) @csrf_exempt def put(self, request): entries = json.loads(request.body.decode()) for entry in entries: Entry(content=entry.encode()).save() res = JsonResponse({'ok': 1}) res.status_code = 201 return res Update the protocol to mirror the return result of get.import json from django.utils.decorators import method_decorator from django.views.decorators.csrf import csrf_exempt from django.views.generic import View from django.http import JsonResponse from .models import Entry @method_decorator(csrf_exempt, name='dispatch') class RestView(View): def get(self, request): last = request.GET.get('last', None) if last is None: entries = Entry.objects.all() else: entries = Entry.objects.filter(id__gt=last) ret = map(lambda x: {'id': x.id, 'content': x.content.decode()}, entries) return JsonResponse({'entries': list(ret)}) @csrf_exempt def put(self, request): entries = json.loads(request.body.decode()) for entry in entries['entries']: Entry(content=entry['content'].encode()).save() res = JsonResponse({'ok': 1}) res.status_code = 201 return res
<commit_before>import json from django.utils.decorators import method_decorator from django.views.decorators.csrf import csrf_exempt from django.views.generic import View from django.http import JsonResponse from .models import Entry @method_decorator(csrf_exempt, name='dispatch') class RestView(View): def get(self, request): last = request.GET.get('last', None) if last is None: entries = Entry.objects.all() else: entries = Entry.objects.filter(id__gt=last) ret = map(lambda x: {'id': x.id, 'content': x.content.decode()}, entries) return JsonResponse({'entries': list(ret)}) @csrf_exempt def put(self, request): entries = json.loads(request.body.decode()) for entry in entries: Entry(content=entry.encode()).save() res = JsonResponse({'ok': 1}) res.status_code = 201 return res <commit_msg>Update the protocol to mirror the return result of get.<commit_after>import json from django.utils.decorators import method_decorator from django.views.decorators.csrf import csrf_exempt from django.views.generic import View from django.http import JsonResponse from .models import Entry @method_decorator(csrf_exempt, name='dispatch') class RestView(View): def get(self, request): last = request.GET.get('last', None) if last is None: entries = Entry.objects.all() else: entries = Entry.objects.filter(id__gt=last) ret = map(lambda x: {'id': x.id, 'content': x.content.decode()}, entries) return JsonResponse({'entries': list(ret)}) @csrf_exempt def put(self, request): entries = json.loads(request.body.decode()) for entry in entries['entries']: Entry(content=entry['content'].encode()).save() res = JsonResponse({'ok': 1}) res.status_code = 201 return res
c87c4a972f0f2d4966142fa666a900112762ed76
scipy/constants/tests/test_codata.py
scipy/constants/tests/test_codata.py
import warnings from scipy.constants import find from numpy.testing import assert_equal def test_find(): warnings.simplefilter('ignore', DeprecationWarning) keys = find('weak mixing', disp=False) assert_equal(keys, ['weak mixing angle']) keys = find('qwertyuiop', disp=False) assert_equal(keys, []) keys = find('natural unit', disp=False) assert_equal(keys, sorted(['natural unit of velocity', 'natural unit of action', 'natural unit of action in eV s', 'natural unit of mass', 'natural unit of energy', 'natural unit of energy in MeV', 'natural unit of momentum', 'natural unit of momentum in MeV/c', 'natural unit of length', 'natural unit of time']))
import warnings from scipy.constants import find from numpy.testing import assert_equal, run_module_suite def test_find(): warnings.simplefilter('ignore', DeprecationWarning) keys = find('weak mixing', disp=False) assert_equal(keys, ['weak mixing angle']) keys = find('qwertyuiop', disp=False) assert_equal(keys, []) keys = find('natural unit', disp=False) assert_equal(keys, sorted(['natural unit of velocity', 'natural unit of action', 'natural unit of action in eV s', 'natural unit of mass', 'natural unit of energy', 'natural unit of energy in MeV', 'natural unit of momentum', 'natural unit of momentum in MeV/c', 'natural unit of length', 'natural unit of time'])) if __name__ == "__main__": run_module_suite()
Allow codata tests to be run as script.
ENH: Allow codata tests to be run as script.
Python
bsd-3-clause
zerothi/scipy,zxsted/scipy,josephcslater/scipy,rgommers/scipy,grlee77/scipy,sargas/scipy,dch312/scipy,ilayn/scipy,apbard/scipy,jakevdp/scipy,niknow/scipy,vanpact/scipy,jakevdp/scipy,rmcgibbo/scipy,zxsted/scipy,pnedunuri/scipy,raoulbq/scipy,lhilt/scipy,mgaitan/scipy,mingwpy/scipy,maciejkula/scipy,njwilson23/scipy,Dapid/scipy,woodscn/scipy,perimosocordiae/scipy,aarchiba/scipy,pbrod/scipy,aarchiba/scipy,trankmichael/scipy,Srisai85/scipy,aarchiba/scipy,pschella/scipy,tylerjereddy/scipy,minhlongdo/scipy,mortonjt/scipy,dominicelse/scipy,jsilter/scipy,sargas/scipy,vanpact/scipy,anielsen001/scipy,mingwpy/scipy,matthew-brett/scipy,aeklant/scipy,mikebenfield/scipy,Srisai85/scipy,sauliusl/scipy,FRidh/scipy,andim/scipy,ortylp/scipy,josephcslater/scipy,mhogg/scipy,WarrenWeckesser/scipy,ndchorley/scipy,josephcslater/scipy,mortonjt/scipy,felipebetancur/scipy,jonycgn/scipy,mtrbean/scipy,juliantaylor/scipy,mdhaber/scipy,mdhaber/scipy,kleskjr/scipy,ogrisel/scipy,teoliphant/scipy,petebachant/scipy,lukauskas/scipy,Gillu13/scipy,sonnyhu/scipy,WillieMaddox/scipy,arokem/scipy,jonycgn/scipy,gdooper/scipy,ChanderG/scipy,pbrod/scipy,surhudm/scipy,rgommers/scipy,anntzer/scipy,gfyoung/scipy,sriki18/scipy,nmayorov/scipy,ogrisel/scipy,nvoron23/scipy,giorgiop/scipy,trankmichael/scipy,scipy/scipy,dominicelse/scipy,petebachant/scipy,ChanderG/scipy,ndchorley/scipy,scipy/scipy,Stefan-Endres/scipy,futurulus/scipy,giorgiop/scipy,behzadnouri/scipy,sargas/scipy,haudren/scipy,Gillu13/scipy,teoliphant/scipy,WillieMaddox/scipy,cpaulik/scipy,juliantaylor/scipy,sriki18/scipy,ortylp/scipy,zerothi/scipy,gdooper/scipy,befelix/scipy,raoulbq/scipy,sargas/scipy,newemailjdm/scipy,piyush0609/scipy,vhaasteren/scipy,sriki18/scipy,sauliusl/scipy,rgommers/scipy,zaxliu/scipy,Stefan-Endres/scipy,dch312/scipy,scipy/scipy,aeklant/scipy,cpaulik/scipy,maniteja123/scipy,zerothi/scipy,cpaulik/scipy,Eric89GXL/scipy,mgaitan/scipy,bkendzior/scipy,jonycgn/scipy,juliantaylor/scipy,felipebetancur/scipy,giorgiop/scipy,gertingold/scipy,befelix/scipy,maciejkula/scipy,vberaudi/scipy,lhilt/scipy,ChanderG/scipy,mgaitan/scipy,hainm/scipy,endolith/scipy,lukauskas/scipy,gdooper/scipy,kalvdans/scipy,richardotis/scipy,jor-/scipy,mgaitan/scipy,mtrbean/scipy,Gillu13/scipy,raoulbq/scipy,lukauskas/scipy,tylerjereddy/scipy,person142/scipy,mortada/scipy,vhaasteren/scipy,Shaswat27/scipy,vigna/scipy,pschella/scipy,surhudm/scipy,kalvdans/scipy,witcxc/scipy,piyush0609/scipy,woodscn/scipy,matthew-brett/scipy,nvoron23/scipy,matthewalbani/scipy,mdhaber/scipy,gfyoung/scipy,pnedunuri/scipy,vanpact/scipy,ilayn/scipy,FRidh/scipy,mingwpy/scipy,efiring/scipy,jakevdp/scipy,Shaswat27/scipy,matthewalbani/scipy,pizzathief/scipy,mdhaber/scipy,jonycgn/scipy,pyramania/scipy,apbard/scipy,Dapid/scipy,rmcgibbo/scipy,vhaasteren/scipy,niknow/scipy,apbard/scipy,sonnyhu/scipy,larsmans/scipy,bkendzior/scipy,perimosocordiae/scipy,maniteja123/scipy,mortada/scipy,felipebetancur/scipy,andyfaff/scipy,mikebenfield/scipy,jseabold/scipy,mtrbean/scipy,jamestwebber/scipy,anielsen001/scipy,ortylp/scipy,Gillu13/scipy,woodscn/scipy,witcxc/scipy,haudren/scipy,witcxc/scipy,andyfaff/scipy,newemailjdm/scipy,vhaasteren/scipy,njwilson23/scipy,grlee77/scipy,surhudm/scipy,zaxliu/scipy,ortylp/scipy,njwilson23/scipy,dch312/scipy,fredrikw/scipy,pnedunuri/scipy,mortada/scipy,mortada/scipy,fredrikw/scipy,gef756/scipy,zaxliu/scipy,chatcannon/scipy,matthew-brett/scipy,jsilter/scipy,ilayn/scipy,mortada/scipy,Stefan-Endres/scipy,jseabold/scipy,minhlongdo/scipy,anntzer/scipy,andyfaff/scipy,futurulus/scipy,sauliusl/scipy,e-q/scipy,mingwpy/scipy,rmcgibbo/scipy,Dapid/scipy,vigna/scipy,minhlongdo/scipy,mdhaber/scipy,raoulbq/scipy,pyramania/scipy,richardotis/scipy,jor-/scipy,befelix/scipy,sriki18/scipy,mortonjt/scipy,endolith/scipy,haudren/scipy,juliantaylor/scipy,hainm/scipy,e-q/scipy,pizzathief/scipy,endolith/scipy,FRidh/scipy,argriffing/scipy,perimosocordiae/scipy,tylerjereddy/scipy,fernand/scipy,tylerjereddy/scipy,argriffing/scipy,vigna/scipy,mhogg/scipy,vanpact/scipy,woodscn/scipy,Eric89GXL/scipy,vberaudi/scipy,ChanderG/scipy,raoulbq/scipy,gef756/scipy,sauliusl/scipy,anntzer/scipy,jseabold/scipy,pbrod/scipy,argriffing/scipy,nvoron23/scipy,ales-erjavec/scipy,ortylp/scipy,Newman101/scipy,jjhelmus/scipy,petebachant/scipy,mingwpy/scipy,chatcannon/scipy,Eric89GXL/scipy,Shaswat27/scipy,ogrisel/scipy,anntzer/scipy,andyfaff/scipy,aman-iitj/scipy,cpaulik/scipy,sauliusl/scipy,zerothi/scipy,hainm/scipy,Srisai85/scipy,scipy/scipy,mtrbean/scipy,fredrikw/scipy,endolith/scipy,Dapid/scipy,ndchorley/scipy,giorgiop/scipy,felipebetancur/scipy,Gillu13/scipy,maciejkula/scipy,jonycgn/scipy,vanpact/scipy,fernand/scipy,argriffing/scipy,aarchiba/scipy,jjhelmus/scipy,newemailjdm/scipy,matthewalbani/scipy,behzadnouri/scipy,jamestwebber/scipy,behzadnouri/scipy,andim/scipy,pschella/scipy,nvoron23/scipy,ndchorley/scipy,Shaswat27/scipy,argriffing/scipy,andyfaff/scipy,juliantaylor/scipy,felipebetancur/scipy,mtrbean/scipy,fredrikw/scipy,sonnyhu/scipy,efiring/scipy,e-q/scipy,jsilter/scipy,mtrbean/scipy,behzadnouri/scipy,grlee77/scipy,ChanderG/scipy,maciejkula/scipy,ogrisel/scipy,maciejkula/scipy,Srisai85/scipy,jamestwebber/scipy,woodscn/scipy,pschella/scipy,vigna/scipy,dch312/scipy,lhilt/scipy,haudren/scipy,nmayorov/scipy,futurulus/scipy,Newman101/scipy,larsmans/scipy,Eric89GXL/scipy,larsmans/scipy,maniteja123/scipy,maniteja123/scipy,gfyoung/scipy,fernand/scipy,sargas/scipy,nmayorov/scipy,lhilt/scipy,richardotis/scipy,vanpact/scipy,teoliphant/scipy,arokem/scipy,aeklant/scipy,pyramania/scipy,jamestwebber/scipy,vberaudi/scipy,dominicelse/scipy,efiring/scipy,rmcgibbo/scipy,ales-erjavec/scipy,Eric89GXL/scipy,niknow/scipy,zaxliu/scipy,surhudm/scipy,maniteja123/scipy,mikebenfield/scipy,gef756/scipy,vberaudi/scipy,efiring/scipy,sriki18/scipy,FRidh/scipy,gdooper/scipy,futurulus/scipy,cpaulik/scipy,WarrenWeckesser/scipy,larsmans/scipy,person142/scipy,teoliphant/scipy,kleskjr/scipy,befelix/scipy,arokem/scipy,Kamp9/scipy,kleskjr/scipy,teoliphant/scipy,kleskjr/scipy,dominicelse/scipy,ndchorley/scipy,anntzer/scipy,jjhelmus/scipy,pbrod/scipy,ilayn/scipy,Shaswat27/scipy,efiring/scipy,njwilson23/scipy,niknow/scipy,dominicelse/scipy,behzadnouri/scipy,mingwpy/scipy,rgommers/scipy,FRidh/scipy,person142/scipy,mortonjt/scipy,vhaasteren/scipy,pbrod/scipy,nvoron23/scipy,jor-/scipy,niknow/scipy,trankmichael/scipy,fernand/scipy,haudren/scipy,njwilson23/scipy,matthew-brett/scipy,nvoron23/scipy,endolith/scipy,e-q/scipy,jonycgn/scipy,dch312/scipy,WillieMaddox/scipy,rmcgibbo/scipy,lukauskas/scipy,andim/scipy,jakevdp/scipy,zerothi/scipy,witcxc/scipy,ales-erjavec/scipy,ales-erjavec/scipy,jor-/scipy,grlee77/scipy,nonhermitian/scipy,aman-iitj/scipy,vhaasteren/scipy,chatcannon/scipy,arokem/scipy,giorgiop/scipy,futurulus/scipy,e-q/scipy,ChanderG/scipy,ilayn/scipy,WarrenWeckesser/scipy,mikebenfield/scipy,Newman101/scipy,richardotis/scipy,jsilter/scipy,perimosocordiae/scipy,gfyoung/scipy,scipy/scipy,kalvdans/scipy,sonnyhu/scipy,lukauskas/scipy,WillieMaddox/scipy,aman-iitj/scipy,Newman101/scipy,zerothi/scipy,apbard/scipy,jsilter/scipy,hainm/scipy,Stefan-Endres/scipy,maniteja123/scipy,petebachant/scipy,efiring/scipy,mhogg/scipy,Dapid/scipy,nmayorov/scipy,anielsen001/scipy,Dapid/scipy,aeklant/scipy,nonhermitian/scipy,perimosocordiae/scipy,gertingold/scipy,chatcannon/scipy,zxsted/scipy,trankmichael/scipy,behzadnouri/scipy,WillieMaddox/scipy,newemailjdm/scipy,bkendzior/scipy,vberaudi/scipy,anielsen001/scipy,argriffing/scipy,minhlongdo/scipy,aman-iitj/scipy,futurulus/scipy,pizzathief/scipy,aeklant/scipy,anielsen001/scipy,cpaulik/scipy,njwilson23/scipy,zaxliu/scipy,surhudm/scipy,Kamp9/scipy,Eric89GXL/scipy,mhogg/scipy,scipy/scipy,gfyoung/scipy,befelix/scipy,Newman101/scipy,ales-erjavec/scipy,raoulbq/scipy,zaxliu/scipy,pyramania/scipy,larsmans/scipy,mortada/scipy,chatcannon/scipy,grlee77/scipy,person142/scipy,Kamp9/scipy,giorgiop/scipy,kalvdans/scipy,Stefan-Endres/scipy,mhogg/scipy,mortonjt/scipy,Kamp9/scipy,zxsted/scipy,hainm/scipy,lhilt/scipy,haudren/scipy,sauliusl/scipy,aman-iitj/scipy,aarchiba/scipy,kleskjr/scipy,andyfaff/scipy,gef756/scipy,WarrenWeckesser/scipy,fredrikw/scipy,gertingold/scipy,petebachant/scipy,petebachant/scipy,endolith/scipy,Newman101/scipy,sonnyhu/scipy,trankmichael/scipy,josephcslater/scipy,vigna/scipy,kleskjr/scipy,matthewalbani/scipy,ortylp/scipy,Gillu13/scipy,gdooper/scipy,Kamp9/scipy,larsmans/scipy,mikebenfield/scipy,matthewalbani/scipy,Shaswat27/scipy,lukauskas/scipy,pizzathief/scipy,josephcslater/scipy,pnedunuri/scipy,matthew-brett/scipy,chatcannon/scipy,bkendzior/scipy,ilayn/scipy,jjhelmus/scipy,perimosocordiae/scipy,apbard/scipy,ogrisel/scipy,anntzer/scipy,rmcgibbo/scipy,surhudm/scipy,mdhaber/scipy,bkendzior/scipy,jseabold/scipy,sriki18/scipy,arokem/scipy,kalvdans/scipy,Kamp9/scipy,Srisai85/scipy,pnedunuri/scipy,pyramania/scipy,andim/scipy,fernand/scipy,gef756/scipy,trankmichael/scipy,nonhermitian/scipy,pnedunuri/scipy,niknow/scipy,piyush0609/scipy,richardotis/scipy,jamestwebber/scipy,Stefan-Endres/scipy,woodscn/scipy,minhlongdo/scipy,fredrikw/scipy,newemailjdm/scipy,WarrenWeckesser/scipy,ales-erjavec/scipy,andim/scipy,richardotis/scipy,piyush0609/scipy,andim/scipy,jor-/scipy,FRidh/scipy,nonhermitian/scipy,tylerjereddy/scipy,jseabold/scipy,gertingold/scipy,pizzathief/scipy,jseabold/scipy,zxsted/scipy,rgommers/scipy,gertingold/scipy,WarrenWeckesser/scipy,jakevdp/scipy,mortonjt/scipy,hainm/scipy,jjhelmus/scipy,piyush0609/scipy,WillieMaddox/scipy,ndchorley/scipy,zxsted/scipy,mgaitan/scipy,person142/scipy,minhlongdo/scipy,fernand/scipy,gef756/scipy,pbrod/scipy,pschella/scipy,anielsen001/scipy,vberaudi/scipy,aman-iitj/scipy,mhogg/scipy,witcxc/scipy,sonnyhu/scipy,piyush0609/scipy,mgaitan/scipy,nmayorov/scipy,newemailjdm/scipy,felipebetancur/scipy,Srisai85/scipy,nonhermitian/scipy
import warnings from scipy.constants import find from numpy.testing import assert_equal def test_find(): warnings.simplefilter('ignore', DeprecationWarning) keys = find('weak mixing', disp=False) assert_equal(keys, ['weak mixing angle']) keys = find('qwertyuiop', disp=False) assert_equal(keys, []) keys = find('natural unit', disp=False) assert_equal(keys, sorted(['natural unit of velocity', 'natural unit of action', 'natural unit of action in eV s', 'natural unit of mass', 'natural unit of energy', 'natural unit of energy in MeV', 'natural unit of momentum', 'natural unit of momentum in MeV/c', 'natural unit of length', 'natural unit of time'])) ENH: Allow codata tests to be run as script.
import warnings from scipy.constants import find from numpy.testing import assert_equal, run_module_suite def test_find(): warnings.simplefilter('ignore', DeprecationWarning) keys = find('weak mixing', disp=False) assert_equal(keys, ['weak mixing angle']) keys = find('qwertyuiop', disp=False) assert_equal(keys, []) keys = find('natural unit', disp=False) assert_equal(keys, sorted(['natural unit of velocity', 'natural unit of action', 'natural unit of action in eV s', 'natural unit of mass', 'natural unit of energy', 'natural unit of energy in MeV', 'natural unit of momentum', 'natural unit of momentum in MeV/c', 'natural unit of length', 'natural unit of time'])) if __name__ == "__main__": run_module_suite()
<commit_before> import warnings from scipy.constants import find from numpy.testing import assert_equal def test_find(): warnings.simplefilter('ignore', DeprecationWarning) keys = find('weak mixing', disp=False) assert_equal(keys, ['weak mixing angle']) keys = find('qwertyuiop', disp=False) assert_equal(keys, []) keys = find('natural unit', disp=False) assert_equal(keys, sorted(['natural unit of velocity', 'natural unit of action', 'natural unit of action in eV s', 'natural unit of mass', 'natural unit of energy', 'natural unit of energy in MeV', 'natural unit of momentum', 'natural unit of momentum in MeV/c', 'natural unit of length', 'natural unit of time'])) <commit_msg>ENH: Allow codata tests to be run as script.<commit_after>
import warnings from scipy.constants import find from numpy.testing import assert_equal, run_module_suite def test_find(): warnings.simplefilter('ignore', DeprecationWarning) keys = find('weak mixing', disp=False) assert_equal(keys, ['weak mixing angle']) keys = find('qwertyuiop', disp=False) assert_equal(keys, []) keys = find('natural unit', disp=False) assert_equal(keys, sorted(['natural unit of velocity', 'natural unit of action', 'natural unit of action in eV s', 'natural unit of mass', 'natural unit of energy', 'natural unit of energy in MeV', 'natural unit of momentum', 'natural unit of momentum in MeV/c', 'natural unit of length', 'natural unit of time'])) if __name__ == "__main__": run_module_suite()
import warnings from scipy.constants import find from numpy.testing import assert_equal def test_find(): warnings.simplefilter('ignore', DeprecationWarning) keys = find('weak mixing', disp=False) assert_equal(keys, ['weak mixing angle']) keys = find('qwertyuiop', disp=False) assert_equal(keys, []) keys = find('natural unit', disp=False) assert_equal(keys, sorted(['natural unit of velocity', 'natural unit of action', 'natural unit of action in eV s', 'natural unit of mass', 'natural unit of energy', 'natural unit of energy in MeV', 'natural unit of momentum', 'natural unit of momentum in MeV/c', 'natural unit of length', 'natural unit of time'])) ENH: Allow codata tests to be run as script. import warnings from scipy.constants import find from numpy.testing import assert_equal, run_module_suite def test_find(): warnings.simplefilter('ignore', DeprecationWarning) keys = find('weak mixing', disp=False) assert_equal(keys, ['weak mixing angle']) keys = find('qwertyuiop', disp=False) assert_equal(keys, []) keys = find('natural unit', disp=False) assert_equal(keys, sorted(['natural unit of velocity', 'natural unit of action', 'natural unit of action in eV s', 'natural unit of mass', 'natural unit of energy', 'natural unit of energy in MeV', 'natural unit of momentum', 'natural unit of momentum in MeV/c', 'natural unit of length', 'natural unit of time'])) if __name__ == "__main__": run_module_suite()
<commit_before> import warnings from scipy.constants import find from numpy.testing import assert_equal def test_find(): warnings.simplefilter('ignore', DeprecationWarning) keys = find('weak mixing', disp=False) assert_equal(keys, ['weak mixing angle']) keys = find('qwertyuiop', disp=False) assert_equal(keys, []) keys = find('natural unit', disp=False) assert_equal(keys, sorted(['natural unit of velocity', 'natural unit of action', 'natural unit of action in eV s', 'natural unit of mass', 'natural unit of energy', 'natural unit of energy in MeV', 'natural unit of momentum', 'natural unit of momentum in MeV/c', 'natural unit of length', 'natural unit of time'])) <commit_msg>ENH: Allow codata tests to be run as script.<commit_after> import warnings from scipy.constants import find from numpy.testing import assert_equal, run_module_suite def test_find(): warnings.simplefilter('ignore', DeprecationWarning) keys = find('weak mixing', disp=False) assert_equal(keys, ['weak mixing angle']) keys = find('qwertyuiop', disp=False) assert_equal(keys, []) keys = find('natural unit', disp=False) assert_equal(keys, sorted(['natural unit of velocity', 'natural unit of action', 'natural unit of action in eV s', 'natural unit of mass', 'natural unit of energy', 'natural unit of energy in MeV', 'natural unit of momentum', 'natural unit of momentum in MeV/c', 'natural unit of length', 'natural unit of time'])) if __name__ == "__main__": run_module_suite()
3a2f4940ff83d3d2645505b82d1207a96f6d209e
linked-list/is-list-palindrome.py
linked-list/is-list-palindrome.py
# Given a singly linked list of integers, determine whether or not it's a palindrome class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None def is_list_palindrome(l): if l.value is None: return True # find center of list fast = l slow = l while fast.next and fast.next.next: fast = fast.next.next slow = slow.next # reverse second half of list p = slow.next current_node = None while p: next = p.next p.next = current_node current_node = p p = next def create_nodes(l): root = Node(-1) current_node = root for value in l: current_node.next = Node(value) current_node = current_node.next return root.next is_list_palindrome(create_nodes([1, 2, 3, 4]))
# Given a singly linked list of integers, determine whether or not it's a palindrome class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None def is_list_palindrome(l): if l is None or l.next is None: return True # find center of list fast = l slow = l while fast.next and fast.next.next: fast = fast.next.next slow = slow.next # reverse second half of list p = slow.next current_node = None while p: next = p.next p.next = current_node current_node = p p = next # check for palindrome part_one = current_node part_two = l while part_one and part_one.value == part_two.value: part_one = part_one.next part_two = part_two.next return part_one is None def create_nodes(l): root = Node(-1) current_node = root for value in l: current_node.next = Node(value) current_node = current_node.next return root.next print is_list_palindrome(create_nodes([1, 2, 3, 4]))
Add check for palindrome component of method
Add check for palindrome component of method
Python
mit
derekmpham/interview-prep,derekmpham/interview-prep
# Given a singly linked list of integers, determine whether or not it's a palindrome class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None def is_list_palindrome(l): if l.value is None: return True # find center of list fast = l slow = l while fast.next and fast.next.next: fast = fast.next.next slow = slow.next # reverse second half of list p = slow.next current_node = None while p: next = p.next p.next = current_node current_node = p p = next def create_nodes(l): root = Node(-1) current_node = root for value in l: current_node.next = Node(value) current_node = current_node.next return root.next is_list_palindrome(create_nodes([1, 2, 3, 4])) Add check for palindrome component of method
# Given a singly linked list of integers, determine whether or not it's a palindrome class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None def is_list_palindrome(l): if l is None or l.next is None: return True # find center of list fast = l slow = l while fast.next and fast.next.next: fast = fast.next.next slow = slow.next # reverse second half of list p = slow.next current_node = None while p: next = p.next p.next = current_node current_node = p p = next # check for palindrome part_one = current_node part_two = l while part_one and part_one.value == part_two.value: part_one = part_one.next part_two = part_two.next return part_one is None def create_nodes(l): root = Node(-1) current_node = root for value in l: current_node.next = Node(value) current_node = current_node.next return root.next print is_list_palindrome(create_nodes([1, 2, 3, 4]))
<commit_before># Given a singly linked list of integers, determine whether or not it's a palindrome class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None def is_list_palindrome(l): if l.value is None: return True # find center of list fast = l slow = l while fast.next and fast.next.next: fast = fast.next.next slow = slow.next # reverse second half of list p = slow.next current_node = None while p: next = p.next p.next = current_node current_node = p p = next def create_nodes(l): root = Node(-1) current_node = root for value in l: current_node.next = Node(value) current_node = current_node.next return root.next is_list_palindrome(create_nodes([1, 2, 3, 4])) <commit_msg>Add check for palindrome component of method<commit_after>
# Given a singly linked list of integers, determine whether or not it's a palindrome class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None def is_list_palindrome(l): if l is None or l.next is None: return True # find center of list fast = l slow = l while fast.next and fast.next.next: fast = fast.next.next slow = slow.next # reverse second half of list p = slow.next current_node = None while p: next = p.next p.next = current_node current_node = p p = next # check for palindrome part_one = current_node part_two = l while part_one and part_one.value == part_two.value: part_one = part_one.next part_two = part_two.next return part_one is None def create_nodes(l): root = Node(-1) current_node = root for value in l: current_node.next = Node(value) current_node = current_node.next return root.next print is_list_palindrome(create_nodes([1, 2, 3, 4]))
# Given a singly linked list of integers, determine whether or not it's a palindrome class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None def is_list_palindrome(l): if l.value is None: return True # find center of list fast = l slow = l while fast.next and fast.next.next: fast = fast.next.next slow = slow.next # reverse second half of list p = slow.next current_node = None while p: next = p.next p.next = current_node current_node = p p = next def create_nodes(l): root = Node(-1) current_node = root for value in l: current_node.next = Node(value) current_node = current_node.next return root.next is_list_palindrome(create_nodes([1, 2, 3, 4])) Add check for palindrome component of method# Given a singly linked list of integers, determine whether or not it's a palindrome class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None def is_list_palindrome(l): if l is None or l.next is None: return True # find center of list fast = l slow = l while fast.next and fast.next.next: fast = fast.next.next slow = slow.next # reverse second half of list p = slow.next current_node = None while p: next = p.next p.next = current_node current_node = p p = next # check for palindrome part_one = current_node part_two = l while part_one and part_one.value == part_two.value: part_one = part_one.next part_two = part_two.next return part_one is None def create_nodes(l): root = Node(-1) current_node = root for value in l: current_node.next = Node(value) current_node = current_node.next return root.next print is_list_palindrome(create_nodes([1, 2, 3, 4]))
<commit_before># Given a singly linked list of integers, determine whether or not it's a palindrome class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None def is_list_palindrome(l): if l.value is None: return True # find center of list fast = l slow = l while fast.next and fast.next.next: fast = fast.next.next slow = slow.next # reverse second half of list p = slow.next current_node = None while p: next = p.next p.next = current_node current_node = p p = next def create_nodes(l): root = Node(-1) current_node = root for value in l: current_node.next = Node(value) current_node = current_node.next return root.next is_list_palindrome(create_nodes([1, 2, 3, 4])) <commit_msg>Add check for palindrome component of method<commit_after># Given a singly linked list of integers, determine whether or not it's a palindrome class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None def is_list_palindrome(l): if l is None or l.next is None: return True # find center of list fast = l slow = l while fast.next and fast.next.next: fast = fast.next.next slow = slow.next # reverse second half of list p = slow.next current_node = None while p: next = p.next p.next = current_node current_node = p p = next # check for palindrome part_one = current_node part_two = l while part_one and part_one.value == part_two.value: part_one = part_one.next part_two = part_two.next return part_one is None def create_nodes(l): root = Node(-1) current_node = root for value in l: current_node.next = Node(value) current_node = current_node.next return root.next print is_list_palindrome(create_nodes([1, 2, 3, 4]))
b43dfa19979dc74efb27e56771535b102547e792
utils.py
utils.py
import sqlite3 import shelve def connect_db(name): """ Open a connection to the database used to store quotes. :param name: (str) Name of database file :return: (shelve.DbfilenameShelf) """ try: return shelve.open(name) except Exception: raise Exception('Unable to connect to database with name {}'.format(name)) class DBClient(object): """Client for interacting with database for the application""" def __init__(self, database_name): self.conn = sqlite3.connect(database_name) self._create_quotes_table() def _create_quotes_table(self): """ Create the table used for storing quotes if it does not exist already """ with self.conn: self.conn.execute(''' CREATE TABLE IF NOT EXISTS quotes ( author TEXT quote TEXT created_at TEXT ) ''') def close_connection(self): """ Close connection to the database """ self.conn.close()
import sqlite3 import shelve def connect_db(name): """ Open a connection to the database used to store quotes. :param name: (str) Name of database file :return: (shelve.DbfilenameShelf) """ try: return shelve.open(name) except Exception: raise Exception('Unable to connect to database with name {}'.format(name)) class DBClient(object): """Client for interacting with database for the application""" def __init__(self, database_name: str): self.conn = sqlite3.connect(database_name) self._create_quotes_table() def _create_quotes_table(self): """ Create the table used for storing quotes if it does not exist already """ with self.conn: self.conn.execute(''' CREATE TABLE IF NOT EXISTS quotes ( author TEXT, quote TEXT, created_at TEXT ); ''') def close_connection(self): """ Close connection to the database """ self.conn.close() def insert_quote(self, author: str, quote: str, created_at: str): """ Insert a quote into the database :param author: (str) Name of the author that said the quote :param quote: (str) The quote for the author :param created_at: (str) Timestamp for when the quote was saved to database """ with self.conn: self.conn.execute(''' INSERT INTO quotes VALUES (?, ?, ?) ''', (author, quote, created_at))
Add method to insert quotes into database
Add method to insert quotes into database Fix schema for quotes table
Python
mit
nickdibari/Get-Quote
import sqlite3 import shelve def connect_db(name): """ Open a connection to the database used to store quotes. :param name: (str) Name of database file :return: (shelve.DbfilenameShelf) """ try: return shelve.open(name) except Exception: raise Exception('Unable to connect to database with name {}'.format(name)) class DBClient(object): """Client for interacting with database for the application""" def __init__(self, database_name): self.conn = sqlite3.connect(database_name) self._create_quotes_table() def _create_quotes_table(self): """ Create the table used for storing quotes if it does not exist already """ with self.conn: self.conn.execute(''' CREATE TABLE IF NOT EXISTS quotes ( author TEXT quote TEXT created_at TEXT ) ''') def close_connection(self): """ Close connection to the database """ self.conn.close() Add method to insert quotes into database Fix schema for quotes table
import sqlite3 import shelve def connect_db(name): """ Open a connection to the database used to store quotes. :param name: (str) Name of database file :return: (shelve.DbfilenameShelf) """ try: return shelve.open(name) except Exception: raise Exception('Unable to connect to database with name {}'.format(name)) class DBClient(object): """Client for interacting with database for the application""" def __init__(self, database_name: str): self.conn = sqlite3.connect(database_name) self._create_quotes_table() def _create_quotes_table(self): """ Create the table used for storing quotes if it does not exist already """ with self.conn: self.conn.execute(''' CREATE TABLE IF NOT EXISTS quotes ( author TEXT, quote TEXT, created_at TEXT ); ''') def close_connection(self): """ Close connection to the database """ self.conn.close() def insert_quote(self, author: str, quote: str, created_at: str): """ Insert a quote into the database :param author: (str) Name of the author that said the quote :param quote: (str) The quote for the author :param created_at: (str) Timestamp for when the quote was saved to database """ with self.conn: self.conn.execute(''' INSERT INTO quotes VALUES (?, ?, ?) ''', (author, quote, created_at))
<commit_before>import sqlite3 import shelve def connect_db(name): """ Open a connection to the database used to store quotes. :param name: (str) Name of database file :return: (shelve.DbfilenameShelf) """ try: return shelve.open(name) except Exception: raise Exception('Unable to connect to database with name {}'.format(name)) class DBClient(object): """Client for interacting with database for the application""" def __init__(self, database_name): self.conn = sqlite3.connect(database_name) self._create_quotes_table() def _create_quotes_table(self): """ Create the table used for storing quotes if it does not exist already """ with self.conn: self.conn.execute(''' CREATE TABLE IF NOT EXISTS quotes ( author TEXT quote TEXT created_at TEXT ) ''') def close_connection(self): """ Close connection to the database """ self.conn.close() <commit_msg>Add method to insert quotes into database Fix schema for quotes table<commit_after>
import sqlite3 import shelve def connect_db(name): """ Open a connection to the database used to store quotes. :param name: (str) Name of database file :return: (shelve.DbfilenameShelf) """ try: return shelve.open(name) except Exception: raise Exception('Unable to connect to database with name {}'.format(name)) class DBClient(object): """Client for interacting with database for the application""" def __init__(self, database_name: str): self.conn = sqlite3.connect(database_name) self._create_quotes_table() def _create_quotes_table(self): """ Create the table used for storing quotes if it does not exist already """ with self.conn: self.conn.execute(''' CREATE TABLE IF NOT EXISTS quotes ( author TEXT, quote TEXT, created_at TEXT ); ''') def close_connection(self): """ Close connection to the database """ self.conn.close() def insert_quote(self, author: str, quote: str, created_at: str): """ Insert a quote into the database :param author: (str) Name of the author that said the quote :param quote: (str) The quote for the author :param created_at: (str) Timestamp for when the quote was saved to database """ with self.conn: self.conn.execute(''' INSERT INTO quotes VALUES (?, ?, ?) ''', (author, quote, created_at))
import sqlite3 import shelve def connect_db(name): """ Open a connection to the database used to store quotes. :param name: (str) Name of database file :return: (shelve.DbfilenameShelf) """ try: return shelve.open(name) except Exception: raise Exception('Unable to connect to database with name {}'.format(name)) class DBClient(object): """Client for interacting with database for the application""" def __init__(self, database_name): self.conn = sqlite3.connect(database_name) self._create_quotes_table() def _create_quotes_table(self): """ Create the table used for storing quotes if it does not exist already """ with self.conn: self.conn.execute(''' CREATE TABLE IF NOT EXISTS quotes ( author TEXT quote TEXT created_at TEXT ) ''') def close_connection(self): """ Close connection to the database """ self.conn.close() Add method to insert quotes into database Fix schema for quotes tableimport sqlite3 import shelve def connect_db(name): """ Open a connection to the database used to store quotes. :param name: (str) Name of database file :return: (shelve.DbfilenameShelf) """ try: return shelve.open(name) except Exception: raise Exception('Unable to connect to database with name {}'.format(name)) class DBClient(object): """Client for interacting with database for the application""" def __init__(self, database_name: str): self.conn = sqlite3.connect(database_name) self._create_quotes_table() def _create_quotes_table(self): """ Create the table used for storing quotes if it does not exist already """ with self.conn: self.conn.execute(''' CREATE TABLE IF NOT EXISTS quotes ( author TEXT, quote TEXT, created_at TEXT ); ''') def close_connection(self): """ Close connection to the database """ self.conn.close() def insert_quote(self, author: str, quote: str, created_at: str): """ Insert a quote into the database :param author: (str) Name of the author that said the quote :param quote: (str) The quote for the author :param created_at: (str) Timestamp for when the quote was saved to database """ with self.conn: self.conn.execute(''' INSERT INTO quotes VALUES (?, ?, ?) ''', (author, quote, created_at))
<commit_before>import sqlite3 import shelve def connect_db(name): """ Open a connection to the database used to store quotes. :param name: (str) Name of database file :return: (shelve.DbfilenameShelf) """ try: return shelve.open(name) except Exception: raise Exception('Unable to connect to database with name {}'.format(name)) class DBClient(object): """Client for interacting with database for the application""" def __init__(self, database_name): self.conn = sqlite3.connect(database_name) self._create_quotes_table() def _create_quotes_table(self): """ Create the table used for storing quotes if it does not exist already """ with self.conn: self.conn.execute(''' CREATE TABLE IF NOT EXISTS quotes ( author TEXT quote TEXT created_at TEXT ) ''') def close_connection(self): """ Close connection to the database """ self.conn.close() <commit_msg>Add method to insert quotes into database Fix schema for quotes table<commit_after>import sqlite3 import shelve def connect_db(name): """ Open a connection to the database used to store quotes. :param name: (str) Name of database file :return: (shelve.DbfilenameShelf) """ try: return shelve.open(name) except Exception: raise Exception('Unable to connect to database with name {}'.format(name)) class DBClient(object): """Client for interacting with database for the application""" def __init__(self, database_name: str): self.conn = sqlite3.connect(database_name) self._create_quotes_table() def _create_quotes_table(self): """ Create the table used for storing quotes if it does not exist already """ with self.conn: self.conn.execute(''' CREATE TABLE IF NOT EXISTS quotes ( author TEXT, quote TEXT, created_at TEXT ); ''') def close_connection(self): """ Close connection to the database """ self.conn.close() def insert_quote(self, author: str, quote: str, created_at: str): """ Insert a quote into the database :param author: (str) Name of the author that said the quote :param quote: (str) The quote for the author :param created_at: (str) Timestamp for when the quote was saved to database """ with self.conn: self.conn.execute(''' INSERT INTO quotes VALUES (?, ?, ?) ''', (author, quote, created_at))
44b3f1d2c3e5a31887454aa82b324f15898122bc
zazu/cli.py
zazu/cli.py
# -*- coding: utf-8 -*- """entry point for zazu""" __author__ = "Nicholas Wiles" __copyright__ = "Copyright 2016, Lily Robotics" import click import git_helper import subprocess import zazu.build import zazu.config import zazu.dev.commands import zazu.repo.commands import zazu.style import zazu.upgrade @click.group() @click.version_option(version=zazu.__version__) @click.pass_context def cli(ctx): try: ctx.obj = zazu.config.Config(git_helper.get_root_path()) required_zazu_version = ctx.obj.zazu_version_required() if required_zazu_version and required_zazu_version != zazu.__version__: click.echo('Warning: this repo has requested zazu {}, which doesn\'t match the installed version ({}). \ Use "zazu upgrade" to fix this'.format(ctx.obj.zazu_version_required(), __version__)) except subprocess.CalledProcessError: pass cli.add_command(zazu.upgrade.upgrade) cli.add_command(zazu.style.style) cli.add_command(zazu.build.build) cli.add_command(zazu.dev.commands.dev) cli.add_command(zazu.repo.commands.repo)
# -*- coding: utf-8 -*- """entry point for zazu""" __author__ = "Nicholas Wiles" __copyright__ = "Copyright 2016, Lily Robotics" import click import git_helper import subprocess import zazu.build import zazu.config import zazu.dev.commands import zazu.repo.commands import zazu.style import zazu.upgrade @click.group() @click.version_option(version=zazu.__version__) @click.pass_context def cli(ctx): try: ctx.obj = zazu.config.Config(git_helper.get_root_path()) required_zazu_version = ctx.obj.zazu_version_required() if required_zazu_version and required_zazu_version != zazu.__version__: click.echo('Warning: this repo has requested zazu {}, which doesn\'t match the installed version ({}). \ Use "zazu upgrade" to fix this'.format(ctx.obj.zazu_version_required(), zazu.__version__)) except subprocess.CalledProcessError: pass cli.add_command(zazu.upgrade.upgrade) cli.add_command(zazu.style.style) cli.add_command(zazu.build.build) cli.add_command(zazu.dev.commands.dev) cli.add_command(zazu.repo.commands.repo)
Fix missing module on __version__
Fix missing module on __version__ (develop)
Python
mit
stopthatcow/zazu,stopthatcow/zazu
# -*- coding: utf-8 -*- """entry point for zazu""" __author__ = "Nicholas Wiles" __copyright__ = "Copyright 2016, Lily Robotics" import click import git_helper import subprocess import zazu.build import zazu.config import zazu.dev.commands import zazu.repo.commands import zazu.style import zazu.upgrade @click.group() @click.version_option(version=zazu.__version__) @click.pass_context def cli(ctx): try: ctx.obj = zazu.config.Config(git_helper.get_root_path()) required_zazu_version = ctx.obj.zazu_version_required() if required_zazu_version and required_zazu_version != zazu.__version__: click.echo('Warning: this repo has requested zazu {}, which doesn\'t match the installed version ({}). \ Use "zazu upgrade" to fix this'.format(ctx.obj.zazu_version_required(), __version__)) except subprocess.CalledProcessError: pass cli.add_command(zazu.upgrade.upgrade) cli.add_command(zazu.style.style) cli.add_command(zazu.build.build) cli.add_command(zazu.dev.commands.dev) cli.add_command(zazu.repo.commands.repo) Fix missing module on __version__ (develop)
# -*- coding: utf-8 -*- """entry point for zazu""" __author__ = "Nicholas Wiles" __copyright__ = "Copyright 2016, Lily Robotics" import click import git_helper import subprocess import zazu.build import zazu.config import zazu.dev.commands import zazu.repo.commands import zazu.style import zazu.upgrade @click.group() @click.version_option(version=zazu.__version__) @click.pass_context def cli(ctx): try: ctx.obj = zazu.config.Config(git_helper.get_root_path()) required_zazu_version = ctx.obj.zazu_version_required() if required_zazu_version and required_zazu_version != zazu.__version__: click.echo('Warning: this repo has requested zazu {}, which doesn\'t match the installed version ({}). \ Use "zazu upgrade" to fix this'.format(ctx.obj.zazu_version_required(), zazu.__version__)) except subprocess.CalledProcessError: pass cli.add_command(zazu.upgrade.upgrade) cli.add_command(zazu.style.style) cli.add_command(zazu.build.build) cli.add_command(zazu.dev.commands.dev) cli.add_command(zazu.repo.commands.repo)
<commit_before># -*- coding: utf-8 -*- """entry point for zazu""" __author__ = "Nicholas Wiles" __copyright__ = "Copyright 2016, Lily Robotics" import click import git_helper import subprocess import zazu.build import zazu.config import zazu.dev.commands import zazu.repo.commands import zazu.style import zazu.upgrade @click.group() @click.version_option(version=zazu.__version__) @click.pass_context def cli(ctx): try: ctx.obj = zazu.config.Config(git_helper.get_root_path()) required_zazu_version = ctx.obj.zazu_version_required() if required_zazu_version and required_zazu_version != zazu.__version__: click.echo('Warning: this repo has requested zazu {}, which doesn\'t match the installed version ({}). \ Use "zazu upgrade" to fix this'.format(ctx.obj.zazu_version_required(), __version__)) except subprocess.CalledProcessError: pass cli.add_command(zazu.upgrade.upgrade) cli.add_command(zazu.style.style) cli.add_command(zazu.build.build) cli.add_command(zazu.dev.commands.dev) cli.add_command(zazu.repo.commands.repo) <commit_msg>Fix missing module on __version__ (develop)<commit_after>
# -*- coding: utf-8 -*- """entry point for zazu""" __author__ = "Nicholas Wiles" __copyright__ = "Copyright 2016, Lily Robotics" import click import git_helper import subprocess import zazu.build import zazu.config import zazu.dev.commands import zazu.repo.commands import zazu.style import zazu.upgrade @click.group() @click.version_option(version=zazu.__version__) @click.pass_context def cli(ctx): try: ctx.obj = zazu.config.Config(git_helper.get_root_path()) required_zazu_version = ctx.obj.zazu_version_required() if required_zazu_version and required_zazu_version != zazu.__version__: click.echo('Warning: this repo has requested zazu {}, which doesn\'t match the installed version ({}). \ Use "zazu upgrade" to fix this'.format(ctx.obj.zazu_version_required(), zazu.__version__)) except subprocess.CalledProcessError: pass cli.add_command(zazu.upgrade.upgrade) cli.add_command(zazu.style.style) cli.add_command(zazu.build.build) cli.add_command(zazu.dev.commands.dev) cli.add_command(zazu.repo.commands.repo)
# -*- coding: utf-8 -*- """entry point for zazu""" __author__ = "Nicholas Wiles" __copyright__ = "Copyright 2016, Lily Robotics" import click import git_helper import subprocess import zazu.build import zazu.config import zazu.dev.commands import zazu.repo.commands import zazu.style import zazu.upgrade @click.group() @click.version_option(version=zazu.__version__) @click.pass_context def cli(ctx): try: ctx.obj = zazu.config.Config(git_helper.get_root_path()) required_zazu_version = ctx.obj.zazu_version_required() if required_zazu_version and required_zazu_version != zazu.__version__: click.echo('Warning: this repo has requested zazu {}, which doesn\'t match the installed version ({}). \ Use "zazu upgrade" to fix this'.format(ctx.obj.zazu_version_required(), __version__)) except subprocess.CalledProcessError: pass cli.add_command(zazu.upgrade.upgrade) cli.add_command(zazu.style.style) cli.add_command(zazu.build.build) cli.add_command(zazu.dev.commands.dev) cli.add_command(zazu.repo.commands.repo) Fix missing module on __version__ (develop)# -*- coding: utf-8 -*- """entry point for zazu""" __author__ = "Nicholas Wiles" __copyright__ = "Copyright 2016, Lily Robotics" import click import git_helper import subprocess import zazu.build import zazu.config import zazu.dev.commands import zazu.repo.commands import zazu.style import zazu.upgrade @click.group() @click.version_option(version=zazu.__version__) @click.pass_context def cli(ctx): try: ctx.obj = zazu.config.Config(git_helper.get_root_path()) required_zazu_version = ctx.obj.zazu_version_required() if required_zazu_version and required_zazu_version != zazu.__version__: click.echo('Warning: this repo has requested zazu {}, which doesn\'t match the installed version ({}). \ Use "zazu upgrade" to fix this'.format(ctx.obj.zazu_version_required(), zazu.__version__)) except subprocess.CalledProcessError: pass cli.add_command(zazu.upgrade.upgrade) cli.add_command(zazu.style.style) cli.add_command(zazu.build.build) cli.add_command(zazu.dev.commands.dev) cli.add_command(zazu.repo.commands.repo)
<commit_before># -*- coding: utf-8 -*- """entry point for zazu""" __author__ = "Nicholas Wiles" __copyright__ = "Copyright 2016, Lily Robotics" import click import git_helper import subprocess import zazu.build import zazu.config import zazu.dev.commands import zazu.repo.commands import zazu.style import zazu.upgrade @click.group() @click.version_option(version=zazu.__version__) @click.pass_context def cli(ctx): try: ctx.obj = zazu.config.Config(git_helper.get_root_path()) required_zazu_version = ctx.obj.zazu_version_required() if required_zazu_version and required_zazu_version != zazu.__version__: click.echo('Warning: this repo has requested zazu {}, which doesn\'t match the installed version ({}). \ Use "zazu upgrade" to fix this'.format(ctx.obj.zazu_version_required(), __version__)) except subprocess.CalledProcessError: pass cli.add_command(zazu.upgrade.upgrade) cli.add_command(zazu.style.style) cli.add_command(zazu.build.build) cli.add_command(zazu.dev.commands.dev) cli.add_command(zazu.repo.commands.repo) <commit_msg>Fix missing module on __version__ (develop)<commit_after># -*- coding: utf-8 -*- """entry point for zazu""" __author__ = "Nicholas Wiles" __copyright__ = "Copyright 2016, Lily Robotics" import click import git_helper import subprocess import zazu.build import zazu.config import zazu.dev.commands import zazu.repo.commands import zazu.style import zazu.upgrade @click.group() @click.version_option(version=zazu.__version__) @click.pass_context def cli(ctx): try: ctx.obj = zazu.config.Config(git_helper.get_root_path()) required_zazu_version = ctx.obj.zazu_version_required() if required_zazu_version and required_zazu_version != zazu.__version__: click.echo('Warning: this repo has requested zazu {}, which doesn\'t match the installed version ({}). \ Use "zazu upgrade" to fix this'.format(ctx.obj.zazu_version_required(), zazu.__version__)) except subprocess.CalledProcessError: pass cli.add_command(zazu.upgrade.upgrade) cli.add_command(zazu.style.style) cli.add_command(zazu.build.build) cli.add_command(zazu.dev.commands.dev) cli.add_command(zazu.repo.commands.repo)
f0b54d2f706912fd4a1e21117d1534170a9ce125
Config.py
Config.py
# Web domain = 'mockbox.io' useSSL = False keyFile = 'keys/privkey.pem' certFile = 'keys/cacert.pem' httpPort = 81 httpsPort = 443 # SMTP bindingPort = 587 bindingIP = '0.0.0.0' # dropSize in bytes dropSize = 10000000 # General # warning! This can fill your disk up really quickly. logEmail = False
# Web domain = 'mockbox.io' useSSL = False keyFile = 'keys/privkey.pem' certFile = 'keys/cacert.pem' httpPort = 80 httpsPort = 443 # SMTP bindingPort = 587 bindingIP = '0.0.0.0' # dropSize in bytes dropSize = 10000000 # General # warning! This can fill your disk up really quickly. logEmail = False
Move default port back to 80
Move default port back to 80
Python
mit
jkingsman/Mockbox,jkingsman/Mockbox,jkingsman/Mockbox
# Web domain = 'mockbox.io' useSSL = False keyFile = 'keys/privkey.pem' certFile = 'keys/cacert.pem' httpPort = 81 httpsPort = 443 # SMTP bindingPort = 587 bindingIP = '0.0.0.0' # dropSize in bytes dropSize = 10000000 # General # warning! This can fill your disk up really quickly. logEmail = False Move default port back to 80
# Web domain = 'mockbox.io' useSSL = False keyFile = 'keys/privkey.pem' certFile = 'keys/cacert.pem' httpPort = 80 httpsPort = 443 # SMTP bindingPort = 587 bindingIP = '0.0.0.0' # dropSize in bytes dropSize = 10000000 # General # warning! This can fill your disk up really quickly. logEmail = False
<commit_before># Web domain = 'mockbox.io' useSSL = False keyFile = 'keys/privkey.pem' certFile = 'keys/cacert.pem' httpPort = 81 httpsPort = 443 # SMTP bindingPort = 587 bindingIP = '0.0.0.0' # dropSize in bytes dropSize = 10000000 # General # warning! This can fill your disk up really quickly. logEmail = False <commit_msg>Move default port back to 80<commit_after>
# Web domain = 'mockbox.io' useSSL = False keyFile = 'keys/privkey.pem' certFile = 'keys/cacert.pem' httpPort = 80 httpsPort = 443 # SMTP bindingPort = 587 bindingIP = '0.0.0.0' # dropSize in bytes dropSize = 10000000 # General # warning! This can fill your disk up really quickly. logEmail = False
# Web domain = 'mockbox.io' useSSL = False keyFile = 'keys/privkey.pem' certFile = 'keys/cacert.pem' httpPort = 81 httpsPort = 443 # SMTP bindingPort = 587 bindingIP = '0.0.0.0' # dropSize in bytes dropSize = 10000000 # General # warning! This can fill your disk up really quickly. logEmail = False Move default port back to 80# Web domain = 'mockbox.io' useSSL = False keyFile = 'keys/privkey.pem' certFile = 'keys/cacert.pem' httpPort = 80 httpsPort = 443 # SMTP bindingPort = 587 bindingIP = '0.0.0.0' # dropSize in bytes dropSize = 10000000 # General # warning! This can fill your disk up really quickly. logEmail = False
<commit_before># Web domain = 'mockbox.io' useSSL = False keyFile = 'keys/privkey.pem' certFile = 'keys/cacert.pem' httpPort = 81 httpsPort = 443 # SMTP bindingPort = 587 bindingIP = '0.0.0.0' # dropSize in bytes dropSize = 10000000 # General # warning! This can fill your disk up really quickly. logEmail = False <commit_msg>Move default port back to 80<commit_after># Web domain = 'mockbox.io' useSSL = False keyFile = 'keys/privkey.pem' certFile = 'keys/cacert.pem' httpPort = 80 httpsPort = 443 # SMTP bindingPort = 587 bindingIP = '0.0.0.0' # dropSize in bytes dropSize = 10000000 # General # warning! This can fill your disk up really quickly. logEmail = False
918b001cb6d9743d3d2ee1b2bab8f14c90e1adf7
src/ice/rom_finder.py
src/ice/rom_finder.py
from console import Console from rom import ROM from functools import reduce class ROMFinder(object): def __init__(self, filesystem): self.filesystem = filesystem def roms_for_console(self, console): """ @param console - A console object @returns A list of ROM objects representing all of the valid ROMs for a given console. Valid ROMs are defined as ROMs for which `console`'s `is_valid_rom` method returns True. Returns an empty list if `console` is not enabled """ if not console.is_enabled(): return [] paths = self.filesystem.files_in_directory(console.roms_directory()) valid_rom_paths = filter(console.is_valid_rom, paths) return map(lambda path: ROM(path, console), valid_rom_paths) def roms_for_consoles(self, consoles): """ @param consoles - An iterable list of consoles @returns A list of all of the ROMs for all of the consoles in `consoles` Equivalent to calling `roms_for_console` on every element of `consoles` and combining the results """ assert hasattr( consoles, '__iter__'), "Expecting an iterable list of consoles" def rom_collector(roms, console): roms.extend(self.roms_for_console(console)) return roms return reduce(rom_collector, consoles, [])
from console import Console from rom import ROM from functools import reduce class ROMFinder(object): def __init__(self, filesystem): self.filesystem = filesystem def roms_for_console(self, console): """ @param console - A console object @returns A list of ROM objects representing all of the valid ROMs for a given console. Valid ROMs are defined as ROMs for which `console`'s `is_valid_rom` method returns True. Returns an empty list if `console` is not enabled """ if not console.is_enabled(): return [] paths = self.filesystem.files_in_directory(console.roms_directory()) valid_rom_paths = filter(console.is_valid_rom, paths) return map(lambda path: ROM(path, console), valid_rom_paths) def roms_for_consoles(self, consoles): """ @param consoles - An iterable list of consoles @returns A list of all of the ROMs for all of the consoles in `consoles` Equivalent to calling `roms_for_console` on every element of `consoles` and combining the results """ return reduce(lambda roms, console: roms + self.roms_for_console(console), consoles, [])
Replace 'list.extend' call with '+' operator
[Cleanup] Replace 'list.extend' call with '+' operator I knew there had to be an easier way for merging lists other than `extend`. Turns out the plus operator does exactly what I need.
Python
mit
rdoyle1978/Ice,scottrice/Ice
from console import Console from rom import ROM from functools import reduce class ROMFinder(object): def __init__(self, filesystem): self.filesystem = filesystem def roms_for_console(self, console): """ @param console - A console object @returns A list of ROM objects representing all of the valid ROMs for a given console. Valid ROMs are defined as ROMs for which `console`'s `is_valid_rom` method returns True. Returns an empty list if `console` is not enabled """ if not console.is_enabled(): return [] paths = self.filesystem.files_in_directory(console.roms_directory()) valid_rom_paths = filter(console.is_valid_rom, paths) return map(lambda path: ROM(path, console), valid_rom_paths) def roms_for_consoles(self, consoles): """ @param consoles - An iterable list of consoles @returns A list of all of the ROMs for all of the consoles in `consoles` Equivalent to calling `roms_for_console` on every element of `consoles` and combining the results """ assert hasattr( consoles, '__iter__'), "Expecting an iterable list of consoles" def rom_collector(roms, console): roms.extend(self.roms_for_console(console)) return roms return reduce(rom_collector, consoles, []) [Cleanup] Replace 'list.extend' call with '+' operator I knew there had to be an easier way for merging lists other than `extend`. Turns out the plus operator does exactly what I need.
from console import Console from rom import ROM from functools import reduce class ROMFinder(object): def __init__(self, filesystem): self.filesystem = filesystem def roms_for_console(self, console): """ @param console - A console object @returns A list of ROM objects representing all of the valid ROMs for a given console. Valid ROMs are defined as ROMs for which `console`'s `is_valid_rom` method returns True. Returns an empty list if `console` is not enabled """ if not console.is_enabled(): return [] paths = self.filesystem.files_in_directory(console.roms_directory()) valid_rom_paths = filter(console.is_valid_rom, paths) return map(lambda path: ROM(path, console), valid_rom_paths) def roms_for_consoles(self, consoles): """ @param consoles - An iterable list of consoles @returns A list of all of the ROMs for all of the consoles in `consoles` Equivalent to calling `roms_for_console` on every element of `consoles` and combining the results """ return reduce(lambda roms, console: roms + self.roms_for_console(console), consoles, [])
<commit_before> from console import Console from rom import ROM from functools import reduce class ROMFinder(object): def __init__(self, filesystem): self.filesystem = filesystem def roms_for_console(self, console): """ @param console - A console object @returns A list of ROM objects representing all of the valid ROMs for a given console. Valid ROMs are defined as ROMs for which `console`'s `is_valid_rom` method returns True. Returns an empty list if `console` is not enabled """ if not console.is_enabled(): return [] paths = self.filesystem.files_in_directory(console.roms_directory()) valid_rom_paths = filter(console.is_valid_rom, paths) return map(lambda path: ROM(path, console), valid_rom_paths) def roms_for_consoles(self, consoles): """ @param consoles - An iterable list of consoles @returns A list of all of the ROMs for all of the consoles in `consoles` Equivalent to calling `roms_for_console` on every element of `consoles` and combining the results """ assert hasattr( consoles, '__iter__'), "Expecting an iterable list of consoles" def rom_collector(roms, console): roms.extend(self.roms_for_console(console)) return roms return reduce(rom_collector, consoles, []) <commit_msg>[Cleanup] Replace 'list.extend' call with '+' operator I knew there had to be an easier way for merging lists other than `extend`. Turns out the plus operator does exactly what I need.<commit_after>
from console import Console from rom import ROM from functools import reduce class ROMFinder(object): def __init__(self, filesystem): self.filesystem = filesystem def roms_for_console(self, console): """ @param console - A console object @returns A list of ROM objects representing all of the valid ROMs for a given console. Valid ROMs are defined as ROMs for which `console`'s `is_valid_rom` method returns True. Returns an empty list if `console` is not enabled """ if not console.is_enabled(): return [] paths = self.filesystem.files_in_directory(console.roms_directory()) valid_rom_paths = filter(console.is_valid_rom, paths) return map(lambda path: ROM(path, console), valid_rom_paths) def roms_for_consoles(self, consoles): """ @param consoles - An iterable list of consoles @returns A list of all of the ROMs for all of the consoles in `consoles` Equivalent to calling `roms_for_console` on every element of `consoles` and combining the results """ return reduce(lambda roms, console: roms + self.roms_for_console(console), consoles, [])
from console import Console from rom import ROM from functools import reduce class ROMFinder(object): def __init__(self, filesystem): self.filesystem = filesystem def roms_for_console(self, console): """ @param console - A console object @returns A list of ROM objects representing all of the valid ROMs for a given console. Valid ROMs are defined as ROMs for which `console`'s `is_valid_rom` method returns True. Returns an empty list if `console` is not enabled """ if not console.is_enabled(): return [] paths = self.filesystem.files_in_directory(console.roms_directory()) valid_rom_paths = filter(console.is_valid_rom, paths) return map(lambda path: ROM(path, console), valid_rom_paths) def roms_for_consoles(self, consoles): """ @param consoles - An iterable list of consoles @returns A list of all of the ROMs for all of the consoles in `consoles` Equivalent to calling `roms_for_console` on every element of `consoles` and combining the results """ assert hasattr( consoles, '__iter__'), "Expecting an iterable list of consoles" def rom_collector(roms, console): roms.extend(self.roms_for_console(console)) return roms return reduce(rom_collector, consoles, []) [Cleanup] Replace 'list.extend' call with '+' operator I knew there had to be an easier way for merging lists other than `extend`. Turns out the plus operator does exactly what I need. from console import Console from rom import ROM from functools import reduce class ROMFinder(object): def __init__(self, filesystem): self.filesystem = filesystem def roms_for_console(self, console): """ @param console - A console object @returns A list of ROM objects representing all of the valid ROMs for a given console. Valid ROMs are defined as ROMs for which `console`'s `is_valid_rom` method returns True. Returns an empty list if `console` is not enabled """ if not console.is_enabled(): return [] paths = self.filesystem.files_in_directory(console.roms_directory()) valid_rom_paths = filter(console.is_valid_rom, paths) return map(lambda path: ROM(path, console), valid_rom_paths) def roms_for_consoles(self, consoles): """ @param consoles - An iterable list of consoles @returns A list of all of the ROMs for all of the consoles in `consoles` Equivalent to calling `roms_for_console` on every element of `consoles` and combining the results """ return reduce(lambda roms, console: roms + self.roms_for_console(console), consoles, [])
<commit_before> from console import Console from rom import ROM from functools import reduce class ROMFinder(object): def __init__(self, filesystem): self.filesystem = filesystem def roms_for_console(self, console): """ @param console - A console object @returns A list of ROM objects representing all of the valid ROMs for a given console. Valid ROMs are defined as ROMs for which `console`'s `is_valid_rom` method returns True. Returns an empty list if `console` is not enabled """ if not console.is_enabled(): return [] paths = self.filesystem.files_in_directory(console.roms_directory()) valid_rom_paths = filter(console.is_valid_rom, paths) return map(lambda path: ROM(path, console), valid_rom_paths) def roms_for_consoles(self, consoles): """ @param consoles - An iterable list of consoles @returns A list of all of the ROMs for all of the consoles in `consoles` Equivalent to calling `roms_for_console` on every element of `consoles` and combining the results """ assert hasattr( consoles, '__iter__'), "Expecting an iterable list of consoles" def rom_collector(roms, console): roms.extend(self.roms_for_console(console)) return roms return reduce(rom_collector, consoles, []) <commit_msg>[Cleanup] Replace 'list.extend' call with '+' operator I knew there had to be an easier way for merging lists other than `extend`. Turns out the plus operator does exactly what I need.<commit_after> from console import Console from rom import ROM from functools import reduce class ROMFinder(object): def __init__(self, filesystem): self.filesystem = filesystem def roms_for_console(self, console): """ @param console - A console object @returns A list of ROM objects representing all of the valid ROMs for a given console. Valid ROMs are defined as ROMs for which `console`'s `is_valid_rom` method returns True. Returns an empty list if `console` is not enabled """ if not console.is_enabled(): return [] paths = self.filesystem.files_in_directory(console.roms_directory()) valid_rom_paths = filter(console.is_valid_rom, paths) return map(lambda path: ROM(path, console), valid_rom_paths) def roms_for_consoles(self, consoles): """ @param consoles - An iterable list of consoles @returns A list of all of the ROMs for all of the consoles in `consoles` Equivalent to calling `roms_for_console` on every element of `consoles` and combining the results """ return reduce(lambda roms, console: roms + self.roms_for_console(console), consoles, [])
bb48fbaa362367c117501953f3f1ba7500ff5735
bqueryd/__init__.py
bqueryd/__init__.py
__version__ = 0.6 import os import logging logger = logging.getLogger('bqueryd') ch = logging.StreamHandler() formatter = logging.Formatter('%(asctime)s %(name)s %(message)s', datefmt="%Y-%m-%d %H:%M:%S") ch.setFormatter(formatter) logger.addHandler(ch) DEFAULT_DATA_DIR = '/srv/bcolz/' INCOMING = os.path.join(DEFAULT_DATA_DIR, 'incoming') if not os.path.exists(INCOMING): os.mkdir(INCOMING) REDIS_SET_KEY = 'bqueryd_controllers' from rpc import RPC, RPCError from controller import ControllerNode from worker import WorkerNode
__version__ = 0.8 import os import logging logger = logging.getLogger('bqueryd') ch = logging.StreamHandler() formatter = logging.Formatter('%(asctime)s %(name)s %(message)s', datefmt="%Y-%m-%d %H:%M:%S") ch.setFormatter(formatter) logger.addHandler(ch) DEFAULT_DATA_DIR = '/srv/bcolz/' INCOMING = os.path.join(DEFAULT_DATA_DIR, 'incoming') if not os.path.exists(INCOMING): os.makedirs(INCOMING) REDIS_SET_KEY = 'bqueryd_controllers' from rpc import RPC, RPCError from controller import ControllerNode from worker import WorkerNode
Use os.makedirs in stead of os.mkdir to handle intermediate paths
Use os.makedirs in stead of os.mkdir to handle intermediate paths
Python
bsd-3-clause
visualfabriq/bqueryd
__version__ = 0.6 import os import logging logger = logging.getLogger('bqueryd') ch = logging.StreamHandler() formatter = logging.Formatter('%(asctime)s %(name)s %(message)s', datefmt="%Y-%m-%d %H:%M:%S") ch.setFormatter(formatter) logger.addHandler(ch) DEFAULT_DATA_DIR = '/srv/bcolz/' INCOMING = os.path.join(DEFAULT_DATA_DIR, 'incoming') if not os.path.exists(INCOMING): os.mkdir(INCOMING) REDIS_SET_KEY = 'bqueryd_controllers' from rpc import RPC, RPCError from controller import ControllerNode from worker import WorkerNode Use os.makedirs in stead of os.mkdir to handle intermediate paths
__version__ = 0.8 import os import logging logger = logging.getLogger('bqueryd') ch = logging.StreamHandler() formatter = logging.Formatter('%(asctime)s %(name)s %(message)s', datefmt="%Y-%m-%d %H:%M:%S") ch.setFormatter(formatter) logger.addHandler(ch) DEFAULT_DATA_DIR = '/srv/bcolz/' INCOMING = os.path.join(DEFAULT_DATA_DIR, 'incoming') if not os.path.exists(INCOMING): os.makedirs(INCOMING) REDIS_SET_KEY = 'bqueryd_controllers' from rpc import RPC, RPCError from controller import ControllerNode from worker import WorkerNode
<commit_before>__version__ = 0.6 import os import logging logger = logging.getLogger('bqueryd') ch = logging.StreamHandler() formatter = logging.Formatter('%(asctime)s %(name)s %(message)s', datefmt="%Y-%m-%d %H:%M:%S") ch.setFormatter(formatter) logger.addHandler(ch) DEFAULT_DATA_DIR = '/srv/bcolz/' INCOMING = os.path.join(DEFAULT_DATA_DIR, 'incoming') if not os.path.exists(INCOMING): os.mkdir(INCOMING) REDIS_SET_KEY = 'bqueryd_controllers' from rpc import RPC, RPCError from controller import ControllerNode from worker import WorkerNode <commit_msg>Use os.makedirs in stead of os.mkdir to handle intermediate paths<commit_after>
__version__ = 0.8 import os import logging logger = logging.getLogger('bqueryd') ch = logging.StreamHandler() formatter = logging.Formatter('%(asctime)s %(name)s %(message)s', datefmt="%Y-%m-%d %H:%M:%S") ch.setFormatter(formatter) logger.addHandler(ch) DEFAULT_DATA_DIR = '/srv/bcolz/' INCOMING = os.path.join(DEFAULT_DATA_DIR, 'incoming') if not os.path.exists(INCOMING): os.makedirs(INCOMING) REDIS_SET_KEY = 'bqueryd_controllers' from rpc import RPC, RPCError from controller import ControllerNode from worker import WorkerNode
__version__ = 0.6 import os import logging logger = logging.getLogger('bqueryd') ch = logging.StreamHandler() formatter = logging.Formatter('%(asctime)s %(name)s %(message)s', datefmt="%Y-%m-%d %H:%M:%S") ch.setFormatter(formatter) logger.addHandler(ch) DEFAULT_DATA_DIR = '/srv/bcolz/' INCOMING = os.path.join(DEFAULT_DATA_DIR, 'incoming') if not os.path.exists(INCOMING): os.mkdir(INCOMING) REDIS_SET_KEY = 'bqueryd_controllers' from rpc import RPC, RPCError from controller import ControllerNode from worker import WorkerNode Use os.makedirs in stead of os.mkdir to handle intermediate paths__version__ = 0.8 import os import logging logger = logging.getLogger('bqueryd') ch = logging.StreamHandler() formatter = logging.Formatter('%(asctime)s %(name)s %(message)s', datefmt="%Y-%m-%d %H:%M:%S") ch.setFormatter(formatter) logger.addHandler(ch) DEFAULT_DATA_DIR = '/srv/bcolz/' INCOMING = os.path.join(DEFAULT_DATA_DIR, 'incoming') if not os.path.exists(INCOMING): os.makedirs(INCOMING) REDIS_SET_KEY = 'bqueryd_controllers' from rpc import RPC, RPCError from controller import ControllerNode from worker import WorkerNode
<commit_before>__version__ = 0.6 import os import logging logger = logging.getLogger('bqueryd') ch = logging.StreamHandler() formatter = logging.Formatter('%(asctime)s %(name)s %(message)s', datefmt="%Y-%m-%d %H:%M:%S") ch.setFormatter(formatter) logger.addHandler(ch) DEFAULT_DATA_DIR = '/srv/bcolz/' INCOMING = os.path.join(DEFAULT_DATA_DIR, 'incoming') if not os.path.exists(INCOMING): os.mkdir(INCOMING) REDIS_SET_KEY = 'bqueryd_controllers' from rpc import RPC, RPCError from controller import ControllerNode from worker import WorkerNode <commit_msg>Use os.makedirs in stead of os.mkdir to handle intermediate paths<commit_after>__version__ = 0.8 import os import logging logger = logging.getLogger('bqueryd') ch = logging.StreamHandler() formatter = logging.Formatter('%(asctime)s %(name)s %(message)s', datefmt="%Y-%m-%d %H:%M:%S") ch.setFormatter(formatter) logger.addHandler(ch) DEFAULT_DATA_DIR = '/srv/bcolz/' INCOMING = os.path.join(DEFAULT_DATA_DIR, 'incoming') if not os.path.exists(INCOMING): os.makedirs(INCOMING) REDIS_SET_KEY = 'bqueryd_controllers' from rpc import RPC, RPCError from controller import ControllerNode from worker import WorkerNode
d2197583c197745ad33fa9c59c6c7f9681b35078
plugins/shorewall/__init__.py
plugins/shorewall/__init__.py
import logging, interfaces, os from datetime import datetime class Shorewall(interfaces.IOpenMesherConfigPlugin): def activate(self): self._register('shorewall/interfaces.conf') self._register('shorewall/rules.conf') def process(self, mesh, cliargs=None): logging.debug('Generating Shorewall config...') self._files = {} for router in mesh.links: self._files[router] = {} self._files[router]['/shorewall/interfaces.mesh'] = self._templates['shorewall/interfaces.conf'].render(links = mesh.links[router], zone='vpn') self._files[router]['/shorewall/rules.mesh'] = self._templates['shorewall/rules.conf'].render(links= mesh.links[router]) return True def files(self): """ Return a dictionary of routers containing a dictionary of filenames and contents """ return self._files
import logging, interfaces, os from datetime import datetime class Shorewall(interfaces.IOpenMesherConfigPlugin): def activate(self): self._register('shorewall/interfaces.conf') self._register('shorewall/rules.conf') def setupargs(self, parser): parser.add_argument('--wanzone', action='store', default='wan', help='Name of the WAN zone in Shorewall') parser.add_argument('--fwzone', action='store', default='fw', help='Name of the Firewall zone in Shorewall') parser.add_argument('--vpnzone', action='store', default='vpn', help='Name of the VPN zone in Shorewall') def process(self, mesh, cliargs=None): logging.debug('Generating Shorewall config...') self._files = {} for router in mesh.links: self._files[router] = {} self._files[router]['/shorewall/interfaces.mesh'] = self._templates['shorewall/interfaces.conf'].render(links = mesh.links[router], vpnzone=cliargs.vpnzone, fwzone=cliargs.fwzone, wanzone=cliargs.wanzone) self._files[router]['/shorewall/rules.mesh'] = self._templates['shorewall/rules.conf'].render(links= mesh.links[router], vpnzone=cliargs.vpnzone, fwzone=cliargs.fwzone, wanzone=cliargs.wanzone) return True def files(self): """ Return a dictionary of routers containing a dictionary of filenames and contents """ return self._files
Add cliargs to shorewall for zone handling
Add cliargs to shorewall for zone handling
Python
bsd-3-clause
heyaaron/openmesher,darkpixel/openmesher,heyaaron/openmesher,darkpixel/openmesher
import logging, interfaces, os from datetime import datetime class Shorewall(interfaces.IOpenMesherConfigPlugin): def activate(self): self._register('shorewall/interfaces.conf') self._register('shorewall/rules.conf') def process(self, mesh, cliargs=None): logging.debug('Generating Shorewall config...') self._files = {} for router in mesh.links: self._files[router] = {} self._files[router]['/shorewall/interfaces.mesh'] = self._templates['shorewall/interfaces.conf'].render(links = mesh.links[router], zone='vpn') self._files[router]['/shorewall/rules.mesh'] = self._templates['shorewall/rules.conf'].render(links= mesh.links[router]) return True def files(self): """ Return a dictionary of routers containing a dictionary of filenames and contents """ return self._files Add cliargs to shorewall for zone handling
import logging, interfaces, os from datetime import datetime class Shorewall(interfaces.IOpenMesherConfigPlugin): def activate(self): self._register('shorewall/interfaces.conf') self._register('shorewall/rules.conf') def setupargs(self, parser): parser.add_argument('--wanzone', action='store', default='wan', help='Name of the WAN zone in Shorewall') parser.add_argument('--fwzone', action='store', default='fw', help='Name of the Firewall zone in Shorewall') parser.add_argument('--vpnzone', action='store', default='vpn', help='Name of the VPN zone in Shorewall') def process(self, mesh, cliargs=None): logging.debug('Generating Shorewall config...') self._files = {} for router in mesh.links: self._files[router] = {} self._files[router]['/shorewall/interfaces.mesh'] = self._templates['shorewall/interfaces.conf'].render(links = mesh.links[router], vpnzone=cliargs.vpnzone, fwzone=cliargs.fwzone, wanzone=cliargs.wanzone) self._files[router]['/shorewall/rules.mesh'] = self._templates['shorewall/rules.conf'].render(links= mesh.links[router], vpnzone=cliargs.vpnzone, fwzone=cliargs.fwzone, wanzone=cliargs.wanzone) return True def files(self): """ Return a dictionary of routers containing a dictionary of filenames and contents """ return self._files
<commit_before>import logging, interfaces, os from datetime import datetime class Shorewall(interfaces.IOpenMesherConfigPlugin): def activate(self): self._register('shorewall/interfaces.conf') self._register('shorewall/rules.conf') def process(self, mesh, cliargs=None): logging.debug('Generating Shorewall config...') self._files = {} for router in mesh.links: self._files[router] = {} self._files[router]['/shorewall/interfaces.mesh'] = self._templates['shorewall/interfaces.conf'].render(links = mesh.links[router], zone='vpn') self._files[router]['/shorewall/rules.mesh'] = self._templates['shorewall/rules.conf'].render(links= mesh.links[router]) return True def files(self): """ Return a dictionary of routers containing a dictionary of filenames and contents """ return self._files <commit_msg>Add cliargs to shorewall for zone handling<commit_after>
import logging, interfaces, os from datetime import datetime class Shorewall(interfaces.IOpenMesherConfigPlugin): def activate(self): self._register('shorewall/interfaces.conf') self._register('shorewall/rules.conf') def setupargs(self, parser): parser.add_argument('--wanzone', action='store', default='wan', help='Name of the WAN zone in Shorewall') parser.add_argument('--fwzone', action='store', default='fw', help='Name of the Firewall zone in Shorewall') parser.add_argument('--vpnzone', action='store', default='vpn', help='Name of the VPN zone in Shorewall') def process(self, mesh, cliargs=None): logging.debug('Generating Shorewall config...') self._files = {} for router in mesh.links: self._files[router] = {} self._files[router]['/shorewall/interfaces.mesh'] = self._templates['shorewall/interfaces.conf'].render(links = mesh.links[router], vpnzone=cliargs.vpnzone, fwzone=cliargs.fwzone, wanzone=cliargs.wanzone) self._files[router]['/shorewall/rules.mesh'] = self._templates['shorewall/rules.conf'].render(links= mesh.links[router], vpnzone=cliargs.vpnzone, fwzone=cliargs.fwzone, wanzone=cliargs.wanzone) return True def files(self): """ Return a dictionary of routers containing a dictionary of filenames and contents """ return self._files
import logging, interfaces, os from datetime import datetime class Shorewall(interfaces.IOpenMesherConfigPlugin): def activate(self): self._register('shorewall/interfaces.conf') self._register('shorewall/rules.conf') def process(self, mesh, cliargs=None): logging.debug('Generating Shorewall config...') self._files = {} for router in mesh.links: self._files[router] = {} self._files[router]['/shorewall/interfaces.mesh'] = self._templates['shorewall/interfaces.conf'].render(links = mesh.links[router], zone='vpn') self._files[router]['/shorewall/rules.mesh'] = self._templates['shorewall/rules.conf'].render(links= mesh.links[router]) return True def files(self): """ Return a dictionary of routers containing a dictionary of filenames and contents """ return self._files Add cliargs to shorewall for zone handlingimport logging, interfaces, os from datetime import datetime class Shorewall(interfaces.IOpenMesherConfigPlugin): def activate(self): self._register('shorewall/interfaces.conf') self._register('shorewall/rules.conf') def setupargs(self, parser): parser.add_argument('--wanzone', action='store', default='wan', help='Name of the WAN zone in Shorewall') parser.add_argument('--fwzone', action='store', default='fw', help='Name of the Firewall zone in Shorewall') parser.add_argument('--vpnzone', action='store', default='vpn', help='Name of the VPN zone in Shorewall') def process(self, mesh, cliargs=None): logging.debug('Generating Shorewall config...') self._files = {} for router in mesh.links: self._files[router] = {} self._files[router]['/shorewall/interfaces.mesh'] = self._templates['shorewall/interfaces.conf'].render(links = mesh.links[router], vpnzone=cliargs.vpnzone, fwzone=cliargs.fwzone, wanzone=cliargs.wanzone) self._files[router]['/shorewall/rules.mesh'] = self._templates['shorewall/rules.conf'].render(links= mesh.links[router], vpnzone=cliargs.vpnzone, fwzone=cliargs.fwzone, wanzone=cliargs.wanzone) return True def files(self): """ Return a dictionary of routers containing a dictionary of filenames and contents """ return self._files
<commit_before>import logging, interfaces, os from datetime import datetime class Shorewall(interfaces.IOpenMesherConfigPlugin): def activate(self): self._register('shorewall/interfaces.conf') self._register('shorewall/rules.conf') def process(self, mesh, cliargs=None): logging.debug('Generating Shorewall config...') self._files = {} for router in mesh.links: self._files[router] = {} self._files[router]['/shorewall/interfaces.mesh'] = self._templates['shorewall/interfaces.conf'].render(links = mesh.links[router], zone='vpn') self._files[router]['/shorewall/rules.mesh'] = self._templates['shorewall/rules.conf'].render(links= mesh.links[router]) return True def files(self): """ Return a dictionary of routers containing a dictionary of filenames and contents """ return self._files <commit_msg>Add cliargs to shorewall for zone handling<commit_after>import logging, interfaces, os from datetime import datetime class Shorewall(interfaces.IOpenMesherConfigPlugin): def activate(self): self._register('shorewall/interfaces.conf') self._register('shorewall/rules.conf') def setupargs(self, parser): parser.add_argument('--wanzone', action='store', default='wan', help='Name of the WAN zone in Shorewall') parser.add_argument('--fwzone', action='store', default='fw', help='Name of the Firewall zone in Shorewall') parser.add_argument('--vpnzone', action='store', default='vpn', help='Name of the VPN zone in Shorewall') def process(self, mesh, cliargs=None): logging.debug('Generating Shorewall config...') self._files = {} for router in mesh.links: self._files[router] = {} self._files[router]['/shorewall/interfaces.mesh'] = self._templates['shorewall/interfaces.conf'].render(links = mesh.links[router], vpnzone=cliargs.vpnzone, fwzone=cliargs.fwzone, wanzone=cliargs.wanzone) self._files[router]['/shorewall/rules.mesh'] = self._templates['shorewall/rules.conf'].render(links= mesh.links[router], vpnzone=cliargs.vpnzone, fwzone=cliargs.fwzone, wanzone=cliargs.wanzone) return True def files(self): """ Return a dictionary of routers containing a dictionary of filenames and contents """ return self._files
cdd79aa60f4ef707714a632188373a5c2c4b0af4
mass_mailing_switzerland/models/crm_event_compassion.py
mass_mailing_switzerland/models/crm_event_compassion.py
############################################################################## # # Copyright (C) 2018 Compassion CH (http://www.compassion.ch) # Releasing children from poverty in Jesus' name # @author: Nathan Fluckiger <nathan.fluckiger@hotmail.ch> # # The licence is in the file __manifest__.py # ############################################################################## from odoo import models, api class EventCompassion(models.Model): _inherit = "crm.event.compassion" @api.model def create(self, vals): event = super().create(vals) if event.campaign_id: event.analytic_id.campaign_id = event.campaign_id event.origin_id.campaign_id = event.campaign_id return event @api.multi def write(self, vals): res = super().write(vals) for new_event in self: if new_event.campaign_id: new_event.analytic_id.campaign_id = new_event.campaign_id new_event.origin_id.campaign_id = new_event.campaign_id return res
############################################################################## # # Copyright (C) 2018 Compassion CH (http://www.compassion.ch) # Releasing children from poverty in Jesus' name # @author: Nathan Fluckiger <nathan.fluckiger@hotmail.ch> # # The licence is in the file __manifest__.py # ############################################################################## from odoo import models, api class EventCompassion(models.Model): _inherit = "crm.event.compassion" @api.model def create(self, vals): event = super().create(vals) if event.campaign_id: event.analytic_id.campaign_id = event.campaign_id event.origin_id.campaign_id = event.campaign_id return event @api.multi def write(self, vals): res = super().write(vals) for new_event in self: if new_event.campaign_id: if new_event.analytic_id: new_event.analytic_id.campaign_id = new_event.campaign_id if new_event.origin_id: new_event.origin_id.campaign_id = new_event.campaign_id return res
FIX bug in event creation
FIX bug in event creation
Python
agpl-3.0
CompassionCH/compassion-switzerland,eicher31/compassion-switzerland,CompassionCH/compassion-switzerland,eicher31/compassion-switzerland,CompassionCH/compassion-switzerland,eicher31/compassion-switzerland
############################################################################## # # Copyright (C) 2018 Compassion CH (http://www.compassion.ch) # Releasing children from poverty in Jesus' name # @author: Nathan Fluckiger <nathan.fluckiger@hotmail.ch> # # The licence is in the file __manifest__.py # ############################################################################## from odoo import models, api class EventCompassion(models.Model): _inherit = "crm.event.compassion" @api.model def create(self, vals): event = super().create(vals) if event.campaign_id: event.analytic_id.campaign_id = event.campaign_id event.origin_id.campaign_id = event.campaign_id return event @api.multi def write(self, vals): res = super().write(vals) for new_event in self: if new_event.campaign_id: new_event.analytic_id.campaign_id = new_event.campaign_id new_event.origin_id.campaign_id = new_event.campaign_id return res FIX bug in event creation
############################################################################## # # Copyright (C) 2018 Compassion CH (http://www.compassion.ch) # Releasing children from poverty in Jesus' name # @author: Nathan Fluckiger <nathan.fluckiger@hotmail.ch> # # The licence is in the file __manifest__.py # ############################################################################## from odoo import models, api class EventCompassion(models.Model): _inherit = "crm.event.compassion" @api.model def create(self, vals): event = super().create(vals) if event.campaign_id: event.analytic_id.campaign_id = event.campaign_id event.origin_id.campaign_id = event.campaign_id return event @api.multi def write(self, vals): res = super().write(vals) for new_event in self: if new_event.campaign_id: if new_event.analytic_id: new_event.analytic_id.campaign_id = new_event.campaign_id if new_event.origin_id: new_event.origin_id.campaign_id = new_event.campaign_id return res
<commit_before>############################################################################## # # Copyright (C) 2018 Compassion CH (http://www.compassion.ch) # Releasing children from poverty in Jesus' name # @author: Nathan Fluckiger <nathan.fluckiger@hotmail.ch> # # The licence is in the file __manifest__.py # ############################################################################## from odoo import models, api class EventCompassion(models.Model): _inherit = "crm.event.compassion" @api.model def create(self, vals): event = super().create(vals) if event.campaign_id: event.analytic_id.campaign_id = event.campaign_id event.origin_id.campaign_id = event.campaign_id return event @api.multi def write(self, vals): res = super().write(vals) for new_event in self: if new_event.campaign_id: new_event.analytic_id.campaign_id = new_event.campaign_id new_event.origin_id.campaign_id = new_event.campaign_id return res <commit_msg>FIX bug in event creation<commit_after>
############################################################################## # # Copyright (C) 2018 Compassion CH (http://www.compassion.ch) # Releasing children from poverty in Jesus' name # @author: Nathan Fluckiger <nathan.fluckiger@hotmail.ch> # # The licence is in the file __manifest__.py # ############################################################################## from odoo import models, api class EventCompassion(models.Model): _inherit = "crm.event.compassion" @api.model def create(self, vals): event = super().create(vals) if event.campaign_id: event.analytic_id.campaign_id = event.campaign_id event.origin_id.campaign_id = event.campaign_id return event @api.multi def write(self, vals): res = super().write(vals) for new_event in self: if new_event.campaign_id: if new_event.analytic_id: new_event.analytic_id.campaign_id = new_event.campaign_id if new_event.origin_id: new_event.origin_id.campaign_id = new_event.campaign_id return res
############################################################################## # # Copyright (C) 2018 Compassion CH (http://www.compassion.ch) # Releasing children from poverty in Jesus' name # @author: Nathan Fluckiger <nathan.fluckiger@hotmail.ch> # # The licence is in the file __manifest__.py # ############################################################################## from odoo import models, api class EventCompassion(models.Model): _inherit = "crm.event.compassion" @api.model def create(self, vals): event = super().create(vals) if event.campaign_id: event.analytic_id.campaign_id = event.campaign_id event.origin_id.campaign_id = event.campaign_id return event @api.multi def write(self, vals): res = super().write(vals) for new_event in self: if new_event.campaign_id: new_event.analytic_id.campaign_id = new_event.campaign_id new_event.origin_id.campaign_id = new_event.campaign_id return res FIX bug in event creation############################################################################## # # Copyright (C) 2018 Compassion CH (http://www.compassion.ch) # Releasing children from poverty in Jesus' name # @author: Nathan Fluckiger <nathan.fluckiger@hotmail.ch> # # The licence is in the file __manifest__.py # ############################################################################## from odoo import models, api class EventCompassion(models.Model): _inherit = "crm.event.compassion" @api.model def create(self, vals): event = super().create(vals) if event.campaign_id: event.analytic_id.campaign_id = event.campaign_id event.origin_id.campaign_id = event.campaign_id return event @api.multi def write(self, vals): res = super().write(vals) for new_event in self: if new_event.campaign_id: if new_event.analytic_id: new_event.analytic_id.campaign_id = new_event.campaign_id if new_event.origin_id: new_event.origin_id.campaign_id = new_event.campaign_id return res
<commit_before>############################################################################## # # Copyright (C) 2018 Compassion CH (http://www.compassion.ch) # Releasing children from poverty in Jesus' name # @author: Nathan Fluckiger <nathan.fluckiger@hotmail.ch> # # The licence is in the file __manifest__.py # ############################################################################## from odoo import models, api class EventCompassion(models.Model): _inherit = "crm.event.compassion" @api.model def create(self, vals): event = super().create(vals) if event.campaign_id: event.analytic_id.campaign_id = event.campaign_id event.origin_id.campaign_id = event.campaign_id return event @api.multi def write(self, vals): res = super().write(vals) for new_event in self: if new_event.campaign_id: new_event.analytic_id.campaign_id = new_event.campaign_id new_event.origin_id.campaign_id = new_event.campaign_id return res <commit_msg>FIX bug in event creation<commit_after>############################################################################## # # Copyright (C) 2018 Compassion CH (http://www.compassion.ch) # Releasing children from poverty in Jesus' name # @author: Nathan Fluckiger <nathan.fluckiger@hotmail.ch> # # The licence is in the file __manifest__.py # ############################################################################## from odoo import models, api class EventCompassion(models.Model): _inherit = "crm.event.compassion" @api.model def create(self, vals): event = super().create(vals) if event.campaign_id: event.analytic_id.campaign_id = event.campaign_id event.origin_id.campaign_id = event.campaign_id return event @api.multi def write(self, vals): res = super().write(vals) for new_event in self: if new_event.campaign_id: if new_event.analytic_id: new_event.analytic_id.campaign_id = new_event.campaign_id if new_event.origin_id: new_event.origin_id.campaign_id = new_event.campaign_id return res
19e6c020bc7d640fe4c8ffbdf7825693d7e4a03a
scripts/missing-qq.py
scripts/missing-qq.py
#!/usr/bin/env python import os import xml.etree.ElementTree as ET RES_FOLDER = os.path.abspath(os.path.join(os.path.dirname(__file__), "../app/src/main/res")) EN_STRINGS = os.path.join(RES_FOLDER, "values/strings.xml") QQ_STRINGS = os.path.join(RES_FOLDER, "values-qq/strings.xml") # Get ElementTree containing all message names in English enroot = ET.parse(EN_STRINGS).getroot() # Get ElementTree containing all documented messages qqroot = ET.parse(QQ_STRINGS).getroot() # Create a set to store all documented messages qqmsgs = set() # Add all documented messages to that set for child in qqroot: qqmsgs.add(child.attrib['name']) # Iterate through all messages and check that they're documented for child in enroot: if child.attrib['name'] not in qqmsgs: print(child.attrib['name'] + " is undocumented!")
#!/usr/bin/env python import os import sys import xml.etree.ElementTree as ET RES_FOLDER = os.path.abspath(os.path.join(os.path.dirname(__file__), "../app/src/main/res")) EN_STRINGS = os.path.join(RES_FOLDER, "values/strings.xml") QQ_STRINGS = os.path.join(RES_FOLDER, "values-qq/strings.xml") # Get ElementTree containing all message names in English enroot = ET.parse(EN_STRINGS).getroot() # Get ElementTree containing all documented messages qqroot = ET.parse(QQ_STRINGS).getroot() # Create a set to store all documented messages qqmsgs = set() # Add all documented messages to that set for child in qqroot: qqmsgs.add(child.attrib['name']) # Iterate through all messages and check that they're documented missing = 0 for child in enroot: if child.attrib['name'] not in qqmsgs: print(child.attrib['name'] + " is undocumented!") missing += 1 sys.exit(1 if missing else 0)
Exit with nonzero when qq strings are missing
Exit with nonzero when qq strings are missing Change-Id: Ife0f114dbe48faa445397aa7a94f74de2309d117
Python
apache-2.0
anirudh24seven/apps-android-wikipedia,wikimedia/apps-android-wikipedia,SAGROUP2/apps-android-wikipedia,SAGROUP2/apps-android-wikipedia,SAGROUP2/apps-android-wikipedia,SAGROUP2/apps-android-wikipedia,wikimedia/apps-android-wikipedia,wikimedia/apps-android-wikipedia,SAGROUP2/apps-android-wikipedia,dbrant/apps-android-wikipedia,anirudh24seven/apps-android-wikipedia,dbrant/apps-android-wikipedia,dbrant/apps-android-wikipedia,anirudh24seven/apps-android-wikipedia,wikimedia/apps-android-wikipedia,dbrant/apps-android-wikipedia,dbrant/apps-android-wikipedia,anirudh24seven/apps-android-wikipedia,anirudh24seven/apps-android-wikipedia
#!/usr/bin/env python import os import xml.etree.ElementTree as ET RES_FOLDER = os.path.abspath(os.path.join(os.path.dirname(__file__), "../app/src/main/res")) EN_STRINGS = os.path.join(RES_FOLDER, "values/strings.xml") QQ_STRINGS = os.path.join(RES_FOLDER, "values-qq/strings.xml") # Get ElementTree containing all message names in English enroot = ET.parse(EN_STRINGS).getroot() # Get ElementTree containing all documented messages qqroot = ET.parse(QQ_STRINGS).getroot() # Create a set to store all documented messages qqmsgs = set() # Add all documented messages to that set for child in qqroot: qqmsgs.add(child.attrib['name']) # Iterate through all messages and check that they're documented for child in enroot: if child.attrib['name'] not in qqmsgs: print(child.attrib['name'] + " is undocumented!") Exit with nonzero when qq strings are missing Change-Id: Ife0f114dbe48faa445397aa7a94f74de2309d117
#!/usr/bin/env python import os import sys import xml.etree.ElementTree as ET RES_FOLDER = os.path.abspath(os.path.join(os.path.dirname(__file__), "../app/src/main/res")) EN_STRINGS = os.path.join(RES_FOLDER, "values/strings.xml") QQ_STRINGS = os.path.join(RES_FOLDER, "values-qq/strings.xml") # Get ElementTree containing all message names in English enroot = ET.parse(EN_STRINGS).getroot() # Get ElementTree containing all documented messages qqroot = ET.parse(QQ_STRINGS).getroot() # Create a set to store all documented messages qqmsgs = set() # Add all documented messages to that set for child in qqroot: qqmsgs.add(child.attrib['name']) # Iterate through all messages and check that they're documented missing = 0 for child in enroot: if child.attrib['name'] not in qqmsgs: print(child.attrib['name'] + " is undocumented!") missing += 1 sys.exit(1 if missing else 0)
<commit_before>#!/usr/bin/env python import os import xml.etree.ElementTree as ET RES_FOLDER = os.path.abspath(os.path.join(os.path.dirname(__file__), "../app/src/main/res")) EN_STRINGS = os.path.join(RES_FOLDER, "values/strings.xml") QQ_STRINGS = os.path.join(RES_FOLDER, "values-qq/strings.xml") # Get ElementTree containing all message names in English enroot = ET.parse(EN_STRINGS).getroot() # Get ElementTree containing all documented messages qqroot = ET.parse(QQ_STRINGS).getroot() # Create a set to store all documented messages qqmsgs = set() # Add all documented messages to that set for child in qqroot: qqmsgs.add(child.attrib['name']) # Iterate through all messages and check that they're documented for child in enroot: if child.attrib['name'] not in qqmsgs: print(child.attrib['name'] + " is undocumented!") <commit_msg>Exit with nonzero when qq strings are missing Change-Id: Ife0f114dbe48faa445397aa7a94f74de2309d117<commit_after>
#!/usr/bin/env python import os import sys import xml.etree.ElementTree as ET RES_FOLDER = os.path.abspath(os.path.join(os.path.dirname(__file__), "../app/src/main/res")) EN_STRINGS = os.path.join(RES_FOLDER, "values/strings.xml") QQ_STRINGS = os.path.join(RES_FOLDER, "values-qq/strings.xml") # Get ElementTree containing all message names in English enroot = ET.parse(EN_STRINGS).getroot() # Get ElementTree containing all documented messages qqroot = ET.parse(QQ_STRINGS).getroot() # Create a set to store all documented messages qqmsgs = set() # Add all documented messages to that set for child in qqroot: qqmsgs.add(child.attrib['name']) # Iterate through all messages and check that they're documented missing = 0 for child in enroot: if child.attrib['name'] not in qqmsgs: print(child.attrib['name'] + " is undocumented!") missing += 1 sys.exit(1 if missing else 0)
#!/usr/bin/env python import os import xml.etree.ElementTree as ET RES_FOLDER = os.path.abspath(os.path.join(os.path.dirname(__file__), "../app/src/main/res")) EN_STRINGS = os.path.join(RES_FOLDER, "values/strings.xml") QQ_STRINGS = os.path.join(RES_FOLDER, "values-qq/strings.xml") # Get ElementTree containing all message names in English enroot = ET.parse(EN_STRINGS).getroot() # Get ElementTree containing all documented messages qqroot = ET.parse(QQ_STRINGS).getroot() # Create a set to store all documented messages qqmsgs = set() # Add all documented messages to that set for child in qqroot: qqmsgs.add(child.attrib['name']) # Iterate through all messages and check that they're documented for child in enroot: if child.attrib['name'] not in qqmsgs: print(child.attrib['name'] + " is undocumented!") Exit with nonzero when qq strings are missing Change-Id: Ife0f114dbe48faa445397aa7a94f74de2309d117#!/usr/bin/env python import os import sys import xml.etree.ElementTree as ET RES_FOLDER = os.path.abspath(os.path.join(os.path.dirname(__file__), "../app/src/main/res")) EN_STRINGS = os.path.join(RES_FOLDER, "values/strings.xml") QQ_STRINGS = os.path.join(RES_FOLDER, "values-qq/strings.xml") # Get ElementTree containing all message names in English enroot = ET.parse(EN_STRINGS).getroot() # Get ElementTree containing all documented messages qqroot = ET.parse(QQ_STRINGS).getroot() # Create a set to store all documented messages qqmsgs = set() # Add all documented messages to that set for child in qqroot: qqmsgs.add(child.attrib['name']) # Iterate through all messages and check that they're documented missing = 0 for child in enroot: if child.attrib['name'] not in qqmsgs: print(child.attrib['name'] + " is undocumented!") missing += 1 sys.exit(1 if missing else 0)
<commit_before>#!/usr/bin/env python import os import xml.etree.ElementTree as ET RES_FOLDER = os.path.abspath(os.path.join(os.path.dirname(__file__), "../app/src/main/res")) EN_STRINGS = os.path.join(RES_FOLDER, "values/strings.xml") QQ_STRINGS = os.path.join(RES_FOLDER, "values-qq/strings.xml") # Get ElementTree containing all message names in English enroot = ET.parse(EN_STRINGS).getroot() # Get ElementTree containing all documented messages qqroot = ET.parse(QQ_STRINGS).getroot() # Create a set to store all documented messages qqmsgs = set() # Add all documented messages to that set for child in qqroot: qqmsgs.add(child.attrib['name']) # Iterate through all messages and check that they're documented for child in enroot: if child.attrib['name'] not in qqmsgs: print(child.attrib['name'] + " is undocumented!") <commit_msg>Exit with nonzero when qq strings are missing Change-Id: Ife0f114dbe48faa445397aa7a94f74de2309d117<commit_after>#!/usr/bin/env python import os import sys import xml.etree.ElementTree as ET RES_FOLDER = os.path.abspath(os.path.join(os.path.dirname(__file__), "../app/src/main/res")) EN_STRINGS = os.path.join(RES_FOLDER, "values/strings.xml") QQ_STRINGS = os.path.join(RES_FOLDER, "values-qq/strings.xml") # Get ElementTree containing all message names in English enroot = ET.parse(EN_STRINGS).getroot() # Get ElementTree containing all documented messages qqroot = ET.parse(QQ_STRINGS).getroot() # Create a set to store all documented messages qqmsgs = set() # Add all documented messages to that set for child in qqroot: qqmsgs.add(child.attrib['name']) # Iterate through all messages and check that they're documented missing = 0 for child in enroot: if child.attrib['name'] not in qqmsgs: print(child.attrib['name'] + " is undocumented!") missing += 1 sys.exit(1 if missing else 0)
91c95c383803b6d25f8f68752ca7d507eaa9c7c8
harvester/tests/test_cubeupload.py
harvester/tests/test_cubeupload.py
#!/usr/bin/env python3 #-*- coding: utf-8 -*- import hashlib import unittest from harvester import harvester class CupeuploadTest(unittest.TestCase): def setUp(self): self.nick = "test" self.chan = '#brotherBot' self.mask = "brotherBox!~brotherBo@unaffiliated/brotherbox" self.h = harvester.HarvesterBot def test_fetch_cubeupload_share(self): msg = "http://cubeupload.com/im/YhUxlj.jpg" test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca" c = self.h._retrieve_content(self.h, self.mask, msg, self.chan) md5 = hashlib.md5() md5.update(c[0]['content']) self.assertEqual(md5.hexdigest(), test_hash) def test_fetch_cubeupload_raw(self): msg = "http://i.cubeupload.com/YhUxlj.jpg" test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca" c = self.h._retrieve_content(self.h, self.mask, msg, self.chan) md5 = hashlib.md5() md5.update(c[0]['content']) self.assertEqual(md5.hexdigest(), test_hash) if __name__ == '__main__': unittest.main()
#!/usr/bin/env python3 #-*- coding: utf-8 -*- import hashlib import unittest from harvester import harvester class CubeuploadTest(unittest.TestCase): def setUp(self): self.nick = "test" self.chan = '#brotherBot' self.mask = "brotherBox!~brotherBo@unaffiliated/brotherbox" self.h = harvester.HarvesterBot def test_fetch_cubeupload_share(self): msg = "http://cubeupload.com/im/YhUxlj.jpg" test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca" c = self.h._retrieve_content(self.h, self.mask, msg, self.chan) md5 = hashlib.md5() md5.update(c[0]['content']) self.assertEqual(md5.hexdigest(), test_hash) def test_fetch_cubeupload_raw(self): msg = "http://i.cubeupload.com/YhUxlj.jpg" test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca" c = self.h._retrieve_content(self.h, self.mask, msg, self.chan) md5 = hashlib.md5() md5.update(c[0]['content']) self.assertEqual(md5.hexdigest(), test_hash) if __name__ == '__main__': unittest.main()
Fix minor spelling error in cubeupload test
Fix minor spelling error in cubeupload test
Python
unlicense
luceatnobis/bravester,luceatnobis/harvester
#!/usr/bin/env python3 #-*- coding: utf-8 -*- import hashlib import unittest from harvester import harvester class CupeuploadTest(unittest.TestCase): def setUp(self): self.nick = "test" self.chan = '#brotherBot' self.mask = "brotherBox!~brotherBo@unaffiliated/brotherbox" self.h = harvester.HarvesterBot def test_fetch_cubeupload_share(self): msg = "http://cubeupload.com/im/YhUxlj.jpg" test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca" c = self.h._retrieve_content(self.h, self.mask, msg, self.chan) md5 = hashlib.md5() md5.update(c[0]['content']) self.assertEqual(md5.hexdigest(), test_hash) def test_fetch_cubeupload_raw(self): msg = "http://i.cubeupload.com/YhUxlj.jpg" test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca" c = self.h._retrieve_content(self.h, self.mask, msg, self.chan) md5 = hashlib.md5() md5.update(c[0]['content']) self.assertEqual(md5.hexdigest(), test_hash) if __name__ == '__main__': unittest.main() Fix minor spelling error in cubeupload test
#!/usr/bin/env python3 #-*- coding: utf-8 -*- import hashlib import unittest from harvester import harvester class CubeuploadTest(unittest.TestCase): def setUp(self): self.nick = "test" self.chan = '#brotherBot' self.mask = "brotherBox!~brotherBo@unaffiliated/brotherbox" self.h = harvester.HarvesterBot def test_fetch_cubeupload_share(self): msg = "http://cubeupload.com/im/YhUxlj.jpg" test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca" c = self.h._retrieve_content(self.h, self.mask, msg, self.chan) md5 = hashlib.md5() md5.update(c[0]['content']) self.assertEqual(md5.hexdigest(), test_hash) def test_fetch_cubeupload_raw(self): msg = "http://i.cubeupload.com/YhUxlj.jpg" test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca" c = self.h._retrieve_content(self.h, self.mask, msg, self.chan) md5 = hashlib.md5() md5.update(c[0]['content']) self.assertEqual(md5.hexdigest(), test_hash) if __name__ == '__main__': unittest.main()
<commit_before>#!/usr/bin/env python3 #-*- coding: utf-8 -*- import hashlib import unittest from harvester import harvester class CupeuploadTest(unittest.TestCase): def setUp(self): self.nick = "test" self.chan = '#brotherBot' self.mask = "brotherBox!~brotherBo@unaffiliated/brotherbox" self.h = harvester.HarvesterBot def test_fetch_cubeupload_share(self): msg = "http://cubeupload.com/im/YhUxlj.jpg" test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca" c = self.h._retrieve_content(self.h, self.mask, msg, self.chan) md5 = hashlib.md5() md5.update(c[0]['content']) self.assertEqual(md5.hexdigest(), test_hash) def test_fetch_cubeupload_raw(self): msg = "http://i.cubeupload.com/YhUxlj.jpg" test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca" c = self.h._retrieve_content(self.h, self.mask, msg, self.chan) md5 = hashlib.md5() md5.update(c[0]['content']) self.assertEqual(md5.hexdigest(), test_hash) if __name__ == '__main__': unittest.main() <commit_msg>Fix minor spelling error in cubeupload test<commit_after>
#!/usr/bin/env python3 #-*- coding: utf-8 -*- import hashlib import unittest from harvester import harvester class CubeuploadTest(unittest.TestCase): def setUp(self): self.nick = "test" self.chan = '#brotherBot' self.mask = "brotherBox!~brotherBo@unaffiliated/brotherbox" self.h = harvester.HarvesterBot def test_fetch_cubeupload_share(self): msg = "http://cubeupload.com/im/YhUxlj.jpg" test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca" c = self.h._retrieve_content(self.h, self.mask, msg, self.chan) md5 = hashlib.md5() md5.update(c[0]['content']) self.assertEqual(md5.hexdigest(), test_hash) def test_fetch_cubeupload_raw(self): msg = "http://i.cubeupload.com/YhUxlj.jpg" test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca" c = self.h._retrieve_content(self.h, self.mask, msg, self.chan) md5 = hashlib.md5() md5.update(c[0]['content']) self.assertEqual(md5.hexdigest(), test_hash) if __name__ == '__main__': unittest.main()
#!/usr/bin/env python3 #-*- coding: utf-8 -*- import hashlib import unittest from harvester import harvester class CupeuploadTest(unittest.TestCase): def setUp(self): self.nick = "test" self.chan = '#brotherBot' self.mask = "brotherBox!~brotherBo@unaffiliated/brotherbox" self.h = harvester.HarvesterBot def test_fetch_cubeupload_share(self): msg = "http://cubeupload.com/im/YhUxlj.jpg" test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca" c = self.h._retrieve_content(self.h, self.mask, msg, self.chan) md5 = hashlib.md5() md5.update(c[0]['content']) self.assertEqual(md5.hexdigest(), test_hash) def test_fetch_cubeupload_raw(self): msg = "http://i.cubeupload.com/YhUxlj.jpg" test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca" c = self.h._retrieve_content(self.h, self.mask, msg, self.chan) md5 = hashlib.md5() md5.update(c[0]['content']) self.assertEqual(md5.hexdigest(), test_hash) if __name__ == '__main__': unittest.main() Fix minor spelling error in cubeupload test#!/usr/bin/env python3 #-*- coding: utf-8 -*- import hashlib import unittest from harvester import harvester class CubeuploadTest(unittest.TestCase): def setUp(self): self.nick = "test" self.chan = '#brotherBot' self.mask = "brotherBox!~brotherBo@unaffiliated/brotherbox" self.h = harvester.HarvesterBot def test_fetch_cubeupload_share(self): msg = "http://cubeupload.com/im/YhUxlj.jpg" test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca" c = self.h._retrieve_content(self.h, self.mask, msg, self.chan) md5 = hashlib.md5() md5.update(c[0]['content']) self.assertEqual(md5.hexdigest(), test_hash) def test_fetch_cubeupload_raw(self): msg = "http://i.cubeupload.com/YhUxlj.jpg" test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca" c = self.h._retrieve_content(self.h, self.mask, msg, self.chan) md5 = hashlib.md5() md5.update(c[0]['content']) self.assertEqual(md5.hexdigest(), test_hash) if __name__ == '__main__': unittest.main()
<commit_before>#!/usr/bin/env python3 #-*- coding: utf-8 -*- import hashlib import unittest from harvester import harvester class CupeuploadTest(unittest.TestCase): def setUp(self): self.nick = "test" self.chan = '#brotherBot' self.mask = "brotherBox!~brotherBo@unaffiliated/brotherbox" self.h = harvester.HarvesterBot def test_fetch_cubeupload_share(self): msg = "http://cubeupload.com/im/YhUxlj.jpg" test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca" c = self.h._retrieve_content(self.h, self.mask, msg, self.chan) md5 = hashlib.md5() md5.update(c[0]['content']) self.assertEqual(md5.hexdigest(), test_hash) def test_fetch_cubeupload_raw(self): msg = "http://i.cubeupload.com/YhUxlj.jpg" test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca" c = self.h._retrieve_content(self.h, self.mask, msg, self.chan) md5 = hashlib.md5() md5.update(c[0]['content']) self.assertEqual(md5.hexdigest(), test_hash) if __name__ == '__main__': unittest.main() <commit_msg>Fix minor spelling error in cubeupload test<commit_after>#!/usr/bin/env python3 #-*- coding: utf-8 -*- import hashlib import unittest from harvester import harvester class CubeuploadTest(unittest.TestCase): def setUp(self): self.nick = "test" self.chan = '#brotherBot' self.mask = "brotherBox!~brotherBo@unaffiliated/brotherbox" self.h = harvester.HarvesterBot def test_fetch_cubeupload_share(self): msg = "http://cubeupload.com/im/YhUxlj.jpg" test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca" c = self.h._retrieve_content(self.h, self.mask, msg, self.chan) md5 = hashlib.md5() md5.update(c[0]['content']) self.assertEqual(md5.hexdigest(), test_hash) def test_fetch_cubeupload_raw(self): msg = "http://i.cubeupload.com/YhUxlj.jpg" test_hash = "3c1a8ef650f3c3c3c2f4dd115931c0ca" c = self.h._retrieve_content(self.h, self.mask, msg, self.chan) md5 = hashlib.md5() md5.update(c[0]['content']) self.assertEqual(md5.hexdigest(), test_hash) if __name__ == '__main__': unittest.main()
174c570d69d0958aa734794ffb7712ea37e70c6f
parse.py
parse.py
import sys import configparser def main(): config = configparser.ConfigParser(strict=False) try: section = sys.argv[1] config_key = sys.argv[2] config_value = sys.argv[3] except IndexError: print("Usage: cat test.ini | python parse.py <section> <option> <value>") sys.exit(1) try: config.read_string(''.join(sys.stdin)) """ Try to update or remove a config option from a section, or add the option to a new section. """ if section in config: if config_key in config[section]: if config_value == 'delete': config.remove_option(section, config_key) else: config[section][config_key] = config_value else: config.add_section(section) config.set(section, config_key, config_value) config.write(sys.stdout) except: print("There was an error parsing the config.") if __name__ == "__main__": main()
import sys import configparser def main(): config = configparser.ConfigParser(strict=False) try: section = sys.argv[1] config_key = sys.argv[2] config_value = sys.argv[3] except IndexError: print("Usage: cat test.ini | python parse.py <section> <option> <value>") sys.exit(1) try: config.read_string(''.join(sys.stdin)) """Try to update or remove a config option from a section, or add the option to a new section.""" if section in config: if config_key in config[section]: if config_value == 'delete': config.remove_option(section, config_key) else: config[section][config_key] = config_value else: config.set(section, config_key, config_value) else: config.add_section(section) config.set(section, config_key, config_value) config.write(sys.stdout) except: print("There was an error parsing the config.") if __name__ == "__main__": main()
Add new key to existing section.
Add new key to existing section.
Python
mit
tonigrigoriu/ini-parser
import sys import configparser def main(): config = configparser.ConfigParser(strict=False) try: section = sys.argv[1] config_key = sys.argv[2] config_value = sys.argv[3] except IndexError: print("Usage: cat test.ini | python parse.py <section> <option> <value>") sys.exit(1) try: config.read_string(''.join(sys.stdin)) """ Try to update or remove a config option from a section, or add the option to a new section. """ if section in config: if config_key in config[section]: if config_value == 'delete': config.remove_option(section, config_key) else: config[section][config_key] = config_value else: config.add_section(section) config.set(section, config_key, config_value) config.write(sys.stdout) except: print("There was an error parsing the config.") if __name__ == "__main__": main() Add new key to existing section.
import sys import configparser def main(): config = configparser.ConfigParser(strict=False) try: section = sys.argv[1] config_key = sys.argv[2] config_value = sys.argv[3] except IndexError: print("Usage: cat test.ini | python parse.py <section> <option> <value>") sys.exit(1) try: config.read_string(''.join(sys.stdin)) """Try to update or remove a config option from a section, or add the option to a new section.""" if section in config: if config_key in config[section]: if config_value == 'delete': config.remove_option(section, config_key) else: config[section][config_key] = config_value else: config.set(section, config_key, config_value) else: config.add_section(section) config.set(section, config_key, config_value) config.write(sys.stdout) except: print("There was an error parsing the config.") if __name__ == "__main__": main()
<commit_before>import sys import configparser def main(): config = configparser.ConfigParser(strict=False) try: section = sys.argv[1] config_key = sys.argv[2] config_value = sys.argv[3] except IndexError: print("Usage: cat test.ini | python parse.py <section> <option> <value>") sys.exit(1) try: config.read_string(''.join(sys.stdin)) """ Try to update or remove a config option from a section, or add the option to a new section. """ if section in config: if config_key in config[section]: if config_value == 'delete': config.remove_option(section, config_key) else: config[section][config_key] = config_value else: config.add_section(section) config.set(section, config_key, config_value) config.write(sys.stdout) except: print("There was an error parsing the config.") if __name__ == "__main__": main() <commit_msg>Add new key to existing section.<commit_after>
import sys import configparser def main(): config = configparser.ConfigParser(strict=False) try: section = sys.argv[1] config_key = sys.argv[2] config_value = sys.argv[3] except IndexError: print("Usage: cat test.ini | python parse.py <section> <option> <value>") sys.exit(1) try: config.read_string(''.join(sys.stdin)) """Try to update or remove a config option from a section, or add the option to a new section.""" if section in config: if config_key in config[section]: if config_value == 'delete': config.remove_option(section, config_key) else: config[section][config_key] = config_value else: config.set(section, config_key, config_value) else: config.add_section(section) config.set(section, config_key, config_value) config.write(sys.stdout) except: print("There was an error parsing the config.") if __name__ == "__main__": main()
import sys import configparser def main(): config = configparser.ConfigParser(strict=False) try: section = sys.argv[1] config_key = sys.argv[2] config_value = sys.argv[3] except IndexError: print("Usage: cat test.ini | python parse.py <section> <option> <value>") sys.exit(1) try: config.read_string(''.join(sys.stdin)) """ Try to update or remove a config option from a section, or add the option to a new section. """ if section in config: if config_key in config[section]: if config_value == 'delete': config.remove_option(section, config_key) else: config[section][config_key] = config_value else: config.add_section(section) config.set(section, config_key, config_value) config.write(sys.stdout) except: print("There was an error parsing the config.") if __name__ == "__main__": main() Add new key to existing section.import sys import configparser def main(): config = configparser.ConfigParser(strict=False) try: section = sys.argv[1] config_key = sys.argv[2] config_value = sys.argv[3] except IndexError: print("Usage: cat test.ini | python parse.py <section> <option> <value>") sys.exit(1) try: config.read_string(''.join(sys.stdin)) """Try to update or remove a config option from a section, or add the option to a new section.""" if section in config: if config_key in config[section]: if config_value == 'delete': config.remove_option(section, config_key) else: config[section][config_key] = config_value else: config.set(section, config_key, config_value) else: config.add_section(section) config.set(section, config_key, config_value) config.write(sys.stdout) except: print("There was an error parsing the config.") if __name__ == "__main__": main()
<commit_before>import sys import configparser def main(): config = configparser.ConfigParser(strict=False) try: section = sys.argv[1] config_key = sys.argv[2] config_value = sys.argv[3] except IndexError: print("Usage: cat test.ini | python parse.py <section> <option> <value>") sys.exit(1) try: config.read_string(''.join(sys.stdin)) """ Try to update or remove a config option from a section, or add the option to a new section. """ if section in config: if config_key in config[section]: if config_value == 'delete': config.remove_option(section, config_key) else: config[section][config_key] = config_value else: config.add_section(section) config.set(section, config_key, config_value) config.write(sys.stdout) except: print("There was an error parsing the config.") if __name__ == "__main__": main() <commit_msg>Add new key to existing section.<commit_after>import sys import configparser def main(): config = configparser.ConfigParser(strict=False) try: section = sys.argv[1] config_key = sys.argv[2] config_value = sys.argv[3] except IndexError: print("Usage: cat test.ini | python parse.py <section> <option> <value>") sys.exit(1) try: config.read_string(''.join(sys.stdin)) """Try to update or remove a config option from a section, or add the option to a new section.""" if section in config: if config_key in config[section]: if config_value == 'delete': config.remove_option(section, config_key) else: config[section][config_key] = config_value else: config.set(section, config_key, config_value) else: config.add_section(section) config.set(section, config_key, config_value) config.write(sys.stdout) except: print("There was an error parsing the config.") if __name__ == "__main__": main()
2e71f1a9deaf160ee666423e94ae526041cd32ff
engine.py
engine.py
# Use x, y coords for unit positions # (97, 56) ... (104, 56) # ... ... # (97, 49) ... (104, 49) # # Algebraic notation for a position is: # algebraic_pos = chr(x) + chr(y)
# Use x, y coords for unit positions # (97, 56) ... (104, 56) # ... ... # (97, 49) ... (104, 49) # # Algebraic notation for a position is: # algebraic_pos = chr(x) + chr(y) def _coord_to_algebraic(coord): x, y = coord return chr(x) + chr(y) def _algebraic_to_coord(algebraic): x, y = algebraic[0], algebraic[1] return ord(x), ord(y)
Add _coord_to_algebraic() and _algebraic_to_coord() to convert positions between (x, y) coords and 'a1' algebraic chess notation
Add _coord_to_algebraic() and _algebraic_to_coord() to convert positions between (x, y) coords and 'a1' algebraic chess notation
Python
mit
EyuelAbebe/gamer,EyuelAbebe/gamer
# Use x, y coords for unit positions # (97, 56) ... (104, 56) # ... ... # (97, 49) ... (104, 49) # # Algebraic notation for a position is: # algebraic_pos = chr(x) + chr(y) Add _coord_to_algebraic() and _algebraic_to_coord() to convert positions between (x, y) coords and 'a1' algebraic chess notation
# Use x, y coords for unit positions # (97, 56) ... (104, 56) # ... ... # (97, 49) ... (104, 49) # # Algebraic notation for a position is: # algebraic_pos = chr(x) + chr(y) def _coord_to_algebraic(coord): x, y = coord return chr(x) + chr(y) def _algebraic_to_coord(algebraic): x, y = algebraic[0], algebraic[1] return ord(x), ord(y)
<commit_before># Use x, y coords for unit positions # (97, 56) ... (104, 56) # ... ... # (97, 49) ... (104, 49) # # Algebraic notation for a position is: # algebraic_pos = chr(x) + chr(y) <commit_msg>Add _coord_to_algebraic() and _algebraic_to_coord() to convert positions between (x, y) coords and 'a1' algebraic chess notation<commit_after>
# Use x, y coords for unit positions # (97, 56) ... (104, 56) # ... ... # (97, 49) ... (104, 49) # # Algebraic notation for a position is: # algebraic_pos = chr(x) + chr(y) def _coord_to_algebraic(coord): x, y = coord return chr(x) + chr(y) def _algebraic_to_coord(algebraic): x, y = algebraic[0], algebraic[1] return ord(x), ord(y)
# Use x, y coords for unit positions # (97, 56) ... (104, 56) # ... ... # (97, 49) ... (104, 49) # # Algebraic notation for a position is: # algebraic_pos = chr(x) + chr(y) Add _coord_to_algebraic() and _algebraic_to_coord() to convert positions between (x, y) coords and 'a1' algebraic chess notation# Use x, y coords for unit positions # (97, 56) ... (104, 56) # ... ... # (97, 49) ... (104, 49) # # Algebraic notation for a position is: # algebraic_pos = chr(x) + chr(y) def _coord_to_algebraic(coord): x, y = coord return chr(x) + chr(y) def _algebraic_to_coord(algebraic): x, y = algebraic[0], algebraic[1] return ord(x), ord(y)
<commit_before># Use x, y coords for unit positions # (97, 56) ... (104, 56) # ... ... # (97, 49) ... (104, 49) # # Algebraic notation for a position is: # algebraic_pos = chr(x) + chr(y) <commit_msg>Add _coord_to_algebraic() and _algebraic_to_coord() to convert positions between (x, y) coords and 'a1' algebraic chess notation<commit_after># Use x, y coords for unit positions # (97, 56) ... (104, 56) # ... ... # (97, 49) ... (104, 49) # # Algebraic notation for a position is: # algebraic_pos = chr(x) + chr(y) def _coord_to_algebraic(coord): x, y = coord return chr(x) + chr(y) def _algebraic_to_coord(algebraic): x, y = algebraic[0], algebraic[1] return ord(x), ord(y)
215e37fce8b3fedf7bf31bf7c6393271c84141a2
src/tapdisk/plugin.py
src/tapdisk/plugin.py
#!/usr/bin/env python import os import sys import xapi import xapi.plugin from xapi.storage.datapath import log class Implementation(xapi.plugin.Plugin_skeleton): def query(self, dbg): return { "plugin": "tapdisk", "name": "The tapdisk user-space datapath plugin", "description": ("This plugin manages and configures tapdisk" " instances backend by either raw or vhd" " format files"), "vendor": "Citrix", "copyright": "(C) 2015 Citrix Inc", "version": "3.0", "required_api_version": "3.0", "features": [ ], "configuration": {}, "required_cluster_stack": []} if __name__ == "__main__": log.log_call_argv() cmd = xapi.plugin.Plugin_commandline(Implementation()) base = os.path.basename(sys.argv[0]) if base == "Plugin.Query": cmd.query() else: raise xapi.plugin.Unimplemented(base)
#!/usr/bin/env python import os import sys import xapi import xapi.storage.api.plugin from xapi.storage import log class Implementation(xapi.storage.api.plugin.Plugin_skeleton): def query(self, dbg): return { "plugin": "tapdisk", "name": "The tapdisk user-space datapath plugin", "description": ("This plugin manages and configures tapdisk" " instances backend by either raw or vhd" " format files"), "vendor": "Citrix", "copyright": "(C) 2015 Citrix Inc", "version": "3.0", "required_api_version": "3.0", "features": [ ], "configuration": {}, "required_cluster_stack": []} if __name__ == "__main__": log.log_call_argv() cmd = xapi.storage.api.plugin.Plugin_commandline(Implementation()) base = os.path.basename(sys.argv[0]) if base == "Plugin.Query": cmd.query() else: raise xapi.storage.api.plugin.Unimplemented(base)
Use the new xapi.storage package structure
Use the new xapi.storage package structure Signed-off-by: David Scott <63c9eb0ea83039690fefa11afe17873ba8278a56@eu.citrix.com>
Python
lgpl-2.1
xapi-project/xapi-storage-datapath-plugins,stefanopanella/xapi-storage-plugins,djs55/xapi-storage-datapath-plugins,stefanopanella/xapi-storage-plugins,stefanopanella/xapi-storage-plugins,jjd27/xapi-storage-datapath-plugins,robertbreker/xapi-storage-datapath-plugins
#!/usr/bin/env python import os import sys import xapi import xapi.plugin from xapi.storage.datapath import log class Implementation(xapi.plugin.Plugin_skeleton): def query(self, dbg): return { "plugin": "tapdisk", "name": "The tapdisk user-space datapath plugin", "description": ("This plugin manages and configures tapdisk" " instances backend by either raw or vhd" " format files"), "vendor": "Citrix", "copyright": "(C) 2015 Citrix Inc", "version": "3.0", "required_api_version": "3.0", "features": [ ], "configuration": {}, "required_cluster_stack": []} if __name__ == "__main__": log.log_call_argv() cmd = xapi.plugin.Plugin_commandline(Implementation()) base = os.path.basename(sys.argv[0]) if base == "Plugin.Query": cmd.query() else: raise xapi.plugin.Unimplemented(base) Use the new xapi.storage package structure Signed-off-by: David Scott <63c9eb0ea83039690fefa11afe17873ba8278a56@eu.citrix.com>
#!/usr/bin/env python import os import sys import xapi import xapi.storage.api.plugin from xapi.storage import log class Implementation(xapi.storage.api.plugin.Plugin_skeleton): def query(self, dbg): return { "plugin": "tapdisk", "name": "The tapdisk user-space datapath plugin", "description": ("This plugin manages and configures tapdisk" " instances backend by either raw or vhd" " format files"), "vendor": "Citrix", "copyright": "(C) 2015 Citrix Inc", "version": "3.0", "required_api_version": "3.0", "features": [ ], "configuration": {}, "required_cluster_stack": []} if __name__ == "__main__": log.log_call_argv() cmd = xapi.storage.api.plugin.Plugin_commandline(Implementation()) base = os.path.basename(sys.argv[0]) if base == "Plugin.Query": cmd.query() else: raise xapi.storage.api.plugin.Unimplemented(base)
<commit_before>#!/usr/bin/env python import os import sys import xapi import xapi.plugin from xapi.storage.datapath import log class Implementation(xapi.plugin.Plugin_skeleton): def query(self, dbg): return { "plugin": "tapdisk", "name": "The tapdisk user-space datapath plugin", "description": ("This plugin manages and configures tapdisk" " instances backend by either raw or vhd" " format files"), "vendor": "Citrix", "copyright": "(C) 2015 Citrix Inc", "version": "3.0", "required_api_version": "3.0", "features": [ ], "configuration": {}, "required_cluster_stack": []} if __name__ == "__main__": log.log_call_argv() cmd = xapi.plugin.Plugin_commandline(Implementation()) base = os.path.basename(sys.argv[0]) if base == "Plugin.Query": cmd.query() else: raise xapi.plugin.Unimplemented(base) <commit_msg>Use the new xapi.storage package structure Signed-off-by: David Scott <63c9eb0ea83039690fefa11afe17873ba8278a56@eu.citrix.com><commit_after>
#!/usr/bin/env python import os import sys import xapi import xapi.storage.api.plugin from xapi.storage import log class Implementation(xapi.storage.api.plugin.Plugin_skeleton): def query(self, dbg): return { "plugin": "tapdisk", "name": "The tapdisk user-space datapath plugin", "description": ("This plugin manages and configures tapdisk" " instances backend by either raw or vhd" " format files"), "vendor": "Citrix", "copyright": "(C) 2015 Citrix Inc", "version": "3.0", "required_api_version": "3.0", "features": [ ], "configuration": {}, "required_cluster_stack": []} if __name__ == "__main__": log.log_call_argv() cmd = xapi.storage.api.plugin.Plugin_commandline(Implementation()) base = os.path.basename(sys.argv[0]) if base == "Plugin.Query": cmd.query() else: raise xapi.storage.api.plugin.Unimplemented(base)
#!/usr/bin/env python import os import sys import xapi import xapi.plugin from xapi.storage.datapath import log class Implementation(xapi.plugin.Plugin_skeleton): def query(self, dbg): return { "plugin": "tapdisk", "name": "The tapdisk user-space datapath plugin", "description": ("This plugin manages and configures tapdisk" " instances backend by either raw or vhd" " format files"), "vendor": "Citrix", "copyright": "(C) 2015 Citrix Inc", "version": "3.0", "required_api_version": "3.0", "features": [ ], "configuration": {}, "required_cluster_stack": []} if __name__ == "__main__": log.log_call_argv() cmd = xapi.plugin.Plugin_commandline(Implementation()) base = os.path.basename(sys.argv[0]) if base == "Plugin.Query": cmd.query() else: raise xapi.plugin.Unimplemented(base) Use the new xapi.storage package structure Signed-off-by: David Scott <63c9eb0ea83039690fefa11afe17873ba8278a56@eu.citrix.com>#!/usr/bin/env python import os import sys import xapi import xapi.storage.api.plugin from xapi.storage import log class Implementation(xapi.storage.api.plugin.Plugin_skeleton): def query(self, dbg): return { "plugin": "tapdisk", "name": "The tapdisk user-space datapath plugin", "description": ("This plugin manages and configures tapdisk" " instances backend by either raw or vhd" " format files"), "vendor": "Citrix", "copyright": "(C) 2015 Citrix Inc", "version": "3.0", "required_api_version": "3.0", "features": [ ], "configuration": {}, "required_cluster_stack": []} if __name__ == "__main__": log.log_call_argv() cmd = xapi.storage.api.plugin.Plugin_commandline(Implementation()) base = os.path.basename(sys.argv[0]) if base == "Plugin.Query": cmd.query() else: raise xapi.storage.api.plugin.Unimplemented(base)
<commit_before>#!/usr/bin/env python import os import sys import xapi import xapi.plugin from xapi.storage.datapath import log class Implementation(xapi.plugin.Plugin_skeleton): def query(self, dbg): return { "plugin": "tapdisk", "name": "The tapdisk user-space datapath plugin", "description": ("This plugin manages and configures tapdisk" " instances backend by either raw or vhd" " format files"), "vendor": "Citrix", "copyright": "(C) 2015 Citrix Inc", "version": "3.0", "required_api_version": "3.0", "features": [ ], "configuration": {}, "required_cluster_stack": []} if __name__ == "__main__": log.log_call_argv() cmd = xapi.plugin.Plugin_commandline(Implementation()) base = os.path.basename(sys.argv[0]) if base == "Plugin.Query": cmd.query() else: raise xapi.plugin.Unimplemented(base) <commit_msg>Use the new xapi.storage package structure Signed-off-by: David Scott <63c9eb0ea83039690fefa11afe17873ba8278a56@eu.citrix.com><commit_after>#!/usr/bin/env python import os import sys import xapi import xapi.storage.api.plugin from xapi.storage import log class Implementation(xapi.storage.api.plugin.Plugin_skeleton): def query(self, dbg): return { "plugin": "tapdisk", "name": "The tapdisk user-space datapath plugin", "description": ("This plugin manages and configures tapdisk" " instances backend by either raw or vhd" " format files"), "vendor": "Citrix", "copyright": "(C) 2015 Citrix Inc", "version": "3.0", "required_api_version": "3.0", "features": [ ], "configuration": {}, "required_cluster_stack": []} if __name__ == "__main__": log.log_call_argv() cmd = xapi.storage.api.plugin.Plugin_commandline(Implementation()) base = os.path.basename(sys.argv[0]) if base == "Plugin.Query": cmd.query() else: raise xapi.storage.api.plugin.Unimplemented(base)
273b463d8c86e60639b0c0ff54dad3e376a93060
correctiv_justizgelder/views.py
correctiv_justizgelder/views.py
from django.views.generic import ListView, DetailView from .forms import OrganisationSearchForm from .models import Organisation from .search_indexes import OrganisationIndex from .search_utils import SearchPaginator class OrganisationSearchView(ListView): template_name = 'justizgelder/search.html' paginate_by = 15 paginator_class = SearchPaginator def get_queryset(self): self.form = OrganisationSearchForm(self.request.GET) self.result = self.form.search(size=self.paginate_by) return self.result def get_context_data(self, **kwargs): context = super(OrganisationSearchView, self).get_context_data(**kwargs) context['result'] = self.result context['query'] = self.request.GET.get('q') context['form'] = self.form context['base_template'] = 'justizgelder/search_base.html' if self.request.GET.get('embed'): context['base_template'] = 'justizgelder/embed_base.html' return context class OrganisationDetail(DetailView): template_name = 'justizgelder/organisation_detail.html' model = Organisation def get_context_data(self, **kwargs): context = super(OrganisationDetail, self).get_context_data(**kwargs) idx = OrganisationIndex() context['mlt'] = idx.search(self.object.name, size=15, sort=False, aggregations=False ) return context
from django.views.generic import ListView, DetailView from .forms import OrganisationSearchForm from .models import Organisation from .search_indexes import OrganisationIndex from .search_utils import SearchPaginator class OrganisationSearchView(ListView): template_name = 'justizgelder/search.html' paginate_by = 25 paginator_class = SearchPaginator def get_queryset(self): self.form = OrganisationSearchForm(self.request.GET) self.result = self.form.search(size=self.paginate_by) return self.result def get_context_data(self, **kwargs): context = super(OrganisationSearchView, self).get_context_data(**kwargs) context['result'] = self.result context['query'] = self.request.GET.get('q') context['form'] = self.form context['base_template'] = 'justizgelder/search_base.html' if self.request.GET.get('embed'): context['base_template'] = 'justizgelder/embed_base.html' return context class OrganisationDetail(DetailView): template_name = 'justizgelder/organisation_detail.html' model = Organisation def get_context_data(self, **kwargs): context = super(OrganisationDetail, self).get_context_data(**kwargs) idx = OrganisationIndex() context['mlt'] = idx.search(self.object.name, size=15, sort=False, aggregations=False ) return context
Increase results to 25 per page to make page look nicer
Increase results to 25 per page to make page look nicer
Python
mit
correctiv/correctiv-justizgelder,correctiv/correctiv-justizgelder
from django.views.generic import ListView, DetailView from .forms import OrganisationSearchForm from .models import Organisation from .search_indexes import OrganisationIndex from .search_utils import SearchPaginator class OrganisationSearchView(ListView): template_name = 'justizgelder/search.html' paginate_by = 15 paginator_class = SearchPaginator def get_queryset(self): self.form = OrganisationSearchForm(self.request.GET) self.result = self.form.search(size=self.paginate_by) return self.result def get_context_data(self, **kwargs): context = super(OrganisationSearchView, self).get_context_data(**kwargs) context['result'] = self.result context['query'] = self.request.GET.get('q') context['form'] = self.form context['base_template'] = 'justizgelder/search_base.html' if self.request.GET.get('embed'): context['base_template'] = 'justizgelder/embed_base.html' return context class OrganisationDetail(DetailView): template_name = 'justizgelder/organisation_detail.html' model = Organisation def get_context_data(self, **kwargs): context = super(OrganisationDetail, self).get_context_data(**kwargs) idx = OrganisationIndex() context['mlt'] = idx.search(self.object.name, size=15, sort=False, aggregations=False ) return context Increase results to 25 per page to make page look nicer
from django.views.generic import ListView, DetailView from .forms import OrganisationSearchForm from .models import Organisation from .search_indexes import OrganisationIndex from .search_utils import SearchPaginator class OrganisationSearchView(ListView): template_name = 'justizgelder/search.html' paginate_by = 25 paginator_class = SearchPaginator def get_queryset(self): self.form = OrganisationSearchForm(self.request.GET) self.result = self.form.search(size=self.paginate_by) return self.result def get_context_data(self, **kwargs): context = super(OrganisationSearchView, self).get_context_data(**kwargs) context['result'] = self.result context['query'] = self.request.GET.get('q') context['form'] = self.form context['base_template'] = 'justizgelder/search_base.html' if self.request.GET.get('embed'): context['base_template'] = 'justizgelder/embed_base.html' return context class OrganisationDetail(DetailView): template_name = 'justizgelder/organisation_detail.html' model = Organisation def get_context_data(self, **kwargs): context = super(OrganisationDetail, self).get_context_data(**kwargs) idx = OrganisationIndex() context['mlt'] = idx.search(self.object.name, size=15, sort=False, aggregations=False ) return context
<commit_before>from django.views.generic import ListView, DetailView from .forms import OrganisationSearchForm from .models import Organisation from .search_indexes import OrganisationIndex from .search_utils import SearchPaginator class OrganisationSearchView(ListView): template_name = 'justizgelder/search.html' paginate_by = 15 paginator_class = SearchPaginator def get_queryset(self): self.form = OrganisationSearchForm(self.request.GET) self.result = self.form.search(size=self.paginate_by) return self.result def get_context_data(self, **kwargs): context = super(OrganisationSearchView, self).get_context_data(**kwargs) context['result'] = self.result context['query'] = self.request.GET.get('q') context['form'] = self.form context['base_template'] = 'justizgelder/search_base.html' if self.request.GET.get('embed'): context['base_template'] = 'justizgelder/embed_base.html' return context class OrganisationDetail(DetailView): template_name = 'justizgelder/organisation_detail.html' model = Organisation def get_context_data(self, **kwargs): context = super(OrganisationDetail, self).get_context_data(**kwargs) idx = OrganisationIndex() context['mlt'] = idx.search(self.object.name, size=15, sort=False, aggregations=False ) return context <commit_msg>Increase results to 25 per page to make page look nicer<commit_after>
from django.views.generic import ListView, DetailView from .forms import OrganisationSearchForm from .models import Organisation from .search_indexes import OrganisationIndex from .search_utils import SearchPaginator class OrganisationSearchView(ListView): template_name = 'justizgelder/search.html' paginate_by = 25 paginator_class = SearchPaginator def get_queryset(self): self.form = OrganisationSearchForm(self.request.GET) self.result = self.form.search(size=self.paginate_by) return self.result def get_context_data(self, **kwargs): context = super(OrganisationSearchView, self).get_context_data(**kwargs) context['result'] = self.result context['query'] = self.request.GET.get('q') context['form'] = self.form context['base_template'] = 'justizgelder/search_base.html' if self.request.GET.get('embed'): context['base_template'] = 'justizgelder/embed_base.html' return context class OrganisationDetail(DetailView): template_name = 'justizgelder/organisation_detail.html' model = Organisation def get_context_data(self, **kwargs): context = super(OrganisationDetail, self).get_context_data(**kwargs) idx = OrganisationIndex() context['mlt'] = idx.search(self.object.name, size=15, sort=False, aggregations=False ) return context
from django.views.generic import ListView, DetailView from .forms import OrganisationSearchForm from .models import Organisation from .search_indexes import OrganisationIndex from .search_utils import SearchPaginator class OrganisationSearchView(ListView): template_name = 'justizgelder/search.html' paginate_by = 15 paginator_class = SearchPaginator def get_queryset(self): self.form = OrganisationSearchForm(self.request.GET) self.result = self.form.search(size=self.paginate_by) return self.result def get_context_data(self, **kwargs): context = super(OrganisationSearchView, self).get_context_data(**kwargs) context['result'] = self.result context['query'] = self.request.GET.get('q') context['form'] = self.form context['base_template'] = 'justizgelder/search_base.html' if self.request.GET.get('embed'): context['base_template'] = 'justizgelder/embed_base.html' return context class OrganisationDetail(DetailView): template_name = 'justizgelder/organisation_detail.html' model = Organisation def get_context_data(self, **kwargs): context = super(OrganisationDetail, self).get_context_data(**kwargs) idx = OrganisationIndex() context['mlt'] = idx.search(self.object.name, size=15, sort=False, aggregations=False ) return context Increase results to 25 per page to make page look nicerfrom django.views.generic import ListView, DetailView from .forms import OrganisationSearchForm from .models import Organisation from .search_indexes import OrganisationIndex from .search_utils import SearchPaginator class OrganisationSearchView(ListView): template_name = 'justizgelder/search.html' paginate_by = 25 paginator_class = SearchPaginator def get_queryset(self): self.form = OrganisationSearchForm(self.request.GET) self.result = self.form.search(size=self.paginate_by) return self.result def get_context_data(self, **kwargs): context = super(OrganisationSearchView, self).get_context_data(**kwargs) context['result'] = self.result context['query'] = self.request.GET.get('q') context['form'] = self.form context['base_template'] = 'justizgelder/search_base.html' if self.request.GET.get('embed'): context['base_template'] = 'justizgelder/embed_base.html' return context class OrganisationDetail(DetailView): template_name = 'justizgelder/organisation_detail.html' model = Organisation def get_context_data(self, **kwargs): context = super(OrganisationDetail, self).get_context_data(**kwargs) idx = OrganisationIndex() context['mlt'] = idx.search(self.object.name, size=15, sort=False, aggregations=False ) return context
<commit_before>from django.views.generic import ListView, DetailView from .forms import OrganisationSearchForm from .models import Organisation from .search_indexes import OrganisationIndex from .search_utils import SearchPaginator class OrganisationSearchView(ListView): template_name = 'justizgelder/search.html' paginate_by = 15 paginator_class = SearchPaginator def get_queryset(self): self.form = OrganisationSearchForm(self.request.GET) self.result = self.form.search(size=self.paginate_by) return self.result def get_context_data(self, **kwargs): context = super(OrganisationSearchView, self).get_context_data(**kwargs) context['result'] = self.result context['query'] = self.request.GET.get('q') context['form'] = self.form context['base_template'] = 'justizgelder/search_base.html' if self.request.GET.get('embed'): context['base_template'] = 'justizgelder/embed_base.html' return context class OrganisationDetail(DetailView): template_name = 'justizgelder/organisation_detail.html' model = Organisation def get_context_data(self, **kwargs): context = super(OrganisationDetail, self).get_context_data(**kwargs) idx = OrganisationIndex() context['mlt'] = idx.search(self.object.name, size=15, sort=False, aggregations=False ) return context <commit_msg>Increase results to 25 per page to make page look nicer<commit_after>from django.views.generic import ListView, DetailView from .forms import OrganisationSearchForm from .models import Organisation from .search_indexes import OrganisationIndex from .search_utils import SearchPaginator class OrganisationSearchView(ListView): template_name = 'justizgelder/search.html' paginate_by = 25 paginator_class = SearchPaginator def get_queryset(self): self.form = OrganisationSearchForm(self.request.GET) self.result = self.form.search(size=self.paginate_by) return self.result def get_context_data(self, **kwargs): context = super(OrganisationSearchView, self).get_context_data(**kwargs) context['result'] = self.result context['query'] = self.request.GET.get('q') context['form'] = self.form context['base_template'] = 'justizgelder/search_base.html' if self.request.GET.get('embed'): context['base_template'] = 'justizgelder/embed_base.html' return context class OrganisationDetail(DetailView): template_name = 'justizgelder/organisation_detail.html' model = Organisation def get_context_data(self, **kwargs): context = super(OrganisationDetail, self).get_context_data(**kwargs) idx = OrganisationIndex() context['mlt'] = idx.search(self.object.name, size=15, sort=False, aggregations=False ) return context
90b1567ee8e1906b1d1724cf63cf8d383530da29
nimp/commands/cis_tomat_mining.py
nimp/commands/cis_tomat_mining.py
# -*- coding: utf-8 -*- from nimp.commands._cis_command import * from nimp.utilities.ue3 import * from nimp.utilities.deployment import * from nimp.utilities.file_mapper import * #------------------------------------------------------------------------------- class CisTomatMining(CisCommand): abstract = 0 def __init__(self): CisCommand.__init__(self, 'cis-tomat-mining', 'Mines UE3 content into Tomat') #--------------------------------------------------------------------------- def cis_configure_arguments(self, env, parser): return True #--------------------------------------------------------------------------- def _cis_run(self, env): if call_process('.', ['pacman', '-S', '--noconfirm', 'repman']) != 0: return False if call_process('.', ['repman', 'add', 'dont-nod', 'http://pacman']) != 0: return False if call_process('.', ['pacman', '-S', '--noconfirm', '--needed', 'tomat-console']) != 0: return False return call_process('.', [ 'TomatConsole', 'ImportFromUnreal', '--RepositoryUri', 'sql://mining@console', '--UnrealEnginePath', 'Binaries/Win64/ExampleGame.exe' ])
# -*- coding: utf-8 -*- from nimp.commands._cis_command import * from nimp.utilities.ue3 import * from nimp.utilities.deployment import * from nimp.utilities.file_mapper import * import tempfile import shutil #------------------------------------------------------------------------------- class CisTomatMining(CisCommand): abstract = 0 def __init__(self): CisCommand.__init__(self, 'cis-tomat-mining', 'Mines UE3 content into Tomat') #--------------------------------------------------------------------------- def cis_configure_arguments(self, env, parser): return True #--------------------------------------------------------------------------- def _cis_run(self, env): if call_process('.', ['pacman', '-S', '--noconfirm', 'repman']) != 0: return False if call_process('.', ['repman', 'add', 'dont-nod', 'http://pacman/']) != 0: return False if call_process('.', ['pacman', '-Scc', '--noconfirm']) != 0: return False if call_process('.', ['pacman', '-S', '--noconfirm', '--force', 'tomat-console']) != 0: return False tmpdir = tempfile.mkdtemp() success = True if success and env.is_ue3: success = ue3_commandlet(env.game, 'dnetomatminingcommandlet', [ tmpdir ]) if success: success = call_process('.', [ 'TomatConsole', 'ImportFromUnreal', '--RepositoryUri', 'sql://mining@console', '--TmpDirectory', tmpdir ]) == 0 # Clean up after ourselves shutil.rmtree(tmpdir) return success
Split the Tomat mining commandlet into an Unreal part and a TomatConsole part.
Split the Tomat mining commandlet into an Unreal part and a TomatConsole part.
Python
mit
dontnod/nimp
# -*- coding: utf-8 -*- from nimp.commands._cis_command import * from nimp.utilities.ue3 import * from nimp.utilities.deployment import * from nimp.utilities.file_mapper import * #------------------------------------------------------------------------------- class CisTomatMining(CisCommand): abstract = 0 def __init__(self): CisCommand.__init__(self, 'cis-tomat-mining', 'Mines UE3 content into Tomat') #--------------------------------------------------------------------------- def cis_configure_arguments(self, env, parser): return True #--------------------------------------------------------------------------- def _cis_run(self, env): if call_process('.', ['pacman', '-S', '--noconfirm', 'repman']) != 0: return False if call_process('.', ['repman', 'add', 'dont-nod', 'http://pacman']) != 0: return False if call_process('.', ['pacman', '-S', '--noconfirm', '--needed', 'tomat-console']) != 0: return False return call_process('.', [ 'TomatConsole', 'ImportFromUnreal', '--RepositoryUri', 'sql://mining@console', '--UnrealEnginePath', 'Binaries/Win64/ExampleGame.exe' ]) Split the Tomat mining commandlet into an Unreal part and a TomatConsole part.
# -*- coding: utf-8 -*- from nimp.commands._cis_command import * from nimp.utilities.ue3 import * from nimp.utilities.deployment import * from nimp.utilities.file_mapper import * import tempfile import shutil #------------------------------------------------------------------------------- class CisTomatMining(CisCommand): abstract = 0 def __init__(self): CisCommand.__init__(self, 'cis-tomat-mining', 'Mines UE3 content into Tomat') #--------------------------------------------------------------------------- def cis_configure_arguments(self, env, parser): return True #--------------------------------------------------------------------------- def _cis_run(self, env): if call_process('.', ['pacman', '-S', '--noconfirm', 'repman']) != 0: return False if call_process('.', ['repman', 'add', 'dont-nod', 'http://pacman/']) != 0: return False if call_process('.', ['pacman', '-Scc', '--noconfirm']) != 0: return False if call_process('.', ['pacman', '-S', '--noconfirm', '--force', 'tomat-console']) != 0: return False tmpdir = tempfile.mkdtemp() success = True if success and env.is_ue3: success = ue3_commandlet(env.game, 'dnetomatminingcommandlet', [ tmpdir ]) if success: success = call_process('.', [ 'TomatConsole', 'ImportFromUnreal', '--RepositoryUri', 'sql://mining@console', '--TmpDirectory', tmpdir ]) == 0 # Clean up after ourselves shutil.rmtree(tmpdir) return success
<commit_before># -*- coding: utf-8 -*- from nimp.commands._cis_command import * from nimp.utilities.ue3 import * from nimp.utilities.deployment import * from nimp.utilities.file_mapper import * #------------------------------------------------------------------------------- class CisTomatMining(CisCommand): abstract = 0 def __init__(self): CisCommand.__init__(self, 'cis-tomat-mining', 'Mines UE3 content into Tomat') #--------------------------------------------------------------------------- def cis_configure_arguments(self, env, parser): return True #--------------------------------------------------------------------------- def _cis_run(self, env): if call_process('.', ['pacman', '-S', '--noconfirm', 'repman']) != 0: return False if call_process('.', ['repman', 'add', 'dont-nod', 'http://pacman']) != 0: return False if call_process('.', ['pacman', '-S', '--noconfirm', '--needed', 'tomat-console']) != 0: return False return call_process('.', [ 'TomatConsole', 'ImportFromUnreal', '--RepositoryUri', 'sql://mining@console', '--UnrealEnginePath', 'Binaries/Win64/ExampleGame.exe' ]) <commit_msg>Split the Tomat mining commandlet into an Unreal part and a TomatConsole part.<commit_after>
# -*- coding: utf-8 -*- from nimp.commands._cis_command import * from nimp.utilities.ue3 import * from nimp.utilities.deployment import * from nimp.utilities.file_mapper import * import tempfile import shutil #------------------------------------------------------------------------------- class CisTomatMining(CisCommand): abstract = 0 def __init__(self): CisCommand.__init__(self, 'cis-tomat-mining', 'Mines UE3 content into Tomat') #--------------------------------------------------------------------------- def cis_configure_arguments(self, env, parser): return True #--------------------------------------------------------------------------- def _cis_run(self, env): if call_process('.', ['pacman', '-S', '--noconfirm', 'repman']) != 0: return False if call_process('.', ['repman', 'add', 'dont-nod', 'http://pacman/']) != 0: return False if call_process('.', ['pacman', '-Scc', '--noconfirm']) != 0: return False if call_process('.', ['pacman', '-S', '--noconfirm', '--force', 'tomat-console']) != 0: return False tmpdir = tempfile.mkdtemp() success = True if success and env.is_ue3: success = ue3_commandlet(env.game, 'dnetomatminingcommandlet', [ tmpdir ]) if success: success = call_process('.', [ 'TomatConsole', 'ImportFromUnreal', '--RepositoryUri', 'sql://mining@console', '--TmpDirectory', tmpdir ]) == 0 # Clean up after ourselves shutil.rmtree(tmpdir) return success
# -*- coding: utf-8 -*- from nimp.commands._cis_command import * from nimp.utilities.ue3 import * from nimp.utilities.deployment import * from nimp.utilities.file_mapper import * #------------------------------------------------------------------------------- class CisTomatMining(CisCommand): abstract = 0 def __init__(self): CisCommand.__init__(self, 'cis-tomat-mining', 'Mines UE3 content into Tomat') #--------------------------------------------------------------------------- def cis_configure_arguments(self, env, parser): return True #--------------------------------------------------------------------------- def _cis_run(self, env): if call_process('.', ['pacman', '-S', '--noconfirm', 'repman']) != 0: return False if call_process('.', ['repman', 'add', 'dont-nod', 'http://pacman']) != 0: return False if call_process('.', ['pacman', '-S', '--noconfirm', '--needed', 'tomat-console']) != 0: return False return call_process('.', [ 'TomatConsole', 'ImportFromUnreal', '--RepositoryUri', 'sql://mining@console', '--UnrealEnginePath', 'Binaries/Win64/ExampleGame.exe' ]) Split the Tomat mining commandlet into an Unreal part and a TomatConsole part.# -*- coding: utf-8 -*- from nimp.commands._cis_command import * from nimp.utilities.ue3 import * from nimp.utilities.deployment import * from nimp.utilities.file_mapper import * import tempfile import shutil #------------------------------------------------------------------------------- class CisTomatMining(CisCommand): abstract = 0 def __init__(self): CisCommand.__init__(self, 'cis-tomat-mining', 'Mines UE3 content into Tomat') #--------------------------------------------------------------------------- def cis_configure_arguments(self, env, parser): return True #--------------------------------------------------------------------------- def _cis_run(self, env): if call_process('.', ['pacman', '-S', '--noconfirm', 'repman']) != 0: return False if call_process('.', ['repman', 'add', 'dont-nod', 'http://pacman/']) != 0: return False if call_process('.', ['pacman', '-Scc', '--noconfirm']) != 0: return False if call_process('.', ['pacman', '-S', '--noconfirm', '--force', 'tomat-console']) != 0: return False tmpdir = tempfile.mkdtemp() success = True if success and env.is_ue3: success = ue3_commandlet(env.game, 'dnetomatminingcommandlet', [ tmpdir ]) if success: success = call_process('.', [ 'TomatConsole', 'ImportFromUnreal', '--RepositoryUri', 'sql://mining@console', '--TmpDirectory', tmpdir ]) == 0 # Clean up after ourselves shutil.rmtree(tmpdir) return success
<commit_before># -*- coding: utf-8 -*- from nimp.commands._cis_command import * from nimp.utilities.ue3 import * from nimp.utilities.deployment import * from nimp.utilities.file_mapper import * #------------------------------------------------------------------------------- class CisTomatMining(CisCommand): abstract = 0 def __init__(self): CisCommand.__init__(self, 'cis-tomat-mining', 'Mines UE3 content into Tomat') #--------------------------------------------------------------------------- def cis_configure_arguments(self, env, parser): return True #--------------------------------------------------------------------------- def _cis_run(self, env): if call_process('.', ['pacman', '-S', '--noconfirm', 'repman']) != 0: return False if call_process('.', ['repman', 'add', 'dont-nod', 'http://pacman']) != 0: return False if call_process('.', ['pacman', '-S', '--noconfirm', '--needed', 'tomat-console']) != 0: return False return call_process('.', [ 'TomatConsole', 'ImportFromUnreal', '--RepositoryUri', 'sql://mining@console', '--UnrealEnginePath', 'Binaries/Win64/ExampleGame.exe' ]) <commit_msg>Split the Tomat mining commandlet into an Unreal part and a TomatConsole part.<commit_after># -*- coding: utf-8 -*- from nimp.commands._cis_command import * from nimp.utilities.ue3 import * from nimp.utilities.deployment import * from nimp.utilities.file_mapper import * import tempfile import shutil #------------------------------------------------------------------------------- class CisTomatMining(CisCommand): abstract = 0 def __init__(self): CisCommand.__init__(self, 'cis-tomat-mining', 'Mines UE3 content into Tomat') #--------------------------------------------------------------------------- def cis_configure_arguments(self, env, parser): return True #--------------------------------------------------------------------------- def _cis_run(self, env): if call_process('.', ['pacman', '-S', '--noconfirm', 'repman']) != 0: return False if call_process('.', ['repman', 'add', 'dont-nod', 'http://pacman/']) != 0: return False if call_process('.', ['pacman', '-Scc', '--noconfirm']) != 0: return False if call_process('.', ['pacman', '-S', '--noconfirm', '--force', 'tomat-console']) != 0: return False tmpdir = tempfile.mkdtemp() success = True if success and env.is_ue3: success = ue3_commandlet(env.game, 'dnetomatminingcommandlet', [ tmpdir ]) if success: success = call_process('.', [ 'TomatConsole', 'ImportFromUnreal', '--RepositoryUri', 'sql://mining@console', '--TmpDirectory', tmpdir ]) == 0 # Clean up after ourselves shutil.rmtree(tmpdir) return success
e356ce2c6fc6a3383a4ab8f7eea1ecb3ef7aa978
linter.py
linter.py
from SublimeLinter.lint import ComposerLinter class Phpcs(ComposerLinter): cmd = ('phpcs', '--report=emacs', '${args}', '-') regex = r'^.*:(?P<line>[0-9]+):(?P<col>[0-9]+): (?:(?P<error>error)|(?P<warning>warning)) - (?P<message>(.(?!\(\S+\)$))+)( \((?P<code>\S+)\)$)?' # noqa: E501 defaults = { 'selector': 'source.php - text.blade, text.html.basic', # we want auto-substitution of the filename, # but `cmd` does not support that yet '--stdin-path=': '${file}' }
from SublimeLinter.lint import ComposerLinter class Phpcs(ComposerLinter): cmd = ('phpcs', '--report=emacs', '${args}', '-') regex = r'^.*:(?P<line>[0-9]+):(?P<col>[0-9]+): (?:(?P<error>error)|(?P<warning>warning)) - (?P<message>(.(?!\(\S+\)$))+)( \((?P<code>\S+)\)$)?' # noqa: E501 defaults = { 'selector': 'embedding.php, source.php - text.blade', # we want auto-substitution of the filename, # but `cmd` does not support that yet '--stdin-path=': '${file}' }
Update selector for Sublime Text >= 4134
Update selector for Sublime Text >= 4134
Python
mit
SublimeLinter/SublimeLinter-phpcs
from SublimeLinter.lint import ComposerLinter class Phpcs(ComposerLinter): cmd = ('phpcs', '--report=emacs', '${args}', '-') regex = r'^.*:(?P<line>[0-9]+):(?P<col>[0-9]+): (?:(?P<error>error)|(?P<warning>warning)) - (?P<message>(.(?!\(\S+\)$))+)( \((?P<code>\S+)\)$)?' # noqa: E501 defaults = { 'selector': 'source.php - text.blade, text.html.basic', # we want auto-substitution of the filename, # but `cmd` does not support that yet '--stdin-path=': '${file}' } Update selector for Sublime Text >= 4134
from SublimeLinter.lint import ComposerLinter class Phpcs(ComposerLinter): cmd = ('phpcs', '--report=emacs', '${args}', '-') regex = r'^.*:(?P<line>[0-9]+):(?P<col>[0-9]+): (?:(?P<error>error)|(?P<warning>warning)) - (?P<message>(.(?!\(\S+\)$))+)( \((?P<code>\S+)\)$)?' # noqa: E501 defaults = { 'selector': 'embedding.php, source.php - text.blade', # we want auto-substitution of the filename, # but `cmd` does not support that yet '--stdin-path=': '${file}' }
<commit_before>from SublimeLinter.lint import ComposerLinter class Phpcs(ComposerLinter): cmd = ('phpcs', '--report=emacs', '${args}', '-') regex = r'^.*:(?P<line>[0-9]+):(?P<col>[0-9]+): (?:(?P<error>error)|(?P<warning>warning)) - (?P<message>(.(?!\(\S+\)$))+)( \((?P<code>\S+)\)$)?' # noqa: E501 defaults = { 'selector': 'source.php - text.blade, text.html.basic', # we want auto-substitution of the filename, # but `cmd` does not support that yet '--stdin-path=': '${file}' } <commit_msg>Update selector for Sublime Text >= 4134<commit_after>
from SublimeLinter.lint import ComposerLinter class Phpcs(ComposerLinter): cmd = ('phpcs', '--report=emacs', '${args}', '-') regex = r'^.*:(?P<line>[0-9]+):(?P<col>[0-9]+): (?:(?P<error>error)|(?P<warning>warning)) - (?P<message>(.(?!\(\S+\)$))+)( \((?P<code>\S+)\)$)?' # noqa: E501 defaults = { 'selector': 'embedding.php, source.php - text.blade', # we want auto-substitution of the filename, # but `cmd` does not support that yet '--stdin-path=': '${file}' }
from SublimeLinter.lint import ComposerLinter class Phpcs(ComposerLinter): cmd = ('phpcs', '--report=emacs', '${args}', '-') regex = r'^.*:(?P<line>[0-9]+):(?P<col>[0-9]+): (?:(?P<error>error)|(?P<warning>warning)) - (?P<message>(.(?!\(\S+\)$))+)( \((?P<code>\S+)\)$)?' # noqa: E501 defaults = { 'selector': 'source.php - text.blade, text.html.basic', # we want auto-substitution of the filename, # but `cmd` does not support that yet '--stdin-path=': '${file}' } Update selector for Sublime Text >= 4134from SublimeLinter.lint import ComposerLinter class Phpcs(ComposerLinter): cmd = ('phpcs', '--report=emacs', '${args}', '-') regex = r'^.*:(?P<line>[0-9]+):(?P<col>[0-9]+): (?:(?P<error>error)|(?P<warning>warning)) - (?P<message>(.(?!\(\S+\)$))+)( \((?P<code>\S+)\)$)?' # noqa: E501 defaults = { 'selector': 'embedding.php, source.php - text.blade', # we want auto-substitution of the filename, # but `cmd` does not support that yet '--stdin-path=': '${file}' }
<commit_before>from SublimeLinter.lint import ComposerLinter class Phpcs(ComposerLinter): cmd = ('phpcs', '--report=emacs', '${args}', '-') regex = r'^.*:(?P<line>[0-9]+):(?P<col>[0-9]+): (?:(?P<error>error)|(?P<warning>warning)) - (?P<message>(.(?!\(\S+\)$))+)( \((?P<code>\S+)\)$)?' # noqa: E501 defaults = { 'selector': 'source.php - text.blade, text.html.basic', # we want auto-substitution of the filename, # but `cmd` does not support that yet '--stdin-path=': '${file}' } <commit_msg>Update selector for Sublime Text >= 4134<commit_after>from SublimeLinter.lint import ComposerLinter class Phpcs(ComposerLinter): cmd = ('phpcs', '--report=emacs', '${args}', '-') regex = r'^.*:(?P<line>[0-9]+):(?P<col>[0-9]+): (?:(?P<error>error)|(?P<warning>warning)) - (?P<message>(.(?!\(\S+\)$))+)( \((?P<code>\S+)\)$)?' # noqa: E501 defaults = { 'selector': 'embedding.php, source.php - text.blade', # we want auto-substitution of the filename, # but `cmd` does not support that yet '--stdin-path=': '${file}' }
f61b81e968384859eb51a2ff14ca7709e8322ae8
yunity/walls/models.py
yunity/walls/models.py
from django.db.models import ForeignKey, TextField from config import settings from yunity.base.models import BaseModel class Wall(BaseModel): pass class WallPost(BaseModel): wall = ForeignKey(Wall) author = ForeignKey(settings.AUTH_USER_MODEL) class WallPostContent(BaseModel): post = ForeignKey(WallPost) author = ForeignKey(settings.AUTH_USER_MODEL) body = TextField()
from django.db.models import ForeignKey, TextField from config import settings from yunity.base.models import BaseModel class Wall(BaseModel): def resolve_permissions(self, collector): h = self.hub if h.target_content_type.model == 'group': g = h.target """:type : Group""" collector.add_hub(h, 'read') if g.is_content_included_in_parent: g = g.parent while g: collector.add_hub(g.hub, 'read') g = g.parent class WallPost(BaseModel): wall = ForeignKey(Wall) author = ForeignKey(settings.AUTH_USER_MODEL) class WallPostContent(BaseModel): post = ForeignKey(WallPost) author = ForeignKey(settings.AUTH_USER_MODEL) body = TextField()
Implement basic permissions resolver for walls
Implement basic permissions resolver for walls To be seen as a poc, collect all hub permissions for a basic permission and settings/inheritance model for reading a wall. with @nicksellen
Python
agpl-3.0
yunity/foodsaving-backend,yunity/yunity-core,yunity/foodsaving-backend,yunity/yunity-core,yunity/foodsaving-backend
from django.db.models import ForeignKey, TextField from config import settings from yunity.base.models import BaseModel class Wall(BaseModel): pass class WallPost(BaseModel): wall = ForeignKey(Wall) author = ForeignKey(settings.AUTH_USER_MODEL) class WallPostContent(BaseModel): post = ForeignKey(WallPost) author = ForeignKey(settings.AUTH_USER_MODEL) body = TextField() Implement basic permissions resolver for walls To be seen as a poc, collect all hub permissions for a basic permission and settings/inheritance model for reading a wall. with @nicksellen
from django.db.models import ForeignKey, TextField from config import settings from yunity.base.models import BaseModel class Wall(BaseModel): def resolve_permissions(self, collector): h = self.hub if h.target_content_type.model == 'group': g = h.target """:type : Group""" collector.add_hub(h, 'read') if g.is_content_included_in_parent: g = g.parent while g: collector.add_hub(g.hub, 'read') g = g.parent class WallPost(BaseModel): wall = ForeignKey(Wall) author = ForeignKey(settings.AUTH_USER_MODEL) class WallPostContent(BaseModel): post = ForeignKey(WallPost) author = ForeignKey(settings.AUTH_USER_MODEL) body = TextField()
<commit_before>from django.db.models import ForeignKey, TextField from config import settings from yunity.base.models import BaseModel class Wall(BaseModel): pass class WallPost(BaseModel): wall = ForeignKey(Wall) author = ForeignKey(settings.AUTH_USER_MODEL) class WallPostContent(BaseModel): post = ForeignKey(WallPost) author = ForeignKey(settings.AUTH_USER_MODEL) body = TextField() <commit_msg>Implement basic permissions resolver for walls To be seen as a poc, collect all hub permissions for a basic permission and settings/inheritance model for reading a wall. with @nicksellen<commit_after>
from django.db.models import ForeignKey, TextField from config import settings from yunity.base.models import BaseModel class Wall(BaseModel): def resolve_permissions(self, collector): h = self.hub if h.target_content_type.model == 'group': g = h.target """:type : Group""" collector.add_hub(h, 'read') if g.is_content_included_in_parent: g = g.parent while g: collector.add_hub(g.hub, 'read') g = g.parent class WallPost(BaseModel): wall = ForeignKey(Wall) author = ForeignKey(settings.AUTH_USER_MODEL) class WallPostContent(BaseModel): post = ForeignKey(WallPost) author = ForeignKey(settings.AUTH_USER_MODEL) body = TextField()
from django.db.models import ForeignKey, TextField from config import settings from yunity.base.models import BaseModel class Wall(BaseModel): pass class WallPost(BaseModel): wall = ForeignKey(Wall) author = ForeignKey(settings.AUTH_USER_MODEL) class WallPostContent(BaseModel): post = ForeignKey(WallPost) author = ForeignKey(settings.AUTH_USER_MODEL) body = TextField() Implement basic permissions resolver for walls To be seen as a poc, collect all hub permissions for a basic permission and settings/inheritance model for reading a wall. with @nicksellenfrom django.db.models import ForeignKey, TextField from config import settings from yunity.base.models import BaseModel class Wall(BaseModel): def resolve_permissions(self, collector): h = self.hub if h.target_content_type.model == 'group': g = h.target """:type : Group""" collector.add_hub(h, 'read') if g.is_content_included_in_parent: g = g.parent while g: collector.add_hub(g.hub, 'read') g = g.parent class WallPost(BaseModel): wall = ForeignKey(Wall) author = ForeignKey(settings.AUTH_USER_MODEL) class WallPostContent(BaseModel): post = ForeignKey(WallPost) author = ForeignKey(settings.AUTH_USER_MODEL) body = TextField()
<commit_before>from django.db.models import ForeignKey, TextField from config import settings from yunity.base.models import BaseModel class Wall(BaseModel): pass class WallPost(BaseModel): wall = ForeignKey(Wall) author = ForeignKey(settings.AUTH_USER_MODEL) class WallPostContent(BaseModel): post = ForeignKey(WallPost) author = ForeignKey(settings.AUTH_USER_MODEL) body = TextField() <commit_msg>Implement basic permissions resolver for walls To be seen as a poc, collect all hub permissions for a basic permission and settings/inheritance model for reading a wall. with @nicksellen<commit_after>from django.db.models import ForeignKey, TextField from config import settings from yunity.base.models import BaseModel class Wall(BaseModel): def resolve_permissions(self, collector): h = self.hub if h.target_content_type.model == 'group': g = h.target """:type : Group""" collector.add_hub(h, 'read') if g.is_content_included_in_parent: g = g.parent while g: collector.add_hub(g.hub, 'read') g = g.parent class WallPost(BaseModel): wall = ForeignKey(Wall) author = ForeignKey(settings.AUTH_USER_MODEL) class WallPostContent(BaseModel): post = ForeignKey(WallPost) author = ForeignKey(settings.AUTH_USER_MODEL) body = TextField()
f83e8de41b4148f483a4de8ab624d89ee25fedb0
soundbot.py
soundbot.py
from slacksocket import SlackSocket import subprocess import config import os def handle_cmd(cmd): if cmd in ('ls','list'): s.send_msg(list_files(), channel_name=config.slack_channel) else: playsound(cmd) def playsound(sound): subprocess.call([config.play_cmd,"{0}.mp3".format(sound)]) def list_files(): all_files = os.listdir('.') mp3s = [parts[0] for file in all_files for parts in [file.split('.')] if parts[1]=='mp3'] mp3s.sort() return ' '.join(mp3s) s = SlackSocket(config.api_key,translate=True) for event in s.events(): if event.event.get('channel')==config.slack_channel and event.event.get('type')=='message': print(event.json) handle_cmd(event.event.get('text'))
from slacksocket import SlackSocket import subprocess import config import os def handle_cmd(cmd): if cmd in ('ls','list'): s.send_msg(list_files(), channel_name=config.slack_channel) else: playsound(cmd) def playsound(sound): subprocess.call([config.play_cmd,"{0}.mp3".format(sound)]) def list_files(): all_files = os.listdir('.') mp3s = [parts[0] for file in all_files for parts in [file.split('.')] if parts[-1]=='mp3'] mp3s.sort() return ' '.join(mp3s) s = SlackSocket(config.api_key,translate=True) for event in s.events(): if event.event.get('channel')==config.slack_channel and event.event.get('type')=='message': print(event.json) handle_cmd(event.event.get('text'))
Make file listing more robust
Make file listing more robust
Python
mit
sanderevers/slack-soundbot
from slacksocket import SlackSocket import subprocess import config import os def handle_cmd(cmd): if cmd in ('ls','list'): s.send_msg(list_files(), channel_name=config.slack_channel) else: playsound(cmd) def playsound(sound): subprocess.call([config.play_cmd,"{0}.mp3".format(sound)]) def list_files(): all_files = os.listdir('.') mp3s = [parts[0] for file in all_files for parts in [file.split('.')] if parts[1]=='mp3'] mp3s.sort() return ' '.join(mp3s) s = SlackSocket(config.api_key,translate=True) for event in s.events(): if event.event.get('channel')==config.slack_channel and event.event.get('type')=='message': print(event.json) handle_cmd(event.event.get('text')) Make file listing more robust
from slacksocket import SlackSocket import subprocess import config import os def handle_cmd(cmd): if cmd in ('ls','list'): s.send_msg(list_files(), channel_name=config.slack_channel) else: playsound(cmd) def playsound(sound): subprocess.call([config.play_cmd,"{0}.mp3".format(sound)]) def list_files(): all_files = os.listdir('.') mp3s = [parts[0] for file in all_files for parts in [file.split('.')] if parts[-1]=='mp3'] mp3s.sort() return ' '.join(mp3s) s = SlackSocket(config.api_key,translate=True) for event in s.events(): if event.event.get('channel')==config.slack_channel and event.event.get('type')=='message': print(event.json) handle_cmd(event.event.get('text'))
<commit_before>from slacksocket import SlackSocket import subprocess import config import os def handle_cmd(cmd): if cmd in ('ls','list'): s.send_msg(list_files(), channel_name=config.slack_channel) else: playsound(cmd) def playsound(sound): subprocess.call([config.play_cmd,"{0}.mp3".format(sound)]) def list_files(): all_files = os.listdir('.') mp3s = [parts[0] for file in all_files for parts in [file.split('.')] if parts[1]=='mp3'] mp3s.sort() return ' '.join(mp3s) s = SlackSocket(config.api_key,translate=True) for event in s.events(): if event.event.get('channel')==config.slack_channel and event.event.get('type')=='message': print(event.json) handle_cmd(event.event.get('text')) <commit_msg>Make file listing more robust<commit_after>
from slacksocket import SlackSocket import subprocess import config import os def handle_cmd(cmd): if cmd in ('ls','list'): s.send_msg(list_files(), channel_name=config.slack_channel) else: playsound(cmd) def playsound(sound): subprocess.call([config.play_cmd,"{0}.mp3".format(sound)]) def list_files(): all_files = os.listdir('.') mp3s = [parts[0] for file in all_files for parts in [file.split('.')] if parts[-1]=='mp3'] mp3s.sort() return ' '.join(mp3s) s = SlackSocket(config.api_key,translate=True) for event in s.events(): if event.event.get('channel')==config.slack_channel and event.event.get('type')=='message': print(event.json) handle_cmd(event.event.get('text'))
from slacksocket import SlackSocket import subprocess import config import os def handle_cmd(cmd): if cmd in ('ls','list'): s.send_msg(list_files(), channel_name=config.slack_channel) else: playsound(cmd) def playsound(sound): subprocess.call([config.play_cmd,"{0}.mp3".format(sound)]) def list_files(): all_files = os.listdir('.') mp3s = [parts[0] for file in all_files for parts in [file.split('.')] if parts[1]=='mp3'] mp3s.sort() return ' '.join(mp3s) s = SlackSocket(config.api_key,translate=True) for event in s.events(): if event.event.get('channel')==config.slack_channel and event.event.get('type')=='message': print(event.json) handle_cmd(event.event.get('text')) Make file listing more robustfrom slacksocket import SlackSocket import subprocess import config import os def handle_cmd(cmd): if cmd in ('ls','list'): s.send_msg(list_files(), channel_name=config.slack_channel) else: playsound(cmd) def playsound(sound): subprocess.call([config.play_cmd,"{0}.mp3".format(sound)]) def list_files(): all_files = os.listdir('.') mp3s = [parts[0] for file in all_files for parts in [file.split('.')] if parts[-1]=='mp3'] mp3s.sort() return ' '.join(mp3s) s = SlackSocket(config.api_key,translate=True) for event in s.events(): if event.event.get('channel')==config.slack_channel and event.event.get('type')=='message': print(event.json) handle_cmd(event.event.get('text'))
<commit_before>from slacksocket import SlackSocket import subprocess import config import os def handle_cmd(cmd): if cmd in ('ls','list'): s.send_msg(list_files(), channel_name=config.slack_channel) else: playsound(cmd) def playsound(sound): subprocess.call([config.play_cmd,"{0}.mp3".format(sound)]) def list_files(): all_files = os.listdir('.') mp3s = [parts[0] for file in all_files for parts in [file.split('.')] if parts[1]=='mp3'] mp3s.sort() return ' '.join(mp3s) s = SlackSocket(config.api_key,translate=True) for event in s.events(): if event.event.get('channel')==config.slack_channel and event.event.get('type')=='message': print(event.json) handle_cmd(event.event.get('text')) <commit_msg>Make file listing more robust<commit_after>from slacksocket import SlackSocket import subprocess import config import os def handle_cmd(cmd): if cmd in ('ls','list'): s.send_msg(list_files(), channel_name=config.slack_channel) else: playsound(cmd) def playsound(sound): subprocess.call([config.play_cmd,"{0}.mp3".format(sound)]) def list_files(): all_files = os.listdir('.') mp3s = [parts[0] for file in all_files for parts in [file.split('.')] if parts[-1]=='mp3'] mp3s.sort() return ' '.join(mp3s) s = SlackSocket(config.api_key,translate=True) for event in s.events(): if event.event.get('channel')==config.slack_channel and event.event.get('type')=='message': print(event.json) handle_cmd(event.event.get('text'))
58970055d4905b50a3ed76a8ead39d9d0b572854
hookit.py
hookit.py
#!/usr/bin/env python # -*- coding: utf-8 -*- from datetime import datetime from flask import Flask, request app = Flask(__name__) @app.route('/') def index(): return 'Hello from Flask! (%s)' % datetime.now().strftime('%Y-%m-%d %H:%M:%S') @app.route('/github') def github(): if request.method != 'POST': return 405 with open('test', 'w') as f: f.write(request.get_json()) return ':)'
#!/usr/bin/env python # -*- coding: utf-8 -*- from datetime import datetime from flask import Flask, request app = Flask(__name__) app.debug = True @app.route('/') def index(): return 'Hello from Flask! (%s)' % datetime.now().strftime('%Y-%m-%d %H:%M:%S') @app.route('/github') def github(): if request.method != 'POST': return 'Not Allowed', 405 with open('test', 'w') as f: f.write(request.get_json()) return ':)' if __name__ == '__main__': app.run()
Add debugging options by default
Add debugging options by default
Python
mit
pcostesi/hookit,pcostesi/hookit
#!/usr/bin/env python # -*- coding: utf-8 -*- from datetime import datetime from flask import Flask, request app = Flask(__name__) @app.route('/') def index(): return 'Hello from Flask! (%s)' % datetime.now().strftime('%Y-%m-%d %H:%M:%S') @app.route('/github') def github(): if request.method != 'POST': return 405 with open('test', 'w') as f: f.write(request.get_json()) return ':)' Add debugging options by default
#!/usr/bin/env python # -*- coding: utf-8 -*- from datetime import datetime from flask import Flask, request app = Flask(__name__) app.debug = True @app.route('/') def index(): return 'Hello from Flask! (%s)' % datetime.now().strftime('%Y-%m-%d %H:%M:%S') @app.route('/github') def github(): if request.method != 'POST': return 'Not Allowed', 405 with open('test', 'w') as f: f.write(request.get_json()) return ':)' if __name__ == '__main__': app.run()
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- from datetime import datetime from flask import Flask, request app = Flask(__name__) @app.route('/') def index(): return 'Hello from Flask! (%s)' % datetime.now().strftime('%Y-%m-%d %H:%M:%S') @app.route('/github') def github(): if request.method != 'POST': return 405 with open('test', 'w') as f: f.write(request.get_json()) return ':)' <commit_msg>Add debugging options by default<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- from datetime import datetime from flask import Flask, request app = Flask(__name__) app.debug = True @app.route('/') def index(): return 'Hello from Flask! (%s)' % datetime.now().strftime('%Y-%m-%d %H:%M:%S') @app.route('/github') def github(): if request.method != 'POST': return 'Not Allowed', 405 with open('test', 'w') as f: f.write(request.get_json()) return ':)' if __name__ == '__main__': app.run()
#!/usr/bin/env python # -*- coding: utf-8 -*- from datetime import datetime from flask import Flask, request app = Flask(__name__) @app.route('/') def index(): return 'Hello from Flask! (%s)' % datetime.now().strftime('%Y-%m-%d %H:%M:%S') @app.route('/github') def github(): if request.method != 'POST': return 405 with open('test', 'w') as f: f.write(request.get_json()) return ':)' Add debugging options by default#!/usr/bin/env python # -*- coding: utf-8 -*- from datetime import datetime from flask import Flask, request app = Flask(__name__) app.debug = True @app.route('/') def index(): return 'Hello from Flask! (%s)' % datetime.now().strftime('%Y-%m-%d %H:%M:%S') @app.route('/github') def github(): if request.method != 'POST': return 'Not Allowed', 405 with open('test', 'w') as f: f.write(request.get_json()) return ':)' if __name__ == '__main__': app.run()
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- from datetime import datetime from flask import Flask, request app = Flask(__name__) @app.route('/') def index(): return 'Hello from Flask! (%s)' % datetime.now().strftime('%Y-%m-%d %H:%M:%S') @app.route('/github') def github(): if request.method != 'POST': return 405 with open('test', 'w') as f: f.write(request.get_json()) return ':)' <commit_msg>Add debugging options by default<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- from datetime import datetime from flask import Flask, request app = Flask(__name__) app.debug = True @app.route('/') def index(): return 'Hello from Flask! (%s)' % datetime.now().strftime('%Y-%m-%d %H:%M:%S') @app.route('/github') def github(): if request.method != 'POST': return 'Not Allowed', 405 with open('test', 'w') as f: f.write(request.get_json()) return ':)' if __name__ == '__main__': app.run()
3d48d62aca0455cf71a841fcfcddabaf770bd048
plumeria/plugins/bible.py
plumeria/plugins/bible.py
from bs4 import BeautifulSoup from bs4 import Comment from bs4 import NavigableString import plumeria.util.http as http from plumeria.command import commands, CommandError from plumeria.command.parse import Text from plumeria.util.ratelimit import rate_limit @commands.register("bible", "esv", category="Search", params=[Text('verse')]) @rate_limit() async def search_esv(message, verse): """ Search for a bible passage from the English Standard Version. Example:: bible Romans 12:16 """ r = await http.get("http://www.esvapi.org/v2/rest/passageQuery", params={ "key": "IP", "passage": verse, "output-format": "crossway-xml-1.0" }) doc = BeautifulSoup(r.text(), features="lxml") if not doc.passage: raise CommandError("Verse not found.") lines = [] for verse_unit in doc.passage.content.find_all('verse-unit'): num = int(verse_unit.find('verse-num').text) woc = verse_unit.find('woc') if woc: text = woc.text else: text = "".join([str(node) for node in verse_unit.children if isinstance(node, NavigableString) and not isinstance(node, Comment)]) lines.append("({}) {}".format(num, text.strip())) return "\n".join(lines)
from bs4 import BeautifulSoup from bs4 import Comment from bs4 import NavigableString import plumeria.util.http as http from plumeria.command import commands, CommandError from plumeria.command.parse import Text from plumeria.util.ratelimit import rate_limit @commands.register("bible", "esv", category="Search", params=[Text('verse')]) @rate_limit() async def search_esv(message, verse): """ Search for a bible passage from the English Standard Version. Example:: bible Romans 12:16 """ r = await http.get("http://www.esvapi.org/v2/rest/passageQuery", params={ "key": "IP", "passage": verse, "output-format": "crossway-xml-1.0" }) doc = BeautifulSoup(r.text(), features="lxml") if not doc.passage: raise CommandError("Verse not found.") lines = [] for verse_unit in doc.passage.content.find_all('verse-unit'): num = int(verse_unit.find('verse-num').text) woc = verse_unit.find('woc') if woc: text = woc.text else: text = "".join([str(node) for node in verse_unit.children if isinstance(node, NavigableString) and not isinstance(node, Comment)]) lines.append("**{}** {}".format(num, text.strip())) return "\n".join(lines)
Use bold line number for verse.
Use bold line number for verse.
Python
mit
sk89q/Plumeria,sk89q/Plumeria,sk89q/Plumeria
from bs4 import BeautifulSoup from bs4 import Comment from bs4 import NavigableString import plumeria.util.http as http from plumeria.command import commands, CommandError from plumeria.command.parse import Text from plumeria.util.ratelimit import rate_limit @commands.register("bible", "esv", category="Search", params=[Text('verse')]) @rate_limit() async def search_esv(message, verse): """ Search for a bible passage from the English Standard Version. Example:: bible Romans 12:16 """ r = await http.get("http://www.esvapi.org/v2/rest/passageQuery", params={ "key": "IP", "passage": verse, "output-format": "crossway-xml-1.0" }) doc = BeautifulSoup(r.text(), features="lxml") if not doc.passage: raise CommandError("Verse not found.") lines = [] for verse_unit in doc.passage.content.find_all('verse-unit'): num = int(verse_unit.find('verse-num').text) woc = verse_unit.find('woc') if woc: text = woc.text else: text = "".join([str(node) for node in verse_unit.children if isinstance(node, NavigableString) and not isinstance(node, Comment)]) lines.append("({}) {}".format(num, text.strip())) return "\n".join(lines) Use bold line number for verse.
from bs4 import BeautifulSoup from bs4 import Comment from bs4 import NavigableString import plumeria.util.http as http from plumeria.command import commands, CommandError from plumeria.command.parse import Text from plumeria.util.ratelimit import rate_limit @commands.register("bible", "esv", category="Search", params=[Text('verse')]) @rate_limit() async def search_esv(message, verse): """ Search for a bible passage from the English Standard Version. Example:: bible Romans 12:16 """ r = await http.get("http://www.esvapi.org/v2/rest/passageQuery", params={ "key": "IP", "passage": verse, "output-format": "crossway-xml-1.0" }) doc = BeautifulSoup(r.text(), features="lxml") if not doc.passage: raise CommandError("Verse not found.") lines = [] for verse_unit in doc.passage.content.find_all('verse-unit'): num = int(verse_unit.find('verse-num').text) woc = verse_unit.find('woc') if woc: text = woc.text else: text = "".join([str(node) for node in verse_unit.children if isinstance(node, NavigableString) and not isinstance(node, Comment)]) lines.append("**{}** {}".format(num, text.strip())) return "\n".join(lines)
<commit_before>from bs4 import BeautifulSoup from bs4 import Comment from bs4 import NavigableString import plumeria.util.http as http from plumeria.command import commands, CommandError from plumeria.command.parse import Text from plumeria.util.ratelimit import rate_limit @commands.register("bible", "esv", category="Search", params=[Text('verse')]) @rate_limit() async def search_esv(message, verse): """ Search for a bible passage from the English Standard Version. Example:: bible Romans 12:16 """ r = await http.get("http://www.esvapi.org/v2/rest/passageQuery", params={ "key": "IP", "passage": verse, "output-format": "crossway-xml-1.0" }) doc = BeautifulSoup(r.text(), features="lxml") if not doc.passage: raise CommandError("Verse not found.") lines = [] for verse_unit in doc.passage.content.find_all('verse-unit'): num = int(verse_unit.find('verse-num').text) woc = verse_unit.find('woc') if woc: text = woc.text else: text = "".join([str(node) for node in verse_unit.children if isinstance(node, NavigableString) and not isinstance(node, Comment)]) lines.append("({}) {}".format(num, text.strip())) return "\n".join(lines) <commit_msg>Use bold line number for verse.<commit_after>
from bs4 import BeautifulSoup from bs4 import Comment from bs4 import NavigableString import plumeria.util.http as http from plumeria.command import commands, CommandError from plumeria.command.parse import Text from plumeria.util.ratelimit import rate_limit @commands.register("bible", "esv", category="Search", params=[Text('verse')]) @rate_limit() async def search_esv(message, verse): """ Search for a bible passage from the English Standard Version. Example:: bible Romans 12:16 """ r = await http.get("http://www.esvapi.org/v2/rest/passageQuery", params={ "key": "IP", "passage": verse, "output-format": "crossway-xml-1.0" }) doc = BeautifulSoup(r.text(), features="lxml") if not doc.passage: raise CommandError("Verse not found.") lines = [] for verse_unit in doc.passage.content.find_all('verse-unit'): num = int(verse_unit.find('verse-num').text) woc = verse_unit.find('woc') if woc: text = woc.text else: text = "".join([str(node) for node in verse_unit.children if isinstance(node, NavigableString) and not isinstance(node, Comment)]) lines.append("**{}** {}".format(num, text.strip())) return "\n".join(lines)
from bs4 import BeautifulSoup from bs4 import Comment from bs4 import NavigableString import plumeria.util.http as http from plumeria.command import commands, CommandError from plumeria.command.parse import Text from plumeria.util.ratelimit import rate_limit @commands.register("bible", "esv", category="Search", params=[Text('verse')]) @rate_limit() async def search_esv(message, verse): """ Search for a bible passage from the English Standard Version. Example:: bible Romans 12:16 """ r = await http.get("http://www.esvapi.org/v2/rest/passageQuery", params={ "key": "IP", "passage": verse, "output-format": "crossway-xml-1.0" }) doc = BeautifulSoup(r.text(), features="lxml") if not doc.passage: raise CommandError("Verse not found.") lines = [] for verse_unit in doc.passage.content.find_all('verse-unit'): num = int(verse_unit.find('verse-num').text) woc = verse_unit.find('woc') if woc: text = woc.text else: text = "".join([str(node) for node in verse_unit.children if isinstance(node, NavigableString) and not isinstance(node, Comment)]) lines.append("({}) {}".format(num, text.strip())) return "\n".join(lines) Use bold line number for verse.from bs4 import BeautifulSoup from bs4 import Comment from bs4 import NavigableString import plumeria.util.http as http from plumeria.command import commands, CommandError from plumeria.command.parse import Text from plumeria.util.ratelimit import rate_limit @commands.register("bible", "esv", category="Search", params=[Text('verse')]) @rate_limit() async def search_esv(message, verse): """ Search for a bible passage from the English Standard Version. Example:: bible Romans 12:16 """ r = await http.get("http://www.esvapi.org/v2/rest/passageQuery", params={ "key": "IP", "passage": verse, "output-format": "crossway-xml-1.0" }) doc = BeautifulSoup(r.text(), features="lxml") if not doc.passage: raise CommandError("Verse not found.") lines = [] for verse_unit in doc.passage.content.find_all('verse-unit'): num = int(verse_unit.find('verse-num').text) woc = verse_unit.find('woc') if woc: text = woc.text else: text = "".join([str(node) for node in verse_unit.children if isinstance(node, NavigableString) and not isinstance(node, Comment)]) lines.append("**{}** {}".format(num, text.strip())) return "\n".join(lines)
<commit_before>from bs4 import BeautifulSoup from bs4 import Comment from bs4 import NavigableString import plumeria.util.http as http from plumeria.command import commands, CommandError from plumeria.command.parse import Text from plumeria.util.ratelimit import rate_limit @commands.register("bible", "esv", category="Search", params=[Text('verse')]) @rate_limit() async def search_esv(message, verse): """ Search for a bible passage from the English Standard Version. Example:: bible Romans 12:16 """ r = await http.get("http://www.esvapi.org/v2/rest/passageQuery", params={ "key": "IP", "passage": verse, "output-format": "crossway-xml-1.0" }) doc = BeautifulSoup(r.text(), features="lxml") if not doc.passage: raise CommandError("Verse not found.") lines = [] for verse_unit in doc.passage.content.find_all('verse-unit'): num = int(verse_unit.find('verse-num').text) woc = verse_unit.find('woc') if woc: text = woc.text else: text = "".join([str(node) for node in verse_unit.children if isinstance(node, NavigableString) and not isinstance(node, Comment)]) lines.append("({}) {}".format(num, text.strip())) return "\n".join(lines) <commit_msg>Use bold line number for verse.<commit_after>from bs4 import BeautifulSoup from bs4 import Comment from bs4 import NavigableString import plumeria.util.http as http from plumeria.command import commands, CommandError from plumeria.command.parse import Text from plumeria.util.ratelimit import rate_limit @commands.register("bible", "esv", category="Search", params=[Text('verse')]) @rate_limit() async def search_esv(message, verse): """ Search for a bible passage from the English Standard Version. Example:: bible Romans 12:16 """ r = await http.get("http://www.esvapi.org/v2/rest/passageQuery", params={ "key": "IP", "passage": verse, "output-format": "crossway-xml-1.0" }) doc = BeautifulSoup(r.text(), features="lxml") if not doc.passage: raise CommandError("Verse not found.") lines = [] for verse_unit in doc.passage.content.find_all('verse-unit'): num = int(verse_unit.find('verse-num').text) woc = verse_unit.find('woc') if woc: text = woc.text else: text = "".join([str(node) for node in verse_unit.children if isinstance(node, NavigableString) and not isinstance(node, Comment)]) lines.append("**{}** {}".format(num, text.strip())) return "\n".join(lines)
f4f5852944d1fd1b9e96a70cb4496ee6e1e66dc0
genome_designer/main/celery_util.py
genome_designer/main/celery_util.py
""" Methods for interfacing with the Celery task queue management library. """ from errno import errorcode from celery.task.control import inspect CELERY_ERROR_KEY = 'ERROR' def get_celery_worker_status(): """Checks whether celery is running and reports the error if not. Source: http://stackoverflow.com/questions/8506914/detect-whether-celery-is-available-running """ try: insp = inspect() d = insp.stats() if not d: d = { CELERY_ERROR_KEY: 'No running Celery workers were found.' } except IOError as e: msg = "Error connecting to the backend: " + str(e) if len(e.args) > 0 and errorcode.get(e.args[0]) == 'ECONNREFUSED': msg += ' Check that the RabbitMQ server is running.' d = { CELERY_ERROR_KEY: msg } except ImportError as e: d = { CELERY_ERROR_KEY: str(e)} return d
""" Methods for interfacing with the Celery task queue management library. """ from errno import errorcode from celery.task.control import inspect from django.conf import settings CELERY_ERROR_KEY = 'ERROR' def get_celery_worker_status(): """Checks whether celery is running and reports the error if not. Source: http://stackoverflow.com/questions/8506914/detect-whether-celery-is-available-running """ if settings.BROKER_BACKEND == 'memory': # We are testing with in-memory celery. Celery is effectively running. return {} try: insp = inspect() d = insp.stats() if not d: d = { CELERY_ERROR_KEY: 'No running Celery workers were found.' } except IOError as e: msg = "Error connecting to the backend: " + str(e) if len(e.args) > 0 and errorcode.get(e.args[0]) == 'ECONNREFUSED': msg += ' Check that the RabbitMQ server is running.' d = { CELERY_ERROR_KEY: msg } except ImportError as e: d = { CELERY_ERROR_KEY: str(e)} return d
Fix tests: Allow for celery not to be running when doing in-memory celery for tests.
Fix tests: Allow for celery not to be running when doing in-memory celery for tests.
Python
mit
churchlab/millstone,churchlab/millstone,churchlab/millstone,churchlab/millstone,woodymit/millstone_accidental_source,woodymit/millstone_accidental_source,woodymit/millstone_accidental_source,woodymit/millstone,woodymit/millstone,woodymit/millstone_accidental_source,woodymit/millstone,woodymit/millstone
""" Methods for interfacing with the Celery task queue management library. """ from errno import errorcode from celery.task.control import inspect CELERY_ERROR_KEY = 'ERROR' def get_celery_worker_status(): """Checks whether celery is running and reports the error if not. Source: http://stackoverflow.com/questions/8506914/detect-whether-celery-is-available-running """ try: insp = inspect() d = insp.stats() if not d: d = { CELERY_ERROR_KEY: 'No running Celery workers were found.' } except IOError as e: msg = "Error connecting to the backend: " + str(e) if len(e.args) > 0 and errorcode.get(e.args[0]) == 'ECONNREFUSED': msg += ' Check that the RabbitMQ server is running.' d = { CELERY_ERROR_KEY: msg } except ImportError as e: d = { CELERY_ERROR_KEY: str(e)} return d Fix tests: Allow for celery not to be running when doing in-memory celery for tests.
""" Methods for interfacing with the Celery task queue management library. """ from errno import errorcode from celery.task.control import inspect from django.conf import settings CELERY_ERROR_KEY = 'ERROR' def get_celery_worker_status(): """Checks whether celery is running and reports the error if not. Source: http://stackoverflow.com/questions/8506914/detect-whether-celery-is-available-running """ if settings.BROKER_BACKEND == 'memory': # We are testing with in-memory celery. Celery is effectively running. return {} try: insp = inspect() d = insp.stats() if not d: d = { CELERY_ERROR_KEY: 'No running Celery workers were found.' } except IOError as e: msg = "Error connecting to the backend: " + str(e) if len(e.args) > 0 and errorcode.get(e.args[0]) == 'ECONNREFUSED': msg += ' Check that the RabbitMQ server is running.' d = { CELERY_ERROR_KEY: msg } except ImportError as e: d = { CELERY_ERROR_KEY: str(e)} return d
<commit_before>""" Methods for interfacing with the Celery task queue management library. """ from errno import errorcode from celery.task.control import inspect CELERY_ERROR_KEY = 'ERROR' def get_celery_worker_status(): """Checks whether celery is running and reports the error if not. Source: http://stackoverflow.com/questions/8506914/detect-whether-celery-is-available-running """ try: insp = inspect() d = insp.stats() if not d: d = { CELERY_ERROR_KEY: 'No running Celery workers were found.' } except IOError as e: msg = "Error connecting to the backend: " + str(e) if len(e.args) > 0 and errorcode.get(e.args[0]) == 'ECONNREFUSED': msg += ' Check that the RabbitMQ server is running.' d = { CELERY_ERROR_KEY: msg } except ImportError as e: d = { CELERY_ERROR_KEY: str(e)} return d <commit_msg>Fix tests: Allow for celery not to be running when doing in-memory celery for tests.<commit_after>
""" Methods for interfacing with the Celery task queue management library. """ from errno import errorcode from celery.task.control import inspect from django.conf import settings CELERY_ERROR_KEY = 'ERROR' def get_celery_worker_status(): """Checks whether celery is running and reports the error if not. Source: http://stackoverflow.com/questions/8506914/detect-whether-celery-is-available-running """ if settings.BROKER_BACKEND == 'memory': # We are testing with in-memory celery. Celery is effectively running. return {} try: insp = inspect() d = insp.stats() if not d: d = { CELERY_ERROR_KEY: 'No running Celery workers were found.' } except IOError as e: msg = "Error connecting to the backend: " + str(e) if len(e.args) > 0 and errorcode.get(e.args[0]) == 'ECONNREFUSED': msg += ' Check that the RabbitMQ server is running.' d = { CELERY_ERROR_KEY: msg } except ImportError as e: d = { CELERY_ERROR_KEY: str(e)} return d
""" Methods for interfacing with the Celery task queue management library. """ from errno import errorcode from celery.task.control import inspect CELERY_ERROR_KEY = 'ERROR' def get_celery_worker_status(): """Checks whether celery is running and reports the error if not. Source: http://stackoverflow.com/questions/8506914/detect-whether-celery-is-available-running """ try: insp = inspect() d = insp.stats() if not d: d = { CELERY_ERROR_KEY: 'No running Celery workers were found.' } except IOError as e: msg = "Error connecting to the backend: " + str(e) if len(e.args) > 0 and errorcode.get(e.args[0]) == 'ECONNREFUSED': msg += ' Check that the RabbitMQ server is running.' d = { CELERY_ERROR_KEY: msg } except ImportError as e: d = { CELERY_ERROR_KEY: str(e)} return d Fix tests: Allow for celery not to be running when doing in-memory celery for tests.""" Methods for interfacing with the Celery task queue management library. """ from errno import errorcode from celery.task.control import inspect from django.conf import settings CELERY_ERROR_KEY = 'ERROR' def get_celery_worker_status(): """Checks whether celery is running and reports the error if not. Source: http://stackoverflow.com/questions/8506914/detect-whether-celery-is-available-running """ if settings.BROKER_BACKEND == 'memory': # We are testing with in-memory celery. Celery is effectively running. return {} try: insp = inspect() d = insp.stats() if not d: d = { CELERY_ERROR_KEY: 'No running Celery workers were found.' } except IOError as e: msg = "Error connecting to the backend: " + str(e) if len(e.args) > 0 and errorcode.get(e.args[0]) == 'ECONNREFUSED': msg += ' Check that the RabbitMQ server is running.' d = { CELERY_ERROR_KEY: msg } except ImportError as e: d = { CELERY_ERROR_KEY: str(e)} return d
<commit_before>""" Methods for interfacing with the Celery task queue management library. """ from errno import errorcode from celery.task.control import inspect CELERY_ERROR_KEY = 'ERROR' def get_celery_worker_status(): """Checks whether celery is running and reports the error if not. Source: http://stackoverflow.com/questions/8506914/detect-whether-celery-is-available-running """ try: insp = inspect() d = insp.stats() if not d: d = { CELERY_ERROR_KEY: 'No running Celery workers were found.' } except IOError as e: msg = "Error connecting to the backend: " + str(e) if len(e.args) > 0 and errorcode.get(e.args[0]) == 'ECONNREFUSED': msg += ' Check that the RabbitMQ server is running.' d = { CELERY_ERROR_KEY: msg } except ImportError as e: d = { CELERY_ERROR_KEY: str(e)} return d <commit_msg>Fix tests: Allow for celery not to be running when doing in-memory celery for tests.<commit_after>""" Methods for interfacing with the Celery task queue management library. """ from errno import errorcode from celery.task.control import inspect from django.conf import settings CELERY_ERROR_KEY = 'ERROR' def get_celery_worker_status(): """Checks whether celery is running and reports the error if not. Source: http://stackoverflow.com/questions/8506914/detect-whether-celery-is-available-running """ if settings.BROKER_BACKEND == 'memory': # We are testing with in-memory celery. Celery is effectively running. return {} try: insp = inspect() d = insp.stats() if not d: d = { CELERY_ERROR_KEY: 'No running Celery workers were found.' } except IOError as e: msg = "Error connecting to the backend: " + str(e) if len(e.args) > 0 and errorcode.get(e.args[0]) == 'ECONNREFUSED': msg += ' Check that the RabbitMQ server is running.' d = { CELERY_ERROR_KEY: msg } except ImportError as e: d = { CELERY_ERROR_KEY: str(e)} return d
0b7f99bcb4e42c50263a7d8a42513876b02b445a
scikits/talkbox/tools/__init__.py
scikits/talkbox/tools/__init__.py
__all__ = [] import correlations from correlations import * __all__ += correlations.__all__ import cffilter from cffilter import cslfilter as slfilter __all__ += ['slfilter']
__all__ = [] import correlations from correlations import * __all__ += correlations.__all__ import cffilter from cffilter import cslfilter as slfilter __all__ += ['slfilter'] from segmentaxis import segment_axis __all__ += ['segment_axis']
Put segment_axis in the main scikits.talkbox namespace.
Put segment_axis in the main scikits.talkbox namespace.
Python
mit
cournape/talkbox,cournape/talkbox
__all__ = [] import correlations from correlations import * __all__ += correlations.__all__ import cffilter from cffilter import cslfilter as slfilter __all__ += ['slfilter'] Put segment_axis in the main scikits.talkbox namespace.
__all__ = [] import correlations from correlations import * __all__ += correlations.__all__ import cffilter from cffilter import cslfilter as slfilter __all__ += ['slfilter'] from segmentaxis import segment_axis __all__ += ['segment_axis']
<commit_before>__all__ = [] import correlations from correlations import * __all__ += correlations.__all__ import cffilter from cffilter import cslfilter as slfilter __all__ += ['slfilter'] <commit_msg>Put segment_axis in the main scikits.talkbox namespace.<commit_after>
__all__ = [] import correlations from correlations import * __all__ += correlations.__all__ import cffilter from cffilter import cslfilter as slfilter __all__ += ['slfilter'] from segmentaxis import segment_axis __all__ += ['segment_axis']
__all__ = [] import correlations from correlations import * __all__ += correlations.__all__ import cffilter from cffilter import cslfilter as slfilter __all__ += ['slfilter'] Put segment_axis in the main scikits.talkbox namespace.__all__ = [] import correlations from correlations import * __all__ += correlations.__all__ import cffilter from cffilter import cslfilter as slfilter __all__ += ['slfilter'] from segmentaxis import segment_axis __all__ += ['segment_axis']
<commit_before>__all__ = [] import correlations from correlations import * __all__ += correlations.__all__ import cffilter from cffilter import cslfilter as slfilter __all__ += ['slfilter'] <commit_msg>Put segment_axis in the main scikits.talkbox namespace.<commit_after>__all__ = [] import correlations from correlations import * __all__ += correlations.__all__ import cffilter from cffilter import cslfilter as slfilter __all__ += ['slfilter'] from segmentaxis import segment_axis __all__ += ['segment_axis']
91fd97d7579673a0c310c734a1c1ef83a07b50d1
phantasy/library/scan/datautil.py
phantasy/library/scan/datautil.py
#!/usr/bin/env python # -*- coding: utf-8 -*- """utils for data crunching, saving. """ import numpy as np class ScanDataFactory(object): """Post processor of data from scan server. Parameters ---------- data : dict Raw data retrieving from scan server regarding scan ID, after completing certain scan task. n_sample : int Sample number for every scan device setup. """ def __init__(self, data, n_sample): self._raw_data = data self._n = n_sample self._rebuild_data() def _rebuild_data(self): """Rebuild raw_data """ self._data = {k:np.array(v.get('value')).reshape(-1, self._n) for k,v in self._raw_data.iteritems()} def get_average(self, name): """Get average. Parameters ---------- name : str Key name of raw_data. """ return self._data.get(name).mean(axis=1) def get_errorbar(self, name): """Get errorbar Parameters ---------- name : str Key name of raw_data. """ return self._data.get(name).std(axis=1) def get_all_names(self): """Get all key names of raw_data. Returns ------- ret : list List of keys. """ return self._data.keys() def save(self, ext='dat'): """ """ pass
#!/usr/bin/env python # -*- coding: utf-8 -*- """utils for data crunching, saving. """ import numpy as np class ScanDataFactory(object): """Post processor of data from scan server. Parameters ---------- data : dict Raw data retrieving from scan server regarding scan ID, after completing certain scan task. n_sample : int Sample number for every scan device setup. """ def __init__(self, data, n_sample): self._raw_data = data self._n = n_sample self._rebuild_data() @property def raw_data(self): """dict: Dict of array, raw scan data.""" return self._raw_data @property def data(self): """dict: Dict of array, raw scan data after postprocessing.""" return self._data def _rebuild_data(self): """Rebuild raw_data """ self._data = {k:np.array(v.get('value')).reshape(-1, self._n) for k,v in self._raw_data.iteritems()} def get_average(self, name): """Get average. Parameters ---------- name : str Key name of raw_data. """ return self._data.get(name).mean(axis=1) def get_errorbar(self, name): """Get errorbar Parameters ---------- name : str Key name of raw_data. """ return self._data.get(name).std(axis=1) def get_all_names(self): """Get all key names of raw_data. Returns ------- ret : list List of keys. """ return self._data.keys() def save(self, ext='dat'): """ """ pass
Make raw_data and data as properties
Make raw_data and data as properties
Python
bsd-3-clause
archman/phantasy,archman/phantasy
#!/usr/bin/env python # -*- coding: utf-8 -*- """utils for data crunching, saving. """ import numpy as np class ScanDataFactory(object): """Post processor of data from scan server. Parameters ---------- data : dict Raw data retrieving from scan server regarding scan ID, after completing certain scan task. n_sample : int Sample number for every scan device setup. """ def __init__(self, data, n_sample): self._raw_data = data self._n = n_sample self._rebuild_data() def _rebuild_data(self): """Rebuild raw_data """ self._data = {k:np.array(v.get('value')).reshape(-1, self._n) for k,v in self._raw_data.iteritems()} def get_average(self, name): """Get average. Parameters ---------- name : str Key name of raw_data. """ return self._data.get(name).mean(axis=1) def get_errorbar(self, name): """Get errorbar Parameters ---------- name : str Key name of raw_data. """ return self._data.get(name).std(axis=1) def get_all_names(self): """Get all key names of raw_data. Returns ------- ret : list List of keys. """ return self._data.keys() def save(self, ext='dat'): """ """ pass Make raw_data and data as properties
#!/usr/bin/env python # -*- coding: utf-8 -*- """utils for data crunching, saving. """ import numpy as np class ScanDataFactory(object): """Post processor of data from scan server. Parameters ---------- data : dict Raw data retrieving from scan server regarding scan ID, after completing certain scan task. n_sample : int Sample number for every scan device setup. """ def __init__(self, data, n_sample): self._raw_data = data self._n = n_sample self._rebuild_data() @property def raw_data(self): """dict: Dict of array, raw scan data.""" return self._raw_data @property def data(self): """dict: Dict of array, raw scan data after postprocessing.""" return self._data def _rebuild_data(self): """Rebuild raw_data """ self._data = {k:np.array(v.get('value')).reshape(-1, self._n) for k,v in self._raw_data.iteritems()} def get_average(self, name): """Get average. Parameters ---------- name : str Key name of raw_data. """ return self._data.get(name).mean(axis=1) def get_errorbar(self, name): """Get errorbar Parameters ---------- name : str Key name of raw_data. """ return self._data.get(name).std(axis=1) def get_all_names(self): """Get all key names of raw_data. Returns ------- ret : list List of keys. """ return self._data.keys() def save(self, ext='dat'): """ """ pass
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- """utils for data crunching, saving. """ import numpy as np class ScanDataFactory(object): """Post processor of data from scan server. Parameters ---------- data : dict Raw data retrieving from scan server regarding scan ID, after completing certain scan task. n_sample : int Sample number for every scan device setup. """ def __init__(self, data, n_sample): self._raw_data = data self._n = n_sample self._rebuild_data() def _rebuild_data(self): """Rebuild raw_data """ self._data = {k:np.array(v.get('value')).reshape(-1, self._n) for k,v in self._raw_data.iteritems()} def get_average(self, name): """Get average. Parameters ---------- name : str Key name of raw_data. """ return self._data.get(name).mean(axis=1) def get_errorbar(self, name): """Get errorbar Parameters ---------- name : str Key name of raw_data. """ return self._data.get(name).std(axis=1) def get_all_names(self): """Get all key names of raw_data. Returns ------- ret : list List of keys. """ return self._data.keys() def save(self, ext='dat'): """ """ pass <commit_msg>Make raw_data and data as properties<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- """utils for data crunching, saving. """ import numpy as np class ScanDataFactory(object): """Post processor of data from scan server. Parameters ---------- data : dict Raw data retrieving from scan server regarding scan ID, after completing certain scan task. n_sample : int Sample number for every scan device setup. """ def __init__(self, data, n_sample): self._raw_data = data self._n = n_sample self._rebuild_data() @property def raw_data(self): """dict: Dict of array, raw scan data.""" return self._raw_data @property def data(self): """dict: Dict of array, raw scan data after postprocessing.""" return self._data def _rebuild_data(self): """Rebuild raw_data """ self._data = {k:np.array(v.get('value')).reshape(-1, self._n) for k,v in self._raw_data.iteritems()} def get_average(self, name): """Get average. Parameters ---------- name : str Key name of raw_data. """ return self._data.get(name).mean(axis=1) def get_errorbar(self, name): """Get errorbar Parameters ---------- name : str Key name of raw_data. """ return self._data.get(name).std(axis=1) def get_all_names(self): """Get all key names of raw_data. Returns ------- ret : list List of keys. """ return self._data.keys() def save(self, ext='dat'): """ """ pass
#!/usr/bin/env python # -*- coding: utf-8 -*- """utils for data crunching, saving. """ import numpy as np class ScanDataFactory(object): """Post processor of data from scan server. Parameters ---------- data : dict Raw data retrieving from scan server regarding scan ID, after completing certain scan task. n_sample : int Sample number for every scan device setup. """ def __init__(self, data, n_sample): self._raw_data = data self._n = n_sample self._rebuild_data() def _rebuild_data(self): """Rebuild raw_data """ self._data = {k:np.array(v.get('value')).reshape(-1, self._n) for k,v in self._raw_data.iteritems()} def get_average(self, name): """Get average. Parameters ---------- name : str Key name of raw_data. """ return self._data.get(name).mean(axis=1) def get_errorbar(self, name): """Get errorbar Parameters ---------- name : str Key name of raw_data. """ return self._data.get(name).std(axis=1) def get_all_names(self): """Get all key names of raw_data. Returns ------- ret : list List of keys. """ return self._data.keys() def save(self, ext='dat'): """ """ pass Make raw_data and data as properties#!/usr/bin/env python # -*- coding: utf-8 -*- """utils for data crunching, saving. """ import numpy as np class ScanDataFactory(object): """Post processor of data from scan server. Parameters ---------- data : dict Raw data retrieving from scan server regarding scan ID, after completing certain scan task. n_sample : int Sample number for every scan device setup. """ def __init__(self, data, n_sample): self._raw_data = data self._n = n_sample self._rebuild_data() @property def raw_data(self): """dict: Dict of array, raw scan data.""" return self._raw_data @property def data(self): """dict: Dict of array, raw scan data after postprocessing.""" return self._data def _rebuild_data(self): """Rebuild raw_data """ self._data = {k:np.array(v.get('value')).reshape(-1, self._n) for k,v in self._raw_data.iteritems()} def get_average(self, name): """Get average. Parameters ---------- name : str Key name of raw_data. """ return self._data.get(name).mean(axis=1) def get_errorbar(self, name): """Get errorbar Parameters ---------- name : str Key name of raw_data. """ return self._data.get(name).std(axis=1) def get_all_names(self): """Get all key names of raw_data. Returns ------- ret : list List of keys. """ return self._data.keys() def save(self, ext='dat'): """ """ pass
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- """utils for data crunching, saving. """ import numpy as np class ScanDataFactory(object): """Post processor of data from scan server. Parameters ---------- data : dict Raw data retrieving from scan server regarding scan ID, after completing certain scan task. n_sample : int Sample number for every scan device setup. """ def __init__(self, data, n_sample): self._raw_data = data self._n = n_sample self._rebuild_data() def _rebuild_data(self): """Rebuild raw_data """ self._data = {k:np.array(v.get('value')).reshape(-1, self._n) for k,v in self._raw_data.iteritems()} def get_average(self, name): """Get average. Parameters ---------- name : str Key name of raw_data. """ return self._data.get(name).mean(axis=1) def get_errorbar(self, name): """Get errorbar Parameters ---------- name : str Key name of raw_data. """ return self._data.get(name).std(axis=1) def get_all_names(self): """Get all key names of raw_data. Returns ------- ret : list List of keys. """ return self._data.keys() def save(self, ext='dat'): """ """ pass <commit_msg>Make raw_data and data as properties<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- """utils for data crunching, saving. """ import numpy as np class ScanDataFactory(object): """Post processor of data from scan server. Parameters ---------- data : dict Raw data retrieving from scan server regarding scan ID, after completing certain scan task. n_sample : int Sample number for every scan device setup. """ def __init__(self, data, n_sample): self._raw_data = data self._n = n_sample self._rebuild_data() @property def raw_data(self): """dict: Dict of array, raw scan data.""" return self._raw_data @property def data(self): """dict: Dict of array, raw scan data after postprocessing.""" return self._data def _rebuild_data(self): """Rebuild raw_data """ self._data = {k:np.array(v.get('value')).reshape(-1, self._n) for k,v in self._raw_data.iteritems()} def get_average(self, name): """Get average. Parameters ---------- name : str Key name of raw_data. """ return self._data.get(name).mean(axis=1) def get_errorbar(self, name): """Get errorbar Parameters ---------- name : str Key name of raw_data. """ return self._data.get(name).std(axis=1) def get_all_names(self): """Get all key names of raw_data. Returns ------- ret : list List of keys. """ return self._data.keys() def save(self, ext='dat'): """ """ pass