commit stringlengths 40 40 | old_file stringlengths 4 118 | new_file stringlengths 4 118 | old_contents stringlengths 0 2.94k | new_contents stringlengths 1 4.43k | subject stringlengths 15 444 | message stringlengths 16 3.45k | lang stringclasses 1 value | license stringclasses 13 values | repos stringlengths 5 43.2k | prompt stringlengths 17 4.58k | response stringlengths 1 4.43k | prompt_tagged stringlengths 58 4.62k | response_tagged stringlengths 1 4.43k | text stringlengths 132 7.29k | text_tagged stringlengths 173 7.33k |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
c0549776224eaceda575c4eea37defa2acc0557b | setup.py | setup.py | import os
try:
from setuptools import setup
except:
from distutils.core import setup
setup(name = "clrsvsim",
version = "0.0.2",
description = "Color Genomics Structural Variant Simulator",
author = "Color Genomics",
author_email = "dev@color.com",
url = "https://github.com/ColorGenomics/clrsvsim",
packages = ["clrsvsim"],
install_requires=[
'cigar==0.1.3',
'mock==2.0.0',
'nose==1.3.7',
'numpy==1.10.1',
'preconditions==0.1',
'pyfasta==0.5.2',
'pysam==0.10.0',
],
license = "Apache-2.0",
)
| try:
from setuptools import setup
except ImportError:
from distutils.core import setup
setup(name = "clrsvsim",
version = "0.0.2",
description = "Color Genomics Structural Variant Simulator",
author = "Color Genomics",
author_email = "dev@color.com",
url = "https://github.com/ColorGenomics/clrsvsim",
packages = ["clrsvsim"],
install_requires=[
'cigar==0.1.3',
'mock==2.0.0',
'nose==1.3.7',
'numpy==1.10.1',
'preconditions==0.1',
'pyfasta==0.5.2',
'pysam==0.10.0',
],
license = "Apache-2.0",
)
| Remove unneeded import, fix bare except | Remove unneeded import, fix bare except
The `os` module isn't needed, so we need not import it.
Additionally, a bare `except` clause is pretty much never desired, since
it includes all exceptions, including sigkills.
Instead, just check for an `ImportError`, since that's what we're really
trying to do: fall back on `distutils` if `setuptools` is not present.
| Python | apache-2.0 | color/clrsvsim | import os
try:
from setuptools import setup
except:
from distutils.core import setup
setup(name = "clrsvsim",
version = "0.0.2",
description = "Color Genomics Structural Variant Simulator",
author = "Color Genomics",
author_email = "dev@color.com",
url = "https://github.com/ColorGenomics/clrsvsim",
packages = ["clrsvsim"],
install_requires=[
'cigar==0.1.3',
'mock==2.0.0',
'nose==1.3.7',
'numpy==1.10.1',
'preconditions==0.1',
'pyfasta==0.5.2',
'pysam==0.10.0',
],
license = "Apache-2.0",
)
Remove unneeded import, fix bare except
The `os` module isn't needed, so we need not import it.
Additionally, a bare `except` clause is pretty much never desired, since
it includes all exceptions, including sigkills.
Instead, just check for an `ImportError`, since that's what we're really
trying to do: fall back on `distutils` if `setuptools` is not present. | try:
from setuptools import setup
except ImportError:
from distutils.core import setup
setup(name = "clrsvsim",
version = "0.0.2",
description = "Color Genomics Structural Variant Simulator",
author = "Color Genomics",
author_email = "dev@color.com",
url = "https://github.com/ColorGenomics/clrsvsim",
packages = ["clrsvsim"],
install_requires=[
'cigar==0.1.3',
'mock==2.0.0',
'nose==1.3.7',
'numpy==1.10.1',
'preconditions==0.1',
'pyfasta==0.5.2',
'pysam==0.10.0',
],
license = "Apache-2.0",
)
| <commit_before>import os
try:
from setuptools import setup
except:
from distutils.core import setup
setup(name = "clrsvsim",
version = "0.0.2",
description = "Color Genomics Structural Variant Simulator",
author = "Color Genomics",
author_email = "dev@color.com",
url = "https://github.com/ColorGenomics/clrsvsim",
packages = ["clrsvsim"],
install_requires=[
'cigar==0.1.3',
'mock==2.0.0',
'nose==1.3.7',
'numpy==1.10.1',
'preconditions==0.1',
'pyfasta==0.5.2',
'pysam==0.10.0',
],
license = "Apache-2.0",
)
<commit_msg>Remove unneeded import, fix bare except
The `os` module isn't needed, so we need not import it.
Additionally, a bare `except` clause is pretty much never desired, since
it includes all exceptions, including sigkills.
Instead, just check for an `ImportError`, since that's what we're really
trying to do: fall back on `distutils` if `setuptools` is not present.<commit_after> | try:
from setuptools import setup
except ImportError:
from distutils.core import setup
setup(name = "clrsvsim",
version = "0.0.2",
description = "Color Genomics Structural Variant Simulator",
author = "Color Genomics",
author_email = "dev@color.com",
url = "https://github.com/ColorGenomics/clrsvsim",
packages = ["clrsvsim"],
install_requires=[
'cigar==0.1.3',
'mock==2.0.0',
'nose==1.3.7',
'numpy==1.10.1',
'preconditions==0.1',
'pyfasta==0.5.2',
'pysam==0.10.0',
],
license = "Apache-2.0",
)
| import os
try:
from setuptools import setup
except:
from distutils.core import setup
setup(name = "clrsvsim",
version = "0.0.2",
description = "Color Genomics Structural Variant Simulator",
author = "Color Genomics",
author_email = "dev@color.com",
url = "https://github.com/ColorGenomics/clrsvsim",
packages = ["clrsvsim"],
install_requires=[
'cigar==0.1.3',
'mock==2.0.0',
'nose==1.3.7',
'numpy==1.10.1',
'preconditions==0.1',
'pyfasta==0.5.2',
'pysam==0.10.0',
],
license = "Apache-2.0",
)
Remove unneeded import, fix bare except
The `os` module isn't needed, so we need not import it.
Additionally, a bare `except` clause is pretty much never desired, since
it includes all exceptions, including sigkills.
Instead, just check for an `ImportError`, since that's what we're really
trying to do: fall back on `distutils` if `setuptools` is not present.try:
from setuptools import setup
except ImportError:
from distutils.core import setup
setup(name = "clrsvsim",
version = "0.0.2",
description = "Color Genomics Structural Variant Simulator",
author = "Color Genomics",
author_email = "dev@color.com",
url = "https://github.com/ColorGenomics/clrsvsim",
packages = ["clrsvsim"],
install_requires=[
'cigar==0.1.3',
'mock==2.0.0',
'nose==1.3.7',
'numpy==1.10.1',
'preconditions==0.1',
'pyfasta==0.5.2',
'pysam==0.10.0',
],
license = "Apache-2.0",
)
| <commit_before>import os
try:
from setuptools import setup
except:
from distutils.core import setup
setup(name = "clrsvsim",
version = "0.0.2",
description = "Color Genomics Structural Variant Simulator",
author = "Color Genomics",
author_email = "dev@color.com",
url = "https://github.com/ColorGenomics/clrsvsim",
packages = ["clrsvsim"],
install_requires=[
'cigar==0.1.3',
'mock==2.0.0',
'nose==1.3.7',
'numpy==1.10.1',
'preconditions==0.1',
'pyfasta==0.5.2',
'pysam==0.10.0',
],
license = "Apache-2.0",
)
<commit_msg>Remove unneeded import, fix bare except
The `os` module isn't needed, so we need not import it.
Additionally, a bare `except` clause is pretty much never desired, since
it includes all exceptions, including sigkills.
Instead, just check for an `ImportError`, since that's what we're really
trying to do: fall back on `distutils` if `setuptools` is not present.<commit_after>try:
from setuptools import setup
except ImportError:
from distutils.core import setup
setup(name = "clrsvsim",
version = "0.0.2",
description = "Color Genomics Structural Variant Simulator",
author = "Color Genomics",
author_email = "dev@color.com",
url = "https://github.com/ColorGenomics/clrsvsim",
packages = ["clrsvsim"],
install_requires=[
'cigar==0.1.3',
'mock==2.0.0',
'nose==1.3.7',
'numpy==1.10.1',
'preconditions==0.1',
'pyfasta==0.5.2',
'pysam==0.10.0',
],
license = "Apache-2.0",
)
|
2d7be7f8344a928aecdb2bbfeb7531bb0c35aeee | setup.py | setup.py | from setuptools import setup
from mdtoc import __version__
long_description = "Adds table of contents to Markdown files"
setup(
name="mdtoc",
version=__version__,
description=long_description,
author="Scott Frazer",
author_email="scott.d.frazer@gmail.com",
packages=["mdtoc"],
install_requires=["xtermcolor", "requests<3.0.0"],
scripts={"scripts/mdtoc"},
license="MIT",
keywords="Markdown, table of contents, toc",
url="http://github.com/scottfrazer/mdtoc",
classifiers=[
"License :: OSI Approved :: MIT License",
"Environment :: Console",
"Topic :: Utilities",
"Topic :: Text Processing",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"Natural Language :: English",
],
)
| import os
from setuptools import setup
from mdtoc import __version__
setup(
name="mdtoc",
version=__version__,
description="Adds table of contents to Markdown files",
long_description=open(
os.path.join(os.path.abspath(os.path.dirname(__file__)), "README.md")
).read(),
author="Scott Frazer",
author_email="scott.d.frazer@gmail.com",
packages=["mdtoc"],
install_requires=["xtermcolor", "requests<3.0.0"],
scripts={"scripts/mdtoc"},
license="MIT",
keywords="Markdown, table of contents, toc",
url="http://github.com/scottfrazer/mdtoc",
classifiers=[
"License :: OSI Approved :: MIT License",
"Environment :: Console",
"Topic :: Utilities",
"Topic :: Text Processing",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"Natural Language :: English",
],
)
| Use a .md long_description for PyPI | Use a .md long_description for PyPI
| Python | mit | scottfrazer/mdtoc | from setuptools import setup
from mdtoc import __version__
long_description = "Adds table of contents to Markdown files"
setup(
name="mdtoc",
version=__version__,
description=long_description,
author="Scott Frazer",
author_email="scott.d.frazer@gmail.com",
packages=["mdtoc"],
install_requires=["xtermcolor", "requests<3.0.0"],
scripts={"scripts/mdtoc"},
license="MIT",
keywords="Markdown, table of contents, toc",
url="http://github.com/scottfrazer/mdtoc",
classifiers=[
"License :: OSI Approved :: MIT License",
"Environment :: Console",
"Topic :: Utilities",
"Topic :: Text Processing",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"Natural Language :: English",
],
)
Use a .md long_description for PyPI | import os
from setuptools import setup
from mdtoc import __version__
setup(
name="mdtoc",
version=__version__,
description="Adds table of contents to Markdown files",
long_description=open(
os.path.join(os.path.abspath(os.path.dirname(__file__)), "README.md")
).read(),
author="Scott Frazer",
author_email="scott.d.frazer@gmail.com",
packages=["mdtoc"],
install_requires=["xtermcolor", "requests<3.0.0"],
scripts={"scripts/mdtoc"},
license="MIT",
keywords="Markdown, table of contents, toc",
url="http://github.com/scottfrazer/mdtoc",
classifiers=[
"License :: OSI Approved :: MIT License",
"Environment :: Console",
"Topic :: Utilities",
"Topic :: Text Processing",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"Natural Language :: English",
],
)
| <commit_before>from setuptools import setup
from mdtoc import __version__
long_description = "Adds table of contents to Markdown files"
setup(
name="mdtoc",
version=__version__,
description=long_description,
author="Scott Frazer",
author_email="scott.d.frazer@gmail.com",
packages=["mdtoc"],
install_requires=["xtermcolor", "requests<3.0.0"],
scripts={"scripts/mdtoc"},
license="MIT",
keywords="Markdown, table of contents, toc",
url="http://github.com/scottfrazer/mdtoc",
classifiers=[
"License :: OSI Approved :: MIT License",
"Environment :: Console",
"Topic :: Utilities",
"Topic :: Text Processing",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"Natural Language :: English",
],
)
<commit_msg>Use a .md long_description for PyPI<commit_after> | import os
from setuptools import setup
from mdtoc import __version__
setup(
name="mdtoc",
version=__version__,
description="Adds table of contents to Markdown files",
long_description=open(
os.path.join(os.path.abspath(os.path.dirname(__file__)), "README.md")
).read(),
author="Scott Frazer",
author_email="scott.d.frazer@gmail.com",
packages=["mdtoc"],
install_requires=["xtermcolor", "requests<3.0.0"],
scripts={"scripts/mdtoc"},
license="MIT",
keywords="Markdown, table of contents, toc",
url="http://github.com/scottfrazer/mdtoc",
classifiers=[
"License :: OSI Approved :: MIT License",
"Environment :: Console",
"Topic :: Utilities",
"Topic :: Text Processing",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"Natural Language :: English",
],
)
| from setuptools import setup
from mdtoc import __version__
long_description = "Adds table of contents to Markdown files"
setup(
name="mdtoc",
version=__version__,
description=long_description,
author="Scott Frazer",
author_email="scott.d.frazer@gmail.com",
packages=["mdtoc"],
install_requires=["xtermcolor", "requests<3.0.0"],
scripts={"scripts/mdtoc"},
license="MIT",
keywords="Markdown, table of contents, toc",
url="http://github.com/scottfrazer/mdtoc",
classifiers=[
"License :: OSI Approved :: MIT License",
"Environment :: Console",
"Topic :: Utilities",
"Topic :: Text Processing",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"Natural Language :: English",
],
)
Use a .md long_description for PyPIimport os
from setuptools import setup
from mdtoc import __version__
setup(
name="mdtoc",
version=__version__,
description="Adds table of contents to Markdown files",
long_description=open(
os.path.join(os.path.abspath(os.path.dirname(__file__)), "README.md")
).read(),
author="Scott Frazer",
author_email="scott.d.frazer@gmail.com",
packages=["mdtoc"],
install_requires=["xtermcolor", "requests<3.0.0"],
scripts={"scripts/mdtoc"},
license="MIT",
keywords="Markdown, table of contents, toc",
url="http://github.com/scottfrazer/mdtoc",
classifiers=[
"License :: OSI Approved :: MIT License",
"Environment :: Console",
"Topic :: Utilities",
"Topic :: Text Processing",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"Natural Language :: English",
],
)
| <commit_before>from setuptools import setup
from mdtoc import __version__
long_description = "Adds table of contents to Markdown files"
setup(
name="mdtoc",
version=__version__,
description=long_description,
author="Scott Frazer",
author_email="scott.d.frazer@gmail.com",
packages=["mdtoc"],
install_requires=["xtermcolor", "requests<3.0.0"],
scripts={"scripts/mdtoc"},
license="MIT",
keywords="Markdown, table of contents, toc",
url="http://github.com/scottfrazer/mdtoc",
classifiers=[
"License :: OSI Approved :: MIT License",
"Environment :: Console",
"Topic :: Utilities",
"Topic :: Text Processing",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"Natural Language :: English",
],
)
<commit_msg>Use a .md long_description for PyPI<commit_after>import os
from setuptools import setup
from mdtoc import __version__
setup(
name="mdtoc",
version=__version__,
description="Adds table of contents to Markdown files",
long_description=open(
os.path.join(os.path.abspath(os.path.dirname(__file__)), "README.md")
).read(),
author="Scott Frazer",
author_email="scott.d.frazer@gmail.com",
packages=["mdtoc"],
install_requires=["xtermcolor", "requests<3.0.0"],
scripts={"scripts/mdtoc"},
license="MIT",
keywords="Markdown, table of contents, toc",
url="http://github.com/scottfrazer/mdtoc",
classifiers=[
"License :: OSI Approved :: MIT License",
"Environment :: Console",
"Topic :: Utilities",
"Topic :: Text Processing",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.4",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"Natural Language :: English",
],
)
|
9a0de77615c943de4344b4c74d5d5114b8baf0ab | eggsclaim.py | eggsclaim.py | import signal
import sys
import serial
import sms
from xbee import XBee
egg_was_present = False
def signal_handler(signal, frame):
xbee.halt()
serial_port.close()
sys.exit(0)
def packet_received(packet):
samples = packet['samples'][0]
egg_is_present = True if 'dio-4' in samples else False
if egg_was_present != egg_is_present and egg_is_present:
sms.send_sms("", "Cock-a-doodle-doo! An egg is waiting for you!")
egg_was_present = egg_is_present
signal.signal(signal.SIGINT, signal_handler)
serial_port = serial.Serial('/dev/ttyp0', 9600)
xbee = XBee(serial_port, callback=packet_received) | import signal
import sys
import serial
import sms
from xbee import XBee
egg_was_present = False
def signal_handler(signal, frame):
xbee.halt()
serial_port.close()
sys.exit(0)
def packet_received(packet):
samples = packet['samples'][0]
egg_is_present = True if 'dio-4' in samples else False
if egg_is_present and egg_is_present != egg_was_present:
sms.send_sms("", "Cock-a-doodle-doo! An egg is waiting for you!")
egg_was_present = egg_is_present
signal.signal(signal.SIGINT, signal_handler)
serial_port = serial.Serial('/dev/ttyp0', 9600)
xbee = XBee(serial_port, callback=packet_received) | Rearrange conditions to make logic clearer | Rearrange conditions to make logic clearer
| Python | mit | jamespettigrew/eggsclaim | import signal
import sys
import serial
import sms
from xbee import XBee
egg_was_present = False
def signal_handler(signal, frame):
xbee.halt()
serial_port.close()
sys.exit(0)
def packet_received(packet):
samples = packet['samples'][0]
egg_is_present = True if 'dio-4' in samples else False
if egg_was_present != egg_is_present and egg_is_present:
sms.send_sms("", "Cock-a-doodle-doo! An egg is waiting for you!")
egg_was_present = egg_is_present
signal.signal(signal.SIGINT, signal_handler)
serial_port = serial.Serial('/dev/ttyp0', 9600)
xbee = XBee(serial_port, callback=packet_received)Rearrange conditions to make logic clearer | import signal
import sys
import serial
import sms
from xbee import XBee
egg_was_present = False
def signal_handler(signal, frame):
xbee.halt()
serial_port.close()
sys.exit(0)
def packet_received(packet):
samples = packet['samples'][0]
egg_is_present = True if 'dio-4' in samples else False
if egg_is_present and egg_is_present != egg_was_present:
sms.send_sms("", "Cock-a-doodle-doo! An egg is waiting for you!")
egg_was_present = egg_is_present
signal.signal(signal.SIGINT, signal_handler)
serial_port = serial.Serial('/dev/ttyp0', 9600)
xbee = XBee(serial_port, callback=packet_received) | <commit_before>import signal
import sys
import serial
import sms
from xbee import XBee
egg_was_present = False
def signal_handler(signal, frame):
xbee.halt()
serial_port.close()
sys.exit(0)
def packet_received(packet):
samples = packet['samples'][0]
egg_is_present = True if 'dio-4' in samples else False
if egg_was_present != egg_is_present and egg_is_present:
sms.send_sms("", "Cock-a-doodle-doo! An egg is waiting for you!")
egg_was_present = egg_is_present
signal.signal(signal.SIGINT, signal_handler)
serial_port = serial.Serial('/dev/ttyp0', 9600)
xbee = XBee(serial_port, callback=packet_received)<commit_msg>Rearrange conditions to make logic clearer<commit_after> | import signal
import sys
import serial
import sms
from xbee import XBee
egg_was_present = False
def signal_handler(signal, frame):
xbee.halt()
serial_port.close()
sys.exit(0)
def packet_received(packet):
samples = packet['samples'][0]
egg_is_present = True if 'dio-4' in samples else False
if egg_is_present and egg_is_present != egg_was_present:
sms.send_sms("", "Cock-a-doodle-doo! An egg is waiting for you!")
egg_was_present = egg_is_present
signal.signal(signal.SIGINT, signal_handler)
serial_port = serial.Serial('/dev/ttyp0', 9600)
xbee = XBee(serial_port, callback=packet_received) | import signal
import sys
import serial
import sms
from xbee import XBee
egg_was_present = False
def signal_handler(signal, frame):
xbee.halt()
serial_port.close()
sys.exit(0)
def packet_received(packet):
samples = packet['samples'][0]
egg_is_present = True if 'dio-4' in samples else False
if egg_was_present != egg_is_present and egg_is_present:
sms.send_sms("", "Cock-a-doodle-doo! An egg is waiting for you!")
egg_was_present = egg_is_present
signal.signal(signal.SIGINT, signal_handler)
serial_port = serial.Serial('/dev/ttyp0', 9600)
xbee = XBee(serial_port, callback=packet_received)Rearrange conditions to make logic clearerimport signal
import sys
import serial
import sms
from xbee import XBee
egg_was_present = False
def signal_handler(signal, frame):
xbee.halt()
serial_port.close()
sys.exit(0)
def packet_received(packet):
samples = packet['samples'][0]
egg_is_present = True if 'dio-4' in samples else False
if egg_is_present and egg_is_present != egg_was_present:
sms.send_sms("", "Cock-a-doodle-doo! An egg is waiting for you!")
egg_was_present = egg_is_present
signal.signal(signal.SIGINT, signal_handler)
serial_port = serial.Serial('/dev/ttyp0', 9600)
xbee = XBee(serial_port, callback=packet_received) | <commit_before>import signal
import sys
import serial
import sms
from xbee import XBee
egg_was_present = False
def signal_handler(signal, frame):
xbee.halt()
serial_port.close()
sys.exit(0)
def packet_received(packet):
samples = packet['samples'][0]
egg_is_present = True if 'dio-4' in samples else False
if egg_was_present != egg_is_present and egg_is_present:
sms.send_sms("", "Cock-a-doodle-doo! An egg is waiting for you!")
egg_was_present = egg_is_present
signal.signal(signal.SIGINT, signal_handler)
serial_port = serial.Serial('/dev/ttyp0', 9600)
xbee = XBee(serial_port, callback=packet_received)<commit_msg>Rearrange conditions to make logic clearer<commit_after>import signal
import sys
import serial
import sms
from xbee import XBee
egg_was_present = False
def signal_handler(signal, frame):
xbee.halt()
serial_port.close()
sys.exit(0)
def packet_received(packet):
samples = packet['samples'][0]
egg_is_present = True if 'dio-4' in samples else False
if egg_is_present and egg_is_present != egg_was_present:
sms.send_sms("", "Cock-a-doodle-doo! An egg is waiting for you!")
egg_was_present = egg_is_present
signal.signal(signal.SIGINT, signal_handler)
serial_port = serial.Serial('/dev/ttyp0', 9600)
xbee = XBee(serial_port, callback=packet_received) |
e8c9c22c7c57ff2de8b9ef9e73ec8f339aa73fd7 | setup.py | setup.py | #!/usr/bin/env python
import os
import sys
if sys.version < '2.7':
print 'Python >= 2.7 required'
sys.exit(1)
from setuptools import setup
long_description = '''
A simple Python package for using Twitter search functionality
that is only available through the Twitter web interface
(such as searching for tweets older than a few weeks).'''.strip()
setup(
name = 'twitterwebsearch',
version='0.0.5',
author = 'Raynor Vliegendhart',
author_email = 'ShinNoNoir@gmail.com',
url = 'https://github.com/ShinNoNoir/twitterwebsearch',
packages=['twitterwebsearch'],
description = "Package for Twitter's web search",
long_description = long_description,
platforms = 'Any',
license = 'MIT (see: LICENSE.txt)',
keywords = 'Twitter, search',
install_requires = open('requirements.txt').readlines(),
)
| #!/usr/bin/env python
import os
import sys
if sys.version < '2.7':
print 'Python >= 2.7 required'
sys.exit(1)
from setuptools import setup
long_description = '''
A simple Python package for using Twitter search functionality
that is only available through the Twitter web interface
(such as searching for tweets older than a few weeks).'''.strip()
setup(
name = 'twitterwebsearch',
version='0.1.1',
author = 'Raynor Vliegendhart',
author_email = 'ShinNoNoir@gmail.com',
url = 'https://github.com/ShinNoNoir/twitterwebsearch',
packages=['twitterwebsearch'],
description = "Package for Twitter's web search",
long_description = long_description,
platforms = 'Any',
license = 'MIT (see: LICENSE.txt)',
keywords = 'Twitter, search',
install_requires = open('requirements.txt').readlines(),
)
| Increase version number to 0.1.1 (breaking API change) | Increase version number to 0.1.1 (breaking API change)
| Python | mit | ShinNoNoir/twitterwebsearch | #!/usr/bin/env python
import os
import sys
if sys.version < '2.7':
print 'Python >= 2.7 required'
sys.exit(1)
from setuptools import setup
long_description = '''
A simple Python package for using Twitter search functionality
that is only available through the Twitter web interface
(such as searching for tweets older than a few weeks).'''.strip()
setup(
name = 'twitterwebsearch',
version='0.0.5',
author = 'Raynor Vliegendhart',
author_email = 'ShinNoNoir@gmail.com',
url = 'https://github.com/ShinNoNoir/twitterwebsearch',
packages=['twitterwebsearch'],
description = "Package for Twitter's web search",
long_description = long_description,
platforms = 'Any',
license = 'MIT (see: LICENSE.txt)',
keywords = 'Twitter, search',
install_requires = open('requirements.txt').readlines(),
)
Increase version number to 0.1.1 (breaking API change) | #!/usr/bin/env python
import os
import sys
if sys.version < '2.7':
print 'Python >= 2.7 required'
sys.exit(1)
from setuptools import setup
long_description = '''
A simple Python package for using Twitter search functionality
that is only available through the Twitter web interface
(such as searching for tweets older than a few weeks).'''.strip()
setup(
name = 'twitterwebsearch',
version='0.1.1',
author = 'Raynor Vliegendhart',
author_email = 'ShinNoNoir@gmail.com',
url = 'https://github.com/ShinNoNoir/twitterwebsearch',
packages=['twitterwebsearch'],
description = "Package for Twitter's web search",
long_description = long_description,
platforms = 'Any',
license = 'MIT (see: LICENSE.txt)',
keywords = 'Twitter, search',
install_requires = open('requirements.txt').readlines(),
)
| <commit_before>#!/usr/bin/env python
import os
import sys
if sys.version < '2.7':
print 'Python >= 2.7 required'
sys.exit(1)
from setuptools import setup
long_description = '''
A simple Python package for using Twitter search functionality
that is only available through the Twitter web interface
(such as searching for tweets older than a few weeks).'''.strip()
setup(
name = 'twitterwebsearch',
version='0.0.5',
author = 'Raynor Vliegendhart',
author_email = 'ShinNoNoir@gmail.com',
url = 'https://github.com/ShinNoNoir/twitterwebsearch',
packages=['twitterwebsearch'],
description = "Package for Twitter's web search",
long_description = long_description,
platforms = 'Any',
license = 'MIT (see: LICENSE.txt)',
keywords = 'Twitter, search',
install_requires = open('requirements.txt').readlines(),
)
<commit_msg>Increase version number to 0.1.1 (breaking API change)<commit_after> | #!/usr/bin/env python
import os
import sys
if sys.version < '2.7':
print 'Python >= 2.7 required'
sys.exit(1)
from setuptools import setup
long_description = '''
A simple Python package for using Twitter search functionality
that is only available through the Twitter web interface
(such as searching for tweets older than a few weeks).'''.strip()
setup(
name = 'twitterwebsearch',
version='0.1.1',
author = 'Raynor Vliegendhart',
author_email = 'ShinNoNoir@gmail.com',
url = 'https://github.com/ShinNoNoir/twitterwebsearch',
packages=['twitterwebsearch'],
description = "Package for Twitter's web search",
long_description = long_description,
platforms = 'Any',
license = 'MIT (see: LICENSE.txt)',
keywords = 'Twitter, search',
install_requires = open('requirements.txt').readlines(),
)
| #!/usr/bin/env python
import os
import sys
if sys.version < '2.7':
print 'Python >= 2.7 required'
sys.exit(1)
from setuptools import setup
long_description = '''
A simple Python package for using Twitter search functionality
that is only available through the Twitter web interface
(such as searching for tweets older than a few weeks).'''.strip()
setup(
name = 'twitterwebsearch',
version='0.0.5',
author = 'Raynor Vliegendhart',
author_email = 'ShinNoNoir@gmail.com',
url = 'https://github.com/ShinNoNoir/twitterwebsearch',
packages=['twitterwebsearch'],
description = "Package for Twitter's web search",
long_description = long_description,
platforms = 'Any',
license = 'MIT (see: LICENSE.txt)',
keywords = 'Twitter, search',
install_requires = open('requirements.txt').readlines(),
)
Increase version number to 0.1.1 (breaking API change)#!/usr/bin/env python
import os
import sys
if sys.version < '2.7':
print 'Python >= 2.7 required'
sys.exit(1)
from setuptools import setup
long_description = '''
A simple Python package for using Twitter search functionality
that is only available through the Twitter web interface
(such as searching for tweets older than a few weeks).'''.strip()
setup(
name = 'twitterwebsearch',
version='0.1.1',
author = 'Raynor Vliegendhart',
author_email = 'ShinNoNoir@gmail.com',
url = 'https://github.com/ShinNoNoir/twitterwebsearch',
packages=['twitterwebsearch'],
description = "Package for Twitter's web search",
long_description = long_description,
platforms = 'Any',
license = 'MIT (see: LICENSE.txt)',
keywords = 'Twitter, search',
install_requires = open('requirements.txt').readlines(),
)
| <commit_before>#!/usr/bin/env python
import os
import sys
if sys.version < '2.7':
print 'Python >= 2.7 required'
sys.exit(1)
from setuptools import setup
long_description = '''
A simple Python package for using Twitter search functionality
that is only available through the Twitter web interface
(such as searching for tweets older than a few weeks).'''.strip()
setup(
name = 'twitterwebsearch',
version='0.0.5',
author = 'Raynor Vliegendhart',
author_email = 'ShinNoNoir@gmail.com',
url = 'https://github.com/ShinNoNoir/twitterwebsearch',
packages=['twitterwebsearch'],
description = "Package for Twitter's web search",
long_description = long_description,
platforms = 'Any',
license = 'MIT (see: LICENSE.txt)',
keywords = 'Twitter, search',
install_requires = open('requirements.txt').readlines(),
)
<commit_msg>Increase version number to 0.1.1 (breaking API change)<commit_after>#!/usr/bin/env python
import os
import sys
if sys.version < '2.7':
print 'Python >= 2.7 required'
sys.exit(1)
from setuptools import setup
long_description = '''
A simple Python package for using Twitter search functionality
that is only available through the Twitter web interface
(such as searching for tweets older than a few weeks).'''.strip()
setup(
name = 'twitterwebsearch',
version='0.1.1',
author = 'Raynor Vliegendhart',
author_email = 'ShinNoNoir@gmail.com',
url = 'https://github.com/ShinNoNoir/twitterwebsearch',
packages=['twitterwebsearch'],
description = "Package for Twitter's web search",
long_description = long_description,
platforms = 'Any',
license = 'MIT (see: LICENSE.txt)',
keywords = 'Twitter, search',
install_requires = open('requirements.txt').readlines(),
)
|
48418ac0fe75bbb331878b80d9d0903dde445838 | setup.py | setup.py | from distutils.core import setup
setup(
name='pyticketswitch',
version='1.6.1',
author='Matt Jared',
author_email='mattjared@ingresso.co.uk',
packages=[
'pyticketswitch',
'pyticketswitch.test',
'pyticketswitch.interface_objects'
],
license='LICENSE.txt',
description='A Python interface for the Ingresso XML Core API',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 5 - Production/Stable',
'Programming Language :: Python :: 2.7',
],
)
| from distutils.core import setup
setup(
name='pyticketswitch',
version='1.6.1',
author='Ingresso',
author_email='systems@ingresso.co.uk',
packages=[
'pyticketswitch',
'pyticketswitch.test',
'pyticketswitch.interface_objects'
],
license='LICENSE.txt',
description='A Python interface for the Ingresso XML Core API',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 5 - Production/Stable',
'Programming Language :: Python :: 2.7',
],
)
| Update author and email address | Update author and email address
| Python | mit | ingresso-group/pyticketswitch,ingtechteam/pyticketswitch,graingert/pyticketswitch | from distutils.core import setup
setup(
name='pyticketswitch',
version='1.6.1',
author='Matt Jared',
author_email='mattjared@ingresso.co.uk',
packages=[
'pyticketswitch',
'pyticketswitch.test',
'pyticketswitch.interface_objects'
],
license='LICENSE.txt',
description='A Python interface for the Ingresso XML Core API',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 5 - Production/Stable',
'Programming Language :: Python :: 2.7',
],
)
Update author and email address | from distutils.core import setup
setup(
name='pyticketswitch',
version='1.6.1',
author='Ingresso',
author_email='systems@ingresso.co.uk',
packages=[
'pyticketswitch',
'pyticketswitch.test',
'pyticketswitch.interface_objects'
],
license='LICENSE.txt',
description='A Python interface for the Ingresso XML Core API',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 5 - Production/Stable',
'Programming Language :: Python :: 2.7',
],
)
| <commit_before>from distutils.core import setup
setup(
name='pyticketswitch',
version='1.6.1',
author='Matt Jared',
author_email='mattjared@ingresso.co.uk',
packages=[
'pyticketswitch',
'pyticketswitch.test',
'pyticketswitch.interface_objects'
],
license='LICENSE.txt',
description='A Python interface for the Ingresso XML Core API',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 5 - Production/Stable',
'Programming Language :: Python :: 2.7',
],
)
<commit_msg>Update author and email address<commit_after> | from distutils.core import setup
setup(
name='pyticketswitch',
version='1.6.1',
author='Ingresso',
author_email='systems@ingresso.co.uk',
packages=[
'pyticketswitch',
'pyticketswitch.test',
'pyticketswitch.interface_objects'
],
license='LICENSE.txt',
description='A Python interface for the Ingresso XML Core API',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 5 - Production/Stable',
'Programming Language :: Python :: 2.7',
],
)
| from distutils.core import setup
setup(
name='pyticketswitch',
version='1.6.1',
author='Matt Jared',
author_email='mattjared@ingresso.co.uk',
packages=[
'pyticketswitch',
'pyticketswitch.test',
'pyticketswitch.interface_objects'
],
license='LICENSE.txt',
description='A Python interface for the Ingresso XML Core API',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 5 - Production/Stable',
'Programming Language :: Python :: 2.7',
],
)
Update author and email addressfrom distutils.core import setup
setup(
name='pyticketswitch',
version='1.6.1',
author='Ingresso',
author_email='systems@ingresso.co.uk',
packages=[
'pyticketswitch',
'pyticketswitch.test',
'pyticketswitch.interface_objects'
],
license='LICENSE.txt',
description='A Python interface for the Ingresso XML Core API',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 5 - Production/Stable',
'Programming Language :: Python :: 2.7',
],
)
| <commit_before>from distutils.core import setup
setup(
name='pyticketswitch',
version='1.6.1',
author='Matt Jared',
author_email='mattjared@ingresso.co.uk',
packages=[
'pyticketswitch',
'pyticketswitch.test',
'pyticketswitch.interface_objects'
],
license='LICENSE.txt',
description='A Python interface for the Ingresso XML Core API',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 5 - Production/Stable',
'Programming Language :: Python :: 2.7',
],
)
<commit_msg>Update author and email address<commit_after>from distutils.core import setup
setup(
name='pyticketswitch',
version='1.6.1',
author='Ingresso',
author_email='systems@ingresso.co.uk',
packages=[
'pyticketswitch',
'pyticketswitch.test',
'pyticketswitch.interface_objects'
],
license='LICENSE.txt',
description='A Python interface for the Ingresso XML Core API',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 5 - Production/Stable',
'Programming Language :: Python :: 2.7',
],
)
|
5782d01fa95624c784c6298486caefbe527fb76f | setup.py | setup.py | from distutils.core import setup
from hal.version import __version__ as version
setup(
name='hal',
packages=['hal'],
version=version,
description='Command Line Assistant',
author='Anup Pokhrel',
author_email='virtualanup@gmail.com',
url='https://github.com/virtualanup/hal',
download_url='https://github.com/virtualanup/hal/archive/{}.tar.gz'.format(version),
keywords=['assistant', 'hal'],
classifiers=[],
install_requires=[
'six==1.10.0',
'simpleeval==0.9.1',
'pytz==2016.10'
],
)
| from distutils.core import setup
from hal.version import __version__ as version
setup(
name='hal-assistant',
packages=['hal'],
version=version,
description='Command Line Assistant',
author='Anup Pokhrel',
author_email='virtualanup@gmail.com',
url='https://github.com/virtualanup/hal',
download_url='https://github.com/virtualanup/hal/archive/{}.tar.gz'.format(version),
keywords=['assistant', 'hal'],
classifiers=[],
install_requires=[
'six==1.10.0',
'simpleeval==0.9.1',
'pytz==2016.10'
],
)
| Change package name to hal-assistant | Change package name to hal-assistant
| Python | mit | virtualanup/hal | from distutils.core import setup
from hal.version import __version__ as version
setup(
name='hal',
packages=['hal'],
version=version,
description='Command Line Assistant',
author='Anup Pokhrel',
author_email='virtualanup@gmail.com',
url='https://github.com/virtualanup/hal',
download_url='https://github.com/virtualanup/hal/archive/{}.tar.gz'.format(version),
keywords=['assistant', 'hal'],
classifiers=[],
install_requires=[
'six==1.10.0',
'simpleeval==0.9.1',
'pytz==2016.10'
],
)
Change package name to hal-assistant | from distutils.core import setup
from hal.version import __version__ as version
setup(
name='hal-assistant',
packages=['hal'],
version=version,
description='Command Line Assistant',
author='Anup Pokhrel',
author_email='virtualanup@gmail.com',
url='https://github.com/virtualanup/hal',
download_url='https://github.com/virtualanup/hal/archive/{}.tar.gz'.format(version),
keywords=['assistant', 'hal'],
classifiers=[],
install_requires=[
'six==1.10.0',
'simpleeval==0.9.1',
'pytz==2016.10'
],
)
| <commit_before>from distutils.core import setup
from hal.version import __version__ as version
setup(
name='hal',
packages=['hal'],
version=version,
description='Command Line Assistant',
author='Anup Pokhrel',
author_email='virtualanup@gmail.com',
url='https://github.com/virtualanup/hal',
download_url='https://github.com/virtualanup/hal/archive/{}.tar.gz'.format(version),
keywords=['assistant', 'hal'],
classifiers=[],
install_requires=[
'six==1.10.0',
'simpleeval==0.9.1',
'pytz==2016.10'
],
)
<commit_msg>Change package name to hal-assistant<commit_after> | from distutils.core import setup
from hal.version import __version__ as version
setup(
name='hal-assistant',
packages=['hal'],
version=version,
description='Command Line Assistant',
author='Anup Pokhrel',
author_email='virtualanup@gmail.com',
url='https://github.com/virtualanup/hal',
download_url='https://github.com/virtualanup/hal/archive/{}.tar.gz'.format(version),
keywords=['assistant', 'hal'],
classifiers=[],
install_requires=[
'six==1.10.0',
'simpleeval==0.9.1',
'pytz==2016.10'
],
)
| from distutils.core import setup
from hal.version import __version__ as version
setup(
name='hal',
packages=['hal'],
version=version,
description='Command Line Assistant',
author='Anup Pokhrel',
author_email='virtualanup@gmail.com',
url='https://github.com/virtualanup/hal',
download_url='https://github.com/virtualanup/hal/archive/{}.tar.gz'.format(version),
keywords=['assistant', 'hal'],
classifiers=[],
install_requires=[
'six==1.10.0',
'simpleeval==0.9.1',
'pytz==2016.10'
],
)
Change package name to hal-assistantfrom distutils.core import setup
from hal.version import __version__ as version
setup(
name='hal-assistant',
packages=['hal'],
version=version,
description='Command Line Assistant',
author='Anup Pokhrel',
author_email='virtualanup@gmail.com',
url='https://github.com/virtualanup/hal',
download_url='https://github.com/virtualanup/hal/archive/{}.tar.gz'.format(version),
keywords=['assistant', 'hal'],
classifiers=[],
install_requires=[
'six==1.10.0',
'simpleeval==0.9.1',
'pytz==2016.10'
],
)
| <commit_before>from distutils.core import setup
from hal.version import __version__ as version
setup(
name='hal',
packages=['hal'],
version=version,
description='Command Line Assistant',
author='Anup Pokhrel',
author_email='virtualanup@gmail.com',
url='https://github.com/virtualanup/hal',
download_url='https://github.com/virtualanup/hal/archive/{}.tar.gz'.format(version),
keywords=['assistant', 'hal'],
classifiers=[],
install_requires=[
'six==1.10.0',
'simpleeval==0.9.1',
'pytz==2016.10'
],
)
<commit_msg>Change package name to hal-assistant<commit_after>from distutils.core import setup
from hal.version import __version__ as version
setup(
name='hal-assistant',
packages=['hal'],
version=version,
description='Command Line Assistant',
author='Anup Pokhrel',
author_email='virtualanup@gmail.com',
url='https://github.com/virtualanup/hal',
download_url='https://github.com/virtualanup/hal/archive/{}.tar.gz'.format(version),
keywords=['assistant', 'hal'],
classifiers=[],
install_requires=[
'six==1.10.0',
'simpleeval==0.9.1',
'pytz==2016.10'
],
)
|
caafe83bd35b0b82135be593b88ec9ed64bfb508 | setup.py | setup.py | #! /usr/bin/env python
# -*- coding: utf-8 -*-
from setuptools import setup, find_packages
setup(
name='OpenFisca-Senegal',
version='0.0.0',
author='OpenFisca Team',
author_email='contact@openfisca.fr',
classifiers=[
"Development Status :: 2 - Pre-Alpha",
"License :: OSI Approved :: GNU Affero General Public License v3",
"Operating System :: POSIX",
"Programming Language :: Python",
"Topic :: Scientific/Engineering :: Information Analysis",
],
description=u'Senegalese tax and benefit system for OpenFisca',
keywords='benefit microsimulation senegal social tax',
license='http://www.fsf.org/licensing/licenses/agpl-3.0.html',
url='https://github.com/openfisca/openfisca-senegal',
data_files=[
('share/openfisca/openfisca-senegal',
['CHANGELOG.md', 'LICENSE', 'README.md']),
],
extras_require={
'test': ['nose'],
},
install_requires=[
'OpenFisca-Core >= 3.0.0, < 4.0',
'notebook',
],
packages=find_packages(exclude=['openfisca_senegal.tests*']),
test_suite='nose.collector',
)
| #! /usr/bin/env python
# -*- coding: utf-8 -*-
from setuptools import setup, find_packages
setup(
name='OpenFisca-Senegal',
version='0.0.0',
author='OpenFisca Team',
author_email='contact@openfisca.fr',
classifiers=[
"Development Status :: 2 - Pre-Alpha",
"License :: OSI Approved :: GNU Affero General Public License v3",
"Operating System :: POSIX",
"Programming Language :: Python",
"Topic :: Scientific/Engineering :: Information Analysis",
],
description=u'Senegalese tax and benefit system for OpenFisca',
keywords='benefit microsimulation senegal social tax',
license='http://www.fsf.org/licensing/licenses/agpl-3.0.html',
url='https://github.com/openfisca/openfisca-senegal',
data_files=[
('share/openfisca/openfisca-senegal',
['CHANGELOG.md', 'LICENSE', 'README.md']),
],
extras_require={
'test': ['nose'],
},
install_requires=[
'OpenFisca-Core >= 4.1.4b1, < 5.0',
'notebook',
],
packages=find_packages(exclude=['openfisca_senegal.tests*']),
test_suite='nose.collector',
)
| Upgrade openfisca core to v4 | Upgrade openfisca core to v4
| Python | agpl-3.0 | openfisca/senegal | #! /usr/bin/env python
# -*- coding: utf-8 -*-
from setuptools import setup, find_packages
setup(
name='OpenFisca-Senegal',
version='0.0.0',
author='OpenFisca Team',
author_email='contact@openfisca.fr',
classifiers=[
"Development Status :: 2 - Pre-Alpha",
"License :: OSI Approved :: GNU Affero General Public License v3",
"Operating System :: POSIX",
"Programming Language :: Python",
"Topic :: Scientific/Engineering :: Information Analysis",
],
description=u'Senegalese tax and benefit system for OpenFisca',
keywords='benefit microsimulation senegal social tax',
license='http://www.fsf.org/licensing/licenses/agpl-3.0.html',
url='https://github.com/openfisca/openfisca-senegal',
data_files=[
('share/openfisca/openfisca-senegal',
['CHANGELOG.md', 'LICENSE', 'README.md']),
],
extras_require={
'test': ['nose'],
},
install_requires=[
'OpenFisca-Core >= 3.0.0, < 4.0',
'notebook',
],
packages=find_packages(exclude=['openfisca_senegal.tests*']),
test_suite='nose.collector',
)
Upgrade openfisca core to v4 | #! /usr/bin/env python
# -*- coding: utf-8 -*-
from setuptools import setup, find_packages
setup(
name='OpenFisca-Senegal',
version='0.0.0',
author='OpenFisca Team',
author_email='contact@openfisca.fr',
classifiers=[
"Development Status :: 2 - Pre-Alpha",
"License :: OSI Approved :: GNU Affero General Public License v3",
"Operating System :: POSIX",
"Programming Language :: Python",
"Topic :: Scientific/Engineering :: Information Analysis",
],
description=u'Senegalese tax and benefit system for OpenFisca',
keywords='benefit microsimulation senegal social tax',
license='http://www.fsf.org/licensing/licenses/agpl-3.0.html',
url='https://github.com/openfisca/openfisca-senegal',
data_files=[
('share/openfisca/openfisca-senegal',
['CHANGELOG.md', 'LICENSE', 'README.md']),
],
extras_require={
'test': ['nose'],
},
install_requires=[
'OpenFisca-Core >= 4.1.4b1, < 5.0',
'notebook',
],
packages=find_packages(exclude=['openfisca_senegal.tests*']),
test_suite='nose.collector',
)
| <commit_before>#! /usr/bin/env python
# -*- coding: utf-8 -*-
from setuptools import setup, find_packages
setup(
name='OpenFisca-Senegal',
version='0.0.0',
author='OpenFisca Team',
author_email='contact@openfisca.fr',
classifiers=[
"Development Status :: 2 - Pre-Alpha",
"License :: OSI Approved :: GNU Affero General Public License v3",
"Operating System :: POSIX",
"Programming Language :: Python",
"Topic :: Scientific/Engineering :: Information Analysis",
],
description=u'Senegalese tax and benefit system for OpenFisca',
keywords='benefit microsimulation senegal social tax',
license='http://www.fsf.org/licensing/licenses/agpl-3.0.html',
url='https://github.com/openfisca/openfisca-senegal',
data_files=[
('share/openfisca/openfisca-senegal',
['CHANGELOG.md', 'LICENSE', 'README.md']),
],
extras_require={
'test': ['nose'],
},
install_requires=[
'OpenFisca-Core >= 3.0.0, < 4.0',
'notebook',
],
packages=find_packages(exclude=['openfisca_senegal.tests*']),
test_suite='nose.collector',
)
<commit_msg>Upgrade openfisca core to v4<commit_after> | #! /usr/bin/env python
# -*- coding: utf-8 -*-
from setuptools import setup, find_packages
setup(
name='OpenFisca-Senegal',
version='0.0.0',
author='OpenFisca Team',
author_email='contact@openfisca.fr',
classifiers=[
"Development Status :: 2 - Pre-Alpha",
"License :: OSI Approved :: GNU Affero General Public License v3",
"Operating System :: POSIX",
"Programming Language :: Python",
"Topic :: Scientific/Engineering :: Information Analysis",
],
description=u'Senegalese tax and benefit system for OpenFisca',
keywords='benefit microsimulation senegal social tax',
license='http://www.fsf.org/licensing/licenses/agpl-3.0.html',
url='https://github.com/openfisca/openfisca-senegal',
data_files=[
('share/openfisca/openfisca-senegal',
['CHANGELOG.md', 'LICENSE', 'README.md']),
],
extras_require={
'test': ['nose'],
},
install_requires=[
'OpenFisca-Core >= 4.1.4b1, < 5.0',
'notebook',
],
packages=find_packages(exclude=['openfisca_senegal.tests*']),
test_suite='nose.collector',
)
| #! /usr/bin/env python
# -*- coding: utf-8 -*-
from setuptools import setup, find_packages
setup(
name='OpenFisca-Senegal',
version='0.0.0',
author='OpenFisca Team',
author_email='contact@openfisca.fr',
classifiers=[
"Development Status :: 2 - Pre-Alpha",
"License :: OSI Approved :: GNU Affero General Public License v3",
"Operating System :: POSIX",
"Programming Language :: Python",
"Topic :: Scientific/Engineering :: Information Analysis",
],
description=u'Senegalese tax and benefit system for OpenFisca',
keywords='benefit microsimulation senegal social tax',
license='http://www.fsf.org/licensing/licenses/agpl-3.0.html',
url='https://github.com/openfisca/openfisca-senegal',
data_files=[
('share/openfisca/openfisca-senegal',
['CHANGELOG.md', 'LICENSE', 'README.md']),
],
extras_require={
'test': ['nose'],
},
install_requires=[
'OpenFisca-Core >= 3.0.0, < 4.0',
'notebook',
],
packages=find_packages(exclude=['openfisca_senegal.tests*']),
test_suite='nose.collector',
)
Upgrade openfisca core to v4#! /usr/bin/env python
# -*- coding: utf-8 -*-
from setuptools import setup, find_packages
setup(
name='OpenFisca-Senegal',
version='0.0.0',
author='OpenFisca Team',
author_email='contact@openfisca.fr',
classifiers=[
"Development Status :: 2 - Pre-Alpha",
"License :: OSI Approved :: GNU Affero General Public License v3",
"Operating System :: POSIX",
"Programming Language :: Python",
"Topic :: Scientific/Engineering :: Information Analysis",
],
description=u'Senegalese tax and benefit system for OpenFisca',
keywords='benefit microsimulation senegal social tax',
license='http://www.fsf.org/licensing/licenses/agpl-3.0.html',
url='https://github.com/openfisca/openfisca-senegal',
data_files=[
('share/openfisca/openfisca-senegal',
['CHANGELOG.md', 'LICENSE', 'README.md']),
],
extras_require={
'test': ['nose'],
},
install_requires=[
'OpenFisca-Core >= 4.1.4b1, < 5.0',
'notebook',
],
packages=find_packages(exclude=['openfisca_senegal.tests*']),
test_suite='nose.collector',
)
| <commit_before>#! /usr/bin/env python
# -*- coding: utf-8 -*-
from setuptools import setup, find_packages
setup(
name='OpenFisca-Senegal',
version='0.0.0',
author='OpenFisca Team',
author_email='contact@openfisca.fr',
classifiers=[
"Development Status :: 2 - Pre-Alpha",
"License :: OSI Approved :: GNU Affero General Public License v3",
"Operating System :: POSIX",
"Programming Language :: Python",
"Topic :: Scientific/Engineering :: Information Analysis",
],
description=u'Senegalese tax and benefit system for OpenFisca',
keywords='benefit microsimulation senegal social tax',
license='http://www.fsf.org/licensing/licenses/agpl-3.0.html',
url='https://github.com/openfisca/openfisca-senegal',
data_files=[
('share/openfisca/openfisca-senegal',
['CHANGELOG.md', 'LICENSE', 'README.md']),
],
extras_require={
'test': ['nose'],
},
install_requires=[
'OpenFisca-Core >= 3.0.0, < 4.0',
'notebook',
],
packages=find_packages(exclude=['openfisca_senegal.tests*']),
test_suite='nose.collector',
)
<commit_msg>Upgrade openfisca core to v4<commit_after>#! /usr/bin/env python
# -*- coding: utf-8 -*-
from setuptools import setup, find_packages
setup(
name='OpenFisca-Senegal',
version='0.0.0',
author='OpenFisca Team',
author_email='contact@openfisca.fr',
classifiers=[
"Development Status :: 2 - Pre-Alpha",
"License :: OSI Approved :: GNU Affero General Public License v3",
"Operating System :: POSIX",
"Programming Language :: Python",
"Topic :: Scientific/Engineering :: Information Analysis",
],
description=u'Senegalese tax and benefit system for OpenFisca',
keywords='benefit microsimulation senegal social tax',
license='http://www.fsf.org/licensing/licenses/agpl-3.0.html',
url='https://github.com/openfisca/openfisca-senegal',
data_files=[
('share/openfisca/openfisca-senegal',
['CHANGELOG.md', 'LICENSE', 'README.md']),
],
extras_require={
'test': ['nose'],
},
install_requires=[
'OpenFisca-Core >= 4.1.4b1, < 5.0',
'notebook',
],
packages=find_packages(exclude=['openfisca_senegal.tests*']),
test_suite='nose.collector',
)
|
5d5f5b2924a238452a19c1035a4d2eee9c857ceb | setup.py | setup.py | #!/usr/bin/env python
try:
from setuptools import setup, find_packages
except ImportError:
from ez_setup import use_setuptools
use_setuptools()
from setuptools import setup, find_packages
setup(
name='django-bleach',
version="0.1.0",
description='Easily use bleach with Django models and templates',
author='Tim Heap',
author_email='heap.tim@gmail.com',
url='https://bitbucket.org/ionata/django-bleach',
packages=['django_bleach',],
package_data={},
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
],
)
| #!/usr/bin/env python
try:
from setuptools import setup, find_packages
except ImportError:
from ez_setup import use_setuptools
use_setuptools()
from setuptools import setup, find_packages
setup(
name='django-bleach',
version="0.1.0",
description='Easily use bleach with Django models and templates',
author='Tim Heap',
author_email='heap.tim@gmail.com',
url='https://bitbucket.org/ionata/django-bleach',
packages=['django_bleach',],
install_requires = ['bleach'],
package_data={},
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
],
)
| Add bleach as a requirement | Add bleach as a requirement
| Python | bsd-2-clause | python-force/django-bleach | #!/usr/bin/env python
try:
from setuptools import setup, find_packages
except ImportError:
from ez_setup import use_setuptools
use_setuptools()
from setuptools import setup, find_packages
setup(
name='django-bleach',
version="0.1.0",
description='Easily use bleach with Django models and templates',
author='Tim Heap',
author_email='heap.tim@gmail.com',
url='https://bitbucket.org/ionata/django-bleach',
packages=['django_bleach',],
package_data={},
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
],
)
Add bleach as a requirement | #!/usr/bin/env python
try:
from setuptools import setup, find_packages
except ImportError:
from ez_setup import use_setuptools
use_setuptools()
from setuptools import setup, find_packages
setup(
name='django-bleach',
version="0.1.0",
description='Easily use bleach with Django models and templates',
author='Tim Heap',
author_email='heap.tim@gmail.com',
url='https://bitbucket.org/ionata/django-bleach',
packages=['django_bleach',],
install_requires = ['bleach'],
package_data={},
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
],
)
| <commit_before>#!/usr/bin/env python
try:
from setuptools import setup, find_packages
except ImportError:
from ez_setup import use_setuptools
use_setuptools()
from setuptools import setup, find_packages
setup(
name='django-bleach',
version="0.1.0",
description='Easily use bleach with Django models and templates',
author='Tim Heap',
author_email='heap.tim@gmail.com',
url='https://bitbucket.org/ionata/django-bleach',
packages=['django_bleach',],
package_data={},
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
],
)
<commit_msg>Add bleach as a requirement<commit_after> | #!/usr/bin/env python
try:
from setuptools import setup, find_packages
except ImportError:
from ez_setup import use_setuptools
use_setuptools()
from setuptools import setup, find_packages
setup(
name='django-bleach',
version="0.1.0",
description='Easily use bleach with Django models and templates',
author='Tim Heap',
author_email='heap.tim@gmail.com',
url='https://bitbucket.org/ionata/django-bleach',
packages=['django_bleach',],
install_requires = ['bleach'],
package_data={},
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
],
)
| #!/usr/bin/env python
try:
from setuptools import setup, find_packages
except ImportError:
from ez_setup import use_setuptools
use_setuptools()
from setuptools import setup, find_packages
setup(
name='django-bleach',
version="0.1.0",
description='Easily use bleach with Django models and templates',
author='Tim Heap',
author_email='heap.tim@gmail.com',
url='https://bitbucket.org/ionata/django-bleach',
packages=['django_bleach',],
package_data={},
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
],
)
Add bleach as a requirement#!/usr/bin/env python
try:
from setuptools import setup, find_packages
except ImportError:
from ez_setup import use_setuptools
use_setuptools()
from setuptools import setup, find_packages
setup(
name='django-bleach',
version="0.1.0",
description='Easily use bleach with Django models and templates',
author='Tim Heap',
author_email='heap.tim@gmail.com',
url='https://bitbucket.org/ionata/django-bleach',
packages=['django_bleach',],
install_requires = ['bleach'],
package_data={},
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
],
)
| <commit_before>#!/usr/bin/env python
try:
from setuptools import setup, find_packages
except ImportError:
from ez_setup import use_setuptools
use_setuptools()
from setuptools import setup, find_packages
setup(
name='django-bleach',
version="0.1.0",
description='Easily use bleach with Django models and templates',
author='Tim Heap',
author_email='heap.tim@gmail.com',
url='https://bitbucket.org/ionata/django-bleach',
packages=['django_bleach',],
package_data={},
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
],
)
<commit_msg>Add bleach as a requirement<commit_after>#!/usr/bin/env python
try:
from setuptools import setup, find_packages
except ImportError:
from ez_setup import use_setuptools
use_setuptools()
from setuptools import setup, find_packages
setup(
name='django-bleach',
version="0.1.0",
description='Easily use bleach with Django models and templates',
author='Tim Heap',
author_email='heap.tim@gmail.com',
url='https://bitbucket.org/ionata/django-bleach',
packages=['django_bleach',],
install_requires = ['bleach'],
package_data={},
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
],
)
|
fa6880c9a1da67097fb495339d1b62b3bcda854d | setup.py | setup.py | from setuptools import setup, find_packages
setup(
name='exchangerates',
version='0.1.01',
description="A module to make it easier to handle historical exchange rates",
long_description="",
classifiers=[
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
'Programming Language :: Python :: 2.7'
],
author='Mark Brough',
author_email='mark@brough.io',
url='http://github.com/markbrough/exchangerates',
license='MIT',
packages=find_packages(exclude=['ez_setup', 'examples']),
namespace_packages=[],
include_package_data=True,
zip_safe=True,
install_requires=[
'lxml == 3.7.3',
'requests == 2.13.0'
],
entry_points={
}
)
| from setuptools import setup, find_packages
setup(
name='exchangerates',
version='0.1.2',
description="A module to make it easier to handle historical exchange rates",
long_description="",
classifiers=[
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
'Programming Language :: Python :: 2.7'
],
author='Mark Brough',
author_email='mark@brough.io',
url='http://github.com/markbrough/exchangerates',
license='MIT',
packages=find_packages(exclude=['ez_setup', 'examples']),
namespace_packages=[],
include_package_data=True,
zip_safe=True,
install_requires=[
'lxml == 3.7.3',
'requests == 2.13.0'
],
entry_points={
}
)
| Use a different version number | Use a different version number
| Python | mit | markbrough/exchangerates | from setuptools import setup, find_packages
setup(
name='exchangerates',
version='0.1.01',
description="A module to make it easier to handle historical exchange rates",
long_description="",
classifiers=[
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
'Programming Language :: Python :: 2.7'
],
author='Mark Brough',
author_email='mark@brough.io',
url='http://github.com/markbrough/exchangerates',
license='MIT',
packages=find_packages(exclude=['ez_setup', 'examples']),
namespace_packages=[],
include_package_data=True,
zip_safe=True,
install_requires=[
'lxml == 3.7.3',
'requests == 2.13.0'
],
entry_points={
}
)
Use a different version number | from setuptools import setup, find_packages
setup(
name='exchangerates',
version='0.1.2',
description="A module to make it easier to handle historical exchange rates",
long_description="",
classifiers=[
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
'Programming Language :: Python :: 2.7'
],
author='Mark Brough',
author_email='mark@brough.io',
url='http://github.com/markbrough/exchangerates',
license='MIT',
packages=find_packages(exclude=['ez_setup', 'examples']),
namespace_packages=[],
include_package_data=True,
zip_safe=True,
install_requires=[
'lxml == 3.7.3',
'requests == 2.13.0'
],
entry_points={
}
)
| <commit_before>from setuptools import setup, find_packages
setup(
name='exchangerates',
version='0.1.01',
description="A module to make it easier to handle historical exchange rates",
long_description="",
classifiers=[
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
'Programming Language :: Python :: 2.7'
],
author='Mark Brough',
author_email='mark@brough.io',
url='http://github.com/markbrough/exchangerates',
license='MIT',
packages=find_packages(exclude=['ez_setup', 'examples']),
namespace_packages=[],
include_package_data=True,
zip_safe=True,
install_requires=[
'lxml == 3.7.3',
'requests == 2.13.0'
],
entry_points={
}
)
<commit_msg>Use a different version number<commit_after> | from setuptools import setup, find_packages
setup(
name='exchangerates',
version='0.1.2',
description="A module to make it easier to handle historical exchange rates",
long_description="",
classifiers=[
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
'Programming Language :: Python :: 2.7'
],
author='Mark Brough',
author_email='mark@brough.io',
url='http://github.com/markbrough/exchangerates',
license='MIT',
packages=find_packages(exclude=['ez_setup', 'examples']),
namespace_packages=[],
include_package_data=True,
zip_safe=True,
install_requires=[
'lxml == 3.7.3',
'requests == 2.13.0'
],
entry_points={
}
)
| from setuptools import setup, find_packages
setup(
name='exchangerates',
version='0.1.01',
description="A module to make it easier to handle historical exchange rates",
long_description="",
classifiers=[
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
'Programming Language :: Python :: 2.7'
],
author='Mark Brough',
author_email='mark@brough.io',
url='http://github.com/markbrough/exchangerates',
license='MIT',
packages=find_packages(exclude=['ez_setup', 'examples']),
namespace_packages=[],
include_package_data=True,
zip_safe=True,
install_requires=[
'lxml == 3.7.3',
'requests == 2.13.0'
],
entry_points={
}
)
Use a different version numberfrom setuptools import setup, find_packages
setup(
name='exchangerates',
version='0.1.2',
description="A module to make it easier to handle historical exchange rates",
long_description="",
classifiers=[
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
'Programming Language :: Python :: 2.7'
],
author='Mark Brough',
author_email='mark@brough.io',
url='http://github.com/markbrough/exchangerates',
license='MIT',
packages=find_packages(exclude=['ez_setup', 'examples']),
namespace_packages=[],
include_package_data=True,
zip_safe=True,
install_requires=[
'lxml == 3.7.3',
'requests == 2.13.0'
],
entry_points={
}
)
| <commit_before>from setuptools import setup, find_packages
setup(
name='exchangerates',
version='0.1.01',
description="A module to make it easier to handle historical exchange rates",
long_description="",
classifiers=[
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
'Programming Language :: Python :: 2.7'
],
author='Mark Brough',
author_email='mark@brough.io',
url='http://github.com/markbrough/exchangerates',
license='MIT',
packages=find_packages(exclude=['ez_setup', 'examples']),
namespace_packages=[],
include_package_data=True,
zip_safe=True,
install_requires=[
'lxml == 3.7.3',
'requests == 2.13.0'
],
entry_points={
}
)
<commit_msg>Use a different version number<commit_after>from setuptools import setup, find_packages
setup(
name='exchangerates',
version='0.1.2',
description="A module to make it easier to handle historical exchange rates",
long_description="",
classifiers=[
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
'Programming Language :: Python :: 2.7'
],
author='Mark Brough',
author_email='mark@brough.io',
url='http://github.com/markbrough/exchangerates',
license='MIT',
packages=find_packages(exclude=['ez_setup', 'examples']),
namespace_packages=[],
include_package_data=True,
zip_safe=True,
install_requires=[
'lxml == 3.7.3',
'requests == 2.13.0'
],
entry_points={
}
)
|
a22bdc548940218c408069218c4351941c68d296 | setup.py | setup.py | """
Flask-InfluxDB
"""
from setuptools import setup
setup(
name="Flask-InfluxDB",
version="0.3",
url="http://github.com/btashton/flask-influxdb",
license="BSD",
author="Brennan Ashton",
author_email="brennan@ombitron.com",
description="Flask bindings for the InfluxDB time series database",
long_description=__doc__,
py_modules=["flask_influxdb"],
zip_safe=False,
include_package_data=True,
platforms="any",
install_requires=["Flask", "influxdb==5.2.2",],
classifiers=[
"Environment :: Web Environment",
"Intended Audience :: Developers",
"License :: OSI Approved :: BSD License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Topic :: Internet :: WWW/HTTP :: Dynamic Content",
"Topic :: Software Development :: Libraries :: Python Modules",
],
)
| """
Flask-InfluxDB
"""
from setuptools import setup
setup(
name="Flask-InfluxDB",
version="0.3.1",
url="http://github.com/btashton/flask-influxdb",
license="BSD",
author="Brennan Ashton",
author_email="brennan@ombitron.com",
description="Flask bindings for the InfluxDB time series database",
long_description=__doc__,
packages=["flask_influxdb"],
zip_safe=False,
include_package_data=True,
platforms="any",
install_requires=["Flask", "influxdb==5.2.2",],
classifiers=[
"Environment :: Web Environment",
"Intended Audience :: Developers",
"License :: OSI Approved :: BSD License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Topic :: Internet :: WWW/HTTP :: Dynamic Content",
"Topic :: Software Development :: Libraries :: Python Modules",
],
)
| Fix packaging issue in 0.3 release | Fix packaging issue in 0.3 release
Signed-off-by: Brennan Ashton <3c2365aa085787349a5327558db844b55eabde30@brennanashton.com>
| Python | bsd-3-clause | Ombitron/flask-influxdb,Ombitron/flask-influxdb | """
Flask-InfluxDB
"""
from setuptools import setup
setup(
name="Flask-InfluxDB",
version="0.3",
url="http://github.com/btashton/flask-influxdb",
license="BSD",
author="Brennan Ashton",
author_email="brennan@ombitron.com",
description="Flask bindings for the InfluxDB time series database",
long_description=__doc__,
py_modules=["flask_influxdb"],
zip_safe=False,
include_package_data=True,
platforms="any",
install_requires=["Flask", "influxdb==5.2.2",],
classifiers=[
"Environment :: Web Environment",
"Intended Audience :: Developers",
"License :: OSI Approved :: BSD License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Topic :: Internet :: WWW/HTTP :: Dynamic Content",
"Topic :: Software Development :: Libraries :: Python Modules",
],
)
Fix packaging issue in 0.3 release
Signed-off-by: Brennan Ashton <3c2365aa085787349a5327558db844b55eabde30@brennanashton.com> | """
Flask-InfluxDB
"""
from setuptools import setup
setup(
name="Flask-InfluxDB",
version="0.3.1",
url="http://github.com/btashton/flask-influxdb",
license="BSD",
author="Brennan Ashton",
author_email="brennan@ombitron.com",
description="Flask bindings for the InfluxDB time series database",
long_description=__doc__,
packages=["flask_influxdb"],
zip_safe=False,
include_package_data=True,
platforms="any",
install_requires=["Flask", "influxdb==5.2.2",],
classifiers=[
"Environment :: Web Environment",
"Intended Audience :: Developers",
"License :: OSI Approved :: BSD License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Topic :: Internet :: WWW/HTTP :: Dynamic Content",
"Topic :: Software Development :: Libraries :: Python Modules",
],
)
| <commit_before>"""
Flask-InfluxDB
"""
from setuptools import setup
setup(
name="Flask-InfluxDB",
version="0.3",
url="http://github.com/btashton/flask-influxdb",
license="BSD",
author="Brennan Ashton",
author_email="brennan@ombitron.com",
description="Flask bindings for the InfluxDB time series database",
long_description=__doc__,
py_modules=["flask_influxdb"],
zip_safe=False,
include_package_data=True,
platforms="any",
install_requires=["Flask", "influxdb==5.2.2",],
classifiers=[
"Environment :: Web Environment",
"Intended Audience :: Developers",
"License :: OSI Approved :: BSD License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Topic :: Internet :: WWW/HTTP :: Dynamic Content",
"Topic :: Software Development :: Libraries :: Python Modules",
],
)
<commit_msg>Fix packaging issue in 0.3 release
Signed-off-by: Brennan Ashton <3c2365aa085787349a5327558db844b55eabde30@brennanashton.com><commit_after> | """
Flask-InfluxDB
"""
from setuptools import setup
setup(
name="Flask-InfluxDB",
version="0.3.1",
url="http://github.com/btashton/flask-influxdb",
license="BSD",
author="Brennan Ashton",
author_email="brennan@ombitron.com",
description="Flask bindings for the InfluxDB time series database",
long_description=__doc__,
packages=["flask_influxdb"],
zip_safe=False,
include_package_data=True,
platforms="any",
install_requires=["Flask", "influxdb==5.2.2",],
classifiers=[
"Environment :: Web Environment",
"Intended Audience :: Developers",
"License :: OSI Approved :: BSD License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Topic :: Internet :: WWW/HTTP :: Dynamic Content",
"Topic :: Software Development :: Libraries :: Python Modules",
],
)
| """
Flask-InfluxDB
"""
from setuptools import setup
setup(
name="Flask-InfluxDB",
version="0.3",
url="http://github.com/btashton/flask-influxdb",
license="BSD",
author="Brennan Ashton",
author_email="brennan@ombitron.com",
description="Flask bindings for the InfluxDB time series database",
long_description=__doc__,
py_modules=["flask_influxdb"],
zip_safe=False,
include_package_data=True,
platforms="any",
install_requires=["Flask", "influxdb==5.2.2",],
classifiers=[
"Environment :: Web Environment",
"Intended Audience :: Developers",
"License :: OSI Approved :: BSD License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Topic :: Internet :: WWW/HTTP :: Dynamic Content",
"Topic :: Software Development :: Libraries :: Python Modules",
],
)
Fix packaging issue in 0.3 release
Signed-off-by: Brennan Ashton <3c2365aa085787349a5327558db844b55eabde30@brennanashton.com>"""
Flask-InfluxDB
"""
from setuptools import setup
setup(
name="Flask-InfluxDB",
version="0.3.1",
url="http://github.com/btashton/flask-influxdb",
license="BSD",
author="Brennan Ashton",
author_email="brennan@ombitron.com",
description="Flask bindings for the InfluxDB time series database",
long_description=__doc__,
packages=["flask_influxdb"],
zip_safe=False,
include_package_data=True,
platforms="any",
install_requires=["Flask", "influxdb==5.2.2",],
classifiers=[
"Environment :: Web Environment",
"Intended Audience :: Developers",
"License :: OSI Approved :: BSD License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Topic :: Internet :: WWW/HTTP :: Dynamic Content",
"Topic :: Software Development :: Libraries :: Python Modules",
],
)
| <commit_before>"""
Flask-InfluxDB
"""
from setuptools import setup
setup(
name="Flask-InfluxDB",
version="0.3",
url="http://github.com/btashton/flask-influxdb",
license="BSD",
author="Brennan Ashton",
author_email="brennan@ombitron.com",
description="Flask bindings for the InfluxDB time series database",
long_description=__doc__,
py_modules=["flask_influxdb"],
zip_safe=False,
include_package_data=True,
platforms="any",
install_requires=["Flask", "influxdb==5.2.2",],
classifiers=[
"Environment :: Web Environment",
"Intended Audience :: Developers",
"License :: OSI Approved :: BSD License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Topic :: Internet :: WWW/HTTP :: Dynamic Content",
"Topic :: Software Development :: Libraries :: Python Modules",
],
)
<commit_msg>Fix packaging issue in 0.3 release
Signed-off-by: Brennan Ashton <3c2365aa085787349a5327558db844b55eabde30@brennanashton.com><commit_after>"""
Flask-InfluxDB
"""
from setuptools import setup
setup(
name="Flask-InfluxDB",
version="0.3.1",
url="http://github.com/btashton/flask-influxdb",
license="BSD",
author="Brennan Ashton",
author_email="brennan@ombitron.com",
description="Flask bindings for the InfluxDB time series database",
long_description=__doc__,
packages=["flask_influxdb"],
zip_safe=False,
include_package_data=True,
platforms="any",
install_requires=["Flask", "influxdb==5.2.2",],
classifiers=[
"Environment :: Web Environment",
"Intended Audience :: Developers",
"License :: OSI Approved :: BSD License",
"Operating System :: OS Independent",
"Programming Language :: Python",
"Topic :: Internet :: WWW/HTTP :: Dynamic Content",
"Topic :: Software Development :: Libraries :: Python Modules",
],
)
|
56a7f8aac203d2d0d685c0472d74090dc2e4da0c | setup.py | setup.py | # -*- coding: utf-8 -*-
from setuptools import setup, find_packages
import sys, os
version = '0.1.9.dev'
setup(name='testdroid',
version=version,
description="Testdroid API client for Python",
long_description="""\
Testdroid API client for Python""",
classifiers=['Operating System :: OS Independent',
'Topic :: Software Development',
'Intended Audience :: Developers'], # Get strings from http://pypi.python.org/pypi?%3Aaction=list_classifiers
keywords='testdroid rest api client',
author='Henri Kivelä <henri.kivela@bitbar.com>, Sakari Rautiainen <sakari.rautiainen@bitbar.com>, Teppo Malinen <teppo.malinen@bitbar.com, Jarno Tuovinen <jarno.tuovinen@bitbar.com>',
author_email='info@bitbar.com',
url='http://www.testdroid.com',
license='Apache License v2.0',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
include_package_data=True,
zip_safe=True,
install_requires=[
'requests', 'pillow',
],
entry_points = {
'console_scripts': [
'testdroid = testdroid:main',
],
},
)
| # -*- coding: utf-8 -*-
from setuptools import setup, find_packages
import sys, os
version = '0.1.9.dev'
setup(name='testdroid',
version=version,
description="Testdroid API client for Python",
long_description="""\
Testdroid API client for Python""",
classifiers=['Operating System :: OS Independent',
'Topic :: Software Development',
'Intended Audience :: Developers'], # Get strings from http://pypi.python.org/pypi?%3Aaction=list_classifiers
keywords='testdroid rest api client',
author='Henri Kivelä <henri.kivela@bitbar.com>, Sakari Rautiainen <sakari.rautiainen@bitbar.com>, Teppo Malinen <teppo.malinen@bitbar.com>, Jarno Tuovinen <jarno.tuovinen@bitbar.com>',
author_email='info@bitbar.com',
url='http://www.testdroid.com',
license='Apache License v2.0',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
include_package_data=True,
zip_safe=True,
install_requires=[
'requests', 'pillow',
],
entry_points = {
'console_scripts': [
'testdroid = testdroid:main',
],
},
)
| Fix typo from authors field | Fix typo from authors field
| Python | apache-2.0 | aknackiron/testdroid-api-client-python,bitbar/testdroid-api-client-python,teppomalinen/testdroid-api-client-python | # -*- coding: utf-8 -*-
from setuptools import setup, find_packages
import sys, os
version = '0.1.9.dev'
setup(name='testdroid',
version=version,
description="Testdroid API client for Python",
long_description="""\
Testdroid API client for Python""",
classifiers=['Operating System :: OS Independent',
'Topic :: Software Development',
'Intended Audience :: Developers'], # Get strings from http://pypi.python.org/pypi?%3Aaction=list_classifiers
keywords='testdroid rest api client',
author='Henri Kivelä <henri.kivela@bitbar.com>, Sakari Rautiainen <sakari.rautiainen@bitbar.com>, Teppo Malinen <teppo.malinen@bitbar.com, Jarno Tuovinen <jarno.tuovinen@bitbar.com>',
author_email='info@bitbar.com',
url='http://www.testdroid.com',
license='Apache License v2.0',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
include_package_data=True,
zip_safe=True,
install_requires=[
'requests', 'pillow',
],
entry_points = {
'console_scripts': [
'testdroid = testdroid:main',
],
},
)
Fix typo from authors field | # -*- coding: utf-8 -*-
from setuptools import setup, find_packages
import sys, os
version = '0.1.9.dev'
setup(name='testdroid',
version=version,
description="Testdroid API client for Python",
long_description="""\
Testdroid API client for Python""",
classifiers=['Operating System :: OS Independent',
'Topic :: Software Development',
'Intended Audience :: Developers'], # Get strings from http://pypi.python.org/pypi?%3Aaction=list_classifiers
keywords='testdroid rest api client',
author='Henri Kivelä <henri.kivela@bitbar.com>, Sakari Rautiainen <sakari.rautiainen@bitbar.com>, Teppo Malinen <teppo.malinen@bitbar.com>, Jarno Tuovinen <jarno.tuovinen@bitbar.com>',
author_email='info@bitbar.com',
url='http://www.testdroid.com',
license='Apache License v2.0',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
include_package_data=True,
zip_safe=True,
install_requires=[
'requests', 'pillow',
],
entry_points = {
'console_scripts': [
'testdroid = testdroid:main',
],
},
)
| <commit_before># -*- coding: utf-8 -*-
from setuptools import setup, find_packages
import sys, os
version = '0.1.9.dev'
setup(name='testdroid',
version=version,
description="Testdroid API client for Python",
long_description="""\
Testdroid API client for Python""",
classifiers=['Operating System :: OS Independent',
'Topic :: Software Development',
'Intended Audience :: Developers'], # Get strings from http://pypi.python.org/pypi?%3Aaction=list_classifiers
keywords='testdroid rest api client',
author='Henri Kivelä <henri.kivela@bitbar.com>, Sakari Rautiainen <sakari.rautiainen@bitbar.com>, Teppo Malinen <teppo.malinen@bitbar.com, Jarno Tuovinen <jarno.tuovinen@bitbar.com>',
author_email='info@bitbar.com',
url='http://www.testdroid.com',
license='Apache License v2.0',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
include_package_data=True,
zip_safe=True,
install_requires=[
'requests', 'pillow',
],
entry_points = {
'console_scripts': [
'testdroid = testdroid:main',
],
},
)
<commit_msg>Fix typo from authors field<commit_after> | # -*- coding: utf-8 -*-
from setuptools import setup, find_packages
import sys, os
version = '0.1.9.dev'
setup(name='testdroid',
version=version,
description="Testdroid API client for Python",
long_description="""\
Testdroid API client for Python""",
classifiers=['Operating System :: OS Independent',
'Topic :: Software Development',
'Intended Audience :: Developers'], # Get strings from http://pypi.python.org/pypi?%3Aaction=list_classifiers
keywords='testdroid rest api client',
author='Henri Kivelä <henri.kivela@bitbar.com>, Sakari Rautiainen <sakari.rautiainen@bitbar.com>, Teppo Malinen <teppo.malinen@bitbar.com>, Jarno Tuovinen <jarno.tuovinen@bitbar.com>',
author_email='info@bitbar.com',
url='http://www.testdroid.com',
license='Apache License v2.0',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
include_package_data=True,
zip_safe=True,
install_requires=[
'requests', 'pillow',
],
entry_points = {
'console_scripts': [
'testdroid = testdroid:main',
],
},
)
| # -*- coding: utf-8 -*-
from setuptools import setup, find_packages
import sys, os
version = '0.1.9.dev'
setup(name='testdroid',
version=version,
description="Testdroid API client for Python",
long_description="""\
Testdroid API client for Python""",
classifiers=['Operating System :: OS Independent',
'Topic :: Software Development',
'Intended Audience :: Developers'], # Get strings from http://pypi.python.org/pypi?%3Aaction=list_classifiers
keywords='testdroid rest api client',
author='Henri Kivelä <henri.kivela@bitbar.com>, Sakari Rautiainen <sakari.rautiainen@bitbar.com>, Teppo Malinen <teppo.malinen@bitbar.com, Jarno Tuovinen <jarno.tuovinen@bitbar.com>',
author_email='info@bitbar.com',
url='http://www.testdroid.com',
license='Apache License v2.0',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
include_package_data=True,
zip_safe=True,
install_requires=[
'requests', 'pillow',
],
entry_points = {
'console_scripts': [
'testdroid = testdroid:main',
],
},
)
Fix typo from authors field# -*- coding: utf-8 -*-
from setuptools import setup, find_packages
import sys, os
version = '0.1.9.dev'
setup(name='testdroid',
version=version,
description="Testdroid API client for Python",
long_description="""\
Testdroid API client for Python""",
classifiers=['Operating System :: OS Independent',
'Topic :: Software Development',
'Intended Audience :: Developers'], # Get strings from http://pypi.python.org/pypi?%3Aaction=list_classifiers
keywords='testdroid rest api client',
author='Henri Kivelä <henri.kivela@bitbar.com>, Sakari Rautiainen <sakari.rautiainen@bitbar.com>, Teppo Malinen <teppo.malinen@bitbar.com>, Jarno Tuovinen <jarno.tuovinen@bitbar.com>',
author_email='info@bitbar.com',
url='http://www.testdroid.com',
license='Apache License v2.0',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
include_package_data=True,
zip_safe=True,
install_requires=[
'requests', 'pillow',
],
entry_points = {
'console_scripts': [
'testdroid = testdroid:main',
],
},
)
| <commit_before># -*- coding: utf-8 -*-
from setuptools import setup, find_packages
import sys, os
version = '0.1.9.dev'
setup(name='testdroid',
version=version,
description="Testdroid API client for Python",
long_description="""\
Testdroid API client for Python""",
classifiers=['Operating System :: OS Independent',
'Topic :: Software Development',
'Intended Audience :: Developers'], # Get strings from http://pypi.python.org/pypi?%3Aaction=list_classifiers
keywords='testdroid rest api client',
author='Henri Kivelä <henri.kivela@bitbar.com>, Sakari Rautiainen <sakari.rautiainen@bitbar.com>, Teppo Malinen <teppo.malinen@bitbar.com, Jarno Tuovinen <jarno.tuovinen@bitbar.com>',
author_email='info@bitbar.com',
url='http://www.testdroid.com',
license='Apache License v2.0',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
include_package_data=True,
zip_safe=True,
install_requires=[
'requests', 'pillow',
],
entry_points = {
'console_scripts': [
'testdroid = testdroid:main',
],
},
)
<commit_msg>Fix typo from authors field<commit_after># -*- coding: utf-8 -*-
from setuptools import setup, find_packages
import sys, os
version = '0.1.9.dev'
setup(name='testdroid',
version=version,
description="Testdroid API client for Python",
long_description="""\
Testdroid API client for Python""",
classifiers=['Operating System :: OS Independent',
'Topic :: Software Development',
'Intended Audience :: Developers'], # Get strings from http://pypi.python.org/pypi?%3Aaction=list_classifiers
keywords='testdroid rest api client',
author='Henri Kivelä <henri.kivela@bitbar.com>, Sakari Rautiainen <sakari.rautiainen@bitbar.com>, Teppo Malinen <teppo.malinen@bitbar.com>, Jarno Tuovinen <jarno.tuovinen@bitbar.com>',
author_email='info@bitbar.com',
url='http://www.testdroid.com',
license='Apache License v2.0',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
include_package_data=True,
zip_safe=True,
install_requires=[
'requests', 'pillow',
],
entry_points = {
'console_scripts': [
'testdroid = testdroid:main',
],
},
)
|
cccca9c2feba5cbcb439f6d829e1e930819cb9c1 | setup.py | setup.py | #!/usr/bin/env python
from setuptools import setup, find_packages
tests_require = []
setup(
name='ashlar',
version='0.0.2',
description='Define and validate schemas for metadata for geotemporal event records',
author='Azavea, Inc.',
author_email='info@azavea.com',
keywords='gis jsonschema',
packages=find_packages(exclude=['tests']),
dependency_links=[
'https://github.com/azavea/djsonb/tarball/develop#egg=djsonb-0.1.4'
],
install_requires=[
'Django ==1.8.6',
'djangorestframework >=3.1.1',
'djangorestframework-gis >=0.8.1',
'django-filter >=0.9.2',
'djsonb >=0.1.4',
'jsonschema >=2.4.0',
'psycopg2 >=2.6',
'django-extensions >=1.5.2',
'python-dateutil >=2.4.2',
'PyYAML >=3.11'
],
extras_require={
'dev': [],
'test': tests_require
},
test_suite='tests',
tests_require=tests_require,
)
| #!/usr/bin/env python
from setuptools import setup, find_packages
tests_require = []
setup(
name='ashlar',
version='0.0.2',
description='Define and validate schemas for metadata for geotemporal event records',
author='Azavea, Inc.',
author_email='info@azavea.com',
keywords='gis jsonschema',
packages=find_packages(exclude=['tests']),
dependency_links=[
'https://github.com/azavea/djsonb/tarball/develop#egg=djsonb-0.1.5'
],
install_requires=[
'Django ==1.8.6',
'djangorestframework >=3.1.1',
'djangorestframework-gis >=0.8.1',
'django-filter >=0.9.2',
'djsonb >=0.1.5',
'jsonschema >=2.4.0',
'psycopg2 >=2.6',
'django-extensions >=1.5.2',
'python-dateutil >=2.4.2',
'PyYAML >=3.11',
'pytz >= 2015.7'
],
extras_require={
'dev': [],
'test': tests_require
},
test_suite='tests',
tests_require=tests_require,
)
| Update djsonb, and add pytz | Update djsonb, and add pytz
| Python | mit | azavea/ashlar,flibbertigibbet/ashlar,azavea/ashlar,flibbertigibbet/ashlar | #!/usr/bin/env python
from setuptools import setup, find_packages
tests_require = []
setup(
name='ashlar',
version='0.0.2',
description='Define and validate schemas for metadata for geotemporal event records',
author='Azavea, Inc.',
author_email='info@azavea.com',
keywords='gis jsonschema',
packages=find_packages(exclude=['tests']),
dependency_links=[
'https://github.com/azavea/djsonb/tarball/develop#egg=djsonb-0.1.4'
],
install_requires=[
'Django ==1.8.6',
'djangorestframework >=3.1.1',
'djangorestframework-gis >=0.8.1',
'django-filter >=0.9.2',
'djsonb >=0.1.4',
'jsonschema >=2.4.0',
'psycopg2 >=2.6',
'django-extensions >=1.5.2',
'python-dateutil >=2.4.2',
'PyYAML >=3.11'
],
extras_require={
'dev': [],
'test': tests_require
},
test_suite='tests',
tests_require=tests_require,
)
Update djsonb, and add pytz | #!/usr/bin/env python
from setuptools import setup, find_packages
tests_require = []
setup(
name='ashlar',
version='0.0.2',
description='Define and validate schemas for metadata for geotemporal event records',
author='Azavea, Inc.',
author_email='info@azavea.com',
keywords='gis jsonschema',
packages=find_packages(exclude=['tests']),
dependency_links=[
'https://github.com/azavea/djsonb/tarball/develop#egg=djsonb-0.1.5'
],
install_requires=[
'Django ==1.8.6',
'djangorestframework >=3.1.1',
'djangorestframework-gis >=0.8.1',
'django-filter >=0.9.2',
'djsonb >=0.1.5',
'jsonschema >=2.4.0',
'psycopg2 >=2.6',
'django-extensions >=1.5.2',
'python-dateutil >=2.4.2',
'PyYAML >=3.11',
'pytz >= 2015.7'
],
extras_require={
'dev': [],
'test': tests_require
},
test_suite='tests',
tests_require=tests_require,
)
| <commit_before>#!/usr/bin/env python
from setuptools import setup, find_packages
tests_require = []
setup(
name='ashlar',
version='0.0.2',
description='Define and validate schemas for metadata for geotemporal event records',
author='Azavea, Inc.',
author_email='info@azavea.com',
keywords='gis jsonschema',
packages=find_packages(exclude=['tests']),
dependency_links=[
'https://github.com/azavea/djsonb/tarball/develop#egg=djsonb-0.1.4'
],
install_requires=[
'Django ==1.8.6',
'djangorestframework >=3.1.1',
'djangorestframework-gis >=0.8.1',
'django-filter >=0.9.2',
'djsonb >=0.1.4',
'jsonschema >=2.4.0',
'psycopg2 >=2.6',
'django-extensions >=1.5.2',
'python-dateutil >=2.4.2',
'PyYAML >=3.11'
],
extras_require={
'dev': [],
'test': tests_require
},
test_suite='tests',
tests_require=tests_require,
)
<commit_msg>Update djsonb, and add pytz<commit_after> | #!/usr/bin/env python
from setuptools import setup, find_packages
tests_require = []
setup(
name='ashlar',
version='0.0.2',
description='Define and validate schemas for metadata for geotemporal event records',
author='Azavea, Inc.',
author_email='info@azavea.com',
keywords='gis jsonschema',
packages=find_packages(exclude=['tests']),
dependency_links=[
'https://github.com/azavea/djsonb/tarball/develop#egg=djsonb-0.1.5'
],
install_requires=[
'Django ==1.8.6',
'djangorestframework >=3.1.1',
'djangorestframework-gis >=0.8.1',
'django-filter >=0.9.2',
'djsonb >=0.1.5',
'jsonschema >=2.4.0',
'psycopg2 >=2.6',
'django-extensions >=1.5.2',
'python-dateutil >=2.4.2',
'PyYAML >=3.11',
'pytz >= 2015.7'
],
extras_require={
'dev': [],
'test': tests_require
},
test_suite='tests',
tests_require=tests_require,
)
| #!/usr/bin/env python
from setuptools import setup, find_packages
tests_require = []
setup(
name='ashlar',
version='0.0.2',
description='Define and validate schemas for metadata for geotemporal event records',
author='Azavea, Inc.',
author_email='info@azavea.com',
keywords='gis jsonschema',
packages=find_packages(exclude=['tests']),
dependency_links=[
'https://github.com/azavea/djsonb/tarball/develop#egg=djsonb-0.1.4'
],
install_requires=[
'Django ==1.8.6',
'djangorestframework >=3.1.1',
'djangorestframework-gis >=0.8.1',
'django-filter >=0.9.2',
'djsonb >=0.1.4',
'jsonschema >=2.4.0',
'psycopg2 >=2.6',
'django-extensions >=1.5.2',
'python-dateutil >=2.4.2',
'PyYAML >=3.11'
],
extras_require={
'dev': [],
'test': tests_require
},
test_suite='tests',
tests_require=tests_require,
)
Update djsonb, and add pytz#!/usr/bin/env python
from setuptools import setup, find_packages
tests_require = []
setup(
name='ashlar',
version='0.0.2',
description='Define and validate schemas for metadata for geotemporal event records',
author='Azavea, Inc.',
author_email='info@azavea.com',
keywords='gis jsonschema',
packages=find_packages(exclude=['tests']),
dependency_links=[
'https://github.com/azavea/djsonb/tarball/develop#egg=djsonb-0.1.5'
],
install_requires=[
'Django ==1.8.6',
'djangorestframework >=3.1.1',
'djangorestframework-gis >=0.8.1',
'django-filter >=0.9.2',
'djsonb >=0.1.5',
'jsonschema >=2.4.0',
'psycopg2 >=2.6',
'django-extensions >=1.5.2',
'python-dateutil >=2.4.2',
'PyYAML >=3.11',
'pytz >= 2015.7'
],
extras_require={
'dev': [],
'test': tests_require
},
test_suite='tests',
tests_require=tests_require,
)
| <commit_before>#!/usr/bin/env python
from setuptools import setup, find_packages
tests_require = []
setup(
name='ashlar',
version='0.0.2',
description='Define and validate schemas for metadata for geotemporal event records',
author='Azavea, Inc.',
author_email='info@azavea.com',
keywords='gis jsonschema',
packages=find_packages(exclude=['tests']),
dependency_links=[
'https://github.com/azavea/djsonb/tarball/develop#egg=djsonb-0.1.4'
],
install_requires=[
'Django ==1.8.6',
'djangorestframework >=3.1.1',
'djangorestframework-gis >=0.8.1',
'django-filter >=0.9.2',
'djsonb >=0.1.4',
'jsonschema >=2.4.0',
'psycopg2 >=2.6',
'django-extensions >=1.5.2',
'python-dateutil >=2.4.2',
'PyYAML >=3.11'
],
extras_require={
'dev': [],
'test': tests_require
},
test_suite='tests',
tests_require=tests_require,
)
<commit_msg>Update djsonb, and add pytz<commit_after>#!/usr/bin/env python
from setuptools import setup, find_packages
tests_require = []
setup(
name='ashlar',
version='0.0.2',
description='Define and validate schemas for metadata for geotemporal event records',
author='Azavea, Inc.',
author_email='info@azavea.com',
keywords='gis jsonschema',
packages=find_packages(exclude=['tests']),
dependency_links=[
'https://github.com/azavea/djsonb/tarball/develop#egg=djsonb-0.1.5'
],
install_requires=[
'Django ==1.8.6',
'djangorestframework >=3.1.1',
'djangorestframework-gis >=0.8.1',
'django-filter >=0.9.2',
'djsonb >=0.1.5',
'jsonschema >=2.4.0',
'psycopg2 >=2.6',
'django-extensions >=1.5.2',
'python-dateutil >=2.4.2',
'PyYAML >=3.11',
'pytz >= 2015.7'
],
extras_require={
'dev': [],
'test': tests_require
},
test_suite='tests',
tests_require=tests_require,
)
|
3c8cea52ba0b4d6aadf34f1323cb54bf0238f394 | setup.py | setup.py | #!/usr/bin/env python
from setuptools import setup, find_packages
import re
# get version from __init__.py
INITFILE = "toytree/__init__.py"
CUR_VERSION = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]",
open(INITFILE, "r").read(),
re.M).group(1)
# run setup
setup(
name="toytree",
version=CUR_VERSION,
url="https://github.com/eaton-lab/toytree",
author="Deren Eaton",
author_email="de2356@columbia.edu",
description="minimalist tree plotting using toyplot",
long_description=open('README.rst').read(),
long_description_content_type='text/x-rst',
packages=find_packages(),
install_requires=[
"toyplot",
"numpy",
"requests",
"future",
],
entry_points={},
license='GPL',
classifiers=[
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
],
)
| #!/usr/bin/env python
from setuptools import setup, find_packages
import re
# get version from __init__.py
INITFILE = "toytree/__init__.py"
CUR_VERSION = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]",
open(INITFILE, "r").read(),
re.M).group(1)
# run setup
setup(
name="toytree",
version=CUR_VERSION,
url="https://github.com/eaton-lab/toytree",
author="Deren Eaton",
author_email="de2356@columbia.edu",
description="minimalist tree plotting using toyplot",
long_description=open('README.md').read(),
long_description_content_type='text/x-rst',
packages=find_packages(),
install_requires=[
"toyplot",
"numpy",
"requests",
"future",
],
entry_points={},
license='GPL',
classifiers=[
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
],
)
| Change readme to md to match. | Change readme to md to match. | Python | bsd-3-clause | eaton-lab/toytree | #!/usr/bin/env python
from setuptools import setup, find_packages
import re
# get version from __init__.py
INITFILE = "toytree/__init__.py"
CUR_VERSION = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]",
open(INITFILE, "r").read(),
re.M).group(1)
# run setup
setup(
name="toytree",
version=CUR_VERSION,
url="https://github.com/eaton-lab/toytree",
author="Deren Eaton",
author_email="de2356@columbia.edu",
description="minimalist tree plotting using toyplot",
long_description=open('README.rst').read(),
long_description_content_type='text/x-rst',
packages=find_packages(),
install_requires=[
"toyplot",
"numpy",
"requests",
"future",
],
entry_points={},
license='GPL',
classifiers=[
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
],
)
Change readme to md to match. | #!/usr/bin/env python
from setuptools import setup, find_packages
import re
# get version from __init__.py
INITFILE = "toytree/__init__.py"
CUR_VERSION = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]",
open(INITFILE, "r").read(),
re.M).group(1)
# run setup
setup(
name="toytree",
version=CUR_VERSION,
url="https://github.com/eaton-lab/toytree",
author="Deren Eaton",
author_email="de2356@columbia.edu",
description="minimalist tree plotting using toyplot",
long_description=open('README.md').read(),
long_description_content_type='text/x-rst',
packages=find_packages(),
install_requires=[
"toyplot",
"numpy",
"requests",
"future",
],
entry_points={},
license='GPL',
classifiers=[
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
],
)
| <commit_before>#!/usr/bin/env python
from setuptools import setup, find_packages
import re
# get version from __init__.py
INITFILE = "toytree/__init__.py"
CUR_VERSION = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]",
open(INITFILE, "r").read(),
re.M).group(1)
# run setup
setup(
name="toytree",
version=CUR_VERSION,
url="https://github.com/eaton-lab/toytree",
author="Deren Eaton",
author_email="de2356@columbia.edu",
description="minimalist tree plotting using toyplot",
long_description=open('README.rst').read(),
long_description_content_type='text/x-rst',
packages=find_packages(),
install_requires=[
"toyplot",
"numpy",
"requests",
"future",
],
entry_points={},
license='GPL',
classifiers=[
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
],
)
<commit_msg>Change readme to md to match.<commit_after> | #!/usr/bin/env python
from setuptools import setup, find_packages
import re
# get version from __init__.py
INITFILE = "toytree/__init__.py"
CUR_VERSION = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]",
open(INITFILE, "r").read(),
re.M).group(1)
# run setup
setup(
name="toytree",
version=CUR_VERSION,
url="https://github.com/eaton-lab/toytree",
author="Deren Eaton",
author_email="de2356@columbia.edu",
description="minimalist tree plotting using toyplot",
long_description=open('README.md').read(),
long_description_content_type='text/x-rst',
packages=find_packages(),
install_requires=[
"toyplot",
"numpy",
"requests",
"future",
],
entry_points={},
license='GPL',
classifiers=[
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
],
)
| #!/usr/bin/env python
from setuptools import setup, find_packages
import re
# get version from __init__.py
INITFILE = "toytree/__init__.py"
CUR_VERSION = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]",
open(INITFILE, "r").read(),
re.M).group(1)
# run setup
setup(
name="toytree",
version=CUR_VERSION,
url="https://github.com/eaton-lab/toytree",
author="Deren Eaton",
author_email="de2356@columbia.edu",
description="minimalist tree plotting using toyplot",
long_description=open('README.rst').read(),
long_description_content_type='text/x-rst',
packages=find_packages(),
install_requires=[
"toyplot",
"numpy",
"requests",
"future",
],
entry_points={},
license='GPL',
classifiers=[
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
],
)
Change readme to md to match.#!/usr/bin/env python
from setuptools import setup, find_packages
import re
# get version from __init__.py
INITFILE = "toytree/__init__.py"
CUR_VERSION = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]",
open(INITFILE, "r").read(),
re.M).group(1)
# run setup
setup(
name="toytree",
version=CUR_VERSION,
url="https://github.com/eaton-lab/toytree",
author="Deren Eaton",
author_email="de2356@columbia.edu",
description="minimalist tree plotting using toyplot",
long_description=open('README.md').read(),
long_description_content_type='text/x-rst',
packages=find_packages(),
install_requires=[
"toyplot",
"numpy",
"requests",
"future",
],
entry_points={},
license='GPL',
classifiers=[
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
],
)
| <commit_before>#!/usr/bin/env python
from setuptools import setup, find_packages
import re
# get version from __init__.py
INITFILE = "toytree/__init__.py"
CUR_VERSION = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]",
open(INITFILE, "r").read(),
re.M).group(1)
# run setup
setup(
name="toytree",
version=CUR_VERSION,
url="https://github.com/eaton-lab/toytree",
author="Deren Eaton",
author_email="de2356@columbia.edu",
description="minimalist tree plotting using toyplot",
long_description=open('README.rst').read(),
long_description_content_type='text/x-rst',
packages=find_packages(),
install_requires=[
"toyplot",
"numpy",
"requests",
"future",
],
entry_points={},
license='GPL',
classifiers=[
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
],
)
<commit_msg>Change readme to md to match.<commit_after>#!/usr/bin/env python
from setuptools import setup, find_packages
import re
# get version from __init__.py
INITFILE = "toytree/__init__.py"
CUR_VERSION = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]",
open(INITFILE, "r").read(),
re.M).group(1)
# run setup
setup(
name="toytree",
version=CUR_VERSION,
url="https://github.com/eaton-lab/toytree",
author="Deren Eaton",
author_email="de2356@columbia.edu",
description="minimalist tree plotting using toyplot",
long_description=open('README.md').read(),
long_description_content_type='text/x-rst',
packages=find_packages(),
install_requires=[
"toyplot",
"numpy",
"requests",
"future",
],
entry_points={},
license='GPL',
classifiers=[
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
],
)
|
9cfa095f5f2aa0a62afc0572eb3605e81d607e10 | setup.py | setup.py | from setuptools import setup
from kafka_info import __version__
setup(
name="kafka_info",
version=__version__,
author="Federico Giraud",
author_email="fgiraud@yelp.com",
description="Shows kafka cluster information and metrics",
packages=["kafka_info", "kafka_info.utils"],
data_files=[("bash_completion.d", ["bash_completion.d/kafka-info"])],
scripts=["kafka-info"],
install_requires=[
"argparse",
"argcomplete",
"kazoo",
"PyYAML",
],
)
| from setuptools import setup
from kafka_info import __version__
setup(
name="kafka_info",
version=__version__,
author="Federico Giraud",
author_email="fgiraud@yelp.com",
description="Shows kafka cluster information and metrics",
packages=["kafka_info", "kafka_info.utils", "kafka_info.commands"],
data_files=[("bash_completion.d", ["bash_completion.d/kafka-info"])],
scripts=["kafka-info"],
install_requires=[
"argparse",
"argcomplete",
"kazoo",
"PyYAML",
],
)
| Fix package, bump to 0.1.3 | Fix package, bump to 0.1.3
| Python | apache-2.0 | anthonysandrin/kafka-utils,Yelp/kafka-utils,anthonysandrin/kafka-utils,Yelp/kafka-utils | from setuptools import setup
from kafka_info import __version__
setup(
name="kafka_info",
version=__version__,
author="Federico Giraud",
author_email="fgiraud@yelp.com",
description="Shows kafka cluster information and metrics",
packages=["kafka_info", "kafka_info.utils"],
data_files=[("bash_completion.d", ["bash_completion.d/kafka-info"])],
scripts=["kafka-info"],
install_requires=[
"argparse",
"argcomplete",
"kazoo",
"PyYAML",
],
)
Fix package, bump to 0.1.3 | from setuptools import setup
from kafka_info import __version__
setup(
name="kafka_info",
version=__version__,
author="Federico Giraud",
author_email="fgiraud@yelp.com",
description="Shows kafka cluster information and metrics",
packages=["kafka_info", "kafka_info.utils", "kafka_info.commands"],
data_files=[("bash_completion.d", ["bash_completion.d/kafka-info"])],
scripts=["kafka-info"],
install_requires=[
"argparse",
"argcomplete",
"kazoo",
"PyYAML",
],
)
| <commit_before>from setuptools import setup
from kafka_info import __version__
setup(
name="kafka_info",
version=__version__,
author="Federico Giraud",
author_email="fgiraud@yelp.com",
description="Shows kafka cluster information and metrics",
packages=["kafka_info", "kafka_info.utils"],
data_files=[("bash_completion.d", ["bash_completion.d/kafka-info"])],
scripts=["kafka-info"],
install_requires=[
"argparse",
"argcomplete",
"kazoo",
"PyYAML",
],
)
<commit_msg>Fix package, bump to 0.1.3<commit_after> | from setuptools import setup
from kafka_info import __version__
setup(
name="kafka_info",
version=__version__,
author="Federico Giraud",
author_email="fgiraud@yelp.com",
description="Shows kafka cluster information and metrics",
packages=["kafka_info", "kafka_info.utils", "kafka_info.commands"],
data_files=[("bash_completion.d", ["bash_completion.d/kafka-info"])],
scripts=["kafka-info"],
install_requires=[
"argparse",
"argcomplete",
"kazoo",
"PyYAML",
],
)
| from setuptools import setup
from kafka_info import __version__
setup(
name="kafka_info",
version=__version__,
author="Federico Giraud",
author_email="fgiraud@yelp.com",
description="Shows kafka cluster information and metrics",
packages=["kafka_info", "kafka_info.utils"],
data_files=[("bash_completion.d", ["bash_completion.d/kafka-info"])],
scripts=["kafka-info"],
install_requires=[
"argparse",
"argcomplete",
"kazoo",
"PyYAML",
],
)
Fix package, bump to 0.1.3from setuptools import setup
from kafka_info import __version__
setup(
name="kafka_info",
version=__version__,
author="Federico Giraud",
author_email="fgiraud@yelp.com",
description="Shows kafka cluster information and metrics",
packages=["kafka_info", "kafka_info.utils", "kafka_info.commands"],
data_files=[("bash_completion.d", ["bash_completion.d/kafka-info"])],
scripts=["kafka-info"],
install_requires=[
"argparse",
"argcomplete",
"kazoo",
"PyYAML",
],
)
| <commit_before>from setuptools import setup
from kafka_info import __version__
setup(
name="kafka_info",
version=__version__,
author="Federico Giraud",
author_email="fgiraud@yelp.com",
description="Shows kafka cluster information and metrics",
packages=["kafka_info", "kafka_info.utils"],
data_files=[("bash_completion.d", ["bash_completion.d/kafka-info"])],
scripts=["kafka-info"],
install_requires=[
"argparse",
"argcomplete",
"kazoo",
"PyYAML",
],
)
<commit_msg>Fix package, bump to 0.1.3<commit_after>from setuptools import setup
from kafka_info import __version__
setup(
name="kafka_info",
version=__version__,
author="Federico Giraud",
author_email="fgiraud@yelp.com",
description="Shows kafka cluster information and metrics",
packages=["kafka_info", "kafka_info.utils", "kafka_info.commands"],
data_files=[("bash_completion.d", ["bash_completion.d/kafka-info"])],
scripts=["kafka-info"],
install_requires=[
"argparse",
"argcomplete",
"kazoo",
"PyYAML",
],
)
|
5a7bf12879c637f72c78d5f0a3e45915dd08711a | setup.py | setup.py | #!/usr/bin/env python
import os
import sys
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
if sys.argv[-1] == 'publish':
os.system('python setup.py sdist bdist_wininst upload -r pypi')
sys.exit()
with open('README.rst') as f:
readme = f.read()
with open('LICENSE') as f:
license = f.read()
setup(
name='django-rest-surveys',
version='0.1.0',
description='A RESTful backend for giving surveys.',
long_description=readme,
author='Designlab',
author_email='hello@trydesignlab.com',
url='https://github.com/danxshap/django-rest-surveys',
packages=['rest_surveys'],
package_data={'': ['LICENSE']},
package_dir={'rest_surveys': 'rest_surveys'},
install_requires=[
'Django>=1.7',
'djangorestframework>=3.0',
'django-inline-ordering',
'djangorestframework-bulk',
],
license=license,
)
| #!/usr/bin/env python
import os
import sys
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
if sys.argv[-1] == 'publish':
os.system('python setup.py sdist bdist_wininst upload -r pypi')
sys.exit()
with open('README.rst') as f:
readme = f.read()
with open('LICENSE') as f:
license = f.read()
setup(
name='django-rest-surveys',
version='0.1.0',
description='A RESTful backend for giving surveys.',
long_description=readme,
author='Designlab',
author_email='hello@trydesignlab.com',
url='https://github.com/danxshap/django-rest-surveys',
packages=['rest_surveys'],
package_data={'': ['LICENSE']},
package_dir={'rest_surveys': 'rest_surveys'},
install_requires=[
'Django>=1.7',
'django-inline-ordering',
'django-filter<=0.11.0',
'djangorestframework>=3.0',
'djangorestframework-bulk',
],
license=license,
)
| Add django-filter to the required packages | Add django-filter to the required packages
| Python | mit | danxshap/django-rest-surveys | #!/usr/bin/env python
import os
import sys
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
if sys.argv[-1] == 'publish':
os.system('python setup.py sdist bdist_wininst upload -r pypi')
sys.exit()
with open('README.rst') as f:
readme = f.read()
with open('LICENSE') as f:
license = f.read()
setup(
name='django-rest-surveys',
version='0.1.0',
description='A RESTful backend for giving surveys.',
long_description=readme,
author='Designlab',
author_email='hello@trydesignlab.com',
url='https://github.com/danxshap/django-rest-surveys',
packages=['rest_surveys'],
package_data={'': ['LICENSE']},
package_dir={'rest_surveys': 'rest_surveys'},
install_requires=[
'Django>=1.7',
'djangorestframework>=3.0',
'django-inline-ordering',
'djangorestframework-bulk',
],
license=license,
)
Add django-filter to the required packages | #!/usr/bin/env python
import os
import sys
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
if sys.argv[-1] == 'publish':
os.system('python setup.py sdist bdist_wininst upload -r pypi')
sys.exit()
with open('README.rst') as f:
readme = f.read()
with open('LICENSE') as f:
license = f.read()
setup(
name='django-rest-surveys',
version='0.1.0',
description='A RESTful backend for giving surveys.',
long_description=readme,
author='Designlab',
author_email='hello@trydesignlab.com',
url='https://github.com/danxshap/django-rest-surveys',
packages=['rest_surveys'],
package_data={'': ['LICENSE']},
package_dir={'rest_surveys': 'rest_surveys'},
install_requires=[
'Django>=1.7',
'django-inline-ordering',
'django-filter<=0.11.0',
'djangorestframework>=3.0',
'djangorestframework-bulk',
],
license=license,
)
| <commit_before>#!/usr/bin/env python
import os
import sys
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
if sys.argv[-1] == 'publish':
os.system('python setup.py sdist bdist_wininst upload -r pypi')
sys.exit()
with open('README.rst') as f:
readme = f.read()
with open('LICENSE') as f:
license = f.read()
setup(
name='django-rest-surveys',
version='0.1.0',
description='A RESTful backend for giving surveys.',
long_description=readme,
author='Designlab',
author_email='hello@trydesignlab.com',
url='https://github.com/danxshap/django-rest-surveys',
packages=['rest_surveys'],
package_data={'': ['LICENSE']},
package_dir={'rest_surveys': 'rest_surveys'},
install_requires=[
'Django>=1.7',
'djangorestframework>=3.0',
'django-inline-ordering',
'djangorestframework-bulk',
],
license=license,
)
<commit_msg>Add django-filter to the required packages<commit_after> | #!/usr/bin/env python
import os
import sys
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
if sys.argv[-1] == 'publish':
os.system('python setup.py sdist bdist_wininst upload -r pypi')
sys.exit()
with open('README.rst') as f:
readme = f.read()
with open('LICENSE') as f:
license = f.read()
setup(
name='django-rest-surveys',
version='0.1.0',
description='A RESTful backend for giving surveys.',
long_description=readme,
author='Designlab',
author_email='hello@trydesignlab.com',
url='https://github.com/danxshap/django-rest-surveys',
packages=['rest_surveys'],
package_data={'': ['LICENSE']},
package_dir={'rest_surveys': 'rest_surveys'},
install_requires=[
'Django>=1.7',
'django-inline-ordering',
'django-filter<=0.11.0',
'djangorestframework>=3.0',
'djangorestframework-bulk',
],
license=license,
)
| #!/usr/bin/env python
import os
import sys
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
if sys.argv[-1] == 'publish':
os.system('python setup.py sdist bdist_wininst upload -r pypi')
sys.exit()
with open('README.rst') as f:
readme = f.read()
with open('LICENSE') as f:
license = f.read()
setup(
name='django-rest-surveys',
version='0.1.0',
description='A RESTful backend for giving surveys.',
long_description=readme,
author='Designlab',
author_email='hello@trydesignlab.com',
url='https://github.com/danxshap/django-rest-surveys',
packages=['rest_surveys'],
package_data={'': ['LICENSE']},
package_dir={'rest_surveys': 'rest_surveys'},
install_requires=[
'Django>=1.7',
'djangorestframework>=3.0',
'django-inline-ordering',
'djangorestframework-bulk',
],
license=license,
)
Add django-filter to the required packages#!/usr/bin/env python
import os
import sys
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
if sys.argv[-1] == 'publish':
os.system('python setup.py sdist bdist_wininst upload -r pypi')
sys.exit()
with open('README.rst') as f:
readme = f.read()
with open('LICENSE') as f:
license = f.read()
setup(
name='django-rest-surveys',
version='0.1.0',
description='A RESTful backend for giving surveys.',
long_description=readme,
author='Designlab',
author_email='hello@trydesignlab.com',
url='https://github.com/danxshap/django-rest-surveys',
packages=['rest_surveys'],
package_data={'': ['LICENSE']},
package_dir={'rest_surveys': 'rest_surveys'},
install_requires=[
'Django>=1.7',
'django-inline-ordering',
'django-filter<=0.11.0',
'djangorestframework>=3.0',
'djangorestframework-bulk',
],
license=license,
)
| <commit_before>#!/usr/bin/env python
import os
import sys
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
if sys.argv[-1] == 'publish':
os.system('python setup.py sdist bdist_wininst upload -r pypi')
sys.exit()
with open('README.rst') as f:
readme = f.read()
with open('LICENSE') as f:
license = f.read()
setup(
name='django-rest-surveys',
version='0.1.0',
description='A RESTful backend for giving surveys.',
long_description=readme,
author='Designlab',
author_email='hello@trydesignlab.com',
url='https://github.com/danxshap/django-rest-surveys',
packages=['rest_surveys'],
package_data={'': ['LICENSE']},
package_dir={'rest_surveys': 'rest_surveys'},
install_requires=[
'Django>=1.7',
'djangorestframework>=3.0',
'django-inline-ordering',
'djangorestframework-bulk',
],
license=license,
)
<commit_msg>Add django-filter to the required packages<commit_after>#!/usr/bin/env python
import os
import sys
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
if sys.argv[-1] == 'publish':
os.system('python setup.py sdist bdist_wininst upload -r pypi')
sys.exit()
with open('README.rst') as f:
readme = f.read()
with open('LICENSE') as f:
license = f.read()
setup(
name='django-rest-surveys',
version='0.1.0',
description='A RESTful backend for giving surveys.',
long_description=readme,
author='Designlab',
author_email='hello@trydesignlab.com',
url='https://github.com/danxshap/django-rest-surveys',
packages=['rest_surveys'],
package_data={'': ['LICENSE']},
package_dir={'rest_surveys': 'rest_surveys'},
install_requires=[
'Django>=1.7',
'django-inline-ordering',
'django-filter<=0.11.0',
'djangorestframework>=3.0',
'djangorestframework-bulk',
],
license=license,
)
|
c36d6b17be66c0dfd0a540205b24b22b97739fb9 | setup.py | setup.py | from setuptools import setup
setup(
name='PyFVCOM',
packages=['PyFVCOM'],
version='2.1.0',
description=("PyFVCOM is a collection of various tools and utilities which can be used to extract, analyse and plot input and output files from FVCOM."),
author='Pierre Cazenave',
author_email='pica@pml.ac.uk',
url='https://gitlab.ecosystem-modelling.pml.ac.uk/fvcom/PyFVCOM',
download_url='http://gitlab.em.pml.ac.uk/fvcom/PyFVCOM/repository/archive.tar.gz?ref=2.1.0',
keywords=['fvcom', 'unstructured grid', 'mesh'],
license='MIT',
platforms='any',
install_requires=['jdcal', 'lxml', 'matplotlib', 'netCDF4', 'networkx', 'numpy>=1.13.0', 'pandas', 'pyproj', 'pytz', 'scipy', 'pyshp', 'UTide', 'shapely'],
classifiers=[]
)
| from setuptools import setup
version = '2.1.0'
setup(name='PyFVCOM',
packages=['PyFVCOM'],
version=version,
description=("PyFVCOM is a collection of various tools and utilities which can be used to extract, analyse and plot input and output files from FVCOM."),
author='Pierre Cazenave',
author_email='pica@pml.ac.uk',
url='https://gitlab.ecosystem-modelling.pml.ac.uk/fvcom/PyFVCOM',
download_url='http://gitlab.em.pml.ac.uk/fvcom/PyFVCOM/repository/archive.tar.gz?ref={}'.format(version),
keywords=['fvcom', 'unstructured grid', 'mesh'],
license='MIT',
platforms='any',
install_requires=['jdcal', 'lxml', 'matplotlib', 'netCDF4', 'networkx', 'numpy>=1.13.0', 'pandas', 'pyproj', 'pytz', 'scipy', 'pyshp', 'UTide', 'shapely'],
classifiers=[])
| Fix formatting and define the version up front. | Fix formatting and define the version up front.
| Python | mit | pwcazenave/PyFVCOM | from setuptools import setup
setup(
name='PyFVCOM',
packages=['PyFVCOM'],
version='2.1.0',
description=("PyFVCOM is a collection of various tools and utilities which can be used to extract, analyse and plot input and output files from FVCOM."),
author='Pierre Cazenave',
author_email='pica@pml.ac.uk',
url='https://gitlab.ecosystem-modelling.pml.ac.uk/fvcom/PyFVCOM',
download_url='http://gitlab.em.pml.ac.uk/fvcom/PyFVCOM/repository/archive.tar.gz?ref=2.1.0',
keywords=['fvcom', 'unstructured grid', 'mesh'],
license='MIT',
platforms='any',
install_requires=['jdcal', 'lxml', 'matplotlib', 'netCDF4', 'networkx', 'numpy>=1.13.0', 'pandas', 'pyproj', 'pytz', 'scipy', 'pyshp', 'UTide', 'shapely'],
classifiers=[]
)
Fix formatting and define the version up front. | from setuptools import setup
version = '2.1.0'
setup(name='PyFVCOM',
packages=['PyFVCOM'],
version=version,
description=("PyFVCOM is a collection of various tools and utilities which can be used to extract, analyse and plot input and output files from FVCOM."),
author='Pierre Cazenave',
author_email='pica@pml.ac.uk',
url='https://gitlab.ecosystem-modelling.pml.ac.uk/fvcom/PyFVCOM',
download_url='http://gitlab.em.pml.ac.uk/fvcom/PyFVCOM/repository/archive.tar.gz?ref={}'.format(version),
keywords=['fvcom', 'unstructured grid', 'mesh'],
license='MIT',
platforms='any',
install_requires=['jdcal', 'lxml', 'matplotlib', 'netCDF4', 'networkx', 'numpy>=1.13.0', 'pandas', 'pyproj', 'pytz', 'scipy', 'pyshp', 'UTide', 'shapely'],
classifiers=[])
| <commit_before>from setuptools import setup
setup(
name='PyFVCOM',
packages=['PyFVCOM'],
version='2.1.0',
description=("PyFVCOM is a collection of various tools and utilities which can be used to extract, analyse and plot input and output files from FVCOM."),
author='Pierre Cazenave',
author_email='pica@pml.ac.uk',
url='https://gitlab.ecosystem-modelling.pml.ac.uk/fvcom/PyFVCOM',
download_url='http://gitlab.em.pml.ac.uk/fvcom/PyFVCOM/repository/archive.tar.gz?ref=2.1.0',
keywords=['fvcom', 'unstructured grid', 'mesh'],
license='MIT',
platforms='any',
install_requires=['jdcal', 'lxml', 'matplotlib', 'netCDF4', 'networkx', 'numpy>=1.13.0', 'pandas', 'pyproj', 'pytz', 'scipy', 'pyshp', 'UTide', 'shapely'],
classifiers=[]
)
<commit_msg>Fix formatting and define the version up front.<commit_after> | from setuptools import setup
version = '2.1.0'
setup(name='PyFVCOM',
packages=['PyFVCOM'],
version=version,
description=("PyFVCOM is a collection of various tools and utilities which can be used to extract, analyse and plot input and output files from FVCOM."),
author='Pierre Cazenave',
author_email='pica@pml.ac.uk',
url='https://gitlab.ecosystem-modelling.pml.ac.uk/fvcom/PyFVCOM',
download_url='http://gitlab.em.pml.ac.uk/fvcom/PyFVCOM/repository/archive.tar.gz?ref={}'.format(version),
keywords=['fvcom', 'unstructured grid', 'mesh'],
license='MIT',
platforms='any',
install_requires=['jdcal', 'lxml', 'matplotlib', 'netCDF4', 'networkx', 'numpy>=1.13.0', 'pandas', 'pyproj', 'pytz', 'scipy', 'pyshp', 'UTide', 'shapely'],
classifiers=[])
| from setuptools import setup
setup(
name='PyFVCOM',
packages=['PyFVCOM'],
version='2.1.0',
description=("PyFVCOM is a collection of various tools and utilities which can be used to extract, analyse and plot input and output files from FVCOM."),
author='Pierre Cazenave',
author_email='pica@pml.ac.uk',
url='https://gitlab.ecosystem-modelling.pml.ac.uk/fvcom/PyFVCOM',
download_url='http://gitlab.em.pml.ac.uk/fvcom/PyFVCOM/repository/archive.tar.gz?ref=2.1.0',
keywords=['fvcom', 'unstructured grid', 'mesh'],
license='MIT',
platforms='any',
install_requires=['jdcal', 'lxml', 'matplotlib', 'netCDF4', 'networkx', 'numpy>=1.13.0', 'pandas', 'pyproj', 'pytz', 'scipy', 'pyshp', 'UTide', 'shapely'],
classifiers=[]
)
Fix formatting and define the version up front.from setuptools import setup
version = '2.1.0'
setup(name='PyFVCOM',
packages=['PyFVCOM'],
version=version,
description=("PyFVCOM is a collection of various tools and utilities which can be used to extract, analyse and plot input and output files from FVCOM."),
author='Pierre Cazenave',
author_email='pica@pml.ac.uk',
url='https://gitlab.ecosystem-modelling.pml.ac.uk/fvcom/PyFVCOM',
download_url='http://gitlab.em.pml.ac.uk/fvcom/PyFVCOM/repository/archive.tar.gz?ref={}'.format(version),
keywords=['fvcom', 'unstructured grid', 'mesh'],
license='MIT',
platforms='any',
install_requires=['jdcal', 'lxml', 'matplotlib', 'netCDF4', 'networkx', 'numpy>=1.13.0', 'pandas', 'pyproj', 'pytz', 'scipy', 'pyshp', 'UTide', 'shapely'],
classifiers=[])
| <commit_before>from setuptools import setup
setup(
name='PyFVCOM',
packages=['PyFVCOM'],
version='2.1.0',
description=("PyFVCOM is a collection of various tools and utilities which can be used to extract, analyse and plot input and output files from FVCOM."),
author='Pierre Cazenave',
author_email='pica@pml.ac.uk',
url='https://gitlab.ecosystem-modelling.pml.ac.uk/fvcom/PyFVCOM',
download_url='http://gitlab.em.pml.ac.uk/fvcom/PyFVCOM/repository/archive.tar.gz?ref=2.1.0',
keywords=['fvcom', 'unstructured grid', 'mesh'],
license='MIT',
platforms='any',
install_requires=['jdcal', 'lxml', 'matplotlib', 'netCDF4', 'networkx', 'numpy>=1.13.0', 'pandas', 'pyproj', 'pytz', 'scipy', 'pyshp', 'UTide', 'shapely'],
classifiers=[]
)
<commit_msg>Fix formatting and define the version up front.<commit_after>from setuptools import setup
version = '2.1.0'
setup(name='PyFVCOM',
packages=['PyFVCOM'],
version=version,
description=("PyFVCOM is a collection of various tools and utilities which can be used to extract, analyse and plot input and output files from FVCOM."),
author='Pierre Cazenave',
author_email='pica@pml.ac.uk',
url='https://gitlab.ecosystem-modelling.pml.ac.uk/fvcom/PyFVCOM',
download_url='http://gitlab.em.pml.ac.uk/fvcom/PyFVCOM/repository/archive.tar.gz?ref={}'.format(version),
keywords=['fvcom', 'unstructured grid', 'mesh'],
license='MIT',
platforms='any',
install_requires=['jdcal', 'lxml', 'matplotlib', 'netCDF4', 'networkx', 'numpy>=1.13.0', 'pandas', 'pyproj', 'pytz', 'scipy', 'pyshp', 'UTide', 'shapely'],
classifiers=[])
|
c1218917e6169c9eaaf821f96610c1e9e6d81862 | setup.py | setup.py | from setuptools import setup
setup(
name = 'tomso',
packages = ['tomso'],
version = '0.0.7',
description = 'Tools for Modelling Stars and their Oscillations',
author = 'Warrick Ball',
author_email = 'W.H.Ball@bham.ac.uk',
url = 'https://github.com/warrickball/tomso',
download_url = 'https://github.com/warrickball/tomso/archive/v0.0.7.tar.gz',
install_requires=['numpy'],
keywords = [],
classifiers = [],
license = 'MIT'
)
| from setuptools import setup
setup(
name = 'tomso',
packages = ['tomso'],
version = '0.0.7',
description = 'Tools for Modelling Stars and their Oscillations',
long_description=open('README.md').read(),
long_description_content_type='text/markdown',
author = 'Warrick Ball',
author_email = 'W.H.Ball@bham.ac.uk',
url = 'https://github.com/warrickball/tomso',
download_url = 'https://github.com/warrickball/tomso/archive/v0.0.7.tar.gz',
install_requires=['numpy'],
keywords = [],
classifiers = [],
license = 'MIT'
)
| Use README.md for PyPI description | Use README.md for PyPI description
| Python | mit | warrickball/tomso | from setuptools import setup
setup(
name = 'tomso',
packages = ['tomso'],
version = '0.0.7',
description = 'Tools for Modelling Stars and their Oscillations',
author = 'Warrick Ball',
author_email = 'W.H.Ball@bham.ac.uk',
url = 'https://github.com/warrickball/tomso',
download_url = 'https://github.com/warrickball/tomso/archive/v0.0.7.tar.gz',
install_requires=['numpy'],
keywords = [],
classifiers = [],
license = 'MIT'
)
Use README.md for PyPI description | from setuptools import setup
setup(
name = 'tomso',
packages = ['tomso'],
version = '0.0.7',
description = 'Tools for Modelling Stars and their Oscillations',
long_description=open('README.md').read(),
long_description_content_type='text/markdown',
author = 'Warrick Ball',
author_email = 'W.H.Ball@bham.ac.uk',
url = 'https://github.com/warrickball/tomso',
download_url = 'https://github.com/warrickball/tomso/archive/v0.0.7.tar.gz',
install_requires=['numpy'],
keywords = [],
classifiers = [],
license = 'MIT'
)
| <commit_before>from setuptools import setup
setup(
name = 'tomso',
packages = ['tomso'],
version = '0.0.7',
description = 'Tools for Modelling Stars and their Oscillations',
author = 'Warrick Ball',
author_email = 'W.H.Ball@bham.ac.uk',
url = 'https://github.com/warrickball/tomso',
download_url = 'https://github.com/warrickball/tomso/archive/v0.0.7.tar.gz',
install_requires=['numpy'],
keywords = [],
classifiers = [],
license = 'MIT'
)
<commit_msg>Use README.md for PyPI description<commit_after> | from setuptools import setup
setup(
name = 'tomso',
packages = ['tomso'],
version = '0.0.7',
description = 'Tools for Modelling Stars and their Oscillations',
long_description=open('README.md').read(),
long_description_content_type='text/markdown',
author = 'Warrick Ball',
author_email = 'W.H.Ball@bham.ac.uk',
url = 'https://github.com/warrickball/tomso',
download_url = 'https://github.com/warrickball/tomso/archive/v0.0.7.tar.gz',
install_requires=['numpy'],
keywords = [],
classifiers = [],
license = 'MIT'
)
| from setuptools import setup
setup(
name = 'tomso',
packages = ['tomso'],
version = '0.0.7',
description = 'Tools for Modelling Stars and their Oscillations',
author = 'Warrick Ball',
author_email = 'W.H.Ball@bham.ac.uk',
url = 'https://github.com/warrickball/tomso',
download_url = 'https://github.com/warrickball/tomso/archive/v0.0.7.tar.gz',
install_requires=['numpy'],
keywords = [],
classifiers = [],
license = 'MIT'
)
Use README.md for PyPI descriptionfrom setuptools import setup
setup(
name = 'tomso',
packages = ['tomso'],
version = '0.0.7',
description = 'Tools for Modelling Stars and their Oscillations',
long_description=open('README.md').read(),
long_description_content_type='text/markdown',
author = 'Warrick Ball',
author_email = 'W.H.Ball@bham.ac.uk',
url = 'https://github.com/warrickball/tomso',
download_url = 'https://github.com/warrickball/tomso/archive/v0.0.7.tar.gz',
install_requires=['numpy'],
keywords = [],
classifiers = [],
license = 'MIT'
)
| <commit_before>from setuptools import setup
setup(
name = 'tomso',
packages = ['tomso'],
version = '0.0.7',
description = 'Tools for Modelling Stars and their Oscillations',
author = 'Warrick Ball',
author_email = 'W.H.Ball@bham.ac.uk',
url = 'https://github.com/warrickball/tomso',
download_url = 'https://github.com/warrickball/tomso/archive/v0.0.7.tar.gz',
install_requires=['numpy'],
keywords = [],
classifiers = [],
license = 'MIT'
)
<commit_msg>Use README.md for PyPI description<commit_after>from setuptools import setup
setup(
name = 'tomso',
packages = ['tomso'],
version = '0.0.7',
description = 'Tools for Modelling Stars and their Oscillations',
long_description=open('README.md').read(),
long_description_content_type='text/markdown',
author = 'Warrick Ball',
author_email = 'W.H.Ball@bham.ac.uk',
url = 'https://github.com/warrickball/tomso',
download_url = 'https://github.com/warrickball/tomso/archive/v0.0.7.tar.gz',
install_requires=['numpy'],
keywords = [],
classifiers = [],
license = 'MIT'
)
|
80d4cd9008d70664e7981a5e6018565d6b63d07a | setup.py | setup.py | from distutils.core import setup
setup(
name='udiskie',
version='0.3.9',
description='Removable disk automounter for udisks',
author='Byron Clark',
author_email='byron@theclarkfamily.name',
url='http://bitbucket.org/byronclark/udiskie',
license='MIT',
packages=[
'udiskie',
],
scripts=[
'bin/udiskie',
'bin/udiskie-umount',
],
)
| from distutils.core import setup
setup(
name='udiskie',
version='0.3.10',
description='Removable disk automounter for udisks',
author='Byron Clark',
author_email='byron@theclarkfamily.name',
url='http://bitbucket.org/byronclark/udiskie',
license='MIT',
packages=[
'udiskie',
],
scripts=[
'bin/udiskie',
'bin/udiskie-umount',
],
)
| Prepare for next development cycle. | Prepare for next development cycle.
| Python | mit | coldfix/udiskie,pstray/udiskie,pstray/udiskie,khardix/udiskie,coldfix/udiskie,mathstuf/udiskie | from distutils.core import setup
setup(
name='udiskie',
version='0.3.9',
description='Removable disk automounter for udisks',
author='Byron Clark',
author_email='byron@theclarkfamily.name',
url='http://bitbucket.org/byronclark/udiskie',
license='MIT',
packages=[
'udiskie',
],
scripts=[
'bin/udiskie',
'bin/udiskie-umount',
],
)
Prepare for next development cycle. | from distutils.core import setup
setup(
name='udiskie',
version='0.3.10',
description='Removable disk automounter for udisks',
author='Byron Clark',
author_email='byron@theclarkfamily.name',
url='http://bitbucket.org/byronclark/udiskie',
license='MIT',
packages=[
'udiskie',
],
scripts=[
'bin/udiskie',
'bin/udiskie-umount',
],
)
| <commit_before>from distutils.core import setup
setup(
name='udiskie',
version='0.3.9',
description='Removable disk automounter for udisks',
author='Byron Clark',
author_email='byron@theclarkfamily.name',
url='http://bitbucket.org/byronclark/udiskie',
license='MIT',
packages=[
'udiskie',
],
scripts=[
'bin/udiskie',
'bin/udiskie-umount',
],
)
<commit_msg>Prepare for next development cycle.<commit_after> | from distutils.core import setup
setup(
name='udiskie',
version='0.3.10',
description='Removable disk automounter for udisks',
author='Byron Clark',
author_email='byron@theclarkfamily.name',
url='http://bitbucket.org/byronclark/udiskie',
license='MIT',
packages=[
'udiskie',
],
scripts=[
'bin/udiskie',
'bin/udiskie-umount',
],
)
| from distutils.core import setup
setup(
name='udiskie',
version='0.3.9',
description='Removable disk automounter for udisks',
author='Byron Clark',
author_email='byron@theclarkfamily.name',
url='http://bitbucket.org/byronclark/udiskie',
license='MIT',
packages=[
'udiskie',
],
scripts=[
'bin/udiskie',
'bin/udiskie-umount',
],
)
Prepare for next development cycle.from distutils.core import setup
setup(
name='udiskie',
version='0.3.10',
description='Removable disk automounter for udisks',
author='Byron Clark',
author_email='byron@theclarkfamily.name',
url='http://bitbucket.org/byronclark/udiskie',
license='MIT',
packages=[
'udiskie',
],
scripts=[
'bin/udiskie',
'bin/udiskie-umount',
],
)
| <commit_before>from distutils.core import setup
setup(
name='udiskie',
version='0.3.9',
description='Removable disk automounter for udisks',
author='Byron Clark',
author_email='byron@theclarkfamily.name',
url='http://bitbucket.org/byronclark/udiskie',
license='MIT',
packages=[
'udiskie',
],
scripts=[
'bin/udiskie',
'bin/udiskie-umount',
],
)
<commit_msg>Prepare for next development cycle.<commit_after>from distutils.core import setup
setup(
name='udiskie',
version='0.3.10',
description='Removable disk automounter for udisks',
author='Byron Clark',
author_email='byron@theclarkfamily.name',
url='http://bitbucket.org/byronclark/udiskie',
license='MIT',
packages=[
'udiskie',
],
scripts=[
'bin/udiskie',
'bin/udiskie-umount',
],
)
|
3e04c7e86d92785ba07f30ed2c0ec4eb575d6218 | setup.py | setup.py | from setuptools import setup, find_packages
import sys, os
setup(name='cc.license',
version='0.01',
description="License selection based on ccREL-based metadata.",
classifiers=[],
keywords='',
author='Creative Commons',
author_email='software@creativecommons.org',
url='http://wiki.creativecommons.org/CcLicense',
license='MIT',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
include_package_data=True,
#package_data={'cc.license': ['*.xml', '*.txt']}, # doesn't work
data_files=[('cc/license/rdf', ['license.rdf/rdf/index.rdf',
'license.rdf/rdf/selectors.rdf',
'license.rdf/rdf/jurisdictions.rdf']),
('cc/license/xml', ['license.rdf/xml/questions.xml'])],
zip_safe=False,
test_suite='nose.collector',
install_requires=[
'setuptools',
'zope.interface',
'nose',
'Genshi',
'pylons', # XXX why does nose throw a RuntimeWarning without this?
],
setup_requires=['setuptools-git',],
entry_points="""
# -*- Entry points: -*-
[nose.plugins]
pylons = pylons.test:PylonsPlugin
""",
)
| from setuptools import setup, find_packages
import sys, os
setup(name='cc.license',
version='0.02',
description="License selection based on ccREL-based metadata.",
classifiers=[],
keywords='',
author='Creative Commons',
author_email='software@creativecommons.org',
url='http://wiki.creativecommons.org/CcLicense',
license='MIT',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
include_package_data=True,
#package_data={'cc.license': ['*.xml', '*.txt']}, # doesn't work
data_files=[('cc/license/rdf', ['license.rdf/rdf/index.rdf',
'license.rdf/rdf/selectors.rdf',
'license.rdf/rdf/jurisdictions.rdf']),
('cc/license/xml', ['license.rdf/xml/questions.xml'])],
zip_safe=False,
test_suite='nose.collector',
install_requires=[
'setuptools',
'zope.interface',
'nose',
'Genshi',
'pylons', # XXX why does nose throw a RuntimeWarning without this?
],
setup_requires=['setuptools-git',],
entry_points="""
# -*- Entry points: -*-
[nose.plugins]
pylons = pylons.test:PylonsPlugin
""",
)
| Increment minor version number for new release (v0.02). | Increment minor version number for new release (v0.02).
| Python | mit | creativecommons/cc.license,creativecommons/cc.license | from setuptools import setup, find_packages
import sys, os
setup(name='cc.license',
version='0.01',
description="License selection based on ccREL-based metadata.",
classifiers=[],
keywords='',
author='Creative Commons',
author_email='software@creativecommons.org',
url='http://wiki.creativecommons.org/CcLicense',
license='MIT',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
include_package_data=True,
#package_data={'cc.license': ['*.xml', '*.txt']}, # doesn't work
data_files=[('cc/license/rdf', ['license.rdf/rdf/index.rdf',
'license.rdf/rdf/selectors.rdf',
'license.rdf/rdf/jurisdictions.rdf']),
('cc/license/xml', ['license.rdf/xml/questions.xml'])],
zip_safe=False,
test_suite='nose.collector',
install_requires=[
'setuptools',
'zope.interface',
'nose',
'Genshi',
'pylons', # XXX why does nose throw a RuntimeWarning without this?
],
setup_requires=['setuptools-git',],
entry_points="""
# -*- Entry points: -*-
[nose.plugins]
pylons = pylons.test:PylonsPlugin
""",
)
Increment minor version number for new release (v0.02). | from setuptools import setup, find_packages
import sys, os
setup(name='cc.license',
version='0.02',
description="License selection based on ccREL-based metadata.",
classifiers=[],
keywords='',
author='Creative Commons',
author_email='software@creativecommons.org',
url='http://wiki.creativecommons.org/CcLicense',
license='MIT',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
include_package_data=True,
#package_data={'cc.license': ['*.xml', '*.txt']}, # doesn't work
data_files=[('cc/license/rdf', ['license.rdf/rdf/index.rdf',
'license.rdf/rdf/selectors.rdf',
'license.rdf/rdf/jurisdictions.rdf']),
('cc/license/xml', ['license.rdf/xml/questions.xml'])],
zip_safe=False,
test_suite='nose.collector',
install_requires=[
'setuptools',
'zope.interface',
'nose',
'Genshi',
'pylons', # XXX why does nose throw a RuntimeWarning without this?
],
setup_requires=['setuptools-git',],
entry_points="""
# -*- Entry points: -*-
[nose.plugins]
pylons = pylons.test:PylonsPlugin
""",
)
| <commit_before>from setuptools import setup, find_packages
import sys, os
setup(name='cc.license',
version='0.01',
description="License selection based on ccREL-based metadata.",
classifiers=[],
keywords='',
author='Creative Commons',
author_email='software@creativecommons.org',
url='http://wiki.creativecommons.org/CcLicense',
license='MIT',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
include_package_data=True,
#package_data={'cc.license': ['*.xml', '*.txt']}, # doesn't work
data_files=[('cc/license/rdf', ['license.rdf/rdf/index.rdf',
'license.rdf/rdf/selectors.rdf',
'license.rdf/rdf/jurisdictions.rdf']),
('cc/license/xml', ['license.rdf/xml/questions.xml'])],
zip_safe=False,
test_suite='nose.collector',
install_requires=[
'setuptools',
'zope.interface',
'nose',
'Genshi',
'pylons', # XXX why does nose throw a RuntimeWarning without this?
],
setup_requires=['setuptools-git',],
entry_points="""
# -*- Entry points: -*-
[nose.plugins]
pylons = pylons.test:PylonsPlugin
""",
)
<commit_msg>Increment minor version number for new release (v0.02).<commit_after> | from setuptools import setup, find_packages
import sys, os
setup(name='cc.license',
version='0.02',
description="License selection based on ccREL-based metadata.",
classifiers=[],
keywords='',
author='Creative Commons',
author_email='software@creativecommons.org',
url='http://wiki.creativecommons.org/CcLicense',
license='MIT',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
include_package_data=True,
#package_data={'cc.license': ['*.xml', '*.txt']}, # doesn't work
data_files=[('cc/license/rdf', ['license.rdf/rdf/index.rdf',
'license.rdf/rdf/selectors.rdf',
'license.rdf/rdf/jurisdictions.rdf']),
('cc/license/xml', ['license.rdf/xml/questions.xml'])],
zip_safe=False,
test_suite='nose.collector',
install_requires=[
'setuptools',
'zope.interface',
'nose',
'Genshi',
'pylons', # XXX why does nose throw a RuntimeWarning without this?
],
setup_requires=['setuptools-git',],
entry_points="""
# -*- Entry points: -*-
[nose.plugins]
pylons = pylons.test:PylonsPlugin
""",
)
| from setuptools import setup, find_packages
import sys, os
setup(name='cc.license',
version='0.01',
description="License selection based on ccREL-based metadata.",
classifiers=[],
keywords='',
author='Creative Commons',
author_email='software@creativecommons.org',
url='http://wiki.creativecommons.org/CcLicense',
license='MIT',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
include_package_data=True,
#package_data={'cc.license': ['*.xml', '*.txt']}, # doesn't work
data_files=[('cc/license/rdf', ['license.rdf/rdf/index.rdf',
'license.rdf/rdf/selectors.rdf',
'license.rdf/rdf/jurisdictions.rdf']),
('cc/license/xml', ['license.rdf/xml/questions.xml'])],
zip_safe=False,
test_suite='nose.collector',
install_requires=[
'setuptools',
'zope.interface',
'nose',
'Genshi',
'pylons', # XXX why does nose throw a RuntimeWarning without this?
],
setup_requires=['setuptools-git',],
entry_points="""
# -*- Entry points: -*-
[nose.plugins]
pylons = pylons.test:PylonsPlugin
""",
)
Increment minor version number for new release (v0.02).from setuptools import setup, find_packages
import sys, os
setup(name='cc.license',
version='0.02',
description="License selection based on ccREL-based metadata.",
classifiers=[],
keywords='',
author='Creative Commons',
author_email='software@creativecommons.org',
url='http://wiki.creativecommons.org/CcLicense',
license='MIT',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
include_package_data=True,
#package_data={'cc.license': ['*.xml', '*.txt']}, # doesn't work
data_files=[('cc/license/rdf', ['license.rdf/rdf/index.rdf',
'license.rdf/rdf/selectors.rdf',
'license.rdf/rdf/jurisdictions.rdf']),
('cc/license/xml', ['license.rdf/xml/questions.xml'])],
zip_safe=False,
test_suite='nose.collector',
install_requires=[
'setuptools',
'zope.interface',
'nose',
'Genshi',
'pylons', # XXX why does nose throw a RuntimeWarning without this?
],
setup_requires=['setuptools-git',],
entry_points="""
# -*- Entry points: -*-
[nose.plugins]
pylons = pylons.test:PylonsPlugin
""",
)
| <commit_before>from setuptools import setup, find_packages
import sys, os
setup(name='cc.license',
version='0.01',
description="License selection based on ccREL-based metadata.",
classifiers=[],
keywords='',
author='Creative Commons',
author_email='software@creativecommons.org',
url='http://wiki.creativecommons.org/CcLicense',
license='MIT',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
include_package_data=True,
#package_data={'cc.license': ['*.xml', '*.txt']}, # doesn't work
data_files=[('cc/license/rdf', ['license.rdf/rdf/index.rdf',
'license.rdf/rdf/selectors.rdf',
'license.rdf/rdf/jurisdictions.rdf']),
('cc/license/xml', ['license.rdf/xml/questions.xml'])],
zip_safe=False,
test_suite='nose.collector',
install_requires=[
'setuptools',
'zope.interface',
'nose',
'Genshi',
'pylons', # XXX why does nose throw a RuntimeWarning without this?
],
setup_requires=['setuptools-git',],
entry_points="""
# -*- Entry points: -*-
[nose.plugins]
pylons = pylons.test:PylonsPlugin
""",
)
<commit_msg>Increment minor version number for new release (v0.02).<commit_after>from setuptools import setup, find_packages
import sys, os
setup(name='cc.license',
version='0.02',
description="License selection based on ccREL-based metadata.",
classifiers=[],
keywords='',
author='Creative Commons',
author_email='software@creativecommons.org',
url='http://wiki.creativecommons.org/CcLicense',
license='MIT',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
include_package_data=True,
#package_data={'cc.license': ['*.xml', '*.txt']}, # doesn't work
data_files=[('cc/license/rdf', ['license.rdf/rdf/index.rdf',
'license.rdf/rdf/selectors.rdf',
'license.rdf/rdf/jurisdictions.rdf']),
('cc/license/xml', ['license.rdf/xml/questions.xml'])],
zip_safe=False,
test_suite='nose.collector',
install_requires=[
'setuptools',
'zope.interface',
'nose',
'Genshi',
'pylons', # XXX why does nose throw a RuntimeWarning without this?
],
setup_requires=['setuptools-git',],
entry_points="""
# -*- Entry points: -*-
[nose.plugins]
pylons = pylons.test:PylonsPlugin
""",
)
|
a1208aeffb57e16f49007f86b144ab6d576cbd0d | setup.py | setup.py | #! /usr/bin/env python
# coding: utf-8
from setuptools import find_packages, setup
setup(name='egoio',
author='NEXT ENERGY, Reiner Lemoine Institut gGmbH, ZNES',
author_email='ulf.p.mueller@hs-flensburg.de',
description='ego input/output repository',
version='0.3.0',
url='https://github.com/openego/ego.io',
packages=find_packages(),
license='GNU Affero General Public License v3.0',
install_requires=[
'geoalchemy2 >= 0.3.0, <= 0.4.0',
'sqlalchemy >= 1.0.11, <= 1.1.15',
'keyring >= 4.0',
'psycopg2'],
extras_require={
"sqlalchemy": 'postgresql'},
package_data={'tools': 'sqlacodegen_oedb.sh'}
)
| #! /usr/bin/env python
# coding: utf-8
from setuptools import find_packages, setup
setup(name='egoio',
author='NEXT ENERGY, Reiner Lemoine Institut gGmbH, ZNES',
author_email='ulf.p.mueller@hs-flensburg.de',
description='ego input/output repository',
version='0.3.0',
url='https://github.com/openego/ego.io',
packages=find_packages(),
license='GNU Affero General Public License v3.0',
install_requires=[
'geoalchemy2 >= 0.3.0, <= 0.4.1',
'sqlalchemy >= 1.0.11, <= 1.2.0',
'keyring >= 4.0',
'psycopg2'],
extras_require={
"sqlalchemy": 'postgresql'},
package_data={'tools': 'sqlacodegen_oedb.sh'}
)
| Update SQLAlchemy and Geoalchemy2 version range | Update SQLAlchemy and Geoalchemy2 version range
| Python | agpl-3.0 | openego/ego.io,openego/ego.io | #! /usr/bin/env python
# coding: utf-8
from setuptools import find_packages, setup
setup(name='egoio',
author='NEXT ENERGY, Reiner Lemoine Institut gGmbH, ZNES',
author_email='ulf.p.mueller@hs-flensburg.de',
description='ego input/output repository',
version='0.3.0',
url='https://github.com/openego/ego.io',
packages=find_packages(),
license='GNU Affero General Public License v3.0',
install_requires=[
'geoalchemy2 >= 0.3.0, <= 0.4.0',
'sqlalchemy >= 1.0.11, <= 1.1.15',
'keyring >= 4.0',
'psycopg2'],
extras_require={
"sqlalchemy": 'postgresql'},
package_data={'tools': 'sqlacodegen_oedb.sh'}
)
Update SQLAlchemy and Geoalchemy2 version range | #! /usr/bin/env python
# coding: utf-8
from setuptools import find_packages, setup
setup(name='egoio',
author='NEXT ENERGY, Reiner Lemoine Institut gGmbH, ZNES',
author_email='ulf.p.mueller@hs-flensburg.de',
description='ego input/output repository',
version='0.3.0',
url='https://github.com/openego/ego.io',
packages=find_packages(),
license='GNU Affero General Public License v3.0',
install_requires=[
'geoalchemy2 >= 0.3.0, <= 0.4.1',
'sqlalchemy >= 1.0.11, <= 1.2.0',
'keyring >= 4.0',
'psycopg2'],
extras_require={
"sqlalchemy": 'postgresql'},
package_data={'tools': 'sqlacodegen_oedb.sh'}
)
| <commit_before>#! /usr/bin/env python
# coding: utf-8
from setuptools import find_packages, setup
setup(name='egoio',
author='NEXT ENERGY, Reiner Lemoine Institut gGmbH, ZNES',
author_email='ulf.p.mueller@hs-flensburg.de',
description='ego input/output repository',
version='0.3.0',
url='https://github.com/openego/ego.io',
packages=find_packages(),
license='GNU Affero General Public License v3.0',
install_requires=[
'geoalchemy2 >= 0.3.0, <= 0.4.0',
'sqlalchemy >= 1.0.11, <= 1.1.15',
'keyring >= 4.0',
'psycopg2'],
extras_require={
"sqlalchemy": 'postgresql'},
package_data={'tools': 'sqlacodegen_oedb.sh'}
)
<commit_msg>Update SQLAlchemy and Geoalchemy2 version range<commit_after> | #! /usr/bin/env python
# coding: utf-8
from setuptools import find_packages, setup
setup(name='egoio',
author='NEXT ENERGY, Reiner Lemoine Institut gGmbH, ZNES',
author_email='ulf.p.mueller@hs-flensburg.de',
description='ego input/output repository',
version='0.3.0',
url='https://github.com/openego/ego.io',
packages=find_packages(),
license='GNU Affero General Public License v3.0',
install_requires=[
'geoalchemy2 >= 0.3.0, <= 0.4.1',
'sqlalchemy >= 1.0.11, <= 1.2.0',
'keyring >= 4.0',
'psycopg2'],
extras_require={
"sqlalchemy": 'postgresql'},
package_data={'tools': 'sqlacodegen_oedb.sh'}
)
| #! /usr/bin/env python
# coding: utf-8
from setuptools import find_packages, setup
setup(name='egoio',
author='NEXT ENERGY, Reiner Lemoine Institut gGmbH, ZNES',
author_email='ulf.p.mueller@hs-flensburg.de',
description='ego input/output repository',
version='0.3.0',
url='https://github.com/openego/ego.io',
packages=find_packages(),
license='GNU Affero General Public License v3.0',
install_requires=[
'geoalchemy2 >= 0.3.0, <= 0.4.0',
'sqlalchemy >= 1.0.11, <= 1.1.15',
'keyring >= 4.0',
'psycopg2'],
extras_require={
"sqlalchemy": 'postgresql'},
package_data={'tools': 'sqlacodegen_oedb.sh'}
)
Update SQLAlchemy and Geoalchemy2 version range#! /usr/bin/env python
# coding: utf-8
from setuptools import find_packages, setup
setup(name='egoio',
author='NEXT ENERGY, Reiner Lemoine Institut gGmbH, ZNES',
author_email='ulf.p.mueller@hs-flensburg.de',
description='ego input/output repository',
version='0.3.0',
url='https://github.com/openego/ego.io',
packages=find_packages(),
license='GNU Affero General Public License v3.0',
install_requires=[
'geoalchemy2 >= 0.3.0, <= 0.4.1',
'sqlalchemy >= 1.0.11, <= 1.2.0',
'keyring >= 4.0',
'psycopg2'],
extras_require={
"sqlalchemy": 'postgresql'},
package_data={'tools': 'sqlacodegen_oedb.sh'}
)
| <commit_before>#! /usr/bin/env python
# coding: utf-8
from setuptools import find_packages, setup
setup(name='egoio',
author='NEXT ENERGY, Reiner Lemoine Institut gGmbH, ZNES',
author_email='ulf.p.mueller@hs-flensburg.de',
description='ego input/output repository',
version='0.3.0',
url='https://github.com/openego/ego.io',
packages=find_packages(),
license='GNU Affero General Public License v3.0',
install_requires=[
'geoalchemy2 >= 0.3.0, <= 0.4.0',
'sqlalchemy >= 1.0.11, <= 1.1.15',
'keyring >= 4.0',
'psycopg2'],
extras_require={
"sqlalchemy": 'postgresql'},
package_data={'tools': 'sqlacodegen_oedb.sh'}
)
<commit_msg>Update SQLAlchemy and Geoalchemy2 version range<commit_after>#! /usr/bin/env python
# coding: utf-8
from setuptools import find_packages, setup
setup(name='egoio',
author='NEXT ENERGY, Reiner Lemoine Institut gGmbH, ZNES',
author_email='ulf.p.mueller@hs-flensburg.de',
description='ego input/output repository',
version='0.3.0',
url='https://github.com/openego/ego.io',
packages=find_packages(),
license='GNU Affero General Public License v3.0',
install_requires=[
'geoalchemy2 >= 0.3.0, <= 0.4.1',
'sqlalchemy >= 1.0.11, <= 1.2.0',
'keyring >= 4.0',
'psycopg2'],
extras_require={
"sqlalchemy": 'postgresql'},
package_data={'tools': 'sqlacodegen_oedb.sh'}
)
|
33f87d824118d07cf8a7379bc46f624da5e3b433 | setup.py | setup.py | import sys
IS_PYTHON3 = sys.version_info[0] == 3
if not IS_PYTHON3:
print("Error: signac requires python version >= 3.x.")
sys.exit(1)
from setuptools import setup, find_packages
setup(
name='signac',
version='0.1.7dev1',
packages=find_packages(),
author='Carl Simon Adorf',
author_email='csadorf@umich.edu',
description="Computational Database.",
keywords='simulation tools mc md monte-carlo mongodb '
'jobmanagement materials database',
classifiers=[
"Development Status :: 3 - Alpha",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Topic :: Scientific/Engineering :: Physics",
],
extras_require={
'db': ['pymongo>=3.0'],
'mpi': ['mpi4py'],
'conversion': ['networkx>=1.1.0'],
},
entry_points={
'console_scripts': [
'signac = signac.contrib.script:main',
'signac_init = signac.contrib.init_project:main',
'signac_configure = signac.contrib.configure:main',
'signac_admin = signac.admin.manage:main',
'signac_server = signac.contrib.server:main',
'signac_user = signac.contrib.admin:main',
'signac_admin_project = signac.admin.manage_project:main',
],
},
)
| import sys
IS_PYTHON3 = sys.version_info[0] == 3
if not IS_PYTHON3:
print("Error: signac requires python version >= 3.x.")
sys.exit(1)
from setuptools import setup, find_packages
setup(
name='signac',
version='0.1.7dev5',
packages=find_packages(),
author='Carl Simon Adorf',
author_email='csadorf@umich.edu',
description="Computational Database.",
keywords='simulation tools mc md monte-carlo mongodb '
'jobmanagement materials database',
classifiers=[
"Development Status :: 3 - Alpha",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Topic :: Scientific/Engineering :: Physics",
],
extras_require={
'db': ['pymongo>=3.0'],
'mpi': ['mpi4py'],
'conversion': ['networkx>=1.1.0'],
},
)
| Remove entry points and bump dev version. | Remove entry points and bump dev version.
| Python | bsd-3-clause | csadorf/signac,csadorf/signac | import sys
IS_PYTHON3 = sys.version_info[0] == 3
if not IS_PYTHON3:
print("Error: signac requires python version >= 3.x.")
sys.exit(1)
from setuptools import setup, find_packages
setup(
name='signac',
version='0.1.7dev1',
packages=find_packages(),
author='Carl Simon Adorf',
author_email='csadorf@umich.edu',
description="Computational Database.",
keywords='simulation tools mc md monte-carlo mongodb '
'jobmanagement materials database',
classifiers=[
"Development Status :: 3 - Alpha",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Topic :: Scientific/Engineering :: Physics",
],
extras_require={
'db': ['pymongo>=3.0'],
'mpi': ['mpi4py'],
'conversion': ['networkx>=1.1.0'],
},
entry_points={
'console_scripts': [
'signac = signac.contrib.script:main',
'signac_init = signac.contrib.init_project:main',
'signac_configure = signac.contrib.configure:main',
'signac_admin = signac.admin.manage:main',
'signac_server = signac.contrib.server:main',
'signac_user = signac.contrib.admin:main',
'signac_admin_project = signac.admin.manage_project:main',
],
},
)
Remove entry points and bump dev version. | import sys
IS_PYTHON3 = sys.version_info[0] == 3
if not IS_PYTHON3:
print("Error: signac requires python version >= 3.x.")
sys.exit(1)
from setuptools import setup, find_packages
setup(
name='signac',
version='0.1.7dev5',
packages=find_packages(),
author='Carl Simon Adorf',
author_email='csadorf@umich.edu',
description="Computational Database.",
keywords='simulation tools mc md monte-carlo mongodb '
'jobmanagement materials database',
classifiers=[
"Development Status :: 3 - Alpha",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Topic :: Scientific/Engineering :: Physics",
],
extras_require={
'db': ['pymongo>=3.0'],
'mpi': ['mpi4py'],
'conversion': ['networkx>=1.1.0'],
},
)
| <commit_before>import sys
IS_PYTHON3 = sys.version_info[0] == 3
if not IS_PYTHON3:
print("Error: signac requires python version >= 3.x.")
sys.exit(1)
from setuptools import setup, find_packages
setup(
name='signac',
version='0.1.7dev1',
packages=find_packages(),
author='Carl Simon Adorf',
author_email='csadorf@umich.edu',
description="Computational Database.",
keywords='simulation tools mc md monte-carlo mongodb '
'jobmanagement materials database',
classifiers=[
"Development Status :: 3 - Alpha",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Topic :: Scientific/Engineering :: Physics",
],
extras_require={
'db': ['pymongo>=3.0'],
'mpi': ['mpi4py'],
'conversion': ['networkx>=1.1.0'],
},
entry_points={
'console_scripts': [
'signac = signac.contrib.script:main',
'signac_init = signac.contrib.init_project:main',
'signac_configure = signac.contrib.configure:main',
'signac_admin = signac.admin.manage:main',
'signac_server = signac.contrib.server:main',
'signac_user = signac.contrib.admin:main',
'signac_admin_project = signac.admin.manage_project:main',
],
},
)
<commit_msg>Remove entry points and bump dev version.<commit_after> | import sys
IS_PYTHON3 = sys.version_info[0] == 3
if not IS_PYTHON3:
print("Error: signac requires python version >= 3.x.")
sys.exit(1)
from setuptools import setup, find_packages
setup(
name='signac',
version='0.1.7dev5',
packages=find_packages(),
author='Carl Simon Adorf',
author_email='csadorf@umich.edu',
description="Computational Database.",
keywords='simulation tools mc md monte-carlo mongodb '
'jobmanagement materials database',
classifiers=[
"Development Status :: 3 - Alpha",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Topic :: Scientific/Engineering :: Physics",
],
extras_require={
'db': ['pymongo>=3.0'],
'mpi': ['mpi4py'],
'conversion': ['networkx>=1.1.0'],
},
)
| import sys
IS_PYTHON3 = sys.version_info[0] == 3
if not IS_PYTHON3:
print("Error: signac requires python version >= 3.x.")
sys.exit(1)
from setuptools import setup, find_packages
setup(
name='signac',
version='0.1.7dev1',
packages=find_packages(),
author='Carl Simon Adorf',
author_email='csadorf@umich.edu',
description="Computational Database.",
keywords='simulation tools mc md monte-carlo mongodb '
'jobmanagement materials database',
classifiers=[
"Development Status :: 3 - Alpha",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Topic :: Scientific/Engineering :: Physics",
],
extras_require={
'db': ['pymongo>=3.0'],
'mpi': ['mpi4py'],
'conversion': ['networkx>=1.1.0'],
},
entry_points={
'console_scripts': [
'signac = signac.contrib.script:main',
'signac_init = signac.contrib.init_project:main',
'signac_configure = signac.contrib.configure:main',
'signac_admin = signac.admin.manage:main',
'signac_server = signac.contrib.server:main',
'signac_user = signac.contrib.admin:main',
'signac_admin_project = signac.admin.manage_project:main',
],
},
)
Remove entry points and bump dev version.import sys
IS_PYTHON3 = sys.version_info[0] == 3
if not IS_PYTHON3:
print("Error: signac requires python version >= 3.x.")
sys.exit(1)
from setuptools import setup, find_packages
setup(
name='signac',
version='0.1.7dev5',
packages=find_packages(),
author='Carl Simon Adorf',
author_email='csadorf@umich.edu',
description="Computational Database.",
keywords='simulation tools mc md monte-carlo mongodb '
'jobmanagement materials database',
classifiers=[
"Development Status :: 3 - Alpha",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Topic :: Scientific/Engineering :: Physics",
],
extras_require={
'db': ['pymongo>=3.0'],
'mpi': ['mpi4py'],
'conversion': ['networkx>=1.1.0'],
},
)
| <commit_before>import sys
IS_PYTHON3 = sys.version_info[0] == 3
if not IS_PYTHON3:
print("Error: signac requires python version >= 3.x.")
sys.exit(1)
from setuptools import setup, find_packages
setup(
name='signac',
version='0.1.7dev1',
packages=find_packages(),
author='Carl Simon Adorf',
author_email='csadorf@umich.edu',
description="Computational Database.",
keywords='simulation tools mc md monte-carlo mongodb '
'jobmanagement materials database',
classifiers=[
"Development Status :: 3 - Alpha",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Topic :: Scientific/Engineering :: Physics",
],
extras_require={
'db': ['pymongo>=3.0'],
'mpi': ['mpi4py'],
'conversion': ['networkx>=1.1.0'],
},
entry_points={
'console_scripts': [
'signac = signac.contrib.script:main',
'signac_init = signac.contrib.init_project:main',
'signac_configure = signac.contrib.configure:main',
'signac_admin = signac.admin.manage:main',
'signac_server = signac.contrib.server:main',
'signac_user = signac.contrib.admin:main',
'signac_admin_project = signac.admin.manage_project:main',
],
},
)
<commit_msg>Remove entry points and bump dev version.<commit_after>import sys
IS_PYTHON3 = sys.version_info[0] == 3
if not IS_PYTHON3:
print("Error: signac requires python version >= 3.x.")
sys.exit(1)
from setuptools import setup, find_packages
setup(
name='signac',
version='0.1.7dev5',
packages=find_packages(),
author='Carl Simon Adorf',
author_email='csadorf@umich.edu',
description="Computational Database.",
keywords='simulation tools mc md monte-carlo mongodb '
'jobmanagement materials database',
classifiers=[
"Development Status :: 3 - Alpha",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Topic :: Scientific/Engineering :: Physics",
],
extras_require={
'db': ['pymongo>=3.0'],
'mpi': ['mpi4py'],
'conversion': ['networkx>=1.1.0'],
},
)
|
dd062287f182c1a4d7d32c3db365c0ee92eb4120 | setup.py | setup.py | # -*- coding: utf-8 -*-
from setuptools import setup, find_packages
classifiers = [
"Development Status :: 4 - Beta",
"Intended Audience :: Developers",
"License :: OSI Approved :: Apache Software License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.3",
"Topic :: Database",
"Topic :: Software Development",
"Topic :: Software Development :: Testing",
]
setup(
name='test.mysql',
version='0.1.0',
description='automatically setups a mysqld instance in a temporary directory, and destroys it after testing',
long_description='',
classifiers=classifiers,
keywords=[],
author='Takeshi Komiya',
author_email='i.tkomiya at gmail.com',
url='http://bitbucket.org/tk0miya/test.mysql',
license='Apache License 2.0',
packages=find_packages('src'),
package_dir={'': 'src'},
package_data = {'': ['buildout.cfg']},
include_package_data=True,
install_requires=[
'pymysql',
],
extras_require=dict(
test=[
'Nose',
'pep8',
],
),
test_suite='nose.collector',
)
| # -*- coding: utf-8 -*-
from setuptools import setup, find_packages
classifiers = [
"Development Status :: 4 - Beta",
"Intended Audience :: Developers",
"License :: OSI Approved :: Apache Software License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.3",
"Topic :: Database",
"Topic :: Software Development",
"Topic :: Software Development :: Testing",
]
setup(
name='test.mysqld',
version='0.1.0',
description='automatically setups a mysqld instance in a temporary directory, and destroys it after testing',
long_description='',
classifiers=classifiers,
keywords=[],
author='Takeshi Komiya',
author_email='i.tkomiya at gmail.com',
url='http://bitbucket.org/tk0miya/test.mysql',
license='Apache License 2.0',
packages=find_packages('src'),
package_dir={'': 'src'},
package_data = {'': ['buildout.cfg']},
include_package_data=True,
install_requires=[
'pymysql',
],
extras_require=dict(
test=[
'Nose',
'pep8',
],
),
test_suite='nose.collector',
)
| Fix module name: test.mysql -> test.mysqld | Fix module name: test.mysql -> test.mysqld
| Python | apache-2.0 | tk0miya/testing.mysqld | # -*- coding: utf-8 -*-
from setuptools import setup, find_packages
classifiers = [
"Development Status :: 4 - Beta",
"Intended Audience :: Developers",
"License :: OSI Approved :: Apache Software License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.3",
"Topic :: Database",
"Topic :: Software Development",
"Topic :: Software Development :: Testing",
]
setup(
name='test.mysql',
version='0.1.0',
description='automatically setups a mysqld instance in a temporary directory, and destroys it after testing',
long_description='',
classifiers=classifiers,
keywords=[],
author='Takeshi Komiya',
author_email='i.tkomiya at gmail.com',
url='http://bitbucket.org/tk0miya/test.mysql',
license='Apache License 2.0',
packages=find_packages('src'),
package_dir={'': 'src'},
package_data = {'': ['buildout.cfg']},
include_package_data=True,
install_requires=[
'pymysql',
],
extras_require=dict(
test=[
'Nose',
'pep8',
],
),
test_suite='nose.collector',
)
Fix module name: test.mysql -> test.mysqld | # -*- coding: utf-8 -*-
from setuptools import setup, find_packages
classifiers = [
"Development Status :: 4 - Beta",
"Intended Audience :: Developers",
"License :: OSI Approved :: Apache Software License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.3",
"Topic :: Database",
"Topic :: Software Development",
"Topic :: Software Development :: Testing",
]
setup(
name='test.mysqld',
version='0.1.0',
description='automatically setups a mysqld instance in a temporary directory, and destroys it after testing',
long_description='',
classifiers=classifiers,
keywords=[],
author='Takeshi Komiya',
author_email='i.tkomiya at gmail.com',
url='http://bitbucket.org/tk0miya/test.mysql',
license='Apache License 2.0',
packages=find_packages('src'),
package_dir={'': 'src'},
package_data = {'': ['buildout.cfg']},
include_package_data=True,
install_requires=[
'pymysql',
],
extras_require=dict(
test=[
'Nose',
'pep8',
],
),
test_suite='nose.collector',
)
| <commit_before># -*- coding: utf-8 -*-
from setuptools import setup, find_packages
classifiers = [
"Development Status :: 4 - Beta",
"Intended Audience :: Developers",
"License :: OSI Approved :: Apache Software License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.3",
"Topic :: Database",
"Topic :: Software Development",
"Topic :: Software Development :: Testing",
]
setup(
name='test.mysql',
version='0.1.0',
description='automatically setups a mysqld instance in a temporary directory, and destroys it after testing',
long_description='',
classifiers=classifiers,
keywords=[],
author='Takeshi Komiya',
author_email='i.tkomiya at gmail.com',
url='http://bitbucket.org/tk0miya/test.mysql',
license='Apache License 2.0',
packages=find_packages('src'),
package_dir={'': 'src'},
package_data = {'': ['buildout.cfg']},
include_package_data=True,
install_requires=[
'pymysql',
],
extras_require=dict(
test=[
'Nose',
'pep8',
],
),
test_suite='nose.collector',
)
<commit_msg>Fix module name: test.mysql -> test.mysqld<commit_after> | # -*- coding: utf-8 -*-
from setuptools import setup, find_packages
classifiers = [
"Development Status :: 4 - Beta",
"Intended Audience :: Developers",
"License :: OSI Approved :: Apache Software License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.3",
"Topic :: Database",
"Topic :: Software Development",
"Topic :: Software Development :: Testing",
]
setup(
name='test.mysqld',
version='0.1.0',
description='automatically setups a mysqld instance in a temporary directory, and destroys it after testing',
long_description='',
classifiers=classifiers,
keywords=[],
author='Takeshi Komiya',
author_email='i.tkomiya at gmail.com',
url='http://bitbucket.org/tk0miya/test.mysql',
license='Apache License 2.0',
packages=find_packages('src'),
package_dir={'': 'src'},
package_data = {'': ['buildout.cfg']},
include_package_data=True,
install_requires=[
'pymysql',
],
extras_require=dict(
test=[
'Nose',
'pep8',
],
),
test_suite='nose.collector',
)
| # -*- coding: utf-8 -*-
from setuptools import setup, find_packages
classifiers = [
"Development Status :: 4 - Beta",
"Intended Audience :: Developers",
"License :: OSI Approved :: Apache Software License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.3",
"Topic :: Database",
"Topic :: Software Development",
"Topic :: Software Development :: Testing",
]
setup(
name='test.mysql',
version='0.1.0',
description='automatically setups a mysqld instance in a temporary directory, and destroys it after testing',
long_description='',
classifiers=classifiers,
keywords=[],
author='Takeshi Komiya',
author_email='i.tkomiya at gmail.com',
url='http://bitbucket.org/tk0miya/test.mysql',
license='Apache License 2.0',
packages=find_packages('src'),
package_dir={'': 'src'},
package_data = {'': ['buildout.cfg']},
include_package_data=True,
install_requires=[
'pymysql',
],
extras_require=dict(
test=[
'Nose',
'pep8',
],
),
test_suite='nose.collector',
)
Fix module name: test.mysql -> test.mysqld# -*- coding: utf-8 -*-
from setuptools import setup, find_packages
classifiers = [
"Development Status :: 4 - Beta",
"Intended Audience :: Developers",
"License :: OSI Approved :: Apache Software License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.3",
"Topic :: Database",
"Topic :: Software Development",
"Topic :: Software Development :: Testing",
]
setup(
name='test.mysqld',
version='0.1.0',
description='automatically setups a mysqld instance in a temporary directory, and destroys it after testing',
long_description='',
classifiers=classifiers,
keywords=[],
author='Takeshi Komiya',
author_email='i.tkomiya at gmail.com',
url='http://bitbucket.org/tk0miya/test.mysql',
license='Apache License 2.0',
packages=find_packages('src'),
package_dir={'': 'src'},
package_data = {'': ['buildout.cfg']},
include_package_data=True,
install_requires=[
'pymysql',
],
extras_require=dict(
test=[
'Nose',
'pep8',
],
),
test_suite='nose.collector',
)
| <commit_before># -*- coding: utf-8 -*-
from setuptools import setup, find_packages
classifiers = [
"Development Status :: 4 - Beta",
"Intended Audience :: Developers",
"License :: OSI Approved :: Apache Software License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.3",
"Topic :: Database",
"Topic :: Software Development",
"Topic :: Software Development :: Testing",
]
setup(
name='test.mysql',
version='0.1.0',
description='automatically setups a mysqld instance in a temporary directory, and destroys it after testing',
long_description='',
classifiers=classifiers,
keywords=[],
author='Takeshi Komiya',
author_email='i.tkomiya at gmail.com',
url='http://bitbucket.org/tk0miya/test.mysql',
license='Apache License 2.0',
packages=find_packages('src'),
package_dir={'': 'src'},
package_data = {'': ['buildout.cfg']},
include_package_data=True,
install_requires=[
'pymysql',
],
extras_require=dict(
test=[
'Nose',
'pep8',
],
),
test_suite='nose.collector',
)
<commit_msg>Fix module name: test.mysql -> test.mysqld<commit_after># -*- coding: utf-8 -*-
from setuptools import setup, find_packages
classifiers = [
"Development Status :: 4 - Beta",
"Intended Audience :: Developers",
"License :: OSI Approved :: Apache Software License",
"Programming Language :: Python",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3.3",
"Topic :: Database",
"Topic :: Software Development",
"Topic :: Software Development :: Testing",
]
setup(
name='test.mysqld',
version='0.1.0',
description='automatically setups a mysqld instance in a temporary directory, and destroys it after testing',
long_description='',
classifiers=classifiers,
keywords=[],
author='Takeshi Komiya',
author_email='i.tkomiya at gmail.com',
url='http://bitbucket.org/tk0miya/test.mysql',
license='Apache License 2.0',
packages=find_packages('src'),
package_dir={'': 'src'},
package_data = {'': ['buildout.cfg']},
include_package_data=True,
install_requires=[
'pymysql',
],
extras_require=dict(
test=[
'Nose',
'pep8',
],
),
test_suite='nose.collector',
)
|
04fb3d2a7d9416fe91b069b92d8fa157ea3d657b | setup.py | setup.py | from __future__ import absolute_import
from setuptools import setup, find_packages
setup(
name='flyingcloud',
version='0.1.9',
description='Build Docker images using SaltStack',
author='CookBrite, Inc.',
author_email='flyingcloud-admin@cookbrite.com',
license='Apache Software License 2.0',
url='https://github.com/cookbrite/flyingcloud',
packages=find_packages(exclude='tests'),
install_requires=['docker-py', 'sh', 'six'],
classifiers=['Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: Apache Software License',
'Operating System :: POSIX :: Linux',
'Topic :: Utilities',
'Topic :: System :: Clustering',
'Topic :: System :: Distributed Computing'],
long_description=open('README.rst').read(),
keywords="docker saltstack devops automation",
)
| from __future__ import absolute_import
from setuptools import setup, find_packages
setup(
name='flyingcloud',
version='0.1.9',
description='Build Docker images using SaltStack',
author='CookBrite, Inc.',
author_email='flyingcloud-admin@cookbrite.com',
license='Apache Software License 2.0',
url='https://github.com/cookbrite/flyingcloud',
packages=find_packages(exclude='tests'),
install_requires=['docker-py', 'requests', 'sh', 'six'],
classifiers=['Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: Apache Software License',
'Operating System :: POSIX :: Linux',
'Topic :: Utilities',
'Topic :: System :: Clustering',
'Topic :: System :: Distributed Computing'],
long_description=open('README.rst').read(),
keywords="docker saltstack devops automation",
)
| Add requests as a dependency | Add requests as a dependency
| Python | apache-2.0 | cookbrite/flyingcloud,cookbrite/flyingcloud | from __future__ import absolute_import
from setuptools import setup, find_packages
setup(
name='flyingcloud',
version='0.1.9',
description='Build Docker images using SaltStack',
author='CookBrite, Inc.',
author_email='flyingcloud-admin@cookbrite.com',
license='Apache Software License 2.0',
url='https://github.com/cookbrite/flyingcloud',
packages=find_packages(exclude='tests'),
install_requires=['docker-py', 'sh', 'six'],
classifiers=['Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: Apache Software License',
'Operating System :: POSIX :: Linux',
'Topic :: Utilities',
'Topic :: System :: Clustering',
'Topic :: System :: Distributed Computing'],
long_description=open('README.rst').read(),
keywords="docker saltstack devops automation",
)
Add requests as a dependency | from __future__ import absolute_import
from setuptools import setup, find_packages
setup(
name='flyingcloud',
version='0.1.9',
description='Build Docker images using SaltStack',
author='CookBrite, Inc.',
author_email='flyingcloud-admin@cookbrite.com',
license='Apache Software License 2.0',
url='https://github.com/cookbrite/flyingcloud',
packages=find_packages(exclude='tests'),
install_requires=['docker-py', 'requests', 'sh', 'six'],
classifiers=['Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: Apache Software License',
'Operating System :: POSIX :: Linux',
'Topic :: Utilities',
'Topic :: System :: Clustering',
'Topic :: System :: Distributed Computing'],
long_description=open('README.rst').read(),
keywords="docker saltstack devops automation",
)
| <commit_before>from __future__ import absolute_import
from setuptools import setup, find_packages
setup(
name='flyingcloud',
version='0.1.9',
description='Build Docker images using SaltStack',
author='CookBrite, Inc.',
author_email='flyingcloud-admin@cookbrite.com',
license='Apache Software License 2.0',
url='https://github.com/cookbrite/flyingcloud',
packages=find_packages(exclude='tests'),
install_requires=['docker-py', 'sh', 'six'],
classifiers=['Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: Apache Software License',
'Operating System :: POSIX :: Linux',
'Topic :: Utilities',
'Topic :: System :: Clustering',
'Topic :: System :: Distributed Computing'],
long_description=open('README.rst').read(),
keywords="docker saltstack devops automation",
)
<commit_msg>Add requests as a dependency<commit_after> | from __future__ import absolute_import
from setuptools import setup, find_packages
setup(
name='flyingcloud',
version='0.1.9',
description='Build Docker images using SaltStack',
author='CookBrite, Inc.',
author_email='flyingcloud-admin@cookbrite.com',
license='Apache Software License 2.0',
url='https://github.com/cookbrite/flyingcloud',
packages=find_packages(exclude='tests'),
install_requires=['docker-py', 'requests', 'sh', 'six'],
classifiers=['Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: Apache Software License',
'Operating System :: POSIX :: Linux',
'Topic :: Utilities',
'Topic :: System :: Clustering',
'Topic :: System :: Distributed Computing'],
long_description=open('README.rst').read(),
keywords="docker saltstack devops automation",
)
| from __future__ import absolute_import
from setuptools import setup, find_packages
setup(
name='flyingcloud',
version='0.1.9',
description='Build Docker images using SaltStack',
author='CookBrite, Inc.',
author_email='flyingcloud-admin@cookbrite.com',
license='Apache Software License 2.0',
url='https://github.com/cookbrite/flyingcloud',
packages=find_packages(exclude='tests'),
install_requires=['docker-py', 'sh', 'six'],
classifiers=['Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: Apache Software License',
'Operating System :: POSIX :: Linux',
'Topic :: Utilities',
'Topic :: System :: Clustering',
'Topic :: System :: Distributed Computing'],
long_description=open('README.rst').read(),
keywords="docker saltstack devops automation",
)
Add requests as a dependencyfrom __future__ import absolute_import
from setuptools import setup, find_packages
setup(
name='flyingcloud',
version='0.1.9',
description='Build Docker images using SaltStack',
author='CookBrite, Inc.',
author_email='flyingcloud-admin@cookbrite.com',
license='Apache Software License 2.0',
url='https://github.com/cookbrite/flyingcloud',
packages=find_packages(exclude='tests'),
install_requires=['docker-py', 'requests', 'sh', 'six'],
classifiers=['Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: Apache Software License',
'Operating System :: POSIX :: Linux',
'Topic :: Utilities',
'Topic :: System :: Clustering',
'Topic :: System :: Distributed Computing'],
long_description=open('README.rst').read(),
keywords="docker saltstack devops automation",
)
| <commit_before>from __future__ import absolute_import
from setuptools import setup, find_packages
setup(
name='flyingcloud',
version='0.1.9',
description='Build Docker images using SaltStack',
author='CookBrite, Inc.',
author_email='flyingcloud-admin@cookbrite.com',
license='Apache Software License 2.0',
url='https://github.com/cookbrite/flyingcloud',
packages=find_packages(exclude='tests'),
install_requires=['docker-py', 'sh', 'six'],
classifiers=['Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: Apache Software License',
'Operating System :: POSIX :: Linux',
'Topic :: Utilities',
'Topic :: System :: Clustering',
'Topic :: System :: Distributed Computing'],
long_description=open('README.rst').read(),
keywords="docker saltstack devops automation",
)
<commit_msg>Add requests as a dependency<commit_after>from __future__ import absolute_import
from setuptools import setup, find_packages
setup(
name='flyingcloud',
version='0.1.9',
description='Build Docker images using SaltStack',
author='CookBrite, Inc.',
author_email='flyingcloud-admin@cookbrite.com',
license='Apache Software License 2.0',
url='https://github.com/cookbrite/flyingcloud',
packages=find_packages(exclude='tests'),
install_requires=['docker-py', 'requests', 'sh', 'six'],
classifiers=['Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Intended Audience :: Developers',
'Intended Audience :: Information Technology',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: Apache Software License',
'Operating System :: POSIX :: Linux',
'Topic :: Utilities',
'Topic :: System :: Clustering',
'Topic :: System :: Distributed Computing'],
long_description=open('README.rst').read(),
keywords="docker saltstack devops automation",
)
|
065cc9310a93af251db7a8464f6562ef62e3961c | setup.py | setup.py | #! /usr/bin/env python
# encoding: utf-8
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
REQUIREMENTS = [i.strip() for i in open("requirements.txt").readlines()]
setup(
name='pagseguro-sdk',
version="0.1.0",
description='SDK para utilização do PagSeguro em Python',
url='https://pagseguro-sdk.readthedocs.com/',
author='Jean O. Rodrigues',
author_email='github@jean.bz',
download_url='https://github.com/jeanmask/pagseguro-sdk/archive/v0.1.0.tar.gz',
license='MIT',
packages=['pagseguro'],
install_requires=REQUIREMENTS,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Environment :: Web Environment',
'Intended Audience :: End Users/Desktop',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
| #! /usr/bin/env python
# encoding: utf-8
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
REQUIREMENTS = [i.strip() for i in open("requirements.txt").readlines()]
setup(
name='pagseguro-sdk',
version="0.1.0",
description='SDK para utilização do PagSeguro em Python',
url='http://pagseguro-sdk.readthedocs.org/',
author='Jean O. Rodrigues',
author_email='github@jean.bz',
download_url='https://github.com/jeanmask/pagseguro-sdk/archive/v0.1.0.tar.gz',
license='MIT',
packages=['pagseguro'],
install_requires=REQUIREMENTS,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Environment :: Web Environment',
'Intended Audience :: End Users/Desktop',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
| Fix read the docs url. | Fix read the docs url.
| Python | mit | jeanmask/pagseguro-sdk | #! /usr/bin/env python
# encoding: utf-8
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
REQUIREMENTS = [i.strip() for i in open("requirements.txt").readlines()]
setup(
name='pagseguro-sdk',
version="0.1.0",
description='SDK para utilização do PagSeguro em Python',
url='https://pagseguro-sdk.readthedocs.com/',
author='Jean O. Rodrigues',
author_email='github@jean.bz',
download_url='https://github.com/jeanmask/pagseguro-sdk/archive/v0.1.0.tar.gz',
license='MIT',
packages=['pagseguro'],
install_requires=REQUIREMENTS,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Environment :: Web Environment',
'Intended Audience :: End Users/Desktop',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
Fix read the docs url. | #! /usr/bin/env python
# encoding: utf-8
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
REQUIREMENTS = [i.strip() for i in open("requirements.txt").readlines()]
setup(
name='pagseguro-sdk',
version="0.1.0",
description='SDK para utilização do PagSeguro em Python',
url='http://pagseguro-sdk.readthedocs.org/',
author='Jean O. Rodrigues',
author_email='github@jean.bz',
download_url='https://github.com/jeanmask/pagseguro-sdk/archive/v0.1.0.tar.gz',
license='MIT',
packages=['pagseguro'],
install_requires=REQUIREMENTS,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Environment :: Web Environment',
'Intended Audience :: End Users/Desktop',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
| <commit_before>#! /usr/bin/env python
# encoding: utf-8
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
REQUIREMENTS = [i.strip() for i in open("requirements.txt").readlines()]
setup(
name='pagseguro-sdk',
version="0.1.0",
description='SDK para utilização do PagSeguro em Python',
url='https://pagseguro-sdk.readthedocs.com/',
author='Jean O. Rodrigues',
author_email='github@jean.bz',
download_url='https://github.com/jeanmask/pagseguro-sdk/archive/v0.1.0.tar.gz',
license='MIT',
packages=['pagseguro'],
install_requires=REQUIREMENTS,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Environment :: Web Environment',
'Intended Audience :: End Users/Desktop',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
<commit_msg>Fix read the docs url.<commit_after> | #! /usr/bin/env python
# encoding: utf-8
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
REQUIREMENTS = [i.strip() for i in open("requirements.txt").readlines()]
setup(
name='pagseguro-sdk',
version="0.1.0",
description='SDK para utilização do PagSeguro em Python',
url='http://pagseguro-sdk.readthedocs.org/',
author='Jean O. Rodrigues',
author_email='github@jean.bz',
download_url='https://github.com/jeanmask/pagseguro-sdk/archive/v0.1.0.tar.gz',
license='MIT',
packages=['pagseguro'],
install_requires=REQUIREMENTS,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Environment :: Web Environment',
'Intended Audience :: End Users/Desktop',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
| #! /usr/bin/env python
# encoding: utf-8
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
REQUIREMENTS = [i.strip() for i in open("requirements.txt").readlines()]
setup(
name='pagseguro-sdk',
version="0.1.0",
description='SDK para utilização do PagSeguro em Python',
url='https://pagseguro-sdk.readthedocs.com/',
author='Jean O. Rodrigues',
author_email='github@jean.bz',
download_url='https://github.com/jeanmask/pagseguro-sdk/archive/v0.1.0.tar.gz',
license='MIT',
packages=['pagseguro'],
install_requires=REQUIREMENTS,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Environment :: Web Environment',
'Intended Audience :: End Users/Desktop',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
Fix read the docs url.#! /usr/bin/env python
# encoding: utf-8
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
REQUIREMENTS = [i.strip() for i in open("requirements.txt").readlines()]
setup(
name='pagseguro-sdk',
version="0.1.0",
description='SDK para utilização do PagSeguro em Python',
url='http://pagseguro-sdk.readthedocs.org/',
author='Jean O. Rodrigues',
author_email='github@jean.bz',
download_url='https://github.com/jeanmask/pagseguro-sdk/archive/v0.1.0.tar.gz',
license='MIT',
packages=['pagseguro'],
install_requires=REQUIREMENTS,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Environment :: Web Environment',
'Intended Audience :: End Users/Desktop',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
| <commit_before>#! /usr/bin/env python
# encoding: utf-8
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
REQUIREMENTS = [i.strip() for i in open("requirements.txt").readlines()]
setup(
name='pagseguro-sdk',
version="0.1.0",
description='SDK para utilização do PagSeguro em Python',
url='https://pagseguro-sdk.readthedocs.com/',
author='Jean O. Rodrigues',
author_email='github@jean.bz',
download_url='https://github.com/jeanmask/pagseguro-sdk/archive/v0.1.0.tar.gz',
license='MIT',
packages=['pagseguro'],
install_requires=REQUIREMENTS,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Environment :: Web Environment',
'Intended Audience :: End Users/Desktop',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
<commit_msg>Fix read the docs url.<commit_after>#! /usr/bin/env python
# encoding: utf-8
try:
from setuptools import setup
except ImportError:
from distutils.core import setup
REQUIREMENTS = [i.strip() for i in open("requirements.txt").readlines()]
setup(
name='pagseguro-sdk',
version="0.1.0",
description='SDK para utilização do PagSeguro em Python',
url='http://pagseguro-sdk.readthedocs.org/',
author='Jean O. Rodrigues',
author_email='github@jean.bz',
download_url='https://github.com/jeanmask/pagseguro-sdk/archive/v0.1.0.tar.gz',
license='MIT',
packages=['pagseguro'],
install_requires=REQUIREMENTS,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Environment :: Web Environment',
'Intended Audience :: End Users/Desktop',
'Intended Audience :: Developers',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet',
'Topic :: Software Development :: Libraries :: Python Modules',
],
)
|
abb49ae96786018c7a8a8cd3c8b30d612f710ed2 | setup.py | setup.py | """Rachiopy setup script."""
from setuptools import find_packages, setup
VERSION = "1.0.3"
GITHUB_USERNAME = "rfverbruggen"
GITHUB_REPOSITORY = "rachiopy"
GITHUB_PATH = f"{GITHUB_USERNAME}/{GITHUB_REPOSITORY}"
GITHUB_URL = f"https://github.com/{GITHUB_PATH}"
DOWNLOAD_URL = f"{GITHUB_URL}/archive/{VERSION}.tar.gz"
PROJECT_URLS = {"Bug Reports": f"{GITHUB_URL}/issues"}
PACKAGES = find_packages(exclude=["tests", "tests.*"])
setup(
name="RachioPy",
version=VERSION,
author="Robbert Verbruggen",
author_email="rfverbruggen@icloud.com",
packages=PACKAGES,
install_requires=["requests"],
url=GITHUB_URL,
download_url=DOWNLOAD_URL,
project_urls=PROJECT_URLS,
license="MIT",
description="A Python module for the Rachio API.",
platforms="Cross Platform",
classifiers=[
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python :: 3",
"Topic :: Software Development",
],
)
| """Rachiopy setup script."""
from setuptools import find_packages, setup
from datetime import datetime
NOW = datetime.now().strftime("%m%d%Y%H%M%S")
VERSION = f"1.0.4-dev{NOW}"
GITHUB_USERNAME = "rfverbruggen"
GITHUB_REPOSITORY = "rachiopy"
GITHUB_PATH = f"{GITHUB_USERNAME}/{GITHUB_REPOSITORY}"
GITHUB_URL = f"https://github.com/{GITHUB_PATH}"
DOWNLOAD_URL = f"{GITHUB_URL}/archive/{VERSION}.tar.gz"
PROJECT_URLS = {"Bug Reports": f"{GITHUB_URL}/issues"}
PACKAGES = find_packages(exclude=["tests", "tests.*"])
setup(
name="RachioPy",
version=VERSION,
author="Robbert Verbruggen",
author_email="rfverbruggen@icloud.com",
packages=PACKAGES,
install_requires=["requests"],
url=GITHUB_URL,
download_url=DOWNLOAD_URL,
project_urls=PROJECT_URLS,
license="MIT",
description="A Python module for the Rachio API.",
platforms="Cross Platform",
classifiers=[
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python :: 3",
"Topic :: Software Development",
],
)
| Set the next dev version number: 1.0.4-dev | Set the next dev version number: 1.0.4-dev | Python | mit | rfverbruggen/rachiopy | """Rachiopy setup script."""
from setuptools import find_packages, setup
VERSION = "1.0.3"
GITHUB_USERNAME = "rfverbruggen"
GITHUB_REPOSITORY = "rachiopy"
GITHUB_PATH = f"{GITHUB_USERNAME}/{GITHUB_REPOSITORY}"
GITHUB_URL = f"https://github.com/{GITHUB_PATH}"
DOWNLOAD_URL = f"{GITHUB_URL}/archive/{VERSION}.tar.gz"
PROJECT_URLS = {"Bug Reports": f"{GITHUB_URL}/issues"}
PACKAGES = find_packages(exclude=["tests", "tests.*"])
setup(
name="RachioPy",
version=VERSION,
author="Robbert Verbruggen",
author_email="rfverbruggen@icloud.com",
packages=PACKAGES,
install_requires=["requests"],
url=GITHUB_URL,
download_url=DOWNLOAD_URL,
project_urls=PROJECT_URLS,
license="MIT",
description="A Python module for the Rachio API.",
platforms="Cross Platform",
classifiers=[
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python :: 3",
"Topic :: Software Development",
],
)
Set the next dev version number: 1.0.4-dev | """Rachiopy setup script."""
from setuptools import find_packages, setup
from datetime import datetime
NOW = datetime.now().strftime("%m%d%Y%H%M%S")
VERSION = f"1.0.4-dev{NOW}"
GITHUB_USERNAME = "rfverbruggen"
GITHUB_REPOSITORY = "rachiopy"
GITHUB_PATH = f"{GITHUB_USERNAME}/{GITHUB_REPOSITORY}"
GITHUB_URL = f"https://github.com/{GITHUB_PATH}"
DOWNLOAD_URL = f"{GITHUB_URL}/archive/{VERSION}.tar.gz"
PROJECT_URLS = {"Bug Reports": f"{GITHUB_URL}/issues"}
PACKAGES = find_packages(exclude=["tests", "tests.*"])
setup(
name="RachioPy",
version=VERSION,
author="Robbert Verbruggen",
author_email="rfverbruggen@icloud.com",
packages=PACKAGES,
install_requires=["requests"],
url=GITHUB_URL,
download_url=DOWNLOAD_URL,
project_urls=PROJECT_URLS,
license="MIT",
description="A Python module for the Rachio API.",
platforms="Cross Platform",
classifiers=[
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python :: 3",
"Topic :: Software Development",
],
)
| <commit_before>"""Rachiopy setup script."""
from setuptools import find_packages, setup
VERSION = "1.0.3"
GITHUB_USERNAME = "rfverbruggen"
GITHUB_REPOSITORY = "rachiopy"
GITHUB_PATH = f"{GITHUB_USERNAME}/{GITHUB_REPOSITORY}"
GITHUB_URL = f"https://github.com/{GITHUB_PATH}"
DOWNLOAD_URL = f"{GITHUB_URL}/archive/{VERSION}.tar.gz"
PROJECT_URLS = {"Bug Reports": f"{GITHUB_URL}/issues"}
PACKAGES = find_packages(exclude=["tests", "tests.*"])
setup(
name="RachioPy",
version=VERSION,
author="Robbert Verbruggen",
author_email="rfverbruggen@icloud.com",
packages=PACKAGES,
install_requires=["requests"],
url=GITHUB_URL,
download_url=DOWNLOAD_URL,
project_urls=PROJECT_URLS,
license="MIT",
description="A Python module for the Rachio API.",
platforms="Cross Platform",
classifiers=[
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python :: 3",
"Topic :: Software Development",
],
)
<commit_msg>Set the next dev version number: 1.0.4-dev<commit_after> | """Rachiopy setup script."""
from setuptools import find_packages, setup
from datetime import datetime
NOW = datetime.now().strftime("%m%d%Y%H%M%S")
VERSION = f"1.0.4-dev{NOW}"
GITHUB_USERNAME = "rfverbruggen"
GITHUB_REPOSITORY = "rachiopy"
GITHUB_PATH = f"{GITHUB_USERNAME}/{GITHUB_REPOSITORY}"
GITHUB_URL = f"https://github.com/{GITHUB_PATH}"
DOWNLOAD_URL = f"{GITHUB_URL}/archive/{VERSION}.tar.gz"
PROJECT_URLS = {"Bug Reports": f"{GITHUB_URL}/issues"}
PACKAGES = find_packages(exclude=["tests", "tests.*"])
setup(
name="RachioPy",
version=VERSION,
author="Robbert Verbruggen",
author_email="rfverbruggen@icloud.com",
packages=PACKAGES,
install_requires=["requests"],
url=GITHUB_URL,
download_url=DOWNLOAD_URL,
project_urls=PROJECT_URLS,
license="MIT",
description="A Python module for the Rachio API.",
platforms="Cross Platform",
classifiers=[
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python :: 3",
"Topic :: Software Development",
],
)
| """Rachiopy setup script."""
from setuptools import find_packages, setup
VERSION = "1.0.3"
GITHUB_USERNAME = "rfverbruggen"
GITHUB_REPOSITORY = "rachiopy"
GITHUB_PATH = f"{GITHUB_USERNAME}/{GITHUB_REPOSITORY}"
GITHUB_URL = f"https://github.com/{GITHUB_PATH}"
DOWNLOAD_URL = f"{GITHUB_URL}/archive/{VERSION}.tar.gz"
PROJECT_URLS = {"Bug Reports": f"{GITHUB_URL}/issues"}
PACKAGES = find_packages(exclude=["tests", "tests.*"])
setup(
name="RachioPy",
version=VERSION,
author="Robbert Verbruggen",
author_email="rfverbruggen@icloud.com",
packages=PACKAGES,
install_requires=["requests"],
url=GITHUB_URL,
download_url=DOWNLOAD_URL,
project_urls=PROJECT_URLS,
license="MIT",
description="A Python module for the Rachio API.",
platforms="Cross Platform",
classifiers=[
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python :: 3",
"Topic :: Software Development",
],
)
Set the next dev version number: 1.0.4-dev"""Rachiopy setup script."""
from setuptools import find_packages, setup
from datetime import datetime
NOW = datetime.now().strftime("%m%d%Y%H%M%S")
VERSION = f"1.0.4-dev{NOW}"
GITHUB_USERNAME = "rfverbruggen"
GITHUB_REPOSITORY = "rachiopy"
GITHUB_PATH = f"{GITHUB_USERNAME}/{GITHUB_REPOSITORY}"
GITHUB_URL = f"https://github.com/{GITHUB_PATH}"
DOWNLOAD_URL = f"{GITHUB_URL}/archive/{VERSION}.tar.gz"
PROJECT_URLS = {"Bug Reports": f"{GITHUB_URL}/issues"}
PACKAGES = find_packages(exclude=["tests", "tests.*"])
setup(
name="RachioPy",
version=VERSION,
author="Robbert Verbruggen",
author_email="rfverbruggen@icloud.com",
packages=PACKAGES,
install_requires=["requests"],
url=GITHUB_URL,
download_url=DOWNLOAD_URL,
project_urls=PROJECT_URLS,
license="MIT",
description="A Python module for the Rachio API.",
platforms="Cross Platform",
classifiers=[
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python :: 3",
"Topic :: Software Development",
],
)
| <commit_before>"""Rachiopy setup script."""
from setuptools import find_packages, setup
VERSION = "1.0.3"
GITHUB_USERNAME = "rfverbruggen"
GITHUB_REPOSITORY = "rachiopy"
GITHUB_PATH = f"{GITHUB_USERNAME}/{GITHUB_REPOSITORY}"
GITHUB_URL = f"https://github.com/{GITHUB_PATH}"
DOWNLOAD_URL = f"{GITHUB_URL}/archive/{VERSION}.tar.gz"
PROJECT_URLS = {"Bug Reports": f"{GITHUB_URL}/issues"}
PACKAGES = find_packages(exclude=["tests", "tests.*"])
setup(
name="RachioPy",
version=VERSION,
author="Robbert Verbruggen",
author_email="rfverbruggen@icloud.com",
packages=PACKAGES,
install_requires=["requests"],
url=GITHUB_URL,
download_url=DOWNLOAD_URL,
project_urls=PROJECT_URLS,
license="MIT",
description="A Python module for the Rachio API.",
platforms="Cross Platform",
classifiers=[
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python :: 3",
"Topic :: Software Development",
],
)
<commit_msg>Set the next dev version number: 1.0.4-dev<commit_after>"""Rachiopy setup script."""
from setuptools import find_packages, setup
from datetime import datetime
NOW = datetime.now().strftime("%m%d%Y%H%M%S")
VERSION = f"1.0.4-dev{NOW}"
GITHUB_USERNAME = "rfverbruggen"
GITHUB_REPOSITORY = "rachiopy"
GITHUB_PATH = f"{GITHUB_USERNAME}/{GITHUB_REPOSITORY}"
GITHUB_URL = f"https://github.com/{GITHUB_PATH}"
DOWNLOAD_URL = f"{GITHUB_URL}/archive/{VERSION}.tar.gz"
PROJECT_URLS = {"Bug Reports": f"{GITHUB_URL}/issues"}
PACKAGES = find_packages(exclude=["tests", "tests.*"])
setup(
name="RachioPy",
version=VERSION,
author="Robbert Verbruggen",
author_email="rfverbruggen@icloud.com",
packages=PACKAGES,
install_requires=["requests"],
url=GITHUB_URL,
download_url=DOWNLOAD_URL,
project_urls=PROJECT_URLS,
license="MIT",
description="A Python module for the Rachio API.",
platforms="Cross Platform",
classifiers=[
"Development Status :: 5 - Production/Stable",
"Intended Audience :: Developers",
"License :: OSI Approved :: MIT License",
"Operating System :: OS Independent",
"Programming Language :: Python :: 3",
"Topic :: Software Development",
],
)
|
de5fbf7d63245e9d14844e66fdf16f88dbfae2e5 | rest_framework/authtoken/migrations/0001_initial.py | rest_framework/authtoken/migrations/0001_initial.py |
from __future__ import unicode_literals
from django.db import models, migrations
from django.conf import settings
class Migration(migrations.Migration):
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
]
operations = [
migrations.CreateModel(
name='Token',
fields=[
('key', models.CharField(max_length=40, serialize=False, primary_key=True)),
('created', models.DateTimeField(auto_now_add=True)),
('user', models.OneToOneField(related_name=b'auth_token', to=settings.AUTH_USER_MODEL)),
],
options={
},
bases=(models.Model,),
),
]
| # -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import models, migrations
from django.conf import settings
class Migration(migrations.Migration):
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
]
operations = [
migrations.CreateModel(
name='Token',
fields=[
('key', models.CharField(primary_key=True, serialize=False, max_length=40)),
('created', models.DateTimeField(auto_now_add=True)),
('user', models.OneToOneField(to=settings.AUTH_USER_MODEL, related_name='auth_token')),
],
options={
},
bases=(models.Model,),
),
]
| Update initial migration to work on Python 3 | Update initial migration to work on Python 3 | Python | bsd-2-clause | ajaali/django-rest-framework,mgaitan/django-rest-framework,xiaotangyuan/django-rest-framework,wedaly/django-rest-framework,edx/django-rest-framework,uploadcare/django-rest-framework,brandoncazander/django-rest-framework,mgaitan/django-rest-framework,werthen/django-rest-framework,akalipetis/django-rest-framework,YBJAY00000/django-rest-framework,douwevandermeij/django-rest-framework,elim/django-rest-framework,jpulec/django-rest-framework,abdulhaq-e/django-rest-framework,waytai/django-rest-framework,sbellem/django-rest-framework,justanr/django-rest-framework,akalipetis/django-rest-framework,antonyc/django-rest-framework,simudream/django-rest-framework,gregmuellegger/django-rest-framework,AlexandreProenca/django-rest-framework,aericson/django-rest-framework,jpadilla/django-rest-framework,johnraz/django-rest-framework,fishky/django-rest-framework,jerryhebert/django-rest-framework,sheppard/django-rest-framework,wwj718/django-rest-framework,raphaelmerx/django-rest-framework,cyberj/django-rest-framework,kgeorgy/django-rest-framework,justanr/django-rest-framework,cheif/django-rest-framework,leeahoward/django-rest-framework,alacritythief/django-rest-framework,cyberj/django-rest-framework,ebsaral/django-rest-framework,MJafarMashhadi/django-rest-framework,nryoung/django-rest-framework,rubendura/django-rest-framework,jness/django-rest-framework,James1345/django-rest-framework,jtiai/django-rest-framework,qsorix/django-rest-framework,raphaelmerx/django-rest-framework,kgeorgy/django-rest-framework,rhblind/django-rest-framework,hunter007/django-rest-framework,davesque/django-rest-framework,jtiai/django-rest-framework,bluedazzle/django-rest-framework,gregmuellegger/django-rest-framework,hnakamur/django-rest-framework,adambain-vokal/django-rest-framework,zeldalink0515/django-rest-framework,tomchristie/django-rest-framework,tigeraniya/django-rest-framework,cheif/django-rest-framework,lubomir/django-rest-framework,iheitlager/django-rest-framework,AlexandreProenca/django-rest-framework,andriy-s/django-rest-framework,VishvajitP/django-rest-framework,krinart/django-rest-framework,abdulhaq-e/django-rest-framework,zeldalink0515/django-rest-framework,ambivalentno/django-rest-framework,ashishfinoit/django-rest-framework,gregmuellegger/django-rest-framework,canassa/django-rest-framework,YBJAY00000/django-rest-framework,vstoykov/django-rest-framework,alacritythief/django-rest-framework,ajaali/django-rest-framework,hunter007/django-rest-framework,uruz/django-rest-framework,delinhabit/django-rest-framework,lubomir/django-rest-framework,pombredanne/django-rest-framework,leeahoward/django-rest-framework,mgaitan/django-rest-framework,tcroiset/django-rest-framework,adambain-vokal/django-rest-framework,d0ugal/django-rest-framework,nryoung/django-rest-framework,ossanna16/django-rest-framework,potpath/django-rest-framework,nryoung/django-rest-framework,nhorelik/django-rest-framework,jpadilla/django-rest-framework,ashishfinoit/django-rest-framework,qsorix/django-rest-framework,rafaelcaricio/django-rest-framework,uruz/django-rest-framework,arpheno/django-rest-framework,cheif/django-rest-framework,kylefox/django-rest-framework,brandoncazander/django-rest-framework,linovia/django-rest-framework,linovia/django-rest-framework,yiyocx/django-rest-framework,jpulec/django-rest-framework,ambivalentno/django-rest-framework,jerryhebert/django-rest-framework,damycra/django-rest-framework,pombredanne/django-rest-framework,andriy-s/django-rest-framework,abdulhaq-e/django-rest-framework,buptlsl/django-rest-framework,leeahoward/django-rest-framework,kylefox/django-rest-framework,delinhabit/django-rest-framework,tcroiset/django-rest-framework,MJafarMashhadi/django-rest-framework,ajaali/django-rest-framework,rhblind/django-rest-framework,agconti/django-rest-framework,ticosax/django-rest-framework,bluedazzle/django-rest-framework,wwj718/django-rest-framework,davesque/django-rest-framework,alacritythief/django-rest-framework,jpulec/django-rest-framework,ashishfinoit/django-rest-framework,tomchristie/django-rest-framework,potpath/django-rest-framework,ossanna16/django-rest-framework,aericson/django-rest-framework,ticosax/django-rest-framework,antonyc/django-rest-framework,iheitlager/django-rest-framework,ticosax/django-rest-framework,rubendura/django-rest-framework,HireAnEsquire/django-rest-framework,paolopaolopaolo/django-rest-framework,ezheidtmann/django-rest-framework,delinhabit/django-rest-framework,aericson/django-rest-framework,d0ugal/django-rest-framework,canassa/django-rest-framework,wangpanjun/django-rest-framework,rafaelang/django-rest-framework,kezabelle/django-rest-framework,paolopaolopaolo/django-rest-framework,wzbozon/django-rest-framework,wangpanjun/django-rest-framework,callorico/django-rest-framework,krinart/django-rest-framework,thedrow/django-rest-framework-1,linovia/django-rest-framework,jpadilla/django-rest-framework,bluedazzle/django-rest-framework,wangpanjun/django-rest-framework,thedrow/django-rest-framework-1,yiyocx/django-rest-framework,brandoncazander/django-rest-framework,potpath/django-rest-framework,canassa/django-rest-framework,dmwyatt/django-rest-framework,pombredanne/django-rest-framework,maryokhin/django-rest-framework,MJafarMashhadi/django-rest-framework,ambivalentno/django-rest-framework,qsorix/django-rest-framework,sheppard/django-rest-framework,hnarayanan/django-rest-framework,arpheno/django-rest-framework,wzbozon/django-rest-framework,lubomir/django-rest-framework,johnraz/django-rest-framework,damycra/django-rest-framework,werthen/django-rest-framework,atombrella/django-rest-framework,uruz/django-rest-framework,James1345/django-rest-framework,tigeraniya/django-rest-framework,waytai/django-rest-framework,sehmaschine/django-rest-framework,fishky/django-rest-framework,yiyocx/django-rest-framework,nhorelik/django-rest-framework,YBJAY00000/django-rest-framework,johnraz/django-rest-framework,kylefox/django-rest-framework,werthen/django-rest-framework,wwj718/django-rest-framework,dmwyatt/django-rest-framework,xiaotangyuan/django-rest-framework,atombrella/django-rest-framework,hnarayanan/django-rest-framework,James1345/django-rest-framework,tigeraniya/django-rest-framework,kgeorgy/django-rest-framework,maryokhin/django-rest-framework,sbellem/django-rest-framework,xiaotangyuan/django-rest-framework,AlexandreProenca/django-rest-framework,douwevandermeij/django-rest-framework,sehmaschine/django-rest-framework,hnakamur/django-rest-framework,edx/django-rest-framework,HireAnEsquire/django-rest-framework,vstoykov/django-rest-framework,rafaelcaricio/django-rest-framework,sbellem/django-rest-framework,vstoykov/django-rest-framework,kezabelle/django-rest-framework,HireAnEsquire/django-rest-framework,kezabelle/django-rest-framework,jness/django-rest-framework,elim/django-rest-framework,raphaelmerx/django-rest-framework,davesque/django-rest-framework,rubendura/django-rest-framework,simudream/django-rest-framework,kennydude/django-rest-framework,ossanna16/django-rest-framework,waytai/django-rest-framework,justanr/django-rest-framework,maryokhin/django-rest-framework,thedrow/django-rest-framework-1,cyberj/django-rest-framework,wedaly/django-rest-framework,rafaelang/django-rest-framework,rafaelcaricio/django-rest-framework,rafaelang/django-rest-framework,tcroiset/django-rest-framework,rhblind/django-rest-framework,jness/django-rest-framework,ebsaral/django-rest-framework,sehmaschine/django-rest-framework,kennydude/django-rest-framework,kennydude/django-rest-framework,ezheidtmann/django-rest-framework,sheppard/django-rest-framework,akalipetis/django-rest-framework,tomchristie/django-rest-framework,adambain-vokal/django-rest-framework,hunter007/django-rest-framework,damycra/django-rest-framework,edx/django-rest-framework,nhorelik/django-rest-framework,dmwyatt/django-rest-framework,jerryhebert/django-rest-framework,d0ugal/django-rest-framework,jtiai/django-rest-framework,callorico/django-rest-framework,callorico/django-rest-framework,agconti/django-rest-framework,uploadcare/django-rest-framework,andriy-s/django-rest-framework,iheitlager/django-rest-framework,krinart/django-rest-framework,agconti/django-rest-framework,zeldalink0515/django-rest-framework,wzbozon/django-rest-framework,fishky/django-rest-framework,arpheno/django-rest-framework,wedaly/django-rest-framework,paolopaolopaolo/django-rest-framework,VishvajitP/django-rest-framework,VishvajitP/django-rest-framework,simudream/django-rest-framework,ezheidtmann/django-rest-framework,ebsaral/django-rest-framework,antonyc/django-rest-framework,elim/django-rest-framework,atombrella/django-rest-framework,hnarayanan/django-rest-framework,buptlsl/django-rest-framework,buptlsl/django-rest-framework,hnakamur/django-rest-framework,douwevandermeij/django-rest-framework,uploadcare/django-rest-framework |
from __future__ import unicode_literals
from django.db import models, migrations
from django.conf import settings
class Migration(migrations.Migration):
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
]
operations = [
migrations.CreateModel(
name='Token',
fields=[
('key', models.CharField(max_length=40, serialize=False, primary_key=True)),
('created', models.DateTimeField(auto_now_add=True)),
('user', models.OneToOneField(related_name=b'auth_token', to=settings.AUTH_USER_MODEL)),
],
options={
},
bases=(models.Model,),
),
]
Update initial migration to work on Python 3 | # -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import models, migrations
from django.conf import settings
class Migration(migrations.Migration):
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
]
operations = [
migrations.CreateModel(
name='Token',
fields=[
('key', models.CharField(primary_key=True, serialize=False, max_length=40)),
('created', models.DateTimeField(auto_now_add=True)),
('user', models.OneToOneField(to=settings.AUTH_USER_MODEL, related_name='auth_token')),
],
options={
},
bases=(models.Model,),
),
]
| <commit_before>
from __future__ import unicode_literals
from django.db import models, migrations
from django.conf import settings
class Migration(migrations.Migration):
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
]
operations = [
migrations.CreateModel(
name='Token',
fields=[
('key', models.CharField(max_length=40, serialize=False, primary_key=True)),
('created', models.DateTimeField(auto_now_add=True)),
('user', models.OneToOneField(related_name=b'auth_token', to=settings.AUTH_USER_MODEL)),
],
options={
},
bases=(models.Model,),
),
]
<commit_msg>Update initial migration to work on Python 3<commit_after> | # -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import models, migrations
from django.conf import settings
class Migration(migrations.Migration):
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
]
operations = [
migrations.CreateModel(
name='Token',
fields=[
('key', models.CharField(primary_key=True, serialize=False, max_length=40)),
('created', models.DateTimeField(auto_now_add=True)),
('user', models.OneToOneField(to=settings.AUTH_USER_MODEL, related_name='auth_token')),
],
options={
},
bases=(models.Model,),
),
]
|
from __future__ import unicode_literals
from django.db import models, migrations
from django.conf import settings
class Migration(migrations.Migration):
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
]
operations = [
migrations.CreateModel(
name='Token',
fields=[
('key', models.CharField(max_length=40, serialize=False, primary_key=True)),
('created', models.DateTimeField(auto_now_add=True)),
('user', models.OneToOneField(related_name=b'auth_token', to=settings.AUTH_USER_MODEL)),
],
options={
},
bases=(models.Model,),
),
]
Update initial migration to work on Python 3# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import models, migrations
from django.conf import settings
class Migration(migrations.Migration):
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
]
operations = [
migrations.CreateModel(
name='Token',
fields=[
('key', models.CharField(primary_key=True, serialize=False, max_length=40)),
('created', models.DateTimeField(auto_now_add=True)),
('user', models.OneToOneField(to=settings.AUTH_USER_MODEL, related_name='auth_token')),
],
options={
},
bases=(models.Model,),
),
]
| <commit_before>
from __future__ import unicode_literals
from django.db import models, migrations
from django.conf import settings
class Migration(migrations.Migration):
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
]
operations = [
migrations.CreateModel(
name='Token',
fields=[
('key', models.CharField(max_length=40, serialize=False, primary_key=True)),
('created', models.DateTimeField(auto_now_add=True)),
('user', models.OneToOneField(related_name=b'auth_token', to=settings.AUTH_USER_MODEL)),
],
options={
},
bases=(models.Model,),
),
]
<commit_msg>Update initial migration to work on Python 3<commit_after># -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import models, migrations
from django.conf import settings
class Migration(migrations.Migration):
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
]
operations = [
migrations.CreateModel(
name='Token',
fields=[
('key', models.CharField(primary_key=True, serialize=False, max_length=40)),
('created', models.DateTimeField(auto_now_add=True)),
('user', models.OneToOneField(to=settings.AUTH_USER_MODEL, related_name='auth_token')),
],
options={
},
bases=(models.Model,),
),
]
|
8e9fd28004c1f8daadc5ce7f51b40543c28720c0 | djangoautoconf/settings_templates/smtp_account_template.py | djangoautoconf/settings_templates/smtp_account_template.py | __author__ = 'q19420'
smtp_username = "test"
smtp_password = "testpass" | __author__ = 'weijia'
smtp_username = None
smtp_password = None | Use None as username and password for SMTP. | Use None as username and password for SMTP.
| Python | bsd-3-clause | weijia/djangoautoconf,weijia/djangoautoconf | __author__ = 'q19420'
smtp_username = "test"
smtp_password = "testpass"Use None as username and password for SMTP. | __author__ = 'weijia'
smtp_username = None
smtp_password = None | <commit_before>__author__ = 'q19420'
smtp_username = "test"
smtp_password = "testpass"<commit_msg>Use None as username and password for SMTP.<commit_after> | __author__ = 'weijia'
smtp_username = None
smtp_password = None | __author__ = 'q19420'
smtp_username = "test"
smtp_password = "testpass"Use None as username and password for SMTP.__author__ = 'weijia'
smtp_username = None
smtp_password = None | <commit_before>__author__ = 'q19420'
smtp_username = "test"
smtp_password = "testpass"<commit_msg>Use None as username and password for SMTP.<commit_after>__author__ = 'weijia'
smtp_username = None
smtp_password = None |
60f87cb4c3523faf5c5cdbc5f16453cae755988b | angr/procedures/java_jni/GetArrayElements.py | angr/procedures/java_jni/GetArrayElements.py | from . import JNISimProcedure
from ...engines.soot.values.arrayref import SimSootValue_ArrayRef
class GetArrayElements(JNISimProcedure):
return_ty = 'reference'
def run(self, ptr_env, array, ptr_isCopy):
array_ref = self.state.jni_references.lookup(array)
values = self.load_java_array(self.state, array_ref)
memory_addr = self.store_in_native_memory(values, array_ref.type)
return memory_addr
def load_java_array(self, array_ref, start_idx=None, end_idx=None):
if start_idx is None:
start_idx = 0
if end_idx is None:
end_idx = self.state.solver.max(array_ref.size)
javavm_memory = self.state.get_javavm_view_of_plugin("memory")
values = []
for idx in range(start_idx, end_idx):
idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx)
value = javavm_memory.load(idx_array_ref)
values.append(value)
return values | from . import JNISimProcedure
from ...engines.soot.values.arrayref import SimSootValue_ArrayRef
class GetArrayElements(JNISimProcedure):
return_ty = 'reference'
def run(self, ptr_env, array, ptr_isCopy):
array_ref = self.state.jni_references.lookup(array)
values = self.load_java_array(self.state, array_ref)
memory_addr = self.store_in_native_memory(values, array_ref.type)
if self.state.solver.eval(ptr_isCopy != 0):
self.store_in_native_memory(data=self.JNI_TRUE, data_type='boolean', addr=ptr_isCopy)
return memory_addr
def load_java_array(self, array_ref, start_idx=None, end_idx=None):
if start_idx is None:
start_idx = 0
if end_idx is None:
end_idx = self.state.solver.max(array_ref.size)
javavm_memory = self.state.get_javavm_view_of_plugin("memory")
values = []
for idx in range(start_idx, end_idx):
idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx)
value = javavm_memory.load(idx_array_ref)
values.append(value)
return values | Fix case if isCopy is null | Fix case if isCopy is null
| Python | bsd-2-clause | schieb/angr,schieb/angr,angr/angr,angr/angr,iamahuman/angr,angr/angr,iamahuman/angr,schieb/angr,iamahuman/angr | from . import JNISimProcedure
from ...engines.soot.values.arrayref import SimSootValue_ArrayRef
class GetArrayElements(JNISimProcedure):
return_ty = 'reference'
def run(self, ptr_env, array, ptr_isCopy):
array_ref = self.state.jni_references.lookup(array)
values = self.load_java_array(self.state, array_ref)
memory_addr = self.store_in_native_memory(values, array_ref.type)
return memory_addr
def load_java_array(self, array_ref, start_idx=None, end_idx=None):
if start_idx is None:
start_idx = 0
if end_idx is None:
end_idx = self.state.solver.max(array_ref.size)
javavm_memory = self.state.get_javavm_view_of_plugin("memory")
values = []
for idx in range(start_idx, end_idx):
idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx)
value = javavm_memory.load(idx_array_ref)
values.append(value)
return valuesFix case if isCopy is null | from . import JNISimProcedure
from ...engines.soot.values.arrayref import SimSootValue_ArrayRef
class GetArrayElements(JNISimProcedure):
return_ty = 'reference'
def run(self, ptr_env, array, ptr_isCopy):
array_ref = self.state.jni_references.lookup(array)
values = self.load_java_array(self.state, array_ref)
memory_addr = self.store_in_native_memory(values, array_ref.type)
if self.state.solver.eval(ptr_isCopy != 0):
self.store_in_native_memory(data=self.JNI_TRUE, data_type='boolean', addr=ptr_isCopy)
return memory_addr
def load_java_array(self, array_ref, start_idx=None, end_idx=None):
if start_idx is None:
start_idx = 0
if end_idx is None:
end_idx = self.state.solver.max(array_ref.size)
javavm_memory = self.state.get_javavm_view_of_plugin("memory")
values = []
for idx in range(start_idx, end_idx):
idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx)
value = javavm_memory.load(idx_array_ref)
values.append(value)
return values | <commit_before>from . import JNISimProcedure
from ...engines.soot.values.arrayref import SimSootValue_ArrayRef
class GetArrayElements(JNISimProcedure):
return_ty = 'reference'
def run(self, ptr_env, array, ptr_isCopy):
array_ref = self.state.jni_references.lookup(array)
values = self.load_java_array(self.state, array_ref)
memory_addr = self.store_in_native_memory(values, array_ref.type)
return memory_addr
def load_java_array(self, array_ref, start_idx=None, end_idx=None):
if start_idx is None:
start_idx = 0
if end_idx is None:
end_idx = self.state.solver.max(array_ref.size)
javavm_memory = self.state.get_javavm_view_of_plugin("memory")
values = []
for idx in range(start_idx, end_idx):
idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx)
value = javavm_memory.load(idx_array_ref)
values.append(value)
return values<commit_msg>Fix case if isCopy is null<commit_after> | from . import JNISimProcedure
from ...engines.soot.values.arrayref import SimSootValue_ArrayRef
class GetArrayElements(JNISimProcedure):
return_ty = 'reference'
def run(self, ptr_env, array, ptr_isCopy):
array_ref = self.state.jni_references.lookup(array)
values = self.load_java_array(self.state, array_ref)
memory_addr = self.store_in_native_memory(values, array_ref.type)
if self.state.solver.eval(ptr_isCopy != 0):
self.store_in_native_memory(data=self.JNI_TRUE, data_type='boolean', addr=ptr_isCopy)
return memory_addr
def load_java_array(self, array_ref, start_idx=None, end_idx=None):
if start_idx is None:
start_idx = 0
if end_idx is None:
end_idx = self.state.solver.max(array_ref.size)
javavm_memory = self.state.get_javavm_view_of_plugin("memory")
values = []
for idx in range(start_idx, end_idx):
idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx)
value = javavm_memory.load(idx_array_ref)
values.append(value)
return values | from . import JNISimProcedure
from ...engines.soot.values.arrayref import SimSootValue_ArrayRef
class GetArrayElements(JNISimProcedure):
return_ty = 'reference'
def run(self, ptr_env, array, ptr_isCopy):
array_ref = self.state.jni_references.lookup(array)
values = self.load_java_array(self.state, array_ref)
memory_addr = self.store_in_native_memory(values, array_ref.type)
return memory_addr
def load_java_array(self, array_ref, start_idx=None, end_idx=None):
if start_idx is None:
start_idx = 0
if end_idx is None:
end_idx = self.state.solver.max(array_ref.size)
javavm_memory = self.state.get_javavm_view_of_plugin("memory")
values = []
for idx in range(start_idx, end_idx):
idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx)
value = javavm_memory.load(idx_array_ref)
values.append(value)
return valuesFix case if isCopy is nullfrom . import JNISimProcedure
from ...engines.soot.values.arrayref import SimSootValue_ArrayRef
class GetArrayElements(JNISimProcedure):
return_ty = 'reference'
def run(self, ptr_env, array, ptr_isCopy):
array_ref = self.state.jni_references.lookup(array)
values = self.load_java_array(self.state, array_ref)
memory_addr = self.store_in_native_memory(values, array_ref.type)
if self.state.solver.eval(ptr_isCopy != 0):
self.store_in_native_memory(data=self.JNI_TRUE, data_type='boolean', addr=ptr_isCopy)
return memory_addr
def load_java_array(self, array_ref, start_idx=None, end_idx=None):
if start_idx is None:
start_idx = 0
if end_idx is None:
end_idx = self.state.solver.max(array_ref.size)
javavm_memory = self.state.get_javavm_view_of_plugin("memory")
values = []
for idx in range(start_idx, end_idx):
idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx)
value = javavm_memory.load(idx_array_ref)
values.append(value)
return values | <commit_before>from . import JNISimProcedure
from ...engines.soot.values.arrayref import SimSootValue_ArrayRef
class GetArrayElements(JNISimProcedure):
return_ty = 'reference'
def run(self, ptr_env, array, ptr_isCopy):
array_ref = self.state.jni_references.lookup(array)
values = self.load_java_array(self.state, array_ref)
memory_addr = self.store_in_native_memory(values, array_ref.type)
return memory_addr
def load_java_array(self, array_ref, start_idx=None, end_idx=None):
if start_idx is None:
start_idx = 0
if end_idx is None:
end_idx = self.state.solver.max(array_ref.size)
javavm_memory = self.state.get_javavm_view_of_plugin("memory")
values = []
for idx in range(start_idx, end_idx):
idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx)
value = javavm_memory.load(idx_array_ref)
values.append(value)
return values<commit_msg>Fix case if isCopy is null<commit_after>from . import JNISimProcedure
from ...engines.soot.values.arrayref import SimSootValue_ArrayRef
class GetArrayElements(JNISimProcedure):
return_ty = 'reference'
def run(self, ptr_env, array, ptr_isCopy):
array_ref = self.state.jni_references.lookup(array)
values = self.load_java_array(self.state, array_ref)
memory_addr = self.store_in_native_memory(values, array_ref.type)
if self.state.solver.eval(ptr_isCopy != 0):
self.store_in_native_memory(data=self.JNI_TRUE, data_type='boolean', addr=ptr_isCopy)
return memory_addr
def load_java_array(self, array_ref, start_idx=None, end_idx=None):
if start_idx is None:
start_idx = 0
if end_idx is None:
end_idx = self.state.solver.max(array_ref.size)
javavm_memory = self.state.get_javavm_view_of_plugin("memory")
values = []
for idx in range(start_idx, end_idx):
idx_array_ref = SimSootValue_ArrayRef.get_arrayref_for_idx(base=array_ref, idx=idx)
value = javavm_memory.load(idx_array_ref)
values.append(value)
return values |
36d7a5f754fef3bdab0103229fe8b5ee267f9376 | scripts/urls-starting-with.py | scripts/urls-starting-with.py | import re
import sys
from xml.sax import make_parser, handler
if len(sys.argv) < 3:
print "This script expects two arguments: \n1. The path to a posts.xml file from a Stack Overflow data dump.\n2. A URL prefix to search for."
else:
start_with = sys.argv[2]
class SOProcessor(handler.ContentHandler):
def startElement(self, name, attrs):
if name == "row":
if attrs["PostTypeId"] == "1":
body = attrs["Body"]
if start_with in body:
matches = re.findall(r'<a href="([^"]+)"', body)
for url in filter(lambda x: x.startswith(start_with), matches):
print url
parser = make_parser()
parser.setContentHandler(SOProcessor())
parser.parse(open(sys.argv[1]))
| import re
import sys
from xml.sax import make_parser, handler
if len(sys.argv) < 3:
print "This script expects two arguments: \n1. The path to a posts.xml file from a Stack Overflow data dump.\n2. A URL prefix to search for."
else:
start_with = sys.argv[2]
class SOProcessor(handler.ContentHandler):
def startElement(self, name, attrs):
if name == "row":
if attrs["PostTypeId"] == "1":
body = attrs["Body"]
if start_with in body:
matches = re.findall(r'<a href="([^"]+)"', body)
for url in filter(lambda x: x.startswith(start_with), matches):
print url, attrs["Tags"]
parser = make_parser()
parser.setContentHandler(SOProcessor())
parser.parse(open(sys.argv[1]))
| Print tags out when scanning for URLs. | Print tags out when scanning for URLs.
| Python | bsd-3-clause | alnorth/stackdoc,alnorth/stackdoc,alnorth/stackdoc | import re
import sys
from xml.sax import make_parser, handler
if len(sys.argv) < 3:
print "This script expects two arguments: \n1. The path to a posts.xml file from a Stack Overflow data dump.\n2. A URL prefix to search for."
else:
start_with = sys.argv[2]
class SOProcessor(handler.ContentHandler):
def startElement(self, name, attrs):
if name == "row":
if attrs["PostTypeId"] == "1":
body = attrs["Body"]
if start_with in body:
matches = re.findall(r'<a href="([^"]+)"', body)
for url in filter(lambda x: x.startswith(start_with), matches):
print url
parser = make_parser()
parser.setContentHandler(SOProcessor())
parser.parse(open(sys.argv[1]))
Print tags out when scanning for URLs. | import re
import sys
from xml.sax import make_parser, handler
if len(sys.argv) < 3:
print "This script expects two arguments: \n1. The path to a posts.xml file from a Stack Overflow data dump.\n2. A URL prefix to search for."
else:
start_with = sys.argv[2]
class SOProcessor(handler.ContentHandler):
def startElement(self, name, attrs):
if name == "row":
if attrs["PostTypeId"] == "1":
body = attrs["Body"]
if start_with in body:
matches = re.findall(r'<a href="([^"]+)"', body)
for url in filter(lambda x: x.startswith(start_with), matches):
print url, attrs["Tags"]
parser = make_parser()
parser.setContentHandler(SOProcessor())
parser.parse(open(sys.argv[1]))
| <commit_before>import re
import sys
from xml.sax import make_parser, handler
if len(sys.argv) < 3:
print "This script expects two arguments: \n1. The path to a posts.xml file from a Stack Overflow data dump.\n2. A URL prefix to search for."
else:
start_with = sys.argv[2]
class SOProcessor(handler.ContentHandler):
def startElement(self, name, attrs):
if name == "row":
if attrs["PostTypeId"] == "1":
body = attrs["Body"]
if start_with in body:
matches = re.findall(r'<a href="([^"]+)"', body)
for url in filter(lambda x: x.startswith(start_with), matches):
print url
parser = make_parser()
parser.setContentHandler(SOProcessor())
parser.parse(open(sys.argv[1]))
<commit_msg>Print tags out when scanning for URLs.<commit_after> | import re
import sys
from xml.sax import make_parser, handler
if len(sys.argv) < 3:
print "This script expects two arguments: \n1. The path to a posts.xml file from a Stack Overflow data dump.\n2. A URL prefix to search for."
else:
start_with = sys.argv[2]
class SOProcessor(handler.ContentHandler):
def startElement(self, name, attrs):
if name == "row":
if attrs["PostTypeId"] == "1":
body = attrs["Body"]
if start_with in body:
matches = re.findall(r'<a href="([^"]+)"', body)
for url in filter(lambda x: x.startswith(start_with), matches):
print url, attrs["Tags"]
parser = make_parser()
parser.setContentHandler(SOProcessor())
parser.parse(open(sys.argv[1]))
| import re
import sys
from xml.sax import make_parser, handler
if len(sys.argv) < 3:
print "This script expects two arguments: \n1. The path to a posts.xml file from a Stack Overflow data dump.\n2. A URL prefix to search for."
else:
start_with = sys.argv[2]
class SOProcessor(handler.ContentHandler):
def startElement(self, name, attrs):
if name == "row":
if attrs["PostTypeId"] == "1":
body = attrs["Body"]
if start_with in body:
matches = re.findall(r'<a href="([^"]+)"', body)
for url in filter(lambda x: x.startswith(start_with), matches):
print url
parser = make_parser()
parser.setContentHandler(SOProcessor())
parser.parse(open(sys.argv[1]))
Print tags out when scanning for URLs.import re
import sys
from xml.sax import make_parser, handler
if len(sys.argv) < 3:
print "This script expects two arguments: \n1. The path to a posts.xml file from a Stack Overflow data dump.\n2. A URL prefix to search for."
else:
start_with = sys.argv[2]
class SOProcessor(handler.ContentHandler):
def startElement(self, name, attrs):
if name == "row":
if attrs["PostTypeId"] == "1":
body = attrs["Body"]
if start_with in body:
matches = re.findall(r'<a href="([^"]+)"', body)
for url in filter(lambda x: x.startswith(start_with), matches):
print url, attrs["Tags"]
parser = make_parser()
parser.setContentHandler(SOProcessor())
parser.parse(open(sys.argv[1]))
| <commit_before>import re
import sys
from xml.sax import make_parser, handler
if len(sys.argv) < 3:
print "This script expects two arguments: \n1. The path to a posts.xml file from a Stack Overflow data dump.\n2. A URL prefix to search for."
else:
start_with = sys.argv[2]
class SOProcessor(handler.ContentHandler):
def startElement(self, name, attrs):
if name == "row":
if attrs["PostTypeId"] == "1":
body = attrs["Body"]
if start_with in body:
matches = re.findall(r'<a href="([^"]+)"', body)
for url in filter(lambda x: x.startswith(start_with), matches):
print url
parser = make_parser()
parser.setContentHandler(SOProcessor())
parser.parse(open(sys.argv[1]))
<commit_msg>Print tags out when scanning for URLs.<commit_after>import re
import sys
from xml.sax import make_parser, handler
if len(sys.argv) < 3:
print "This script expects two arguments: \n1. The path to a posts.xml file from a Stack Overflow data dump.\n2. A URL prefix to search for."
else:
start_with = sys.argv[2]
class SOProcessor(handler.ContentHandler):
def startElement(self, name, attrs):
if name == "row":
if attrs["PostTypeId"] == "1":
body = attrs["Body"]
if start_with in body:
matches = re.findall(r'<a href="([^"]+)"', body)
for url in filter(lambda x: x.startswith(start_with), matches):
print url, attrs["Tags"]
parser = make_parser()
parser.setContentHandler(SOProcessor())
parser.parse(open(sys.argv[1]))
|
fe4f2fa1c64d40a15c49cc4183a59c912574fdff | ibmcnx/doc/DataSources.py | ibmcnx/doc/DataSources.py | ######
# Check ExId (GUID) by Email through JDBC
#
# Author: Christoph Stoettner
# Mail: christoph.stoettner@stoeps.de
# Documentation: http://scripting101.stoeps.de
#
# Version: 2.0
# Date: 2014-06-04
#
# License: Apache 2.0
#
# Check ExId of a User in all Connections Applications
import ibmcnx.functions
cell = "'/Cell:" + AdminControl.getCell() + "/'"
cellid = AdminConfig.getid( cell )
dbs = AdminConfig.list( 'DataSource', str(cellid) )
for db in dbs:
t1 = ibmcnx.functions.getDSId( db )
AdminConfig.list( t1 ) | ######
# Check ExId (GUID) by Email through JDBC
#
# Author: Christoph Stoettner
# Mail: christoph.stoettner@stoeps.de
# Documentation: http://scripting101.stoeps.de
#
# Version: 2.0
# Date: 2014-06-04
#
# License: Apache 2.0
#
# Check ExId of a User in all Connections Applications
import ibmcnx.functions
print AdminControl.getCell()
cell = "'/Cell:" + AdminControl.getCell() + "/'"
cellid = AdminConfig.getid( cell )
dbs = AdminConfig.list( 'DataSource', str(cellid) )
for db in dbs:
t1 = ibmcnx.functions.getDSId( db )
AdminConfig.list( t1 ) | Create script to save documentation to a file | 4: Create script to save documentation to a file
Task-Url: http://github.com/stoeps13/ibmcnx2/issues/issue/4 | Python | apache-2.0 | stoeps13/ibmcnx2,stoeps13/ibmcnx2 | ######
# Check ExId (GUID) by Email through JDBC
#
# Author: Christoph Stoettner
# Mail: christoph.stoettner@stoeps.de
# Documentation: http://scripting101.stoeps.de
#
# Version: 2.0
# Date: 2014-06-04
#
# License: Apache 2.0
#
# Check ExId of a User in all Connections Applications
import ibmcnx.functions
cell = "'/Cell:" + AdminControl.getCell() + "/'"
cellid = AdminConfig.getid( cell )
dbs = AdminConfig.list( 'DataSource', str(cellid) )
for db in dbs:
t1 = ibmcnx.functions.getDSId( db )
AdminConfig.list( t1 )4: Create script to save documentation to a file
Task-Url: http://github.com/stoeps13/ibmcnx2/issues/issue/4 | ######
# Check ExId (GUID) by Email through JDBC
#
# Author: Christoph Stoettner
# Mail: christoph.stoettner@stoeps.de
# Documentation: http://scripting101.stoeps.de
#
# Version: 2.0
# Date: 2014-06-04
#
# License: Apache 2.0
#
# Check ExId of a User in all Connections Applications
import ibmcnx.functions
print AdminControl.getCell()
cell = "'/Cell:" + AdminControl.getCell() + "/'"
cellid = AdminConfig.getid( cell )
dbs = AdminConfig.list( 'DataSource', str(cellid) )
for db in dbs:
t1 = ibmcnx.functions.getDSId( db )
AdminConfig.list( t1 ) | <commit_before>######
# Check ExId (GUID) by Email through JDBC
#
# Author: Christoph Stoettner
# Mail: christoph.stoettner@stoeps.de
# Documentation: http://scripting101.stoeps.de
#
# Version: 2.0
# Date: 2014-06-04
#
# License: Apache 2.0
#
# Check ExId of a User in all Connections Applications
import ibmcnx.functions
cell = "'/Cell:" + AdminControl.getCell() + "/'"
cellid = AdminConfig.getid( cell )
dbs = AdminConfig.list( 'DataSource', str(cellid) )
for db in dbs:
t1 = ibmcnx.functions.getDSId( db )
AdminConfig.list( t1 )<commit_msg>4: Create script to save documentation to a file
Task-Url: http://github.com/stoeps13/ibmcnx2/issues/issue/4<commit_after> | ######
# Check ExId (GUID) by Email through JDBC
#
# Author: Christoph Stoettner
# Mail: christoph.stoettner@stoeps.de
# Documentation: http://scripting101.stoeps.de
#
# Version: 2.0
# Date: 2014-06-04
#
# License: Apache 2.0
#
# Check ExId of a User in all Connections Applications
import ibmcnx.functions
print AdminControl.getCell()
cell = "'/Cell:" + AdminControl.getCell() + "/'"
cellid = AdminConfig.getid( cell )
dbs = AdminConfig.list( 'DataSource', str(cellid) )
for db in dbs:
t1 = ibmcnx.functions.getDSId( db )
AdminConfig.list( t1 ) | ######
# Check ExId (GUID) by Email through JDBC
#
# Author: Christoph Stoettner
# Mail: christoph.stoettner@stoeps.de
# Documentation: http://scripting101.stoeps.de
#
# Version: 2.0
# Date: 2014-06-04
#
# License: Apache 2.0
#
# Check ExId of a User in all Connections Applications
import ibmcnx.functions
cell = "'/Cell:" + AdminControl.getCell() + "/'"
cellid = AdminConfig.getid( cell )
dbs = AdminConfig.list( 'DataSource', str(cellid) )
for db in dbs:
t1 = ibmcnx.functions.getDSId( db )
AdminConfig.list( t1 )4: Create script to save documentation to a file
Task-Url: http://github.com/stoeps13/ibmcnx2/issues/issue/4######
# Check ExId (GUID) by Email through JDBC
#
# Author: Christoph Stoettner
# Mail: christoph.stoettner@stoeps.de
# Documentation: http://scripting101.stoeps.de
#
# Version: 2.0
# Date: 2014-06-04
#
# License: Apache 2.0
#
# Check ExId of a User in all Connections Applications
import ibmcnx.functions
print AdminControl.getCell()
cell = "'/Cell:" + AdminControl.getCell() + "/'"
cellid = AdminConfig.getid( cell )
dbs = AdminConfig.list( 'DataSource', str(cellid) )
for db in dbs:
t1 = ibmcnx.functions.getDSId( db )
AdminConfig.list( t1 ) | <commit_before>######
# Check ExId (GUID) by Email through JDBC
#
# Author: Christoph Stoettner
# Mail: christoph.stoettner@stoeps.de
# Documentation: http://scripting101.stoeps.de
#
# Version: 2.0
# Date: 2014-06-04
#
# License: Apache 2.0
#
# Check ExId of a User in all Connections Applications
import ibmcnx.functions
cell = "'/Cell:" + AdminControl.getCell() + "/'"
cellid = AdminConfig.getid( cell )
dbs = AdminConfig.list( 'DataSource', str(cellid) )
for db in dbs:
t1 = ibmcnx.functions.getDSId( db )
AdminConfig.list( t1 )<commit_msg>4: Create script to save documentation to a file
Task-Url: http://github.com/stoeps13/ibmcnx2/issues/issue/4<commit_after>######
# Check ExId (GUID) by Email through JDBC
#
# Author: Christoph Stoettner
# Mail: christoph.stoettner@stoeps.de
# Documentation: http://scripting101.stoeps.de
#
# Version: 2.0
# Date: 2014-06-04
#
# License: Apache 2.0
#
# Check ExId of a User in all Connections Applications
import ibmcnx.functions
print AdminControl.getCell()
cell = "'/Cell:" + AdminControl.getCell() + "/'"
cellid = AdminConfig.getid( cell )
dbs = AdminConfig.list( 'DataSource', str(cellid) )
for db in dbs:
t1 = ibmcnx.functions.getDSId( db )
AdminConfig.list( t1 ) |
ed59db63ab5832468b1348f6cd9bf00880fbbdbc | busstops/management/commands/import_areas.py | busstops/management/commands/import_areas.py | """
Import administrative areas from the NPTG.
Usage:
import_areas < AdminAreas.csv
"""
from ..import_from_csv import ImportFromCSVCommand
from ...models import AdminArea
class Command(ImportFromCSVCommand):
def handle_row(self, row):
AdminArea.objects.update_or_create(
id=row['AdministrativeAreaCode'],
defaults={
'atco_code': row['AtcoAreaCode'],
'name': row['AreaName'],
'short_name': row['ShortName'],
'country': row['Country'],
'region_id': row['RegionCode'],
}
)
| """
Import administrative areas from the NPTG.
Usage:
import_areas < AdminAreas.csv
"""
from ..import_from_csv import ImportFromCSVCommand
from ...models import AdminArea
class Command(ImportFromCSVCommand):
def handle_row(self, row):
AdminArea.objects.update_or_create(
id=row['AdministrativeAreaCode'],
defaults={
'atco_code': row['AtcoAreaCode'],
'name': row['AreaName'],
'short_name': row['ShortName'],
'country': row['Country'],
'region_id': row['RegionCode'],
}
)
def handle(self, *args, **options):
super(Command, self).handle(*args, **options)
# Move Cumbria to the North West.
# There is the legacy of the confusing 'North East and Cumbria' Traveline region,
# but actually Cumbrian bus services are in the North West now
AdminArea.objects.filter(name='Cumbria').update(region_id='NW')
| Move Cumbria to the North West | Move Cumbria to the North West
| Python | mpl-2.0 | jclgoodwin/bustimes.org.uk,jclgoodwin/bustimes.org.uk,stev-0/bustimes.org.uk,jclgoodwin/bustimes.org.uk,stev-0/bustimes.org.uk,jclgoodwin/bustimes.org.uk,stev-0/bustimes.org.uk,stev-0/bustimes.org.uk,stev-0/bustimes.org.uk | """
Import administrative areas from the NPTG.
Usage:
import_areas < AdminAreas.csv
"""
from ..import_from_csv import ImportFromCSVCommand
from ...models import AdminArea
class Command(ImportFromCSVCommand):
def handle_row(self, row):
AdminArea.objects.update_or_create(
id=row['AdministrativeAreaCode'],
defaults={
'atco_code': row['AtcoAreaCode'],
'name': row['AreaName'],
'short_name': row['ShortName'],
'country': row['Country'],
'region_id': row['RegionCode'],
}
)
Move Cumbria to the North West | """
Import administrative areas from the NPTG.
Usage:
import_areas < AdminAreas.csv
"""
from ..import_from_csv import ImportFromCSVCommand
from ...models import AdminArea
class Command(ImportFromCSVCommand):
def handle_row(self, row):
AdminArea.objects.update_or_create(
id=row['AdministrativeAreaCode'],
defaults={
'atco_code': row['AtcoAreaCode'],
'name': row['AreaName'],
'short_name': row['ShortName'],
'country': row['Country'],
'region_id': row['RegionCode'],
}
)
def handle(self, *args, **options):
super(Command, self).handle(*args, **options)
# Move Cumbria to the North West.
# There is the legacy of the confusing 'North East and Cumbria' Traveline region,
# but actually Cumbrian bus services are in the North West now
AdminArea.objects.filter(name='Cumbria').update(region_id='NW')
| <commit_before>"""
Import administrative areas from the NPTG.
Usage:
import_areas < AdminAreas.csv
"""
from ..import_from_csv import ImportFromCSVCommand
from ...models import AdminArea
class Command(ImportFromCSVCommand):
def handle_row(self, row):
AdminArea.objects.update_or_create(
id=row['AdministrativeAreaCode'],
defaults={
'atco_code': row['AtcoAreaCode'],
'name': row['AreaName'],
'short_name': row['ShortName'],
'country': row['Country'],
'region_id': row['RegionCode'],
}
)
<commit_msg>Move Cumbria to the North West<commit_after> | """
Import administrative areas from the NPTG.
Usage:
import_areas < AdminAreas.csv
"""
from ..import_from_csv import ImportFromCSVCommand
from ...models import AdminArea
class Command(ImportFromCSVCommand):
def handle_row(self, row):
AdminArea.objects.update_or_create(
id=row['AdministrativeAreaCode'],
defaults={
'atco_code': row['AtcoAreaCode'],
'name': row['AreaName'],
'short_name': row['ShortName'],
'country': row['Country'],
'region_id': row['RegionCode'],
}
)
def handle(self, *args, **options):
super(Command, self).handle(*args, **options)
# Move Cumbria to the North West.
# There is the legacy of the confusing 'North East and Cumbria' Traveline region,
# but actually Cumbrian bus services are in the North West now
AdminArea.objects.filter(name='Cumbria').update(region_id='NW')
| """
Import administrative areas from the NPTG.
Usage:
import_areas < AdminAreas.csv
"""
from ..import_from_csv import ImportFromCSVCommand
from ...models import AdminArea
class Command(ImportFromCSVCommand):
def handle_row(self, row):
AdminArea.objects.update_or_create(
id=row['AdministrativeAreaCode'],
defaults={
'atco_code': row['AtcoAreaCode'],
'name': row['AreaName'],
'short_name': row['ShortName'],
'country': row['Country'],
'region_id': row['RegionCode'],
}
)
Move Cumbria to the North West"""
Import administrative areas from the NPTG.
Usage:
import_areas < AdminAreas.csv
"""
from ..import_from_csv import ImportFromCSVCommand
from ...models import AdminArea
class Command(ImportFromCSVCommand):
def handle_row(self, row):
AdminArea.objects.update_or_create(
id=row['AdministrativeAreaCode'],
defaults={
'atco_code': row['AtcoAreaCode'],
'name': row['AreaName'],
'short_name': row['ShortName'],
'country': row['Country'],
'region_id': row['RegionCode'],
}
)
def handle(self, *args, **options):
super(Command, self).handle(*args, **options)
# Move Cumbria to the North West.
# There is the legacy of the confusing 'North East and Cumbria' Traveline region,
# but actually Cumbrian bus services are in the North West now
AdminArea.objects.filter(name='Cumbria').update(region_id='NW')
| <commit_before>"""
Import administrative areas from the NPTG.
Usage:
import_areas < AdminAreas.csv
"""
from ..import_from_csv import ImportFromCSVCommand
from ...models import AdminArea
class Command(ImportFromCSVCommand):
def handle_row(self, row):
AdminArea.objects.update_or_create(
id=row['AdministrativeAreaCode'],
defaults={
'atco_code': row['AtcoAreaCode'],
'name': row['AreaName'],
'short_name': row['ShortName'],
'country': row['Country'],
'region_id': row['RegionCode'],
}
)
<commit_msg>Move Cumbria to the North West<commit_after>"""
Import administrative areas from the NPTG.
Usage:
import_areas < AdminAreas.csv
"""
from ..import_from_csv import ImportFromCSVCommand
from ...models import AdminArea
class Command(ImportFromCSVCommand):
def handle_row(self, row):
AdminArea.objects.update_or_create(
id=row['AdministrativeAreaCode'],
defaults={
'atco_code': row['AtcoAreaCode'],
'name': row['AreaName'],
'short_name': row['ShortName'],
'country': row['Country'],
'region_id': row['RegionCode'],
}
)
def handle(self, *args, **options):
super(Command, self).handle(*args, **options)
# Move Cumbria to the North West.
# There is the legacy of the confusing 'North East and Cumbria' Traveline region,
# but actually Cumbrian bus services are in the North West now
AdminArea.objects.filter(name='Cumbria').update(region_id='NW')
|
8974832551f48ef3fbd3023cc2c26836aa01de1c | derrida/__init__.py | derrida/__init__.py | __version_info__ = (1, 1, 0, None)
# Dot-connect all but the last. Last is dash-connected if not None.
__version__ = '.'.join([str(i) for i in __version_info__[:-1]])
if __version_info__[-1] is not None:
__version__ += ('-%s' % (__version_info__[-1],))
# context processor to add version to the template environment
def context_extras(request):
return {
# software version
'SW_VERSION': __version__
}
| __version_info__ = (1, 2, 0, 'dev')
# Dot-connect all but the last. Last is dash-connected if not None.
__version__ = '.'.join([str(i) for i in __version_info__[:-1]])
if __version_info__[-1] is not None:
__version__ += ('-%s' % (__version_info__[-1],))
# context processor to add version to the template environment
def context_extras(request):
return {
# software version
'SW_VERSION': __version__
}
| Set develop version to 1.2-dev | Set develop version to 1.2-dev
| Python | apache-2.0 | Princeton-CDH/derrida-django,Princeton-CDH/derrida-django,Princeton-CDH/derrida-django,Princeton-CDH/derrida-django | __version_info__ = (1, 1, 0, None)
# Dot-connect all but the last. Last is dash-connected if not None.
__version__ = '.'.join([str(i) for i in __version_info__[:-1]])
if __version_info__[-1] is not None:
__version__ += ('-%s' % (__version_info__[-1],))
# context processor to add version to the template environment
def context_extras(request):
return {
# software version
'SW_VERSION': __version__
}
Set develop version to 1.2-dev | __version_info__ = (1, 2, 0, 'dev')
# Dot-connect all but the last. Last is dash-connected if not None.
__version__ = '.'.join([str(i) for i in __version_info__[:-1]])
if __version_info__[-1] is not None:
__version__ += ('-%s' % (__version_info__[-1],))
# context processor to add version to the template environment
def context_extras(request):
return {
# software version
'SW_VERSION': __version__
}
| <commit_before>__version_info__ = (1, 1, 0, None)
# Dot-connect all but the last. Last is dash-connected if not None.
__version__ = '.'.join([str(i) for i in __version_info__[:-1]])
if __version_info__[-1] is not None:
__version__ += ('-%s' % (__version_info__[-1],))
# context processor to add version to the template environment
def context_extras(request):
return {
# software version
'SW_VERSION': __version__
}
<commit_msg>Set develop version to 1.2-dev<commit_after> | __version_info__ = (1, 2, 0, 'dev')
# Dot-connect all but the last. Last is dash-connected if not None.
__version__ = '.'.join([str(i) for i in __version_info__[:-1]])
if __version_info__[-1] is not None:
__version__ += ('-%s' % (__version_info__[-1],))
# context processor to add version to the template environment
def context_extras(request):
return {
# software version
'SW_VERSION': __version__
}
| __version_info__ = (1, 1, 0, None)
# Dot-connect all but the last. Last is dash-connected if not None.
__version__ = '.'.join([str(i) for i in __version_info__[:-1]])
if __version_info__[-1] is not None:
__version__ += ('-%s' % (__version_info__[-1],))
# context processor to add version to the template environment
def context_extras(request):
return {
# software version
'SW_VERSION': __version__
}
Set develop version to 1.2-dev__version_info__ = (1, 2, 0, 'dev')
# Dot-connect all but the last. Last is dash-connected if not None.
__version__ = '.'.join([str(i) for i in __version_info__[:-1]])
if __version_info__[-1] is not None:
__version__ += ('-%s' % (__version_info__[-1],))
# context processor to add version to the template environment
def context_extras(request):
return {
# software version
'SW_VERSION': __version__
}
| <commit_before>__version_info__ = (1, 1, 0, None)
# Dot-connect all but the last. Last is dash-connected if not None.
__version__ = '.'.join([str(i) for i in __version_info__[:-1]])
if __version_info__[-1] is not None:
__version__ += ('-%s' % (__version_info__[-1],))
# context processor to add version to the template environment
def context_extras(request):
return {
# software version
'SW_VERSION': __version__
}
<commit_msg>Set develop version to 1.2-dev<commit_after>__version_info__ = (1, 2, 0, 'dev')
# Dot-connect all but the last. Last is dash-connected if not None.
__version__ = '.'.join([str(i) for i in __version_info__[:-1]])
if __version_info__[-1] is not None:
__version__ += ('-%s' % (__version_info__[-1],))
# context processor to add version to the template environment
def context_extras(request):
return {
# software version
'SW_VERSION': __version__
}
|
c98ac4ca313606c966dc45dbe7861898177f2f04 | api/tests/test_delete_bucket_list.py | api/tests/test_delete_bucket_list.py | import json
from api.test import BaseTestCase
from api.models import BucketList
class TestDeleteBucketList(BaseTestCase):
def test_delete_bucket_list(self):
bucket_list_one = {
"description": "Movies i have to watch by the end of the week",
"status": "Pending",
"title": "Entertainment",
"user_id": 1
}
self.client.post('/api/v1/bucketlists',
headers={
'Authorization': 'JWT ' + self.token
},
data=json.dumps(bucket_list_one),
content_type='application/json')
count = len(BucketList.query.all())
self.client.delete('/api/v1/bucketlists/1',
headers={
'Authorization': 'JWT ' + self.token
},)
new_count = len(BucketList.query.all())
self.assertEqual(new_count - count, -1)
| import json
from api.test import BaseTestCase
from api.models import BucketList
class TestDeleteBucketList(BaseTestCase):
def test_delete_bucket_list(self):
bucket_list_one = {
"description": "Movies i have to watch by the end of the week",
"status": "Pending",
"title": "Entertainment",
"user_id": 1
}
self.client.post('/api/v1/bucketlists',
headers={
'Authorization': 'JWT ' + self.token
},
data=json.dumps(bucket_list_one),
content_type='application/json')
count = len(BucketList.query.all())
self.client.delete('/api/v1/bucketlists/1',
headers={
'Authorization': 'JWT ' + self.token
},)
new_count = len(BucketList.query.all())
self.assertEqual(new_count - count, -1)
response = self.client.get(
'/api/v1/bucketlists/1',
headers=dict(
Authorization='Bearer ' + self.token
)
)
self.assertIn("Bucket list not found", str(response.data))
self.assertEqual(response.status_code, 404)
| Modify test to test that bucketlist nolonger exists in system | Modify test to test that bucketlist nolonger exists in system
| Python | mit | EdwinKato/bucket-list,EdwinKato/bucket-list,EdwinKato/bucket-list,EdwinKato/bucket-list,EdwinKato/bucket-list | import json
from api.test import BaseTestCase
from api.models import BucketList
class TestDeleteBucketList(BaseTestCase):
def test_delete_bucket_list(self):
bucket_list_one = {
"description": "Movies i have to watch by the end of the week",
"status": "Pending",
"title": "Entertainment",
"user_id": 1
}
self.client.post('/api/v1/bucketlists',
headers={
'Authorization': 'JWT ' + self.token
},
data=json.dumps(bucket_list_one),
content_type='application/json')
count = len(BucketList.query.all())
self.client.delete('/api/v1/bucketlists/1',
headers={
'Authorization': 'JWT ' + self.token
},)
new_count = len(BucketList.query.all())
self.assertEqual(new_count - count, -1)
Modify test to test that bucketlist nolonger exists in system | import json
from api.test import BaseTestCase
from api.models import BucketList
class TestDeleteBucketList(BaseTestCase):
def test_delete_bucket_list(self):
bucket_list_one = {
"description": "Movies i have to watch by the end of the week",
"status": "Pending",
"title": "Entertainment",
"user_id": 1
}
self.client.post('/api/v1/bucketlists',
headers={
'Authorization': 'JWT ' + self.token
},
data=json.dumps(bucket_list_one),
content_type='application/json')
count = len(BucketList.query.all())
self.client.delete('/api/v1/bucketlists/1',
headers={
'Authorization': 'JWT ' + self.token
},)
new_count = len(BucketList.query.all())
self.assertEqual(new_count - count, -1)
response = self.client.get(
'/api/v1/bucketlists/1',
headers=dict(
Authorization='Bearer ' + self.token
)
)
self.assertIn("Bucket list not found", str(response.data))
self.assertEqual(response.status_code, 404)
| <commit_before>import json
from api.test import BaseTestCase
from api.models import BucketList
class TestDeleteBucketList(BaseTestCase):
def test_delete_bucket_list(self):
bucket_list_one = {
"description": "Movies i have to watch by the end of the week",
"status": "Pending",
"title": "Entertainment",
"user_id": 1
}
self.client.post('/api/v1/bucketlists',
headers={
'Authorization': 'JWT ' + self.token
},
data=json.dumps(bucket_list_one),
content_type='application/json')
count = len(BucketList.query.all())
self.client.delete('/api/v1/bucketlists/1',
headers={
'Authorization': 'JWT ' + self.token
},)
new_count = len(BucketList.query.all())
self.assertEqual(new_count - count, -1)
<commit_msg>Modify test to test that bucketlist nolonger exists in system<commit_after> | import json
from api.test import BaseTestCase
from api.models import BucketList
class TestDeleteBucketList(BaseTestCase):
def test_delete_bucket_list(self):
bucket_list_one = {
"description": "Movies i have to watch by the end of the week",
"status": "Pending",
"title": "Entertainment",
"user_id": 1
}
self.client.post('/api/v1/bucketlists',
headers={
'Authorization': 'JWT ' + self.token
},
data=json.dumps(bucket_list_one),
content_type='application/json')
count = len(BucketList.query.all())
self.client.delete('/api/v1/bucketlists/1',
headers={
'Authorization': 'JWT ' + self.token
},)
new_count = len(BucketList.query.all())
self.assertEqual(new_count - count, -1)
response = self.client.get(
'/api/v1/bucketlists/1',
headers=dict(
Authorization='Bearer ' + self.token
)
)
self.assertIn("Bucket list not found", str(response.data))
self.assertEqual(response.status_code, 404)
| import json
from api.test import BaseTestCase
from api.models import BucketList
class TestDeleteBucketList(BaseTestCase):
def test_delete_bucket_list(self):
bucket_list_one = {
"description": "Movies i have to watch by the end of the week",
"status": "Pending",
"title": "Entertainment",
"user_id": 1
}
self.client.post('/api/v1/bucketlists',
headers={
'Authorization': 'JWT ' + self.token
},
data=json.dumps(bucket_list_one),
content_type='application/json')
count = len(BucketList.query.all())
self.client.delete('/api/v1/bucketlists/1',
headers={
'Authorization': 'JWT ' + self.token
},)
new_count = len(BucketList.query.all())
self.assertEqual(new_count - count, -1)
Modify test to test that bucketlist nolonger exists in systemimport json
from api.test import BaseTestCase
from api.models import BucketList
class TestDeleteBucketList(BaseTestCase):
def test_delete_bucket_list(self):
bucket_list_one = {
"description": "Movies i have to watch by the end of the week",
"status": "Pending",
"title": "Entertainment",
"user_id": 1
}
self.client.post('/api/v1/bucketlists',
headers={
'Authorization': 'JWT ' + self.token
},
data=json.dumps(bucket_list_one),
content_type='application/json')
count = len(BucketList.query.all())
self.client.delete('/api/v1/bucketlists/1',
headers={
'Authorization': 'JWT ' + self.token
},)
new_count = len(BucketList.query.all())
self.assertEqual(new_count - count, -1)
response = self.client.get(
'/api/v1/bucketlists/1',
headers=dict(
Authorization='Bearer ' + self.token
)
)
self.assertIn("Bucket list not found", str(response.data))
self.assertEqual(response.status_code, 404)
| <commit_before>import json
from api.test import BaseTestCase
from api.models import BucketList
class TestDeleteBucketList(BaseTestCase):
def test_delete_bucket_list(self):
bucket_list_one = {
"description": "Movies i have to watch by the end of the week",
"status": "Pending",
"title": "Entertainment",
"user_id": 1
}
self.client.post('/api/v1/bucketlists',
headers={
'Authorization': 'JWT ' + self.token
},
data=json.dumps(bucket_list_one),
content_type='application/json')
count = len(BucketList.query.all())
self.client.delete('/api/v1/bucketlists/1',
headers={
'Authorization': 'JWT ' + self.token
},)
new_count = len(BucketList.query.all())
self.assertEqual(new_count - count, -1)
<commit_msg>Modify test to test that bucketlist nolonger exists in system<commit_after>import json
from api.test import BaseTestCase
from api.models import BucketList
class TestDeleteBucketList(BaseTestCase):
def test_delete_bucket_list(self):
bucket_list_one = {
"description": "Movies i have to watch by the end of the week",
"status": "Pending",
"title": "Entertainment",
"user_id": 1
}
self.client.post('/api/v1/bucketlists',
headers={
'Authorization': 'JWT ' + self.token
},
data=json.dumps(bucket_list_one),
content_type='application/json')
count = len(BucketList.query.all())
self.client.delete('/api/v1/bucketlists/1',
headers={
'Authorization': 'JWT ' + self.token
},)
new_count = len(BucketList.query.all())
self.assertEqual(new_count - count, -1)
response = self.client.get(
'/api/v1/bucketlists/1',
headers=dict(
Authorization='Bearer ' + self.token
)
)
self.assertIn("Bucket list not found", str(response.data))
self.assertEqual(response.status_code, 404)
|
a2aceffa0133756d833ba0057b3d22e7f4d95406 | utils.py | utils.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
import os
def is_image_file(f):
known_extensions = ('.jpg', '.jpeg', '.png')
return os.path.isfile(f) and f.lower().endswith(known_extensions)
def get_files_list(paths, recursive):
result = []
for path in paths:
if os.path.isdir(path):
if recursive:
for root, dir_names, file_names in os.walk(path):
for filename in file_names:
f = os.path.join(root, filename)
if is_image_file(f):
result.append(f)
else:
for filename in os.listdir(path):
f = os.path.join(path, filename)
if is_image_file(f):
result.append(f)
elif is_image_file(path):
result.append(path)
else:
raise RuntimeError('unknown image format ', path)
return result
def get_file_list(path):
return get_files_list((path,), True)
def make_dir_by_file_name(file_name):
dir_name, file_name = os.path.split(file_name)
file_name_wo_ext = os.path.splitext(file_name)[0]
out_dir = os.path.join(dir_name, file_name_wo_ext)
if not os.path.exists(out_dir):
os.makedirs(out_dir)
return out_dir
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
import os
def is_image_file(f):
known_extensions = ('.bmp', '.jpg', '.jpeg', '.png')
return os.path.isfile(f) and f.lower().endswith(known_extensions)
def get_files_list(paths, recursive):
result = []
for path in paths:
if os.path.isdir(path):
if recursive:
for root, dir_names, file_names in os.walk(path):
for filename in file_names:
f = os.path.join(root, filename)
if is_image_file(f):
result.append(f)
else:
for filename in os.listdir(path):
f = os.path.join(path, filename)
if is_image_file(f):
result.append(f)
elif is_image_file(path):
result.append(path)
else:
raise RuntimeError('unknown image format ', path)
return result
def get_file_list(path):
return get_files_list((path,), True)
def make_dir_by_file_name(file_name):
dir_name, file_name = os.path.split(file_name)
file_name_wo_ext = os.path.splitext(file_name)[0]
out_dir = os.path.join(dir_name, file_name_wo_ext)
if not os.path.exists(out_dir):
os.makedirs(out_dir)
return out_dir
| Add BMP files to supported list | Add BMP files to supported list
| Python | mit | vladimirgamalian/pictools | #!/usr/bin/env python
# -*- coding: utf-8 -*-
import os
def is_image_file(f):
known_extensions = ('.jpg', '.jpeg', '.png')
return os.path.isfile(f) and f.lower().endswith(known_extensions)
def get_files_list(paths, recursive):
result = []
for path in paths:
if os.path.isdir(path):
if recursive:
for root, dir_names, file_names in os.walk(path):
for filename in file_names:
f = os.path.join(root, filename)
if is_image_file(f):
result.append(f)
else:
for filename in os.listdir(path):
f = os.path.join(path, filename)
if is_image_file(f):
result.append(f)
elif is_image_file(path):
result.append(path)
else:
raise RuntimeError('unknown image format ', path)
return result
def get_file_list(path):
return get_files_list((path,), True)
def make_dir_by_file_name(file_name):
dir_name, file_name = os.path.split(file_name)
file_name_wo_ext = os.path.splitext(file_name)[0]
out_dir = os.path.join(dir_name, file_name_wo_ext)
if not os.path.exists(out_dir):
os.makedirs(out_dir)
return out_dir
Add BMP files to supported list | #!/usr/bin/env python
# -*- coding: utf-8 -*-
import os
def is_image_file(f):
known_extensions = ('.bmp', '.jpg', '.jpeg', '.png')
return os.path.isfile(f) and f.lower().endswith(known_extensions)
def get_files_list(paths, recursive):
result = []
for path in paths:
if os.path.isdir(path):
if recursive:
for root, dir_names, file_names in os.walk(path):
for filename in file_names:
f = os.path.join(root, filename)
if is_image_file(f):
result.append(f)
else:
for filename in os.listdir(path):
f = os.path.join(path, filename)
if is_image_file(f):
result.append(f)
elif is_image_file(path):
result.append(path)
else:
raise RuntimeError('unknown image format ', path)
return result
def get_file_list(path):
return get_files_list((path,), True)
def make_dir_by_file_name(file_name):
dir_name, file_name = os.path.split(file_name)
file_name_wo_ext = os.path.splitext(file_name)[0]
out_dir = os.path.join(dir_name, file_name_wo_ext)
if not os.path.exists(out_dir):
os.makedirs(out_dir)
return out_dir
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
import os
def is_image_file(f):
known_extensions = ('.jpg', '.jpeg', '.png')
return os.path.isfile(f) and f.lower().endswith(known_extensions)
def get_files_list(paths, recursive):
result = []
for path in paths:
if os.path.isdir(path):
if recursive:
for root, dir_names, file_names in os.walk(path):
for filename in file_names:
f = os.path.join(root, filename)
if is_image_file(f):
result.append(f)
else:
for filename in os.listdir(path):
f = os.path.join(path, filename)
if is_image_file(f):
result.append(f)
elif is_image_file(path):
result.append(path)
else:
raise RuntimeError('unknown image format ', path)
return result
def get_file_list(path):
return get_files_list((path,), True)
def make_dir_by_file_name(file_name):
dir_name, file_name = os.path.split(file_name)
file_name_wo_ext = os.path.splitext(file_name)[0]
out_dir = os.path.join(dir_name, file_name_wo_ext)
if not os.path.exists(out_dir):
os.makedirs(out_dir)
return out_dir
<commit_msg>Add BMP files to supported list<commit_after> | #!/usr/bin/env python
# -*- coding: utf-8 -*-
import os
def is_image_file(f):
known_extensions = ('.bmp', '.jpg', '.jpeg', '.png')
return os.path.isfile(f) and f.lower().endswith(known_extensions)
def get_files_list(paths, recursive):
result = []
for path in paths:
if os.path.isdir(path):
if recursive:
for root, dir_names, file_names in os.walk(path):
for filename in file_names:
f = os.path.join(root, filename)
if is_image_file(f):
result.append(f)
else:
for filename in os.listdir(path):
f = os.path.join(path, filename)
if is_image_file(f):
result.append(f)
elif is_image_file(path):
result.append(path)
else:
raise RuntimeError('unknown image format ', path)
return result
def get_file_list(path):
return get_files_list((path,), True)
def make_dir_by_file_name(file_name):
dir_name, file_name = os.path.split(file_name)
file_name_wo_ext = os.path.splitext(file_name)[0]
out_dir = os.path.join(dir_name, file_name_wo_ext)
if not os.path.exists(out_dir):
os.makedirs(out_dir)
return out_dir
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
import os
def is_image_file(f):
known_extensions = ('.jpg', '.jpeg', '.png')
return os.path.isfile(f) and f.lower().endswith(known_extensions)
def get_files_list(paths, recursive):
result = []
for path in paths:
if os.path.isdir(path):
if recursive:
for root, dir_names, file_names in os.walk(path):
for filename in file_names:
f = os.path.join(root, filename)
if is_image_file(f):
result.append(f)
else:
for filename in os.listdir(path):
f = os.path.join(path, filename)
if is_image_file(f):
result.append(f)
elif is_image_file(path):
result.append(path)
else:
raise RuntimeError('unknown image format ', path)
return result
def get_file_list(path):
return get_files_list((path,), True)
def make_dir_by_file_name(file_name):
dir_name, file_name = os.path.split(file_name)
file_name_wo_ext = os.path.splitext(file_name)[0]
out_dir = os.path.join(dir_name, file_name_wo_ext)
if not os.path.exists(out_dir):
os.makedirs(out_dir)
return out_dir
Add BMP files to supported list#!/usr/bin/env python
# -*- coding: utf-8 -*-
import os
def is_image_file(f):
known_extensions = ('.bmp', '.jpg', '.jpeg', '.png')
return os.path.isfile(f) and f.lower().endswith(known_extensions)
def get_files_list(paths, recursive):
result = []
for path in paths:
if os.path.isdir(path):
if recursive:
for root, dir_names, file_names in os.walk(path):
for filename in file_names:
f = os.path.join(root, filename)
if is_image_file(f):
result.append(f)
else:
for filename in os.listdir(path):
f = os.path.join(path, filename)
if is_image_file(f):
result.append(f)
elif is_image_file(path):
result.append(path)
else:
raise RuntimeError('unknown image format ', path)
return result
def get_file_list(path):
return get_files_list((path,), True)
def make_dir_by_file_name(file_name):
dir_name, file_name = os.path.split(file_name)
file_name_wo_ext = os.path.splitext(file_name)[0]
out_dir = os.path.join(dir_name, file_name_wo_ext)
if not os.path.exists(out_dir):
os.makedirs(out_dir)
return out_dir
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
import os
def is_image_file(f):
known_extensions = ('.jpg', '.jpeg', '.png')
return os.path.isfile(f) and f.lower().endswith(known_extensions)
def get_files_list(paths, recursive):
result = []
for path in paths:
if os.path.isdir(path):
if recursive:
for root, dir_names, file_names in os.walk(path):
for filename in file_names:
f = os.path.join(root, filename)
if is_image_file(f):
result.append(f)
else:
for filename in os.listdir(path):
f = os.path.join(path, filename)
if is_image_file(f):
result.append(f)
elif is_image_file(path):
result.append(path)
else:
raise RuntimeError('unknown image format ', path)
return result
def get_file_list(path):
return get_files_list((path,), True)
def make_dir_by_file_name(file_name):
dir_name, file_name = os.path.split(file_name)
file_name_wo_ext = os.path.splitext(file_name)[0]
out_dir = os.path.join(dir_name, file_name_wo_ext)
if not os.path.exists(out_dir):
os.makedirs(out_dir)
return out_dir
<commit_msg>Add BMP files to supported list<commit_after>#!/usr/bin/env python
# -*- coding: utf-8 -*-
import os
def is_image_file(f):
known_extensions = ('.bmp', '.jpg', '.jpeg', '.png')
return os.path.isfile(f) and f.lower().endswith(known_extensions)
def get_files_list(paths, recursive):
result = []
for path in paths:
if os.path.isdir(path):
if recursive:
for root, dir_names, file_names in os.walk(path):
for filename in file_names:
f = os.path.join(root, filename)
if is_image_file(f):
result.append(f)
else:
for filename in os.listdir(path):
f = os.path.join(path, filename)
if is_image_file(f):
result.append(f)
elif is_image_file(path):
result.append(path)
else:
raise RuntimeError('unknown image format ', path)
return result
def get_file_list(path):
return get_files_list((path,), True)
def make_dir_by_file_name(file_name):
dir_name, file_name = os.path.split(file_name)
file_name_wo_ext = os.path.splitext(file_name)[0]
out_dir = os.path.join(dir_name, file_name_wo_ext)
if not os.path.exists(out_dir):
os.makedirs(out_dir)
return out_dir
|
ae2be1dc39baa8f8cd73e574d384619290b0c707 | tests/api/views/users/read_test.py | tests/api/views/users/read_test.py | from tests.data import add_fixtures, users
def test_read_user(db_session, client):
john = users.john()
add_fixtures(db_session, john)
res = client.get('/users/{id}'.format(id=john.id))
assert res.status_code == 200
assert res.json == {
u'id': john.id,
u'firstName': u'John',
u'lastName': u'Doe',
u'name': u'John Doe',
u'club': None,
u'trackingCallsign': None,
u'trackingDelay': 0,
u'followers': 0,
u'following': 0,
}
def test_read_missing_user(client):
res = client.get('/users/1000000000000')
assert res.status_code == 404
def test_read_user_with_invalid_id(client):
res = client.get('/users/abc')
assert res.status_code == 404
| from skylines.model import Follower
from tests.api import auth_for
from tests.data import add_fixtures, users
def test_read_user(db_session, client):
john = users.john()
add_fixtures(db_session, john)
res = client.get('/users/{id}'.format(id=john.id))
assert res.status_code == 200
assert res.json == {
u'id': john.id,
u'firstName': u'John',
u'lastName': u'Doe',
u'name': u'John Doe',
u'club': None,
u'trackingCallsign': None,
u'trackingDelay': 0,
u'followers': 0,
u'following': 0,
}
def test_following(db_session, client):
john = users.john()
jane = users.jane()
Follower.follow(john, jane)
add_fixtures(db_session, john, jane)
res = client.get('/users/{id}'.format(id=john.id))
assert res.status_code == 200
assert res.json['following'] == 1
res = client.get('/users/{id}'.format(id=jane.id))
assert res.status_code == 200
assert res.json['followers'] == 1
assert 'followed' not in res.json
res = client.get('/users/{id}'.format(id=jane.id), headers=auth_for(john))
assert res.status_code == 200
assert res.json['followers'] == 1
assert res.json['followed'] == True
def test_read_missing_user(client):
res = client.get('/users/1000000000000')
assert res.status_code == 404
def test_read_user_with_invalid_id(client):
res = client.get('/users/abc')
assert res.status_code == 404
| Add more "GET /users/:id" tests | tests/api: Add more "GET /users/:id" tests
| Python | agpl-3.0 | Turbo87/skylines,skylines-project/skylines,shadowoneau/skylines,Harry-R/skylines,shadowoneau/skylines,RBE-Avionik/skylines,skylines-project/skylines,RBE-Avionik/skylines,Turbo87/skylines,RBE-Avionik/skylines,shadowoneau/skylines,Harry-R/skylines,Turbo87/skylines,skylines-project/skylines,Harry-R/skylines,Harry-R/skylines,shadowoneau/skylines,skylines-project/skylines,RBE-Avionik/skylines,Turbo87/skylines | from tests.data import add_fixtures, users
def test_read_user(db_session, client):
john = users.john()
add_fixtures(db_session, john)
res = client.get('/users/{id}'.format(id=john.id))
assert res.status_code == 200
assert res.json == {
u'id': john.id,
u'firstName': u'John',
u'lastName': u'Doe',
u'name': u'John Doe',
u'club': None,
u'trackingCallsign': None,
u'trackingDelay': 0,
u'followers': 0,
u'following': 0,
}
def test_read_missing_user(client):
res = client.get('/users/1000000000000')
assert res.status_code == 404
def test_read_user_with_invalid_id(client):
res = client.get('/users/abc')
assert res.status_code == 404
tests/api: Add more "GET /users/:id" tests | from skylines.model import Follower
from tests.api import auth_for
from tests.data import add_fixtures, users
def test_read_user(db_session, client):
john = users.john()
add_fixtures(db_session, john)
res = client.get('/users/{id}'.format(id=john.id))
assert res.status_code == 200
assert res.json == {
u'id': john.id,
u'firstName': u'John',
u'lastName': u'Doe',
u'name': u'John Doe',
u'club': None,
u'trackingCallsign': None,
u'trackingDelay': 0,
u'followers': 0,
u'following': 0,
}
def test_following(db_session, client):
john = users.john()
jane = users.jane()
Follower.follow(john, jane)
add_fixtures(db_session, john, jane)
res = client.get('/users/{id}'.format(id=john.id))
assert res.status_code == 200
assert res.json['following'] == 1
res = client.get('/users/{id}'.format(id=jane.id))
assert res.status_code == 200
assert res.json['followers'] == 1
assert 'followed' not in res.json
res = client.get('/users/{id}'.format(id=jane.id), headers=auth_for(john))
assert res.status_code == 200
assert res.json['followers'] == 1
assert res.json['followed'] == True
def test_read_missing_user(client):
res = client.get('/users/1000000000000')
assert res.status_code == 404
def test_read_user_with_invalid_id(client):
res = client.get('/users/abc')
assert res.status_code == 404
| <commit_before>from tests.data import add_fixtures, users
def test_read_user(db_session, client):
john = users.john()
add_fixtures(db_session, john)
res = client.get('/users/{id}'.format(id=john.id))
assert res.status_code == 200
assert res.json == {
u'id': john.id,
u'firstName': u'John',
u'lastName': u'Doe',
u'name': u'John Doe',
u'club': None,
u'trackingCallsign': None,
u'trackingDelay': 0,
u'followers': 0,
u'following': 0,
}
def test_read_missing_user(client):
res = client.get('/users/1000000000000')
assert res.status_code == 404
def test_read_user_with_invalid_id(client):
res = client.get('/users/abc')
assert res.status_code == 404
<commit_msg>tests/api: Add more "GET /users/:id" tests<commit_after> | from skylines.model import Follower
from tests.api import auth_for
from tests.data import add_fixtures, users
def test_read_user(db_session, client):
john = users.john()
add_fixtures(db_session, john)
res = client.get('/users/{id}'.format(id=john.id))
assert res.status_code == 200
assert res.json == {
u'id': john.id,
u'firstName': u'John',
u'lastName': u'Doe',
u'name': u'John Doe',
u'club': None,
u'trackingCallsign': None,
u'trackingDelay': 0,
u'followers': 0,
u'following': 0,
}
def test_following(db_session, client):
john = users.john()
jane = users.jane()
Follower.follow(john, jane)
add_fixtures(db_session, john, jane)
res = client.get('/users/{id}'.format(id=john.id))
assert res.status_code == 200
assert res.json['following'] == 1
res = client.get('/users/{id}'.format(id=jane.id))
assert res.status_code == 200
assert res.json['followers'] == 1
assert 'followed' not in res.json
res = client.get('/users/{id}'.format(id=jane.id), headers=auth_for(john))
assert res.status_code == 200
assert res.json['followers'] == 1
assert res.json['followed'] == True
def test_read_missing_user(client):
res = client.get('/users/1000000000000')
assert res.status_code == 404
def test_read_user_with_invalid_id(client):
res = client.get('/users/abc')
assert res.status_code == 404
| from tests.data import add_fixtures, users
def test_read_user(db_session, client):
john = users.john()
add_fixtures(db_session, john)
res = client.get('/users/{id}'.format(id=john.id))
assert res.status_code == 200
assert res.json == {
u'id': john.id,
u'firstName': u'John',
u'lastName': u'Doe',
u'name': u'John Doe',
u'club': None,
u'trackingCallsign': None,
u'trackingDelay': 0,
u'followers': 0,
u'following': 0,
}
def test_read_missing_user(client):
res = client.get('/users/1000000000000')
assert res.status_code == 404
def test_read_user_with_invalid_id(client):
res = client.get('/users/abc')
assert res.status_code == 404
tests/api: Add more "GET /users/:id" testsfrom skylines.model import Follower
from tests.api import auth_for
from tests.data import add_fixtures, users
def test_read_user(db_session, client):
john = users.john()
add_fixtures(db_session, john)
res = client.get('/users/{id}'.format(id=john.id))
assert res.status_code == 200
assert res.json == {
u'id': john.id,
u'firstName': u'John',
u'lastName': u'Doe',
u'name': u'John Doe',
u'club': None,
u'trackingCallsign': None,
u'trackingDelay': 0,
u'followers': 0,
u'following': 0,
}
def test_following(db_session, client):
john = users.john()
jane = users.jane()
Follower.follow(john, jane)
add_fixtures(db_session, john, jane)
res = client.get('/users/{id}'.format(id=john.id))
assert res.status_code == 200
assert res.json['following'] == 1
res = client.get('/users/{id}'.format(id=jane.id))
assert res.status_code == 200
assert res.json['followers'] == 1
assert 'followed' not in res.json
res = client.get('/users/{id}'.format(id=jane.id), headers=auth_for(john))
assert res.status_code == 200
assert res.json['followers'] == 1
assert res.json['followed'] == True
def test_read_missing_user(client):
res = client.get('/users/1000000000000')
assert res.status_code == 404
def test_read_user_with_invalid_id(client):
res = client.get('/users/abc')
assert res.status_code == 404
| <commit_before>from tests.data import add_fixtures, users
def test_read_user(db_session, client):
john = users.john()
add_fixtures(db_session, john)
res = client.get('/users/{id}'.format(id=john.id))
assert res.status_code == 200
assert res.json == {
u'id': john.id,
u'firstName': u'John',
u'lastName': u'Doe',
u'name': u'John Doe',
u'club': None,
u'trackingCallsign': None,
u'trackingDelay': 0,
u'followers': 0,
u'following': 0,
}
def test_read_missing_user(client):
res = client.get('/users/1000000000000')
assert res.status_code == 404
def test_read_user_with_invalid_id(client):
res = client.get('/users/abc')
assert res.status_code == 404
<commit_msg>tests/api: Add more "GET /users/:id" tests<commit_after>from skylines.model import Follower
from tests.api import auth_for
from tests.data import add_fixtures, users
def test_read_user(db_session, client):
john = users.john()
add_fixtures(db_session, john)
res = client.get('/users/{id}'.format(id=john.id))
assert res.status_code == 200
assert res.json == {
u'id': john.id,
u'firstName': u'John',
u'lastName': u'Doe',
u'name': u'John Doe',
u'club': None,
u'trackingCallsign': None,
u'trackingDelay': 0,
u'followers': 0,
u'following': 0,
}
def test_following(db_session, client):
john = users.john()
jane = users.jane()
Follower.follow(john, jane)
add_fixtures(db_session, john, jane)
res = client.get('/users/{id}'.format(id=john.id))
assert res.status_code == 200
assert res.json['following'] == 1
res = client.get('/users/{id}'.format(id=jane.id))
assert res.status_code == 200
assert res.json['followers'] == 1
assert 'followed' not in res.json
res = client.get('/users/{id}'.format(id=jane.id), headers=auth_for(john))
assert res.status_code == 200
assert res.json['followers'] == 1
assert res.json['followed'] == True
def test_read_missing_user(client):
res = client.get('/users/1000000000000')
assert res.status_code == 404
def test_read_user_with_invalid_id(client):
res = client.get('/users/abc')
assert res.status_code == 404
|
788284da7d586b538fefb3beb751938fc555d923 | ovp_users/models/profile.py | ovp_users/models/profile.py | from django.db import models
from django.utils.translation import ugettext_lazy as _
from ovp_users.helpers import get_settings, import_from_string
gender_choices = (
("male", "Male"),
("female", "Female"),
("unspecified", "Unspecified"),
)
class UserProfile(models.Model):
user = models.OneToOneField("User", blank=True, null=True, related_name="%(app_label)s_%(class)s_profile")
full_name = models.CharField(_("Full name"), max_length=300, null=True, blank=True)
skills = models.ManyToManyField("ovp_core.Skill")
causes = models.ManyToManyField("ovp_core.Cause")
about = models.TextField(_("About me"), null=True, blank=True)
gender = models.CharField(_("Gender"), max_length=10, choices=gender_choices, default='unspecified')
def get_profile_model():
s = get_settings()
class_path = s.get("PROFILE_MODEL", None)
if class_path:
return import_from_string(class_path)
return UserProfile
| from django.db import models
from django.utils.translation import ugettext_lazy as _
from ovp_users.helpers import get_settings, import_from_string
gender_choices = (
("male", "Male"),
("female", "Female"),
("unspecified", "Unspecified"),
)
class UserProfile(models.Model):
user = models.OneToOneField("User", blank=True, null=True, related_name="%(app_label)s_%(class)s_profile")
full_name = models.CharField(_("Full name"), max_length=300, null=True, blank=True)
skills = models.ManyToManyField("ovp_core.Skill")
causes = models.ManyToManyField("ovp_core.Cause")
about = models.TextField(_("About me"), null=True, blank=True)
gender = models.CharField(_("Gender"), max_length=20, choices=gender_choices, default='unspecified')
def get_profile_model():
s = get_settings()
class_path = s.get("PROFILE_MODEL", None)
if class_path:
return import_from_string(class_path)
return UserProfile
| Set Profile.gender maxlength to 20 | Set Profile.gender maxlength to 20
| Python | agpl-3.0 | OpenVolunteeringPlatform/django-ovp-users,OpenVolunteeringPlatform/django-ovp-users | from django.db import models
from django.utils.translation import ugettext_lazy as _
from ovp_users.helpers import get_settings, import_from_string
gender_choices = (
("male", "Male"),
("female", "Female"),
("unspecified", "Unspecified"),
)
class UserProfile(models.Model):
user = models.OneToOneField("User", blank=True, null=True, related_name="%(app_label)s_%(class)s_profile")
full_name = models.CharField(_("Full name"), max_length=300, null=True, blank=True)
skills = models.ManyToManyField("ovp_core.Skill")
causes = models.ManyToManyField("ovp_core.Cause")
about = models.TextField(_("About me"), null=True, blank=True)
gender = models.CharField(_("Gender"), max_length=10, choices=gender_choices, default='unspecified')
def get_profile_model():
s = get_settings()
class_path = s.get("PROFILE_MODEL", None)
if class_path:
return import_from_string(class_path)
return UserProfile
Set Profile.gender maxlength to 20 | from django.db import models
from django.utils.translation import ugettext_lazy as _
from ovp_users.helpers import get_settings, import_from_string
gender_choices = (
("male", "Male"),
("female", "Female"),
("unspecified", "Unspecified"),
)
class UserProfile(models.Model):
user = models.OneToOneField("User", blank=True, null=True, related_name="%(app_label)s_%(class)s_profile")
full_name = models.CharField(_("Full name"), max_length=300, null=True, blank=True)
skills = models.ManyToManyField("ovp_core.Skill")
causes = models.ManyToManyField("ovp_core.Cause")
about = models.TextField(_("About me"), null=True, blank=True)
gender = models.CharField(_("Gender"), max_length=20, choices=gender_choices, default='unspecified')
def get_profile_model():
s = get_settings()
class_path = s.get("PROFILE_MODEL", None)
if class_path:
return import_from_string(class_path)
return UserProfile
| <commit_before>from django.db import models
from django.utils.translation import ugettext_lazy as _
from ovp_users.helpers import get_settings, import_from_string
gender_choices = (
("male", "Male"),
("female", "Female"),
("unspecified", "Unspecified"),
)
class UserProfile(models.Model):
user = models.OneToOneField("User", blank=True, null=True, related_name="%(app_label)s_%(class)s_profile")
full_name = models.CharField(_("Full name"), max_length=300, null=True, blank=True)
skills = models.ManyToManyField("ovp_core.Skill")
causes = models.ManyToManyField("ovp_core.Cause")
about = models.TextField(_("About me"), null=True, blank=True)
gender = models.CharField(_("Gender"), max_length=10, choices=gender_choices, default='unspecified')
def get_profile_model():
s = get_settings()
class_path = s.get("PROFILE_MODEL", None)
if class_path:
return import_from_string(class_path)
return UserProfile
<commit_msg>Set Profile.gender maxlength to 20<commit_after> | from django.db import models
from django.utils.translation import ugettext_lazy as _
from ovp_users.helpers import get_settings, import_from_string
gender_choices = (
("male", "Male"),
("female", "Female"),
("unspecified", "Unspecified"),
)
class UserProfile(models.Model):
user = models.OneToOneField("User", blank=True, null=True, related_name="%(app_label)s_%(class)s_profile")
full_name = models.CharField(_("Full name"), max_length=300, null=True, blank=True)
skills = models.ManyToManyField("ovp_core.Skill")
causes = models.ManyToManyField("ovp_core.Cause")
about = models.TextField(_("About me"), null=True, blank=True)
gender = models.CharField(_("Gender"), max_length=20, choices=gender_choices, default='unspecified')
def get_profile_model():
s = get_settings()
class_path = s.get("PROFILE_MODEL", None)
if class_path:
return import_from_string(class_path)
return UserProfile
| from django.db import models
from django.utils.translation import ugettext_lazy as _
from ovp_users.helpers import get_settings, import_from_string
gender_choices = (
("male", "Male"),
("female", "Female"),
("unspecified", "Unspecified"),
)
class UserProfile(models.Model):
user = models.OneToOneField("User", blank=True, null=True, related_name="%(app_label)s_%(class)s_profile")
full_name = models.CharField(_("Full name"), max_length=300, null=True, blank=True)
skills = models.ManyToManyField("ovp_core.Skill")
causes = models.ManyToManyField("ovp_core.Cause")
about = models.TextField(_("About me"), null=True, blank=True)
gender = models.CharField(_("Gender"), max_length=10, choices=gender_choices, default='unspecified')
def get_profile_model():
s = get_settings()
class_path = s.get("PROFILE_MODEL", None)
if class_path:
return import_from_string(class_path)
return UserProfile
Set Profile.gender maxlength to 20from django.db import models
from django.utils.translation import ugettext_lazy as _
from ovp_users.helpers import get_settings, import_from_string
gender_choices = (
("male", "Male"),
("female", "Female"),
("unspecified", "Unspecified"),
)
class UserProfile(models.Model):
user = models.OneToOneField("User", blank=True, null=True, related_name="%(app_label)s_%(class)s_profile")
full_name = models.CharField(_("Full name"), max_length=300, null=True, blank=True)
skills = models.ManyToManyField("ovp_core.Skill")
causes = models.ManyToManyField("ovp_core.Cause")
about = models.TextField(_("About me"), null=True, blank=True)
gender = models.CharField(_("Gender"), max_length=20, choices=gender_choices, default='unspecified')
def get_profile_model():
s = get_settings()
class_path = s.get("PROFILE_MODEL", None)
if class_path:
return import_from_string(class_path)
return UserProfile
| <commit_before>from django.db import models
from django.utils.translation import ugettext_lazy as _
from ovp_users.helpers import get_settings, import_from_string
gender_choices = (
("male", "Male"),
("female", "Female"),
("unspecified", "Unspecified"),
)
class UserProfile(models.Model):
user = models.OneToOneField("User", blank=True, null=True, related_name="%(app_label)s_%(class)s_profile")
full_name = models.CharField(_("Full name"), max_length=300, null=True, blank=True)
skills = models.ManyToManyField("ovp_core.Skill")
causes = models.ManyToManyField("ovp_core.Cause")
about = models.TextField(_("About me"), null=True, blank=True)
gender = models.CharField(_("Gender"), max_length=10, choices=gender_choices, default='unspecified')
def get_profile_model():
s = get_settings()
class_path = s.get("PROFILE_MODEL", None)
if class_path:
return import_from_string(class_path)
return UserProfile
<commit_msg>Set Profile.gender maxlength to 20<commit_after>from django.db import models
from django.utils.translation import ugettext_lazy as _
from ovp_users.helpers import get_settings, import_from_string
gender_choices = (
("male", "Male"),
("female", "Female"),
("unspecified", "Unspecified"),
)
class UserProfile(models.Model):
user = models.OneToOneField("User", blank=True, null=True, related_name="%(app_label)s_%(class)s_profile")
full_name = models.CharField(_("Full name"), max_length=300, null=True, blank=True)
skills = models.ManyToManyField("ovp_core.Skill")
causes = models.ManyToManyField("ovp_core.Cause")
about = models.TextField(_("About me"), null=True, blank=True)
gender = models.CharField(_("Gender"), max_length=20, choices=gender_choices, default='unspecified')
def get_profile_model():
s = get_settings()
class_path = s.get("PROFILE_MODEL", None)
if class_path:
return import_from_string(class_path)
return UserProfile
|
da3c17d9142161c9dd9136e604fb9d0f82355044 | tests/tools_tests.py | tests/tools_tests.py | """Tests for ifcfg.tools."""
import logging
import os
import unittest
import ifcfg
from ifcfg.tools import exec_cmd
from nose.tools import eq_
class IfcfgToolsTestCase(unittest.TestCase):
def test_minimal_logger(self):
os.environ['IFCFG_DEBUG'] = '1'
log = ifcfg.tools.minimal_logger(__name__)
eq_(log.level, logging.DEBUG)
os.environ['IFCFG_DEBUG'] = '0'
def test_command(self):
output, __, __ = exec_cmd("echo -n 'this is a test'")
self.assertEqual(output, "this is a test")
| """Tests for ifcfg.tools."""
import locale
import logging
import os
import sys
import unittest
import ifcfg
from ifcfg.tools import exec_cmd
from nose.tools import eq_
class IfcfgToolsTestCase(unittest.TestCase):
def test_minimal_logger(self):
os.environ['IFCFG_DEBUG'] = '1'
log = ifcfg.tools.minimal_logger(__name__)
eq_(log.level, logging.DEBUG)
os.environ['IFCFG_DEBUG'] = '0'
def test_command(self):
output, __, __ = exec_cmd("echo -n 'this is a test'")
self.assertEqual(output, "this is a test")
@unittest.skipIf(sys.version[0] != '2',
"Python 2 only supports non-unicode stuff")
def test_command_non_unicode(self):
getpreferredencoding_orig = locale.getpreferredencoding
locale.getpreferredencoding = lambda: "ISO-8859-1"
output, __, __ = exec_cmd("echo -n 'this is a test'")
self.assertEqual(output, "this is a test")
locale.getpreferredencoding = getpreferredencoding_orig
| Add a Python2 test for non-unicode commands | Add a Python2 test for non-unicode commands
| Python | bsd-3-clause | ftao/python-ifcfg | """Tests for ifcfg.tools."""
import logging
import os
import unittest
import ifcfg
from ifcfg.tools import exec_cmd
from nose.tools import eq_
class IfcfgToolsTestCase(unittest.TestCase):
def test_minimal_logger(self):
os.environ['IFCFG_DEBUG'] = '1'
log = ifcfg.tools.minimal_logger(__name__)
eq_(log.level, logging.DEBUG)
os.environ['IFCFG_DEBUG'] = '0'
def test_command(self):
output, __, __ = exec_cmd("echo -n 'this is a test'")
self.assertEqual(output, "this is a test")
Add a Python2 test for non-unicode commands | """Tests for ifcfg.tools."""
import locale
import logging
import os
import sys
import unittest
import ifcfg
from ifcfg.tools import exec_cmd
from nose.tools import eq_
class IfcfgToolsTestCase(unittest.TestCase):
def test_minimal_logger(self):
os.environ['IFCFG_DEBUG'] = '1'
log = ifcfg.tools.minimal_logger(__name__)
eq_(log.level, logging.DEBUG)
os.environ['IFCFG_DEBUG'] = '0'
def test_command(self):
output, __, __ = exec_cmd("echo -n 'this is a test'")
self.assertEqual(output, "this is a test")
@unittest.skipIf(sys.version[0] != '2',
"Python 2 only supports non-unicode stuff")
def test_command_non_unicode(self):
getpreferredencoding_orig = locale.getpreferredencoding
locale.getpreferredencoding = lambda: "ISO-8859-1"
output, __, __ = exec_cmd("echo -n 'this is a test'")
self.assertEqual(output, "this is a test")
locale.getpreferredencoding = getpreferredencoding_orig
| <commit_before>"""Tests for ifcfg.tools."""
import logging
import os
import unittest
import ifcfg
from ifcfg.tools import exec_cmd
from nose.tools import eq_
class IfcfgToolsTestCase(unittest.TestCase):
def test_minimal_logger(self):
os.environ['IFCFG_DEBUG'] = '1'
log = ifcfg.tools.minimal_logger(__name__)
eq_(log.level, logging.DEBUG)
os.environ['IFCFG_DEBUG'] = '0'
def test_command(self):
output, __, __ = exec_cmd("echo -n 'this is a test'")
self.assertEqual(output, "this is a test")
<commit_msg>Add a Python2 test for non-unicode commands<commit_after> | """Tests for ifcfg.tools."""
import locale
import logging
import os
import sys
import unittest
import ifcfg
from ifcfg.tools import exec_cmd
from nose.tools import eq_
class IfcfgToolsTestCase(unittest.TestCase):
def test_minimal_logger(self):
os.environ['IFCFG_DEBUG'] = '1'
log = ifcfg.tools.minimal_logger(__name__)
eq_(log.level, logging.DEBUG)
os.environ['IFCFG_DEBUG'] = '0'
def test_command(self):
output, __, __ = exec_cmd("echo -n 'this is a test'")
self.assertEqual(output, "this is a test")
@unittest.skipIf(sys.version[0] != '2',
"Python 2 only supports non-unicode stuff")
def test_command_non_unicode(self):
getpreferredencoding_orig = locale.getpreferredencoding
locale.getpreferredencoding = lambda: "ISO-8859-1"
output, __, __ = exec_cmd("echo -n 'this is a test'")
self.assertEqual(output, "this is a test")
locale.getpreferredencoding = getpreferredencoding_orig
| """Tests for ifcfg.tools."""
import logging
import os
import unittest
import ifcfg
from ifcfg.tools import exec_cmd
from nose.tools import eq_
class IfcfgToolsTestCase(unittest.TestCase):
def test_minimal_logger(self):
os.environ['IFCFG_DEBUG'] = '1'
log = ifcfg.tools.minimal_logger(__name__)
eq_(log.level, logging.DEBUG)
os.environ['IFCFG_DEBUG'] = '0'
def test_command(self):
output, __, __ = exec_cmd("echo -n 'this is a test'")
self.assertEqual(output, "this is a test")
Add a Python2 test for non-unicode commands"""Tests for ifcfg.tools."""
import locale
import logging
import os
import sys
import unittest
import ifcfg
from ifcfg.tools import exec_cmd
from nose.tools import eq_
class IfcfgToolsTestCase(unittest.TestCase):
def test_minimal_logger(self):
os.environ['IFCFG_DEBUG'] = '1'
log = ifcfg.tools.minimal_logger(__name__)
eq_(log.level, logging.DEBUG)
os.environ['IFCFG_DEBUG'] = '0'
def test_command(self):
output, __, __ = exec_cmd("echo -n 'this is a test'")
self.assertEqual(output, "this is a test")
@unittest.skipIf(sys.version[0] != '2',
"Python 2 only supports non-unicode stuff")
def test_command_non_unicode(self):
getpreferredencoding_orig = locale.getpreferredencoding
locale.getpreferredencoding = lambda: "ISO-8859-1"
output, __, __ = exec_cmd("echo -n 'this is a test'")
self.assertEqual(output, "this is a test")
locale.getpreferredencoding = getpreferredencoding_orig
| <commit_before>"""Tests for ifcfg.tools."""
import logging
import os
import unittest
import ifcfg
from ifcfg.tools import exec_cmd
from nose.tools import eq_
class IfcfgToolsTestCase(unittest.TestCase):
def test_minimal_logger(self):
os.environ['IFCFG_DEBUG'] = '1'
log = ifcfg.tools.minimal_logger(__name__)
eq_(log.level, logging.DEBUG)
os.environ['IFCFG_DEBUG'] = '0'
def test_command(self):
output, __, __ = exec_cmd("echo -n 'this is a test'")
self.assertEqual(output, "this is a test")
<commit_msg>Add a Python2 test for non-unicode commands<commit_after>"""Tests for ifcfg.tools."""
import locale
import logging
import os
import sys
import unittest
import ifcfg
from ifcfg.tools import exec_cmd
from nose.tools import eq_
class IfcfgToolsTestCase(unittest.TestCase):
def test_minimal_logger(self):
os.environ['IFCFG_DEBUG'] = '1'
log = ifcfg.tools.minimal_logger(__name__)
eq_(log.level, logging.DEBUG)
os.environ['IFCFG_DEBUG'] = '0'
def test_command(self):
output, __, __ = exec_cmd("echo -n 'this is a test'")
self.assertEqual(output, "this is a test")
@unittest.skipIf(sys.version[0] != '2',
"Python 2 only supports non-unicode stuff")
def test_command_non_unicode(self):
getpreferredencoding_orig = locale.getpreferredencoding
locale.getpreferredencoding = lambda: "ISO-8859-1"
output, __, __ = exec_cmd("echo -n 'this is a test'")
self.assertEqual(output, "this is a test")
locale.getpreferredencoding = getpreferredencoding_orig
|
dfa94ee2f7712ed66157ff0024989025831bf6ac | calaccess_website/urls.py | calaccess_website/urls.py | from django.conf.urls import url
from calaccess_website import views
urlpatterns = [
# The homepage
url(
r'^$',
views.VersionArchiveIndex.as_view(),
name="version_index",
),
# Version archive views
url(
r'^versions/archive/(?P<year>[0-9]{4})/$',
views.VersionYearArchiveList.as_view(),
name="version_year_archive"
),
url(
r'^versions/(?P<pk>[0-9]{1,})/$',
views.VersionDetail.as_view(),
name="version_detail"
),
url(
r'^versions/latest/$',
views.LatestVersion.as_view(),
name='version_latest_redirect'
),
# Raw data file archive views
url(
r'^raw-data-files/$',
views.RawDataFileList.as_view(),
name='rawdatafiles_list'
),
url(
r'^raw-data-files/(?P<file_name>\w+)/$',
views.RawDataFileDetail.as_view(),
name='rawdatafile_detail',
),
]
| from django.conf.urls import url
from calaccess_website import views
urlpatterns = [
# The homepage
url(
r'^$',
views.VersionArchiveIndex.as_view(),
name="version_index",
),
# Version archive views
url(
r'^archive/(?P<year>[0-9]{4})/$',
views.VersionYearArchiveList.as_view(),
name="version_year_archive"
),
url(
r'^versions/(?P<pk>[0-9]{1,})/$',
views.VersionDetail.as_view(),
name="version_detail"
),
url(
r'^versions/latest/$',
views.LatestVersion.as_view(),
name='version_latest_redirect'
),
# Raw data file archive views
url(
r'^raw-data-files/$',
views.RawDataFileList.as_view(),
name='rawdatafiles_list'
),
url(
r'^raw-data-files/(?P<file_name>\w+)/$',
views.RawDataFileDetail.as_view(),
name='rawdatafile_detail',
),
]
| Remove 'versions/' from archive url path | Remove 'versions/' from archive url path
| Python | mit | california-civic-data-coalition/django-calaccess-downloads-website,california-civic-data-coalition/django-calaccess-downloads-website,california-civic-data-coalition/django-calaccess-downloads-website | from django.conf.urls import url
from calaccess_website import views
urlpatterns = [
# The homepage
url(
r'^$',
views.VersionArchiveIndex.as_view(),
name="version_index",
),
# Version archive views
url(
r'^versions/archive/(?P<year>[0-9]{4})/$',
views.VersionYearArchiveList.as_view(),
name="version_year_archive"
),
url(
r'^versions/(?P<pk>[0-9]{1,})/$',
views.VersionDetail.as_view(),
name="version_detail"
),
url(
r'^versions/latest/$',
views.LatestVersion.as_view(),
name='version_latest_redirect'
),
# Raw data file archive views
url(
r'^raw-data-files/$',
views.RawDataFileList.as_view(),
name='rawdatafiles_list'
),
url(
r'^raw-data-files/(?P<file_name>\w+)/$',
views.RawDataFileDetail.as_view(),
name='rawdatafile_detail',
),
]
Remove 'versions/' from archive url path | from django.conf.urls import url
from calaccess_website import views
urlpatterns = [
# The homepage
url(
r'^$',
views.VersionArchiveIndex.as_view(),
name="version_index",
),
# Version archive views
url(
r'^archive/(?P<year>[0-9]{4})/$',
views.VersionYearArchiveList.as_view(),
name="version_year_archive"
),
url(
r'^versions/(?P<pk>[0-9]{1,})/$',
views.VersionDetail.as_view(),
name="version_detail"
),
url(
r'^versions/latest/$',
views.LatestVersion.as_view(),
name='version_latest_redirect'
),
# Raw data file archive views
url(
r'^raw-data-files/$',
views.RawDataFileList.as_view(),
name='rawdatafiles_list'
),
url(
r'^raw-data-files/(?P<file_name>\w+)/$',
views.RawDataFileDetail.as_view(),
name='rawdatafile_detail',
),
]
| <commit_before>from django.conf.urls import url
from calaccess_website import views
urlpatterns = [
# The homepage
url(
r'^$',
views.VersionArchiveIndex.as_view(),
name="version_index",
),
# Version archive views
url(
r'^versions/archive/(?P<year>[0-9]{4})/$',
views.VersionYearArchiveList.as_view(),
name="version_year_archive"
),
url(
r'^versions/(?P<pk>[0-9]{1,})/$',
views.VersionDetail.as_view(),
name="version_detail"
),
url(
r'^versions/latest/$',
views.LatestVersion.as_view(),
name='version_latest_redirect'
),
# Raw data file archive views
url(
r'^raw-data-files/$',
views.RawDataFileList.as_view(),
name='rawdatafiles_list'
),
url(
r'^raw-data-files/(?P<file_name>\w+)/$',
views.RawDataFileDetail.as_view(),
name='rawdatafile_detail',
),
]
<commit_msg>Remove 'versions/' from archive url path<commit_after> | from django.conf.urls import url
from calaccess_website import views
urlpatterns = [
# The homepage
url(
r'^$',
views.VersionArchiveIndex.as_view(),
name="version_index",
),
# Version archive views
url(
r'^archive/(?P<year>[0-9]{4})/$',
views.VersionYearArchiveList.as_view(),
name="version_year_archive"
),
url(
r'^versions/(?P<pk>[0-9]{1,})/$',
views.VersionDetail.as_view(),
name="version_detail"
),
url(
r'^versions/latest/$',
views.LatestVersion.as_view(),
name='version_latest_redirect'
),
# Raw data file archive views
url(
r'^raw-data-files/$',
views.RawDataFileList.as_view(),
name='rawdatafiles_list'
),
url(
r'^raw-data-files/(?P<file_name>\w+)/$',
views.RawDataFileDetail.as_view(),
name='rawdatafile_detail',
),
]
| from django.conf.urls import url
from calaccess_website import views
urlpatterns = [
# The homepage
url(
r'^$',
views.VersionArchiveIndex.as_view(),
name="version_index",
),
# Version archive views
url(
r'^versions/archive/(?P<year>[0-9]{4})/$',
views.VersionYearArchiveList.as_view(),
name="version_year_archive"
),
url(
r'^versions/(?P<pk>[0-9]{1,})/$',
views.VersionDetail.as_view(),
name="version_detail"
),
url(
r'^versions/latest/$',
views.LatestVersion.as_view(),
name='version_latest_redirect'
),
# Raw data file archive views
url(
r'^raw-data-files/$',
views.RawDataFileList.as_view(),
name='rawdatafiles_list'
),
url(
r'^raw-data-files/(?P<file_name>\w+)/$',
views.RawDataFileDetail.as_view(),
name='rawdatafile_detail',
),
]
Remove 'versions/' from archive url pathfrom django.conf.urls import url
from calaccess_website import views
urlpatterns = [
# The homepage
url(
r'^$',
views.VersionArchiveIndex.as_view(),
name="version_index",
),
# Version archive views
url(
r'^archive/(?P<year>[0-9]{4})/$',
views.VersionYearArchiveList.as_view(),
name="version_year_archive"
),
url(
r'^versions/(?P<pk>[0-9]{1,})/$',
views.VersionDetail.as_view(),
name="version_detail"
),
url(
r'^versions/latest/$',
views.LatestVersion.as_view(),
name='version_latest_redirect'
),
# Raw data file archive views
url(
r'^raw-data-files/$',
views.RawDataFileList.as_view(),
name='rawdatafiles_list'
),
url(
r'^raw-data-files/(?P<file_name>\w+)/$',
views.RawDataFileDetail.as_view(),
name='rawdatafile_detail',
),
]
| <commit_before>from django.conf.urls import url
from calaccess_website import views
urlpatterns = [
# The homepage
url(
r'^$',
views.VersionArchiveIndex.as_view(),
name="version_index",
),
# Version archive views
url(
r'^versions/archive/(?P<year>[0-9]{4})/$',
views.VersionYearArchiveList.as_view(),
name="version_year_archive"
),
url(
r'^versions/(?P<pk>[0-9]{1,})/$',
views.VersionDetail.as_view(),
name="version_detail"
),
url(
r'^versions/latest/$',
views.LatestVersion.as_view(),
name='version_latest_redirect'
),
# Raw data file archive views
url(
r'^raw-data-files/$',
views.RawDataFileList.as_view(),
name='rawdatafiles_list'
),
url(
r'^raw-data-files/(?P<file_name>\w+)/$',
views.RawDataFileDetail.as_view(),
name='rawdatafile_detail',
),
]
<commit_msg>Remove 'versions/' from archive url path<commit_after>from django.conf.urls import url
from calaccess_website import views
urlpatterns = [
# The homepage
url(
r'^$',
views.VersionArchiveIndex.as_view(),
name="version_index",
),
# Version archive views
url(
r'^archive/(?P<year>[0-9]{4})/$',
views.VersionYearArchiveList.as_view(),
name="version_year_archive"
),
url(
r'^versions/(?P<pk>[0-9]{1,})/$',
views.VersionDetail.as_view(),
name="version_detail"
),
url(
r'^versions/latest/$',
views.LatestVersion.as_view(),
name='version_latest_redirect'
),
# Raw data file archive views
url(
r'^raw-data-files/$',
views.RawDataFileList.as_view(),
name='rawdatafiles_list'
),
url(
r'^raw-data-files/(?P<file_name>\w+)/$',
views.RawDataFileDetail.as_view(),
name='rawdatafile_detail',
),
]
|
4b488c8d0842bb25c719fcd93ee0ae46978b5680 | meta/util.py | meta/util.py | import os
import sys
import time
import math
from contextlib import contextmanager
@contextmanager
def timeit_msg(msg):
print '{}...'.format(msg),
sys.stdout.flush()
start = time.time()
yield
stop = time.time()
print ' ({:1.3f} s)'.format((stop-start))
def get_file_size(fname):
fstats = os.stat(fname)
size = fstats.st_size
exponent = int(math.log(size, 10))
divisor = {'B': 0, 'kB': 3, 'MB': 6, 'GB': 9}
if exponent < 3:
res = 'B'
if exponent >= 3 and exponent < 6:
res = 'kB'
elif exponent >= 6 and exponent < 9:
res = 'MB'
else:
res = 'GB'
return '{:1.2f} {}'.format(float(size) / 10**divisor[res], res)
| import os
import sys
import time
import math
import sqlite3
from contextlib import contextmanager
import meta
@contextmanager
def timeit_msg(msg):
print '{}...'.format(msg),
sys.stdout.flush()
start = time.time()
yield
stop = time.time()
print ' ({:1.3f} s)'.format((stop-start))
def get_file_size(fname):
fstats = os.stat(fname)
size = fstats.st_size
exponent = int(math.log(size, 10))
divisor = {'B': 0, 'kB': 3, 'MB': 6, 'GB': 9}
if exponent < 3:
res = 'B'
if exponent >= 3 and exponent < 6:
res = 'kB'
elif exponent >= 6 and exponent < 9:
res = 'MB'
else:
res = 'GB'
return '{:1.2f} {}'.format(float(size) / 10**divisor[res], res)
def get_ncbi_db_conn():
dbfile = os.path.join(meta.__path__[0], 'data', 'NCBI.db')
conn = sqlite3.connect(dbfile)
conn.text_factory = str
return conn
| Add support for connecting to NCBI database | Add support for connecting to NCBI database
| Python | mit | abulovic/pgnd-meta | import os
import sys
import time
import math
from contextlib import contextmanager
@contextmanager
def timeit_msg(msg):
print '{}...'.format(msg),
sys.stdout.flush()
start = time.time()
yield
stop = time.time()
print ' ({:1.3f} s)'.format((stop-start))
def get_file_size(fname):
fstats = os.stat(fname)
size = fstats.st_size
exponent = int(math.log(size, 10))
divisor = {'B': 0, 'kB': 3, 'MB': 6, 'GB': 9}
if exponent < 3:
res = 'B'
if exponent >= 3 and exponent < 6:
res = 'kB'
elif exponent >= 6 and exponent < 9:
res = 'MB'
else:
res = 'GB'
return '{:1.2f} {}'.format(float(size) / 10**divisor[res], res)
Add support for connecting to NCBI database | import os
import sys
import time
import math
import sqlite3
from contextlib import contextmanager
import meta
@contextmanager
def timeit_msg(msg):
print '{}...'.format(msg),
sys.stdout.flush()
start = time.time()
yield
stop = time.time()
print ' ({:1.3f} s)'.format((stop-start))
def get_file_size(fname):
fstats = os.stat(fname)
size = fstats.st_size
exponent = int(math.log(size, 10))
divisor = {'B': 0, 'kB': 3, 'MB': 6, 'GB': 9}
if exponent < 3:
res = 'B'
if exponent >= 3 and exponent < 6:
res = 'kB'
elif exponent >= 6 and exponent < 9:
res = 'MB'
else:
res = 'GB'
return '{:1.2f} {}'.format(float(size) / 10**divisor[res], res)
def get_ncbi_db_conn():
dbfile = os.path.join(meta.__path__[0], 'data', 'NCBI.db')
conn = sqlite3.connect(dbfile)
conn.text_factory = str
return conn
| <commit_before>import os
import sys
import time
import math
from contextlib import contextmanager
@contextmanager
def timeit_msg(msg):
print '{}...'.format(msg),
sys.stdout.flush()
start = time.time()
yield
stop = time.time()
print ' ({:1.3f} s)'.format((stop-start))
def get_file_size(fname):
fstats = os.stat(fname)
size = fstats.st_size
exponent = int(math.log(size, 10))
divisor = {'B': 0, 'kB': 3, 'MB': 6, 'GB': 9}
if exponent < 3:
res = 'B'
if exponent >= 3 and exponent < 6:
res = 'kB'
elif exponent >= 6 and exponent < 9:
res = 'MB'
else:
res = 'GB'
return '{:1.2f} {}'.format(float(size) / 10**divisor[res], res)
<commit_msg>Add support for connecting to NCBI database<commit_after> | import os
import sys
import time
import math
import sqlite3
from contextlib import contextmanager
import meta
@contextmanager
def timeit_msg(msg):
print '{}...'.format(msg),
sys.stdout.flush()
start = time.time()
yield
stop = time.time()
print ' ({:1.3f} s)'.format((stop-start))
def get_file_size(fname):
fstats = os.stat(fname)
size = fstats.st_size
exponent = int(math.log(size, 10))
divisor = {'B': 0, 'kB': 3, 'MB': 6, 'GB': 9}
if exponent < 3:
res = 'B'
if exponent >= 3 and exponent < 6:
res = 'kB'
elif exponent >= 6 and exponent < 9:
res = 'MB'
else:
res = 'GB'
return '{:1.2f} {}'.format(float(size) / 10**divisor[res], res)
def get_ncbi_db_conn():
dbfile = os.path.join(meta.__path__[0], 'data', 'NCBI.db')
conn = sqlite3.connect(dbfile)
conn.text_factory = str
return conn
| import os
import sys
import time
import math
from contextlib import contextmanager
@contextmanager
def timeit_msg(msg):
print '{}...'.format(msg),
sys.stdout.flush()
start = time.time()
yield
stop = time.time()
print ' ({:1.3f} s)'.format((stop-start))
def get_file_size(fname):
fstats = os.stat(fname)
size = fstats.st_size
exponent = int(math.log(size, 10))
divisor = {'B': 0, 'kB': 3, 'MB': 6, 'GB': 9}
if exponent < 3:
res = 'B'
if exponent >= 3 and exponent < 6:
res = 'kB'
elif exponent >= 6 and exponent < 9:
res = 'MB'
else:
res = 'GB'
return '{:1.2f} {}'.format(float(size) / 10**divisor[res], res)
Add support for connecting to NCBI databaseimport os
import sys
import time
import math
import sqlite3
from contextlib import contextmanager
import meta
@contextmanager
def timeit_msg(msg):
print '{}...'.format(msg),
sys.stdout.flush()
start = time.time()
yield
stop = time.time()
print ' ({:1.3f} s)'.format((stop-start))
def get_file_size(fname):
fstats = os.stat(fname)
size = fstats.st_size
exponent = int(math.log(size, 10))
divisor = {'B': 0, 'kB': 3, 'MB': 6, 'GB': 9}
if exponent < 3:
res = 'B'
if exponent >= 3 and exponent < 6:
res = 'kB'
elif exponent >= 6 and exponent < 9:
res = 'MB'
else:
res = 'GB'
return '{:1.2f} {}'.format(float(size) / 10**divisor[res], res)
def get_ncbi_db_conn():
dbfile = os.path.join(meta.__path__[0], 'data', 'NCBI.db')
conn = sqlite3.connect(dbfile)
conn.text_factory = str
return conn
| <commit_before>import os
import sys
import time
import math
from contextlib import contextmanager
@contextmanager
def timeit_msg(msg):
print '{}...'.format(msg),
sys.stdout.flush()
start = time.time()
yield
stop = time.time()
print ' ({:1.3f} s)'.format((stop-start))
def get_file_size(fname):
fstats = os.stat(fname)
size = fstats.st_size
exponent = int(math.log(size, 10))
divisor = {'B': 0, 'kB': 3, 'MB': 6, 'GB': 9}
if exponent < 3:
res = 'B'
if exponent >= 3 and exponent < 6:
res = 'kB'
elif exponent >= 6 and exponent < 9:
res = 'MB'
else:
res = 'GB'
return '{:1.2f} {}'.format(float(size) / 10**divisor[res], res)
<commit_msg>Add support for connecting to NCBI database<commit_after>import os
import sys
import time
import math
import sqlite3
from contextlib import contextmanager
import meta
@contextmanager
def timeit_msg(msg):
print '{}...'.format(msg),
sys.stdout.flush()
start = time.time()
yield
stop = time.time()
print ' ({:1.3f} s)'.format((stop-start))
def get_file_size(fname):
fstats = os.stat(fname)
size = fstats.st_size
exponent = int(math.log(size, 10))
divisor = {'B': 0, 'kB': 3, 'MB': 6, 'GB': 9}
if exponent < 3:
res = 'B'
if exponent >= 3 and exponent < 6:
res = 'kB'
elif exponent >= 6 and exponent < 9:
res = 'MB'
else:
res = 'GB'
return '{:1.2f} {}'.format(float(size) / 10**divisor[res], res)
def get_ncbi_db_conn():
dbfile = os.path.join(meta.__path__[0], 'data', 'NCBI.db')
conn = sqlite3.connect(dbfile)
conn.text_factory = str
return conn
|
dac3cedaee583db4cc3c05a9cb2c4f15a707123e | pylib/mapit/middleware.py | pylib/mapit/middleware.py | import re
class JSONPMiddleware(object):
def process_response(self, request, response):
if request.GET.get('callback') and re.match('[a-zA-Z0-9_]+$', request.GET.get('callback')):
response.content = request.GET.get('callback') + '(' + response.content + ')'
return response
| import re
class JSONPMiddleware(object):
def process_response(self, request, response):
if request.GET.get('callback') and re.match('[a-zA-Z0-9_]+$', request.GET.get('callback')):
response.content = request.GET.get('callback') + '(' + response.content + ')'
response.status_code = 200 # Must return OK for JSONP to be processed
return response
| Set up JSONP requests to always return 200. | Set up JSONP requests to always return 200.
| Python | agpl-3.0 | Sinar/mapit,Code4SA/mapit,New-Bamboo/mapit,opencorato/mapit,Sinar/mapit,opencorato/mapit,opencorato/mapit,chris48s/mapit,chris48s/mapit,Code4SA/mapit,chris48s/mapit,New-Bamboo/mapit,Code4SA/mapit | import re
class JSONPMiddleware(object):
def process_response(self, request, response):
if request.GET.get('callback') and re.match('[a-zA-Z0-9_]+$', request.GET.get('callback')):
response.content = request.GET.get('callback') + '(' + response.content + ')'
return response
Set up JSONP requests to always return 200. | import re
class JSONPMiddleware(object):
def process_response(self, request, response):
if request.GET.get('callback') and re.match('[a-zA-Z0-9_]+$', request.GET.get('callback')):
response.content = request.GET.get('callback') + '(' + response.content + ')'
response.status_code = 200 # Must return OK for JSONP to be processed
return response
| <commit_before>import re
class JSONPMiddleware(object):
def process_response(self, request, response):
if request.GET.get('callback') and re.match('[a-zA-Z0-9_]+$', request.GET.get('callback')):
response.content = request.GET.get('callback') + '(' + response.content + ')'
return response
<commit_msg>Set up JSONP requests to always return 200.<commit_after> | import re
class JSONPMiddleware(object):
def process_response(self, request, response):
if request.GET.get('callback') and re.match('[a-zA-Z0-9_]+$', request.GET.get('callback')):
response.content = request.GET.get('callback') + '(' + response.content + ')'
response.status_code = 200 # Must return OK for JSONP to be processed
return response
| import re
class JSONPMiddleware(object):
def process_response(self, request, response):
if request.GET.get('callback') and re.match('[a-zA-Z0-9_]+$', request.GET.get('callback')):
response.content = request.GET.get('callback') + '(' + response.content + ')'
return response
Set up JSONP requests to always return 200.import re
class JSONPMiddleware(object):
def process_response(self, request, response):
if request.GET.get('callback') and re.match('[a-zA-Z0-9_]+$', request.GET.get('callback')):
response.content = request.GET.get('callback') + '(' + response.content + ')'
response.status_code = 200 # Must return OK for JSONP to be processed
return response
| <commit_before>import re
class JSONPMiddleware(object):
def process_response(self, request, response):
if request.GET.get('callback') and re.match('[a-zA-Z0-9_]+$', request.GET.get('callback')):
response.content = request.GET.get('callback') + '(' + response.content + ')'
return response
<commit_msg>Set up JSONP requests to always return 200.<commit_after>import re
class JSONPMiddleware(object):
def process_response(self, request, response):
if request.GET.get('callback') and re.match('[a-zA-Z0-9_]+$', request.GET.get('callback')):
response.content = request.GET.get('callback') + '(' + response.content + ')'
response.status_code = 200 # Must return OK for JSONP to be processed
return response
|
c80893c5789998b6d6068703bb2434919abc65a3 | sqjobs/tests/django_test.py | sqjobs/tests/django_test.py | #!/usr/bin/env python
import os
import sys
from django.conf import settings
import django
DEFAULT_SETTINGS = {
'INSTALLED_APPS': (
'sqjobs',
'sqjobs.tests',
'sqjobs.contrib.django.djsqjobs'
),
'DATABASES': {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
}
},
'USE_TZ': 'true',
'SILENCED_SYSTEM_CHECKS': ["1_7.W001"],
}
if not settings.configured:
settings.configure(**DEFAULT_SETTINGS)
if hasattr(django, 'setup'):
django.setup()
parent = os.path.dirname(os.path.dirname(os.path.dirname(
os.path.abspath(__file__)
)))
sys.path.insert(0, parent)
from django.test.runner import DiscoverRunner
DiscoverRunner(failfast=False).run_tests([
'sqjobs.tests'
], verbosity=1)
| #!/usr/bin/env python
import os
import sys
from django.conf import settings
import django
DEFAULT_SETTINGS = {
'INSTALLED_APPS': (
'sqjobs',
'sqjobs.tests',
'sqjobs.contrib.django.djsqjobs'
),
'DATABASES': {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
}
},
'USE_TZ': 'true',
'SILENCED_SYSTEM_CHECKS': ["1_7.W001"],
}
if not settings.configured:
settings.configure(**DEFAULT_SETTINGS)
if hasattr(django, 'setup'):
django.setup()
parent = os.path.dirname(os.path.dirname(os.path.dirname(
os.path.abspath(__file__)
)))
sys.path.insert(0, parent)
from django.test.runner import DiscoverRunner
res = DiscoverRunner(failfast=False).run_tests([
'sqjobs.tests'
], verbosity=1)
os._exit(res)
| Return status when executing tests with DiscoverRunner | Return status when executing tests with DiscoverRunner
| Python | bsd-3-clause | gnufede/sqjobs,gnufede/sqjobs | #!/usr/bin/env python
import os
import sys
from django.conf import settings
import django
DEFAULT_SETTINGS = {
'INSTALLED_APPS': (
'sqjobs',
'sqjobs.tests',
'sqjobs.contrib.django.djsqjobs'
),
'DATABASES': {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
}
},
'USE_TZ': 'true',
'SILENCED_SYSTEM_CHECKS': ["1_7.W001"],
}
if not settings.configured:
settings.configure(**DEFAULT_SETTINGS)
if hasattr(django, 'setup'):
django.setup()
parent = os.path.dirname(os.path.dirname(os.path.dirname(
os.path.abspath(__file__)
)))
sys.path.insert(0, parent)
from django.test.runner import DiscoverRunner
DiscoverRunner(failfast=False).run_tests([
'sqjobs.tests'
], verbosity=1)
Return status when executing tests with DiscoverRunner | #!/usr/bin/env python
import os
import sys
from django.conf import settings
import django
DEFAULT_SETTINGS = {
'INSTALLED_APPS': (
'sqjobs',
'sqjobs.tests',
'sqjobs.contrib.django.djsqjobs'
),
'DATABASES': {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
}
},
'USE_TZ': 'true',
'SILENCED_SYSTEM_CHECKS': ["1_7.W001"],
}
if not settings.configured:
settings.configure(**DEFAULT_SETTINGS)
if hasattr(django, 'setup'):
django.setup()
parent = os.path.dirname(os.path.dirname(os.path.dirname(
os.path.abspath(__file__)
)))
sys.path.insert(0, parent)
from django.test.runner import DiscoverRunner
res = DiscoverRunner(failfast=False).run_tests([
'sqjobs.tests'
], verbosity=1)
os._exit(res)
| <commit_before>#!/usr/bin/env python
import os
import sys
from django.conf import settings
import django
DEFAULT_SETTINGS = {
'INSTALLED_APPS': (
'sqjobs',
'sqjobs.tests',
'sqjobs.contrib.django.djsqjobs'
),
'DATABASES': {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
}
},
'USE_TZ': 'true',
'SILENCED_SYSTEM_CHECKS': ["1_7.W001"],
}
if not settings.configured:
settings.configure(**DEFAULT_SETTINGS)
if hasattr(django, 'setup'):
django.setup()
parent = os.path.dirname(os.path.dirname(os.path.dirname(
os.path.abspath(__file__)
)))
sys.path.insert(0, parent)
from django.test.runner import DiscoverRunner
DiscoverRunner(failfast=False).run_tests([
'sqjobs.tests'
], verbosity=1)
<commit_msg>Return status when executing tests with DiscoverRunner<commit_after> | #!/usr/bin/env python
import os
import sys
from django.conf import settings
import django
DEFAULT_SETTINGS = {
'INSTALLED_APPS': (
'sqjobs',
'sqjobs.tests',
'sqjobs.contrib.django.djsqjobs'
),
'DATABASES': {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
}
},
'USE_TZ': 'true',
'SILENCED_SYSTEM_CHECKS': ["1_7.W001"],
}
if not settings.configured:
settings.configure(**DEFAULT_SETTINGS)
if hasattr(django, 'setup'):
django.setup()
parent = os.path.dirname(os.path.dirname(os.path.dirname(
os.path.abspath(__file__)
)))
sys.path.insert(0, parent)
from django.test.runner import DiscoverRunner
res = DiscoverRunner(failfast=False).run_tests([
'sqjobs.tests'
], verbosity=1)
os._exit(res)
| #!/usr/bin/env python
import os
import sys
from django.conf import settings
import django
DEFAULT_SETTINGS = {
'INSTALLED_APPS': (
'sqjobs',
'sqjobs.tests',
'sqjobs.contrib.django.djsqjobs'
),
'DATABASES': {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
}
},
'USE_TZ': 'true',
'SILENCED_SYSTEM_CHECKS': ["1_7.W001"],
}
if not settings.configured:
settings.configure(**DEFAULT_SETTINGS)
if hasattr(django, 'setup'):
django.setup()
parent = os.path.dirname(os.path.dirname(os.path.dirname(
os.path.abspath(__file__)
)))
sys.path.insert(0, parent)
from django.test.runner import DiscoverRunner
DiscoverRunner(failfast=False).run_tests([
'sqjobs.tests'
], verbosity=1)
Return status when executing tests with DiscoverRunner#!/usr/bin/env python
import os
import sys
from django.conf import settings
import django
DEFAULT_SETTINGS = {
'INSTALLED_APPS': (
'sqjobs',
'sqjobs.tests',
'sqjobs.contrib.django.djsqjobs'
),
'DATABASES': {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
}
},
'USE_TZ': 'true',
'SILENCED_SYSTEM_CHECKS': ["1_7.W001"],
}
if not settings.configured:
settings.configure(**DEFAULT_SETTINGS)
if hasattr(django, 'setup'):
django.setup()
parent = os.path.dirname(os.path.dirname(os.path.dirname(
os.path.abspath(__file__)
)))
sys.path.insert(0, parent)
from django.test.runner import DiscoverRunner
res = DiscoverRunner(failfast=False).run_tests([
'sqjobs.tests'
], verbosity=1)
os._exit(res)
| <commit_before>#!/usr/bin/env python
import os
import sys
from django.conf import settings
import django
DEFAULT_SETTINGS = {
'INSTALLED_APPS': (
'sqjobs',
'sqjobs.tests',
'sqjobs.contrib.django.djsqjobs'
),
'DATABASES': {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
}
},
'USE_TZ': 'true',
'SILENCED_SYSTEM_CHECKS': ["1_7.W001"],
}
if not settings.configured:
settings.configure(**DEFAULT_SETTINGS)
if hasattr(django, 'setup'):
django.setup()
parent = os.path.dirname(os.path.dirname(os.path.dirname(
os.path.abspath(__file__)
)))
sys.path.insert(0, parent)
from django.test.runner import DiscoverRunner
DiscoverRunner(failfast=False).run_tests([
'sqjobs.tests'
], verbosity=1)
<commit_msg>Return status when executing tests with DiscoverRunner<commit_after>#!/usr/bin/env python
import os
import sys
from django.conf import settings
import django
DEFAULT_SETTINGS = {
'INSTALLED_APPS': (
'sqjobs',
'sqjobs.tests',
'sqjobs.contrib.django.djsqjobs'
),
'DATABASES': {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
}
},
'USE_TZ': 'true',
'SILENCED_SYSTEM_CHECKS': ["1_7.W001"],
}
if not settings.configured:
settings.configure(**DEFAULT_SETTINGS)
if hasattr(django, 'setup'):
django.setup()
parent = os.path.dirname(os.path.dirname(os.path.dirname(
os.path.abspath(__file__)
)))
sys.path.insert(0, parent)
from django.test.runner import DiscoverRunner
res = DiscoverRunner(failfast=False).run_tests([
'sqjobs.tests'
], verbosity=1)
os._exit(res)
|
2548a0c5e108fa22867c6a0e4f5b06ceba52dac0 | bottery/message.py | bottery/message.py | import os
from datetime import datetime
import attr
from jinja2 import Environment, FileSystemLoader, select_autoescape
@attr.s
class Message:
id = attr.ib()
platform = attr.ib()
user = attr.ib()
chat = attr.ib()
text = attr.ib()
timestamp = attr.ib()
raw = attr.ib()
@property
def datetime(self):
return datetime.utcfromtimestamp(self.timestamp)
def render(message, template_name, context={}):
base_dir = os.path.join(os.getcwd(), 'templates')
paths = [base_dir]
# Include paths on settings
# paths.extend(settings.TEMPLATES)
env = Environment(
loader=FileSystemLoader(paths),
autoescape=select_autoescape(['html']))
template = env.get_template(template_name)
default_context = {
'user': message.user,
'platform': message.platform,
}
default_context.update(context)
return template.render(**default_context)
| import os
from datetime import datetime
import attr
from jinja2 import Environment, FileSystemLoader, select_autoescape
from bottery.conf import settings
@attr.s
class Message:
id = attr.ib()
platform = attr.ib()
user = attr.ib()
chat = attr.ib()
text = attr.ib()
timestamp = attr.ib()
raw = attr.ib()
@property
def datetime(self):
return datetime.utcfromtimestamp(self.timestamp)
def render(message, template_name, context=None):
context = context or {}
base_dir = os.path.join(os.getcwd(), 'templates')
paths = [base_dir]
paths.extend(settings.TEMPLATES)
env = Environment(
loader=FileSystemLoader(paths),
autoescape=select_autoescape(['html']))
template = env.get_template(template_name)
default_context = {
'user': message.user,
'platform': message.platform,
}
default_context.update(context)
return template.render(**default_context)
| Extend templates dirs on settings.py | Extend templates dirs on settings.py
| Python | mit | rougeth/bottery | import os
from datetime import datetime
import attr
from jinja2 import Environment, FileSystemLoader, select_autoescape
@attr.s
class Message:
id = attr.ib()
platform = attr.ib()
user = attr.ib()
chat = attr.ib()
text = attr.ib()
timestamp = attr.ib()
raw = attr.ib()
@property
def datetime(self):
return datetime.utcfromtimestamp(self.timestamp)
def render(message, template_name, context={}):
base_dir = os.path.join(os.getcwd(), 'templates')
paths = [base_dir]
# Include paths on settings
# paths.extend(settings.TEMPLATES)
env = Environment(
loader=FileSystemLoader(paths),
autoescape=select_autoescape(['html']))
template = env.get_template(template_name)
default_context = {
'user': message.user,
'platform': message.platform,
}
default_context.update(context)
return template.render(**default_context)
Extend templates dirs on settings.py | import os
from datetime import datetime
import attr
from jinja2 import Environment, FileSystemLoader, select_autoescape
from bottery.conf import settings
@attr.s
class Message:
id = attr.ib()
platform = attr.ib()
user = attr.ib()
chat = attr.ib()
text = attr.ib()
timestamp = attr.ib()
raw = attr.ib()
@property
def datetime(self):
return datetime.utcfromtimestamp(self.timestamp)
def render(message, template_name, context=None):
context = context or {}
base_dir = os.path.join(os.getcwd(), 'templates')
paths = [base_dir]
paths.extend(settings.TEMPLATES)
env = Environment(
loader=FileSystemLoader(paths),
autoescape=select_autoescape(['html']))
template = env.get_template(template_name)
default_context = {
'user': message.user,
'platform': message.platform,
}
default_context.update(context)
return template.render(**default_context)
| <commit_before>import os
from datetime import datetime
import attr
from jinja2 import Environment, FileSystemLoader, select_autoescape
@attr.s
class Message:
id = attr.ib()
platform = attr.ib()
user = attr.ib()
chat = attr.ib()
text = attr.ib()
timestamp = attr.ib()
raw = attr.ib()
@property
def datetime(self):
return datetime.utcfromtimestamp(self.timestamp)
def render(message, template_name, context={}):
base_dir = os.path.join(os.getcwd(), 'templates')
paths = [base_dir]
# Include paths on settings
# paths.extend(settings.TEMPLATES)
env = Environment(
loader=FileSystemLoader(paths),
autoescape=select_autoescape(['html']))
template = env.get_template(template_name)
default_context = {
'user': message.user,
'platform': message.platform,
}
default_context.update(context)
return template.render(**default_context)
<commit_msg>Extend templates dirs on settings.py<commit_after> | import os
from datetime import datetime
import attr
from jinja2 import Environment, FileSystemLoader, select_autoescape
from bottery.conf import settings
@attr.s
class Message:
id = attr.ib()
platform = attr.ib()
user = attr.ib()
chat = attr.ib()
text = attr.ib()
timestamp = attr.ib()
raw = attr.ib()
@property
def datetime(self):
return datetime.utcfromtimestamp(self.timestamp)
def render(message, template_name, context=None):
context = context or {}
base_dir = os.path.join(os.getcwd(), 'templates')
paths = [base_dir]
paths.extend(settings.TEMPLATES)
env = Environment(
loader=FileSystemLoader(paths),
autoescape=select_autoescape(['html']))
template = env.get_template(template_name)
default_context = {
'user': message.user,
'platform': message.platform,
}
default_context.update(context)
return template.render(**default_context)
| import os
from datetime import datetime
import attr
from jinja2 import Environment, FileSystemLoader, select_autoescape
@attr.s
class Message:
id = attr.ib()
platform = attr.ib()
user = attr.ib()
chat = attr.ib()
text = attr.ib()
timestamp = attr.ib()
raw = attr.ib()
@property
def datetime(self):
return datetime.utcfromtimestamp(self.timestamp)
def render(message, template_name, context={}):
base_dir = os.path.join(os.getcwd(), 'templates')
paths = [base_dir]
# Include paths on settings
# paths.extend(settings.TEMPLATES)
env = Environment(
loader=FileSystemLoader(paths),
autoescape=select_autoescape(['html']))
template = env.get_template(template_name)
default_context = {
'user': message.user,
'platform': message.platform,
}
default_context.update(context)
return template.render(**default_context)
Extend templates dirs on settings.pyimport os
from datetime import datetime
import attr
from jinja2 import Environment, FileSystemLoader, select_autoescape
from bottery.conf import settings
@attr.s
class Message:
id = attr.ib()
platform = attr.ib()
user = attr.ib()
chat = attr.ib()
text = attr.ib()
timestamp = attr.ib()
raw = attr.ib()
@property
def datetime(self):
return datetime.utcfromtimestamp(self.timestamp)
def render(message, template_name, context=None):
context = context or {}
base_dir = os.path.join(os.getcwd(), 'templates')
paths = [base_dir]
paths.extend(settings.TEMPLATES)
env = Environment(
loader=FileSystemLoader(paths),
autoescape=select_autoescape(['html']))
template = env.get_template(template_name)
default_context = {
'user': message.user,
'platform': message.platform,
}
default_context.update(context)
return template.render(**default_context)
| <commit_before>import os
from datetime import datetime
import attr
from jinja2 import Environment, FileSystemLoader, select_autoescape
@attr.s
class Message:
id = attr.ib()
platform = attr.ib()
user = attr.ib()
chat = attr.ib()
text = attr.ib()
timestamp = attr.ib()
raw = attr.ib()
@property
def datetime(self):
return datetime.utcfromtimestamp(self.timestamp)
def render(message, template_name, context={}):
base_dir = os.path.join(os.getcwd(), 'templates')
paths = [base_dir]
# Include paths on settings
# paths.extend(settings.TEMPLATES)
env = Environment(
loader=FileSystemLoader(paths),
autoescape=select_autoescape(['html']))
template = env.get_template(template_name)
default_context = {
'user': message.user,
'platform': message.platform,
}
default_context.update(context)
return template.render(**default_context)
<commit_msg>Extend templates dirs on settings.py<commit_after>import os
from datetime import datetime
import attr
from jinja2 import Environment, FileSystemLoader, select_autoescape
from bottery.conf import settings
@attr.s
class Message:
id = attr.ib()
platform = attr.ib()
user = attr.ib()
chat = attr.ib()
text = attr.ib()
timestamp = attr.ib()
raw = attr.ib()
@property
def datetime(self):
return datetime.utcfromtimestamp(self.timestamp)
def render(message, template_name, context=None):
context = context or {}
base_dir = os.path.join(os.getcwd(), 'templates')
paths = [base_dir]
paths.extend(settings.TEMPLATES)
env = Environment(
loader=FileSystemLoader(paths),
autoescape=select_autoescape(['html']))
template = env.get_template(template_name)
default_context = {
'user': message.user,
'platform': message.platform,
}
default_context.update(context)
return template.render(**default_context)
|
0d73f5d18a927ffebc3fa32180b608f0c96dcdf1 | Exe_04.py | Exe_04.py | cars = 100
space_in_a_car = 4.0
drivers = 30
passengers = 90
cars_not_driven = cars - drivers
cars_driven = drivers
carpool_capacity = cars_driven * space_in_a_car
average_passengers_per_car = passengers / cars_driven
print("There are", cars, "cars available.")
print("There are only", drivers, "drivers available.")
print("There will be", cars_not_driven, "empty cars today.")
print("We can transport", carpool_capacity, "people today.")
print("We have", passengers, "to carpool today.")
print("We need to put about", average_passengers_per_car, "in each car.") | cars = 100
space_in_a_car = 4.0
drivers = 30
passengers = 90
cars_not_driven = cars - drivers
cars_driven = drivers
carpool_capacity = cars_driven * space_in_a_car
average_passengers_per_car = passengers / cars_driven
print("There are", cars, "cars available.")
print("There are only", drivers, "drivers available.")
print("There will be", cars_not_driven, "empty cars today.")
print("We can transport", carpool_capacity, "people today.")
print("We have", passengers, "to carpool today.")
print("We need to put about", average_passengers_per_car, "in each car.")
# Upgrade part
my_name = 'Binh D. Nguyen'
my_age = 22
my_height = 74 # inches
my_weight = 180 # lbs
my_eyes = 'Black'
my_teeth = 'White'
my_hair = 'Black'
#print "Let's talk about %s." % my_name
print(f"Let's talk about {my_name:}.")
#print "He's %d inches tall." % my_height
print("He's {:d} inches tall.".format(my_height))
#print "He's %d pounds heavy." % my_weight
#print "Actually that's not too heavy."
#print "He's got %s eyes and %s hair." % (my_eyes, my_hair)
print(f"He's got { my_eyes:} eyes and {my_hair:} hair.")
#print "His teeth are usually %s depending on the coffee." % my_teeth
# this line is tricky, try to get it exactly right
#print "If I add %d, %d, and %d I get %d." % (
# my_age, my_height, my_weight, my_age + my_height + my_weight)
print(f"If I add {my_age}, {my_height}, and {my_weight} I get {my_age + my_height + my_weight}.")
| Update exercise of variables and names | Update exercise of variables and names
| Python | mit | Oreder/PythonSelfStudy | cars = 100
space_in_a_car = 4.0
drivers = 30
passengers = 90
cars_not_driven = cars - drivers
cars_driven = drivers
carpool_capacity = cars_driven * space_in_a_car
average_passengers_per_car = passengers / cars_driven
print("There are", cars, "cars available.")
print("There are only", drivers, "drivers available.")
print("There will be", cars_not_driven, "empty cars today.")
print("We can transport", carpool_capacity, "people today.")
print("We have", passengers, "to carpool today.")
print("We need to put about", average_passengers_per_car, "in each car.")Update exercise of variables and names | cars = 100
space_in_a_car = 4.0
drivers = 30
passengers = 90
cars_not_driven = cars - drivers
cars_driven = drivers
carpool_capacity = cars_driven * space_in_a_car
average_passengers_per_car = passengers / cars_driven
print("There are", cars, "cars available.")
print("There are only", drivers, "drivers available.")
print("There will be", cars_not_driven, "empty cars today.")
print("We can transport", carpool_capacity, "people today.")
print("We have", passengers, "to carpool today.")
print("We need to put about", average_passengers_per_car, "in each car.")
# Upgrade part
my_name = 'Binh D. Nguyen'
my_age = 22
my_height = 74 # inches
my_weight = 180 # lbs
my_eyes = 'Black'
my_teeth = 'White'
my_hair = 'Black'
#print "Let's talk about %s." % my_name
print(f"Let's talk about {my_name:}.")
#print "He's %d inches tall." % my_height
print("He's {:d} inches tall.".format(my_height))
#print "He's %d pounds heavy." % my_weight
#print "Actually that's not too heavy."
#print "He's got %s eyes and %s hair." % (my_eyes, my_hair)
print(f"He's got { my_eyes:} eyes and {my_hair:} hair.")
#print "His teeth are usually %s depending on the coffee." % my_teeth
# this line is tricky, try to get it exactly right
#print "If I add %d, %d, and %d I get %d." % (
# my_age, my_height, my_weight, my_age + my_height + my_weight)
print(f"If I add {my_age}, {my_height}, and {my_weight} I get {my_age + my_height + my_weight}.")
| <commit_before>cars = 100
space_in_a_car = 4.0
drivers = 30
passengers = 90
cars_not_driven = cars - drivers
cars_driven = drivers
carpool_capacity = cars_driven * space_in_a_car
average_passengers_per_car = passengers / cars_driven
print("There are", cars, "cars available.")
print("There are only", drivers, "drivers available.")
print("There will be", cars_not_driven, "empty cars today.")
print("We can transport", carpool_capacity, "people today.")
print("We have", passengers, "to carpool today.")
print("We need to put about", average_passengers_per_car, "in each car.")<commit_msg>Update exercise of variables and names<commit_after> | cars = 100
space_in_a_car = 4.0
drivers = 30
passengers = 90
cars_not_driven = cars - drivers
cars_driven = drivers
carpool_capacity = cars_driven * space_in_a_car
average_passengers_per_car = passengers / cars_driven
print("There are", cars, "cars available.")
print("There are only", drivers, "drivers available.")
print("There will be", cars_not_driven, "empty cars today.")
print("We can transport", carpool_capacity, "people today.")
print("We have", passengers, "to carpool today.")
print("We need to put about", average_passengers_per_car, "in each car.")
# Upgrade part
my_name = 'Binh D. Nguyen'
my_age = 22
my_height = 74 # inches
my_weight = 180 # lbs
my_eyes = 'Black'
my_teeth = 'White'
my_hair = 'Black'
#print "Let's talk about %s." % my_name
print(f"Let's talk about {my_name:}.")
#print "He's %d inches tall." % my_height
print("He's {:d} inches tall.".format(my_height))
#print "He's %d pounds heavy." % my_weight
#print "Actually that's not too heavy."
#print "He's got %s eyes and %s hair." % (my_eyes, my_hair)
print(f"He's got { my_eyes:} eyes and {my_hair:} hair.")
#print "His teeth are usually %s depending on the coffee." % my_teeth
# this line is tricky, try to get it exactly right
#print "If I add %d, %d, and %d I get %d." % (
# my_age, my_height, my_weight, my_age + my_height + my_weight)
print(f"If I add {my_age}, {my_height}, and {my_weight} I get {my_age + my_height + my_weight}.")
| cars = 100
space_in_a_car = 4.0
drivers = 30
passengers = 90
cars_not_driven = cars - drivers
cars_driven = drivers
carpool_capacity = cars_driven * space_in_a_car
average_passengers_per_car = passengers / cars_driven
print("There are", cars, "cars available.")
print("There are only", drivers, "drivers available.")
print("There will be", cars_not_driven, "empty cars today.")
print("We can transport", carpool_capacity, "people today.")
print("We have", passengers, "to carpool today.")
print("We need to put about", average_passengers_per_car, "in each car.")Update exercise of variables and namescars = 100
space_in_a_car = 4.0
drivers = 30
passengers = 90
cars_not_driven = cars - drivers
cars_driven = drivers
carpool_capacity = cars_driven * space_in_a_car
average_passengers_per_car = passengers / cars_driven
print("There are", cars, "cars available.")
print("There are only", drivers, "drivers available.")
print("There will be", cars_not_driven, "empty cars today.")
print("We can transport", carpool_capacity, "people today.")
print("We have", passengers, "to carpool today.")
print("We need to put about", average_passengers_per_car, "in each car.")
# Upgrade part
my_name = 'Binh D. Nguyen'
my_age = 22
my_height = 74 # inches
my_weight = 180 # lbs
my_eyes = 'Black'
my_teeth = 'White'
my_hair = 'Black'
#print "Let's talk about %s." % my_name
print(f"Let's talk about {my_name:}.")
#print "He's %d inches tall." % my_height
print("He's {:d} inches tall.".format(my_height))
#print "He's %d pounds heavy." % my_weight
#print "Actually that's not too heavy."
#print "He's got %s eyes and %s hair." % (my_eyes, my_hair)
print(f"He's got { my_eyes:} eyes and {my_hair:} hair.")
#print "His teeth are usually %s depending on the coffee." % my_teeth
# this line is tricky, try to get it exactly right
#print "If I add %d, %d, and %d I get %d." % (
# my_age, my_height, my_weight, my_age + my_height + my_weight)
print(f"If I add {my_age}, {my_height}, and {my_weight} I get {my_age + my_height + my_weight}.")
| <commit_before>cars = 100
space_in_a_car = 4.0
drivers = 30
passengers = 90
cars_not_driven = cars - drivers
cars_driven = drivers
carpool_capacity = cars_driven * space_in_a_car
average_passengers_per_car = passengers / cars_driven
print("There are", cars, "cars available.")
print("There are only", drivers, "drivers available.")
print("There will be", cars_not_driven, "empty cars today.")
print("We can transport", carpool_capacity, "people today.")
print("We have", passengers, "to carpool today.")
print("We need to put about", average_passengers_per_car, "in each car.")<commit_msg>Update exercise of variables and names<commit_after>cars = 100
space_in_a_car = 4.0
drivers = 30
passengers = 90
cars_not_driven = cars - drivers
cars_driven = drivers
carpool_capacity = cars_driven * space_in_a_car
average_passengers_per_car = passengers / cars_driven
print("There are", cars, "cars available.")
print("There are only", drivers, "drivers available.")
print("There will be", cars_not_driven, "empty cars today.")
print("We can transport", carpool_capacity, "people today.")
print("We have", passengers, "to carpool today.")
print("We need to put about", average_passengers_per_car, "in each car.")
# Upgrade part
my_name = 'Binh D. Nguyen'
my_age = 22
my_height = 74 # inches
my_weight = 180 # lbs
my_eyes = 'Black'
my_teeth = 'White'
my_hair = 'Black'
#print "Let's talk about %s." % my_name
print(f"Let's talk about {my_name:}.")
#print "He's %d inches tall." % my_height
print("He's {:d} inches tall.".format(my_height))
#print "He's %d pounds heavy." % my_weight
#print "Actually that's not too heavy."
#print "He's got %s eyes and %s hair." % (my_eyes, my_hair)
print(f"He's got { my_eyes:} eyes and {my_hair:} hair.")
#print "His teeth are usually %s depending on the coffee." % my_teeth
# this line is tricky, try to get it exactly right
#print "If I add %d, %d, and %d I get %d." % (
# my_age, my_height, my_weight, my_age + my_height + my_weight)
print(f"If I add {my_age}, {my_height}, and {my_weight} I get {my_age + my_height + my_weight}.")
|
cc51207881be9c76690bcbc1ce14d048ecc71d76 | commweb/cartridge_hook.py | commweb/cartridge_hook.py | from cartridge.shop.checkout import CheckoutError
from cartridge.shop.models import Cart
from commweb.exc import PaymentDeclinedError
from commweb.purchase import Purchase
def cartridge_payment_handler(request, order_form, order):
cart = Cart.objects.from_request(request)
trans_id = 'WFS_%d' % order.id
p = Purchase(cart.total_price(), trans_id,
order_form.cleaned_data['card_number'],
order_form.cleaned_data['card_expiry_year'][2:4],
order_form.cleaned_data['card_expiry_month'],
order_form.cleaned_data['card_ccv'])
try:
p.process()
return trans_id
except PaymentDeclinedError, e:
raise CheckoutError('Payment declined: %s' % str(e))
| from cartridge.shop.checkout import CheckoutError
from commweb.exc import PaymentDeclinedError
from commweb.purchase import Purchase
def cartridge_payment_handler(request, order_form, order):
trans_id = 'WFS_%d' % order.id
p = Purchase(order.total, trans_id,
order_form.cleaned_data['card_number'],
order_form.cleaned_data['card_expiry_year'][2:4],
order_form.cleaned_data['card_expiry_month'],
order_form.cleaned_data['card_ccv'])
try:
p.process()
return trans_id
except PaymentDeclinedError, e:
raise CheckoutError('Payment declined: %s' % str(e))
| Fix bug where shipping wasn't included in payment amount with cartridge handler | Fix bug where shipping wasn't included in payment amount with cartridge handler
| Python | bsd-2-clause | sjkingo/django-commweb | from cartridge.shop.checkout import CheckoutError
from cartridge.shop.models import Cart
from commweb.exc import PaymentDeclinedError
from commweb.purchase import Purchase
def cartridge_payment_handler(request, order_form, order):
cart = Cart.objects.from_request(request)
trans_id = 'WFS_%d' % order.id
p = Purchase(cart.total_price(), trans_id,
order_form.cleaned_data['card_number'],
order_form.cleaned_data['card_expiry_year'][2:4],
order_form.cleaned_data['card_expiry_month'],
order_form.cleaned_data['card_ccv'])
try:
p.process()
return trans_id
except PaymentDeclinedError, e:
raise CheckoutError('Payment declined: %s' % str(e))
Fix bug where shipping wasn't included in payment amount with cartridge handler | from cartridge.shop.checkout import CheckoutError
from commweb.exc import PaymentDeclinedError
from commweb.purchase import Purchase
def cartridge_payment_handler(request, order_form, order):
trans_id = 'WFS_%d' % order.id
p = Purchase(order.total, trans_id,
order_form.cleaned_data['card_number'],
order_form.cleaned_data['card_expiry_year'][2:4],
order_form.cleaned_data['card_expiry_month'],
order_form.cleaned_data['card_ccv'])
try:
p.process()
return trans_id
except PaymentDeclinedError, e:
raise CheckoutError('Payment declined: %s' % str(e))
| <commit_before>from cartridge.shop.checkout import CheckoutError
from cartridge.shop.models import Cart
from commweb.exc import PaymentDeclinedError
from commweb.purchase import Purchase
def cartridge_payment_handler(request, order_form, order):
cart = Cart.objects.from_request(request)
trans_id = 'WFS_%d' % order.id
p = Purchase(cart.total_price(), trans_id,
order_form.cleaned_data['card_number'],
order_form.cleaned_data['card_expiry_year'][2:4],
order_form.cleaned_data['card_expiry_month'],
order_form.cleaned_data['card_ccv'])
try:
p.process()
return trans_id
except PaymentDeclinedError, e:
raise CheckoutError('Payment declined: %s' % str(e))
<commit_msg>Fix bug where shipping wasn't included in payment amount with cartridge handler<commit_after> | from cartridge.shop.checkout import CheckoutError
from commweb.exc import PaymentDeclinedError
from commweb.purchase import Purchase
def cartridge_payment_handler(request, order_form, order):
trans_id = 'WFS_%d' % order.id
p = Purchase(order.total, trans_id,
order_form.cleaned_data['card_number'],
order_form.cleaned_data['card_expiry_year'][2:4],
order_form.cleaned_data['card_expiry_month'],
order_form.cleaned_data['card_ccv'])
try:
p.process()
return trans_id
except PaymentDeclinedError, e:
raise CheckoutError('Payment declined: %s' % str(e))
| from cartridge.shop.checkout import CheckoutError
from cartridge.shop.models import Cart
from commweb.exc import PaymentDeclinedError
from commweb.purchase import Purchase
def cartridge_payment_handler(request, order_form, order):
cart = Cart.objects.from_request(request)
trans_id = 'WFS_%d' % order.id
p = Purchase(cart.total_price(), trans_id,
order_form.cleaned_data['card_number'],
order_form.cleaned_data['card_expiry_year'][2:4],
order_form.cleaned_data['card_expiry_month'],
order_form.cleaned_data['card_ccv'])
try:
p.process()
return trans_id
except PaymentDeclinedError, e:
raise CheckoutError('Payment declined: %s' % str(e))
Fix bug where shipping wasn't included in payment amount with cartridge handlerfrom cartridge.shop.checkout import CheckoutError
from commweb.exc import PaymentDeclinedError
from commweb.purchase import Purchase
def cartridge_payment_handler(request, order_form, order):
trans_id = 'WFS_%d' % order.id
p = Purchase(order.total, trans_id,
order_form.cleaned_data['card_number'],
order_form.cleaned_data['card_expiry_year'][2:4],
order_form.cleaned_data['card_expiry_month'],
order_form.cleaned_data['card_ccv'])
try:
p.process()
return trans_id
except PaymentDeclinedError, e:
raise CheckoutError('Payment declined: %s' % str(e))
| <commit_before>from cartridge.shop.checkout import CheckoutError
from cartridge.shop.models import Cart
from commweb.exc import PaymentDeclinedError
from commweb.purchase import Purchase
def cartridge_payment_handler(request, order_form, order):
cart = Cart.objects.from_request(request)
trans_id = 'WFS_%d' % order.id
p = Purchase(cart.total_price(), trans_id,
order_form.cleaned_data['card_number'],
order_form.cleaned_data['card_expiry_year'][2:4],
order_form.cleaned_data['card_expiry_month'],
order_form.cleaned_data['card_ccv'])
try:
p.process()
return trans_id
except PaymentDeclinedError, e:
raise CheckoutError('Payment declined: %s' % str(e))
<commit_msg>Fix bug where shipping wasn't included in payment amount with cartridge handler<commit_after>from cartridge.shop.checkout import CheckoutError
from commweb.exc import PaymentDeclinedError
from commweb.purchase import Purchase
def cartridge_payment_handler(request, order_form, order):
trans_id = 'WFS_%d' % order.id
p = Purchase(order.total, trans_id,
order_form.cleaned_data['card_number'],
order_form.cleaned_data['card_expiry_year'][2:4],
order_form.cleaned_data['card_expiry_month'],
order_form.cleaned_data['card_ccv'])
try:
p.process()
return trans_id
except PaymentDeclinedError, e:
raise CheckoutError('Payment declined: %s' % str(e))
|
09371cf8f27c49dec97b05e26ede4709fb01aa81 | examples/filter/simple_example.py | examples/filter/simple_example.py | import pyrox.filtering as filtering
class SimpleFilter(filtering.HttpFilter):
"""
This is an example of a simple filter that simply prints out the
user-agent value from the header
"""
@filtering.handles_request_head
def on_request_head(self, request_message):
user_agent_header = request_message.get_header('user-agent')
if user_agent_header and len(user_agent_header.values) > 0:
# If there is a user-agent value then print it out and pass
# the request upstream
print(user_agent_header.values[0])
return filtering.pass_event()
else:
# If there is no user-agent, then reject the request
return filtering.reject()
| import pyrox.filtering as filtering
class SimpleFilter(filtering.HttpFilter):
"""
This is an example of a simple filter that simply prints out the
user-agent value from the header
"""
@filtering.handles_request_head
def on_request_head(self, request_message):
user_agent_header = request_message.get_header('user-agent')
if user_agent_header and len(user_agent_header.values) > 0:
# If there is a user-agent value then print it out and pass
# the request upstream
print(user_agent_header.values[0])
return filtering.next()
else:
# If there is no user-agent, then reject the request
return filtering.reject()
| Remove pass_event (deprecated) and replace with next | Remove pass_event (deprecated) and replace with next | Python | mit | zinic/pyrox,jon-armstrong/pyrox,jon-armstrong/pyrox,akatrevorjay/pyrox,zinic/pyrox,jon-armstrong/pyrox,zinic/pyrox,akatrevorjay/pyrox,akatrevorjay/pyrox | import pyrox.filtering as filtering
class SimpleFilter(filtering.HttpFilter):
"""
This is an example of a simple filter that simply prints out the
user-agent value from the header
"""
@filtering.handles_request_head
def on_request_head(self, request_message):
user_agent_header = request_message.get_header('user-agent')
if user_agent_header and len(user_agent_header.values) > 0:
# If there is a user-agent value then print it out and pass
# the request upstream
print(user_agent_header.values[0])
return filtering.pass_event()
else:
# If there is no user-agent, then reject the request
return filtering.reject()
Remove pass_event (deprecated) and replace with next | import pyrox.filtering as filtering
class SimpleFilter(filtering.HttpFilter):
"""
This is an example of a simple filter that simply prints out the
user-agent value from the header
"""
@filtering.handles_request_head
def on_request_head(self, request_message):
user_agent_header = request_message.get_header('user-agent')
if user_agent_header and len(user_agent_header.values) > 0:
# If there is a user-agent value then print it out and pass
# the request upstream
print(user_agent_header.values[0])
return filtering.next()
else:
# If there is no user-agent, then reject the request
return filtering.reject()
| <commit_before>import pyrox.filtering as filtering
class SimpleFilter(filtering.HttpFilter):
"""
This is an example of a simple filter that simply prints out the
user-agent value from the header
"""
@filtering.handles_request_head
def on_request_head(self, request_message):
user_agent_header = request_message.get_header('user-agent')
if user_agent_header and len(user_agent_header.values) > 0:
# If there is a user-agent value then print it out and pass
# the request upstream
print(user_agent_header.values[0])
return filtering.pass_event()
else:
# If there is no user-agent, then reject the request
return filtering.reject()
<commit_msg>Remove pass_event (deprecated) and replace with next<commit_after> | import pyrox.filtering as filtering
class SimpleFilter(filtering.HttpFilter):
"""
This is an example of a simple filter that simply prints out the
user-agent value from the header
"""
@filtering.handles_request_head
def on_request_head(self, request_message):
user_agent_header = request_message.get_header('user-agent')
if user_agent_header and len(user_agent_header.values) > 0:
# If there is a user-agent value then print it out and pass
# the request upstream
print(user_agent_header.values[0])
return filtering.next()
else:
# If there is no user-agent, then reject the request
return filtering.reject()
| import pyrox.filtering as filtering
class SimpleFilter(filtering.HttpFilter):
"""
This is an example of a simple filter that simply prints out the
user-agent value from the header
"""
@filtering.handles_request_head
def on_request_head(self, request_message):
user_agent_header = request_message.get_header('user-agent')
if user_agent_header and len(user_agent_header.values) > 0:
# If there is a user-agent value then print it out and pass
# the request upstream
print(user_agent_header.values[0])
return filtering.pass_event()
else:
# If there is no user-agent, then reject the request
return filtering.reject()
Remove pass_event (deprecated) and replace with nextimport pyrox.filtering as filtering
class SimpleFilter(filtering.HttpFilter):
"""
This is an example of a simple filter that simply prints out the
user-agent value from the header
"""
@filtering.handles_request_head
def on_request_head(self, request_message):
user_agent_header = request_message.get_header('user-agent')
if user_agent_header and len(user_agent_header.values) > 0:
# If there is a user-agent value then print it out and pass
# the request upstream
print(user_agent_header.values[0])
return filtering.next()
else:
# If there is no user-agent, then reject the request
return filtering.reject()
| <commit_before>import pyrox.filtering as filtering
class SimpleFilter(filtering.HttpFilter):
"""
This is an example of a simple filter that simply prints out the
user-agent value from the header
"""
@filtering.handles_request_head
def on_request_head(self, request_message):
user_agent_header = request_message.get_header('user-agent')
if user_agent_header and len(user_agent_header.values) > 0:
# If there is a user-agent value then print it out and pass
# the request upstream
print(user_agent_header.values[0])
return filtering.pass_event()
else:
# If there is no user-agent, then reject the request
return filtering.reject()
<commit_msg>Remove pass_event (deprecated) and replace with next<commit_after>import pyrox.filtering as filtering
class SimpleFilter(filtering.HttpFilter):
"""
This is an example of a simple filter that simply prints out the
user-agent value from the header
"""
@filtering.handles_request_head
def on_request_head(self, request_message):
user_agent_header = request_message.get_header('user-agent')
if user_agent_header and len(user_agent_header.values) > 0:
# If there is a user-agent value then print it out and pass
# the request upstream
print(user_agent_header.values[0])
return filtering.next()
else:
# If there is no user-agent, then reject the request
return filtering.reject()
|
83cbec4ccf669c997bbe6c7131f63ad08a482c39 | examples/mnist-deepautoencoder.py | examples/mnist-deepautoencoder.py | #!/usr/bin/env python
import matplotlib.pyplot as plt
import theanets
from utils import load_mnist, plot_layers, plot_images
train, valid, _ = load_mnist()
e = theanets.Experiment(
theanets.Autoencoder,
layers=(784, 256, 64, 36, 64, 256, 784),
train_batches=100,
tied_weights=True,
)
e.train(train, valid, optimize='layerwise', patience=1, min_improvement=0.1)
e.train(train, valid)
plot_layers([e.network.find(i, 0) for i in (1, 2, 3)], tied_weights=True)
plt.tight_layout()
plt.show()
valid = valid[:16*16]
plot_images(valid, 121, 'Sample data')
plot_images(e.network.predict(valid), 122, 'Reconstructed data')
plt.tight_layout()
plt.show()
| #!/usr/bin/env python
import matplotlib.pyplot as plt
import theanets
from utils import load_mnist, plot_layers, plot_images
train, valid, _ = load_mnist()
e = theanets.Experiment(
theanets.Autoencoder,
layers=(784, 256, 64, 36, ('tied', 64), ('tied', 256), ('tied', 784)),
train_batches=100,
)
e.train(train, valid, optimize='layerwise', patience=1, min_improvement=0.1)
e.train(train, valid)
plot_layers([e.network.find(i, 'w') for i in (1, 2, 3)], tied_weights=True)
plt.tight_layout()
plt.show()
valid = valid[:16*16]
plot_images(valid, 121, 'Sample data')
plot_images(e.network.predict(valid), 122, 'Reconstructed data')
plt.tight_layout()
plt.show()
| Update example with newer tied layer specification. | Update example with newer tied layer specification.
| Python | mit | lmjohns3/theanets,chrinide/theanets,devdoer/theanets | #!/usr/bin/env python
import matplotlib.pyplot as plt
import theanets
from utils import load_mnist, plot_layers, plot_images
train, valid, _ = load_mnist()
e = theanets.Experiment(
theanets.Autoencoder,
layers=(784, 256, 64, 36, 64, 256, 784),
train_batches=100,
tied_weights=True,
)
e.train(train, valid, optimize='layerwise', patience=1, min_improvement=0.1)
e.train(train, valid)
plot_layers([e.network.find(i, 0) for i in (1, 2, 3)], tied_weights=True)
plt.tight_layout()
plt.show()
valid = valid[:16*16]
plot_images(valid, 121, 'Sample data')
plot_images(e.network.predict(valid), 122, 'Reconstructed data')
plt.tight_layout()
plt.show()
Update example with newer tied layer specification. | #!/usr/bin/env python
import matplotlib.pyplot as plt
import theanets
from utils import load_mnist, plot_layers, plot_images
train, valid, _ = load_mnist()
e = theanets.Experiment(
theanets.Autoencoder,
layers=(784, 256, 64, 36, ('tied', 64), ('tied', 256), ('tied', 784)),
train_batches=100,
)
e.train(train, valid, optimize='layerwise', patience=1, min_improvement=0.1)
e.train(train, valid)
plot_layers([e.network.find(i, 'w') for i in (1, 2, 3)], tied_weights=True)
plt.tight_layout()
plt.show()
valid = valid[:16*16]
plot_images(valid, 121, 'Sample data')
plot_images(e.network.predict(valid), 122, 'Reconstructed data')
plt.tight_layout()
plt.show()
| <commit_before>#!/usr/bin/env python
import matplotlib.pyplot as plt
import theanets
from utils import load_mnist, plot_layers, plot_images
train, valid, _ = load_mnist()
e = theanets.Experiment(
theanets.Autoencoder,
layers=(784, 256, 64, 36, 64, 256, 784),
train_batches=100,
tied_weights=True,
)
e.train(train, valid, optimize='layerwise', patience=1, min_improvement=0.1)
e.train(train, valid)
plot_layers([e.network.find(i, 0) for i in (1, 2, 3)], tied_weights=True)
plt.tight_layout()
plt.show()
valid = valid[:16*16]
plot_images(valid, 121, 'Sample data')
plot_images(e.network.predict(valid), 122, 'Reconstructed data')
plt.tight_layout()
plt.show()
<commit_msg>Update example with newer tied layer specification.<commit_after> | #!/usr/bin/env python
import matplotlib.pyplot as plt
import theanets
from utils import load_mnist, plot_layers, plot_images
train, valid, _ = load_mnist()
e = theanets.Experiment(
theanets.Autoencoder,
layers=(784, 256, 64, 36, ('tied', 64), ('tied', 256), ('tied', 784)),
train_batches=100,
)
e.train(train, valid, optimize='layerwise', patience=1, min_improvement=0.1)
e.train(train, valid)
plot_layers([e.network.find(i, 'w') for i in (1, 2, 3)], tied_weights=True)
plt.tight_layout()
plt.show()
valid = valid[:16*16]
plot_images(valid, 121, 'Sample data')
plot_images(e.network.predict(valid), 122, 'Reconstructed data')
plt.tight_layout()
plt.show()
| #!/usr/bin/env python
import matplotlib.pyplot as plt
import theanets
from utils import load_mnist, plot_layers, plot_images
train, valid, _ = load_mnist()
e = theanets.Experiment(
theanets.Autoencoder,
layers=(784, 256, 64, 36, 64, 256, 784),
train_batches=100,
tied_weights=True,
)
e.train(train, valid, optimize='layerwise', patience=1, min_improvement=0.1)
e.train(train, valid)
plot_layers([e.network.find(i, 0) for i in (1, 2, 3)], tied_weights=True)
plt.tight_layout()
plt.show()
valid = valid[:16*16]
plot_images(valid, 121, 'Sample data')
plot_images(e.network.predict(valid), 122, 'Reconstructed data')
plt.tight_layout()
plt.show()
Update example with newer tied layer specification.#!/usr/bin/env python
import matplotlib.pyplot as plt
import theanets
from utils import load_mnist, plot_layers, plot_images
train, valid, _ = load_mnist()
e = theanets.Experiment(
theanets.Autoencoder,
layers=(784, 256, 64, 36, ('tied', 64), ('tied', 256), ('tied', 784)),
train_batches=100,
)
e.train(train, valid, optimize='layerwise', patience=1, min_improvement=0.1)
e.train(train, valid)
plot_layers([e.network.find(i, 'w') for i in (1, 2, 3)], tied_weights=True)
plt.tight_layout()
plt.show()
valid = valid[:16*16]
plot_images(valid, 121, 'Sample data')
plot_images(e.network.predict(valid), 122, 'Reconstructed data')
plt.tight_layout()
plt.show()
| <commit_before>#!/usr/bin/env python
import matplotlib.pyplot as plt
import theanets
from utils import load_mnist, plot_layers, plot_images
train, valid, _ = load_mnist()
e = theanets.Experiment(
theanets.Autoencoder,
layers=(784, 256, 64, 36, 64, 256, 784),
train_batches=100,
tied_weights=True,
)
e.train(train, valid, optimize='layerwise', patience=1, min_improvement=0.1)
e.train(train, valid)
plot_layers([e.network.find(i, 0) for i in (1, 2, 3)], tied_weights=True)
plt.tight_layout()
plt.show()
valid = valid[:16*16]
plot_images(valid, 121, 'Sample data')
plot_images(e.network.predict(valid), 122, 'Reconstructed data')
plt.tight_layout()
plt.show()
<commit_msg>Update example with newer tied layer specification.<commit_after>#!/usr/bin/env python
import matplotlib.pyplot as plt
import theanets
from utils import load_mnist, plot_layers, plot_images
train, valid, _ = load_mnist()
e = theanets.Experiment(
theanets.Autoencoder,
layers=(784, 256, 64, 36, ('tied', 64), ('tied', 256), ('tied', 784)),
train_batches=100,
)
e.train(train, valid, optimize='layerwise', patience=1, min_improvement=0.1)
e.train(train, valid)
plot_layers([e.network.find(i, 'w') for i in (1, 2, 3)], tied_weights=True)
plt.tight_layout()
plt.show()
valid = valid[:16*16]
plot_images(valid, 121, 'Sample data')
plot_images(e.network.predict(valid), 122, 'Reconstructed data')
plt.tight_layout()
plt.show()
|
f10ac5cb7e01feeaaab5b8d308cb0640afaa895c | tests/test_encoding.py | tests/test_encoding.py | # -*- encoding: utf-8 -*-
import pytest
from owslib.fes import PropertyIsEqualTo
from pydov.search.boring import BoringSearch
from tests.abstract import (
AbstractTestSearch,
service_ok,
)
class TestEncoding(AbstractTestSearch):
"""Class grouping tests related to encoding issues."""
@pytest.mark.online
@pytest.mark.skipif(not service_ok(), reason="DOV service is unreachable")
def test_search(self):
"""Test the search method with strange character in the output.
Test whether the output has the correct encoding.
"""
boringsearch = BoringSearch()
query = PropertyIsEqualTo(
propertyname='pkey_boring',
literal='https://www.dov.vlaanderen.be/data/boring/1928-031159')
df = boringsearch.search(query=query,
return_fields=('pkey_boring', 'uitvoerder'))
assert df.uitvoerder[0] == 'Societé Belge des Bétons'
| # -*- encoding: utf-8 -*-
import pytest
from owslib.fes import PropertyIsEqualTo
from pydov.search.boring import BoringSearch
from tests.abstract import (
AbstractTestSearch,
service_ok,
)
class TestEncoding(AbstractTestSearch):
"""Class grouping tests related to encoding issues."""
@pytest.mark.online
@pytest.mark.skipif(not service_ok(), reason="DOV service is unreachable")
def test_search(self):
"""Test the search method with strange character in the output.
Test whether the output has the correct encoding.
"""
boringsearch = BoringSearch()
query = PropertyIsEqualTo(
propertyname='pkey_boring',
literal='https://www.dov.vlaanderen.be/data/boring/1928-031159')
df = boringsearch.search(query=query,
return_fields=('pkey_boring', 'uitvoerder'))
assert df.uitvoerder[0] == u'Societé Belge des Bétons'
| Fix encoding test: compare with unicode string. | Fix encoding test: compare with unicode string.
| Python | mit | DOV-Vlaanderen/pydov | # -*- encoding: utf-8 -*-
import pytest
from owslib.fes import PropertyIsEqualTo
from pydov.search.boring import BoringSearch
from tests.abstract import (
AbstractTestSearch,
service_ok,
)
class TestEncoding(AbstractTestSearch):
"""Class grouping tests related to encoding issues."""
@pytest.mark.online
@pytest.mark.skipif(not service_ok(), reason="DOV service is unreachable")
def test_search(self):
"""Test the search method with strange character in the output.
Test whether the output has the correct encoding.
"""
boringsearch = BoringSearch()
query = PropertyIsEqualTo(
propertyname='pkey_boring',
literal='https://www.dov.vlaanderen.be/data/boring/1928-031159')
df = boringsearch.search(query=query,
return_fields=('pkey_boring', 'uitvoerder'))
assert df.uitvoerder[0] == 'Societé Belge des Bétons'
Fix encoding test: compare with unicode string. | # -*- encoding: utf-8 -*-
import pytest
from owslib.fes import PropertyIsEqualTo
from pydov.search.boring import BoringSearch
from tests.abstract import (
AbstractTestSearch,
service_ok,
)
class TestEncoding(AbstractTestSearch):
"""Class grouping tests related to encoding issues."""
@pytest.mark.online
@pytest.mark.skipif(not service_ok(), reason="DOV service is unreachable")
def test_search(self):
"""Test the search method with strange character in the output.
Test whether the output has the correct encoding.
"""
boringsearch = BoringSearch()
query = PropertyIsEqualTo(
propertyname='pkey_boring',
literal='https://www.dov.vlaanderen.be/data/boring/1928-031159')
df = boringsearch.search(query=query,
return_fields=('pkey_boring', 'uitvoerder'))
assert df.uitvoerder[0] == u'Societé Belge des Bétons'
| <commit_before># -*- encoding: utf-8 -*-
import pytest
from owslib.fes import PropertyIsEqualTo
from pydov.search.boring import BoringSearch
from tests.abstract import (
AbstractTestSearch,
service_ok,
)
class TestEncoding(AbstractTestSearch):
"""Class grouping tests related to encoding issues."""
@pytest.mark.online
@pytest.mark.skipif(not service_ok(), reason="DOV service is unreachable")
def test_search(self):
"""Test the search method with strange character in the output.
Test whether the output has the correct encoding.
"""
boringsearch = BoringSearch()
query = PropertyIsEqualTo(
propertyname='pkey_boring',
literal='https://www.dov.vlaanderen.be/data/boring/1928-031159')
df = boringsearch.search(query=query,
return_fields=('pkey_boring', 'uitvoerder'))
assert df.uitvoerder[0] == 'Societé Belge des Bétons'
<commit_msg>Fix encoding test: compare with unicode string.<commit_after> | # -*- encoding: utf-8 -*-
import pytest
from owslib.fes import PropertyIsEqualTo
from pydov.search.boring import BoringSearch
from tests.abstract import (
AbstractTestSearch,
service_ok,
)
class TestEncoding(AbstractTestSearch):
"""Class grouping tests related to encoding issues."""
@pytest.mark.online
@pytest.mark.skipif(not service_ok(), reason="DOV service is unreachable")
def test_search(self):
"""Test the search method with strange character in the output.
Test whether the output has the correct encoding.
"""
boringsearch = BoringSearch()
query = PropertyIsEqualTo(
propertyname='pkey_boring',
literal='https://www.dov.vlaanderen.be/data/boring/1928-031159')
df = boringsearch.search(query=query,
return_fields=('pkey_boring', 'uitvoerder'))
assert df.uitvoerder[0] == u'Societé Belge des Bétons'
| # -*- encoding: utf-8 -*-
import pytest
from owslib.fes import PropertyIsEqualTo
from pydov.search.boring import BoringSearch
from tests.abstract import (
AbstractTestSearch,
service_ok,
)
class TestEncoding(AbstractTestSearch):
"""Class grouping tests related to encoding issues."""
@pytest.mark.online
@pytest.mark.skipif(not service_ok(), reason="DOV service is unreachable")
def test_search(self):
"""Test the search method with strange character in the output.
Test whether the output has the correct encoding.
"""
boringsearch = BoringSearch()
query = PropertyIsEqualTo(
propertyname='pkey_boring',
literal='https://www.dov.vlaanderen.be/data/boring/1928-031159')
df = boringsearch.search(query=query,
return_fields=('pkey_boring', 'uitvoerder'))
assert df.uitvoerder[0] == 'Societé Belge des Bétons'
Fix encoding test: compare with unicode string.# -*- encoding: utf-8 -*-
import pytest
from owslib.fes import PropertyIsEqualTo
from pydov.search.boring import BoringSearch
from tests.abstract import (
AbstractTestSearch,
service_ok,
)
class TestEncoding(AbstractTestSearch):
"""Class grouping tests related to encoding issues."""
@pytest.mark.online
@pytest.mark.skipif(not service_ok(), reason="DOV service is unreachable")
def test_search(self):
"""Test the search method with strange character in the output.
Test whether the output has the correct encoding.
"""
boringsearch = BoringSearch()
query = PropertyIsEqualTo(
propertyname='pkey_boring',
literal='https://www.dov.vlaanderen.be/data/boring/1928-031159')
df = boringsearch.search(query=query,
return_fields=('pkey_boring', 'uitvoerder'))
assert df.uitvoerder[0] == u'Societé Belge des Bétons'
| <commit_before># -*- encoding: utf-8 -*-
import pytest
from owslib.fes import PropertyIsEqualTo
from pydov.search.boring import BoringSearch
from tests.abstract import (
AbstractTestSearch,
service_ok,
)
class TestEncoding(AbstractTestSearch):
"""Class grouping tests related to encoding issues."""
@pytest.mark.online
@pytest.mark.skipif(not service_ok(), reason="DOV service is unreachable")
def test_search(self):
"""Test the search method with strange character in the output.
Test whether the output has the correct encoding.
"""
boringsearch = BoringSearch()
query = PropertyIsEqualTo(
propertyname='pkey_boring',
literal='https://www.dov.vlaanderen.be/data/boring/1928-031159')
df = boringsearch.search(query=query,
return_fields=('pkey_boring', 'uitvoerder'))
assert df.uitvoerder[0] == 'Societé Belge des Bétons'
<commit_msg>Fix encoding test: compare with unicode string.<commit_after># -*- encoding: utf-8 -*-
import pytest
from owslib.fes import PropertyIsEqualTo
from pydov.search.boring import BoringSearch
from tests.abstract import (
AbstractTestSearch,
service_ok,
)
class TestEncoding(AbstractTestSearch):
"""Class grouping tests related to encoding issues."""
@pytest.mark.online
@pytest.mark.skipif(not service_ok(), reason="DOV service is unreachable")
def test_search(self):
"""Test the search method with strange character in the output.
Test whether the output has the correct encoding.
"""
boringsearch = BoringSearch()
query = PropertyIsEqualTo(
propertyname='pkey_boring',
literal='https://www.dov.vlaanderen.be/data/boring/1928-031159')
df = boringsearch.search(query=query,
return_fields=('pkey_boring', 'uitvoerder'))
assert df.uitvoerder[0] == u'Societé Belge des Bétons'
|
5c2b5b4ad973717ab35c75ff8d5d63d87c15cf79 | twphotos/settings.py | twphotos/settings.py | import ConfigParser
import os
import sys
USER_DIR = os.path.join(os.path.expanduser('~'))
USER_CONFIG = os.path.join(USER_DIR, '.twphotos')
d = os.path.dirname(__file__)
PROJECT_PATH = os.path.abspath(os.path.join(d, os.pardir))
TEST_CONFIG = os.path.join(PROJECT_PATH, '.twphotos')
sys.path.append(os.path.join(PROJECT_PATH, 'python-twitter'))
config = ConfigParser.ConfigParser()
if os.environ.get('TWPHOTOS_TEST_CONFIG'):
config.read(TEST_CONFIG)
else:
config.read(USER_CONFIG)
items = {}
item_names = [
'consumer_key',
'consumer_secret',
'access_token_key',
'access_token_secret',
]
if config.has_section('twphotos'):
items = dict(config.items('twphotos'))
for name in items:
if name not in item_names:
raise Exception('Unknown name "%s" in credential file.' % name)
if len(items) < 4:
raise Exception('No credentials found.')
CONSUMER_KEY = items.get('consumer_key')
CONSUMER_SECRET = items.get('consumer_secret')
ACCESS_TOKEN = items.get('access_token_key')
ACCESS_TOKEN_SECRET = items.get('access_token_secret')
COUNT_PER_GET = 200
MEDIA_SIZES = ['large', 'medium', 'small', 'thumb']
| import ConfigParser
import os
import sys
USER_DIR = os.path.join(os.path.expanduser('~'))
USER_CONFIG = os.path.join(USER_DIR, '.twphotos')
d = os.path.dirname(__file__)
PROJECT_PATH = os.path.abspath(os.path.join(d, os.pardir))
TEST_CONFIG = os.path.join(PROJECT_PATH, '.twphotos')
sys.path.insert(1, os.path.join(PROJECT_PATH, 'python-twitter'))
config = ConfigParser.ConfigParser()
if os.environ.get('TWPHOTOS_TEST_CONFIG'):
config.read(TEST_CONFIG)
else:
config.read(USER_CONFIG)
items = {}
item_names = [
'consumer_key',
'consumer_secret',
'access_token_key',
'access_token_secret',
]
if config.has_section('twphotos'):
items = dict(config.items('twphotos'))
for name in items:
if name not in item_names:
raise Exception('Unknown name "%s" in credential file.' % name)
if len(items) < 4:
raise Exception('No credentials found.')
CONSUMER_KEY = items.get('consumer_key')
CONSUMER_SECRET = items.get('consumer_secret')
ACCESS_TOKEN = items.get('access_token_key')
ACCESS_TOKEN_SECRET = items.get('access_token_secret')
COUNT_PER_GET = 200
MEDIA_SIZES = ['large', 'medium', 'small', 'thumb']
| Add sys.path for python-twitter for local development | Add sys.path for python-twitter for local development
| Python | bsd-2-clause | shichao-an/twitter-photos | import ConfigParser
import os
import sys
USER_DIR = os.path.join(os.path.expanduser('~'))
USER_CONFIG = os.path.join(USER_DIR, '.twphotos')
d = os.path.dirname(__file__)
PROJECT_PATH = os.path.abspath(os.path.join(d, os.pardir))
TEST_CONFIG = os.path.join(PROJECT_PATH, '.twphotos')
sys.path.append(os.path.join(PROJECT_PATH, 'python-twitter'))
config = ConfigParser.ConfigParser()
if os.environ.get('TWPHOTOS_TEST_CONFIG'):
config.read(TEST_CONFIG)
else:
config.read(USER_CONFIG)
items = {}
item_names = [
'consumer_key',
'consumer_secret',
'access_token_key',
'access_token_secret',
]
if config.has_section('twphotos'):
items = dict(config.items('twphotos'))
for name in items:
if name not in item_names:
raise Exception('Unknown name "%s" in credential file.' % name)
if len(items) < 4:
raise Exception('No credentials found.')
CONSUMER_KEY = items.get('consumer_key')
CONSUMER_SECRET = items.get('consumer_secret')
ACCESS_TOKEN = items.get('access_token_key')
ACCESS_TOKEN_SECRET = items.get('access_token_secret')
COUNT_PER_GET = 200
MEDIA_SIZES = ['large', 'medium', 'small', 'thumb']
Add sys.path for python-twitter for local development | import ConfigParser
import os
import sys
USER_DIR = os.path.join(os.path.expanduser('~'))
USER_CONFIG = os.path.join(USER_DIR, '.twphotos')
d = os.path.dirname(__file__)
PROJECT_PATH = os.path.abspath(os.path.join(d, os.pardir))
TEST_CONFIG = os.path.join(PROJECT_PATH, '.twphotos')
sys.path.insert(1, os.path.join(PROJECT_PATH, 'python-twitter'))
config = ConfigParser.ConfigParser()
if os.environ.get('TWPHOTOS_TEST_CONFIG'):
config.read(TEST_CONFIG)
else:
config.read(USER_CONFIG)
items = {}
item_names = [
'consumer_key',
'consumer_secret',
'access_token_key',
'access_token_secret',
]
if config.has_section('twphotos'):
items = dict(config.items('twphotos'))
for name in items:
if name not in item_names:
raise Exception('Unknown name "%s" in credential file.' % name)
if len(items) < 4:
raise Exception('No credentials found.')
CONSUMER_KEY = items.get('consumer_key')
CONSUMER_SECRET = items.get('consumer_secret')
ACCESS_TOKEN = items.get('access_token_key')
ACCESS_TOKEN_SECRET = items.get('access_token_secret')
COUNT_PER_GET = 200
MEDIA_SIZES = ['large', 'medium', 'small', 'thumb']
| <commit_before>import ConfigParser
import os
import sys
USER_DIR = os.path.join(os.path.expanduser('~'))
USER_CONFIG = os.path.join(USER_DIR, '.twphotos')
d = os.path.dirname(__file__)
PROJECT_PATH = os.path.abspath(os.path.join(d, os.pardir))
TEST_CONFIG = os.path.join(PROJECT_PATH, '.twphotos')
sys.path.append(os.path.join(PROJECT_PATH, 'python-twitter'))
config = ConfigParser.ConfigParser()
if os.environ.get('TWPHOTOS_TEST_CONFIG'):
config.read(TEST_CONFIG)
else:
config.read(USER_CONFIG)
items = {}
item_names = [
'consumer_key',
'consumer_secret',
'access_token_key',
'access_token_secret',
]
if config.has_section('twphotos'):
items = dict(config.items('twphotos'))
for name in items:
if name not in item_names:
raise Exception('Unknown name "%s" in credential file.' % name)
if len(items) < 4:
raise Exception('No credentials found.')
CONSUMER_KEY = items.get('consumer_key')
CONSUMER_SECRET = items.get('consumer_secret')
ACCESS_TOKEN = items.get('access_token_key')
ACCESS_TOKEN_SECRET = items.get('access_token_secret')
COUNT_PER_GET = 200
MEDIA_SIZES = ['large', 'medium', 'small', 'thumb']
<commit_msg>Add sys.path for python-twitter for local development<commit_after> | import ConfigParser
import os
import sys
USER_DIR = os.path.join(os.path.expanduser('~'))
USER_CONFIG = os.path.join(USER_DIR, '.twphotos')
d = os.path.dirname(__file__)
PROJECT_PATH = os.path.abspath(os.path.join(d, os.pardir))
TEST_CONFIG = os.path.join(PROJECT_PATH, '.twphotos')
sys.path.insert(1, os.path.join(PROJECT_PATH, 'python-twitter'))
config = ConfigParser.ConfigParser()
if os.environ.get('TWPHOTOS_TEST_CONFIG'):
config.read(TEST_CONFIG)
else:
config.read(USER_CONFIG)
items = {}
item_names = [
'consumer_key',
'consumer_secret',
'access_token_key',
'access_token_secret',
]
if config.has_section('twphotos'):
items = dict(config.items('twphotos'))
for name in items:
if name not in item_names:
raise Exception('Unknown name "%s" in credential file.' % name)
if len(items) < 4:
raise Exception('No credentials found.')
CONSUMER_KEY = items.get('consumer_key')
CONSUMER_SECRET = items.get('consumer_secret')
ACCESS_TOKEN = items.get('access_token_key')
ACCESS_TOKEN_SECRET = items.get('access_token_secret')
COUNT_PER_GET = 200
MEDIA_SIZES = ['large', 'medium', 'small', 'thumb']
| import ConfigParser
import os
import sys
USER_DIR = os.path.join(os.path.expanduser('~'))
USER_CONFIG = os.path.join(USER_DIR, '.twphotos')
d = os.path.dirname(__file__)
PROJECT_PATH = os.path.abspath(os.path.join(d, os.pardir))
TEST_CONFIG = os.path.join(PROJECT_PATH, '.twphotos')
sys.path.append(os.path.join(PROJECT_PATH, 'python-twitter'))
config = ConfigParser.ConfigParser()
if os.environ.get('TWPHOTOS_TEST_CONFIG'):
config.read(TEST_CONFIG)
else:
config.read(USER_CONFIG)
items = {}
item_names = [
'consumer_key',
'consumer_secret',
'access_token_key',
'access_token_secret',
]
if config.has_section('twphotos'):
items = dict(config.items('twphotos'))
for name in items:
if name not in item_names:
raise Exception('Unknown name "%s" in credential file.' % name)
if len(items) < 4:
raise Exception('No credentials found.')
CONSUMER_KEY = items.get('consumer_key')
CONSUMER_SECRET = items.get('consumer_secret')
ACCESS_TOKEN = items.get('access_token_key')
ACCESS_TOKEN_SECRET = items.get('access_token_secret')
COUNT_PER_GET = 200
MEDIA_SIZES = ['large', 'medium', 'small', 'thumb']
Add sys.path for python-twitter for local developmentimport ConfigParser
import os
import sys
USER_DIR = os.path.join(os.path.expanduser('~'))
USER_CONFIG = os.path.join(USER_DIR, '.twphotos')
d = os.path.dirname(__file__)
PROJECT_PATH = os.path.abspath(os.path.join(d, os.pardir))
TEST_CONFIG = os.path.join(PROJECT_PATH, '.twphotos')
sys.path.insert(1, os.path.join(PROJECT_PATH, 'python-twitter'))
config = ConfigParser.ConfigParser()
if os.environ.get('TWPHOTOS_TEST_CONFIG'):
config.read(TEST_CONFIG)
else:
config.read(USER_CONFIG)
items = {}
item_names = [
'consumer_key',
'consumer_secret',
'access_token_key',
'access_token_secret',
]
if config.has_section('twphotos'):
items = dict(config.items('twphotos'))
for name in items:
if name not in item_names:
raise Exception('Unknown name "%s" in credential file.' % name)
if len(items) < 4:
raise Exception('No credentials found.')
CONSUMER_KEY = items.get('consumer_key')
CONSUMER_SECRET = items.get('consumer_secret')
ACCESS_TOKEN = items.get('access_token_key')
ACCESS_TOKEN_SECRET = items.get('access_token_secret')
COUNT_PER_GET = 200
MEDIA_SIZES = ['large', 'medium', 'small', 'thumb']
| <commit_before>import ConfigParser
import os
import sys
USER_DIR = os.path.join(os.path.expanduser('~'))
USER_CONFIG = os.path.join(USER_DIR, '.twphotos')
d = os.path.dirname(__file__)
PROJECT_PATH = os.path.abspath(os.path.join(d, os.pardir))
TEST_CONFIG = os.path.join(PROJECT_PATH, '.twphotos')
sys.path.append(os.path.join(PROJECT_PATH, 'python-twitter'))
config = ConfigParser.ConfigParser()
if os.environ.get('TWPHOTOS_TEST_CONFIG'):
config.read(TEST_CONFIG)
else:
config.read(USER_CONFIG)
items = {}
item_names = [
'consumer_key',
'consumer_secret',
'access_token_key',
'access_token_secret',
]
if config.has_section('twphotos'):
items = dict(config.items('twphotos'))
for name in items:
if name not in item_names:
raise Exception('Unknown name "%s" in credential file.' % name)
if len(items) < 4:
raise Exception('No credentials found.')
CONSUMER_KEY = items.get('consumer_key')
CONSUMER_SECRET = items.get('consumer_secret')
ACCESS_TOKEN = items.get('access_token_key')
ACCESS_TOKEN_SECRET = items.get('access_token_secret')
COUNT_PER_GET = 200
MEDIA_SIZES = ['large', 'medium', 'small', 'thumb']
<commit_msg>Add sys.path for python-twitter for local development<commit_after>import ConfigParser
import os
import sys
USER_DIR = os.path.join(os.path.expanduser('~'))
USER_CONFIG = os.path.join(USER_DIR, '.twphotos')
d = os.path.dirname(__file__)
PROJECT_PATH = os.path.abspath(os.path.join(d, os.pardir))
TEST_CONFIG = os.path.join(PROJECT_PATH, '.twphotos')
sys.path.insert(1, os.path.join(PROJECT_PATH, 'python-twitter'))
config = ConfigParser.ConfigParser()
if os.environ.get('TWPHOTOS_TEST_CONFIG'):
config.read(TEST_CONFIG)
else:
config.read(USER_CONFIG)
items = {}
item_names = [
'consumer_key',
'consumer_secret',
'access_token_key',
'access_token_secret',
]
if config.has_section('twphotos'):
items = dict(config.items('twphotos'))
for name in items:
if name not in item_names:
raise Exception('Unknown name "%s" in credential file.' % name)
if len(items) < 4:
raise Exception('No credentials found.')
CONSUMER_KEY = items.get('consumer_key')
CONSUMER_SECRET = items.get('consumer_secret')
ACCESS_TOKEN = items.get('access_token_key')
ACCESS_TOKEN_SECRET = items.get('access_token_secret')
COUNT_PER_GET = 200
MEDIA_SIZES = ['large', 'medium', 'small', 'thumb']
|
79fd5586625d2d7873bc71514eda121325a9646a | linkedin_scraper/commands/people_search.py | linkedin_scraper/commands/people_search.py | from getpass import getpass
from scrapy.commands.crawl import Command as BaseCommand
def sanitize_query(query):
return query.replace(' ', '+')
class Command(BaseCommand):
def short_desc(self):
return "Scrap people from LinkedIn"
def syntax(self):
return "[options] <query>"
def add_options(self, parser):
super().add_options(parser)
parser.add_option('-u', '--username', help='Name of LinkedIn account')
parser.add_option('-p', '--password',
help='Password for LinkedIn account')
def process_options(self, args, opts):
opts.output = opts.output or 'results.csv'
opts.loglevel = opts.loglevel or 'INFO'
super().process_options(args, opts)
people_search_options = {
'query': sanitize_query(args[0]),
'username': opts.username or input(
'Please provide your LinkedIn username: '),
'password': opts.password or getpass(
'Please provide password for your LinkedIn account: ')
}
opts.spargs.update(people_search_options)
def run(self, args, opts):
# Run people_search spider
args = ['people_search']
super().run(args, opts)
| from getpass import getpass
from scrapy.commands.crawl import Command as BaseCommand
from scrapy.exceptions import UsageError
def sanitize_query(query):
return query.replace(' ', '+')
class Command(BaseCommand):
def short_desc(self):
return "Scrap people from LinkedIn"
def syntax(self):
return "[options] <query>"
def add_options(self, parser):
super().add_options(parser)
parser.add_option('-u', '--username', help='Name of LinkedIn account')
parser.add_option('-p', '--password',
help='Password for LinkedIn account')
def process_options(self, args, opts):
opts.output = opts.output or 'results.csv'
opts.loglevel = opts.loglevel or 'INFO'
super().process_options(args, opts)
try:
query = sanitize_query(args[0])
except IndexError:
raise UsageError()
people_search_options = {
'query': query,
'username': opts.username or input(
'Please provide your LinkedIn username: '),
'password': opts.password or getpass(
'Please provide password for your LinkedIn account: ')
}
opts.spargs.update(people_search_options)
def run(self, args, opts):
# Run people_search spider
args = ['people_search']
super().run(args, opts)
| Fix IndexError when running command without arguments. | Fix IndexError when running command without arguments.
| Python | mit | nihn/linkedin-scraper,nihn/linkedin-scraper | from getpass import getpass
from scrapy.commands.crawl import Command as BaseCommand
def sanitize_query(query):
return query.replace(' ', '+')
class Command(BaseCommand):
def short_desc(self):
return "Scrap people from LinkedIn"
def syntax(self):
return "[options] <query>"
def add_options(self, parser):
super().add_options(parser)
parser.add_option('-u', '--username', help='Name of LinkedIn account')
parser.add_option('-p', '--password',
help='Password for LinkedIn account')
def process_options(self, args, opts):
opts.output = opts.output or 'results.csv'
opts.loglevel = opts.loglevel or 'INFO'
super().process_options(args, opts)
people_search_options = {
'query': sanitize_query(args[0]),
'username': opts.username or input(
'Please provide your LinkedIn username: '),
'password': opts.password or getpass(
'Please provide password for your LinkedIn account: ')
}
opts.spargs.update(people_search_options)
def run(self, args, opts):
# Run people_search spider
args = ['people_search']
super().run(args, opts)
Fix IndexError when running command without arguments. | from getpass import getpass
from scrapy.commands.crawl import Command as BaseCommand
from scrapy.exceptions import UsageError
def sanitize_query(query):
return query.replace(' ', '+')
class Command(BaseCommand):
def short_desc(self):
return "Scrap people from LinkedIn"
def syntax(self):
return "[options] <query>"
def add_options(self, parser):
super().add_options(parser)
parser.add_option('-u', '--username', help='Name of LinkedIn account')
parser.add_option('-p', '--password',
help='Password for LinkedIn account')
def process_options(self, args, opts):
opts.output = opts.output or 'results.csv'
opts.loglevel = opts.loglevel or 'INFO'
super().process_options(args, opts)
try:
query = sanitize_query(args[0])
except IndexError:
raise UsageError()
people_search_options = {
'query': query,
'username': opts.username or input(
'Please provide your LinkedIn username: '),
'password': opts.password or getpass(
'Please provide password for your LinkedIn account: ')
}
opts.spargs.update(people_search_options)
def run(self, args, opts):
# Run people_search spider
args = ['people_search']
super().run(args, opts)
| <commit_before>from getpass import getpass
from scrapy.commands.crawl import Command as BaseCommand
def sanitize_query(query):
return query.replace(' ', '+')
class Command(BaseCommand):
def short_desc(self):
return "Scrap people from LinkedIn"
def syntax(self):
return "[options] <query>"
def add_options(self, parser):
super().add_options(parser)
parser.add_option('-u', '--username', help='Name of LinkedIn account')
parser.add_option('-p', '--password',
help='Password for LinkedIn account')
def process_options(self, args, opts):
opts.output = opts.output or 'results.csv'
opts.loglevel = opts.loglevel or 'INFO'
super().process_options(args, opts)
people_search_options = {
'query': sanitize_query(args[0]),
'username': opts.username or input(
'Please provide your LinkedIn username: '),
'password': opts.password or getpass(
'Please provide password for your LinkedIn account: ')
}
opts.spargs.update(people_search_options)
def run(self, args, opts):
# Run people_search spider
args = ['people_search']
super().run(args, opts)
<commit_msg>Fix IndexError when running command without arguments.<commit_after> | from getpass import getpass
from scrapy.commands.crawl import Command as BaseCommand
from scrapy.exceptions import UsageError
def sanitize_query(query):
return query.replace(' ', '+')
class Command(BaseCommand):
def short_desc(self):
return "Scrap people from LinkedIn"
def syntax(self):
return "[options] <query>"
def add_options(self, parser):
super().add_options(parser)
parser.add_option('-u', '--username', help='Name of LinkedIn account')
parser.add_option('-p', '--password',
help='Password for LinkedIn account')
def process_options(self, args, opts):
opts.output = opts.output or 'results.csv'
opts.loglevel = opts.loglevel or 'INFO'
super().process_options(args, opts)
try:
query = sanitize_query(args[0])
except IndexError:
raise UsageError()
people_search_options = {
'query': query,
'username': opts.username or input(
'Please provide your LinkedIn username: '),
'password': opts.password or getpass(
'Please provide password for your LinkedIn account: ')
}
opts.spargs.update(people_search_options)
def run(self, args, opts):
# Run people_search spider
args = ['people_search']
super().run(args, opts)
| from getpass import getpass
from scrapy.commands.crawl import Command as BaseCommand
def sanitize_query(query):
return query.replace(' ', '+')
class Command(BaseCommand):
def short_desc(self):
return "Scrap people from LinkedIn"
def syntax(self):
return "[options] <query>"
def add_options(self, parser):
super().add_options(parser)
parser.add_option('-u', '--username', help='Name of LinkedIn account')
parser.add_option('-p', '--password',
help='Password for LinkedIn account')
def process_options(self, args, opts):
opts.output = opts.output or 'results.csv'
opts.loglevel = opts.loglevel or 'INFO'
super().process_options(args, opts)
people_search_options = {
'query': sanitize_query(args[0]),
'username': opts.username or input(
'Please provide your LinkedIn username: '),
'password': opts.password or getpass(
'Please provide password for your LinkedIn account: ')
}
opts.spargs.update(people_search_options)
def run(self, args, opts):
# Run people_search spider
args = ['people_search']
super().run(args, opts)
Fix IndexError when running command without arguments.from getpass import getpass
from scrapy.commands.crawl import Command as BaseCommand
from scrapy.exceptions import UsageError
def sanitize_query(query):
return query.replace(' ', '+')
class Command(BaseCommand):
def short_desc(self):
return "Scrap people from LinkedIn"
def syntax(self):
return "[options] <query>"
def add_options(self, parser):
super().add_options(parser)
parser.add_option('-u', '--username', help='Name of LinkedIn account')
parser.add_option('-p', '--password',
help='Password for LinkedIn account')
def process_options(self, args, opts):
opts.output = opts.output or 'results.csv'
opts.loglevel = opts.loglevel or 'INFO'
super().process_options(args, opts)
try:
query = sanitize_query(args[0])
except IndexError:
raise UsageError()
people_search_options = {
'query': query,
'username': opts.username or input(
'Please provide your LinkedIn username: '),
'password': opts.password or getpass(
'Please provide password for your LinkedIn account: ')
}
opts.spargs.update(people_search_options)
def run(self, args, opts):
# Run people_search spider
args = ['people_search']
super().run(args, opts)
| <commit_before>from getpass import getpass
from scrapy.commands.crawl import Command as BaseCommand
def sanitize_query(query):
return query.replace(' ', '+')
class Command(BaseCommand):
def short_desc(self):
return "Scrap people from LinkedIn"
def syntax(self):
return "[options] <query>"
def add_options(self, parser):
super().add_options(parser)
parser.add_option('-u', '--username', help='Name of LinkedIn account')
parser.add_option('-p', '--password',
help='Password for LinkedIn account')
def process_options(self, args, opts):
opts.output = opts.output or 'results.csv'
opts.loglevel = opts.loglevel or 'INFO'
super().process_options(args, opts)
people_search_options = {
'query': sanitize_query(args[0]),
'username': opts.username or input(
'Please provide your LinkedIn username: '),
'password': opts.password or getpass(
'Please provide password for your LinkedIn account: ')
}
opts.spargs.update(people_search_options)
def run(self, args, opts):
# Run people_search spider
args = ['people_search']
super().run(args, opts)
<commit_msg>Fix IndexError when running command without arguments.<commit_after>from getpass import getpass
from scrapy.commands.crawl import Command as BaseCommand
from scrapy.exceptions import UsageError
def sanitize_query(query):
return query.replace(' ', '+')
class Command(BaseCommand):
def short_desc(self):
return "Scrap people from LinkedIn"
def syntax(self):
return "[options] <query>"
def add_options(self, parser):
super().add_options(parser)
parser.add_option('-u', '--username', help='Name of LinkedIn account')
parser.add_option('-p', '--password',
help='Password for LinkedIn account')
def process_options(self, args, opts):
opts.output = opts.output or 'results.csv'
opts.loglevel = opts.loglevel or 'INFO'
super().process_options(args, opts)
try:
query = sanitize_query(args[0])
except IndexError:
raise UsageError()
people_search_options = {
'query': query,
'username': opts.username or input(
'Please provide your LinkedIn username: '),
'password': opts.password or getpass(
'Please provide password for your LinkedIn account: ')
}
opts.spargs.update(people_search_options)
def run(self, args, opts):
# Run people_search spider
args = ['people_search']
super().run(args, opts)
|
f9abd5434dded655591029ae45859e8608b4e5d6 | django_facebook/decorators.py | django_facebook/decorators.py | from functools import update_wrapper, wraps
from django.contrib.auth import REDIRECT_FIELD_NAME
from django.http import HttpResponse, HttpResponseRedirect
from django.utils.decorators import available_attrs
from django.utils.http import urlquote
def facebook_required(function=None, redirect_field_name=REDIRECT_FIELD_NAME):
"""
Decorator for views that checks that the user is logged in, redirecting
to the log-in page if necessary.
"""
def _passes_test(test_func, login_url=None, redirect_field_name=REDIRECT_FIELD_NAME):
if not login_url:
from django.conf import settings
login_url = settings.LOGIN_URL
def decorator(view_func):
def _wrapped_view(request, *args, **kwargs):
if test_func(request):
return view_func(request, *args, **kwargs)
path = urlquote(request.get_full_path())
tup = login_url, redirect_field_name, path
return HttpResponseRedirect('%s?%s=%s' % tup)
return wraps(view_func, assigned=available_attrs(view_func))(_wrapped_view)
return decorator
actual_decorator = _passes_test(
lambda r: r.facebook.uid,
redirect_field_name=redirect_field_name
)
if function:
return actual_decorator(function)
return actual_decorator
| from functools import update_wrapper, wraps
from django.contrib.auth import REDIRECT_FIELD_NAME
from django.http import HttpResponse, HttpResponseRedirect
from django.utils.decorators import available_attrs
from django.utils.http import urlquote
def facebook_required(function=None, redirect_field_name=REDIRECT_FIELD_NAME):
"""
Decorator for views that checks that the user is logged in, redirecting
to the log-in page if necessary.
"""
def _passes_test(test_func, login_url=None, redirect_field_name=REDIRECT_FIELD_NAME):
if not login_url:
from django.conf import settings
login_url = settings.LOGIN_URL
def decorator(view_func):
def _wrapped_view(request, *args, **kwargs):
if test_func(request):
return view_func(request, *args, **kwargs)
path = urlquote(request.get_full_path())
tup = login_url, redirect_field_name, path
return HttpResponseRedirect('%s?%s=%s' % tup)
return wraps(view_func, assigned=available_attrs(view_func))(_wrapped_view)
return decorator
actual_decorator = _passes_test(
lambda r: r.facebook,
redirect_field_name=redirect_field_name
)
if function:
return actual_decorator(function)
return actual_decorator
| Fix possible problem with middleware | Fix possible problem with middleware
| Python | mit | tino/django-facebook2,srijanmishra/django-facebook,aidanlister/django-facebook,tino/django-facebook2,vstoykov/django4facebook | from functools import update_wrapper, wraps
from django.contrib.auth import REDIRECT_FIELD_NAME
from django.http import HttpResponse, HttpResponseRedirect
from django.utils.decorators import available_attrs
from django.utils.http import urlquote
def facebook_required(function=None, redirect_field_name=REDIRECT_FIELD_NAME):
"""
Decorator for views that checks that the user is logged in, redirecting
to the log-in page if necessary.
"""
def _passes_test(test_func, login_url=None, redirect_field_name=REDIRECT_FIELD_NAME):
if not login_url:
from django.conf import settings
login_url = settings.LOGIN_URL
def decorator(view_func):
def _wrapped_view(request, *args, **kwargs):
if test_func(request):
return view_func(request, *args, **kwargs)
path = urlquote(request.get_full_path())
tup = login_url, redirect_field_name, path
return HttpResponseRedirect('%s?%s=%s' % tup)
return wraps(view_func, assigned=available_attrs(view_func))(_wrapped_view)
return decorator
actual_decorator = _passes_test(
lambda r: r.facebook.uid,
redirect_field_name=redirect_field_name
)
if function:
return actual_decorator(function)
return actual_decorator
Fix possible problem with middleware | from functools import update_wrapper, wraps
from django.contrib.auth import REDIRECT_FIELD_NAME
from django.http import HttpResponse, HttpResponseRedirect
from django.utils.decorators import available_attrs
from django.utils.http import urlquote
def facebook_required(function=None, redirect_field_name=REDIRECT_FIELD_NAME):
"""
Decorator for views that checks that the user is logged in, redirecting
to the log-in page if necessary.
"""
def _passes_test(test_func, login_url=None, redirect_field_name=REDIRECT_FIELD_NAME):
if not login_url:
from django.conf import settings
login_url = settings.LOGIN_URL
def decorator(view_func):
def _wrapped_view(request, *args, **kwargs):
if test_func(request):
return view_func(request, *args, **kwargs)
path = urlquote(request.get_full_path())
tup = login_url, redirect_field_name, path
return HttpResponseRedirect('%s?%s=%s' % tup)
return wraps(view_func, assigned=available_attrs(view_func))(_wrapped_view)
return decorator
actual_decorator = _passes_test(
lambda r: r.facebook,
redirect_field_name=redirect_field_name
)
if function:
return actual_decorator(function)
return actual_decorator
| <commit_before>from functools import update_wrapper, wraps
from django.contrib.auth import REDIRECT_FIELD_NAME
from django.http import HttpResponse, HttpResponseRedirect
from django.utils.decorators import available_attrs
from django.utils.http import urlquote
def facebook_required(function=None, redirect_field_name=REDIRECT_FIELD_NAME):
"""
Decorator for views that checks that the user is logged in, redirecting
to the log-in page if necessary.
"""
def _passes_test(test_func, login_url=None, redirect_field_name=REDIRECT_FIELD_NAME):
if not login_url:
from django.conf import settings
login_url = settings.LOGIN_URL
def decorator(view_func):
def _wrapped_view(request, *args, **kwargs):
if test_func(request):
return view_func(request, *args, **kwargs)
path = urlquote(request.get_full_path())
tup = login_url, redirect_field_name, path
return HttpResponseRedirect('%s?%s=%s' % tup)
return wraps(view_func, assigned=available_attrs(view_func))(_wrapped_view)
return decorator
actual_decorator = _passes_test(
lambda r: r.facebook.uid,
redirect_field_name=redirect_field_name
)
if function:
return actual_decorator(function)
return actual_decorator
<commit_msg>Fix possible problem with middleware<commit_after> | from functools import update_wrapper, wraps
from django.contrib.auth import REDIRECT_FIELD_NAME
from django.http import HttpResponse, HttpResponseRedirect
from django.utils.decorators import available_attrs
from django.utils.http import urlquote
def facebook_required(function=None, redirect_field_name=REDIRECT_FIELD_NAME):
"""
Decorator for views that checks that the user is logged in, redirecting
to the log-in page if necessary.
"""
def _passes_test(test_func, login_url=None, redirect_field_name=REDIRECT_FIELD_NAME):
if not login_url:
from django.conf import settings
login_url = settings.LOGIN_URL
def decorator(view_func):
def _wrapped_view(request, *args, **kwargs):
if test_func(request):
return view_func(request, *args, **kwargs)
path = urlquote(request.get_full_path())
tup = login_url, redirect_field_name, path
return HttpResponseRedirect('%s?%s=%s' % tup)
return wraps(view_func, assigned=available_attrs(view_func))(_wrapped_view)
return decorator
actual_decorator = _passes_test(
lambda r: r.facebook,
redirect_field_name=redirect_field_name
)
if function:
return actual_decorator(function)
return actual_decorator
| from functools import update_wrapper, wraps
from django.contrib.auth import REDIRECT_FIELD_NAME
from django.http import HttpResponse, HttpResponseRedirect
from django.utils.decorators import available_attrs
from django.utils.http import urlquote
def facebook_required(function=None, redirect_field_name=REDIRECT_FIELD_NAME):
"""
Decorator for views that checks that the user is logged in, redirecting
to the log-in page if necessary.
"""
def _passes_test(test_func, login_url=None, redirect_field_name=REDIRECT_FIELD_NAME):
if not login_url:
from django.conf import settings
login_url = settings.LOGIN_URL
def decorator(view_func):
def _wrapped_view(request, *args, **kwargs):
if test_func(request):
return view_func(request, *args, **kwargs)
path = urlquote(request.get_full_path())
tup = login_url, redirect_field_name, path
return HttpResponseRedirect('%s?%s=%s' % tup)
return wraps(view_func, assigned=available_attrs(view_func))(_wrapped_view)
return decorator
actual_decorator = _passes_test(
lambda r: r.facebook.uid,
redirect_field_name=redirect_field_name
)
if function:
return actual_decorator(function)
return actual_decorator
Fix possible problem with middlewarefrom functools import update_wrapper, wraps
from django.contrib.auth import REDIRECT_FIELD_NAME
from django.http import HttpResponse, HttpResponseRedirect
from django.utils.decorators import available_attrs
from django.utils.http import urlquote
def facebook_required(function=None, redirect_field_name=REDIRECT_FIELD_NAME):
"""
Decorator for views that checks that the user is logged in, redirecting
to the log-in page if necessary.
"""
def _passes_test(test_func, login_url=None, redirect_field_name=REDIRECT_FIELD_NAME):
if not login_url:
from django.conf import settings
login_url = settings.LOGIN_URL
def decorator(view_func):
def _wrapped_view(request, *args, **kwargs):
if test_func(request):
return view_func(request, *args, **kwargs)
path = urlquote(request.get_full_path())
tup = login_url, redirect_field_name, path
return HttpResponseRedirect('%s?%s=%s' % tup)
return wraps(view_func, assigned=available_attrs(view_func))(_wrapped_view)
return decorator
actual_decorator = _passes_test(
lambda r: r.facebook,
redirect_field_name=redirect_field_name
)
if function:
return actual_decorator(function)
return actual_decorator
| <commit_before>from functools import update_wrapper, wraps
from django.contrib.auth import REDIRECT_FIELD_NAME
from django.http import HttpResponse, HttpResponseRedirect
from django.utils.decorators import available_attrs
from django.utils.http import urlquote
def facebook_required(function=None, redirect_field_name=REDIRECT_FIELD_NAME):
"""
Decorator for views that checks that the user is logged in, redirecting
to the log-in page if necessary.
"""
def _passes_test(test_func, login_url=None, redirect_field_name=REDIRECT_FIELD_NAME):
if not login_url:
from django.conf import settings
login_url = settings.LOGIN_URL
def decorator(view_func):
def _wrapped_view(request, *args, **kwargs):
if test_func(request):
return view_func(request, *args, **kwargs)
path = urlquote(request.get_full_path())
tup = login_url, redirect_field_name, path
return HttpResponseRedirect('%s?%s=%s' % tup)
return wraps(view_func, assigned=available_attrs(view_func))(_wrapped_view)
return decorator
actual_decorator = _passes_test(
lambda r: r.facebook.uid,
redirect_field_name=redirect_field_name
)
if function:
return actual_decorator(function)
return actual_decorator
<commit_msg>Fix possible problem with middleware<commit_after>from functools import update_wrapper, wraps
from django.contrib.auth import REDIRECT_FIELD_NAME
from django.http import HttpResponse, HttpResponseRedirect
from django.utils.decorators import available_attrs
from django.utils.http import urlquote
def facebook_required(function=None, redirect_field_name=REDIRECT_FIELD_NAME):
"""
Decorator for views that checks that the user is logged in, redirecting
to the log-in page if necessary.
"""
def _passes_test(test_func, login_url=None, redirect_field_name=REDIRECT_FIELD_NAME):
if not login_url:
from django.conf import settings
login_url = settings.LOGIN_URL
def decorator(view_func):
def _wrapped_view(request, *args, **kwargs):
if test_func(request):
return view_func(request, *args, **kwargs)
path = urlquote(request.get_full_path())
tup = login_url, redirect_field_name, path
return HttpResponseRedirect('%s?%s=%s' % tup)
return wraps(view_func, assigned=available_attrs(view_func))(_wrapped_view)
return decorator
actual_decorator = _passes_test(
lambda r: r.facebook,
redirect_field_name=redirect_field_name
)
if function:
return actual_decorator(function)
return actual_decorator
|
5dc4a262771e616458feeaa9bf4ca8568736761a | docs/contributors/generate.py | docs/contributors/generate.py | """
Generate snippets to copy-paste.
"""
import sys
from jinja2 import Template
from fetch import HERE, load_awesome_people
TPL_FILE = HERE / 'snippet.jinja2'
HTTPIE_TEAM = {
'BoboTiG',
'claudiatd',
'jakubroztocil',
'jkbr',
}
def generate_snippets(release: str) -> str:
people = load_awesome_people()
contributors = {
name: details
for name, details in people.items()
if details['github'] not in HTTPIE_TEAM
and (release in details['committed'] or release in details['reported'])
}
template = Template(source=TPL_FILE.read_text(encoding='utf-8'))
output = template.render(contributors=contributors, release=release)
print(output)
return 0
if __name__ == '__main__':
ret = 1
try:
ret = generate_snippets(sys.argv[1])
except (IndexError, TypeError):
ret = 2
print(f'''
Generate snippets for contributors to a release.
Usage:
python {sys.argv[0]} {sys.argv[0]} <RELEASE>
''')
sys.exit(ret)
| """
Generate snippets to copy-paste.
"""
import sys
from jinja2 import Template
from fetch import HERE, load_awesome_people
TPL_FILE = HERE / 'snippet.jinja2'
HTTPIE_TEAM = {
'claudiatd',
'jakubroztocil',
'jkbr',
}
def generate_snippets(release: str) -> str:
people = load_awesome_people()
contributors = {
name: details
for name, details in people.items()
if details['github'] not in HTTPIE_TEAM
and (release in details['committed'] or release in details['reported'])
}
template = Template(source=TPL_FILE.read_text(encoding='utf-8'))
output = template.render(contributors=contributors, release=release)
print(output)
return 0
if __name__ == '__main__':
ret = 1
try:
ret = generate_snippets(sys.argv[1])
except (IndexError, TypeError):
ret = 2
print(f'''
Generate snippets for contributors to a release.
Usage:
python {sys.argv[0]} {sys.argv[0]} <RELEASE>
''')
sys.exit(ret)
| Remove myself from the HTTPie team | Remove myself from the HTTPie team | Python | bsd-3-clause | jakubroztocil/httpie,PKRoma/httpie,jakubroztocil/httpie,jakubroztocil/httpie,PKRoma/httpie | """
Generate snippets to copy-paste.
"""
import sys
from jinja2 import Template
from fetch import HERE, load_awesome_people
TPL_FILE = HERE / 'snippet.jinja2'
HTTPIE_TEAM = {
'BoboTiG',
'claudiatd',
'jakubroztocil',
'jkbr',
}
def generate_snippets(release: str) -> str:
people = load_awesome_people()
contributors = {
name: details
for name, details in people.items()
if details['github'] not in HTTPIE_TEAM
and (release in details['committed'] or release in details['reported'])
}
template = Template(source=TPL_FILE.read_text(encoding='utf-8'))
output = template.render(contributors=contributors, release=release)
print(output)
return 0
if __name__ == '__main__':
ret = 1
try:
ret = generate_snippets(sys.argv[1])
except (IndexError, TypeError):
ret = 2
print(f'''
Generate snippets for contributors to a release.
Usage:
python {sys.argv[0]} {sys.argv[0]} <RELEASE>
''')
sys.exit(ret)
Remove myself from the HTTPie team | """
Generate snippets to copy-paste.
"""
import sys
from jinja2 import Template
from fetch import HERE, load_awesome_people
TPL_FILE = HERE / 'snippet.jinja2'
HTTPIE_TEAM = {
'claudiatd',
'jakubroztocil',
'jkbr',
}
def generate_snippets(release: str) -> str:
people = load_awesome_people()
contributors = {
name: details
for name, details in people.items()
if details['github'] not in HTTPIE_TEAM
and (release in details['committed'] or release in details['reported'])
}
template = Template(source=TPL_FILE.read_text(encoding='utf-8'))
output = template.render(contributors=contributors, release=release)
print(output)
return 0
if __name__ == '__main__':
ret = 1
try:
ret = generate_snippets(sys.argv[1])
except (IndexError, TypeError):
ret = 2
print(f'''
Generate snippets for contributors to a release.
Usage:
python {sys.argv[0]} {sys.argv[0]} <RELEASE>
''')
sys.exit(ret)
| <commit_before>"""
Generate snippets to copy-paste.
"""
import sys
from jinja2 import Template
from fetch import HERE, load_awesome_people
TPL_FILE = HERE / 'snippet.jinja2'
HTTPIE_TEAM = {
'BoboTiG',
'claudiatd',
'jakubroztocil',
'jkbr',
}
def generate_snippets(release: str) -> str:
people = load_awesome_people()
contributors = {
name: details
for name, details in people.items()
if details['github'] not in HTTPIE_TEAM
and (release in details['committed'] or release in details['reported'])
}
template = Template(source=TPL_FILE.read_text(encoding='utf-8'))
output = template.render(contributors=contributors, release=release)
print(output)
return 0
if __name__ == '__main__':
ret = 1
try:
ret = generate_snippets(sys.argv[1])
except (IndexError, TypeError):
ret = 2
print(f'''
Generate snippets for contributors to a release.
Usage:
python {sys.argv[0]} {sys.argv[0]} <RELEASE>
''')
sys.exit(ret)
<commit_msg>Remove myself from the HTTPie team<commit_after> | """
Generate snippets to copy-paste.
"""
import sys
from jinja2 import Template
from fetch import HERE, load_awesome_people
TPL_FILE = HERE / 'snippet.jinja2'
HTTPIE_TEAM = {
'claudiatd',
'jakubroztocil',
'jkbr',
}
def generate_snippets(release: str) -> str:
people = load_awesome_people()
contributors = {
name: details
for name, details in people.items()
if details['github'] not in HTTPIE_TEAM
and (release in details['committed'] or release in details['reported'])
}
template = Template(source=TPL_FILE.read_text(encoding='utf-8'))
output = template.render(contributors=contributors, release=release)
print(output)
return 0
if __name__ == '__main__':
ret = 1
try:
ret = generate_snippets(sys.argv[1])
except (IndexError, TypeError):
ret = 2
print(f'''
Generate snippets for contributors to a release.
Usage:
python {sys.argv[0]} {sys.argv[0]} <RELEASE>
''')
sys.exit(ret)
| """
Generate snippets to copy-paste.
"""
import sys
from jinja2 import Template
from fetch import HERE, load_awesome_people
TPL_FILE = HERE / 'snippet.jinja2'
HTTPIE_TEAM = {
'BoboTiG',
'claudiatd',
'jakubroztocil',
'jkbr',
}
def generate_snippets(release: str) -> str:
people = load_awesome_people()
contributors = {
name: details
for name, details in people.items()
if details['github'] not in HTTPIE_TEAM
and (release in details['committed'] or release in details['reported'])
}
template = Template(source=TPL_FILE.read_text(encoding='utf-8'))
output = template.render(contributors=contributors, release=release)
print(output)
return 0
if __name__ == '__main__':
ret = 1
try:
ret = generate_snippets(sys.argv[1])
except (IndexError, TypeError):
ret = 2
print(f'''
Generate snippets for contributors to a release.
Usage:
python {sys.argv[0]} {sys.argv[0]} <RELEASE>
''')
sys.exit(ret)
Remove myself from the HTTPie team"""
Generate snippets to copy-paste.
"""
import sys
from jinja2 import Template
from fetch import HERE, load_awesome_people
TPL_FILE = HERE / 'snippet.jinja2'
HTTPIE_TEAM = {
'claudiatd',
'jakubroztocil',
'jkbr',
}
def generate_snippets(release: str) -> str:
people = load_awesome_people()
contributors = {
name: details
for name, details in people.items()
if details['github'] not in HTTPIE_TEAM
and (release in details['committed'] or release in details['reported'])
}
template = Template(source=TPL_FILE.read_text(encoding='utf-8'))
output = template.render(contributors=contributors, release=release)
print(output)
return 0
if __name__ == '__main__':
ret = 1
try:
ret = generate_snippets(sys.argv[1])
except (IndexError, TypeError):
ret = 2
print(f'''
Generate snippets for contributors to a release.
Usage:
python {sys.argv[0]} {sys.argv[0]} <RELEASE>
''')
sys.exit(ret)
| <commit_before>"""
Generate snippets to copy-paste.
"""
import sys
from jinja2 import Template
from fetch import HERE, load_awesome_people
TPL_FILE = HERE / 'snippet.jinja2'
HTTPIE_TEAM = {
'BoboTiG',
'claudiatd',
'jakubroztocil',
'jkbr',
}
def generate_snippets(release: str) -> str:
people = load_awesome_people()
contributors = {
name: details
for name, details in people.items()
if details['github'] not in HTTPIE_TEAM
and (release in details['committed'] or release in details['reported'])
}
template = Template(source=TPL_FILE.read_text(encoding='utf-8'))
output = template.render(contributors=contributors, release=release)
print(output)
return 0
if __name__ == '__main__':
ret = 1
try:
ret = generate_snippets(sys.argv[1])
except (IndexError, TypeError):
ret = 2
print(f'''
Generate snippets for contributors to a release.
Usage:
python {sys.argv[0]} {sys.argv[0]} <RELEASE>
''')
sys.exit(ret)
<commit_msg>Remove myself from the HTTPie team<commit_after>"""
Generate snippets to copy-paste.
"""
import sys
from jinja2 import Template
from fetch import HERE, load_awesome_people
TPL_FILE = HERE / 'snippet.jinja2'
HTTPIE_TEAM = {
'claudiatd',
'jakubroztocil',
'jkbr',
}
def generate_snippets(release: str) -> str:
people = load_awesome_people()
contributors = {
name: details
for name, details in people.items()
if details['github'] not in HTTPIE_TEAM
and (release in details['committed'] or release in details['reported'])
}
template = Template(source=TPL_FILE.read_text(encoding='utf-8'))
output = template.render(contributors=contributors, release=release)
print(output)
return 0
if __name__ == '__main__':
ret = 1
try:
ret = generate_snippets(sys.argv[1])
except (IndexError, TypeError):
ret = 2
print(f'''
Generate snippets for contributors to a release.
Usage:
python {sys.argv[0]} {sys.argv[0]} <RELEASE>
''')
sys.exit(ret)
|
f16d93216e1f0890b0551ca3b741130bb12781ef | gold_digger/settings/__init__.py | gold_digger/settings/__init__.py | # -*- coding: utf-8 -*-
from os import environ, path
from ._settings_default import *
from ..exceptions import ImproperlyConfigured
profile = environ.get("GOLD_DIGGER_PROFILE", "local")
if profile == "master":
from ._settings_master import *
elif profile == "local":
try:
from ._settings_local import *
except ImportError:
raise ImproperlyConfigured(
"Local configuration not found. Create file _settings_local.py in {} directory according to README.".format(
path.abspath(path.join(__file__, path.pardir))
)
)
else:
raise ValueError("Unsupported settings profile. Got: {}. Use one of: master, staging, local.".format(profile))
| # -*- coding: utf-8 -*-
from os import environ, path
from ._settings_default import *
from ..exceptions import ImproperlyConfigured
PROFILE = environ.get("GOLD_DIGGER_PROFILE", "local")
if PROFILE == "master":
from ._settings_master import *
elif PROFILE == "local":
try:
from ._settings_local import *
except ImportError:
raise ImproperlyConfigured(
f"Local configuration not found. Create file _settings_local.py in {path.abspath(path.join(__file__, path.pardir))} directory according to README."
)
else:
raise ValueError(f"Unsupported settings profile. Got: {PROFILE}. Use one of: master, staging, local.")
| Make global variable upper-case and use f-strings | Make global variable upper-case and use f-strings
| Python | apache-2.0 | business-factory/gold-digger | # -*- coding: utf-8 -*-
from os import environ, path
from ._settings_default import *
from ..exceptions import ImproperlyConfigured
profile = environ.get("GOLD_DIGGER_PROFILE", "local")
if profile == "master":
from ._settings_master import *
elif profile == "local":
try:
from ._settings_local import *
except ImportError:
raise ImproperlyConfigured(
"Local configuration not found. Create file _settings_local.py in {} directory according to README.".format(
path.abspath(path.join(__file__, path.pardir))
)
)
else:
raise ValueError("Unsupported settings profile. Got: {}. Use one of: master, staging, local.".format(profile))
Make global variable upper-case and use f-strings | # -*- coding: utf-8 -*-
from os import environ, path
from ._settings_default import *
from ..exceptions import ImproperlyConfigured
PROFILE = environ.get("GOLD_DIGGER_PROFILE", "local")
if PROFILE == "master":
from ._settings_master import *
elif PROFILE == "local":
try:
from ._settings_local import *
except ImportError:
raise ImproperlyConfigured(
f"Local configuration not found. Create file _settings_local.py in {path.abspath(path.join(__file__, path.pardir))} directory according to README."
)
else:
raise ValueError(f"Unsupported settings profile. Got: {PROFILE}. Use one of: master, staging, local.")
| <commit_before># -*- coding: utf-8 -*-
from os import environ, path
from ._settings_default import *
from ..exceptions import ImproperlyConfigured
profile = environ.get("GOLD_DIGGER_PROFILE", "local")
if profile == "master":
from ._settings_master import *
elif profile == "local":
try:
from ._settings_local import *
except ImportError:
raise ImproperlyConfigured(
"Local configuration not found. Create file _settings_local.py in {} directory according to README.".format(
path.abspath(path.join(__file__, path.pardir))
)
)
else:
raise ValueError("Unsupported settings profile. Got: {}. Use one of: master, staging, local.".format(profile))
<commit_msg>Make global variable upper-case and use f-strings<commit_after> | # -*- coding: utf-8 -*-
from os import environ, path
from ._settings_default import *
from ..exceptions import ImproperlyConfigured
PROFILE = environ.get("GOLD_DIGGER_PROFILE", "local")
if PROFILE == "master":
from ._settings_master import *
elif PROFILE == "local":
try:
from ._settings_local import *
except ImportError:
raise ImproperlyConfigured(
f"Local configuration not found. Create file _settings_local.py in {path.abspath(path.join(__file__, path.pardir))} directory according to README."
)
else:
raise ValueError(f"Unsupported settings profile. Got: {PROFILE}. Use one of: master, staging, local.")
| # -*- coding: utf-8 -*-
from os import environ, path
from ._settings_default import *
from ..exceptions import ImproperlyConfigured
profile = environ.get("GOLD_DIGGER_PROFILE", "local")
if profile == "master":
from ._settings_master import *
elif profile == "local":
try:
from ._settings_local import *
except ImportError:
raise ImproperlyConfigured(
"Local configuration not found. Create file _settings_local.py in {} directory according to README.".format(
path.abspath(path.join(__file__, path.pardir))
)
)
else:
raise ValueError("Unsupported settings profile. Got: {}. Use one of: master, staging, local.".format(profile))
Make global variable upper-case and use f-strings# -*- coding: utf-8 -*-
from os import environ, path
from ._settings_default import *
from ..exceptions import ImproperlyConfigured
PROFILE = environ.get("GOLD_DIGGER_PROFILE", "local")
if PROFILE == "master":
from ._settings_master import *
elif PROFILE == "local":
try:
from ._settings_local import *
except ImportError:
raise ImproperlyConfigured(
f"Local configuration not found. Create file _settings_local.py in {path.abspath(path.join(__file__, path.pardir))} directory according to README."
)
else:
raise ValueError(f"Unsupported settings profile. Got: {PROFILE}. Use one of: master, staging, local.")
| <commit_before># -*- coding: utf-8 -*-
from os import environ, path
from ._settings_default import *
from ..exceptions import ImproperlyConfigured
profile = environ.get("GOLD_DIGGER_PROFILE", "local")
if profile == "master":
from ._settings_master import *
elif profile == "local":
try:
from ._settings_local import *
except ImportError:
raise ImproperlyConfigured(
"Local configuration not found. Create file _settings_local.py in {} directory according to README.".format(
path.abspath(path.join(__file__, path.pardir))
)
)
else:
raise ValueError("Unsupported settings profile. Got: {}. Use one of: master, staging, local.".format(profile))
<commit_msg>Make global variable upper-case and use f-strings<commit_after># -*- coding: utf-8 -*-
from os import environ, path
from ._settings_default import *
from ..exceptions import ImproperlyConfigured
PROFILE = environ.get("GOLD_DIGGER_PROFILE", "local")
if PROFILE == "master":
from ._settings_master import *
elif PROFILE == "local":
try:
from ._settings_local import *
except ImportError:
raise ImproperlyConfigured(
f"Local configuration not found. Create file _settings_local.py in {path.abspath(path.join(__file__, path.pardir))} directory according to README."
)
else:
raise ValueError(f"Unsupported settings profile. Got: {PROFILE}. Use one of: master, staging, local.")
|
763051db8cb4efe9b3e85fdb3d974674cf435607 | us_ignite/snippets/management/commands/snippets_load_fixtures.py | us_ignite/snippets/management/commands/snippets_load_fixtures.py | from django.core.management.base import BaseCommand
from us_ignite.snippets.models import Snippet
FIXTURES = [
{
'slug': 'home-box',
'name': 'UP NEXT: LOREM IPSUM',
'body': '',
'url_text': 'GET INVOLVED',
'url': '',
},
{
'slug': 'featured',
'name': 'FEATURED CONTENT',
'body': '',
'url_text': 'FEATURED',
'url': '',
},
{
'slug': 'welcome-email',
'name': 'Welcome to US Ignite',
'body': '',
'url_text': '',
'url': '',
},
]
class Command(BaseCommand):
def handle(self, *args, **options):
for data in FIXTURES:
try:
# Ignore existing snippets:
Snippet.objects.get(slug=data['slug'])
continue
except Snippet.DoesNotExist:
pass
data.update({
'status': Snippet.PUBLISHED,
})
Snippet.objects.create(**data)
print u'Importing %s' % data['slug']
print "Done!"
| from django.core.management.base import BaseCommand
from us_ignite.snippets.models import Snippet
FIXTURES = [
{
'slug': 'home-box',
'name': 'UP NEXT: LOREM IPSUM',
'body': '',
'url_text': 'GET INVOLVED',
'url': '',
},
{
'slug': 'featured',
'name': 'FEATURED CONTENT',
'body': '',
'url_text': 'FEATURED',
'url': '',
},
{
'slug': 'welcome-email',
'name': 'Welcome to US Ignite',
'body': '',
'url_text': '',
'url': '',
},
{
'slug': 'blog-sidebar',
'name': 'Dynamic content',
'body': '',
'url_text': '',
'url': '',
},
]
class Command(BaseCommand):
def handle(self, *args, **options):
for data in FIXTURES:
try:
# Ignore existing snippets:
Snippet.objects.get(slug=data['slug'])
continue
except Snippet.DoesNotExist:
pass
data.update({
'status': Snippet.PUBLISHED,
})
Snippet.objects.create(**data)
print u'Importing %s' % data['slug']
print "Done!"
| Add fixture for the advert in the blog sidebar advert. | Add fixture for the advert in the blog sidebar advert.
| Python | bsd-3-clause | us-ignite/us_ignite,us-ignite/us_ignite,us-ignite/us_ignite,us-ignite/us_ignite,us-ignite/us_ignite | from django.core.management.base import BaseCommand
from us_ignite.snippets.models import Snippet
FIXTURES = [
{
'slug': 'home-box',
'name': 'UP NEXT: LOREM IPSUM',
'body': '',
'url_text': 'GET INVOLVED',
'url': '',
},
{
'slug': 'featured',
'name': 'FEATURED CONTENT',
'body': '',
'url_text': 'FEATURED',
'url': '',
},
{
'slug': 'welcome-email',
'name': 'Welcome to US Ignite',
'body': '',
'url_text': '',
'url': '',
},
]
class Command(BaseCommand):
def handle(self, *args, **options):
for data in FIXTURES:
try:
# Ignore existing snippets:
Snippet.objects.get(slug=data['slug'])
continue
except Snippet.DoesNotExist:
pass
data.update({
'status': Snippet.PUBLISHED,
})
Snippet.objects.create(**data)
print u'Importing %s' % data['slug']
print "Done!"
Add fixture for the advert in the blog sidebar advert. | from django.core.management.base import BaseCommand
from us_ignite.snippets.models import Snippet
FIXTURES = [
{
'slug': 'home-box',
'name': 'UP NEXT: LOREM IPSUM',
'body': '',
'url_text': 'GET INVOLVED',
'url': '',
},
{
'slug': 'featured',
'name': 'FEATURED CONTENT',
'body': '',
'url_text': 'FEATURED',
'url': '',
},
{
'slug': 'welcome-email',
'name': 'Welcome to US Ignite',
'body': '',
'url_text': '',
'url': '',
},
{
'slug': 'blog-sidebar',
'name': 'Dynamic content',
'body': '',
'url_text': '',
'url': '',
},
]
class Command(BaseCommand):
def handle(self, *args, **options):
for data in FIXTURES:
try:
# Ignore existing snippets:
Snippet.objects.get(slug=data['slug'])
continue
except Snippet.DoesNotExist:
pass
data.update({
'status': Snippet.PUBLISHED,
})
Snippet.objects.create(**data)
print u'Importing %s' % data['slug']
print "Done!"
| <commit_before>from django.core.management.base import BaseCommand
from us_ignite.snippets.models import Snippet
FIXTURES = [
{
'slug': 'home-box',
'name': 'UP NEXT: LOREM IPSUM',
'body': '',
'url_text': 'GET INVOLVED',
'url': '',
},
{
'slug': 'featured',
'name': 'FEATURED CONTENT',
'body': '',
'url_text': 'FEATURED',
'url': '',
},
{
'slug': 'welcome-email',
'name': 'Welcome to US Ignite',
'body': '',
'url_text': '',
'url': '',
},
]
class Command(BaseCommand):
def handle(self, *args, **options):
for data in FIXTURES:
try:
# Ignore existing snippets:
Snippet.objects.get(slug=data['slug'])
continue
except Snippet.DoesNotExist:
pass
data.update({
'status': Snippet.PUBLISHED,
})
Snippet.objects.create(**data)
print u'Importing %s' % data['slug']
print "Done!"
<commit_msg>Add fixture for the advert in the blog sidebar advert.<commit_after> | from django.core.management.base import BaseCommand
from us_ignite.snippets.models import Snippet
FIXTURES = [
{
'slug': 'home-box',
'name': 'UP NEXT: LOREM IPSUM',
'body': '',
'url_text': 'GET INVOLVED',
'url': '',
},
{
'slug': 'featured',
'name': 'FEATURED CONTENT',
'body': '',
'url_text': 'FEATURED',
'url': '',
},
{
'slug': 'welcome-email',
'name': 'Welcome to US Ignite',
'body': '',
'url_text': '',
'url': '',
},
{
'slug': 'blog-sidebar',
'name': 'Dynamic content',
'body': '',
'url_text': '',
'url': '',
},
]
class Command(BaseCommand):
def handle(self, *args, **options):
for data in FIXTURES:
try:
# Ignore existing snippets:
Snippet.objects.get(slug=data['slug'])
continue
except Snippet.DoesNotExist:
pass
data.update({
'status': Snippet.PUBLISHED,
})
Snippet.objects.create(**data)
print u'Importing %s' % data['slug']
print "Done!"
| from django.core.management.base import BaseCommand
from us_ignite.snippets.models import Snippet
FIXTURES = [
{
'slug': 'home-box',
'name': 'UP NEXT: LOREM IPSUM',
'body': '',
'url_text': 'GET INVOLVED',
'url': '',
},
{
'slug': 'featured',
'name': 'FEATURED CONTENT',
'body': '',
'url_text': 'FEATURED',
'url': '',
},
{
'slug': 'welcome-email',
'name': 'Welcome to US Ignite',
'body': '',
'url_text': '',
'url': '',
},
]
class Command(BaseCommand):
def handle(self, *args, **options):
for data in FIXTURES:
try:
# Ignore existing snippets:
Snippet.objects.get(slug=data['slug'])
continue
except Snippet.DoesNotExist:
pass
data.update({
'status': Snippet.PUBLISHED,
})
Snippet.objects.create(**data)
print u'Importing %s' % data['slug']
print "Done!"
Add fixture for the advert in the blog sidebar advert.from django.core.management.base import BaseCommand
from us_ignite.snippets.models import Snippet
FIXTURES = [
{
'slug': 'home-box',
'name': 'UP NEXT: LOREM IPSUM',
'body': '',
'url_text': 'GET INVOLVED',
'url': '',
},
{
'slug': 'featured',
'name': 'FEATURED CONTENT',
'body': '',
'url_text': 'FEATURED',
'url': '',
},
{
'slug': 'welcome-email',
'name': 'Welcome to US Ignite',
'body': '',
'url_text': '',
'url': '',
},
{
'slug': 'blog-sidebar',
'name': 'Dynamic content',
'body': '',
'url_text': '',
'url': '',
},
]
class Command(BaseCommand):
def handle(self, *args, **options):
for data in FIXTURES:
try:
# Ignore existing snippets:
Snippet.objects.get(slug=data['slug'])
continue
except Snippet.DoesNotExist:
pass
data.update({
'status': Snippet.PUBLISHED,
})
Snippet.objects.create(**data)
print u'Importing %s' % data['slug']
print "Done!"
| <commit_before>from django.core.management.base import BaseCommand
from us_ignite.snippets.models import Snippet
FIXTURES = [
{
'slug': 'home-box',
'name': 'UP NEXT: LOREM IPSUM',
'body': '',
'url_text': 'GET INVOLVED',
'url': '',
},
{
'slug': 'featured',
'name': 'FEATURED CONTENT',
'body': '',
'url_text': 'FEATURED',
'url': '',
},
{
'slug': 'welcome-email',
'name': 'Welcome to US Ignite',
'body': '',
'url_text': '',
'url': '',
},
]
class Command(BaseCommand):
def handle(self, *args, **options):
for data in FIXTURES:
try:
# Ignore existing snippets:
Snippet.objects.get(slug=data['slug'])
continue
except Snippet.DoesNotExist:
pass
data.update({
'status': Snippet.PUBLISHED,
})
Snippet.objects.create(**data)
print u'Importing %s' % data['slug']
print "Done!"
<commit_msg>Add fixture for the advert in the blog sidebar advert.<commit_after>from django.core.management.base import BaseCommand
from us_ignite.snippets.models import Snippet
FIXTURES = [
{
'slug': 'home-box',
'name': 'UP NEXT: LOREM IPSUM',
'body': '',
'url_text': 'GET INVOLVED',
'url': '',
},
{
'slug': 'featured',
'name': 'FEATURED CONTENT',
'body': '',
'url_text': 'FEATURED',
'url': '',
},
{
'slug': 'welcome-email',
'name': 'Welcome to US Ignite',
'body': '',
'url_text': '',
'url': '',
},
{
'slug': 'blog-sidebar',
'name': 'Dynamic content',
'body': '',
'url_text': '',
'url': '',
},
]
class Command(BaseCommand):
def handle(self, *args, **options):
for data in FIXTURES:
try:
# Ignore existing snippets:
Snippet.objects.get(slug=data['slug'])
continue
except Snippet.DoesNotExist:
pass
data.update({
'status': Snippet.PUBLISHED,
})
Snippet.objects.create(**data)
print u'Importing %s' % data['slug']
print "Done!"
|
49a371728a2e9167494264e0c07c6dd90abec0ff | saleor/core/views.py | saleor/core/views.py | from django.template.response import TemplateResponse
from ..product.utils import products_with_availability, products_with_details
def home(request):
products = Product.objects.get_available_products()[:6]
products = products.prefetch_related('categories', 'images',
'variants__stock')
return TemplateResponse(
request, 'home.html',
{'products': products, 'parent': None})
| from django.template.response import TemplateResponse
from ..product.utils import products_with_availability, products_with_details
def home(request):
products = products_with_details(request.user)[:6]
products = products_with_availability(
products, discounts=request.discounts, local_currency=request.currency)
return TemplateResponse(
request, 'home.html',
{'products': products, 'parent': None})
| Fix homepage after wrong rebase | Fix homepage after wrong rebase
| Python | bsd-3-clause | jreigel/saleor,car3oon/saleor,HyperManTT/ECommerceSaleor,tfroehlich82/saleor,HyperManTT/ECommerceSaleor,UITools/saleor,mociepka/saleor,car3oon/saleor,KenMutemi/saleor,jreigel/saleor,UITools/saleor,UITools/saleor,mociepka/saleor,maferelo/saleor,itbabu/saleor,UITools/saleor,KenMutemi/saleor,mociepka/saleor,itbabu/saleor,maferelo/saleor,tfroehlich82/saleor,HyperManTT/ECommerceSaleor,jreigel/saleor,itbabu/saleor,tfroehlich82/saleor,UITools/saleor,KenMutemi/saleor,car3oon/saleor,maferelo/saleor | from django.template.response import TemplateResponse
from ..product.utils import products_with_availability, products_with_details
def home(request):
products = Product.objects.get_available_products()[:6]
products = products.prefetch_related('categories', 'images',
'variants__stock')
return TemplateResponse(
request, 'home.html',
{'products': products, 'parent': None})
Fix homepage after wrong rebase | from django.template.response import TemplateResponse
from ..product.utils import products_with_availability, products_with_details
def home(request):
products = products_with_details(request.user)[:6]
products = products_with_availability(
products, discounts=request.discounts, local_currency=request.currency)
return TemplateResponse(
request, 'home.html',
{'products': products, 'parent': None})
| <commit_before>from django.template.response import TemplateResponse
from ..product.utils import products_with_availability, products_with_details
def home(request):
products = Product.objects.get_available_products()[:6]
products = products.prefetch_related('categories', 'images',
'variants__stock')
return TemplateResponse(
request, 'home.html',
{'products': products, 'parent': None})
<commit_msg>Fix homepage after wrong rebase<commit_after> | from django.template.response import TemplateResponse
from ..product.utils import products_with_availability, products_with_details
def home(request):
products = products_with_details(request.user)[:6]
products = products_with_availability(
products, discounts=request.discounts, local_currency=request.currency)
return TemplateResponse(
request, 'home.html',
{'products': products, 'parent': None})
| from django.template.response import TemplateResponse
from ..product.utils import products_with_availability, products_with_details
def home(request):
products = Product.objects.get_available_products()[:6]
products = products.prefetch_related('categories', 'images',
'variants__stock')
return TemplateResponse(
request, 'home.html',
{'products': products, 'parent': None})
Fix homepage after wrong rebasefrom django.template.response import TemplateResponse
from ..product.utils import products_with_availability, products_with_details
def home(request):
products = products_with_details(request.user)[:6]
products = products_with_availability(
products, discounts=request.discounts, local_currency=request.currency)
return TemplateResponse(
request, 'home.html',
{'products': products, 'parent': None})
| <commit_before>from django.template.response import TemplateResponse
from ..product.utils import products_with_availability, products_with_details
def home(request):
products = Product.objects.get_available_products()[:6]
products = products.prefetch_related('categories', 'images',
'variants__stock')
return TemplateResponse(
request, 'home.html',
{'products': products, 'parent': None})
<commit_msg>Fix homepage after wrong rebase<commit_after>from django.template.response import TemplateResponse
from ..product.utils import products_with_availability, products_with_details
def home(request):
products = products_with_details(request.user)[:6]
products = products_with_availability(
products, discounts=request.discounts, local_currency=request.currency)
return TemplateResponse(
request, 'home.html',
{'products': products, 'parent': None})
|
ef7163a18ee1cf11c1290f2a8832d8cf39fb552c | fjord/base/tests/test_commands.py | fjord/base/tests/test_commands.py | from django.core.management import call_command
from fjord.base.tests import TestCase
class TestGenerateData(TestCase):
def test_generate_data(self):
"""Make sure ./manage.py generatedata runs."""
call_command('generatedata')
call_command('generatedata', bigsample=True)
class TestPOLint(TestCase):
def test_polint(self):
"""Make sure ./manage.py polint runs."""
# Note: This doesn't make sure it works--just that it doesn't kick
# up obvious errors when it runs like if Dennis has changed in
# some way that prevents it from working correctly.
try:
call_command('polint')
except SystemExit:
# WOAH! WTF ARE YOU DOING? The lint command calls
# sys.exit() because it needs to return correct exit codes
# so we catch that here and ignore it. Otherwise testing
# it will kill the test suite.
pass
| from django.core.management import call_command
from fjord.base.tests import TestCase
class TestGenerateData(TestCase):
def test_generate_data(self):
"""Make sure ./manage.py generatedata runs."""
call_command('generatedata')
call_command('generatedata', bigsample=True)
class TestPOLint(TestCase):
def test_polint(self):
"""Make sure ./manage.py polint runs."""
# Note: This doesn't make sure it works--just that it doesn't kick
# up obvious errors when it runs like if Dennis has changed in
# some way that prevents it from working correctly.
try:
call_command('polint', '--version')
except SystemExit:
# WOAH! WTF ARE YOU DOING? The lint command calls
# sys.exit() because it needs to return correct exit codes
# so we catch that here and ignore it. Otherwise testing
# it will kill the test suite.
pass
| Adjust test_polint to be less stdout-spammy | Adjust test_polint to be less stdout-spammy
| Python | bsd-3-clause | hoosteeno/fjord,lgp171188/fjord,mozilla/fjord,Ritsyy/fjord,lgp171188/fjord,rlr/fjord,staranjeet/fjord,DESHRAJ/fjord,hoosteeno/fjord,Ritsyy/fjord,rlr/fjord,lgp171188/fjord,DESHRAJ/fjord,mozilla/fjord,hoosteeno/fjord,hoosteeno/fjord,rlr/fjord,Ritsyy/fjord,staranjeet/fjord,lgp171188/fjord,mozilla/fjord,Ritsyy/fjord,DESHRAJ/fjord,rlr/fjord,staranjeet/fjord,mozilla/fjord,staranjeet/fjord | from django.core.management import call_command
from fjord.base.tests import TestCase
class TestGenerateData(TestCase):
def test_generate_data(self):
"""Make sure ./manage.py generatedata runs."""
call_command('generatedata')
call_command('generatedata', bigsample=True)
class TestPOLint(TestCase):
def test_polint(self):
"""Make sure ./manage.py polint runs."""
# Note: This doesn't make sure it works--just that it doesn't kick
# up obvious errors when it runs like if Dennis has changed in
# some way that prevents it from working correctly.
try:
call_command('polint')
except SystemExit:
# WOAH! WTF ARE YOU DOING? The lint command calls
# sys.exit() because it needs to return correct exit codes
# so we catch that here and ignore it. Otherwise testing
# it will kill the test suite.
pass
Adjust test_polint to be less stdout-spammy | from django.core.management import call_command
from fjord.base.tests import TestCase
class TestGenerateData(TestCase):
def test_generate_data(self):
"""Make sure ./manage.py generatedata runs."""
call_command('generatedata')
call_command('generatedata', bigsample=True)
class TestPOLint(TestCase):
def test_polint(self):
"""Make sure ./manage.py polint runs."""
# Note: This doesn't make sure it works--just that it doesn't kick
# up obvious errors when it runs like if Dennis has changed in
# some way that prevents it from working correctly.
try:
call_command('polint', '--version')
except SystemExit:
# WOAH! WTF ARE YOU DOING? The lint command calls
# sys.exit() because it needs to return correct exit codes
# so we catch that here and ignore it. Otherwise testing
# it will kill the test suite.
pass
| <commit_before>from django.core.management import call_command
from fjord.base.tests import TestCase
class TestGenerateData(TestCase):
def test_generate_data(self):
"""Make sure ./manage.py generatedata runs."""
call_command('generatedata')
call_command('generatedata', bigsample=True)
class TestPOLint(TestCase):
def test_polint(self):
"""Make sure ./manage.py polint runs."""
# Note: This doesn't make sure it works--just that it doesn't kick
# up obvious errors when it runs like if Dennis has changed in
# some way that prevents it from working correctly.
try:
call_command('polint')
except SystemExit:
# WOAH! WTF ARE YOU DOING? The lint command calls
# sys.exit() because it needs to return correct exit codes
# so we catch that here and ignore it. Otherwise testing
# it will kill the test suite.
pass
<commit_msg>Adjust test_polint to be less stdout-spammy<commit_after> | from django.core.management import call_command
from fjord.base.tests import TestCase
class TestGenerateData(TestCase):
def test_generate_data(self):
"""Make sure ./manage.py generatedata runs."""
call_command('generatedata')
call_command('generatedata', bigsample=True)
class TestPOLint(TestCase):
def test_polint(self):
"""Make sure ./manage.py polint runs."""
# Note: This doesn't make sure it works--just that it doesn't kick
# up obvious errors when it runs like if Dennis has changed in
# some way that prevents it from working correctly.
try:
call_command('polint', '--version')
except SystemExit:
# WOAH! WTF ARE YOU DOING? The lint command calls
# sys.exit() because it needs to return correct exit codes
# so we catch that here and ignore it. Otherwise testing
# it will kill the test suite.
pass
| from django.core.management import call_command
from fjord.base.tests import TestCase
class TestGenerateData(TestCase):
def test_generate_data(self):
"""Make sure ./manage.py generatedata runs."""
call_command('generatedata')
call_command('generatedata', bigsample=True)
class TestPOLint(TestCase):
def test_polint(self):
"""Make sure ./manage.py polint runs."""
# Note: This doesn't make sure it works--just that it doesn't kick
# up obvious errors when it runs like if Dennis has changed in
# some way that prevents it from working correctly.
try:
call_command('polint')
except SystemExit:
# WOAH! WTF ARE YOU DOING? The lint command calls
# sys.exit() because it needs to return correct exit codes
# so we catch that here and ignore it. Otherwise testing
# it will kill the test suite.
pass
Adjust test_polint to be less stdout-spammyfrom django.core.management import call_command
from fjord.base.tests import TestCase
class TestGenerateData(TestCase):
def test_generate_data(self):
"""Make sure ./manage.py generatedata runs."""
call_command('generatedata')
call_command('generatedata', bigsample=True)
class TestPOLint(TestCase):
def test_polint(self):
"""Make sure ./manage.py polint runs."""
# Note: This doesn't make sure it works--just that it doesn't kick
# up obvious errors when it runs like if Dennis has changed in
# some way that prevents it from working correctly.
try:
call_command('polint', '--version')
except SystemExit:
# WOAH! WTF ARE YOU DOING? The lint command calls
# sys.exit() because it needs to return correct exit codes
# so we catch that here and ignore it. Otherwise testing
# it will kill the test suite.
pass
| <commit_before>from django.core.management import call_command
from fjord.base.tests import TestCase
class TestGenerateData(TestCase):
def test_generate_data(self):
"""Make sure ./manage.py generatedata runs."""
call_command('generatedata')
call_command('generatedata', bigsample=True)
class TestPOLint(TestCase):
def test_polint(self):
"""Make sure ./manage.py polint runs."""
# Note: This doesn't make sure it works--just that it doesn't kick
# up obvious errors when it runs like if Dennis has changed in
# some way that prevents it from working correctly.
try:
call_command('polint')
except SystemExit:
# WOAH! WTF ARE YOU DOING? The lint command calls
# sys.exit() because it needs to return correct exit codes
# so we catch that here and ignore it. Otherwise testing
# it will kill the test suite.
pass
<commit_msg>Adjust test_polint to be less stdout-spammy<commit_after>from django.core.management import call_command
from fjord.base.tests import TestCase
class TestGenerateData(TestCase):
def test_generate_data(self):
"""Make sure ./manage.py generatedata runs."""
call_command('generatedata')
call_command('generatedata', bigsample=True)
class TestPOLint(TestCase):
def test_polint(self):
"""Make sure ./manage.py polint runs."""
# Note: This doesn't make sure it works--just that it doesn't kick
# up obvious errors when it runs like if Dennis has changed in
# some way that prevents it from working correctly.
try:
call_command('polint', '--version')
except SystemExit:
# WOAH! WTF ARE YOU DOING? The lint command calls
# sys.exit() because it needs to return correct exit codes
# so we catch that here and ignore it. Otherwise testing
# it will kill the test suite.
pass
|
c0c822df243894106a7fd376582598e4aebb4c24 | kobo/hub/decorators.py | kobo/hub/decorators.py | # -*- coding: utf-8 -*-
from django.core.exceptions import PermissionDenied, SuspiciousOperation
from kobo.decorators import decorator_with_args
from kobo.django.xmlrpc.decorators import *
def validate_worker(func):
def _new_func(request, *args, **kwargs):
if not request.user.is_authenticated():
raise PermissionDenied("Login required.")
if getattr(request, 'worker', None) is None:
raise SuspiciousOperation("User doesn't match any worker: %s" % request.user.username)
return func(request, *args, **kwargs)
_new_func.__name__ = func.__name__
_new_func.__doc__ = func.__doc__
_new_func.__dict__.update(func.__dict__)
return _new_func
| # -*- coding: utf-8 -*-
import socket
from django.core.exceptions import PermissionDenied, SuspiciousOperation
from kobo.decorators import decorator_with_args
from kobo.django.xmlrpc.decorators import *
def validate_worker(func):
def _new_func(request, *args, **kwargs):
if not request.user.is_authenticated():
raise PermissionDenied("Login required.")
if getattr(request, 'worker', None) is None:
raise SuspiciousOperation("User doesn't match any worker: %s" % request.user.username)
fqdn = socket.getfqdn(request.META["REMOTE_ADDR"])
prefix, hostname = request.user.username.split("/", 1)
if hostname != fqdn:
raise SuspiciousOperation("Worker's FQDN (%s) doesn't match username (%s)" % (fqdn, hostname))
return func(request, *args, **kwargs)
_new_func.__name__ = func.__name__
_new_func.__doc__ = func.__doc__
_new_func.__dict__.update(func.__dict__)
return _new_func
| Check worker's FQDN against username. | Check worker's FQDN against username.
| Python | lgpl-2.1 | pombredanne/https-git.fedorahosted.org-git-kobo,pombredanne/https-git.fedorahosted.org-git-kobo,pombredanne/https-git.fedorahosted.org-git-kobo,release-engineering/kobo,release-engineering/kobo,release-engineering/kobo,release-engineering/kobo,pombredanne/https-git.fedorahosted.org-git-kobo | # -*- coding: utf-8 -*-
from django.core.exceptions import PermissionDenied, SuspiciousOperation
from kobo.decorators import decorator_with_args
from kobo.django.xmlrpc.decorators import *
def validate_worker(func):
def _new_func(request, *args, **kwargs):
if not request.user.is_authenticated():
raise PermissionDenied("Login required.")
if getattr(request, 'worker', None) is None:
raise SuspiciousOperation("User doesn't match any worker: %s" % request.user.username)
return func(request, *args, **kwargs)
_new_func.__name__ = func.__name__
_new_func.__doc__ = func.__doc__
_new_func.__dict__.update(func.__dict__)
return _new_func
Check worker's FQDN against username. | # -*- coding: utf-8 -*-
import socket
from django.core.exceptions import PermissionDenied, SuspiciousOperation
from kobo.decorators import decorator_with_args
from kobo.django.xmlrpc.decorators import *
def validate_worker(func):
def _new_func(request, *args, **kwargs):
if not request.user.is_authenticated():
raise PermissionDenied("Login required.")
if getattr(request, 'worker', None) is None:
raise SuspiciousOperation("User doesn't match any worker: %s" % request.user.username)
fqdn = socket.getfqdn(request.META["REMOTE_ADDR"])
prefix, hostname = request.user.username.split("/", 1)
if hostname != fqdn:
raise SuspiciousOperation("Worker's FQDN (%s) doesn't match username (%s)" % (fqdn, hostname))
return func(request, *args, **kwargs)
_new_func.__name__ = func.__name__
_new_func.__doc__ = func.__doc__
_new_func.__dict__.update(func.__dict__)
return _new_func
| <commit_before># -*- coding: utf-8 -*-
from django.core.exceptions import PermissionDenied, SuspiciousOperation
from kobo.decorators import decorator_with_args
from kobo.django.xmlrpc.decorators import *
def validate_worker(func):
def _new_func(request, *args, **kwargs):
if not request.user.is_authenticated():
raise PermissionDenied("Login required.")
if getattr(request, 'worker', None) is None:
raise SuspiciousOperation("User doesn't match any worker: %s" % request.user.username)
return func(request, *args, **kwargs)
_new_func.__name__ = func.__name__
_new_func.__doc__ = func.__doc__
_new_func.__dict__.update(func.__dict__)
return _new_func
<commit_msg>Check worker's FQDN against username.<commit_after> | # -*- coding: utf-8 -*-
import socket
from django.core.exceptions import PermissionDenied, SuspiciousOperation
from kobo.decorators import decorator_with_args
from kobo.django.xmlrpc.decorators import *
def validate_worker(func):
def _new_func(request, *args, **kwargs):
if not request.user.is_authenticated():
raise PermissionDenied("Login required.")
if getattr(request, 'worker', None) is None:
raise SuspiciousOperation("User doesn't match any worker: %s" % request.user.username)
fqdn = socket.getfqdn(request.META["REMOTE_ADDR"])
prefix, hostname = request.user.username.split("/", 1)
if hostname != fqdn:
raise SuspiciousOperation("Worker's FQDN (%s) doesn't match username (%s)" % (fqdn, hostname))
return func(request, *args, **kwargs)
_new_func.__name__ = func.__name__
_new_func.__doc__ = func.__doc__
_new_func.__dict__.update(func.__dict__)
return _new_func
| # -*- coding: utf-8 -*-
from django.core.exceptions import PermissionDenied, SuspiciousOperation
from kobo.decorators import decorator_with_args
from kobo.django.xmlrpc.decorators import *
def validate_worker(func):
def _new_func(request, *args, **kwargs):
if not request.user.is_authenticated():
raise PermissionDenied("Login required.")
if getattr(request, 'worker', None) is None:
raise SuspiciousOperation("User doesn't match any worker: %s" % request.user.username)
return func(request, *args, **kwargs)
_new_func.__name__ = func.__name__
_new_func.__doc__ = func.__doc__
_new_func.__dict__.update(func.__dict__)
return _new_func
Check worker's FQDN against username.# -*- coding: utf-8 -*-
import socket
from django.core.exceptions import PermissionDenied, SuspiciousOperation
from kobo.decorators import decorator_with_args
from kobo.django.xmlrpc.decorators import *
def validate_worker(func):
def _new_func(request, *args, **kwargs):
if not request.user.is_authenticated():
raise PermissionDenied("Login required.")
if getattr(request, 'worker', None) is None:
raise SuspiciousOperation("User doesn't match any worker: %s" % request.user.username)
fqdn = socket.getfqdn(request.META["REMOTE_ADDR"])
prefix, hostname = request.user.username.split("/", 1)
if hostname != fqdn:
raise SuspiciousOperation("Worker's FQDN (%s) doesn't match username (%s)" % (fqdn, hostname))
return func(request, *args, **kwargs)
_new_func.__name__ = func.__name__
_new_func.__doc__ = func.__doc__
_new_func.__dict__.update(func.__dict__)
return _new_func
| <commit_before># -*- coding: utf-8 -*-
from django.core.exceptions import PermissionDenied, SuspiciousOperation
from kobo.decorators import decorator_with_args
from kobo.django.xmlrpc.decorators import *
def validate_worker(func):
def _new_func(request, *args, **kwargs):
if not request.user.is_authenticated():
raise PermissionDenied("Login required.")
if getattr(request, 'worker', None) is None:
raise SuspiciousOperation("User doesn't match any worker: %s" % request.user.username)
return func(request, *args, **kwargs)
_new_func.__name__ = func.__name__
_new_func.__doc__ = func.__doc__
_new_func.__dict__.update(func.__dict__)
return _new_func
<commit_msg>Check worker's FQDN against username.<commit_after># -*- coding: utf-8 -*-
import socket
from django.core.exceptions import PermissionDenied, SuspiciousOperation
from kobo.decorators import decorator_with_args
from kobo.django.xmlrpc.decorators import *
def validate_worker(func):
def _new_func(request, *args, **kwargs):
if not request.user.is_authenticated():
raise PermissionDenied("Login required.")
if getattr(request, 'worker', None) is None:
raise SuspiciousOperation("User doesn't match any worker: %s" % request.user.username)
fqdn = socket.getfqdn(request.META["REMOTE_ADDR"])
prefix, hostname = request.user.username.split("/", 1)
if hostname != fqdn:
raise SuspiciousOperation("Worker's FQDN (%s) doesn't match username (%s)" % (fqdn, hostname))
return func(request, *args, **kwargs)
_new_func.__name__ = func.__name__
_new_func.__doc__ = func.__doc__
_new_func.__dict__.update(func.__dict__)
return _new_func
|
337e7ae58abc7c192633144cc3913078ae3d38bf | hc2002/plugin/symbolic_values.py | hc2002/plugin/symbolic_values.py | import hc2002.plugin as plugin
import hc2002.config as config
plugin.register_for_resource(__name__, 'hc2002.resource.instance')
_prefixes = ('availability-zone:', 'image:', 'kernel:', 'key:',
'load-balancers:', 'ramdisk:', 'security-groups:', 'spot-price:',
'subnet:')
def apply(instance):
def _resolve_symbol(value):
visited = set()
while isinstance(value, basestring) \
and value.startswith(prefix):
value = value.format(region=config.region, **instance)
if value in instance \
and value not in visited:
visited.add(value)
value = instance[value]
return value
# Resolve symbols
for prefix in _prefixes:
key = prefix[:-1]
if key not in instance:
continue
if isinstance(instance[key], basestring):
instance[key] = _resolve_symbol(instance[key])
elif isinstance(instance[key], list):
instance[key] = map(_resolve_symbol, instance[key])
# Drop resolvable symbols
for key in instance.keys():
if key.startswith(_prefixes):
del instance[key]
| import hc2002.plugin as plugin
import hc2002.config as config
plugin.register_for_resource(__name__, 'hc2002.resource.instance')
_prefixes = ('availability-zone:', 'image:', 'kernel:', 'key:',
'load-balancers:', 'ramdisk:', 'security-groups:', 'spot-price:',
'subnet:')
def apply(instance):
def resolve_symbol(original_value):
value = original_value
visited = set()
while isinstance(value, basestring) \
and value.startswith(prefix):
value = value.format(region=config.region, **instance)
if value in instance \
and value not in visited:
visited.add(value)
value = instance[value]
else:
if original_value == value:
raise Exception("Unable to resolve '%s'" % value)
else:
raise Exception(
"While resolving '%s': unable to resolve '%s'"
% (original_value, value))
return value
# Resolve symbols
for prefix in _prefixes:
key = prefix[:-1]
if key not in instance:
continue
if isinstance(instance[key], basestring):
instance[key] = resolve_symbol(instance[key])
elif isinstance(instance[key], list):
instance[key] = map(resolve_symbol, instance[key])
# Drop resolvable symbols
for key in instance.keys():
if key.startswith(_prefixes):
del instance[key]
| Fix infinite loop with missing symbol definition | Fix infinite loop with missing symbol definition
The local function _resolve_symbol is renamed for improved tracebacks.
| Python | apache-2.0 | biochimia/hc2000 | import hc2002.plugin as plugin
import hc2002.config as config
plugin.register_for_resource(__name__, 'hc2002.resource.instance')
_prefixes = ('availability-zone:', 'image:', 'kernel:', 'key:',
'load-balancers:', 'ramdisk:', 'security-groups:', 'spot-price:',
'subnet:')
def apply(instance):
def _resolve_symbol(value):
visited = set()
while isinstance(value, basestring) \
and value.startswith(prefix):
value = value.format(region=config.region, **instance)
if value in instance \
and value not in visited:
visited.add(value)
value = instance[value]
return value
# Resolve symbols
for prefix in _prefixes:
key = prefix[:-1]
if key not in instance:
continue
if isinstance(instance[key], basestring):
instance[key] = _resolve_symbol(instance[key])
elif isinstance(instance[key], list):
instance[key] = map(_resolve_symbol, instance[key])
# Drop resolvable symbols
for key in instance.keys():
if key.startswith(_prefixes):
del instance[key]
Fix infinite loop with missing symbol definition
The local function _resolve_symbol is renamed for improved tracebacks. | import hc2002.plugin as plugin
import hc2002.config as config
plugin.register_for_resource(__name__, 'hc2002.resource.instance')
_prefixes = ('availability-zone:', 'image:', 'kernel:', 'key:',
'load-balancers:', 'ramdisk:', 'security-groups:', 'spot-price:',
'subnet:')
def apply(instance):
def resolve_symbol(original_value):
value = original_value
visited = set()
while isinstance(value, basestring) \
and value.startswith(prefix):
value = value.format(region=config.region, **instance)
if value in instance \
and value not in visited:
visited.add(value)
value = instance[value]
else:
if original_value == value:
raise Exception("Unable to resolve '%s'" % value)
else:
raise Exception(
"While resolving '%s': unable to resolve '%s'"
% (original_value, value))
return value
# Resolve symbols
for prefix in _prefixes:
key = prefix[:-1]
if key not in instance:
continue
if isinstance(instance[key], basestring):
instance[key] = resolve_symbol(instance[key])
elif isinstance(instance[key], list):
instance[key] = map(resolve_symbol, instance[key])
# Drop resolvable symbols
for key in instance.keys():
if key.startswith(_prefixes):
del instance[key]
| <commit_before>import hc2002.plugin as plugin
import hc2002.config as config
plugin.register_for_resource(__name__, 'hc2002.resource.instance')
_prefixes = ('availability-zone:', 'image:', 'kernel:', 'key:',
'load-balancers:', 'ramdisk:', 'security-groups:', 'spot-price:',
'subnet:')
def apply(instance):
def _resolve_symbol(value):
visited = set()
while isinstance(value, basestring) \
and value.startswith(prefix):
value = value.format(region=config.region, **instance)
if value in instance \
and value not in visited:
visited.add(value)
value = instance[value]
return value
# Resolve symbols
for prefix in _prefixes:
key = prefix[:-1]
if key not in instance:
continue
if isinstance(instance[key], basestring):
instance[key] = _resolve_symbol(instance[key])
elif isinstance(instance[key], list):
instance[key] = map(_resolve_symbol, instance[key])
# Drop resolvable symbols
for key in instance.keys():
if key.startswith(_prefixes):
del instance[key]
<commit_msg>Fix infinite loop with missing symbol definition
The local function _resolve_symbol is renamed for improved tracebacks.<commit_after> | import hc2002.plugin as plugin
import hc2002.config as config
plugin.register_for_resource(__name__, 'hc2002.resource.instance')
_prefixes = ('availability-zone:', 'image:', 'kernel:', 'key:',
'load-balancers:', 'ramdisk:', 'security-groups:', 'spot-price:',
'subnet:')
def apply(instance):
def resolve_symbol(original_value):
value = original_value
visited = set()
while isinstance(value, basestring) \
and value.startswith(prefix):
value = value.format(region=config.region, **instance)
if value in instance \
and value not in visited:
visited.add(value)
value = instance[value]
else:
if original_value == value:
raise Exception("Unable to resolve '%s'" % value)
else:
raise Exception(
"While resolving '%s': unable to resolve '%s'"
% (original_value, value))
return value
# Resolve symbols
for prefix in _prefixes:
key = prefix[:-1]
if key not in instance:
continue
if isinstance(instance[key], basestring):
instance[key] = resolve_symbol(instance[key])
elif isinstance(instance[key], list):
instance[key] = map(resolve_symbol, instance[key])
# Drop resolvable symbols
for key in instance.keys():
if key.startswith(_prefixes):
del instance[key]
| import hc2002.plugin as plugin
import hc2002.config as config
plugin.register_for_resource(__name__, 'hc2002.resource.instance')
_prefixes = ('availability-zone:', 'image:', 'kernel:', 'key:',
'load-balancers:', 'ramdisk:', 'security-groups:', 'spot-price:',
'subnet:')
def apply(instance):
def _resolve_symbol(value):
visited = set()
while isinstance(value, basestring) \
and value.startswith(prefix):
value = value.format(region=config.region, **instance)
if value in instance \
and value not in visited:
visited.add(value)
value = instance[value]
return value
# Resolve symbols
for prefix in _prefixes:
key = prefix[:-1]
if key not in instance:
continue
if isinstance(instance[key], basestring):
instance[key] = _resolve_symbol(instance[key])
elif isinstance(instance[key], list):
instance[key] = map(_resolve_symbol, instance[key])
# Drop resolvable symbols
for key in instance.keys():
if key.startswith(_prefixes):
del instance[key]
Fix infinite loop with missing symbol definition
The local function _resolve_symbol is renamed for improved tracebacks.import hc2002.plugin as plugin
import hc2002.config as config
plugin.register_for_resource(__name__, 'hc2002.resource.instance')
_prefixes = ('availability-zone:', 'image:', 'kernel:', 'key:',
'load-balancers:', 'ramdisk:', 'security-groups:', 'spot-price:',
'subnet:')
def apply(instance):
def resolve_symbol(original_value):
value = original_value
visited = set()
while isinstance(value, basestring) \
and value.startswith(prefix):
value = value.format(region=config.region, **instance)
if value in instance \
and value not in visited:
visited.add(value)
value = instance[value]
else:
if original_value == value:
raise Exception("Unable to resolve '%s'" % value)
else:
raise Exception(
"While resolving '%s': unable to resolve '%s'"
% (original_value, value))
return value
# Resolve symbols
for prefix in _prefixes:
key = prefix[:-1]
if key not in instance:
continue
if isinstance(instance[key], basestring):
instance[key] = resolve_symbol(instance[key])
elif isinstance(instance[key], list):
instance[key] = map(resolve_symbol, instance[key])
# Drop resolvable symbols
for key in instance.keys():
if key.startswith(_prefixes):
del instance[key]
| <commit_before>import hc2002.plugin as plugin
import hc2002.config as config
plugin.register_for_resource(__name__, 'hc2002.resource.instance')
_prefixes = ('availability-zone:', 'image:', 'kernel:', 'key:',
'load-balancers:', 'ramdisk:', 'security-groups:', 'spot-price:',
'subnet:')
def apply(instance):
def _resolve_symbol(value):
visited = set()
while isinstance(value, basestring) \
and value.startswith(prefix):
value = value.format(region=config.region, **instance)
if value in instance \
and value not in visited:
visited.add(value)
value = instance[value]
return value
# Resolve symbols
for prefix in _prefixes:
key = prefix[:-1]
if key not in instance:
continue
if isinstance(instance[key], basestring):
instance[key] = _resolve_symbol(instance[key])
elif isinstance(instance[key], list):
instance[key] = map(_resolve_symbol, instance[key])
# Drop resolvable symbols
for key in instance.keys():
if key.startswith(_prefixes):
del instance[key]
<commit_msg>Fix infinite loop with missing symbol definition
The local function _resolve_symbol is renamed for improved tracebacks.<commit_after>import hc2002.plugin as plugin
import hc2002.config as config
plugin.register_for_resource(__name__, 'hc2002.resource.instance')
_prefixes = ('availability-zone:', 'image:', 'kernel:', 'key:',
'load-balancers:', 'ramdisk:', 'security-groups:', 'spot-price:',
'subnet:')
def apply(instance):
def resolve_symbol(original_value):
value = original_value
visited = set()
while isinstance(value, basestring) \
and value.startswith(prefix):
value = value.format(region=config.region, **instance)
if value in instance \
and value not in visited:
visited.add(value)
value = instance[value]
else:
if original_value == value:
raise Exception("Unable to resolve '%s'" % value)
else:
raise Exception(
"While resolving '%s': unable to resolve '%s'"
% (original_value, value))
return value
# Resolve symbols
for prefix in _prefixes:
key = prefix[:-1]
if key not in instance:
continue
if isinstance(instance[key], basestring):
instance[key] = resolve_symbol(instance[key])
elif isinstance(instance[key], list):
instance[key] = map(resolve_symbol, instance[key])
# Drop resolvable symbols
for key in instance.keys():
if key.startswith(_prefixes):
del instance[key]
|
470f089a764185350b698725c6720e602c1eb804 | neutron/plugins/ml2/drivers/mlnx/config.py | neutron/plugins/ml2/drivers/mlnx/config.py | # Copyright (c) 2014 OpenStack Foundation
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from oslo.config import cfg
from neutron.extensions import portbindings
eswitch_opts = [
cfg.StrOpt('vnic_type',
default=portbindings.VIF_TYPE_MLNX_DIRECT,
help=_("Type of VM network interface: mlnx_direct or "
"hostdev")),
cfg.BoolOpt('apply_profile_patch',
default=False,
help=_("Enable server compatibility with old nova ")),
]
cfg.CONF.register_opts(eswitch_opts, "ESWITCH")
| # Copyright (c) 2014 OpenStack Foundation
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from oslo.config import cfg
from neutron.extensions import portbindings
eswitch_opts = [
cfg.StrOpt('vnic_type',
default=portbindings.VIF_TYPE_MLNX_DIRECT,
help=_("Type of VM network interface: mlnx_direct or "
"hostdev")),
cfg.BoolOpt('apply_profile_patch',
default=False,
help=_("Enable server compatibility with old nova")),
]
cfg.CONF.register_opts(eswitch_opts, "ESWITCH")
| Remove extra space in help string | Remove extra space in help string
Extra spaces make the openstack-manuals tests fail with a niceness
error. This patch removes an extra space at the end of a help string.
Change-Id: I29bab90ea5a6f648c4539c7cd20cd9b2b63055c2
| Python | apache-2.0 | apporc/neutron,magic0704/neutron,wenhuizhang/neutron,glove747/liberty-neutron,asgard-lab/neutron,neoareslinux/neutron,adelina-t/neutron,silenci/neutron,chitr/neutron,CiscoSystems/neutron,waltBB/neutron_read,paninetworks/neutron,openstack/neutron,sasukeh/neutron,yamahata/neutron,eayunstack/neutron,aristanetworks/neutron,eonpatapon/neutron,swdream/neutron,yamahata/neutron,barnsnake351/neutron,swdream/neutron,miyakz1192/neutron,gkotton/neutron,cisco-openstack/neutron,sasukeh/neutron,alexandrucoman/vbox-neutron-agent,leeseuljeong/leeseulstack_neutron,paninetworks/neutron,noironetworks/neutron,sebrandon1/neutron,vijayendrabvs/hap,cisco-openstack/neutron,leeseulstack/openstack,vijayendrabvs/hap,vveerava/Openstack,gkotton/neutron,projectcalico/calico-neutron,gopal1cloud/neutron,antonioUnina/neutron,leeseulstack/openstack,mahak/neutron,watonyweng/neutron,wolverineav/neutron,redhat-openstack/neutron,projectcalico/calico-neutron,wolverineav/neutron,takeshineshiro/neutron,mattt416/neutron,javaos74/neutron,Stavitsky/neutron,blueboxgroup/neutron,SmartInfrastructures/neutron,vivekanand1101/neutron,vveerava/Openstack,rdo-management/neutron,infobloxopen/neutron,klmitch/neutron,vivekanand1101/neutron,gopal1cloud/neutron,adelina-t/neutron,cloudbase/neutron-virtualbox,eayunstack/neutron,leeseulstack/openstack,chitr/neutron,yamahata/tacker,wenhuizhang/neutron,mmnelemane/neutron,jumpojoy/neutron,eonpatapon/neutron,SmartInfrastructures/neutron,takeshineshiro/neutron,Stavitsky/neutron,CiscoSystems/neutron,suneeth51/neutron,bgxavier/neutron,mattt416/neutron,miyakz1192/neutron,Metaswitch/calico-neutron,skyddv/neutron,jerryz1982/neutron,huntxu/neutron,cloudbase/neutron,watonyweng/neutron,shahbazn/neutron,igor-toga/local-snat,JianyuWang/neutron,antonioUnina/neutron,mandeepdhami/neutron,sajuptpm/neutron-ipam,barnsnake351/neutron,cernops/neutron,huntxu/neutron,JioCloud/neutron,vijayendrabvs/hap,silenci/neutron,blueboxgroup/neutron,bgxavier/neutron,yuewko/neutron,dhanunjaya/neutron,cernops/neutron,SamYaple/neutron,asgard-lab/neutron,neoareslinux/neutron,sajuptpm/neutron-ipam,virtualopensystems/neutron,redhat-openstack/neutron,shahbazn/neutron,mahak/neutron,mahak/neutron,pnavarro/neutron,cloudbase/neutron,CiscoSystems/neutron,vveerava/Openstack,leeseuljeong/leeseulstack_neutron,yamahata/tacker,suneeth51/neutron,dims/neutron,magic0704/neutron,MaximNevrov/neutron,leeseuljeong/leeseulstack_neutron,MaximNevrov/neutron,skyddv/neutron,waltBB/neutron_read,blueboxgroup/neutron,pnavarro/neutron,bigswitch/neutron,virtualopensystems/neutron,cloudbase/neutron-virtualbox,sajuptpm/neutron-ipam,vbannai/neutron,klmitch/neutron,mandeepdhami/neutron,aristanetworks/neutron,virtualopensystems/neutron,openstack/neutron,igor-toga/local-snat,mmnelemane/neutron,yamahata/tacker,Metaswitch/calico-neutron,apporc/neutron,sebrandon1/neutron,alexandrucoman/vbox-neutron-agent,NeCTAR-RC/neutron,yamahata/neutron,infobloxopen/neutron,noironetworks/neutron,jumpojoy/neutron,vbannai/neutron,yanheven/neutron,SamYaple/neutron,JioCloud/neutron,JianyuWang/neutron,yuewko/neutron,jacknjzhou/neutron,NeCTAR-RC/neutron,dhanunjaya/neutron,javaos74/neutron,glove747/liberty-neutron,jerryz1982/neutron,vbannai/neutron,yanheven/neutron,dims/neutron,bigswitch/neutron,gkotton/neutron,rdo-management/neutron,jacknjzhou/neutron,openstack/neutron | # Copyright (c) 2014 OpenStack Foundation
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from oslo.config import cfg
from neutron.extensions import portbindings
eswitch_opts = [
cfg.StrOpt('vnic_type',
default=portbindings.VIF_TYPE_MLNX_DIRECT,
help=_("Type of VM network interface: mlnx_direct or "
"hostdev")),
cfg.BoolOpt('apply_profile_patch',
default=False,
help=_("Enable server compatibility with old nova ")),
]
cfg.CONF.register_opts(eswitch_opts, "ESWITCH")
Remove extra space in help string
Extra spaces make the openstack-manuals tests fail with a niceness
error. This patch removes an extra space at the end of a help string.
Change-Id: I29bab90ea5a6f648c4539c7cd20cd9b2b63055c2 | # Copyright (c) 2014 OpenStack Foundation
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from oslo.config import cfg
from neutron.extensions import portbindings
eswitch_opts = [
cfg.StrOpt('vnic_type',
default=portbindings.VIF_TYPE_MLNX_DIRECT,
help=_("Type of VM network interface: mlnx_direct or "
"hostdev")),
cfg.BoolOpt('apply_profile_patch',
default=False,
help=_("Enable server compatibility with old nova")),
]
cfg.CONF.register_opts(eswitch_opts, "ESWITCH")
| <commit_before># Copyright (c) 2014 OpenStack Foundation
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from oslo.config import cfg
from neutron.extensions import portbindings
eswitch_opts = [
cfg.StrOpt('vnic_type',
default=portbindings.VIF_TYPE_MLNX_DIRECT,
help=_("Type of VM network interface: mlnx_direct or "
"hostdev")),
cfg.BoolOpt('apply_profile_patch',
default=False,
help=_("Enable server compatibility with old nova ")),
]
cfg.CONF.register_opts(eswitch_opts, "ESWITCH")
<commit_msg>Remove extra space in help string
Extra spaces make the openstack-manuals tests fail with a niceness
error. This patch removes an extra space at the end of a help string.
Change-Id: I29bab90ea5a6f648c4539c7cd20cd9b2b63055c2<commit_after> | # Copyright (c) 2014 OpenStack Foundation
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from oslo.config import cfg
from neutron.extensions import portbindings
eswitch_opts = [
cfg.StrOpt('vnic_type',
default=portbindings.VIF_TYPE_MLNX_DIRECT,
help=_("Type of VM network interface: mlnx_direct or "
"hostdev")),
cfg.BoolOpt('apply_profile_patch',
default=False,
help=_("Enable server compatibility with old nova")),
]
cfg.CONF.register_opts(eswitch_opts, "ESWITCH")
| # Copyright (c) 2014 OpenStack Foundation
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from oslo.config import cfg
from neutron.extensions import portbindings
eswitch_opts = [
cfg.StrOpt('vnic_type',
default=portbindings.VIF_TYPE_MLNX_DIRECT,
help=_("Type of VM network interface: mlnx_direct or "
"hostdev")),
cfg.BoolOpt('apply_profile_patch',
default=False,
help=_("Enable server compatibility with old nova ")),
]
cfg.CONF.register_opts(eswitch_opts, "ESWITCH")
Remove extra space in help string
Extra spaces make the openstack-manuals tests fail with a niceness
error. This patch removes an extra space at the end of a help string.
Change-Id: I29bab90ea5a6f648c4539c7cd20cd9b2b63055c2# Copyright (c) 2014 OpenStack Foundation
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from oslo.config import cfg
from neutron.extensions import portbindings
eswitch_opts = [
cfg.StrOpt('vnic_type',
default=portbindings.VIF_TYPE_MLNX_DIRECT,
help=_("Type of VM network interface: mlnx_direct or "
"hostdev")),
cfg.BoolOpt('apply_profile_patch',
default=False,
help=_("Enable server compatibility with old nova")),
]
cfg.CONF.register_opts(eswitch_opts, "ESWITCH")
| <commit_before># Copyright (c) 2014 OpenStack Foundation
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from oslo.config import cfg
from neutron.extensions import portbindings
eswitch_opts = [
cfg.StrOpt('vnic_type',
default=portbindings.VIF_TYPE_MLNX_DIRECT,
help=_("Type of VM network interface: mlnx_direct or "
"hostdev")),
cfg.BoolOpt('apply_profile_patch',
default=False,
help=_("Enable server compatibility with old nova ")),
]
cfg.CONF.register_opts(eswitch_opts, "ESWITCH")
<commit_msg>Remove extra space in help string
Extra spaces make the openstack-manuals tests fail with a niceness
error. This patch removes an extra space at the end of a help string.
Change-Id: I29bab90ea5a6f648c4539c7cd20cd9b2b63055c2<commit_after># Copyright (c) 2014 OpenStack Foundation
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from oslo.config import cfg
from neutron.extensions import portbindings
eswitch_opts = [
cfg.StrOpt('vnic_type',
default=portbindings.VIF_TYPE_MLNX_DIRECT,
help=_("Type of VM network interface: mlnx_direct or "
"hostdev")),
cfg.BoolOpt('apply_profile_patch',
default=False,
help=_("Enable server compatibility with old nova")),
]
cfg.CONF.register_opts(eswitch_opts, "ESWITCH")
|
ad766ba20db73ceb433b9afe3b5db3e52cf1494a | addons/sale_stock/models/stock_config_settings.py | addons/sale_stock/models/stock_config_settings.py | # -*- coding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
from odoo import fields, models
class StockConfigSettings(models.TransientModel):
_inherit = 'stock.config.settings'
security_lead = fields.Float(related='company_id.security_lead')
default_new_security_lead = fields.Boolean(string="Security Lead Time for Sales", default_model="stock.config.settings", help="Margin of error for dates promised to customers. Products will be scheduled for procurement and delivery that many days earlier than the actual promised date, to cope with unexpected delays in the supply chain.")
default_picking_policy = fields.Selection([
('direct', 'Ship products as soon as available, with back orders'),
('one', 'Ship all products at once')
], "Shipping Management", default='direct', default_model="sale.order")
| # -*- coding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
from odoo import fields, models
class StockConfigSettings(models.TransientModel):
_inherit = 'stock.config.settings'
security_lead = fields.Float(related='company_id.security_lead')
default_new_security_lead = fields.Boolean(string="Security Lead Time for Sales", default_model="stock.config.settings", help="Margin of error for dates promised to customers. Products will be scheduled for procurement and delivery that many days earlier than the actual promised date, to cope with unexpected delays in the supply chain.")
default_picking_policy = fields.Selection([
('direct', 'Ship products as soon as available, with back orders'),
('one', 'Ship all products at once')
], "Shipping Management", default='direct', default_model="sale.order", required=True)
| Make default_picking_policy a required field in settings | [IMP] stock: Make default_picking_policy a required field in settings
| Python | agpl-3.0 | ygol/odoo,ygol/odoo,ygol/odoo,ygol/odoo,ygol/odoo,ygol/odoo,ygol/odoo | # -*- coding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
from odoo import fields, models
class StockConfigSettings(models.TransientModel):
_inherit = 'stock.config.settings'
security_lead = fields.Float(related='company_id.security_lead')
default_new_security_lead = fields.Boolean(string="Security Lead Time for Sales", default_model="stock.config.settings", help="Margin of error for dates promised to customers. Products will be scheduled for procurement and delivery that many days earlier than the actual promised date, to cope with unexpected delays in the supply chain.")
default_picking_policy = fields.Selection([
('direct', 'Ship products as soon as available, with back orders'),
('one', 'Ship all products at once')
], "Shipping Management", default='direct', default_model="sale.order")
[IMP] stock: Make default_picking_policy a required field in settings | # -*- coding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
from odoo import fields, models
class StockConfigSettings(models.TransientModel):
_inherit = 'stock.config.settings'
security_lead = fields.Float(related='company_id.security_lead')
default_new_security_lead = fields.Boolean(string="Security Lead Time for Sales", default_model="stock.config.settings", help="Margin of error for dates promised to customers. Products will be scheduled for procurement and delivery that many days earlier than the actual promised date, to cope with unexpected delays in the supply chain.")
default_picking_policy = fields.Selection([
('direct', 'Ship products as soon as available, with back orders'),
('one', 'Ship all products at once')
], "Shipping Management", default='direct', default_model="sale.order", required=True)
| <commit_before># -*- coding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
from odoo import fields, models
class StockConfigSettings(models.TransientModel):
_inherit = 'stock.config.settings'
security_lead = fields.Float(related='company_id.security_lead')
default_new_security_lead = fields.Boolean(string="Security Lead Time for Sales", default_model="stock.config.settings", help="Margin of error for dates promised to customers. Products will be scheduled for procurement and delivery that many days earlier than the actual promised date, to cope with unexpected delays in the supply chain.")
default_picking_policy = fields.Selection([
('direct', 'Ship products as soon as available, with back orders'),
('one', 'Ship all products at once')
], "Shipping Management", default='direct', default_model="sale.order")
<commit_msg>[IMP] stock: Make default_picking_policy a required field in settings<commit_after> | # -*- coding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
from odoo import fields, models
class StockConfigSettings(models.TransientModel):
_inherit = 'stock.config.settings'
security_lead = fields.Float(related='company_id.security_lead')
default_new_security_lead = fields.Boolean(string="Security Lead Time for Sales", default_model="stock.config.settings", help="Margin of error for dates promised to customers. Products will be scheduled for procurement and delivery that many days earlier than the actual promised date, to cope with unexpected delays in the supply chain.")
default_picking_policy = fields.Selection([
('direct', 'Ship products as soon as available, with back orders'),
('one', 'Ship all products at once')
], "Shipping Management", default='direct', default_model="sale.order", required=True)
| # -*- coding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
from odoo import fields, models
class StockConfigSettings(models.TransientModel):
_inherit = 'stock.config.settings'
security_lead = fields.Float(related='company_id.security_lead')
default_new_security_lead = fields.Boolean(string="Security Lead Time for Sales", default_model="stock.config.settings", help="Margin of error for dates promised to customers. Products will be scheduled for procurement and delivery that many days earlier than the actual promised date, to cope with unexpected delays in the supply chain.")
default_picking_policy = fields.Selection([
('direct', 'Ship products as soon as available, with back orders'),
('one', 'Ship all products at once')
], "Shipping Management", default='direct', default_model="sale.order")
[IMP] stock: Make default_picking_policy a required field in settings# -*- coding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
from odoo import fields, models
class StockConfigSettings(models.TransientModel):
_inherit = 'stock.config.settings'
security_lead = fields.Float(related='company_id.security_lead')
default_new_security_lead = fields.Boolean(string="Security Lead Time for Sales", default_model="stock.config.settings", help="Margin of error for dates promised to customers. Products will be scheduled for procurement and delivery that many days earlier than the actual promised date, to cope with unexpected delays in the supply chain.")
default_picking_policy = fields.Selection([
('direct', 'Ship products as soon as available, with back orders'),
('one', 'Ship all products at once')
], "Shipping Management", default='direct', default_model="sale.order", required=True)
| <commit_before># -*- coding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
from odoo import fields, models
class StockConfigSettings(models.TransientModel):
_inherit = 'stock.config.settings'
security_lead = fields.Float(related='company_id.security_lead')
default_new_security_lead = fields.Boolean(string="Security Lead Time for Sales", default_model="stock.config.settings", help="Margin of error for dates promised to customers. Products will be scheduled for procurement and delivery that many days earlier than the actual promised date, to cope with unexpected delays in the supply chain.")
default_picking_policy = fields.Selection([
('direct', 'Ship products as soon as available, with back orders'),
('one', 'Ship all products at once')
], "Shipping Management", default='direct', default_model="sale.order")
<commit_msg>[IMP] stock: Make default_picking_policy a required field in settings<commit_after># -*- coding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
from odoo import fields, models
class StockConfigSettings(models.TransientModel):
_inherit = 'stock.config.settings'
security_lead = fields.Float(related='company_id.security_lead')
default_new_security_lead = fields.Boolean(string="Security Lead Time for Sales", default_model="stock.config.settings", help="Margin of error for dates promised to customers. Products will be scheduled for procurement and delivery that many days earlier than the actual promised date, to cope with unexpected delays in the supply chain.")
default_picking_policy = fields.Selection([
('direct', 'Ship products as soon as available, with back orders'),
('one', 'Ship all products at once')
], "Shipping Management", default='direct', default_model="sale.order", required=True)
|
96dd8312cded04e5f5758495e72238d58905be7b | count.py | count.py | import json
from collections import Counter
from Bio import SeqIO
from Bio.SeqUtils import GC
for seq_record in SeqIO.parse("samples/test.fasta", "fasta"):
cnt = Counter()
line = str(seq_record.seq)
i = 0
for j in range(((len(line))/ 3)):
cnt[line[i:i+3]] += 1
i += 3
print json.dumps(
[{
"id" : seq_record.id,
"ratios" : cnt
}], sort_keys=True, indent=4, separators=(',', ': '))
| #!/usr/bin/python
import getopt
import json
import sys
from collections import Counter
from Bio import SeqIO
from Bio.SeqUtils import GC
def compute(inputfile):
for seq_record in SeqIO.parse(inputfile, "fasta"):
cnt = Counter()
line = str(seq_record.seq)
i = 0
for j in range(((len(line))/ 3)):
cnt[line[i:i+3]] += 1
i += 3
print json.dumps(
[{
"id" : seq_record.id,
"ratios" : cnt
}], sort_keys=True, indent=4, separators=(',', ': '))
def main(argv):
inputfile = 'samples/test.fasta'
try:
opts, args = getopt.getopt(argv,"hi:o:",["ifile=","ofile="])
except getopt.GetoptError:
print 'count.py -i <inputfile> -o <outputfile>'
sys.exit(2)
for opt, arg in opts:
if opt == '-h':
print 'count.py -i <inputfile>'
sys.exit()
elif opt in ("-i", "--ifile"):
inputfile = arg
print 'Input file is', inputfile
compute(inputfile)
if __name__ == "__main__":
main(sys.argv[1:])
| Add -i flag for input | Add -i flag for input
| Python | apache-2.0 | PDX-Flamingo/codonpdx-python,PDX-Flamingo/codonpdx-python | import json
from collections import Counter
from Bio import SeqIO
from Bio.SeqUtils import GC
for seq_record in SeqIO.parse("samples/test.fasta", "fasta"):
cnt = Counter()
line = str(seq_record.seq)
i = 0
for j in range(((len(line))/ 3)):
cnt[line[i:i+3]] += 1
i += 3
print json.dumps(
[{
"id" : seq_record.id,
"ratios" : cnt
}], sort_keys=True, indent=4, separators=(',', ': '))
Add -i flag for input | #!/usr/bin/python
import getopt
import json
import sys
from collections import Counter
from Bio import SeqIO
from Bio.SeqUtils import GC
def compute(inputfile):
for seq_record in SeqIO.parse(inputfile, "fasta"):
cnt = Counter()
line = str(seq_record.seq)
i = 0
for j in range(((len(line))/ 3)):
cnt[line[i:i+3]] += 1
i += 3
print json.dumps(
[{
"id" : seq_record.id,
"ratios" : cnt
}], sort_keys=True, indent=4, separators=(',', ': '))
def main(argv):
inputfile = 'samples/test.fasta'
try:
opts, args = getopt.getopt(argv,"hi:o:",["ifile=","ofile="])
except getopt.GetoptError:
print 'count.py -i <inputfile> -o <outputfile>'
sys.exit(2)
for opt, arg in opts:
if opt == '-h':
print 'count.py -i <inputfile>'
sys.exit()
elif opt in ("-i", "--ifile"):
inputfile = arg
print 'Input file is', inputfile
compute(inputfile)
if __name__ == "__main__":
main(sys.argv[1:])
| <commit_before>import json
from collections import Counter
from Bio import SeqIO
from Bio.SeqUtils import GC
for seq_record in SeqIO.parse("samples/test.fasta", "fasta"):
cnt = Counter()
line = str(seq_record.seq)
i = 0
for j in range(((len(line))/ 3)):
cnt[line[i:i+3]] += 1
i += 3
print json.dumps(
[{
"id" : seq_record.id,
"ratios" : cnt
}], sort_keys=True, indent=4, separators=(',', ': '))
<commit_msg>Add -i flag for input<commit_after> | #!/usr/bin/python
import getopt
import json
import sys
from collections import Counter
from Bio import SeqIO
from Bio.SeqUtils import GC
def compute(inputfile):
for seq_record in SeqIO.parse(inputfile, "fasta"):
cnt = Counter()
line = str(seq_record.seq)
i = 0
for j in range(((len(line))/ 3)):
cnt[line[i:i+3]] += 1
i += 3
print json.dumps(
[{
"id" : seq_record.id,
"ratios" : cnt
}], sort_keys=True, indent=4, separators=(',', ': '))
def main(argv):
inputfile = 'samples/test.fasta'
try:
opts, args = getopt.getopt(argv,"hi:o:",["ifile=","ofile="])
except getopt.GetoptError:
print 'count.py -i <inputfile> -o <outputfile>'
sys.exit(2)
for opt, arg in opts:
if opt == '-h':
print 'count.py -i <inputfile>'
sys.exit()
elif opt in ("-i", "--ifile"):
inputfile = arg
print 'Input file is', inputfile
compute(inputfile)
if __name__ == "__main__":
main(sys.argv[1:])
| import json
from collections import Counter
from Bio import SeqIO
from Bio.SeqUtils import GC
for seq_record in SeqIO.parse("samples/test.fasta", "fasta"):
cnt = Counter()
line = str(seq_record.seq)
i = 0
for j in range(((len(line))/ 3)):
cnt[line[i:i+3]] += 1
i += 3
print json.dumps(
[{
"id" : seq_record.id,
"ratios" : cnt
}], sort_keys=True, indent=4, separators=(',', ': '))
Add -i flag for input#!/usr/bin/python
import getopt
import json
import sys
from collections import Counter
from Bio import SeqIO
from Bio.SeqUtils import GC
def compute(inputfile):
for seq_record in SeqIO.parse(inputfile, "fasta"):
cnt = Counter()
line = str(seq_record.seq)
i = 0
for j in range(((len(line))/ 3)):
cnt[line[i:i+3]] += 1
i += 3
print json.dumps(
[{
"id" : seq_record.id,
"ratios" : cnt
}], sort_keys=True, indent=4, separators=(',', ': '))
def main(argv):
inputfile = 'samples/test.fasta'
try:
opts, args = getopt.getopt(argv,"hi:o:",["ifile=","ofile="])
except getopt.GetoptError:
print 'count.py -i <inputfile> -o <outputfile>'
sys.exit(2)
for opt, arg in opts:
if opt == '-h':
print 'count.py -i <inputfile>'
sys.exit()
elif opt in ("-i", "--ifile"):
inputfile = arg
print 'Input file is', inputfile
compute(inputfile)
if __name__ == "__main__":
main(sys.argv[1:])
| <commit_before>import json
from collections import Counter
from Bio import SeqIO
from Bio.SeqUtils import GC
for seq_record in SeqIO.parse("samples/test.fasta", "fasta"):
cnt = Counter()
line = str(seq_record.seq)
i = 0
for j in range(((len(line))/ 3)):
cnt[line[i:i+3]] += 1
i += 3
print json.dumps(
[{
"id" : seq_record.id,
"ratios" : cnt
}], sort_keys=True, indent=4, separators=(',', ': '))
<commit_msg>Add -i flag for input<commit_after>#!/usr/bin/python
import getopt
import json
import sys
from collections import Counter
from Bio import SeqIO
from Bio.SeqUtils import GC
def compute(inputfile):
for seq_record in SeqIO.parse(inputfile, "fasta"):
cnt = Counter()
line = str(seq_record.seq)
i = 0
for j in range(((len(line))/ 3)):
cnt[line[i:i+3]] += 1
i += 3
print json.dumps(
[{
"id" : seq_record.id,
"ratios" : cnt
}], sort_keys=True, indent=4, separators=(',', ': '))
def main(argv):
inputfile = 'samples/test.fasta'
try:
opts, args = getopt.getopt(argv,"hi:o:",["ifile=","ofile="])
except getopt.GetoptError:
print 'count.py -i <inputfile> -o <outputfile>'
sys.exit(2)
for opt, arg in opts:
if opt == '-h':
print 'count.py -i <inputfile>'
sys.exit()
elif opt in ("-i", "--ifile"):
inputfile = arg
print 'Input file is', inputfile
compute(inputfile)
if __name__ == "__main__":
main(sys.argv[1:])
|
29562b08e436abc8465404e49d9193537721b717 | src/odin/contrib/money/fields.py | src/odin/contrib/money/fields.py | # -*- coding: utf-8 -*-
from __future__ import absolute_import, print_function
from odin import exceptions
from odin.fields import ScalarField
from odin.validators import EMPTY_VALUES
from .datatypes import Amount
__all__ = ('AmountField', )
class AmountField(ScalarField):
"""
Field that contains a monetary amount (with an optional currency).
"""
default_error_messages = {
'invalid': "'%s' value must be a (amount, currency).",
'invalid_currency': "'%s' currency is not supported.",
}
data_type_name = "Amount"
def __init__(self, allowed_currencies=None, **kwargs):
super(AmountField, self).__init__(**kwargs)
self.allowed_currencies = allowed_currencies
def to_python(self, value):
if value in EMPTY_VALUES:
return
if isinstance(value, Amount):
return value
try:
return Amount(value)
except (ValueError, TypeError):
msg = self.error_messages['invalid'] % value
raise exceptions.ValidationError(msg)
def validate(self, value):
super(AmountField, self).validate(value)
if self.allowed_currencies and value not in EMPTY_VALUES:
if value.currency not in self.allowed_currencies:
msg = self.error_messages['invalid_currency'] % str(value.currency)
raise exceptions.ValidationError(msg)
def prepare(self, value):
if value in EMPTY_VALUES:
return
return float(value), value.currency.code
| # -*- coding: utf-8 -*-
from __future__ import absolute_import, print_function
from odin import exceptions
from odin.fields import ScalarField
from odin.validators import EMPTY_VALUES
from .datatypes import Amount
__all__ = ("AmountField",)
class AmountField(ScalarField):
"""
Field that contains a monetary amount (with an optional currency).
"""
default_error_messages = {
"invalid": "'%s' value must be a (amount, currency).",
"invalid_currency": "'%s' currency is not supported.",
}
data_type_name = "Amount"
def __init__(self, allowed_currencies=None, **kwargs):
super(AmountField, self).__init__(**kwargs)
self.allowed_currencies = allowed_currencies
def to_python(self, value):
if value in EMPTY_VALUES:
return
if isinstance(value, Amount):
return value
try:
return Amount(value)
except (ValueError, TypeError):
msg = self.error_messages["invalid"] % value
raise exceptions.ValidationError(msg)
def validate(self, value):
super(AmountField, self).validate(value)
if (
self.allowed_currencies
and (value not in EMPTY_VALUES)
and (value.currency not in self.allowed_currencies)
):
msg = self.error_messages["invalid_currency"] % str(value.currency)
raise exceptions.ValidationError(msg)
def prepare(self, value):
if value in EMPTY_VALUES:
return
return float(value), value.currency.code
| Correct issue from Sonar (and black file) | Correct issue from Sonar (and black file)
| Python | bsd-3-clause | python-odin/odin | # -*- coding: utf-8 -*-
from __future__ import absolute_import, print_function
from odin import exceptions
from odin.fields import ScalarField
from odin.validators import EMPTY_VALUES
from .datatypes import Amount
__all__ = ('AmountField', )
class AmountField(ScalarField):
"""
Field that contains a monetary amount (with an optional currency).
"""
default_error_messages = {
'invalid': "'%s' value must be a (amount, currency).",
'invalid_currency': "'%s' currency is not supported.",
}
data_type_name = "Amount"
def __init__(self, allowed_currencies=None, **kwargs):
super(AmountField, self).__init__(**kwargs)
self.allowed_currencies = allowed_currencies
def to_python(self, value):
if value in EMPTY_VALUES:
return
if isinstance(value, Amount):
return value
try:
return Amount(value)
except (ValueError, TypeError):
msg = self.error_messages['invalid'] % value
raise exceptions.ValidationError(msg)
def validate(self, value):
super(AmountField, self).validate(value)
if self.allowed_currencies and value not in EMPTY_VALUES:
if value.currency not in self.allowed_currencies:
msg = self.error_messages['invalid_currency'] % str(value.currency)
raise exceptions.ValidationError(msg)
def prepare(self, value):
if value in EMPTY_VALUES:
return
return float(value), value.currency.code
Correct issue from Sonar (and black file) | # -*- coding: utf-8 -*-
from __future__ import absolute_import, print_function
from odin import exceptions
from odin.fields import ScalarField
from odin.validators import EMPTY_VALUES
from .datatypes import Amount
__all__ = ("AmountField",)
class AmountField(ScalarField):
"""
Field that contains a monetary amount (with an optional currency).
"""
default_error_messages = {
"invalid": "'%s' value must be a (amount, currency).",
"invalid_currency": "'%s' currency is not supported.",
}
data_type_name = "Amount"
def __init__(self, allowed_currencies=None, **kwargs):
super(AmountField, self).__init__(**kwargs)
self.allowed_currencies = allowed_currencies
def to_python(self, value):
if value in EMPTY_VALUES:
return
if isinstance(value, Amount):
return value
try:
return Amount(value)
except (ValueError, TypeError):
msg = self.error_messages["invalid"] % value
raise exceptions.ValidationError(msg)
def validate(self, value):
super(AmountField, self).validate(value)
if (
self.allowed_currencies
and (value not in EMPTY_VALUES)
and (value.currency not in self.allowed_currencies)
):
msg = self.error_messages["invalid_currency"] % str(value.currency)
raise exceptions.ValidationError(msg)
def prepare(self, value):
if value in EMPTY_VALUES:
return
return float(value), value.currency.code
| <commit_before># -*- coding: utf-8 -*-
from __future__ import absolute_import, print_function
from odin import exceptions
from odin.fields import ScalarField
from odin.validators import EMPTY_VALUES
from .datatypes import Amount
__all__ = ('AmountField', )
class AmountField(ScalarField):
"""
Field that contains a monetary amount (with an optional currency).
"""
default_error_messages = {
'invalid': "'%s' value must be a (amount, currency).",
'invalid_currency': "'%s' currency is not supported.",
}
data_type_name = "Amount"
def __init__(self, allowed_currencies=None, **kwargs):
super(AmountField, self).__init__(**kwargs)
self.allowed_currencies = allowed_currencies
def to_python(self, value):
if value in EMPTY_VALUES:
return
if isinstance(value, Amount):
return value
try:
return Amount(value)
except (ValueError, TypeError):
msg = self.error_messages['invalid'] % value
raise exceptions.ValidationError(msg)
def validate(self, value):
super(AmountField, self).validate(value)
if self.allowed_currencies and value not in EMPTY_VALUES:
if value.currency not in self.allowed_currencies:
msg = self.error_messages['invalid_currency'] % str(value.currency)
raise exceptions.ValidationError(msg)
def prepare(self, value):
if value in EMPTY_VALUES:
return
return float(value), value.currency.code
<commit_msg>Correct issue from Sonar (and black file)<commit_after> | # -*- coding: utf-8 -*-
from __future__ import absolute_import, print_function
from odin import exceptions
from odin.fields import ScalarField
from odin.validators import EMPTY_VALUES
from .datatypes import Amount
__all__ = ("AmountField",)
class AmountField(ScalarField):
"""
Field that contains a monetary amount (with an optional currency).
"""
default_error_messages = {
"invalid": "'%s' value must be a (amount, currency).",
"invalid_currency": "'%s' currency is not supported.",
}
data_type_name = "Amount"
def __init__(self, allowed_currencies=None, **kwargs):
super(AmountField, self).__init__(**kwargs)
self.allowed_currencies = allowed_currencies
def to_python(self, value):
if value in EMPTY_VALUES:
return
if isinstance(value, Amount):
return value
try:
return Amount(value)
except (ValueError, TypeError):
msg = self.error_messages["invalid"] % value
raise exceptions.ValidationError(msg)
def validate(self, value):
super(AmountField, self).validate(value)
if (
self.allowed_currencies
and (value not in EMPTY_VALUES)
and (value.currency not in self.allowed_currencies)
):
msg = self.error_messages["invalid_currency"] % str(value.currency)
raise exceptions.ValidationError(msg)
def prepare(self, value):
if value in EMPTY_VALUES:
return
return float(value), value.currency.code
| # -*- coding: utf-8 -*-
from __future__ import absolute_import, print_function
from odin import exceptions
from odin.fields import ScalarField
from odin.validators import EMPTY_VALUES
from .datatypes import Amount
__all__ = ('AmountField', )
class AmountField(ScalarField):
"""
Field that contains a monetary amount (with an optional currency).
"""
default_error_messages = {
'invalid': "'%s' value must be a (amount, currency).",
'invalid_currency': "'%s' currency is not supported.",
}
data_type_name = "Amount"
def __init__(self, allowed_currencies=None, **kwargs):
super(AmountField, self).__init__(**kwargs)
self.allowed_currencies = allowed_currencies
def to_python(self, value):
if value in EMPTY_VALUES:
return
if isinstance(value, Amount):
return value
try:
return Amount(value)
except (ValueError, TypeError):
msg = self.error_messages['invalid'] % value
raise exceptions.ValidationError(msg)
def validate(self, value):
super(AmountField, self).validate(value)
if self.allowed_currencies and value not in EMPTY_VALUES:
if value.currency not in self.allowed_currencies:
msg = self.error_messages['invalid_currency'] % str(value.currency)
raise exceptions.ValidationError(msg)
def prepare(self, value):
if value in EMPTY_VALUES:
return
return float(value), value.currency.code
Correct issue from Sonar (and black file)# -*- coding: utf-8 -*-
from __future__ import absolute_import, print_function
from odin import exceptions
from odin.fields import ScalarField
from odin.validators import EMPTY_VALUES
from .datatypes import Amount
__all__ = ("AmountField",)
class AmountField(ScalarField):
"""
Field that contains a monetary amount (with an optional currency).
"""
default_error_messages = {
"invalid": "'%s' value must be a (amount, currency).",
"invalid_currency": "'%s' currency is not supported.",
}
data_type_name = "Amount"
def __init__(self, allowed_currencies=None, **kwargs):
super(AmountField, self).__init__(**kwargs)
self.allowed_currencies = allowed_currencies
def to_python(self, value):
if value in EMPTY_VALUES:
return
if isinstance(value, Amount):
return value
try:
return Amount(value)
except (ValueError, TypeError):
msg = self.error_messages["invalid"] % value
raise exceptions.ValidationError(msg)
def validate(self, value):
super(AmountField, self).validate(value)
if (
self.allowed_currencies
and (value not in EMPTY_VALUES)
and (value.currency not in self.allowed_currencies)
):
msg = self.error_messages["invalid_currency"] % str(value.currency)
raise exceptions.ValidationError(msg)
def prepare(self, value):
if value in EMPTY_VALUES:
return
return float(value), value.currency.code
| <commit_before># -*- coding: utf-8 -*-
from __future__ import absolute_import, print_function
from odin import exceptions
from odin.fields import ScalarField
from odin.validators import EMPTY_VALUES
from .datatypes import Amount
__all__ = ('AmountField', )
class AmountField(ScalarField):
"""
Field that contains a monetary amount (with an optional currency).
"""
default_error_messages = {
'invalid': "'%s' value must be a (amount, currency).",
'invalid_currency': "'%s' currency is not supported.",
}
data_type_name = "Amount"
def __init__(self, allowed_currencies=None, **kwargs):
super(AmountField, self).__init__(**kwargs)
self.allowed_currencies = allowed_currencies
def to_python(self, value):
if value in EMPTY_VALUES:
return
if isinstance(value, Amount):
return value
try:
return Amount(value)
except (ValueError, TypeError):
msg = self.error_messages['invalid'] % value
raise exceptions.ValidationError(msg)
def validate(self, value):
super(AmountField, self).validate(value)
if self.allowed_currencies and value not in EMPTY_VALUES:
if value.currency not in self.allowed_currencies:
msg = self.error_messages['invalid_currency'] % str(value.currency)
raise exceptions.ValidationError(msg)
def prepare(self, value):
if value in EMPTY_VALUES:
return
return float(value), value.currency.code
<commit_msg>Correct issue from Sonar (and black file)<commit_after># -*- coding: utf-8 -*-
from __future__ import absolute_import, print_function
from odin import exceptions
from odin.fields import ScalarField
from odin.validators import EMPTY_VALUES
from .datatypes import Amount
__all__ = ("AmountField",)
class AmountField(ScalarField):
"""
Field that contains a monetary amount (with an optional currency).
"""
default_error_messages = {
"invalid": "'%s' value must be a (amount, currency).",
"invalid_currency": "'%s' currency is not supported.",
}
data_type_name = "Amount"
def __init__(self, allowed_currencies=None, **kwargs):
super(AmountField, self).__init__(**kwargs)
self.allowed_currencies = allowed_currencies
def to_python(self, value):
if value in EMPTY_VALUES:
return
if isinstance(value, Amount):
return value
try:
return Amount(value)
except (ValueError, TypeError):
msg = self.error_messages["invalid"] % value
raise exceptions.ValidationError(msg)
def validate(self, value):
super(AmountField, self).validate(value)
if (
self.allowed_currencies
and (value not in EMPTY_VALUES)
and (value.currency not in self.allowed_currencies)
):
msg = self.error_messages["invalid_currency"] % str(value.currency)
raise exceptions.ValidationError(msg)
def prepare(self, value):
if value in EMPTY_VALUES:
return
return float(value), value.currency.code
|
6b93f6a6bedf875d4bad1af2493c91b28a625ea9 | chempy/electrochemistry/tests/test_nernst.py | chempy/electrochemistry/tests/test_nernst.py | # -*- coding: utf-8 -*-
from __future__ import (absolute_import, division, print_function)
from ..nernst import nernst_potential
def test_nernst_potential():
# Sodium in cells
assert abs(1000 * nernst_potential(145, 15, 1, 310) - 60.605) < 1e-4
# Potassium in cells
assert abs(1000 * nernst_potential(4, 150, 1, 310) - (-96.8196)) < 1e-4
| # -*- coding: utf-8 -*-
from __future__ import (absolute_import, division, print_function)
from ..nernst import nernst_potential
from chempy.util.testing import requires
from chempy.units import default_units, default_constants, units_library
def test_nernst_potential():
"""
Test cases obtained from textbook examples of Nernst potential in cellular
membranes. 310K = 37C, typical mammalian cell environment temperature.
"""
# Sodium in cells
assert abs(1000 * nernst_potential(145, 15, 1, 310) - 60.605) < 1e-4
# Potassium in cells
assert abs(1000 * nernst_potential(4, 150, 1, 310) - (-96.8196)) < 1e-4
# Calcium in cells
assert abs(1000 * nernst_potential(2, 7e-5, 2, 310) - 137.0436) < 1e-4
# Chloride in cells
assert abs(1000 * nernst_potential(110, 10, -1, 310) - (-64.0567)) < 1e-4
@requires(units_library)
def test_nernst_potential_units():
v = nernst_potential(145, 15, 1, 310, default_constants, default_units)
assert (1000 * v - 60.605) < 1e-4
| Add additional testing to electrochemistry/Nernst | Add additional testing to electrochemistry/Nernst
| Python | bsd-2-clause | bjodah/aqchem,bjodah/aqchem,bjodah/chempy,bjodah/chempy,bjodah/aqchem | # -*- coding: utf-8 -*-
from __future__ import (absolute_import, division, print_function)
from ..nernst import nernst_potential
def test_nernst_potential():
# Sodium in cells
assert abs(1000 * nernst_potential(145, 15, 1, 310) - 60.605) < 1e-4
# Potassium in cells
assert abs(1000 * nernst_potential(4, 150, 1, 310) - (-96.8196)) < 1e-4
Add additional testing to electrochemistry/Nernst | # -*- coding: utf-8 -*-
from __future__ import (absolute_import, division, print_function)
from ..nernst import nernst_potential
from chempy.util.testing import requires
from chempy.units import default_units, default_constants, units_library
def test_nernst_potential():
"""
Test cases obtained from textbook examples of Nernst potential in cellular
membranes. 310K = 37C, typical mammalian cell environment temperature.
"""
# Sodium in cells
assert abs(1000 * nernst_potential(145, 15, 1, 310) - 60.605) < 1e-4
# Potassium in cells
assert abs(1000 * nernst_potential(4, 150, 1, 310) - (-96.8196)) < 1e-4
# Calcium in cells
assert abs(1000 * nernst_potential(2, 7e-5, 2, 310) - 137.0436) < 1e-4
# Chloride in cells
assert abs(1000 * nernst_potential(110, 10, -1, 310) - (-64.0567)) < 1e-4
@requires(units_library)
def test_nernst_potential_units():
v = nernst_potential(145, 15, 1, 310, default_constants, default_units)
assert (1000 * v - 60.605) < 1e-4
| <commit_before># -*- coding: utf-8 -*-
from __future__ import (absolute_import, division, print_function)
from ..nernst import nernst_potential
def test_nernst_potential():
# Sodium in cells
assert abs(1000 * nernst_potential(145, 15, 1, 310) - 60.605) < 1e-4
# Potassium in cells
assert abs(1000 * nernst_potential(4, 150, 1, 310) - (-96.8196)) < 1e-4
<commit_msg>Add additional testing to electrochemistry/Nernst<commit_after> | # -*- coding: utf-8 -*-
from __future__ import (absolute_import, division, print_function)
from ..nernst import nernst_potential
from chempy.util.testing import requires
from chempy.units import default_units, default_constants, units_library
def test_nernst_potential():
"""
Test cases obtained from textbook examples of Nernst potential in cellular
membranes. 310K = 37C, typical mammalian cell environment temperature.
"""
# Sodium in cells
assert abs(1000 * nernst_potential(145, 15, 1, 310) - 60.605) < 1e-4
# Potassium in cells
assert abs(1000 * nernst_potential(4, 150, 1, 310) - (-96.8196)) < 1e-4
# Calcium in cells
assert abs(1000 * nernst_potential(2, 7e-5, 2, 310) - 137.0436) < 1e-4
# Chloride in cells
assert abs(1000 * nernst_potential(110, 10, -1, 310) - (-64.0567)) < 1e-4
@requires(units_library)
def test_nernst_potential_units():
v = nernst_potential(145, 15, 1, 310, default_constants, default_units)
assert (1000 * v - 60.605) < 1e-4
| # -*- coding: utf-8 -*-
from __future__ import (absolute_import, division, print_function)
from ..nernst import nernst_potential
def test_nernst_potential():
# Sodium in cells
assert abs(1000 * nernst_potential(145, 15, 1, 310) - 60.605) < 1e-4
# Potassium in cells
assert abs(1000 * nernst_potential(4, 150, 1, 310) - (-96.8196)) < 1e-4
Add additional testing to electrochemistry/Nernst# -*- coding: utf-8 -*-
from __future__ import (absolute_import, division, print_function)
from ..nernst import nernst_potential
from chempy.util.testing import requires
from chempy.units import default_units, default_constants, units_library
def test_nernst_potential():
"""
Test cases obtained from textbook examples of Nernst potential in cellular
membranes. 310K = 37C, typical mammalian cell environment temperature.
"""
# Sodium in cells
assert abs(1000 * nernst_potential(145, 15, 1, 310) - 60.605) < 1e-4
# Potassium in cells
assert abs(1000 * nernst_potential(4, 150, 1, 310) - (-96.8196)) < 1e-4
# Calcium in cells
assert abs(1000 * nernst_potential(2, 7e-5, 2, 310) - 137.0436) < 1e-4
# Chloride in cells
assert abs(1000 * nernst_potential(110, 10, -1, 310) - (-64.0567)) < 1e-4
@requires(units_library)
def test_nernst_potential_units():
v = nernst_potential(145, 15, 1, 310, default_constants, default_units)
assert (1000 * v - 60.605) < 1e-4
| <commit_before># -*- coding: utf-8 -*-
from __future__ import (absolute_import, division, print_function)
from ..nernst import nernst_potential
def test_nernst_potential():
# Sodium in cells
assert abs(1000 * nernst_potential(145, 15, 1, 310) - 60.605) < 1e-4
# Potassium in cells
assert abs(1000 * nernst_potential(4, 150, 1, 310) - (-96.8196)) < 1e-4
<commit_msg>Add additional testing to electrochemistry/Nernst<commit_after># -*- coding: utf-8 -*-
from __future__ import (absolute_import, division, print_function)
from ..nernst import nernst_potential
from chempy.util.testing import requires
from chempy.units import default_units, default_constants, units_library
def test_nernst_potential():
"""
Test cases obtained from textbook examples of Nernst potential in cellular
membranes. 310K = 37C, typical mammalian cell environment temperature.
"""
# Sodium in cells
assert abs(1000 * nernst_potential(145, 15, 1, 310) - 60.605) < 1e-4
# Potassium in cells
assert abs(1000 * nernst_potential(4, 150, 1, 310) - (-96.8196)) < 1e-4
# Calcium in cells
assert abs(1000 * nernst_potential(2, 7e-5, 2, 310) - 137.0436) < 1e-4
# Chloride in cells
assert abs(1000 * nernst_potential(110, 10, -1, 310) - (-64.0567)) < 1e-4
@requires(units_library)
def test_nernst_potential_units():
v = nernst_potential(145, 15, 1, 310, default_constants, default_units)
assert (1000 * v - 60.605) < 1e-4
|
2735d4a7b1ae0af0d58ea0accb973ab11477783e | django_hosts/tests/urls/simple.py | django_hosts/tests/urls/simple.py | from django.conf.urls.defaults import patterns, url
from django.views.generic import TemplateView
urlpatterns = patterns('django.views.generic.simple',
url(r'^simple/$', TemplateView.as_view(), name='simple-direct'),
)
| from django.conf.urls.defaults import patterns, url
urlpatterns = patterns('django.views.generic.simple',
url(r'^simple/$', 'django.shortcuts.render', name='simple-direct'),
)
| Use django.shortcuts.render instead of TemplateView | Use django.shortcuts.render instead of TemplateView | Python | bsd-3-clause | jezdez/django-hosts | from django.conf.urls.defaults import patterns, url
from django.views.generic import TemplateView
urlpatterns = patterns('django.views.generic.simple',
url(r'^simple/$', TemplateView.as_view(), name='simple-direct'),
)
Use django.shortcuts.render instead of TemplateView | from django.conf.urls.defaults import patterns, url
urlpatterns = patterns('django.views.generic.simple',
url(r'^simple/$', 'django.shortcuts.render', name='simple-direct'),
)
| <commit_before>from django.conf.urls.defaults import patterns, url
from django.views.generic import TemplateView
urlpatterns = patterns('django.views.generic.simple',
url(r'^simple/$', TemplateView.as_view(), name='simple-direct'),
)
<commit_msg>Use django.shortcuts.render instead of TemplateView<commit_after> | from django.conf.urls.defaults import patterns, url
urlpatterns = patterns('django.views.generic.simple',
url(r'^simple/$', 'django.shortcuts.render', name='simple-direct'),
)
| from django.conf.urls.defaults import patterns, url
from django.views.generic import TemplateView
urlpatterns = patterns('django.views.generic.simple',
url(r'^simple/$', TemplateView.as_view(), name='simple-direct'),
)
Use django.shortcuts.render instead of TemplateViewfrom django.conf.urls.defaults import patterns, url
urlpatterns = patterns('django.views.generic.simple',
url(r'^simple/$', 'django.shortcuts.render', name='simple-direct'),
)
| <commit_before>from django.conf.urls.defaults import patterns, url
from django.views.generic import TemplateView
urlpatterns = patterns('django.views.generic.simple',
url(r'^simple/$', TemplateView.as_view(), name='simple-direct'),
)
<commit_msg>Use django.shortcuts.render instead of TemplateView<commit_after>from django.conf.urls.defaults import patterns, url
urlpatterns = patterns('django.views.generic.simple',
url(r'^simple/$', 'django.shortcuts.render', name='simple-direct'),
)
|
cb533d74f2e634d9e8c7d0515307fad107da36ef | {{cookiecutter.repo_name}}/{{cookiecutter.repo_name}}/apps/site/urls.py | {{cookiecutter.repo_name}}/{{cookiecutter.repo_name}}/apps/site/urls.py | from django.conf import settings
from django.conf.urls import url
from django.http import HttpResponse
from django.views import generic
from . import views
urlpatterns = (
url(r'^robots\.txt$', views.RobotsView.as_view(), name="robots"),
)
if getattr(settings, 'LOADERIO_TOKEN', None):
loaderio_token = settings.LOADERIO_TOKEN
class LoaderioTokenView(generic.View):
def get(self, request, *args, **kwargs):
return HttpResponse('{}'.format(loaderio_token))
urlpatterns += (
url(r'^{}/$'.format(loaderio_token), LoaderioTokenView.as_view()),
)
| from django.conf import settings
from django.conf.urls import url
from django.http import HttpResponse
from django.views import generic
from . import views
urlpatterns = (
url(r'^robots\.txt$', views.RobotsView.as_view(), name="robots"),
)
if getattr(settings, 'LOADERIO_TOKEN', None):
loaderio_token = settings.LOADERIO_TOKEN
class LoaderioTokenView(generic.View):
def get(self, request, *args, **kwargs):
return HttpResponse('{}'.format(loaderio_token))
urlpatterns += (
url(r'^{}/$'.format(loaderio_token), LoaderioTokenView.as_view()),
)
| Add a missing line between imports | Add a missing line between imports
| Python | mit | dulaccc/cookiecutter-django-herokuapp,dulaccc/cookiecutter-django-herokuapp,dulaccc/cookiecutter-django-herokuapp,dulaccc/cookiecutter-django-herokuapp | from django.conf import settings
from django.conf.urls import url
from django.http import HttpResponse
from django.views import generic
from . import views
urlpatterns = (
url(r'^robots\.txt$', views.RobotsView.as_view(), name="robots"),
)
if getattr(settings, 'LOADERIO_TOKEN', None):
loaderio_token = settings.LOADERIO_TOKEN
class LoaderioTokenView(generic.View):
def get(self, request, *args, **kwargs):
return HttpResponse('{}'.format(loaderio_token))
urlpatterns += (
url(r'^{}/$'.format(loaderio_token), LoaderioTokenView.as_view()),
)
Add a missing line between imports | from django.conf import settings
from django.conf.urls import url
from django.http import HttpResponse
from django.views import generic
from . import views
urlpatterns = (
url(r'^robots\.txt$', views.RobotsView.as_view(), name="robots"),
)
if getattr(settings, 'LOADERIO_TOKEN', None):
loaderio_token = settings.LOADERIO_TOKEN
class LoaderioTokenView(generic.View):
def get(self, request, *args, **kwargs):
return HttpResponse('{}'.format(loaderio_token))
urlpatterns += (
url(r'^{}/$'.format(loaderio_token), LoaderioTokenView.as_view()),
)
| <commit_before>from django.conf import settings
from django.conf.urls import url
from django.http import HttpResponse
from django.views import generic
from . import views
urlpatterns = (
url(r'^robots\.txt$', views.RobotsView.as_view(), name="robots"),
)
if getattr(settings, 'LOADERIO_TOKEN', None):
loaderio_token = settings.LOADERIO_TOKEN
class LoaderioTokenView(generic.View):
def get(self, request, *args, **kwargs):
return HttpResponse('{}'.format(loaderio_token))
urlpatterns += (
url(r'^{}/$'.format(loaderio_token), LoaderioTokenView.as_view()),
)
<commit_msg>Add a missing line between imports<commit_after> | from django.conf import settings
from django.conf.urls import url
from django.http import HttpResponse
from django.views import generic
from . import views
urlpatterns = (
url(r'^robots\.txt$', views.RobotsView.as_view(), name="robots"),
)
if getattr(settings, 'LOADERIO_TOKEN', None):
loaderio_token = settings.LOADERIO_TOKEN
class LoaderioTokenView(generic.View):
def get(self, request, *args, **kwargs):
return HttpResponse('{}'.format(loaderio_token))
urlpatterns += (
url(r'^{}/$'.format(loaderio_token), LoaderioTokenView.as_view()),
)
| from django.conf import settings
from django.conf.urls import url
from django.http import HttpResponse
from django.views import generic
from . import views
urlpatterns = (
url(r'^robots\.txt$', views.RobotsView.as_view(), name="robots"),
)
if getattr(settings, 'LOADERIO_TOKEN', None):
loaderio_token = settings.LOADERIO_TOKEN
class LoaderioTokenView(generic.View):
def get(self, request, *args, **kwargs):
return HttpResponse('{}'.format(loaderio_token))
urlpatterns += (
url(r'^{}/$'.format(loaderio_token), LoaderioTokenView.as_view()),
)
Add a missing line between importsfrom django.conf import settings
from django.conf.urls import url
from django.http import HttpResponse
from django.views import generic
from . import views
urlpatterns = (
url(r'^robots\.txt$', views.RobotsView.as_view(), name="robots"),
)
if getattr(settings, 'LOADERIO_TOKEN', None):
loaderio_token = settings.LOADERIO_TOKEN
class LoaderioTokenView(generic.View):
def get(self, request, *args, **kwargs):
return HttpResponse('{}'.format(loaderio_token))
urlpatterns += (
url(r'^{}/$'.format(loaderio_token), LoaderioTokenView.as_view()),
)
| <commit_before>from django.conf import settings
from django.conf.urls import url
from django.http import HttpResponse
from django.views import generic
from . import views
urlpatterns = (
url(r'^robots\.txt$', views.RobotsView.as_view(), name="robots"),
)
if getattr(settings, 'LOADERIO_TOKEN', None):
loaderio_token = settings.LOADERIO_TOKEN
class LoaderioTokenView(generic.View):
def get(self, request, *args, **kwargs):
return HttpResponse('{}'.format(loaderio_token))
urlpatterns += (
url(r'^{}/$'.format(loaderio_token), LoaderioTokenView.as_view()),
)
<commit_msg>Add a missing line between imports<commit_after>from django.conf import settings
from django.conf.urls import url
from django.http import HttpResponse
from django.views import generic
from . import views
urlpatterns = (
url(r'^robots\.txt$', views.RobotsView.as_view(), name="robots"),
)
if getattr(settings, 'LOADERIO_TOKEN', None):
loaderio_token = settings.LOADERIO_TOKEN
class LoaderioTokenView(generic.View):
def get(self, request, *args, **kwargs):
return HttpResponse('{}'.format(loaderio_token))
urlpatterns += (
url(r'^{}/$'.format(loaderio_token), LoaderioTokenView.as_view()),
)
|
f0cf4b51987befb25f605ee421554da7615c9472 | readthedocs/builds/admin.py | readthedocs/builds/admin.py | """Django admin interface for `~builds.models.Build` and related models.
"""
from django.contrib import admin
from readthedocs.builds.models import Build, VersionAlias, Version, BuildCommandResult
from guardian.admin import GuardedModelAdmin
class BuildCommandResultInline(admin.TabularInline):
model = BuildCommandResult
fields = ('command', 'exit_code')
class BuildAdmin(admin.ModelAdmin):
fields = ('project', 'version', 'type', 'state', 'success', 'length')
list_display = ('project', 'success', 'type', 'state')
raw_id_fields = ('project', 'version')
inlines = (BuildCommandResultInline,)
class VersionAdmin(GuardedModelAdmin):
search_fields = ('slug', 'project__name')
list_filter = ('project', 'privacy_level')
admin.site.register(Build, BuildAdmin)
admin.site.register(VersionAlias)
admin.site.register(Version, VersionAdmin)
| """Django admin interface for `~builds.models.Build` and related models.
"""
from django.contrib import admin
from readthedocs.builds.models import Build, VersionAlias, Version, BuildCommandResult
from guardian.admin import GuardedModelAdmin
class BuildCommandResultInline(admin.TabularInline):
model = BuildCommandResult
fields = ('command', 'exit_code')
class BuildAdmin(admin.ModelAdmin):
fields = ('project', 'version', 'type', 'state', 'success', 'length')
list_display = ('project', 'success', 'type', 'state', 'date')
raw_id_fields = ('project', 'version')
inlines = (BuildCommandResultInline,)
class VersionAdmin(GuardedModelAdmin):
search_fields = ('slug', 'project__name')
list_filter = ('project', 'privacy_level')
admin.site.register(Build, BuildAdmin)
admin.site.register(VersionAlias)
admin.site.register(Version, VersionAdmin)
| Add date back to build display | Add date back to build display
| Python | mit | davidfischer/readthedocs.org,safwanrahman/readthedocs.org,espdev/readthedocs.org,safwanrahman/readthedocs.org,rtfd/readthedocs.org,tddv/readthedocs.org,davidfischer/readthedocs.org,espdev/readthedocs.org,rtfd/readthedocs.org,pombredanne/readthedocs.org,espdev/readthedocs.org,espdev/readthedocs.org,davidfischer/readthedocs.org,espdev/readthedocs.org,pombredanne/readthedocs.org,davidfischer/readthedocs.org,tddv/readthedocs.org,safwanrahman/readthedocs.org,safwanrahman/readthedocs.org,rtfd/readthedocs.org,tddv/readthedocs.org,pombredanne/readthedocs.org,rtfd/readthedocs.org | """Django admin interface for `~builds.models.Build` and related models.
"""
from django.contrib import admin
from readthedocs.builds.models import Build, VersionAlias, Version, BuildCommandResult
from guardian.admin import GuardedModelAdmin
class BuildCommandResultInline(admin.TabularInline):
model = BuildCommandResult
fields = ('command', 'exit_code')
class BuildAdmin(admin.ModelAdmin):
fields = ('project', 'version', 'type', 'state', 'success', 'length')
list_display = ('project', 'success', 'type', 'state')
raw_id_fields = ('project', 'version')
inlines = (BuildCommandResultInline,)
class VersionAdmin(GuardedModelAdmin):
search_fields = ('slug', 'project__name')
list_filter = ('project', 'privacy_level')
admin.site.register(Build, BuildAdmin)
admin.site.register(VersionAlias)
admin.site.register(Version, VersionAdmin)
Add date back to build display | """Django admin interface for `~builds.models.Build` and related models.
"""
from django.contrib import admin
from readthedocs.builds.models import Build, VersionAlias, Version, BuildCommandResult
from guardian.admin import GuardedModelAdmin
class BuildCommandResultInline(admin.TabularInline):
model = BuildCommandResult
fields = ('command', 'exit_code')
class BuildAdmin(admin.ModelAdmin):
fields = ('project', 'version', 'type', 'state', 'success', 'length')
list_display = ('project', 'success', 'type', 'state', 'date')
raw_id_fields = ('project', 'version')
inlines = (BuildCommandResultInline,)
class VersionAdmin(GuardedModelAdmin):
search_fields = ('slug', 'project__name')
list_filter = ('project', 'privacy_level')
admin.site.register(Build, BuildAdmin)
admin.site.register(VersionAlias)
admin.site.register(Version, VersionAdmin)
| <commit_before>"""Django admin interface for `~builds.models.Build` and related models.
"""
from django.contrib import admin
from readthedocs.builds.models import Build, VersionAlias, Version, BuildCommandResult
from guardian.admin import GuardedModelAdmin
class BuildCommandResultInline(admin.TabularInline):
model = BuildCommandResult
fields = ('command', 'exit_code')
class BuildAdmin(admin.ModelAdmin):
fields = ('project', 'version', 'type', 'state', 'success', 'length')
list_display = ('project', 'success', 'type', 'state')
raw_id_fields = ('project', 'version')
inlines = (BuildCommandResultInline,)
class VersionAdmin(GuardedModelAdmin):
search_fields = ('slug', 'project__name')
list_filter = ('project', 'privacy_level')
admin.site.register(Build, BuildAdmin)
admin.site.register(VersionAlias)
admin.site.register(Version, VersionAdmin)
<commit_msg>Add date back to build display<commit_after> | """Django admin interface for `~builds.models.Build` and related models.
"""
from django.contrib import admin
from readthedocs.builds.models import Build, VersionAlias, Version, BuildCommandResult
from guardian.admin import GuardedModelAdmin
class BuildCommandResultInline(admin.TabularInline):
model = BuildCommandResult
fields = ('command', 'exit_code')
class BuildAdmin(admin.ModelAdmin):
fields = ('project', 'version', 'type', 'state', 'success', 'length')
list_display = ('project', 'success', 'type', 'state', 'date')
raw_id_fields = ('project', 'version')
inlines = (BuildCommandResultInline,)
class VersionAdmin(GuardedModelAdmin):
search_fields = ('slug', 'project__name')
list_filter = ('project', 'privacy_level')
admin.site.register(Build, BuildAdmin)
admin.site.register(VersionAlias)
admin.site.register(Version, VersionAdmin)
| """Django admin interface for `~builds.models.Build` and related models.
"""
from django.contrib import admin
from readthedocs.builds.models import Build, VersionAlias, Version, BuildCommandResult
from guardian.admin import GuardedModelAdmin
class BuildCommandResultInline(admin.TabularInline):
model = BuildCommandResult
fields = ('command', 'exit_code')
class BuildAdmin(admin.ModelAdmin):
fields = ('project', 'version', 'type', 'state', 'success', 'length')
list_display = ('project', 'success', 'type', 'state')
raw_id_fields = ('project', 'version')
inlines = (BuildCommandResultInline,)
class VersionAdmin(GuardedModelAdmin):
search_fields = ('slug', 'project__name')
list_filter = ('project', 'privacy_level')
admin.site.register(Build, BuildAdmin)
admin.site.register(VersionAlias)
admin.site.register(Version, VersionAdmin)
Add date back to build display"""Django admin interface for `~builds.models.Build` and related models.
"""
from django.contrib import admin
from readthedocs.builds.models import Build, VersionAlias, Version, BuildCommandResult
from guardian.admin import GuardedModelAdmin
class BuildCommandResultInline(admin.TabularInline):
model = BuildCommandResult
fields = ('command', 'exit_code')
class BuildAdmin(admin.ModelAdmin):
fields = ('project', 'version', 'type', 'state', 'success', 'length')
list_display = ('project', 'success', 'type', 'state', 'date')
raw_id_fields = ('project', 'version')
inlines = (BuildCommandResultInline,)
class VersionAdmin(GuardedModelAdmin):
search_fields = ('slug', 'project__name')
list_filter = ('project', 'privacy_level')
admin.site.register(Build, BuildAdmin)
admin.site.register(VersionAlias)
admin.site.register(Version, VersionAdmin)
| <commit_before>"""Django admin interface for `~builds.models.Build` and related models.
"""
from django.contrib import admin
from readthedocs.builds.models import Build, VersionAlias, Version, BuildCommandResult
from guardian.admin import GuardedModelAdmin
class BuildCommandResultInline(admin.TabularInline):
model = BuildCommandResult
fields = ('command', 'exit_code')
class BuildAdmin(admin.ModelAdmin):
fields = ('project', 'version', 'type', 'state', 'success', 'length')
list_display = ('project', 'success', 'type', 'state')
raw_id_fields = ('project', 'version')
inlines = (BuildCommandResultInline,)
class VersionAdmin(GuardedModelAdmin):
search_fields = ('slug', 'project__name')
list_filter = ('project', 'privacy_level')
admin.site.register(Build, BuildAdmin)
admin.site.register(VersionAlias)
admin.site.register(Version, VersionAdmin)
<commit_msg>Add date back to build display<commit_after>"""Django admin interface for `~builds.models.Build` and related models.
"""
from django.contrib import admin
from readthedocs.builds.models import Build, VersionAlias, Version, BuildCommandResult
from guardian.admin import GuardedModelAdmin
class BuildCommandResultInline(admin.TabularInline):
model = BuildCommandResult
fields = ('command', 'exit_code')
class BuildAdmin(admin.ModelAdmin):
fields = ('project', 'version', 'type', 'state', 'success', 'length')
list_display = ('project', 'success', 'type', 'state', 'date')
raw_id_fields = ('project', 'version')
inlines = (BuildCommandResultInline,)
class VersionAdmin(GuardedModelAdmin):
search_fields = ('slug', 'project__name')
list_filter = ('project', 'privacy_level')
admin.site.register(Build, BuildAdmin)
admin.site.register(VersionAlias)
admin.site.register(Version, VersionAdmin)
|
0ca07405b864f761ae1d7ed659cac67c799bf39a | src/core/queue.py | src/core/queue.py | # -*- coding: utf-8 -*-
"""ICU (LEGO Island Configuration Utility).
Created 2015 Triangle717
<http://le717.github.io/>
Licensed under The MIT License
<http://opensource.org/licenses/MIT/>
"""
class ActionsQueue:
def __init__(self):
self.queue = []
# Normal buttons
def btnBrowse(self, val):
print(val)
def actionReset(self, val):
print(val)
def btnRedirect(self, val):
print(val)
# Radio buttons
def radioColor256(self, val):
print(val)
def radioColor16b(self, val):
print(val)
def radioModelLow(self, val):
print(val)
def radioModelFast(self, val):
print(val)
def radioModelHigh(self, val):
print(val)
def radioTexFast(self, val):
print(val)
def radioTexHigh(self, val):
print(val)
# Check boxes
def chkCursor(self, val):
print(val)
def chkDraw3D(self, val):
print(val)
def chkFlipVideo(self, val):
print(val)
def chkFullscreen(self, val):
print(val)
def chkJoystick(self, val):
print(val)
def chkMusic(self, val):
print(val)
def chkSound(self, val):
print(val)
# Direct3D dropdown selection
def comboD3D(self, val):
print(val)
| # -*- coding: utf-8 -*-
"""ICU (LEGO Island Configuration Utility).
Created 2015 Triangle717
<http://le717.github.io/>
Licensed under The MIT License
<http://opensource.org/licenses/MIT/>
"""
class ActionsQueue:
def __init__(self):
self.queue = []
class Responses:
def __init__(self):
pass
# Normal buttons
def btnBrowse(self, val):
print(val)
def actionReset(self, val):
print(val)
def btnRedirect(self, val):
print(val)
# Radio buttons
def radioColor256(self, val):
print(val)
def radioColor16b(self, val):
print(val)
def radioModelLow(self, val):
print(val)
def radioModelFast(self, val):
print(val)
def radioModelHigh(self, val):
print(val)
def radioTexFast(self, val):
print(val)
def radioTexHigh(self, val):
print(val)
# Check boxes
def chkCursor(self, val):
print(val)
def chkDraw3D(self, val):
print(val)
def chkFlipVideo(self, val):
print(val)
def chkFullscreen(self, val):
print(val)
def chkJoystick(self, val):
print(val)
def chkMusic(self, val):
print(val)
def chkSound(self, val):
print(val)
# Direct3D dropdown selection
def comboD3D(self, val):
print(val)
| Move those methods to own class | Move those methods to own class [ci skip]
| Python | mit | le717/ICU | # -*- coding: utf-8 -*-
"""ICU (LEGO Island Configuration Utility).
Created 2015 Triangle717
<http://le717.github.io/>
Licensed under The MIT License
<http://opensource.org/licenses/MIT/>
"""
class ActionsQueue:
def __init__(self):
self.queue = []
# Normal buttons
def btnBrowse(self, val):
print(val)
def actionReset(self, val):
print(val)
def btnRedirect(self, val):
print(val)
# Radio buttons
def radioColor256(self, val):
print(val)
def radioColor16b(self, val):
print(val)
def radioModelLow(self, val):
print(val)
def radioModelFast(self, val):
print(val)
def radioModelHigh(self, val):
print(val)
def radioTexFast(self, val):
print(val)
def radioTexHigh(self, val):
print(val)
# Check boxes
def chkCursor(self, val):
print(val)
def chkDraw3D(self, val):
print(val)
def chkFlipVideo(self, val):
print(val)
def chkFullscreen(self, val):
print(val)
def chkJoystick(self, val):
print(val)
def chkMusic(self, val):
print(val)
def chkSound(self, val):
print(val)
# Direct3D dropdown selection
def comboD3D(self, val):
print(val)
Move those methods to own class [ci skip] | # -*- coding: utf-8 -*-
"""ICU (LEGO Island Configuration Utility).
Created 2015 Triangle717
<http://le717.github.io/>
Licensed under The MIT License
<http://opensource.org/licenses/MIT/>
"""
class ActionsQueue:
def __init__(self):
self.queue = []
class Responses:
def __init__(self):
pass
# Normal buttons
def btnBrowse(self, val):
print(val)
def actionReset(self, val):
print(val)
def btnRedirect(self, val):
print(val)
# Radio buttons
def radioColor256(self, val):
print(val)
def radioColor16b(self, val):
print(val)
def radioModelLow(self, val):
print(val)
def radioModelFast(self, val):
print(val)
def radioModelHigh(self, val):
print(val)
def radioTexFast(self, val):
print(val)
def radioTexHigh(self, val):
print(val)
# Check boxes
def chkCursor(self, val):
print(val)
def chkDraw3D(self, val):
print(val)
def chkFlipVideo(self, val):
print(val)
def chkFullscreen(self, val):
print(val)
def chkJoystick(self, val):
print(val)
def chkMusic(self, val):
print(val)
def chkSound(self, val):
print(val)
# Direct3D dropdown selection
def comboD3D(self, val):
print(val)
| <commit_before># -*- coding: utf-8 -*-
"""ICU (LEGO Island Configuration Utility).
Created 2015 Triangle717
<http://le717.github.io/>
Licensed under The MIT License
<http://opensource.org/licenses/MIT/>
"""
class ActionsQueue:
def __init__(self):
self.queue = []
# Normal buttons
def btnBrowse(self, val):
print(val)
def actionReset(self, val):
print(val)
def btnRedirect(self, val):
print(val)
# Radio buttons
def radioColor256(self, val):
print(val)
def radioColor16b(self, val):
print(val)
def radioModelLow(self, val):
print(val)
def radioModelFast(self, val):
print(val)
def radioModelHigh(self, val):
print(val)
def radioTexFast(self, val):
print(val)
def radioTexHigh(self, val):
print(val)
# Check boxes
def chkCursor(self, val):
print(val)
def chkDraw3D(self, val):
print(val)
def chkFlipVideo(self, val):
print(val)
def chkFullscreen(self, val):
print(val)
def chkJoystick(self, val):
print(val)
def chkMusic(self, val):
print(val)
def chkSound(self, val):
print(val)
# Direct3D dropdown selection
def comboD3D(self, val):
print(val)
<commit_msg>Move those methods to own class [ci skip]<commit_after> | # -*- coding: utf-8 -*-
"""ICU (LEGO Island Configuration Utility).
Created 2015 Triangle717
<http://le717.github.io/>
Licensed under The MIT License
<http://opensource.org/licenses/MIT/>
"""
class ActionsQueue:
def __init__(self):
self.queue = []
class Responses:
def __init__(self):
pass
# Normal buttons
def btnBrowse(self, val):
print(val)
def actionReset(self, val):
print(val)
def btnRedirect(self, val):
print(val)
# Radio buttons
def radioColor256(self, val):
print(val)
def radioColor16b(self, val):
print(val)
def radioModelLow(self, val):
print(val)
def radioModelFast(self, val):
print(val)
def radioModelHigh(self, val):
print(val)
def radioTexFast(self, val):
print(val)
def radioTexHigh(self, val):
print(val)
# Check boxes
def chkCursor(self, val):
print(val)
def chkDraw3D(self, val):
print(val)
def chkFlipVideo(self, val):
print(val)
def chkFullscreen(self, val):
print(val)
def chkJoystick(self, val):
print(val)
def chkMusic(self, val):
print(val)
def chkSound(self, val):
print(val)
# Direct3D dropdown selection
def comboD3D(self, val):
print(val)
| # -*- coding: utf-8 -*-
"""ICU (LEGO Island Configuration Utility).
Created 2015 Triangle717
<http://le717.github.io/>
Licensed under The MIT License
<http://opensource.org/licenses/MIT/>
"""
class ActionsQueue:
def __init__(self):
self.queue = []
# Normal buttons
def btnBrowse(self, val):
print(val)
def actionReset(self, val):
print(val)
def btnRedirect(self, val):
print(val)
# Radio buttons
def radioColor256(self, val):
print(val)
def radioColor16b(self, val):
print(val)
def radioModelLow(self, val):
print(val)
def radioModelFast(self, val):
print(val)
def radioModelHigh(self, val):
print(val)
def radioTexFast(self, val):
print(val)
def radioTexHigh(self, val):
print(val)
# Check boxes
def chkCursor(self, val):
print(val)
def chkDraw3D(self, val):
print(val)
def chkFlipVideo(self, val):
print(val)
def chkFullscreen(self, val):
print(val)
def chkJoystick(self, val):
print(val)
def chkMusic(self, val):
print(val)
def chkSound(self, val):
print(val)
# Direct3D dropdown selection
def comboD3D(self, val):
print(val)
Move those methods to own class [ci skip]# -*- coding: utf-8 -*-
"""ICU (LEGO Island Configuration Utility).
Created 2015 Triangle717
<http://le717.github.io/>
Licensed under The MIT License
<http://opensource.org/licenses/MIT/>
"""
class ActionsQueue:
def __init__(self):
self.queue = []
class Responses:
def __init__(self):
pass
# Normal buttons
def btnBrowse(self, val):
print(val)
def actionReset(self, val):
print(val)
def btnRedirect(self, val):
print(val)
# Radio buttons
def radioColor256(self, val):
print(val)
def radioColor16b(self, val):
print(val)
def radioModelLow(self, val):
print(val)
def radioModelFast(self, val):
print(val)
def radioModelHigh(self, val):
print(val)
def radioTexFast(self, val):
print(val)
def radioTexHigh(self, val):
print(val)
# Check boxes
def chkCursor(self, val):
print(val)
def chkDraw3D(self, val):
print(val)
def chkFlipVideo(self, val):
print(val)
def chkFullscreen(self, val):
print(val)
def chkJoystick(self, val):
print(val)
def chkMusic(self, val):
print(val)
def chkSound(self, val):
print(val)
# Direct3D dropdown selection
def comboD3D(self, val):
print(val)
| <commit_before># -*- coding: utf-8 -*-
"""ICU (LEGO Island Configuration Utility).
Created 2015 Triangle717
<http://le717.github.io/>
Licensed under The MIT License
<http://opensource.org/licenses/MIT/>
"""
class ActionsQueue:
def __init__(self):
self.queue = []
# Normal buttons
def btnBrowse(self, val):
print(val)
def actionReset(self, val):
print(val)
def btnRedirect(self, val):
print(val)
# Radio buttons
def radioColor256(self, val):
print(val)
def radioColor16b(self, val):
print(val)
def radioModelLow(self, val):
print(val)
def radioModelFast(self, val):
print(val)
def radioModelHigh(self, val):
print(val)
def radioTexFast(self, val):
print(val)
def radioTexHigh(self, val):
print(val)
# Check boxes
def chkCursor(self, val):
print(val)
def chkDraw3D(self, val):
print(val)
def chkFlipVideo(self, val):
print(val)
def chkFullscreen(self, val):
print(val)
def chkJoystick(self, val):
print(val)
def chkMusic(self, val):
print(val)
def chkSound(self, val):
print(val)
# Direct3D dropdown selection
def comboD3D(self, val):
print(val)
<commit_msg>Move those methods to own class [ci skip]<commit_after># -*- coding: utf-8 -*-
"""ICU (LEGO Island Configuration Utility).
Created 2015 Triangle717
<http://le717.github.io/>
Licensed under The MIT License
<http://opensource.org/licenses/MIT/>
"""
class ActionsQueue:
def __init__(self):
self.queue = []
class Responses:
def __init__(self):
pass
# Normal buttons
def btnBrowse(self, val):
print(val)
def actionReset(self, val):
print(val)
def btnRedirect(self, val):
print(val)
# Radio buttons
def radioColor256(self, val):
print(val)
def radioColor16b(self, val):
print(val)
def radioModelLow(self, val):
print(val)
def radioModelFast(self, val):
print(val)
def radioModelHigh(self, val):
print(val)
def radioTexFast(self, val):
print(val)
def radioTexHigh(self, val):
print(val)
# Check boxes
def chkCursor(self, val):
print(val)
def chkDraw3D(self, val):
print(val)
def chkFlipVideo(self, val):
print(val)
def chkFullscreen(self, val):
print(val)
def chkJoystick(self, val):
print(val)
def chkMusic(self, val):
print(val)
def chkSound(self, val):
print(val)
# Direct3D dropdown selection
def comboD3D(self, val):
print(val)
|
facfa3bcd7d35163e0504ef4b6f9b3b15e778993 | modeltranslation/management/commands/update_translation_fields.py | modeltranslation/management/commands/update_translation_fields.py | # -*- coding: utf-8 -*-
from django.db.models import F, Q
from django.core.management.base import NoArgsCommand
from modeltranslation.settings import DEFAULT_LANGUAGE
from modeltranslation.translator import translator
from modeltranslation.utils import build_localized_fieldname
class Command(NoArgsCommand):
help = ('Updates the default translation fields of all or the specified'
'translated application using the value of the original field.')
def handle(self, **options):
verbosity = int(options['verbosity'])
if verbosity > 0:
self.stdout.write("Using default language: %s\n" % DEFAULT_LANGUAGE)
for model, trans_opts in translator._registry.items():
if model._meta.abstract:
continue
if verbosity > 0:
self.stdout.write("Updating data of model '%s'\n" % model)
for fieldname in trans_opts.fields:
def_lang_fieldname = build_localized_fieldname(
fieldname, DEFAULT_LANGUAGE)
# We'll only update fields which do not have an existing value
q = Q(**{def_lang_fieldname: None})
field = model._meta.get_field(fieldname)
if field.empty_strings_allowed:
q |= Q(**{def_lang_fieldname: ""})
try:
model.objects.filter(q).rewrite(False).update(
**{def_lang_fieldname: F(fieldname)})
except AttributeError:
# FIXME: Workaround for abstract models. See issue #123 for details.
model.objects.filter(q).update(**{def_lang_fieldname: F(fieldname)})
| # -*- coding: utf-8 -*-
from django.db.models import F, Q
from django.core.management.base import NoArgsCommand
from modeltranslation.settings import DEFAULT_LANGUAGE
from modeltranslation.translator import translator
from modeltranslation.utils import build_localized_fieldname
class Command(NoArgsCommand):
help = ('Updates the default translation fields of all or the specified'
'translated application using the value of the original field.')
def handle(self, **options):
verbosity = int(options['verbosity'])
if verbosity > 0:
self.stdout.write("Using default language: %s\n" % DEFAULT_LANGUAGE)
for model, trans_opts in translator._registry.items():
if model._meta.abstract:
continue
if verbosity > 0:
self.stdout.write("Updating data of model '%s'\n" % model)
for fieldname in trans_opts.fields:
def_lang_fieldname = build_localized_fieldname(
fieldname, DEFAULT_LANGUAGE)
# We'll only update fields which do not have an existing value
q = Q(**{def_lang_fieldname: None})
field = model._meta.get_field(fieldname)
if field.empty_strings_allowed:
q |= Q(**{def_lang_fieldname: ""})
model.objects.filter(q).rewrite(False).update(**{def_lang_fieldname: F(fieldname)})
| Revert "Added a workaround for abstract models not being handled correctly." | Revert "Added a workaround for abstract models not being handled correctly."
This reverts commit a3e44c187b5abfa6d9b360cecc5c1daa746134f5.
| Python | bsd-3-clause | marctc/django-modeltranslation,akheron/django-modeltranslation,nanuxbe/django-modeltranslation,marctc/django-modeltranslation,nanuxbe/django-modeltranslation,SideStudios/django-modeltranslation,akheron/django-modeltranslation,yoza/django-modeltranslation,extertioner/django-modeltranslation,extertioner/django-modeltranslation,deschler/django-modeltranslation,acdha/django-modeltranslation,acdha/django-modeltranslation,vstoykov/django-modeltranslation,deschler/django-modeltranslation,yoza/django-modeltranslation,SideStudios/django-modeltranslation,vstoykov/django-modeltranslation | # -*- coding: utf-8 -*-
from django.db.models import F, Q
from django.core.management.base import NoArgsCommand
from modeltranslation.settings import DEFAULT_LANGUAGE
from modeltranslation.translator import translator
from modeltranslation.utils import build_localized_fieldname
class Command(NoArgsCommand):
help = ('Updates the default translation fields of all or the specified'
'translated application using the value of the original field.')
def handle(self, **options):
verbosity = int(options['verbosity'])
if verbosity > 0:
self.stdout.write("Using default language: %s\n" % DEFAULT_LANGUAGE)
for model, trans_opts in translator._registry.items():
if model._meta.abstract:
continue
if verbosity > 0:
self.stdout.write("Updating data of model '%s'\n" % model)
for fieldname in trans_opts.fields:
def_lang_fieldname = build_localized_fieldname(
fieldname, DEFAULT_LANGUAGE)
# We'll only update fields which do not have an existing value
q = Q(**{def_lang_fieldname: None})
field = model._meta.get_field(fieldname)
if field.empty_strings_allowed:
q |= Q(**{def_lang_fieldname: ""})
try:
model.objects.filter(q).rewrite(False).update(
**{def_lang_fieldname: F(fieldname)})
except AttributeError:
# FIXME: Workaround for abstract models. See issue #123 for details.
model.objects.filter(q).update(**{def_lang_fieldname: F(fieldname)})
Revert "Added a workaround for abstract models not being handled correctly."
This reverts commit a3e44c187b5abfa6d9b360cecc5c1daa746134f5. | # -*- coding: utf-8 -*-
from django.db.models import F, Q
from django.core.management.base import NoArgsCommand
from modeltranslation.settings import DEFAULT_LANGUAGE
from modeltranslation.translator import translator
from modeltranslation.utils import build_localized_fieldname
class Command(NoArgsCommand):
help = ('Updates the default translation fields of all or the specified'
'translated application using the value of the original field.')
def handle(self, **options):
verbosity = int(options['verbosity'])
if verbosity > 0:
self.stdout.write("Using default language: %s\n" % DEFAULT_LANGUAGE)
for model, trans_opts in translator._registry.items():
if model._meta.abstract:
continue
if verbosity > 0:
self.stdout.write("Updating data of model '%s'\n" % model)
for fieldname in trans_opts.fields:
def_lang_fieldname = build_localized_fieldname(
fieldname, DEFAULT_LANGUAGE)
# We'll only update fields which do not have an existing value
q = Q(**{def_lang_fieldname: None})
field = model._meta.get_field(fieldname)
if field.empty_strings_allowed:
q |= Q(**{def_lang_fieldname: ""})
model.objects.filter(q).rewrite(False).update(**{def_lang_fieldname: F(fieldname)})
| <commit_before># -*- coding: utf-8 -*-
from django.db.models import F, Q
from django.core.management.base import NoArgsCommand
from modeltranslation.settings import DEFAULT_LANGUAGE
from modeltranslation.translator import translator
from modeltranslation.utils import build_localized_fieldname
class Command(NoArgsCommand):
help = ('Updates the default translation fields of all or the specified'
'translated application using the value of the original field.')
def handle(self, **options):
verbosity = int(options['verbosity'])
if verbosity > 0:
self.stdout.write("Using default language: %s\n" % DEFAULT_LANGUAGE)
for model, trans_opts in translator._registry.items():
if model._meta.abstract:
continue
if verbosity > 0:
self.stdout.write("Updating data of model '%s'\n" % model)
for fieldname in trans_opts.fields:
def_lang_fieldname = build_localized_fieldname(
fieldname, DEFAULT_LANGUAGE)
# We'll only update fields which do not have an existing value
q = Q(**{def_lang_fieldname: None})
field = model._meta.get_field(fieldname)
if field.empty_strings_allowed:
q |= Q(**{def_lang_fieldname: ""})
try:
model.objects.filter(q).rewrite(False).update(
**{def_lang_fieldname: F(fieldname)})
except AttributeError:
# FIXME: Workaround for abstract models. See issue #123 for details.
model.objects.filter(q).update(**{def_lang_fieldname: F(fieldname)})
<commit_msg>Revert "Added a workaround for abstract models not being handled correctly."
This reverts commit a3e44c187b5abfa6d9b360cecc5c1daa746134f5.<commit_after> | # -*- coding: utf-8 -*-
from django.db.models import F, Q
from django.core.management.base import NoArgsCommand
from modeltranslation.settings import DEFAULT_LANGUAGE
from modeltranslation.translator import translator
from modeltranslation.utils import build_localized_fieldname
class Command(NoArgsCommand):
help = ('Updates the default translation fields of all or the specified'
'translated application using the value of the original field.')
def handle(self, **options):
verbosity = int(options['verbosity'])
if verbosity > 0:
self.stdout.write("Using default language: %s\n" % DEFAULT_LANGUAGE)
for model, trans_opts in translator._registry.items():
if model._meta.abstract:
continue
if verbosity > 0:
self.stdout.write("Updating data of model '%s'\n" % model)
for fieldname in trans_opts.fields:
def_lang_fieldname = build_localized_fieldname(
fieldname, DEFAULT_LANGUAGE)
# We'll only update fields which do not have an existing value
q = Q(**{def_lang_fieldname: None})
field = model._meta.get_field(fieldname)
if field.empty_strings_allowed:
q |= Q(**{def_lang_fieldname: ""})
model.objects.filter(q).rewrite(False).update(**{def_lang_fieldname: F(fieldname)})
| # -*- coding: utf-8 -*-
from django.db.models import F, Q
from django.core.management.base import NoArgsCommand
from modeltranslation.settings import DEFAULT_LANGUAGE
from modeltranslation.translator import translator
from modeltranslation.utils import build_localized_fieldname
class Command(NoArgsCommand):
help = ('Updates the default translation fields of all or the specified'
'translated application using the value of the original field.')
def handle(self, **options):
verbosity = int(options['verbosity'])
if verbosity > 0:
self.stdout.write("Using default language: %s\n" % DEFAULT_LANGUAGE)
for model, trans_opts in translator._registry.items():
if model._meta.abstract:
continue
if verbosity > 0:
self.stdout.write("Updating data of model '%s'\n" % model)
for fieldname in trans_opts.fields:
def_lang_fieldname = build_localized_fieldname(
fieldname, DEFAULT_LANGUAGE)
# We'll only update fields which do not have an existing value
q = Q(**{def_lang_fieldname: None})
field = model._meta.get_field(fieldname)
if field.empty_strings_allowed:
q |= Q(**{def_lang_fieldname: ""})
try:
model.objects.filter(q).rewrite(False).update(
**{def_lang_fieldname: F(fieldname)})
except AttributeError:
# FIXME: Workaround for abstract models. See issue #123 for details.
model.objects.filter(q).update(**{def_lang_fieldname: F(fieldname)})
Revert "Added a workaround for abstract models not being handled correctly."
This reverts commit a3e44c187b5abfa6d9b360cecc5c1daa746134f5.# -*- coding: utf-8 -*-
from django.db.models import F, Q
from django.core.management.base import NoArgsCommand
from modeltranslation.settings import DEFAULT_LANGUAGE
from modeltranslation.translator import translator
from modeltranslation.utils import build_localized_fieldname
class Command(NoArgsCommand):
help = ('Updates the default translation fields of all or the specified'
'translated application using the value of the original field.')
def handle(self, **options):
verbosity = int(options['verbosity'])
if verbosity > 0:
self.stdout.write("Using default language: %s\n" % DEFAULT_LANGUAGE)
for model, trans_opts in translator._registry.items():
if model._meta.abstract:
continue
if verbosity > 0:
self.stdout.write("Updating data of model '%s'\n" % model)
for fieldname in trans_opts.fields:
def_lang_fieldname = build_localized_fieldname(
fieldname, DEFAULT_LANGUAGE)
# We'll only update fields which do not have an existing value
q = Q(**{def_lang_fieldname: None})
field = model._meta.get_field(fieldname)
if field.empty_strings_allowed:
q |= Q(**{def_lang_fieldname: ""})
model.objects.filter(q).rewrite(False).update(**{def_lang_fieldname: F(fieldname)})
| <commit_before># -*- coding: utf-8 -*-
from django.db.models import F, Q
from django.core.management.base import NoArgsCommand
from modeltranslation.settings import DEFAULT_LANGUAGE
from modeltranslation.translator import translator
from modeltranslation.utils import build_localized_fieldname
class Command(NoArgsCommand):
help = ('Updates the default translation fields of all or the specified'
'translated application using the value of the original field.')
def handle(self, **options):
verbosity = int(options['verbosity'])
if verbosity > 0:
self.stdout.write("Using default language: %s\n" % DEFAULT_LANGUAGE)
for model, trans_opts in translator._registry.items():
if model._meta.abstract:
continue
if verbosity > 0:
self.stdout.write("Updating data of model '%s'\n" % model)
for fieldname in trans_opts.fields:
def_lang_fieldname = build_localized_fieldname(
fieldname, DEFAULT_LANGUAGE)
# We'll only update fields which do not have an existing value
q = Q(**{def_lang_fieldname: None})
field = model._meta.get_field(fieldname)
if field.empty_strings_allowed:
q |= Q(**{def_lang_fieldname: ""})
try:
model.objects.filter(q).rewrite(False).update(
**{def_lang_fieldname: F(fieldname)})
except AttributeError:
# FIXME: Workaround for abstract models. See issue #123 for details.
model.objects.filter(q).update(**{def_lang_fieldname: F(fieldname)})
<commit_msg>Revert "Added a workaround for abstract models not being handled correctly."
This reverts commit a3e44c187b5abfa6d9b360cecc5c1daa746134f5.<commit_after># -*- coding: utf-8 -*-
from django.db.models import F, Q
from django.core.management.base import NoArgsCommand
from modeltranslation.settings import DEFAULT_LANGUAGE
from modeltranslation.translator import translator
from modeltranslation.utils import build_localized_fieldname
class Command(NoArgsCommand):
help = ('Updates the default translation fields of all or the specified'
'translated application using the value of the original field.')
def handle(self, **options):
verbosity = int(options['verbosity'])
if verbosity > 0:
self.stdout.write("Using default language: %s\n" % DEFAULT_LANGUAGE)
for model, trans_opts in translator._registry.items():
if model._meta.abstract:
continue
if verbosity > 0:
self.stdout.write("Updating data of model '%s'\n" % model)
for fieldname in trans_opts.fields:
def_lang_fieldname = build_localized_fieldname(
fieldname, DEFAULT_LANGUAGE)
# We'll only update fields which do not have an existing value
q = Q(**{def_lang_fieldname: None})
field = model._meta.get_field(fieldname)
if field.empty_strings_allowed:
q |= Q(**{def_lang_fieldname: ""})
model.objects.filter(q).rewrite(False).update(**{def_lang_fieldname: F(fieldname)})
|
ae6ed4e7dc6510637d322eb6403f43b9d4aa5d25 | karteikarten/helpers/exporters.py | karteikarten/helpers/exporters.py | # Copyright (C) 2013 Braden Walters
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
from karteikarten.models import Card
class AnkiExporter(object):
'''
Exports cards to ANKI Text File.
'''
@staticmethod
def export(cardset):
result = u'front\tback\n'
for card in Card.objects.filter(parent_card_set = cardset):
result += card.front.replace('\n', '<br />') + '\t' + \
card.back.replace('\n', '<br />') + '\n'
return result
@staticmethod
def getExtension():
return '.txt'
| # Copyright (C) 2013 Braden Walters
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
from karteikarten.models import Card
class AnkiExporter(object):
'''
Exports cards to ANKI Text File.
'''
@staticmethod
def export(cardset):
result = u'front\tback\n'
for card in Card.objects.filter(parent_card_set = cardset):
result += card.front.replace('\n', '<br />').replace('\r', '') + '\t' + \
card.back.replace('\n', '<br />').replace('\r', '') + '\n'
return result
@staticmethod
def getExtension():
return '.txt'
| Remove carriage returns in Anki exporter | Remove carriage returns in Anki exporter
| Python | agpl-3.0 | meoblast001/kksystem,meoblast001/kksystem,meoblast001/kksystem | # Copyright (C) 2013 Braden Walters
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
from karteikarten.models import Card
class AnkiExporter(object):
'''
Exports cards to ANKI Text File.
'''
@staticmethod
def export(cardset):
result = u'front\tback\n'
for card in Card.objects.filter(parent_card_set = cardset):
result += card.front.replace('\n', '<br />') + '\t' + \
card.back.replace('\n', '<br />') + '\n'
return result
@staticmethod
def getExtension():
return '.txt'
Remove carriage returns in Anki exporter | # Copyright (C) 2013 Braden Walters
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
from karteikarten.models import Card
class AnkiExporter(object):
'''
Exports cards to ANKI Text File.
'''
@staticmethod
def export(cardset):
result = u'front\tback\n'
for card in Card.objects.filter(parent_card_set = cardset):
result += card.front.replace('\n', '<br />').replace('\r', '') + '\t' + \
card.back.replace('\n', '<br />').replace('\r', '') + '\n'
return result
@staticmethod
def getExtension():
return '.txt'
| <commit_before># Copyright (C) 2013 Braden Walters
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
from karteikarten.models import Card
class AnkiExporter(object):
'''
Exports cards to ANKI Text File.
'''
@staticmethod
def export(cardset):
result = u'front\tback\n'
for card in Card.objects.filter(parent_card_set = cardset):
result += card.front.replace('\n', '<br />') + '\t' + \
card.back.replace('\n', '<br />') + '\n'
return result
@staticmethod
def getExtension():
return '.txt'
<commit_msg>Remove carriage returns in Anki exporter<commit_after> | # Copyright (C) 2013 Braden Walters
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
from karteikarten.models import Card
class AnkiExporter(object):
'''
Exports cards to ANKI Text File.
'''
@staticmethod
def export(cardset):
result = u'front\tback\n'
for card in Card.objects.filter(parent_card_set = cardset):
result += card.front.replace('\n', '<br />').replace('\r', '') + '\t' + \
card.back.replace('\n', '<br />').replace('\r', '') + '\n'
return result
@staticmethod
def getExtension():
return '.txt'
| # Copyright (C) 2013 Braden Walters
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
from karteikarten.models import Card
class AnkiExporter(object):
'''
Exports cards to ANKI Text File.
'''
@staticmethod
def export(cardset):
result = u'front\tback\n'
for card in Card.objects.filter(parent_card_set = cardset):
result += card.front.replace('\n', '<br />') + '\t' + \
card.back.replace('\n', '<br />') + '\n'
return result
@staticmethod
def getExtension():
return '.txt'
Remove carriage returns in Anki exporter# Copyright (C) 2013 Braden Walters
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
from karteikarten.models import Card
class AnkiExporter(object):
'''
Exports cards to ANKI Text File.
'''
@staticmethod
def export(cardset):
result = u'front\tback\n'
for card in Card.objects.filter(parent_card_set = cardset):
result += card.front.replace('\n', '<br />').replace('\r', '') + '\t' + \
card.back.replace('\n', '<br />').replace('\r', '') + '\n'
return result
@staticmethod
def getExtension():
return '.txt'
| <commit_before># Copyright (C) 2013 Braden Walters
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
from karteikarten.models import Card
class AnkiExporter(object):
'''
Exports cards to ANKI Text File.
'''
@staticmethod
def export(cardset):
result = u'front\tback\n'
for card in Card.objects.filter(parent_card_set = cardset):
result += card.front.replace('\n', '<br />') + '\t' + \
card.back.replace('\n', '<br />') + '\n'
return result
@staticmethod
def getExtension():
return '.txt'
<commit_msg>Remove carriage returns in Anki exporter<commit_after># Copyright (C) 2013 Braden Walters
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
from karteikarten.models import Card
class AnkiExporter(object):
'''
Exports cards to ANKI Text File.
'''
@staticmethod
def export(cardset):
result = u'front\tback\n'
for card in Card.objects.filter(parent_card_set = cardset):
result += card.front.replace('\n', '<br />').replace('\r', '') + '\t' + \
card.back.replace('\n', '<br />').replace('\r', '') + '\n'
return result
@staticmethod
def getExtension():
return '.txt'
|
24d3af9288102e5061c3a1f9e6fe2d7d578f6cc5 | astro.py | astro.py | import ephem
m = getattr(ephem, (raw_input('Planet: ')))
print ephem.constellation(m(raw_input('yyyy/mm/dd: '))) | import ephem
def const(planet_name, date_string): # function name and parameters
planet_class = getattr(ephem, planet_name) # sets ephem object class
planet = planet_class() # sets planet variable
south_bend = ephem.Observer() # Creates the Observer object
south_bend.lat = '41.67' # latitude
south_bend.lon = '-86.26' # west is negative
south_bend.date = date_string # sets date parameter
planet.compute(south_bend) # calculates the location data
print ephem.constellation((planet.ra, planet.dec))
return planet.alt, planet.az
print const(raw_input('Planet: '), getattr(ephem, now()))
| Select planet and display alt/az data | Select planet and display alt/az data
Created a function localized to South Bend, IN, to display altitude and
azimuth data for a planet selected.
| Python | mit | bennettscience/PySky | import ephem
m = getattr(ephem, (raw_input('Planet: ')))
print ephem.constellation(m(raw_input('yyyy/mm/dd: ')))Select planet and display alt/az data
Created a function localized to South Bend, IN, to display altitude and
azimuth data for a planet selected. | import ephem
def const(planet_name, date_string): # function name and parameters
planet_class = getattr(ephem, planet_name) # sets ephem object class
planet = planet_class() # sets planet variable
south_bend = ephem.Observer() # Creates the Observer object
south_bend.lat = '41.67' # latitude
south_bend.lon = '-86.26' # west is negative
south_bend.date = date_string # sets date parameter
planet.compute(south_bend) # calculates the location data
print ephem.constellation((planet.ra, planet.dec))
return planet.alt, planet.az
print const(raw_input('Planet: '), getattr(ephem, now()))
| <commit_before>import ephem
m = getattr(ephem, (raw_input('Planet: ')))
print ephem.constellation(m(raw_input('yyyy/mm/dd: ')))<commit_msg>Select planet and display alt/az data
Created a function localized to South Bend, IN, to display altitude and
azimuth data for a planet selected.<commit_after> | import ephem
def const(planet_name, date_string): # function name and parameters
planet_class = getattr(ephem, planet_name) # sets ephem object class
planet = planet_class() # sets planet variable
south_bend = ephem.Observer() # Creates the Observer object
south_bend.lat = '41.67' # latitude
south_bend.lon = '-86.26' # west is negative
south_bend.date = date_string # sets date parameter
planet.compute(south_bend) # calculates the location data
print ephem.constellation((planet.ra, planet.dec))
return planet.alt, planet.az
print const(raw_input('Planet: '), getattr(ephem, now()))
| import ephem
m = getattr(ephem, (raw_input('Planet: ')))
print ephem.constellation(m(raw_input('yyyy/mm/dd: ')))Select planet and display alt/az data
Created a function localized to South Bend, IN, to display altitude and
azimuth data for a planet selected.import ephem
def const(planet_name, date_string): # function name and parameters
planet_class = getattr(ephem, planet_name) # sets ephem object class
planet = planet_class() # sets planet variable
south_bend = ephem.Observer() # Creates the Observer object
south_bend.lat = '41.67' # latitude
south_bend.lon = '-86.26' # west is negative
south_bend.date = date_string # sets date parameter
planet.compute(south_bend) # calculates the location data
print ephem.constellation((planet.ra, planet.dec))
return planet.alt, planet.az
print const(raw_input('Planet: '), getattr(ephem, now()))
| <commit_before>import ephem
m = getattr(ephem, (raw_input('Planet: ')))
print ephem.constellation(m(raw_input('yyyy/mm/dd: ')))<commit_msg>Select planet and display alt/az data
Created a function localized to South Bend, IN, to display altitude and
azimuth data for a planet selected.<commit_after>import ephem
def const(planet_name, date_string): # function name and parameters
planet_class = getattr(ephem, planet_name) # sets ephem object class
planet = planet_class() # sets planet variable
south_bend = ephem.Observer() # Creates the Observer object
south_bend.lat = '41.67' # latitude
south_bend.lon = '-86.26' # west is negative
south_bend.date = date_string # sets date parameter
planet.compute(south_bend) # calculates the location data
print ephem.constellation((planet.ra, planet.dec))
return planet.alt, planet.az
print const(raw_input('Planet: '), getattr(ephem, now()))
|
071926edc64241b0359c9a0148fc0825a09cb6ba | marionette/__init__.py | marionette/__init__.py |
from cgi import parse_header
import json
from django.http import HttpResponse, Http404
RPC_MARKER = '_rpc'
class Resource(object):
def __init__(self, request, *args, **kwargs):
self.request = request
self.args = args
self.kwargs = kwargs
@classmethod
def as_view(cls):
def view(request, *args, **kwargs):
self = cls(request, *args, **kwargs)
return self.dispatch(request)
return view
def dispatch(self, request):
method = request.META['HTTP_X_RPC_ACTION']
func = getattr(self, method, None)
if not getattr(func, RPC_MARKER, True):
raise Http404
data = self.get_request_data(request)
resp = func(data)
return HttpResponse(json.dumps(resp), content_type='application/json')
def get_request_data(self, default=None):
'''Retrieve data from request'''
c_type, _ = parse_header(self.request.META.get('CONTENT_TYPE', ''))
if c_type in ['application/json', 'text/json']:
if not self.request.body:
return default
return self.loads(self.request.body)
if self.request.method == 'GET':
return self.request.GET
return self.request.POST
def rpc(view):
'''Mark a view as accessible via RPC'''
setattr(view, '_rpc', True)
return view
|
from cgi import parse_header
import json
from django.http import HttpResponse, Http404
RPC_MARKER = '_rpc'
class Resource(object):
def __init__(self, request, *args, **kwargs):
self.request = request
self.args = args
self.kwargs = kwargs
@classmethod
def as_view(cls):
def view(request, *args, **kwargs):
self = cls(request, *args, **kwargs)
return self.dispatch(request)
return view
def dispatch(self, request):
method = request.META['HTTP_X_RPC_ACTION']
func = getattr(self, method, None)
if not getattr(func, RPC_MARKER, True):
raise Http404
data = self.get_request_data(request)
resp = self.execute(func, data)
return HttpResponse(json.dumps(resp), content_type='application/json')
def execute(self, handler, data):
'''Helpful hook to ease wrapping the handler'''
return handler(**data)
def get_request_data(self, default=None):
'''Retrieve data from request'''
c_type, _ = parse_header(self.request.META.get('CONTENT_TYPE', ''))
if c_type in ['application/json', 'text/json']:
if not self.request.body:
return default
return self.loads(self.request.body)
if self.request.method == 'GET':
return self.request.GET
return self.request.POST
def rpc(view):
'''Mark a view as accessible via RPC'''
setattr(view, '_rpc', True)
return view
| Add execute hook to allow wrapping handler calls | Add execute hook to allow wrapping handler calls
| Python | mit | funkybob/django-marionette |
from cgi import parse_header
import json
from django.http import HttpResponse, Http404
RPC_MARKER = '_rpc'
class Resource(object):
def __init__(self, request, *args, **kwargs):
self.request = request
self.args = args
self.kwargs = kwargs
@classmethod
def as_view(cls):
def view(request, *args, **kwargs):
self = cls(request, *args, **kwargs)
return self.dispatch(request)
return view
def dispatch(self, request):
method = request.META['HTTP_X_RPC_ACTION']
func = getattr(self, method, None)
if not getattr(func, RPC_MARKER, True):
raise Http404
data = self.get_request_data(request)
resp = func(data)
return HttpResponse(json.dumps(resp), content_type='application/json')
def get_request_data(self, default=None):
'''Retrieve data from request'''
c_type, _ = parse_header(self.request.META.get('CONTENT_TYPE', ''))
if c_type in ['application/json', 'text/json']:
if not self.request.body:
return default
return self.loads(self.request.body)
if self.request.method == 'GET':
return self.request.GET
return self.request.POST
def rpc(view):
'''Mark a view as accessible via RPC'''
setattr(view, '_rpc', True)
return view
Add execute hook to allow wrapping handler calls |
from cgi import parse_header
import json
from django.http import HttpResponse, Http404
RPC_MARKER = '_rpc'
class Resource(object):
def __init__(self, request, *args, **kwargs):
self.request = request
self.args = args
self.kwargs = kwargs
@classmethod
def as_view(cls):
def view(request, *args, **kwargs):
self = cls(request, *args, **kwargs)
return self.dispatch(request)
return view
def dispatch(self, request):
method = request.META['HTTP_X_RPC_ACTION']
func = getattr(self, method, None)
if not getattr(func, RPC_MARKER, True):
raise Http404
data = self.get_request_data(request)
resp = self.execute(func, data)
return HttpResponse(json.dumps(resp), content_type='application/json')
def execute(self, handler, data):
'''Helpful hook to ease wrapping the handler'''
return handler(**data)
def get_request_data(self, default=None):
'''Retrieve data from request'''
c_type, _ = parse_header(self.request.META.get('CONTENT_TYPE', ''))
if c_type in ['application/json', 'text/json']:
if not self.request.body:
return default
return self.loads(self.request.body)
if self.request.method == 'GET':
return self.request.GET
return self.request.POST
def rpc(view):
'''Mark a view as accessible via RPC'''
setattr(view, '_rpc', True)
return view
| <commit_before>
from cgi import parse_header
import json
from django.http import HttpResponse, Http404
RPC_MARKER = '_rpc'
class Resource(object):
def __init__(self, request, *args, **kwargs):
self.request = request
self.args = args
self.kwargs = kwargs
@classmethod
def as_view(cls):
def view(request, *args, **kwargs):
self = cls(request, *args, **kwargs)
return self.dispatch(request)
return view
def dispatch(self, request):
method = request.META['HTTP_X_RPC_ACTION']
func = getattr(self, method, None)
if not getattr(func, RPC_MARKER, True):
raise Http404
data = self.get_request_data(request)
resp = func(data)
return HttpResponse(json.dumps(resp), content_type='application/json')
def get_request_data(self, default=None):
'''Retrieve data from request'''
c_type, _ = parse_header(self.request.META.get('CONTENT_TYPE', ''))
if c_type in ['application/json', 'text/json']:
if not self.request.body:
return default
return self.loads(self.request.body)
if self.request.method == 'GET':
return self.request.GET
return self.request.POST
def rpc(view):
'''Mark a view as accessible via RPC'''
setattr(view, '_rpc', True)
return view
<commit_msg>Add execute hook to allow wrapping handler calls<commit_after> |
from cgi import parse_header
import json
from django.http import HttpResponse, Http404
RPC_MARKER = '_rpc'
class Resource(object):
def __init__(self, request, *args, **kwargs):
self.request = request
self.args = args
self.kwargs = kwargs
@classmethod
def as_view(cls):
def view(request, *args, **kwargs):
self = cls(request, *args, **kwargs)
return self.dispatch(request)
return view
def dispatch(self, request):
method = request.META['HTTP_X_RPC_ACTION']
func = getattr(self, method, None)
if not getattr(func, RPC_MARKER, True):
raise Http404
data = self.get_request_data(request)
resp = self.execute(func, data)
return HttpResponse(json.dumps(resp), content_type='application/json')
def execute(self, handler, data):
'''Helpful hook to ease wrapping the handler'''
return handler(**data)
def get_request_data(self, default=None):
'''Retrieve data from request'''
c_type, _ = parse_header(self.request.META.get('CONTENT_TYPE', ''))
if c_type in ['application/json', 'text/json']:
if not self.request.body:
return default
return self.loads(self.request.body)
if self.request.method == 'GET':
return self.request.GET
return self.request.POST
def rpc(view):
'''Mark a view as accessible via RPC'''
setattr(view, '_rpc', True)
return view
|
from cgi import parse_header
import json
from django.http import HttpResponse, Http404
RPC_MARKER = '_rpc'
class Resource(object):
def __init__(self, request, *args, **kwargs):
self.request = request
self.args = args
self.kwargs = kwargs
@classmethod
def as_view(cls):
def view(request, *args, **kwargs):
self = cls(request, *args, **kwargs)
return self.dispatch(request)
return view
def dispatch(self, request):
method = request.META['HTTP_X_RPC_ACTION']
func = getattr(self, method, None)
if not getattr(func, RPC_MARKER, True):
raise Http404
data = self.get_request_data(request)
resp = func(data)
return HttpResponse(json.dumps(resp), content_type='application/json')
def get_request_data(self, default=None):
'''Retrieve data from request'''
c_type, _ = parse_header(self.request.META.get('CONTENT_TYPE', ''))
if c_type in ['application/json', 'text/json']:
if not self.request.body:
return default
return self.loads(self.request.body)
if self.request.method == 'GET':
return self.request.GET
return self.request.POST
def rpc(view):
'''Mark a view as accessible via RPC'''
setattr(view, '_rpc', True)
return view
Add execute hook to allow wrapping handler calls
from cgi import parse_header
import json
from django.http import HttpResponse, Http404
RPC_MARKER = '_rpc'
class Resource(object):
def __init__(self, request, *args, **kwargs):
self.request = request
self.args = args
self.kwargs = kwargs
@classmethod
def as_view(cls):
def view(request, *args, **kwargs):
self = cls(request, *args, **kwargs)
return self.dispatch(request)
return view
def dispatch(self, request):
method = request.META['HTTP_X_RPC_ACTION']
func = getattr(self, method, None)
if not getattr(func, RPC_MARKER, True):
raise Http404
data = self.get_request_data(request)
resp = self.execute(func, data)
return HttpResponse(json.dumps(resp), content_type='application/json')
def execute(self, handler, data):
'''Helpful hook to ease wrapping the handler'''
return handler(**data)
def get_request_data(self, default=None):
'''Retrieve data from request'''
c_type, _ = parse_header(self.request.META.get('CONTENT_TYPE', ''))
if c_type in ['application/json', 'text/json']:
if not self.request.body:
return default
return self.loads(self.request.body)
if self.request.method == 'GET':
return self.request.GET
return self.request.POST
def rpc(view):
'''Mark a view as accessible via RPC'''
setattr(view, '_rpc', True)
return view
| <commit_before>
from cgi import parse_header
import json
from django.http import HttpResponse, Http404
RPC_MARKER = '_rpc'
class Resource(object):
def __init__(self, request, *args, **kwargs):
self.request = request
self.args = args
self.kwargs = kwargs
@classmethod
def as_view(cls):
def view(request, *args, **kwargs):
self = cls(request, *args, **kwargs)
return self.dispatch(request)
return view
def dispatch(self, request):
method = request.META['HTTP_X_RPC_ACTION']
func = getattr(self, method, None)
if not getattr(func, RPC_MARKER, True):
raise Http404
data = self.get_request_data(request)
resp = func(data)
return HttpResponse(json.dumps(resp), content_type='application/json')
def get_request_data(self, default=None):
'''Retrieve data from request'''
c_type, _ = parse_header(self.request.META.get('CONTENT_TYPE', ''))
if c_type in ['application/json', 'text/json']:
if not self.request.body:
return default
return self.loads(self.request.body)
if self.request.method == 'GET':
return self.request.GET
return self.request.POST
def rpc(view):
'''Mark a view as accessible via RPC'''
setattr(view, '_rpc', True)
return view
<commit_msg>Add execute hook to allow wrapping handler calls<commit_after>
from cgi import parse_header
import json
from django.http import HttpResponse, Http404
RPC_MARKER = '_rpc'
class Resource(object):
def __init__(self, request, *args, **kwargs):
self.request = request
self.args = args
self.kwargs = kwargs
@classmethod
def as_view(cls):
def view(request, *args, **kwargs):
self = cls(request, *args, **kwargs)
return self.dispatch(request)
return view
def dispatch(self, request):
method = request.META['HTTP_X_RPC_ACTION']
func = getattr(self, method, None)
if not getattr(func, RPC_MARKER, True):
raise Http404
data = self.get_request_data(request)
resp = self.execute(func, data)
return HttpResponse(json.dumps(resp), content_type='application/json')
def execute(self, handler, data):
'''Helpful hook to ease wrapping the handler'''
return handler(**data)
def get_request_data(self, default=None):
'''Retrieve data from request'''
c_type, _ = parse_header(self.request.META.get('CONTENT_TYPE', ''))
if c_type in ['application/json', 'text/json']:
if not self.request.body:
return default
return self.loads(self.request.body)
if self.request.method == 'GET':
return self.request.GET
return self.request.POST
def rpc(view):
'''Mark a view as accessible via RPC'''
setattr(view, '_rpc', True)
return view
|
6f947c99411692f8fe4899203ed9bf202b0412a3 | cihai/__about__.py | cihai/__about__.py | __title__ = 'cihai'
__package_name__ = 'cihai'
__version__ = '0.9.0a3'
__description__ = 'Library for CJK (chinese, japanese, korean) language data.'
__author__ = 'Tony Narlock'
__email__ = 'tony@git-pull.com'
__github__ = 'https://github.com/cihai/cihai'
__pypi__ = 'https://pypi.org/project/cihai/'
__license__ = 'MIT'
__copyright__ = 'Copyright 2013- cihai software foundation'
| __title__ = 'cihai'
__package_name__ = 'cihai'
__version__ = '0.9.0a3'
__description__ = 'Library for CJK (chinese, japanese, korean) language data.'
__author__ = 'Tony Narlock'
__email__ = 'tony@git-pull.com'
__github__ = 'https://github.com/cihai/cihai'
__docs__ = 'https://cihai.git-pull.com'
__tracker__ = 'https://github.com/cihai/cihai/issues'
__pypi__ = 'https://pypi.org/project/cihai/'
__license__ = 'MIT'
__copyright__ = 'Copyright 2013- cihai software foundation'
| Add tracker and doc URL to metadata | Add tracker and doc URL to metadata
| Python | mit | cihai/cihai,cihai/cihai | __title__ = 'cihai'
__package_name__ = 'cihai'
__version__ = '0.9.0a3'
__description__ = 'Library for CJK (chinese, japanese, korean) language data.'
__author__ = 'Tony Narlock'
__email__ = 'tony@git-pull.com'
__github__ = 'https://github.com/cihai/cihai'
__pypi__ = 'https://pypi.org/project/cihai/'
__license__ = 'MIT'
__copyright__ = 'Copyright 2013- cihai software foundation'
Add tracker and doc URL to metadata | __title__ = 'cihai'
__package_name__ = 'cihai'
__version__ = '0.9.0a3'
__description__ = 'Library for CJK (chinese, japanese, korean) language data.'
__author__ = 'Tony Narlock'
__email__ = 'tony@git-pull.com'
__github__ = 'https://github.com/cihai/cihai'
__docs__ = 'https://cihai.git-pull.com'
__tracker__ = 'https://github.com/cihai/cihai/issues'
__pypi__ = 'https://pypi.org/project/cihai/'
__license__ = 'MIT'
__copyright__ = 'Copyright 2013- cihai software foundation'
| <commit_before>__title__ = 'cihai'
__package_name__ = 'cihai'
__version__ = '0.9.0a3'
__description__ = 'Library for CJK (chinese, japanese, korean) language data.'
__author__ = 'Tony Narlock'
__email__ = 'tony@git-pull.com'
__github__ = 'https://github.com/cihai/cihai'
__pypi__ = 'https://pypi.org/project/cihai/'
__license__ = 'MIT'
__copyright__ = 'Copyright 2013- cihai software foundation'
<commit_msg>Add tracker and doc URL to metadata<commit_after> | __title__ = 'cihai'
__package_name__ = 'cihai'
__version__ = '0.9.0a3'
__description__ = 'Library for CJK (chinese, japanese, korean) language data.'
__author__ = 'Tony Narlock'
__email__ = 'tony@git-pull.com'
__github__ = 'https://github.com/cihai/cihai'
__docs__ = 'https://cihai.git-pull.com'
__tracker__ = 'https://github.com/cihai/cihai/issues'
__pypi__ = 'https://pypi.org/project/cihai/'
__license__ = 'MIT'
__copyright__ = 'Copyright 2013- cihai software foundation'
| __title__ = 'cihai'
__package_name__ = 'cihai'
__version__ = '0.9.0a3'
__description__ = 'Library for CJK (chinese, japanese, korean) language data.'
__author__ = 'Tony Narlock'
__email__ = 'tony@git-pull.com'
__github__ = 'https://github.com/cihai/cihai'
__pypi__ = 'https://pypi.org/project/cihai/'
__license__ = 'MIT'
__copyright__ = 'Copyright 2013- cihai software foundation'
Add tracker and doc URL to metadata__title__ = 'cihai'
__package_name__ = 'cihai'
__version__ = '0.9.0a3'
__description__ = 'Library for CJK (chinese, japanese, korean) language data.'
__author__ = 'Tony Narlock'
__email__ = 'tony@git-pull.com'
__github__ = 'https://github.com/cihai/cihai'
__docs__ = 'https://cihai.git-pull.com'
__tracker__ = 'https://github.com/cihai/cihai/issues'
__pypi__ = 'https://pypi.org/project/cihai/'
__license__ = 'MIT'
__copyright__ = 'Copyright 2013- cihai software foundation'
| <commit_before>__title__ = 'cihai'
__package_name__ = 'cihai'
__version__ = '0.9.0a3'
__description__ = 'Library for CJK (chinese, japanese, korean) language data.'
__author__ = 'Tony Narlock'
__email__ = 'tony@git-pull.com'
__github__ = 'https://github.com/cihai/cihai'
__pypi__ = 'https://pypi.org/project/cihai/'
__license__ = 'MIT'
__copyright__ = 'Copyright 2013- cihai software foundation'
<commit_msg>Add tracker and doc URL to metadata<commit_after>__title__ = 'cihai'
__package_name__ = 'cihai'
__version__ = '0.9.0a3'
__description__ = 'Library for CJK (chinese, japanese, korean) language data.'
__author__ = 'Tony Narlock'
__email__ = 'tony@git-pull.com'
__github__ = 'https://github.com/cihai/cihai'
__docs__ = 'https://cihai.git-pull.com'
__tracker__ = 'https://github.com/cihai/cihai/issues'
__pypi__ = 'https://pypi.org/project/cihai/'
__license__ = 'MIT'
__copyright__ = 'Copyright 2013- cihai software foundation'
|
d6ddc3b41040a374c61d4624e052fa8f1e58ee37 | metakernel/__init__.py | metakernel/__init__.py | from .metakernel import MetaKernel
from .process_metakernel import ProcessMetaKernel
from .magic import Magic, option
from .parser import Parser
__all__ = ['Magic', 'MetaKernel', 'option']
__version__ = '0.3'
del magic, metakernel, parser, process_metakernel
| from .metakernel import MetaKernel
import .pyexpect
from .process_metakernel import ProcessMetaKernel
from .magic import Magic, option
from .parser import Parser
__all__ = ['Magic', 'MetaKernel', 'option']
__version__ = '0.3'
del magic, metakernel, parser, process_metakernel, pyexpect
| Fix import error in Python 2 | Fix import error in Python 2
| Python | bsd-3-clause | Calysto/metakernel | from .metakernel import MetaKernel
from .process_metakernel import ProcessMetaKernel
from .magic import Magic, option
from .parser import Parser
__all__ = ['Magic', 'MetaKernel', 'option']
__version__ = '0.3'
del magic, metakernel, parser, process_metakernel
Fix import error in Python 2 | from .metakernel import MetaKernel
import .pyexpect
from .process_metakernel import ProcessMetaKernel
from .magic import Magic, option
from .parser import Parser
__all__ = ['Magic', 'MetaKernel', 'option']
__version__ = '0.3'
del magic, metakernel, parser, process_metakernel, pyexpect
| <commit_before>from .metakernel import MetaKernel
from .process_metakernel import ProcessMetaKernel
from .magic import Magic, option
from .parser import Parser
__all__ = ['Magic', 'MetaKernel', 'option']
__version__ = '0.3'
del magic, metakernel, parser, process_metakernel
<commit_msg>Fix import error in Python 2<commit_after> | from .metakernel import MetaKernel
import .pyexpect
from .process_metakernel import ProcessMetaKernel
from .magic import Magic, option
from .parser import Parser
__all__ = ['Magic', 'MetaKernel', 'option']
__version__ = '0.3'
del magic, metakernel, parser, process_metakernel, pyexpect
| from .metakernel import MetaKernel
from .process_metakernel import ProcessMetaKernel
from .magic import Magic, option
from .parser import Parser
__all__ = ['Magic', 'MetaKernel', 'option']
__version__ = '0.3'
del magic, metakernel, parser, process_metakernel
Fix import error in Python 2from .metakernel import MetaKernel
import .pyexpect
from .process_metakernel import ProcessMetaKernel
from .magic import Magic, option
from .parser import Parser
__all__ = ['Magic', 'MetaKernel', 'option']
__version__ = '0.3'
del magic, metakernel, parser, process_metakernel, pyexpect
| <commit_before>from .metakernel import MetaKernel
from .process_metakernel import ProcessMetaKernel
from .magic import Magic, option
from .parser import Parser
__all__ = ['Magic', 'MetaKernel', 'option']
__version__ = '0.3'
del magic, metakernel, parser, process_metakernel
<commit_msg>Fix import error in Python 2<commit_after>from .metakernel import MetaKernel
import .pyexpect
from .process_metakernel import ProcessMetaKernel
from .magic import Magic, option
from .parser import Parser
__all__ = ['Magic', 'MetaKernel', 'option']
__version__ = '0.3'
del magic, metakernel, parser, process_metakernel, pyexpect
|
ec9239fda210f3d71a81045fdba9d72bab04a05b | indra/sources/drugbank/__init__.py | indra/sources/drugbank/__init__.py | """This module provides an API and processor for DrugBank content.
It builds on the XML-formatted data schema of DrugBank and expects
the XML file to be available locally. The full DrugBank download
can be obtained at: https://www.drugbank.ca/releases/latest. Once the
XML file is decompressed, it can be processed using the process_xml
function."""
from .api import get_drugbank_processor, process_xml
| """This module provides an API and processor for DrugBank content.
It builds on the XML-formatted data schema of DrugBank and expects
the XML file to be available locally. The full DrugBank download
can be obtained at: https://www.drugbank.ca/releases/latest. Once the
XML file is decompressed, it can be processed using the process_xml
function.
Alternatively, the latest DrugBank data can be automatically loaded via
:mod:`drugbank_downloader` with the following code after doing
``pip install drugbank_downloader bioversions``:
.. code-block:: python
import pickle
import indra.sources.drugbank
processor = indra.sources.drugbank.get_drugbank_processor()
with open('drugbank_indra_statements.pkl', 'wb') as file:
pickle.dump(processor.statements, file, protocol=pickle.HIGHEST_PROTOCOL)
"""
from .api import get_drugbank_processor, process_xml
| Add more docs on how to get drugbank quickly | Add more docs on how to get drugbank quickly
| Python | bsd-2-clause | sorgerlab/belpy,sorgerlab/indra,bgyori/indra,sorgerlab/indra,johnbachman/indra,bgyori/indra,sorgerlab/belpy,sorgerlab/indra,johnbachman/indra,johnbachman/indra,sorgerlab/belpy,bgyori/indra | """This module provides an API and processor for DrugBank content.
It builds on the XML-formatted data schema of DrugBank and expects
the XML file to be available locally. The full DrugBank download
can be obtained at: https://www.drugbank.ca/releases/latest. Once the
XML file is decompressed, it can be processed using the process_xml
function."""
from .api import get_drugbank_processor, process_xml
Add more docs on how to get drugbank quickly | """This module provides an API and processor for DrugBank content.
It builds on the XML-formatted data schema of DrugBank and expects
the XML file to be available locally. The full DrugBank download
can be obtained at: https://www.drugbank.ca/releases/latest. Once the
XML file is decompressed, it can be processed using the process_xml
function.
Alternatively, the latest DrugBank data can be automatically loaded via
:mod:`drugbank_downloader` with the following code after doing
``pip install drugbank_downloader bioversions``:
.. code-block:: python
import pickle
import indra.sources.drugbank
processor = indra.sources.drugbank.get_drugbank_processor()
with open('drugbank_indra_statements.pkl', 'wb') as file:
pickle.dump(processor.statements, file, protocol=pickle.HIGHEST_PROTOCOL)
"""
from .api import get_drugbank_processor, process_xml
| <commit_before>"""This module provides an API and processor for DrugBank content.
It builds on the XML-formatted data schema of DrugBank and expects
the XML file to be available locally. The full DrugBank download
can be obtained at: https://www.drugbank.ca/releases/latest. Once the
XML file is decompressed, it can be processed using the process_xml
function."""
from .api import get_drugbank_processor, process_xml
<commit_msg>Add more docs on how to get drugbank quickly<commit_after> | """This module provides an API and processor for DrugBank content.
It builds on the XML-formatted data schema of DrugBank and expects
the XML file to be available locally. The full DrugBank download
can be obtained at: https://www.drugbank.ca/releases/latest. Once the
XML file is decompressed, it can be processed using the process_xml
function.
Alternatively, the latest DrugBank data can be automatically loaded via
:mod:`drugbank_downloader` with the following code after doing
``pip install drugbank_downloader bioversions``:
.. code-block:: python
import pickle
import indra.sources.drugbank
processor = indra.sources.drugbank.get_drugbank_processor()
with open('drugbank_indra_statements.pkl', 'wb') as file:
pickle.dump(processor.statements, file, protocol=pickle.HIGHEST_PROTOCOL)
"""
from .api import get_drugbank_processor, process_xml
| """This module provides an API and processor for DrugBank content.
It builds on the XML-formatted data schema of DrugBank and expects
the XML file to be available locally. The full DrugBank download
can be obtained at: https://www.drugbank.ca/releases/latest. Once the
XML file is decompressed, it can be processed using the process_xml
function."""
from .api import get_drugbank_processor, process_xml
Add more docs on how to get drugbank quickly"""This module provides an API and processor for DrugBank content.
It builds on the XML-formatted data schema of DrugBank and expects
the XML file to be available locally. The full DrugBank download
can be obtained at: https://www.drugbank.ca/releases/latest. Once the
XML file is decompressed, it can be processed using the process_xml
function.
Alternatively, the latest DrugBank data can be automatically loaded via
:mod:`drugbank_downloader` with the following code after doing
``pip install drugbank_downloader bioversions``:
.. code-block:: python
import pickle
import indra.sources.drugbank
processor = indra.sources.drugbank.get_drugbank_processor()
with open('drugbank_indra_statements.pkl', 'wb') as file:
pickle.dump(processor.statements, file, protocol=pickle.HIGHEST_PROTOCOL)
"""
from .api import get_drugbank_processor, process_xml
| <commit_before>"""This module provides an API and processor for DrugBank content.
It builds on the XML-formatted data schema of DrugBank and expects
the XML file to be available locally. The full DrugBank download
can be obtained at: https://www.drugbank.ca/releases/latest. Once the
XML file is decompressed, it can be processed using the process_xml
function."""
from .api import get_drugbank_processor, process_xml
<commit_msg>Add more docs on how to get drugbank quickly<commit_after>"""This module provides an API and processor for DrugBank content.
It builds on the XML-formatted data schema of DrugBank and expects
the XML file to be available locally. The full DrugBank download
can be obtained at: https://www.drugbank.ca/releases/latest. Once the
XML file is decompressed, it can be processed using the process_xml
function.
Alternatively, the latest DrugBank data can be automatically loaded via
:mod:`drugbank_downloader` with the following code after doing
``pip install drugbank_downloader bioversions``:
.. code-block:: python
import pickle
import indra.sources.drugbank
processor = indra.sources.drugbank.get_drugbank_processor()
with open('drugbank_indra_statements.pkl', 'wb') as file:
pickle.dump(processor.statements, file, protocol=pickle.HIGHEST_PROTOCOL)
"""
from .api import get_drugbank_processor, process_xml
|
922db591ca726acae07e2628119b95aa705f414c | leetcode/ds_string_word_pattern.py | leetcode/ds_string_word_pattern.py |
# @file Word Pattern
# @brief Given 2 sets check if it is a bijection
# https://leetcode.com/problems/word-pattern/
'''
Given a pattern and a string str, find if str follows the same pattern.
Here follow means a full match, such that there is a bijection between
a letter in pattern and a non-empty word in str.
Examples:
pattern = "abba", str = "dog cat cat dog" should return true.
pattern = "abba", str = "dog cat cat fish" should return false.
pattern = "aaaa", str = "dog cat cat dog" should return false.
pattern = "abba", str = "dog dog dog dog" should return false.
Notes:
You may assume pattern contains only lowercase letters, and str contains
lowercase letters separated by a single space.
'''
|
# @file Word Pattern
# @brief Given 2 sets check if it is a bijection
# https://leetcode.com/problems/word-pattern/
'''
Given a pattern and a string str, find if str follows the same pattern.
Here follow means a full match, such that there is a bijection between
a letter in pattern and a non-empty word in str.
Examples:
pattern = "abba", str = "dog cat cat dog" should return true.
pattern = "abba", str = "dog cat cat fish" should return false.
pattern = "aaaa", str = "dog cat cat dog" should return false.
pattern = "abba", str = "dog dog dog dog" should return false.
Notes:
You may assume pattern contains only lowercase letters, and str contains
lowercase letters separated by a single space.
'''
# Approach 1:
def wordPattern(self, pattern, str):
clist = pattern #treat string as a list of chars
wlist = str.split() #split string into a list of words
# map(function, sequence): map applies the given function to every element in the sequence and returns a list
# index - finds the index of the first occurence of every element in both list and string
return map(clist.index, clist) == map(wlist.index, wlist)
# Approach 2:
def wordPattern(self, pattern, str):
clist = pattern
wlist = str.split()
# zip returns a tuple, cpupling the ith elements from both lists
return len(clist) == len(wlist) and len(set(clist)) == len(set(wlist)) == len(set(zip(clist, wlist)))
# "abba", "dog cat cat dog", True.
# "abba", "dog cat cat fish" False.
# "aaaa", "dog cat cat dog" False.
# "abba", "dog dog dog dog" False.
| Add two approaches for string word pattern | Add two approaches for string word pattern | Python | mit | ngovindaraj/Python |
# @file Word Pattern
# @brief Given 2 sets check if it is a bijection
# https://leetcode.com/problems/word-pattern/
'''
Given a pattern and a string str, find if str follows the same pattern.
Here follow means a full match, such that there is a bijection between
a letter in pattern and a non-empty word in str.
Examples:
pattern = "abba", str = "dog cat cat dog" should return true.
pattern = "abba", str = "dog cat cat fish" should return false.
pattern = "aaaa", str = "dog cat cat dog" should return false.
pattern = "abba", str = "dog dog dog dog" should return false.
Notes:
You may assume pattern contains only lowercase letters, and str contains
lowercase letters separated by a single space.
'''
Add two approaches for string word pattern |
# @file Word Pattern
# @brief Given 2 sets check if it is a bijection
# https://leetcode.com/problems/word-pattern/
'''
Given a pattern and a string str, find if str follows the same pattern.
Here follow means a full match, such that there is a bijection between
a letter in pattern and a non-empty word in str.
Examples:
pattern = "abba", str = "dog cat cat dog" should return true.
pattern = "abba", str = "dog cat cat fish" should return false.
pattern = "aaaa", str = "dog cat cat dog" should return false.
pattern = "abba", str = "dog dog dog dog" should return false.
Notes:
You may assume pattern contains only lowercase letters, and str contains
lowercase letters separated by a single space.
'''
# Approach 1:
def wordPattern(self, pattern, str):
clist = pattern #treat string as a list of chars
wlist = str.split() #split string into a list of words
# map(function, sequence): map applies the given function to every element in the sequence and returns a list
# index - finds the index of the first occurence of every element in both list and string
return map(clist.index, clist) == map(wlist.index, wlist)
# Approach 2:
def wordPattern(self, pattern, str):
clist = pattern
wlist = str.split()
# zip returns a tuple, cpupling the ith elements from both lists
return len(clist) == len(wlist) and len(set(clist)) == len(set(wlist)) == len(set(zip(clist, wlist)))
# "abba", "dog cat cat dog", True.
# "abba", "dog cat cat fish" False.
# "aaaa", "dog cat cat dog" False.
# "abba", "dog dog dog dog" False.
| <commit_before>
# @file Word Pattern
# @brief Given 2 sets check if it is a bijection
# https://leetcode.com/problems/word-pattern/
'''
Given a pattern and a string str, find if str follows the same pattern.
Here follow means a full match, such that there is a bijection between
a letter in pattern and a non-empty word in str.
Examples:
pattern = "abba", str = "dog cat cat dog" should return true.
pattern = "abba", str = "dog cat cat fish" should return false.
pattern = "aaaa", str = "dog cat cat dog" should return false.
pattern = "abba", str = "dog dog dog dog" should return false.
Notes:
You may assume pattern contains only lowercase letters, and str contains
lowercase letters separated by a single space.
'''
<commit_msg>Add two approaches for string word pattern<commit_after> |
# @file Word Pattern
# @brief Given 2 sets check if it is a bijection
# https://leetcode.com/problems/word-pattern/
'''
Given a pattern and a string str, find if str follows the same pattern.
Here follow means a full match, such that there is a bijection between
a letter in pattern and a non-empty word in str.
Examples:
pattern = "abba", str = "dog cat cat dog" should return true.
pattern = "abba", str = "dog cat cat fish" should return false.
pattern = "aaaa", str = "dog cat cat dog" should return false.
pattern = "abba", str = "dog dog dog dog" should return false.
Notes:
You may assume pattern contains only lowercase letters, and str contains
lowercase letters separated by a single space.
'''
# Approach 1:
def wordPattern(self, pattern, str):
clist = pattern #treat string as a list of chars
wlist = str.split() #split string into a list of words
# map(function, sequence): map applies the given function to every element in the sequence and returns a list
# index - finds the index of the first occurence of every element in both list and string
return map(clist.index, clist) == map(wlist.index, wlist)
# Approach 2:
def wordPattern(self, pattern, str):
clist = pattern
wlist = str.split()
# zip returns a tuple, cpupling the ith elements from both lists
return len(clist) == len(wlist) and len(set(clist)) == len(set(wlist)) == len(set(zip(clist, wlist)))
# "abba", "dog cat cat dog", True.
# "abba", "dog cat cat fish" False.
# "aaaa", "dog cat cat dog" False.
# "abba", "dog dog dog dog" False.
|
# @file Word Pattern
# @brief Given 2 sets check if it is a bijection
# https://leetcode.com/problems/word-pattern/
'''
Given a pattern and a string str, find if str follows the same pattern.
Here follow means a full match, such that there is a bijection between
a letter in pattern and a non-empty word in str.
Examples:
pattern = "abba", str = "dog cat cat dog" should return true.
pattern = "abba", str = "dog cat cat fish" should return false.
pattern = "aaaa", str = "dog cat cat dog" should return false.
pattern = "abba", str = "dog dog dog dog" should return false.
Notes:
You may assume pattern contains only lowercase letters, and str contains
lowercase letters separated by a single space.
'''
Add two approaches for string word pattern
# @file Word Pattern
# @brief Given 2 sets check if it is a bijection
# https://leetcode.com/problems/word-pattern/
'''
Given a pattern and a string str, find if str follows the same pattern.
Here follow means a full match, such that there is a bijection between
a letter in pattern and a non-empty word in str.
Examples:
pattern = "abba", str = "dog cat cat dog" should return true.
pattern = "abba", str = "dog cat cat fish" should return false.
pattern = "aaaa", str = "dog cat cat dog" should return false.
pattern = "abba", str = "dog dog dog dog" should return false.
Notes:
You may assume pattern contains only lowercase letters, and str contains
lowercase letters separated by a single space.
'''
# Approach 1:
def wordPattern(self, pattern, str):
clist = pattern #treat string as a list of chars
wlist = str.split() #split string into a list of words
# map(function, sequence): map applies the given function to every element in the sequence and returns a list
# index - finds the index of the first occurence of every element in both list and string
return map(clist.index, clist) == map(wlist.index, wlist)
# Approach 2:
def wordPattern(self, pattern, str):
clist = pattern
wlist = str.split()
# zip returns a tuple, cpupling the ith elements from both lists
return len(clist) == len(wlist) and len(set(clist)) == len(set(wlist)) == len(set(zip(clist, wlist)))
# "abba", "dog cat cat dog", True.
# "abba", "dog cat cat fish" False.
# "aaaa", "dog cat cat dog" False.
# "abba", "dog dog dog dog" False.
| <commit_before>
# @file Word Pattern
# @brief Given 2 sets check if it is a bijection
# https://leetcode.com/problems/word-pattern/
'''
Given a pattern and a string str, find if str follows the same pattern.
Here follow means a full match, such that there is a bijection between
a letter in pattern and a non-empty word in str.
Examples:
pattern = "abba", str = "dog cat cat dog" should return true.
pattern = "abba", str = "dog cat cat fish" should return false.
pattern = "aaaa", str = "dog cat cat dog" should return false.
pattern = "abba", str = "dog dog dog dog" should return false.
Notes:
You may assume pattern contains only lowercase letters, and str contains
lowercase letters separated by a single space.
'''
<commit_msg>Add two approaches for string word pattern<commit_after>
# @file Word Pattern
# @brief Given 2 sets check if it is a bijection
# https://leetcode.com/problems/word-pattern/
'''
Given a pattern and a string str, find if str follows the same pattern.
Here follow means a full match, such that there is a bijection between
a letter in pattern and a non-empty word in str.
Examples:
pattern = "abba", str = "dog cat cat dog" should return true.
pattern = "abba", str = "dog cat cat fish" should return false.
pattern = "aaaa", str = "dog cat cat dog" should return false.
pattern = "abba", str = "dog dog dog dog" should return false.
Notes:
You may assume pattern contains only lowercase letters, and str contains
lowercase letters separated by a single space.
'''
# Approach 1:
def wordPattern(self, pattern, str):
clist = pattern #treat string as a list of chars
wlist = str.split() #split string into a list of words
# map(function, sequence): map applies the given function to every element in the sequence and returns a list
# index - finds the index of the first occurence of every element in both list and string
return map(clist.index, clist) == map(wlist.index, wlist)
# Approach 2:
def wordPattern(self, pattern, str):
clist = pattern
wlist = str.split()
# zip returns a tuple, cpupling the ith elements from both lists
return len(clist) == len(wlist) and len(set(clist)) == len(set(wlist)) == len(set(zip(clist, wlist)))
# "abba", "dog cat cat dog", True.
# "abba", "dog cat cat fish" False.
# "aaaa", "dog cat cat dog" False.
# "abba", "dog dog dog dog" False.
|
c1f270700d9de209577b64c40b71b5f5b69c5aae | cards.py | cards.py | from pymongo import MongoClient
class Cards:
def __init__(self, dbname='cards'):
"""Instantiate this class.
Set up a connection to the given Mongo database.
Get to the collection we'll store cards in.
Args:
dbname (str): Database name.
"""
self.client = MongoClient()
self.db = self.client[dbname]
self.cards_coll = self.db['cards']
@property
def sets(self):
"""Return a list of all the card sets in the database.
Args:
None
Returns:
list: List of all card sets in the database.
"""
return self.cards_coll.distinct('set')
def retrieve_set(self, setname):
"""Return a list of all the cards in the given set.
Args:
setname (str): Name of set.
Returns:
list: List of all cards in the the given set.
"""
return list(self.cards_coll.find({'set': setname}))
def create_card(self, setname, color, text, creator):
"""Insert a new card with the given properties into the database.
Args:
setname (str): Name of set the card will belong to.
color (str): Color the card will have.
text (str): Text that will appear on the card.
creator (str): Creator to attribute the card to.
Returns:
None
"""
card = {
'set': setname,
'color': color,
'text': text,
'creator': creator,
}
self.cards_coll.insert_one(card)
| from pymongo import MongoClient
class Cards:
def __init__(self, dbname='cards'):
"""Instantiate this class.
Set up a connection to the given Mongo database.
Get to the collection we'll store cards in.
Args:
dbname (str): Database name.
"""
self.client = MongoClient()
self.db = self.client[dbname]
self.cards_coll = self.db['cards']
@property
def sets(self):
"""Return a list of all the card sets in the database.
Args:
None
Returns:
list: List of all card sets in the database.
"""
return self.cards_coll.distinct('set')
def retrieve_set(self, setname):
"""Return a list of all the cards in the given set.
Args:
setname (str): Name of set.
Returns:
list: List of all cards in the the given set.
"""
return list(self.cards_coll.find({'set': setname}))
def create_cards(self, cards):
"""Insert a new card with the given properties into the database.
Args:
cards: List of dictionaries with set, color, text, and creator keys.
Returns:
None
"""
keys = ['set', 'color', 'text', 'creator']
filtered = [ { k: card[k] for k in keys if k in card} for card in cards]
self.cards_coll.insert_many(filtered)
| Refactor create_card method to take a list of card dictionaries. Rename method accordingly. | Refactor create_card method to take a list of card dictionaries. Rename method accordingly.
| Python | isc | wwu-nosql/cards | from pymongo import MongoClient
class Cards:
def __init__(self, dbname='cards'):
"""Instantiate this class.
Set up a connection to the given Mongo database.
Get to the collection we'll store cards in.
Args:
dbname (str): Database name.
"""
self.client = MongoClient()
self.db = self.client[dbname]
self.cards_coll = self.db['cards']
@property
def sets(self):
"""Return a list of all the card sets in the database.
Args:
None
Returns:
list: List of all card sets in the database.
"""
return self.cards_coll.distinct('set')
def retrieve_set(self, setname):
"""Return a list of all the cards in the given set.
Args:
setname (str): Name of set.
Returns:
list: List of all cards in the the given set.
"""
return list(self.cards_coll.find({'set': setname}))
def create_card(self, setname, color, text, creator):
"""Insert a new card with the given properties into the database.
Args:
setname (str): Name of set the card will belong to.
color (str): Color the card will have.
text (str): Text that will appear on the card.
creator (str): Creator to attribute the card to.
Returns:
None
"""
card = {
'set': setname,
'color': color,
'text': text,
'creator': creator,
}
self.cards_coll.insert_one(card)
Refactor create_card method to take a list of card dictionaries. Rename method accordingly. | from pymongo import MongoClient
class Cards:
def __init__(self, dbname='cards'):
"""Instantiate this class.
Set up a connection to the given Mongo database.
Get to the collection we'll store cards in.
Args:
dbname (str): Database name.
"""
self.client = MongoClient()
self.db = self.client[dbname]
self.cards_coll = self.db['cards']
@property
def sets(self):
"""Return a list of all the card sets in the database.
Args:
None
Returns:
list: List of all card sets in the database.
"""
return self.cards_coll.distinct('set')
def retrieve_set(self, setname):
"""Return a list of all the cards in the given set.
Args:
setname (str): Name of set.
Returns:
list: List of all cards in the the given set.
"""
return list(self.cards_coll.find({'set': setname}))
def create_cards(self, cards):
"""Insert a new card with the given properties into the database.
Args:
cards: List of dictionaries with set, color, text, and creator keys.
Returns:
None
"""
keys = ['set', 'color', 'text', 'creator']
filtered = [ { k: card[k] for k in keys if k in card} for card in cards]
self.cards_coll.insert_many(filtered)
| <commit_before>from pymongo import MongoClient
class Cards:
def __init__(self, dbname='cards'):
"""Instantiate this class.
Set up a connection to the given Mongo database.
Get to the collection we'll store cards in.
Args:
dbname (str): Database name.
"""
self.client = MongoClient()
self.db = self.client[dbname]
self.cards_coll = self.db['cards']
@property
def sets(self):
"""Return a list of all the card sets in the database.
Args:
None
Returns:
list: List of all card sets in the database.
"""
return self.cards_coll.distinct('set')
def retrieve_set(self, setname):
"""Return a list of all the cards in the given set.
Args:
setname (str): Name of set.
Returns:
list: List of all cards in the the given set.
"""
return list(self.cards_coll.find({'set': setname}))
def create_card(self, setname, color, text, creator):
"""Insert a new card with the given properties into the database.
Args:
setname (str): Name of set the card will belong to.
color (str): Color the card will have.
text (str): Text that will appear on the card.
creator (str): Creator to attribute the card to.
Returns:
None
"""
card = {
'set': setname,
'color': color,
'text': text,
'creator': creator,
}
self.cards_coll.insert_one(card)
<commit_msg>Refactor create_card method to take a list of card dictionaries. Rename method accordingly.<commit_after> | from pymongo import MongoClient
class Cards:
def __init__(self, dbname='cards'):
"""Instantiate this class.
Set up a connection to the given Mongo database.
Get to the collection we'll store cards in.
Args:
dbname (str): Database name.
"""
self.client = MongoClient()
self.db = self.client[dbname]
self.cards_coll = self.db['cards']
@property
def sets(self):
"""Return a list of all the card sets in the database.
Args:
None
Returns:
list: List of all card sets in the database.
"""
return self.cards_coll.distinct('set')
def retrieve_set(self, setname):
"""Return a list of all the cards in the given set.
Args:
setname (str): Name of set.
Returns:
list: List of all cards in the the given set.
"""
return list(self.cards_coll.find({'set': setname}))
def create_cards(self, cards):
"""Insert a new card with the given properties into the database.
Args:
cards: List of dictionaries with set, color, text, and creator keys.
Returns:
None
"""
keys = ['set', 'color', 'text', 'creator']
filtered = [ { k: card[k] for k in keys if k in card} for card in cards]
self.cards_coll.insert_many(filtered)
| from pymongo import MongoClient
class Cards:
def __init__(self, dbname='cards'):
"""Instantiate this class.
Set up a connection to the given Mongo database.
Get to the collection we'll store cards in.
Args:
dbname (str): Database name.
"""
self.client = MongoClient()
self.db = self.client[dbname]
self.cards_coll = self.db['cards']
@property
def sets(self):
"""Return a list of all the card sets in the database.
Args:
None
Returns:
list: List of all card sets in the database.
"""
return self.cards_coll.distinct('set')
def retrieve_set(self, setname):
"""Return a list of all the cards in the given set.
Args:
setname (str): Name of set.
Returns:
list: List of all cards in the the given set.
"""
return list(self.cards_coll.find({'set': setname}))
def create_card(self, setname, color, text, creator):
"""Insert a new card with the given properties into the database.
Args:
setname (str): Name of set the card will belong to.
color (str): Color the card will have.
text (str): Text that will appear on the card.
creator (str): Creator to attribute the card to.
Returns:
None
"""
card = {
'set': setname,
'color': color,
'text': text,
'creator': creator,
}
self.cards_coll.insert_one(card)
Refactor create_card method to take a list of card dictionaries. Rename method accordingly.from pymongo import MongoClient
class Cards:
def __init__(self, dbname='cards'):
"""Instantiate this class.
Set up a connection to the given Mongo database.
Get to the collection we'll store cards in.
Args:
dbname (str): Database name.
"""
self.client = MongoClient()
self.db = self.client[dbname]
self.cards_coll = self.db['cards']
@property
def sets(self):
"""Return a list of all the card sets in the database.
Args:
None
Returns:
list: List of all card sets in the database.
"""
return self.cards_coll.distinct('set')
def retrieve_set(self, setname):
"""Return a list of all the cards in the given set.
Args:
setname (str): Name of set.
Returns:
list: List of all cards in the the given set.
"""
return list(self.cards_coll.find({'set': setname}))
def create_cards(self, cards):
"""Insert a new card with the given properties into the database.
Args:
cards: List of dictionaries with set, color, text, and creator keys.
Returns:
None
"""
keys = ['set', 'color', 'text', 'creator']
filtered = [ { k: card[k] for k in keys if k in card} for card in cards]
self.cards_coll.insert_many(filtered)
| <commit_before>from pymongo import MongoClient
class Cards:
def __init__(self, dbname='cards'):
"""Instantiate this class.
Set up a connection to the given Mongo database.
Get to the collection we'll store cards in.
Args:
dbname (str): Database name.
"""
self.client = MongoClient()
self.db = self.client[dbname]
self.cards_coll = self.db['cards']
@property
def sets(self):
"""Return a list of all the card sets in the database.
Args:
None
Returns:
list: List of all card sets in the database.
"""
return self.cards_coll.distinct('set')
def retrieve_set(self, setname):
"""Return a list of all the cards in the given set.
Args:
setname (str): Name of set.
Returns:
list: List of all cards in the the given set.
"""
return list(self.cards_coll.find({'set': setname}))
def create_card(self, setname, color, text, creator):
"""Insert a new card with the given properties into the database.
Args:
setname (str): Name of set the card will belong to.
color (str): Color the card will have.
text (str): Text that will appear on the card.
creator (str): Creator to attribute the card to.
Returns:
None
"""
card = {
'set': setname,
'color': color,
'text': text,
'creator': creator,
}
self.cards_coll.insert_one(card)
<commit_msg>Refactor create_card method to take a list of card dictionaries. Rename method accordingly.<commit_after>from pymongo import MongoClient
class Cards:
def __init__(self, dbname='cards'):
"""Instantiate this class.
Set up a connection to the given Mongo database.
Get to the collection we'll store cards in.
Args:
dbname (str): Database name.
"""
self.client = MongoClient()
self.db = self.client[dbname]
self.cards_coll = self.db['cards']
@property
def sets(self):
"""Return a list of all the card sets in the database.
Args:
None
Returns:
list: List of all card sets in the database.
"""
return self.cards_coll.distinct('set')
def retrieve_set(self, setname):
"""Return a list of all the cards in the given set.
Args:
setname (str): Name of set.
Returns:
list: List of all cards in the the given set.
"""
return list(self.cards_coll.find({'set': setname}))
def create_cards(self, cards):
"""Insert a new card with the given properties into the database.
Args:
cards: List of dictionaries with set, color, text, and creator keys.
Returns:
None
"""
keys = ['set', 'color', 'text', 'creator']
filtered = [ { k: card[k] for k in keys if k in card} for card in cards]
self.cards_coll.insert_many(filtered)
|
f70b2a187c274685cb19def1b48256822b0e3f9f | tests/conftest.py | tests/conftest.py | import py
import pytest
from tests.lib.path import Path
@pytest.fixture
def tmpdir(request):
"""
Return a temporary directory path object which is unique to each test
function invocation, created as a sub directory of the base temporary
directory. The returned object is a ``tests.lib.path.Path`` object.
This is taken from pytest itself but modified to return our typical
path object instead of py.path.local.
"""
name = request.node.name
name = py.std.re.sub("[\W]", "_", name)
tmp = request.config._tmpdirhandler.mktemp(name, numbered=True)
return Path(tmp)
| import py
import pytest
from tests.lib.path import Path
from tests.lib.scripttest import PipTestEnvironment
from tests.lib.venv import VirtualEnvironment
@pytest.fixture
def tmpdir(request):
"""
Return a temporary directory path object which is unique to each test
function invocation, created as a sub directory of the base temporary
directory. The returned object is a ``tests.lib.path.Path`` object.
This is taken from pytest itself but modified to return our typical
path object instead of py.path.local.
"""
name = request.node.name
name = py.std.re.sub("[\W]", "_", name)
tmp = request.config._tmpdirhandler.mktemp(name, numbered=True)
return Path(tmp)
@pytest.fixture
def virtualenv(tmpdir):
"""
Return a virtual environment which is unique to each test function
invocation created inside of a sub directory of the test function's
temporary directory. The returned object is a
``tests.lib.venv.VirtualEnvironment`` object.
"""
return VirtualEnvironment.create(tmpdir.join(".venv"))
@pytest.fixture
def script(tmpdir, virtualenv):
"""
Return a PipTestEnvironment which is unique to each test function and
will execute all commands inside of the unique virtual environment for this
test function. The returned object is a
``tests.lib.scripttest.PipTestEnvironment``.
"""
return PipTestEnvironment(
# The base location for our test environment
tmpdir,
# Tell the Test Environment where our virtualenv is located
virtualenv=virtualenv.location,
# Do not ignore hidden files, they need to be checked as well
ignore_hidden=False,
# We are starting with an already empty directory
start_clear=False,
# We want to ensure no temporary files are left behind, so the
# PipTestEnvironment needs to capture and assert against temp
capture_temp=True,
assert_no_temp=True,
)
| Use our classes from tests.lib.* in pytest fixtures | Use our classes from tests.lib.* in pytest fixtures
| Python | mit | KarelJakubec/pip,qbdsoft/pip,alquerci/pip,nthall/pip,dstufft/pip,nthall/pip,esc/pip,blarghmatey/pip,chaoallsome/pip,wkeyword/pip,alex/pip,jmagnusson/pip,James-Firth/pip,msabramo/pip,atdaemon/pip,natefoo/pip,rouge8/pip,h4ck3rm1k3/pip,zvezdan/pip,haridsv/pip,rbtcollins/pip,Ivoz/pip,luzfcb/pip,harrisonfeng/pip,willingc/pip,qbdsoft/pip,natefoo/pip,harrisonfeng/pip,Gabriel439/pip,Gabriel439/pip,caosmo/pip,wkeyword/pip,RonnyPfannschmidt/pip,habnabit/pip,yati-sagade/pip,squidsoup/pip,squidsoup/pip,fiber-space/pip,pjdelport/pip,haridsv/pip,h4ck3rm1k3/pip,erikrose/pip,ianw/pip,techtonik/pip,benesch/pip,supriyantomaftuh/pip,atdaemon/pip,mindw/pip,nthall/pip,zorosteven/pip,rbtcollins/pip,habnabit/pip,alquerci/pip,cjerdonek/pip,tdsmith/pip,graingert/pip,squidsoup/pip,supriyantomaftuh/pip,harrisonfeng/pip,qwcode/pip,jamezpolley/pip,qwcode/pip,pradyunsg/pip,dstufft/pip,fiber-space/pip,alex/pip,rouge8/pip,pypa/pip,ncoghlan/pip,graingert/pip,Gabriel439/pip,rbtcollins/pip,fiber-space/pip,ChristopherHogan/pip,prasaianooz/pip,mattrobenolt/pip,jmagnusson/pip,RonnyPfannschmidt/pip,esc/pip,yati-sagade/pip,prasaianooz/pip,jythontools/pip,sigmavirus24/pip,chaoallsome/pip,qbdsoft/pip,minrk/pip,haridsv/pip,dstufft/pip,alex/pip,willingc/pip,ncoghlan/pip,RonnyPfannschmidt/pip,h4ck3rm1k3/pip,zvezdan/pip,patricklaw/pip,mindw/pip,jasonkying/pip,sigmavirus24/pip,mattrobenolt/pip,jythontools/pip,benesch/pip,xavfernandez/pip,supriyantomaftuh/pip,atdaemon/pip,sbidoul/pip,rouge8/pip,mujiansu/pip,mujiansu/pip,jythontools/pip,pjdelport/pip,ncoghlan/pip,blarghmatey/pip,jamezpolley/pip,patricklaw/pip,mindw/pip,sbidoul/pip,jasonkying/pip,zenlambda/pip,prasaianooz/pip,chaoallsome/pip,sigmavirus24/pip,tdsmith/pip,pjdelport/pip,Carreau/pip,graingert/pip,ChristopherHogan/pip,erikrose/pip,tdsmith/pip,zenlambda/pip,ianw/pip,pradyunsg/pip,mujiansu/pip,zorosteven/pip,davidovich/pip,techtonik/pip,benesch/pip,habnabit/pip,zvezdan/pip,pypa/pip,zorosteven/pip,yati-sagade/pip,luzfcb/pip,davidovich/pip,jmagnusson/pip,techtonik/pip,xavfernandez/pip,erikrose/pip,cjerdonek/pip,esc/pip,caosmo/pip,jamezpolley/pip,Carreau/pip,James-Firth/pip,caosmo/pip,wkeyword/pip,KarelJakubec/pip,jasonkying/pip,minrk/pip,luzfcb/pip,KarelJakubec/pip,xavfernandez/pip,Ivoz/pip,ChristopherHogan/pip,pfmoore/pip,blarghmatey/pip,pfmoore/pip,natefoo/pip,davidovich/pip,James-Firth/pip,willingc/pip,msabramo/pip,zenlambda/pip | import py
import pytest
from tests.lib.path import Path
@pytest.fixture
def tmpdir(request):
"""
Return a temporary directory path object which is unique to each test
function invocation, created as a sub directory of the base temporary
directory. The returned object is a ``tests.lib.path.Path`` object.
This is taken from pytest itself but modified to return our typical
path object instead of py.path.local.
"""
name = request.node.name
name = py.std.re.sub("[\W]", "_", name)
tmp = request.config._tmpdirhandler.mktemp(name, numbered=True)
return Path(tmp)
Use our classes from tests.lib.* in pytest fixtures | import py
import pytest
from tests.lib.path import Path
from tests.lib.scripttest import PipTestEnvironment
from tests.lib.venv import VirtualEnvironment
@pytest.fixture
def tmpdir(request):
"""
Return a temporary directory path object which is unique to each test
function invocation, created as a sub directory of the base temporary
directory. The returned object is a ``tests.lib.path.Path`` object.
This is taken from pytest itself but modified to return our typical
path object instead of py.path.local.
"""
name = request.node.name
name = py.std.re.sub("[\W]", "_", name)
tmp = request.config._tmpdirhandler.mktemp(name, numbered=True)
return Path(tmp)
@pytest.fixture
def virtualenv(tmpdir):
"""
Return a virtual environment which is unique to each test function
invocation created inside of a sub directory of the test function's
temporary directory. The returned object is a
``tests.lib.venv.VirtualEnvironment`` object.
"""
return VirtualEnvironment.create(tmpdir.join(".venv"))
@pytest.fixture
def script(tmpdir, virtualenv):
"""
Return a PipTestEnvironment which is unique to each test function and
will execute all commands inside of the unique virtual environment for this
test function. The returned object is a
``tests.lib.scripttest.PipTestEnvironment``.
"""
return PipTestEnvironment(
# The base location for our test environment
tmpdir,
# Tell the Test Environment where our virtualenv is located
virtualenv=virtualenv.location,
# Do not ignore hidden files, they need to be checked as well
ignore_hidden=False,
# We are starting with an already empty directory
start_clear=False,
# We want to ensure no temporary files are left behind, so the
# PipTestEnvironment needs to capture and assert against temp
capture_temp=True,
assert_no_temp=True,
)
| <commit_before>import py
import pytest
from tests.lib.path import Path
@pytest.fixture
def tmpdir(request):
"""
Return a temporary directory path object which is unique to each test
function invocation, created as a sub directory of the base temporary
directory. The returned object is a ``tests.lib.path.Path`` object.
This is taken from pytest itself but modified to return our typical
path object instead of py.path.local.
"""
name = request.node.name
name = py.std.re.sub("[\W]", "_", name)
tmp = request.config._tmpdirhandler.mktemp(name, numbered=True)
return Path(tmp)
<commit_msg>Use our classes from tests.lib.* in pytest fixtures<commit_after> | import py
import pytest
from tests.lib.path import Path
from tests.lib.scripttest import PipTestEnvironment
from tests.lib.venv import VirtualEnvironment
@pytest.fixture
def tmpdir(request):
"""
Return a temporary directory path object which is unique to each test
function invocation, created as a sub directory of the base temporary
directory. The returned object is a ``tests.lib.path.Path`` object.
This is taken from pytest itself but modified to return our typical
path object instead of py.path.local.
"""
name = request.node.name
name = py.std.re.sub("[\W]", "_", name)
tmp = request.config._tmpdirhandler.mktemp(name, numbered=True)
return Path(tmp)
@pytest.fixture
def virtualenv(tmpdir):
"""
Return a virtual environment which is unique to each test function
invocation created inside of a sub directory of the test function's
temporary directory. The returned object is a
``tests.lib.venv.VirtualEnvironment`` object.
"""
return VirtualEnvironment.create(tmpdir.join(".venv"))
@pytest.fixture
def script(tmpdir, virtualenv):
"""
Return a PipTestEnvironment which is unique to each test function and
will execute all commands inside of the unique virtual environment for this
test function. The returned object is a
``tests.lib.scripttest.PipTestEnvironment``.
"""
return PipTestEnvironment(
# The base location for our test environment
tmpdir,
# Tell the Test Environment where our virtualenv is located
virtualenv=virtualenv.location,
# Do not ignore hidden files, they need to be checked as well
ignore_hidden=False,
# We are starting with an already empty directory
start_clear=False,
# We want to ensure no temporary files are left behind, so the
# PipTestEnvironment needs to capture and assert against temp
capture_temp=True,
assert_no_temp=True,
)
| import py
import pytest
from tests.lib.path import Path
@pytest.fixture
def tmpdir(request):
"""
Return a temporary directory path object which is unique to each test
function invocation, created as a sub directory of the base temporary
directory. The returned object is a ``tests.lib.path.Path`` object.
This is taken from pytest itself but modified to return our typical
path object instead of py.path.local.
"""
name = request.node.name
name = py.std.re.sub("[\W]", "_", name)
tmp = request.config._tmpdirhandler.mktemp(name, numbered=True)
return Path(tmp)
Use our classes from tests.lib.* in pytest fixturesimport py
import pytest
from tests.lib.path import Path
from tests.lib.scripttest import PipTestEnvironment
from tests.lib.venv import VirtualEnvironment
@pytest.fixture
def tmpdir(request):
"""
Return a temporary directory path object which is unique to each test
function invocation, created as a sub directory of the base temporary
directory. The returned object is a ``tests.lib.path.Path`` object.
This is taken from pytest itself but modified to return our typical
path object instead of py.path.local.
"""
name = request.node.name
name = py.std.re.sub("[\W]", "_", name)
tmp = request.config._tmpdirhandler.mktemp(name, numbered=True)
return Path(tmp)
@pytest.fixture
def virtualenv(tmpdir):
"""
Return a virtual environment which is unique to each test function
invocation created inside of a sub directory of the test function's
temporary directory. The returned object is a
``tests.lib.venv.VirtualEnvironment`` object.
"""
return VirtualEnvironment.create(tmpdir.join(".venv"))
@pytest.fixture
def script(tmpdir, virtualenv):
"""
Return a PipTestEnvironment which is unique to each test function and
will execute all commands inside of the unique virtual environment for this
test function. The returned object is a
``tests.lib.scripttest.PipTestEnvironment``.
"""
return PipTestEnvironment(
# The base location for our test environment
tmpdir,
# Tell the Test Environment where our virtualenv is located
virtualenv=virtualenv.location,
# Do not ignore hidden files, they need to be checked as well
ignore_hidden=False,
# We are starting with an already empty directory
start_clear=False,
# We want to ensure no temporary files are left behind, so the
# PipTestEnvironment needs to capture and assert against temp
capture_temp=True,
assert_no_temp=True,
)
| <commit_before>import py
import pytest
from tests.lib.path import Path
@pytest.fixture
def tmpdir(request):
"""
Return a temporary directory path object which is unique to each test
function invocation, created as a sub directory of the base temporary
directory. The returned object is a ``tests.lib.path.Path`` object.
This is taken from pytest itself but modified to return our typical
path object instead of py.path.local.
"""
name = request.node.name
name = py.std.re.sub("[\W]", "_", name)
tmp = request.config._tmpdirhandler.mktemp(name, numbered=True)
return Path(tmp)
<commit_msg>Use our classes from tests.lib.* in pytest fixtures<commit_after>import py
import pytest
from tests.lib.path import Path
from tests.lib.scripttest import PipTestEnvironment
from tests.lib.venv import VirtualEnvironment
@pytest.fixture
def tmpdir(request):
"""
Return a temporary directory path object which is unique to each test
function invocation, created as a sub directory of the base temporary
directory. The returned object is a ``tests.lib.path.Path`` object.
This is taken from pytest itself but modified to return our typical
path object instead of py.path.local.
"""
name = request.node.name
name = py.std.re.sub("[\W]", "_", name)
tmp = request.config._tmpdirhandler.mktemp(name, numbered=True)
return Path(tmp)
@pytest.fixture
def virtualenv(tmpdir):
"""
Return a virtual environment which is unique to each test function
invocation created inside of a sub directory of the test function's
temporary directory. The returned object is a
``tests.lib.venv.VirtualEnvironment`` object.
"""
return VirtualEnvironment.create(tmpdir.join(".venv"))
@pytest.fixture
def script(tmpdir, virtualenv):
"""
Return a PipTestEnvironment which is unique to each test function and
will execute all commands inside of the unique virtual environment for this
test function. The returned object is a
``tests.lib.scripttest.PipTestEnvironment``.
"""
return PipTestEnvironment(
# The base location for our test environment
tmpdir,
# Tell the Test Environment where our virtualenv is located
virtualenv=virtualenv.location,
# Do not ignore hidden files, they need to be checked as well
ignore_hidden=False,
# We are starting with an already empty directory
start_clear=False,
# We want to ensure no temporary files are left behind, so the
# PipTestEnvironment needs to capture and assert against temp
capture_temp=True,
assert_no_temp=True,
)
|
ff3123e21e366a5908655dbd8130ac60ec5eee10 | uvt_user/utils.py | uvt_user/utils.py | from __future__ import unicode_literals
import re
from ldap3 import Server, Connection
class LDAPError(Exception):
pass
def search_ldap(username):
'''Searches the Tilburg University LDAP server for the given username and returns a tuple of first name, last name, full name, ANR, emplId and email address. Permission has been granted by TiU's legal department for retrieving this data. Raises LDAPError on any kind of error.'''
result = ()
baseDN = "o=Universiteit van Tilburg,c=NL"
searchFilter = '(uid={})'.format(username)
attributes = ['givenName', 'sortableSurname', 'cn', 'employeeNumber', 'emplId', 'mail']
try:
server = Server('ldaps.uvt.nl', use_ssl=True)
conn = Connection(server, auto_bind=True)
conn.search(baseDN, searchFilter, attributes=attributes)
response = conn.response[0]['attributes']
for a in attributes:
if a in response:
result += (response[a][0],)
else:
result += ('',)
except IndexError:
raise LDAPError('Username not found')
except Exception:
raise LDAPError('Unknown error in LDAP query')
return result
| from __future__ import unicode_literals
import re
from ldap3 import Server, Connection
class LDAPError(Exception):
pass
def search_ldap(username):
'''Searches the Tilburg University LDAP server for the given username and returns a tuple of first name, last name, full name, ANR, emplId and email address. Permission has been granted by TiU's legal department for retrieving this data. Raises LDAPError on any kind of error.'''
result = ()
baseDN = "o=Universiteit van Tilburg,c=NL"
searchFilter = '(uid={})'.format(username)
attributes = ['givenName', 'sortableSurname', 'cn', 'employeeNumber', 'emplId', 'mail']
try:
server = Server('ldaps.uvt.nl', use_ssl=True)
conn = Connection(server, auto_bind=True)
conn.search(baseDN, searchFilter, attributes=attributes)
response = conn.response[0]['attributes']
for a in attributes:
if a in response and response[a]:
if type(response[a]) is list:
result += (response[a][0],)
else:
result += (response[a],)
else:
result += ('',)
except IndexError:
raise LDAPError('Username not found')
except Exception:
raise LDAPError('Unknown error in LDAP query')
return result
| Fix potential bug in case LDAP does not return lists | Fix potential bug in case LDAP does not return lists
| Python | agpl-3.0 | JaapJoris/bps,JaapJoris/bps | from __future__ import unicode_literals
import re
from ldap3 import Server, Connection
class LDAPError(Exception):
pass
def search_ldap(username):
'''Searches the Tilburg University LDAP server for the given username and returns a tuple of first name, last name, full name, ANR, emplId and email address. Permission has been granted by TiU's legal department for retrieving this data. Raises LDAPError on any kind of error.'''
result = ()
baseDN = "o=Universiteit van Tilburg,c=NL"
searchFilter = '(uid={})'.format(username)
attributes = ['givenName', 'sortableSurname', 'cn', 'employeeNumber', 'emplId', 'mail']
try:
server = Server('ldaps.uvt.nl', use_ssl=True)
conn = Connection(server, auto_bind=True)
conn.search(baseDN, searchFilter, attributes=attributes)
response = conn.response[0]['attributes']
for a in attributes:
if a in response:
result += (response[a][0],)
else:
result += ('',)
except IndexError:
raise LDAPError('Username not found')
except Exception:
raise LDAPError('Unknown error in LDAP query')
return result
Fix potential bug in case LDAP does not return lists | from __future__ import unicode_literals
import re
from ldap3 import Server, Connection
class LDAPError(Exception):
pass
def search_ldap(username):
'''Searches the Tilburg University LDAP server for the given username and returns a tuple of first name, last name, full name, ANR, emplId and email address. Permission has been granted by TiU's legal department for retrieving this data. Raises LDAPError on any kind of error.'''
result = ()
baseDN = "o=Universiteit van Tilburg,c=NL"
searchFilter = '(uid={})'.format(username)
attributes = ['givenName', 'sortableSurname', 'cn', 'employeeNumber', 'emplId', 'mail']
try:
server = Server('ldaps.uvt.nl', use_ssl=True)
conn = Connection(server, auto_bind=True)
conn.search(baseDN, searchFilter, attributes=attributes)
response = conn.response[0]['attributes']
for a in attributes:
if a in response and response[a]:
if type(response[a]) is list:
result += (response[a][0],)
else:
result += (response[a],)
else:
result += ('',)
except IndexError:
raise LDAPError('Username not found')
except Exception:
raise LDAPError('Unknown error in LDAP query')
return result
| <commit_before>from __future__ import unicode_literals
import re
from ldap3 import Server, Connection
class LDAPError(Exception):
pass
def search_ldap(username):
'''Searches the Tilburg University LDAP server for the given username and returns a tuple of first name, last name, full name, ANR, emplId and email address. Permission has been granted by TiU's legal department for retrieving this data. Raises LDAPError on any kind of error.'''
result = ()
baseDN = "o=Universiteit van Tilburg,c=NL"
searchFilter = '(uid={})'.format(username)
attributes = ['givenName', 'sortableSurname', 'cn', 'employeeNumber', 'emplId', 'mail']
try:
server = Server('ldaps.uvt.nl', use_ssl=True)
conn = Connection(server, auto_bind=True)
conn.search(baseDN, searchFilter, attributes=attributes)
response = conn.response[0]['attributes']
for a in attributes:
if a in response:
result += (response[a][0],)
else:
result += ('',)
except IndexError:
raise LDAPError('Username not found')
except Exception:
raise LDAPError('Unknown error in LDAP query')
return result
<commit_msg>Fix potential bug in case LDAP does not return lists<commit_after> | from __future__ import unicode_literals
import re
from ldap3 import Server, Connection
class LDAPError(Exception):
pass
def search_ldap(username):
'''Searches the Tilburg University LDAP server for the given username and returns a tuple of first name, last name, full name, ANR, emplId and email address. Permission has been granted by TiU's legal department for retrieving this data. Raises LDAPError on any kind of error.'''
result = ()
baseDN = "o=Universiteit van Tilburg,c=NL"
searchFilter = '(uid={})'.format(username)
attributes = ['givenName', 'sortableSurname', 'cn', 'employeeNumber', 'emplId', 'mail']
try:
server = Server('ldaps.uvt.nl', use_ssl=True)
conn = Connection(server, auto_bind=True)
conn.search(baseDN, searchFilter, attributes=attributes)
response = conn.response[0]['attributes']
for a in attributes:
if a in response and response[a]:
if type(response[a]) is list:
result += (response[a][0],)
else:
result += (response[a],)
else:
result += ('',)
except IndexError:
raise LDAPError('Username not found')
except Exception:
raise LDAPError('Unknown error in LDAP query')
return result
| from __future__ import unicode_literals
import re
from ldap3 import Server, Connection
class LDAPError(Exception):
pass
def search_ldap(username):
'''Searches the Tilburg University LDAP server for the given username and returns a tuple of first name, last name, full name, ANR, emplId and email address. Permission has been granted by TiU's legal department for retrieving this data. Raises LDAPError on any kind of error.'''
result = ()
baseDN = "o=Universiteit van Tilburg,c=NL"
searchFilter = '(uid={})'.format(username)
attributes = ['givenName', 'sortableSurname', 'cn', 'employeeNumber', 'emplId', 'mail']
try:
server = Server('ldaps.uvt.nl', use_ssl=True)
conn = Connection(server, auto_bind=True)
conn.search(baseDN, searchFilter, attributes=attributes)
response = conn.response[0]['attributes']
for a in attributes:
if a in response:
result += (response[a][0],)
else:
result += ('',)
except IndexError:
raise LDAPError('Username not found')
except Exception:
raise LDAPError('Unknown error in LDAP query')
return result
Fix potential bug in case LDAP does not return listsfrom __future__ import unicode_literals
import re
from ldap3 import Server, Connection
class LDAPError(Exception):
pass
def search_ldap(username):
'''Searches the Tilburg University LDAP server for the given username and returns a tuple of first name, last name, full name, ANR, emplId and email address. Permission has been granted by TiU's legal department for retrieving this data. Raises LDAPError on any kind of error.'''
result = ()
baseDN = "o=Universiteit van Tilburg,c=NL"
searchFilter = '(uid={})'.format(username)
attributes = ['givenName', 'sortableSurname', 'cn', 'employeeNumber', 'emplId', 'mail']
try:
server = Server('ldaps.uvt.nl', use_ssl=True)
conn = Connection(server, auto_bind=True)
conn.search(baseDN, searchFilter, attributes=attributes)
response = conn.response[0]['attributes']
for a in attributes:
if a in response and response[a]:
if type(response[a]) is list:
result += (response[a][0],)
else:
result += (response[a],)
else:
result += ('',)
except IndexError:
raise LDAPError('Username not found')
except Exception:
raise LDAPError('Unknown error in LDAP query')
return result
| <commit_before>from __future__ import unicode_literals
import re
from ldap3 import Server, Connection
class LDAPError(Exception):
pass
def search_ldap(username):
'''Searches the Tilburg University LDAP server for the given username and returns a tuple of first name, last name, full name, ANR, emplId and email address. Permission has been granted by TiU's legal department for retrieving this data. Raises LDAPError on any kind of error.'''
result = ()
baseDN = "o=Universiteit van Tilburg,c=NL"
searchFilter = '(uid={})'.format(username)
attributes = ['givenName', 'sortableSurname', 'cn', 'employeeNumber', 'emplId', 'mail']
try:
server = Server('ldaps.uvt.nl', use_ssl=True)
conn = Connection(server, auto_bind=True)
conn.search(baseDN, searchFilter, attributes=attributes)
response = conn.response[0]['attributes']
for a in attributes:
if a in response:
result += (response[a][0],)
else:
result += ('',)
except IndexError:
raise LDAPError('Username not found')
except Exception:
raise LDAPError('Unknown error in LDAP query')
return result
<commit_msg>Fix potential bug in case LDAP does not return lists<commit_after>from __future__ import unicode_literals
import re
from ldap3 import Server, Connection
class LDAPError(Exception):
pass
def search_ldap(username):
'''Searches the Tilburg University LDAP server for the given username and returns a tuple of first name, last name, full name, ANR, emplId and email address. Permission has been granted by TiU's legal department for retrieving this data. Raises LDAPError on any kind of error.'''
result = ()
baseDN = "o=Universiteit van Tilburg,c=NL"
searchFilter = '(uid={})'.format(username)
attributes = ['givenName', 'sortableSurname', 'cn', 'employeeNumber', 'emplId', 'mail']
try:
server = Server('ldaps.uvt.nl', use_ssl=True)
conn = Connection(server, auto_bind=True)
conn.search(baseDN, searchFilter, attributes=attributes)
response = conn.response[0]['attributes']
for a in attributes:
if a in response and response[a]:
if type(response[a]) is list:
result += (response[a][0],)
else:
result += (response[a],)
else:
result += ('',)
except IndexError:
raise LDAPError('Username not found')
except Exception:
raise LDAPError('Unknown error in LDAP query')
return result
|
32dae9d8af362c5ec00af069d70272a125aa02c5 | firmata/__init__.py | firmata/__init__.py | """Provides an API wrapper around the Firmata wire protocol.
There are two major pieces to the firmata module. When FirmataInit() is called, a thread is spun up to handle serial
port IO. Its sole function is to read bytes into the read queue and write bytes from the write queue. These queues are
then used by the main body of code to respond to API calls made by the host application.
The API presented to the host program is encapsulated in the `Board` class, instances of which are obtained by calling
the previously mentioned `FirmataInit()` function. You can create as many Board classes as you wish, but you will not
go to space today if you create more than on on the same serial port.
"""
from Queue import Queue, Empty
import serial
import threading
from firmata.constants import *
IO_TIMEOUT = 0.2 # Number of seconds to block on IO. Set to None for infinite timeout.
BYTES_IN_CHUNK = 100
class _IOThread(threading.Thread):
def __init__(self, port, baud):
self._port = port
self._baud = baud
self.shutdown = False
self.from_board = Queue()
self.to_board = Queue()
super(_IOThread, self).__init__()
def run(self):
serial_port = serial.Serial(port=self._port, baudrate=self._baud, timeout=0.2)
while not self.shutdown:
r = serial_port.read(BYTES_IN_CHUNK)
[self.from_board.put(i) for i in r]
bytes_written = 0
while not self.to_board.empty() and bytes_written < BYTES_IN_CHUNK:
try:
w = self.to_board.get(block=False)
serial_port.write(w)
self.to_board.task_done()
bytes_written += 1
except Empty:
break
class Board(object):
def __init__(self, port, baud, start_serial=False):
"""Board object constructor. Should not be called directly.
Args:
port: The serial port to use. Expressed as either a string or an integer (see pyserial docs for more info.)
baud: A number representing the baud rate to use for serial communication.
start_serial: If True, starts the serial IO thread right away. Default: False.
"""
self._io_thread = _IOThread(port, baud)
if start_serial:
self._io_thread.start()
self._out = self._io_thread.to_board
self._in = self._io_thread.from_board
def FirmataInit(port, baud=57600):
"""Instantiate a `Board` object for a given serial port.
Args:
port: The serial port to use. Expressed as either a string or an integer (see pyserial docs for more info.)
baud: A number representing the baud rate to use for serial communication.
Returns:
A Board object which implements the firmata protocol over the specified serial port.
"""
return Board(port, baud, start_serial=True)
| Add basic skeleton of the library, including the implementation of the IO thread system. | Add basic skeleton of the library, including the implementation of the IO thread system.
| Python | apache-2.0 | google/firmata.py | Add basic skeleton of the library, including the implementation of the IO thread system. | """Provides an API wrapper around the Firmata wire protocol.
There are two major pieces to the firmata module. When FirmataInit() is called, a thread is spun up to handle serial
port IO. Its sole function is to read bytes into the read queue and write bytes from the write queue. These queues are
then used by the main body of code to respond to API calls made by the host application.
The API presented to the host program is encapsulated in the `Board` class, instances of which are obtained by calling
the previously mentioned `FirmataInit()` function. You can create as many Board classes as you wish, but you will not
go to space today if you create more than on on the same serial port.
"""
from Queue import Queue, Empty
import serial
import threading
from firmata.constants import *
IO_TIMEOUT = 0.2 # Number of seconds to block on IO. Set to None for infinite timeout.
BYTES_IN_CHUNK = 100
class _IOThread(threading.Thread):
def __init__(self, port, baud):
self._port = port
self._baud = baud
self.shutdown = False
self.from_board = Queue()
self.to_board = Queue()
super(_IOThread, self).__init__()
def run(self):
serial_port = serial.Serial(port=self._port, baudrate=self._baud, timeout=0.2)
while not self.shutdown:
r = serial_port.read(BYTES_IN_CHUNK)
[self.from_board.put(i) for i in r]
bytes_written = 0
while not self.to_board.empty() and bytes_written < BYTES_IN_CHUNK:
try:
w = self.to_board.get(block=False)
serial_port.write(w)
self.to_board.task_done()
bytes_written += 1
except Empty:
break
class Board(object):
def __init__(self, port, baud, start_serial=False):
"""Board object constructor. Should not be called directly.
Args:
port: The serial port to use. Expressed as either a string or an integer (see pyserial docs for more info.)
baud: A number representing the baud rate to use for serial communication.
start_serial: If True, starts the serial IO thread right away. Default: False.
"""
self._io_thread = _IOThread(port, baud)
if start_serial:
self._io_thread.start()
self._out = self._io_thread.to_board
self._in = self._io_thread.from_board
def FirmataInit(port, baud=57600):
"""Instantiate a `Board` object for a given serial port.
Args:
port: The serial port to use. Expressed as either a string or an integer (see pyserial docs for more info.)
baud: A number representing the baud rate to use for serial communication.
Returns:
A Board object which implements the firmata protocol over the specified serial port.
"""
return Board(port, baud, start_serial=True)
| <commit_before><commit_msg>Add basic skeleton of the library, including the implementation of the IO thread system.<commit_after> | """Provides an API wrapper around the Firmata wire protocol.
There are two major pieces to the firmata module. When FirmataInit() is called, a thread is spun up to handle serial
port IO. Its sole function is to read bytes into the read queue and write bytes from the write queue. These queues are
then used by the main body of code to respond to API calls made by the host application.
The API presented to the host program is encapsulated in the `Board` class, instances of which are obtained by calling
the previously mentioned `FirmataInit()` function. You can create as many Board classes as you wish, but you will not
go to space today if you create more than on on the same serial port.
"""
from Queue import Queue, Empty
import serial
import threading
from firmata.constants import *
IO_TIMEOUT = 0.2 # Number of seconds to block on IO. Set to None for infinite timeout.
BYTES_IN_CHUNK = 100
class _IOThread(threading.Thread):
def __init__(self, port, baud):
self._port = port
self._baud = baud
self.shutdown = False
self.from_board = Queue()
self.to_board = Queue()
super(_IOThread, self).__init__()
def run(self):
serial_port = serial.Serial(port=self._port, baudrate=self._baud, timeout=0.2)
while not self.shutdown:
r = serial_port.read(BYTES_IN_CHUNK)
[self.from_board.put(i) for i in r]
bytes_written = 0
while not self.to_board.empty() and bytes_written < BYTES_IN_CHUNK:
try:
w = self.to_board.get(block=False)
serial_port.write(w)
self.to_board.task_done()
bytes_written += 1
except Empty:
break
class Board(object):
def __init__(self, port, baud, start_serial=False):
"""Board object constructor. Should not be called directly.
Args:
port: The serial port to use. Expressed as either a string or an integer (see pyserial docs for more info.)
baud: A number representing the baud rate to use for serial communication.
start_serial: If True, starts the serial IO thread right away. Default: False.
"""
self._io_thread = _IOThread(port, baud)
if start_serial:
self._io_thread.start()
self._out = self._io_thread.to_board
self._in = self._io_thread.from_board
def FirmataInit(port, baud=57600):
"""Instantiate a `Board` object for a given serial port.
Args:
port: The serial port to use. Expressed as either a string or an integer (see pyserial docs for more info.)
baud: A number representing the baud rate to use for serial communication.
Returns:
A Board object which implements the firmata protocol over the specified serial port.
"""
return Board(port, baud, start_serial=True)
| Add basic skeleton of the library, including the implementation of the IO thread system."""Provides an API wrapper around the Firmata wire protocol.
There are two major pieces to the firmata module. When FirmataInit() is called, a thread is spun up to handle serial
port IO. Its sole function is to read bytes into the read queue and write bytes from the write queue. These queues are
then used by the main body of code to respond to API calls made by the host application.
The API presented to the host program is encapsulated in the `Board` class, instances of which are obtained by calling
the previously mentioned `FirmataInit()` function. You can create as many Board classes as you wish, but you will not
go to space today if you create more than on on the same serial port.
"""
from Queue import Queue, Empty
import serial
import threading
from firmata.constants import *
IO_TIMEOUT = 0.2 # Number of seconds to block on IO. Set to None for infinite timeout.
BYTES_IN_CHUNK = 100
class _IOThread(threading.Thread):
def __init__(self, port, baud):
self._port = port
self._baud = baud
self.shutdown = False
self.from_board = Queue()
self.to_board = Queue()
super(_IOThread, self).__init__()
def run(self):
serial_port = serial.Serial(port=self._port, baudrate=self._baud, timeout=0.2)
while not self.shutdown:
r = serial_port.read(BYTES_IN_CHUNK)
[self.from_board.put(i) for i in r]
bytes_written = 0
while not self.to_board.empty() and bytes_written < BYTES_IN_CHUNK:
try:
w = self.to_board.get(block=False)
serial_port.write(w)
self.to_board.task_done()
bytes_written += 1
except Empty:
break
class Board(object):
def __init__(self, port, baud, start_serial=False):
"""Board object constructor. Should not be called directly.
Args:
port: The serial port to use. Expressed as either a string or an integer (see pyserial docs for more info.)
baud: A number representing the baud rate to use for serial communication.
start_serial: If True, starts the serial IO thread right away. Default: False.
"""
self._io_thread = _IOThread(port, baud)
if start_serial:
self._io_thread.start()
self._out = self._io_thread.to_board
self._in = self._io_thread.from_board
def FirmataInit(port, baud=57600):
"""Instantiate a `Board` object for a given serial port.
Args:
port: The serial port to use. Expressed as either a string or an integer (see pyserial docs for more info.)
baud: A number representing the baud rate to use for serial communication.
Returns:
A Board object which implements the firmata protocol over the specified serial port.
"""
return Board(port, baud, start_serial=True)
| <commit_before><commit_msg>Add basic skeleton of the library, including the implementation of the IO thread system.<commit_after>"""Provides an API wrapper around the Firmata wire protocol.
There are two major pieces to the firmata module. When FirmataInit() is called, a thread is spun up to handle serial
port IO. Its sole function is to read bytes into the read queue and write bytes from the write queue. These queues are
then used by the main body of code to respond to API calls made by the host application.
The API presented to the host program is encapsulated in the `Board` class, instances of which are obtained by calling
the previously mentioned `FirmataInit()` function. You can create as many Board classes as you wish, but you will not
go to space today if you create more than on on the same serial port.
"""
from Queue import Queue, Empty
import serial
import threading
from firmata.constants import *
IO_TIMEOUT = 0.2 # Number of seconds to block on IO. Set to None for infinite timeout.
BYTES_IN_CHUNK = 100
class _IOThread(threading.Thread):
def __init__(self, port, baud):
self._port = port
self._baud = baud
self.shutdown = False
self.from_board = Queue()
self.to_board = Queue()
super(_IOThread, self).__init__()
def run(self):
serial_port = serial.Serial(port=self._port, baudrate=self._baud, timeout=0.2)
while not self.shutdown:
r = serial_port.read(BYTES_IN_CHUNK)
[self.from_board.put(i) for i in r]
bytes_written = 0
while not self.to_board.empty() and bytes_written < BYTES_IN_CHUNK:
try:
w = self.to_board.get(block=False)
serial_port.write(w)
self.to_board.task_done()
bytes_written += 1
except Empty:
break
class Board(object):
def __init__(self, port, baud, start_serial=False):
"""Board object constructor. Should not be called directly.
Args:
port: The serial port to use. Expressed as either a string or an integer (see pyserial docs for more info.)
baud: A number representing the baud rate to use for serial communication.
start_serial: If True, starts the serial IO thread right away. Default: False.
"""
self._io_thread = _IOThread(port, baud)
if start_serial:
self._io_thread.start()
self._out = self._io_thread.to_board
self._in = self._io_thread.from_board
def FirmataInit(port, baud=57600):
"""Instantiate a `Board` object for a given serial port.
Args:
port: The serial port to use. Expressed as either a string or an integer (see pyserial docs for more info.)
baud: A number representing the baud rate to use for serial communication.
Returns:
A Board object which implements the firmata protocol over the specified serial port.
"""
return Board(port, baud, start_serial=True)
| |
a7622fc3d996407799cec166968c1e56baf07ea9 | wqflask/wqflask/markdown_routes.py | wqflask/wqflask/markdown_routes.py | """Markdown routes
Render pages from github, or if they are unavailable, look for it else where
"""
import requests
import mistune
from flask import Blueprint
from flask import render_template
glossary_blueprint = Blueprint('glossary_blueprint', __name__)
@glossary_blueprint.route('/')
def glossary():
markdown_url = ("https://raw.githubusercontent.com"
"/genenetwork/genenetwork2/"
"wqflask/wqflask/static"
"/glossary.md")
md_content = requests.get(markdown_url)
if md_content.status_code == 200:
return render_template(
"glossary_html",
rendered_markdown=mistune.html(
md_content.content.decode("utf-8"))), 200
return render_template(
"glossary.html",
rendered_markdown=mistune.html("# Github Down!")), 200
| """Markdown routes
Render pages from github, or if they are unavailable, look for it else where
"""
import os
import requests
import mistune
from flask import Blueprint
from flask import render_template
glossary_blueprint = Blueprint('glossary_blueprint', __name__)
def render_markdown(file_name):
"""Try to fetch the file name from Github and if that fails, try to
look for it inside the file system
"""
markdown_url = (f"https://raw.githubusercontent.com"
f"/genenetwork/genenetwork2/"
f"wqflask/wqflask/static/"
f"{file_name}")
md_content = requests.get(markdown_url)
if md_content.status_code == 200:
return mistune.html(md_content.content.decode("utf-8"))
with open(os.path.join(os.path.abspath(os.path.dirname(__file__)),
f"static/markdown/{file_name}")) as md_file:
markdown = md_file.read()
return mistune.html(markdown)
@glossary_blueprint.route('/')
def glossary():
return render_template(
"glossary.html",
rendered_markdown=render_markdown("glossary.md")), 200
| Move logic for fetching md files to it's own function | Move logic for fetching md files to it's own function
* wqflask/wqflask/markdown_routes.py
(render_markdown): New function.
(glossary): use render_markdown function.
| Python | agpl-3.0 | genenetwork/genenetwork2,zsloan/genenetwork2,zsloan/genenetwork2,genenetwork/genenetwork2,genenetwork/genenetwork2,genenetwork/genenetwork2,pjotrp/genenetwork2,pjotrp/genenetwork2,zsloan/genenetwork2,pjotrp/genenetwork2,pjotrp/genenetwork2,pjotrp/genenetwork2,zsloan/genenetwork2 | """Markdown routes
Render pages from github, or if they are unavailable, look for it else where
"""
import requests
import mistune
from flask import Blueprint
from flask import render_template
glossary_blueprint = Blueprint('glossary_blueprint', __name__)
@glossary_blueprint.route('/')
def glossary():
markdown_url = ("https://raw.githubusercontent.com"
"/genenetwork/genenetwork2/"
"wqflask/wqflask/static"
"/glossary.md")
md_content = requests.get(markdown_url)
if md_content.status_code == 200:
return render_template(
"glossary_html",
rendered_markdown=mistune.html(
md_content.content.decode("utf-8"))), 200
return render_template(
"glossary.html",
rendered_markdown=mistune.html("# Github Down!")), 200
Move logic for fetching md files to it's own function
* wqflask/wqflask/markdown_routes.py
(render_markdown): New function.
(glossary): use render_markdown function. | """Markdown routes
Render pages from github, or if they are unavailable, look for it else where
"""
import os
import requests
import mistune
from flask import Blueprint
from flask import render_template
glossary_blueprint = Blueprint('glossary_blueprint', __name__)
def render_markdown(file_name):
"""Try to fetch the file name from Github and if that fails, try to
look for it inside the file system
"""
markdown_url = (f"https://raw.githubusercontent.com"
f"/genenetwork/genenetwork2/"
f"wqflask/wqflask/static/"
f"{file_name}")
md_content = requests.get(markdown_url)
if md_content.status_code == 200:
return mistune.html(md_content.content.decode("utf-8"))
with open(os.path.join(os.path.abspath(os.path.dirname(__file__)),
f"static/markdown/{file_name}")) as md_file:
markdown = md_file.read()
return mistune.html(markdown)
@glossary_blueprint.route('/')
def glossary():
return render_template(
"glossary.html",
rendered_markdown=render_markdown("glossary.md")), 200
| <commit_before>"""Markdown routes
Render pages from github, or if they are unavailable, look for it else where
"""
import requests
import mistune
from flask import Blueprint
from flask import render_template
glossary_blueprint = Blueprint('glossary_blueprint', __name__)
@glossary_blueprint.route('/')
def glossary():
markdown_url = ("https://raw.githubusercontent.com"
"/genenetwork/genenetwork2/"
"wqflask/wqflask/static"
"/glossary.md")
md_content = requests.get(markdown_url)
if md_content.status_code == 200:
return render_template(
"glossary_html",
rendered_markdown=mistune.html(
md_content.content.decode("utf-8"))), 200
return render_template(
"glossary.html",
rendered_markdown=mistune.html("# Github Down!")), 200
<commit_msg>Move logic for fetching md files to it's own function
* wqflask/wqflask/markdown_routes.py
(render_markdown): New function.
(glossary): use render_markdown function.<commit_after> | """Markdown routes
Render pages from github, or if they are unavailable, look for it else where
"""
import os
import requests
import mistune
from flask import Blueprint
from flask import render_template
glossary_blueprint = Blueprint('glossary_blueprint', __name__)
def render_markdown(file_name):
"""Try to fetch the file name from Github and if that fails, try to
look for it inside the file system
"""
markdown_url = (f"https://raw.githubusercontent.com"
f"/genenetwork/genenetwork2/"
f"wqflask/wqflask/static/"
f"{file_name}")
md_content = requests.get(markdown_url)
if md_content.status_code == 200:
return mistune.html(md_content.content.decode("utf-8"))
with open(os.path.join(os.path.abspath(os.path.dirname(__file__)),
f"static/markdown/{file_name}")) as md_file:
markdown = md_file.read()
return mistune.html(markdown)
@glossary_blueprint.route('/')
def glossary():
return render_template(
"glossary.html",
rendered_markdown=render_markdown("glossary.md")), 200
| """Markdown routes
Render pages from github, or if they are unavailable, look for it else where
"""
import requests
import mistune
from flask import Blueprint
from flask import render_template
glossary_blueprint = Blueprint('glossary_blueprint', __name__)
@glossary_blueprint.route('/')
def glossary():
markdown_url = ("https://raw.githubusercontent.com"
"/genenetwork/genenetwork2/"
"wqflask/wqflask/static"
"/glossary.md")
md_content = requests.get(markdown_url)
if md_content.status_code == 200:
return render_template(
"glossary_html",
rendered_markdown=mistune.html(
md_content.content.decode("utf-8"))), 200
return render_template(
"glossary.html",
rendered_markdown=mistune.html("# Github Down!")), 200
Move logic for fetching md files to it's own function
* wqflask/wqflask/markdown_routes.py
(render_markdown): New function.
(glossary): use render_markdown function."""Markdown routes
Render pages from github, or if they are unavailable, look for it else where
"""
import os
import requests
import mistune
from flask import Blueprint
from flask import render_template
glossary_blueprint = Blueprint('glossary_blueprint', __name__)
def render_markdown(file_name):
"""Try to fetch the file name from Github and if that fails, try to
look for it inside the file system
"""
markdown_url = (f"https://raw.githubusercontent.com"
f"/genenetwork/genenetwork2/"
f"wqflask/wqflask/static/"
f"{file_name}")
md_content = requests.get(markdown_url)
if md_content.status_code == 200:
return mistune.html(md_content.content.decode("utf-8"))
with open(os.path.join(os.path.abspath(os.path.dirname(__file__)),
f"static/markdown/{file_name}")) as md_file:
markdown = md_file.read()
return mistune.html(markdown)
@glossary_blueprint.route('/')
def glossary():
return render_template(
"glossary.html",
rendered_markdown=render_markdown("glossary.md")), 200
| <commit_before>"""Markdown routes
Render pages from github, or if they are unavailable, look for it else where
"""
import requests
import mistune
from flask import Blueprint
from flask import render_template
glossary_blueprint = Blueprint('glossary_blueprint', __name__)
@glossary_blueprint.route('/')
def glossary():
markdown_url = ("https://raw.githubusercontent.com"
"/genenetwork/genenetwork2/"
"wqflask/wqflask/static"
"/glossary.md")
md_content = requests.get(markdown_url)
if md_content.status_code == 200:
return render_template(
"glossary_html",
rendered_markdown=mistune.html(
md_content.content.decode("utf-8"))), 200
return render_template(
"glossary.html",
rendered_markdown=mistune.html("# Github Down!")), 200
<commit_msg>Move logic for fetching md files to it's own function
* wqflask/wqflask/markdown_routes.py
(render_markdown): New function.
(glossary): use render_markdown function.<commit_after>"""Markdown routes
Render pages from github, or if they are unavailable, look for it else where
"""
import os
import requests
import mistune
from flask import Blueprint
from flask import render_template
glossary_blueprint = Blueprint('glossary_blueprint', __name__)
def render_markdown(file_name):
"""Try to fetch the file name from Github and if that fails, try to
look for it inside the file system
"""
markdown_url = (f"https://raw.githubusercontent.com"
f"/genenetwork/genenetwork2/"
f"wqflask/wqflask/static/"
f"{file_name}")
md_content = requests.get(markdown_url)
if md_content.status_code == 200:
return mistune.html(md_content.content.decode("utf-8"))
with open(os.path.join(os.path.abspath(os.path.dirname(__file__)),
f"static/markdown/{file_name}")) as md_file:
markdown = md_file.read()
return mistune.html(markdown)
@glossary_blueprint.route('/')
def glossary():
return render_template(
"glossary.html",
rendered_markdown=render_markdown("glossary.md")), 200
|
1abfdea38e868d68c532961459d2b4cbef5a9b71 | src/zeit/website/section.py | src/zeit/website/section.py | import zeit.website.interfaces
from zeit.cms.section.interfaces import ISectionMarker
import grokcore.component as grok
import zeit.cms.checkout.interfaces
import zeit.cms.content.interfaces
import zope.interface
@grok.subscribe(
zeit.cms.content.interfaces.ICommonMetadata,
zeit.cms.checkout.interfaces.IBeforeCheckinEvent)
def provide_website_content(content, event):
content = zope.security.proxy.getObject(content)
if not content.rebrush_website_content:
return
for iface in zope.interface.providedBy(content):
if issubclass(iface, ISectionMarker):
zope.interface.noLongerProvides(content, iface)
zope.interface.alsoProvides(content, zeit.website.interfaces.IWebsiteSection)
| import zeit.website.interfaces
from zeit.cms.section.interfaces import ISectionMarker
import grokcore.component as grok
import zeit.cms.checkout.interfaces
import zeit.cms.content.interfaces
import zope.interface
@grok.subscribe(
zeit.cms.content.interfaces.ICommonMetadata,
zeit.cms.checkout.interfaces.IBeforeCheckinEvent)
def provide_website_content(content, event):
content = zope.security.proxy.getObject(content)
if not content.rebrush_website_content:
zope.interface.noLongerProvides(content,
zeit.website.interfaces.IWebsiteSection)
return
for iface in zope.interface.providedBy(content):
if issubclass(iface, ISectionMarker):
zope.interface.noLongerProvides(content, iface)
zope.interface.alsoProvides(content, zeit.website.interfaces.IWebsiteSection)
| Remove iface, if rebrush_contetn ist not set | Remove iface, if rebrush_contetn ist not set
| Python | bsd-3-clause | ZeitOnline/zeit.website | import zeit.website.interfaces
from zeit.cms.section.interfaces import ISectionMarker
import grokcore.component as grok
import zeit.cms.checkout.interfaces
import zeit.cms.content.interfaces
import zope.interface
@grok.subscribe(
zeit.cms.content.interfaces.ICommonMetadata,
zeit.cms.checkout.interfaces.IBeforeCheckinEvent)
def provide_website_content(content, event):
content = zope.security.proxy.getObject(content)
if not content.rebrush_website_content:
return
for iface in zope.interface.providedBy(content):
if issubclass(iface, ISectionMarker):
zope.interface.noLongerProvides(content, iface)
zope.interface.alsoProvides(content, zeit.website.interfaces.IWebsiteSection)
Remove iface, if rebrush_contetn ist not set | import zeit.website.interfaces
from zeit.cms.section.interfaces import ISectionMarker
import grokcore.component as grok
import zeit.cms.checkout.interfaces
import zeit.cms.content.interfaces
import zope.interface
@grok.subscribe(
zeit.cms.content.interfaces.ICommonMetadata,
zeit.cms.checkout.interfaces.IBeforeCheckinEvent)
def provide_website_content(content, event):
content = zope.security.proxy.getObject(content)
if not content.rebrush_website_content:
zope.interface.noLongerProvides(content,
zeit.website.interfaces.IWebsiteSection)
return
for iface in zope.interface.providedBy(content):
if issubclass(iface, ISectionMarker):
zope.interface.noLongerProvides(content, iface)
zope.interface.alsoProvides(content, zeit.website.interfaces.IWebsiteSection)
| <commit_before>import zeit.website.interfaces
from zeit.cms.section.interfaces import ISectionMarker
import grokcore.component as grok
import zeit.cms.checkout.interfaces
import zeit.cms.content.interfaces
import zope.interface
@grok.subscribe(
zeit.cms.content.interfaces.ICommonMetadata,
zeit.cms.checkout.interfaces.IBeforeCheckinEvent)
def provide_website_content(content, event):
content = zope.security.proxy.getObject(content)
if not content.rebrush_website_content:
return
for iface in zope.interface.providedBy(content):
if issubclass(iface, ISectionMarker):
zope.interface.noLongerProvides(content, iface)
zope.interface.alsoProvides(content, zeit.website.interfaces.IWebsiteSection)
<commit_msg>Remove iface, if rebrush_contetn ist not set<commit_after> | import zeit.website.interfaces
from zeit.cms.section.interfaces import ISectionMarker
import grokcore.component as grok
import zeit.cms.checkout.interfaces
import zeit.cms.content.interfaces
import zope.interface
@grok.subscribe(
zeit.cms.content.interfaces.ICommonMetadata,
zeit.cms.checkout.interfaces.IBeforeCheckinEvent)
def provide_website_content(content, event):
content = zope.security.proxy.getObject(content)
if not content.rebrush_website_content:
zope.interface.noLongerProvides(content,
zeit.website.interfaces.IWebsiteSection)
return
for iface in zope.interface.providedBy(content):
if issubclass(iface, ISectionMarker):
zope.interface.noLongerProvides(content, iface)
zope.interface.alsoProvides(content, zeit.website.interfaces.IWebsiteSection)
| import zeit.website.interfaces
from zeit.cms.section.interfaces import ISectionMarker
import grokcore.component as grok
import zeit.cms.checkout.interfaces
import zeit.cms.content.interfaces
import zope.interface
@grok.subscribe(
zeit.cms.content.interfaces.ICommonMetadata,
zeit.cms.checkout.interfaces.IBeforeCheckinEvent)
def provide_website_content(content, event):
content = zope.security.proxy.getObject(content)
if not content.rebrush_website_content:
return
for iface in zope.interface.providedBy(content):
if issubclass(iface, ISectionMarker):
zope.interface.noLongerProvides(content, iface)
zope.interface.alsoProvides(content, zeit.website.interfaces.IWebsiteSection)
Remove iface, if rebrush_contetn ist not setimport zeit.website.interfaces
from zeit.cms.section.interfaces import ISectionMarker
import grokcore.component as grok
import zeit.cms.checkout.interfaces
import zeit.cms.content.interfaces
import zope.interface
@grok.subscribe(
zeit.cms.content.interfaces.ICommonMetadata,
zeit.cms.checkout.interfaces.IBeforeCheckinEvent)
def provide_website_content(content, event):
content = zope.security.proxy.getObject(content)
if not content.rebrush_website_content:
zope.interface.noLongerProvides(content,
zeit.website.interfaces.IWebsiteSection)
return
for iface in zope.interface.providedBy(content):
if issubclass(iface, ISectionMarker):
zope.interface.noLongerProvides(content, iface)
zope.interface.alsoProvides(content, zeit.website.interfaces.IWebsiteSection)
| <commit_before>import zeit.website.interfaces
from zeit.cms.section.interfaces import ISectionMarker
import grokcore.component as grok
import zeit.cms.checkout.interfaces
import zeit.cms.content.interfaces
import zope.interface
@grok.subscribe(
zeit.cms.content.interfaces.ICommonMetadata,
zeit.cms.checkout.interfaces.IBeforeCheckinEvent)
def provide_website_content(content, event):
content = zope.security.proxy.getObject(content)
if not content.rebrush_website_content:
return
for iface in zope.interface.providedBy(content):
if issubclass(iface, ISectionMarker):
zope.interface.noLongerProvides(content, iface)
zope.interface.alsoProvides(content, zeit.website.interfaces.IWebsiteSection)
<commit_msg>Remove iface, if rebrush_contetn ist not set<commit_after>import zeit.website.interfaces
from zeit.cms.section.interfaces import ISectionMarker
import grokcore.component as grok
import zeit.cms.checkout.interfaces
import zeit.cms.content.interfaces
import zope.interface
@grok.subscribe(
zeit.cms.content.interfaces.ICommonMetadata,
zeit.cms.checkout.interfaces.IBeforeCheckinEvent)
def provide_website_content(content, event):
content = zope.security.proxy.getObject(content)
if not content.rebrush_website_content:
zope.interface.noLongerProvides(content,
zeit.website.interfaces.IWebsiteSection)
return
for iface in zope.interface.providedBy(content):
if issubclass(iface, ISectionMarker):
zope.interface.noLongerProvides(content, iface)
zope.interface.alsoProvides(content, zeit.website.interfaces.IWebsiteSection)
|
358f31fcc8155e15f28f77aee6b434fad2a54935 | iris_sdk/models/import_tn_checker.py | iris_sdk/models/import_tn_checker.py | #!/usr/bin/env python
from __future__ import division, absolute_import, print_function
from future.builtins import super
from iris_sdk.models.base_resource import BaseResource
from iris_sdk.models.data.import_tn_checker import ImportTnCheckerData
from iris_sdk.models.import_tn_checker_response import ImportTnCheckerResponse
XML_NAME_IMPORTTN_CHECKER = "ImportTnCheckerPayload"
XPATH_IMPORTTN_CHECKER = "/importTnChecker"
class ImportTnChecker(BaseResource, ImportTnCheckerData):
"""Request portability information for hosted messaging on a set of TNs"""
_save_post = True
_node_name = XML_NAME_IMPORTTN_CHECKER
_xpath = XPATH_IMPORTTN_CHECKER
def __call__(self, numbers):
# self.clear()
self.telephone_numbers.items.extend(numbers)
return self._post_data(ImportTnCheckerResponse())
def __init__(self, parent=None, client=None):
super().__init__(parent, client)
ImportTnCheckerData.__init__(self)
| #!/usr/bin/env python
from __future__ import division, absolute_import, print_function
from future.builtins import super
from iris_sdk.models.base_resource import BaseResource
from iris_sdk.models.data.import_tn_checker import ImportTnCheckerData
from iris_sdk.models.import_tn_checker_response import ImportTnCheckerResponse
XML_NAME_IMPORTTN_CHECKER = "ImportTnCheckerPayload"
XPATH_IMPORTTN_CHECKER = "/importTnChecker"
class ImportTnChecker(BaseResource, ImportTnCheckerData):
"""Request portability information for hosted messaging on a set of TNs"""
_save_post = True
_node_name = XML_NAME_IMPORTTN_CHECKER
_xpath = XPATH_IMPORTTN_CHECKER
def __call__(self, numbers):
self.clear()
self.telephone_numbers.items.extend(numbers)
return self._post_data(ImportTnCheckerResponse())
def __init__(self, parent=None, client=None):
super().__init__(parent, client)
ImportTnCheckerData.__init__(self)
| Reset the ImportTnChecker on every call | Reset the ImportTnChecker on every call
The list of phone numbers was being retained, potentially resulting in
duplicates with every call.
| Python | mit | bandwidthcom/python-bandwidth-iris | #!/usr/bin/env python
from __future__ import division, absolute_import, print_function
from future.builtins import super
from iris_sdk.models.base_resource import BaseResource
from iris_sdk.models.data.import_tn_checker import ImportTnCheckerData
from iris_sdk.models.import_tn_checker_response import ImportTnCheckerResponse
XML_NAME_IMPORTTN_CHECKER = "ImportTnCheckerPayload"
XPATH_IMPORTTN_CHECKER = "/importTnChecker"
class ImportTnChecker(BaseResource, ImportTnCheckerData):
"""Request portability information for hosted messaging on a set of TNs"""
_save_post = True
_node_name = XML_NAME_IMPORTTN_CHECKER
_xpath = XPATH_IMPORTTN_CHECKER
def __call__(self, numbers):
# self.clear()
self.telephone_numbers.items.extend(numbers)
return self._post_data(ImportTnCheckerResponse())
def __init__(self, parent=None, client=None):
super().__init__(parent, client)
ImportTnCheckerData.__init__(self)
Reset the ImportTnChecker on every call
The list of phone numbers was being retained, potentially resulting in
duplicates with every call. | #!/usr/bin/env python
from __future__ import division, absolute_import, print_function
from future.builtins import super
from iris_sdk.models.base_resource import BaseResource
from iris_sdk.models.data.import_tn_checker import ImportTnCheckerData
from iris_sdk.models.import_tn_checker_response import ImportTnCheckerResponse
XML_NAME_IMPORTTN_CHECKER = "ImportTnCheckerPayload"
XPATH_IMPORTTN_CHECKER = "/importTnChecker"
class ImportTnChecker(BaseResource, ImportTnCheckerData):
"""Request portability information for hosted messaging on a set of TNs"""
_save_post = True
_node_name = XML_NAME_IMPORTTN_CHECKER
_xpath = XPATH_IMPORTTN_CHECKER
def __call__(self, numbers):
self.clear()
self.telephone_numbers.items.extend(numbers)
return self._post_data(ImportTnCheckerResponse())
def __init__(self, parent=None, client=None):
super().__init__(parent, client)
ImportTnCheckerData.__init__(self)
| <commit_before>#!/usr/bin/env python
from __future__ import division, absolute_import, print_function
from future.builtins import super
from iris_sdk.models.base_resource import BaseResource
from iris_sdk.models.data.import_tn_checker import ImportTnCheckerData
from iris_sdk.models.import_tn_checker_response import ImportTnCheckerResponse
XML_NAME_IMPORTTN_CHECKER = "ImportTnCheckerPayload"
XPATH_IMPORTTN_CHECKER = "/importTnChecker"
class ImportTnChecker(BaseResource, ImportTnCheckerData):
"""Request portability information for hosted messaging on a set of TNs"""
_save_post = True
_node_name = XML_NAME_IMPORTTN_CHECKER
_xpath = XPATH_IMPORTTN_CHECKER
def __call__(self, numbers):
# self.clear()
self.telephone_numbers.items.extend(numbers)
return self._post_data(ImportTnCheckerResponse())
def __init__(self, parent=None, client=None):
super().__init__(parent, client)
ImportTnCheckerData.__init__(self)
<commit_msg>Reset the ImportTnChecker on every call
The list of phone numbers was being retained, potentially resulting in
duplicates with every call.<commit_after> | #!/usr/bin/env python
from __future__ import division, absolute_import, print_function
from future.builtins import super
from iris_sdk.models.base_resource import BaseResource
from iris_sdk.models.data.import_tn_checker import ImportTnCheckerData
from iris_sdk.models.import_tn_checker_response import ImportTnCheckerResponse
XML_NAME_IMPORTTN_CHECKER = "ImportTnCheckerPayload"
XPATH_IMPORTTN_CHECKER = "/importTnChecker"
class ImportTnChecker(BaseResource, ImportTnCheckerData):
"""Request portability information for hosted messaging on a set of TNs"""
_save_post = True
_node_name = XML_NAME_IMPORTTN_CHECKER
_xpath = XPATH_IMPORTTN_CHECKER
def __call__(self, numbers):
self.clear()
self.telephone_numbers.items.extend(numbers)
return self._post_data(ImportTnCheckerResponse())
def __init__(self, parent=None, client=None):
super().__init__(parent, client)
ImportTnCheckerData.__init__(self)
| #!/usr/bin/env python
from __future__ import division, absolute_import, print_function
from future.builtins import super
from iris_sdk.models.base_resource import BaseResource
from iris_sdk.models.data.import_tn_checker import ImportTnCheckerData
from iris_sdk.models.import_tn_checker_response import ImportTnCheckerResponse
XML_NAME_IMPORTTN_CHECKER = "ImportTnCheckerPayload"
XPATH_IMPORTTN_CHECKER = "/importTnChecker"
class ImportTnChecker(BaseResource, ImportTnCheckerData):
"""Request portability information for hosted messaging on a set of TNs"""
_save_post = True
_node_name = XML_NAME_IMPORTTN_CHECKER
_xpath = XPATH_IMPORTTN_CHECKER
def __call__(self, numbers):
# self.clear()
self.telephone_numbers.items.extend(numbers)
return self._post_data(ImportTnCheckerResponse())
def __init__(self, parent=None, client=None):
super().__init__(parent, client)
ImportTnCheckerData.__init__(self)
Reset the ImportTnChecker on every call
The list of phone numbers was being retained, potentially resulting in
duplicates with every call.#!/usr/bin/env python
from __future__ import division, absolute_import, print_function
from future.builtins import super
from iris_sdk.models.base_resource import BaseResource
from iris_sdk.models.data.import_tn_checker import ImportTnCheckerData
from iris_sdk.models.import_tn_checker_response import ImportTnCheckerResponse
XML_NAME_IMPORTTN_CHECKER = "ImportTnCheckerPayload"
XPATH_IMPORTTN_CHECKER = "/importTnChecker"
class ImportTnChecker(BaseResource, ImportTnCheckerData):
"""Request portability information for hosted messaging on a set of TNs"""
_save_post = True
_node_name = XML_NAME_IMPORTTN_CHECKER
_xpath = XPATH_IMPORTTN_CHECKER
def __call__(self, numbers):
self.clear()
self.telephone_numbers.items.extend(numbers)
return self._post_data(ImportTnCheckerResponse())
def __init__(self, parent=None, client=None):
super().__init__(parent, client)
ImportTnCheckerData.__init__(self)
| <commit_before>#!/usr/bin/env python
from __future__ import division, absolute_import, print_function
from future.builtins import super
from iris_sdk.models.base_resource import BaseResource
from iris_sdk.models.data.import_tn_checker import ImportTnCheckerData
from iris_sdk.models.import_tn_checker_response import ImportTnCheckerResponse
XML_NAME_IMPORTTN_CHECKER = "ImportTnCheckerPayload"
XPATH_IMPORTTN_CHECKER = "/importTnChecker"
class ImportTnChecker(BaseResource, ImportTnCheckerData):
"""Request portability information for hosted messaging on a set of TNs"""
_save_post = True
_node_name = XML_NAME_IMPORTTN_CHECKER
_xpath = XPATH_IMPORTTN_CHECKER
def __call__(self, numbers):
# self.clear()
self.telephone_numbers.items.extend(numbers)
return self._post_data(ImportTnCheckerResponse())
def __init__(self, parent=None, client=None):
super().__init__(parent, client)
ImportTnCheckerData.__init__(self)
<commit_msg>Reset the ImportTnChecker on every call
The list of phone numbers was being retained, potentially resulting in
duplicates with every call.<commit_after>#!/usr/bin/env python
from __future__ import division, absolute_import, print_function
from future.builtins import super
from iris_sdk.models.base_resource import BaseResource
from iris_sdk.models.data.import_tn_checker import ImportTnCheckerData
from iris_sdk.models.import_tn_checker_response import ImportTnCheckerResponse
XML_NAME_IMPORTTN_CHECKER = "ImportTnCheckerPayload"
XPATH_IMPORTTN_CHECKER = "/importTnChecker"
class ImportTnChecker(BaseResource, ImportTnCheckerData):
"""Request portability information for hosted messaging on a set of TNs"""
_save_post = True
_node_name = XML_NAME_IMPORTTN_CHECKER
_xpath = XPATH_IMPORTTN_CHECKER
def __call__(self, numbers):
self.clear()
self.telephone_numbers.items.extend(numbers)
return self._post_data(ImportTnCheckerResponse())
def __init__(self, parent=None, client=None):
super().__init__(parent, client)
ImportTnCheckerData.__init__(self)
|
533d6294a47a4a974dfda1743e9fcd6146ede27f | codonpdx/insert.py | codonpdx/insert.py | #!/usr/bin/env python
import json
import sys
from db import dbManager
# insert an organism into a database table
def insert(args):
if args.json:
data = json.loads(args.json)
else:
data = json.load(args.infile)
with dbManager('config/db.cfg') as db:
for org in data:
db.insertOrganism(org, args.dbname, args.job)
return data
| #!/usr/bin/env python
import json
import sys
from db import dbManager
# insert an organism into a database table
def insert(args):
if hasattr(args, 'json'):
data = json.loads(args.json)
else:
data = json.load(args.infile)
with dbManager('config/db.cfg') as db:
for org in data:
db.insertOrganism(org, args.dbname, args.job)
return data
| Fix testing for JSON parameter existence. | Fix testing for JSON parameter existence.
| Python | apache-2.0 | PDX-Flamingo/codonpdx-python,PDX-Flamingo/codonpdx-python | #!/usr/bin/env python
import json
import sys
from db import dbManager
# insert an organism into a database table
def insert(args):
if args.json:
data = json.loads(args.json)
else:
data = json.load(args.infile)
with dbManager('config/db.cfg') as db:
for org in data:
db.insertOrganism(org, args.dbname, args.job)
return data
Fix testing for JSON parameter existence. | #!/usr/bin/env python
import json
import sys
from db import dbManager
# insert an organism into a database table
def insert(args):
if hasattr(args, 'json'):
data = json.loads(args.json)
else:
data = json.load(args.infile)
with dbManager('config/db.cfg') as db:
for org in data:
db.insertOrganism(org, args.dbname, args.job)
return data
| <commit_before>#!/usr/bin/env python
import json
import sys
from db import dbManager
# insert an organism into a database table
def insert(args):
if args.json:
data = json.loads(args.json)
else:
data = json.load(args.infile)
with dbManager('config/db.cfg') as db:
for org in data:
db.insertOrganism(org, args.dbname, args.job)
return data
<commit_msg>Fix testing for JSON parameter existence.<commit_after> | #!/usr/bin/env python
import json
import sys
from db import dbManager
# insert an organism into a database table
def insert(args):
if hasattr(args, 'json'):
data = json.loads(args.json)
else:
data = json.load(args.infile)
with dbManager('config/db.cfg') as db:
for org in data:
db.insertOrganism(org, args.dbname, args.job)
return data
| #!/usr/bin/env python
import json
import sys
from db import dbManager
# insert an organism into a database table
def insert(args):
if args.json:
data = json.loads(args.json)
else:
data = json.load(args.infile)
with dbManager('config/db.cfg') as db:
for org in data:
db.insertOrganism(org, args.dbname, args.job)
return data
Fix testing for JSON parameter existence.#!/usr/bin/env python
import json
import sys
from db import dbManager
# insert an organism into a database table
def insert(args):
if hasattr(args, 'json'):
data = json.loads(args.json)
else:
data = json.load(args.infile)
with dbManager('config/db.cfg') as db:
for org in data:
db.insertOrganism(org, args.dbname, args.job)
return data
| <commit_before>#!/usr/bin/env python
import json
import sys
from db import dbManager
# insert an organism into a database table
def insert(args):
if args.json:
data = json.loads(args.json)
else:
data = json.load(args.infile)
with dbManager('config/db.cfg') as db:
for org in data:
db.insertOrganism(org, args.dbname, args.job)
return data
<commit_msg>Fix testing for JSON parameter existence.<commit_after>#!/usr/bin/env python
import json
import sys
from db import dbManager
# insert an organism into a database table
def insert(args):
if hasattr(args, 'json'):
data = json.loads(args.json)
else:
data = json.load(args.infile)
with dbManager('config/db.cfg') as db:
for org in data:
db.insertOrganism(org, args.dbname, args.job)
return data
|
215c6d714df53f6f52f2bf819f2a01f1c1eab294 | learntris.py | learntris.py | #!/usr/bin/env python
import sys
class Grid(object):
def __init__(self):
self.board = ['. '*10 for row in range(0,22)]
self.score = 0
self.lines_clear = 0
def draw_board(self):
for cell in self.board:
print cell
def given(self):
self.board = []
for row in range(0,22):
self.board.append(raw_input())
def clear(self):
self.board = ['. '*10 for row in range(0,22)]
def show_score(self):
print self.score
def show_clear_lines(self):
print self.lines_clear
def main():
grid = Grid()
commands = {'p': grid.draw_board,
'g': grid.given,
'c': grid.clear,
'?s': grid.show_score,
'?n': grid.show_clear_lines}
while True:
command = raw_input()
if command == 'q':
break
commands[command]()
if __name__ == '__main__':
main() | #!/usr/bin/env python
import sys
class Grid(object):
def __init__(self):
self.board = [[None] * 10 for i in range(22)]
self.score = 0
self.lines_clear = 0
def draw_board(self):
current_board = self.board
for row in current_board:
row = map(lambda cell: '.' if cell == None else y, row)
print ' '.join(row)
def given(self):
self.board = []
for row in range(0,22):
self.board.append(raw_input())
def clear(self):
self.board = ['. '*10 for row in range(0,22)]
def show_score(self):
print self.score
def show_clear_lines(self):
print self.lines_clear
def main():
grid = Grid()
commands = {'p': grid.draw_board,
'g': grid.given,
'c': grid.clear,
'?s': grid.show_score,
'?n': grid.show_clear_lines}
while True:
command = raw_input()
if command == 'q':
break
commands[command]()
if __name__ == '__main__':
main() | Change data structure to multidimensional array Now passing Test 2 | Change data structure to multidimensional array
Now passing Test 2
| Python | mit | mosegontar/learntris | #!/usr/bin/env python
import sys
class Grid(object):
def __init__(self):
self.board = ['. '*10 for row in range(0,22)]
self.score = 0
self.lines_clear = 0
def draw_board(self):
for cell in self.board:
print cell
def given(self):
self.board = []
for row in range(0,22):
self.board.append(raw_input())
def clear(self):
self.board = ['. '*10 for row in range(0,22)]
def show_score(self):
print self.score
def show_clear_lines(self):
print self.lines_clear
def main():
grid = Grid()
commands = {'p': grid.draw_board,
'g': grid.given,
'c': grid.clear,
'?s': grid.show_score,
'?n': grid.show_clear_lines}
while True:
command = raw_input()
if command == 'q':
break
commands[command]()
if __name__ == '__main__':
main()Change data structure to multidimensional array
Now passing Test 2 | #!/usr/bin/env python
import sys
class Grid(object):
def __init__(self):
self.board = [[None] * 10 for i in range(22)]
self.score = 0
self.lines_clear = 0
def draw_board(self):
current_board = self.board
for row in current_board:
row = map(lambda cell: '.' if cell == None else y, row)
print ' '.join(row)
def given(self):
self.board = []
for row in range(0,22):
self.board.append(raw_input())
def clear(self):
self.board = ['. '*10 for row in range(0,22)]
def show_score(self):
print self.score
def show_clear_lines(self):
print self.lines_clear
def main():
grid = Grid()
commands = {'p': grid.draw_board,
'g': grid.given,
'c': grid.clear,
'?s': grid.show_score,
'?n': grid.show_clear_lines}
while True:
command = raw_input()
if command == 'q':
break
commands[command]()
if __name__ == '__main__':
main() | <commit_before>#!/usr/bin/env python
import sys
class Grid(object):
def __init__(self):
self.board = ['. '*10 for row in range(0,22)]
self.score = 0
self.lines_clear = 0
def draw_board(self):
for cell in self.board:
print cell
def given(self):
self.board = []
for row in range(0,22):
self.board.append(raw_input())
def clear(self):
self.board = ['. '*10 for row in range(0,22)]
def show_score(self):
print self.score
def show_clear_lines(self):
print self.lines_clear
def main():
grid = Grid()
commands = {'p': grid.draw_board,
'g': grid.given,
'c': grid.clear,
'?s': grid.show_score,
'?n': grid.show_clear_lines}
while True:
command = raw_input()
if command == 'q':
break
commands[command]()
if __name__ == '__main__':
main()<commit_msg>Change data structure to multidimensional array
Now passing Test 2<commit_after> | #!/usr/bin/env python
import sys
class Grid(object):
def __init__(self):
self.board = [[None] * 10 for i in range(22)]
self.score = 0
self.lines_clear = 0
def draw_board(self):
current_board = self.board
for row in current_board:
row = map(lambda cell: '.' if cell == None else y, row)
print ' '.join(row)
def given(self):
self.board = []
for row in range(0,22):
self.board.append(raw_input())
def clear(self):
self.board = ['. '*10 for row in range(0,22)]
def show_score(self):
print self.score
def show_clear_lines(self):
print self.lines_clear
def main():
grid = Grid()
commands = {'p': grid.draw_board,
'g': grid.given,
'c': grid.clear,
'?s': grid.show_score,
'?n': grid.show_clear_lines}
while True:
command = raw_input()
if command == 'q':
break
commands[command]()
if __name__ == '__main__':
main() | #!/usr/bin/env python
import sys
class Grid(object):
def __init__(self):
self.board = ['. '*10 for row in range(0,22)]
self.score = 0
self.lines_clear = 0
def draw_board(self):
for cell in self.board:
print cell
def given(self):
self.board = []
for row in range(0,22):
self.board.append(raw_input())
def clear(self):
self.board = ['. '*10 for row in range(0,22)]
def show_score(self):
print self.score
def show_clear_lines(self):
print self.lines_clear
def main():
grid = Grid()
commands = {'p': grid.draw_board,
'g': grid.given,
'c': grid.clear,
'?s': grid.show_score,
'?n': grid.show_clear_lines}
while True:
command = raw_input()
if command == 'q':
break
commands[command]()
if __name__ == '__main__':
main()Change data structure to multidimensional array
Now passing Test 2#!/usr/bin/env python
import sys
class Grid(object):
def __init__(self):
self.board = [[None] * 10 for i in range(22)]
self.score = 0
self.lines_clear = 0
def draw_board(self):
current_board = self.board
for row in current_board:
row = map(lambda cell: '.' if cell == None else y, row)
print ' '.join(row)
def given(self):
self.board = []
for row in range(0,22):
self.board.append(raw_input())
def clear(self):
self.board = ['. '*10 for row in range(0,22)]
def show_score(self):
print self.score
def show_clear_lines(self):
print self.lines_clear
def main():
grid = Grid()
commands = {'p': grid.draw_board,
'g': grid.given,
'c': grid.clear,
'?s': grid.show_score,
'?n': grid.show_clear_lines}
while True:
command = raw_input()
if command == 'q':
break
commands[command]()
if __name__ == '__main__':
main() | <commit_before>#!/usr/bin/env python
import sys
class Grid(object):
def __init__(self):
self.board = ['. '*10 for row in range(0,22)]
self.score = 0
self.lines_clear = 0
def draw_board(self):
for cell in self.board:
print cell
def given(self):
self.board = []
for row in range(0,22):
self.board.append(raw_input())
def clear(self):
self.board = ['. '*10 for row in range(0,22)]
def show_score(self):
print self.score
def show_clear_lines(self):
print self.lines_clear
def main():
grid = Grid()
commands = {'p': grid.draw_board,
'g': grid.given,
'c': grid.clear,
'?s': grid.show_score,
'?n': grid.show_clear_lines}
while True:
command = raw_input()
if command == 'q':
break
commands[command]()
if __name__ == '__main__':
main()<commit_msg>Change data structure to multidimensional array
Now passing Test 2<commit_after>#!/usr/bin/env python
import sys
class Grid(object):
def __init__(self):
self.board = [[None] * 10 for i in range(22)]
self.score = 0
self.lines_clear = 0
def draw_board(self):
current_board = self.board
for row in current_board:
row = map(lambda cell: '.' if cell == None else y, row)
print ' '.join(row)
def given(self):
self.board = []
for row in range(0,22):
self.board.append(raw_input())
def clear(self):
self.board = ['. '*10 for row in range(0,22)]
def show_score(self):
print self.score
def show_clear_lines(self):
print self.lines_clear
def main():
grid = Grid()
commands = {'p': grid.draw_board,
'g': grid.given,
'c': grid.clear,
'?s': grid.show_score,
'?n': grid.show_clear_lines}
while True:
command = raw_input()
if command == 'q':
break
commands[command]()
if __name__ == '__main__':
main() |
73e28db67c8e2ea897790844dd3eb65e6c4c5c98 | extensions/rules/coord_two_dim.py | extensions/rules/coord_two_dim.py | # coding: utf-8
#
# Copyright 2014 The Oppia Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, softwar
# distributed under the License is distributed on an "AS-IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Rules for CoordTwoDim objects."""
__author__ = 'Sean Lip'
from extensions.rules import base
class Within(base.CoordTwoDimRule):
description = 'is within {{d|Real}} km of {{p|CoordTwoDim}}'
class NotWithin(base.CoordTwoDimRule):
description = 'is not within {{d|Real}} km of {{p|CoordTwoDim}}'
class FuzzyMatches(base.CoordTwoDimRule):
description = 'is similar to {{training_data|ListOfCoordTwoDim}}'
| # coding: utf-8
#
# Copyright 2014 The Oppia Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, softwar
# distributed under the License is distributed on an "AS-IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Rules for CoordTwoDim objects."""
__author__ = 'Sean Lip'
from extensions.rules import base
class Within(base.CoordTwoDimRule):
description = 'is within {{d|Real}} km of {{p|CoordTwoDim}}'
class NotWithin(base.CoordTwoDimRule):
description = 'is not within {{d|Real}} km of {{p|CoordTwoDim}}'
class FuzzyMatches(base.CoordTwoDimRule):
description = 'is similar to {{training_data|ListOfCoordTwoDim}}'
# TODO(wxy): Create a better classifier for this interaction. Currently,
# the frontend implementation of this rule returns a booleen value,
# checking if the answer is close to any point in the training data.
# If this fails, the answer should then go to a backend classifier that
# picks the answer group with the best matching answer group.
| Add TODO for interactive map classifier | Add TODO for interactive map classifier
| Python | apache-2.0 | zgchizi/oppia-uc,kevinlee12/oppia,michaelWagner/oppia,oppia/oppia,kennho/oppia,jestapinski/oppia,prasanna08/oppia,prasanna08/oppia,amitdeutsch/oppia,DewarM/oppia,kingctan/oppia,rackstar17/oppia,sdulal/oppia,zgchizi/oppia-uc,prasanna08/oppia,kennho/oppia,prasanna08/oppia,AllanYangZhou/oppia,souravbadami/oppia,AllanYangZhou/oppia,sbhowmik89/oppia,oppia/oppia,jestapinski/oppia,terrameijar/oppia,sbhowmik89/oppia,himanshu-dixit/oppia,MAKOSCAFEE/oppia,kennho/oppia,sdulal/oppia,DewarM/oppia,brianrodri/oppia,raju249/oppia,MAKOSCAFEE/oppia,amitdeutsch/oppia,AllanYangZhou/oppia,himanshu-dixit/oppia,bjvoth/oppia,michaelWagner/oppia,amitdeutsch/oppia,mit0110/oppia,mit0110/oppia,zgchizi/oppia-uc,anthkris/oppia,sdulal/oppia,michaelWagner/oppia,sbhowmik89/oppia,kingctan/oppia,edallison/oppia,amgowano/oppia,bjvoth/oppia,kevinlee12/oppia,anggorodewanto/oppia,jestapinski/oppia,brianrodri/oppia,souravbadami/oppia,sbhowmik89/oppia,oppia/oppia,MaximLich/oppia,DewarM/oppia,raju249/oppia,shaz13/oppia,bjvoth/oppia,oppia/oppia,rackstar17/oppia,rackstar17/oppia,kingctan/oppia,kevinlee12/oppia,anggorodewanto/oppia,MaximLich/oppia,DewarM/oppia,mit0110/oppia,MaximLich/oppia,zgchizi/oppia-uc,souravbadami/oppia,rackstar17/oppia,prasanna08/oppia,kennho/oppia,amgowano/oppia,kevinlee12/oppia,sbhowmik89/oppia,amgowano/oppia,souravbadami/oppia,sdulal/oppia,amitdeutsch/oppia,bjvoth/oppia,edallison/oppia,anggorodewanto/oppia,edallison/oppia,anthkris/oppia,MAKOSCAFEE/oppia,bjvoth/oppia,brianrodri/oppia,terrameijar/oppia,souravbadami/oppia,anggorodewanto/oppia,kennho/oppia,anthkris/oppia,brianrodri/oppia,MaximLich/oppia,terrameijar/oppia,oppia/oppia,himanshu-dixit/oppia,raju249/oppia,jestapinski/oppia,mit0110/oppia,AllanYangZhou/oppia,terrameijar/oppia,anthkris/oppia,amgowano/oppia,michaelWagner/oppia,brianrodri/oppia,edallison/oppia,michaelWagner/oppia,amitdeutsch/oppia,kevinlee12/oppia,sdulal/oppia,kingctan/oppia,mit0110/oppia,shaz13/oppia,himanshu-dixit/oppia,shaz13/oppia,raju249/oppia,kingctan/oppia,shaz13/oppia,DewarM/oppia,MAKOSCAFEE/oppia | # coding: utf-8
#
# Copyright 2014 The Oppia Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, softwar
# distributed under the License is distributed on an "AS-IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Rules for CoordTwoDim objects."""
__author__ = 'Sean Lip'
from extensions.rules import base
class Within(base.CoordTwoDimRule):
description = 'is within {{d|Real}} km of {{p|CoordTwoDim}}'
class NotWithin(base.CoordTwoDimRule):
description = 'is not within {{d|Real}} km of {{p|CoordTwoDim}}'
class FuzzyMatches(base.CoordTwoDimRule):
description = 'is similar to {{training_data|ListOfCoordTwoDim}}'
Add TODO for interactive map classifier | # coding: utf-8
#
# Copyright 2014 The Oppia Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, softwar
# distributed under the License is distributed on an "AS-IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Rules for CoordTwoDim objects."""
__author__ = 'Sean Lip'
from extensions.rules import base
class Within(base.CoordTwoDimRule):
description = 'is within {{d|Real}} km of {{p|CoordTwoDim}}'
class NotWithin(base.CoordTwoDimRule):
description = 'is not within {{d|Real}} km of {{p|CoordTwoDim}}'
class FuzzyMatches(base.CoordTwoDimRule):
description = 'is similar to {{training_data|ListOfCoordTwoDim}}'
# TODO(wxy): Create a better classifier for this interaction. Currently,
# the frontend implementation of this rule returns a booleen value,
# checking if the answer is close to any point in the training data.
# If this fails, the answer should then go to a backend classifier that
# picks the answer group with the best matching answer group.
| <commit_before># coding: utf-8
#
# Copyright 2014 The Oppia Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, softwar
# distributed under the License is distributed on an "AS-IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Rules for CoordTwoDim objects."""
__author__ = 'Sean Lip'
from extensions.rules import base
class Within(base.CoordTwoDimRule):
description = 'is within {{d|Real}} km of {{p|CoordTwoDim}}'
class NotWithin(base.CoordTwoDimRule):
description = 'is not within {{d|Real}} km of {{p|CoordTwoDim}}'
class FuzzyMatches(base.CoordTwoDimRule):
description = 'is similar to {{training_data|ListOfCoordTwoDim}}'
<commit_msg>Add TODO for interactive map classifier<commit_after> | # coding: utf-8
#
# Copyright 2014 The Oppia Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, softwar
# distributed under the License is distributed on an "AS-IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Rules for CoordTwoDim objects."""
__author__ = 'Sean Lip'
from extensions.rules import base
class Within(base.CoordTwoDimRule):
description = 'is within {{d|Real}} km of {{p|CoordTwoDim}}'
class NotWithin(base.CoordTwoDimRule):
description = 'is not within {{d|Real}} km of {{p|CoordTwoDim}}'
class FuzzyMatches(base.CoordTwoDimRule):
description = 'is similar to {{training_data|ListOfCoordTwoDim}}'
# TODO(wxy): Create a better classifier for this interaction. Currently,
# the frontend implementation of this rule returns a booleen value,
# checking if the answer is close to any point in the training data.
# If this fails, the answer should then go to a backend classifier that
# picks the answer group with the best matching answer group.
| # coding: utf-8
#
# Copyright 2014 The Oppia Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, softwar
# distributed under the License is distributed on an "AS-IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Rules for CoordTwoDim objects."""
__author__ = 'Sean Lip'
from extensions.rules import base
class Within(base.CoordTwoDimRule):
description = 'is within {{d|Real}} km of {{p|CoordTwoDim}}'
class NotWithin(base.CoordTwoDimRule):
description = 'is not within {{d|Real}} km of {{p|CoordTwoDim}}'
class FuzzyMatches(base.CoordTwoDimRule):
description = 'is similar to {{training_data|ListOfCoordTwoDim}}'
Add TODO for interactive map classifier# coding: utf-8
#
# Copyright 2014 The Oppia Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, softwar
# distributed under the License is distributed on an "AS-IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Rules for CoordTwoDim objects."""
__author__ = 'Sean Lip'
from extensions.rules import base
class Within(base.CoordTwoDimRule):
description = 'is within {{d|Real}} km of {{p|CoordTwoDim}}'
class NotWithin(base.CoordTwoDimRule):
description = 'is not within {{d|Real}} km of {{p|CoordTwoDim}}'
class FuzzyMatches(base.CoordTwoDimRule):
description = 'is similar to {{training_data|ListOfCoordTwoDim}}'
# TODO(wxy): Create a better classifier for this interaction. Currently,
# the frontend implementation of this rule returns a booleen value,
# checking if the answer is close to any point in the training data.
# If this fails, the answer should then go to a backend classifier that
# picks the answer group with the best matching answer group.
| <commit_before># coding: utf-8
#
# Copyright 2014 The Oppia Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, softwar
# distributed under the License is distributed on an "AS-IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Rules for CoordTwoDim objects."""
__author__ = 'Sean Lip'
from extensions.rules import base
class Within(base.CoordTwoDimRule):
description = 'is within {{d|Real}} km of {{p|CoordTwoDim}}'
class NotWithin(base.CoordTwoDimRule):
description = 'is not within {{d|Real}} km of {{p|CoordTwoDim}}'
class FuzzyMatches(base.CoordTwoDimRule):
description = 'is similar to {{training_data|ListOfCoordTwoDim}}'
<commit_msg>Add TODO for interactive map classifier<commit_after># coding: utf-8
#
# Copyright 2014 The Oppia Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, softwar
# distributed under the License is distributed on an "AS-IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Rules for CoordTwoDim objects."""
__author__ = 'Sean Lip'
from extensions.rules import base
class Within(base.CoordTwoDimRule):
description = 'is within {{d|Real}} km of {{p|CoordTwoDim}}'
class NotWithin(base.CoordTwoDimRule):
description = 'is not within {{d|Real}} km of {{p|CoordTwoDim}}'
class FuzzyMatches(base.CoordTwoDimRule):
description = 'is similar to {{training_data|ListOfCoordTwoDim}}'
# TODO(wxy): Create a better classifier for this interaction. Currently,
# the frontend implementation of this rule returns a booleen value,
# checking if the answer is close to any point in the training data.
# If this fails, the answer should then go to a backend classifier that
# picks the answer group with the best matching answer group.
|
ba745d03c11d7478c4da5a68246ec9d461077365 | experiments/rnnencdec/__init__.py | experiments/rnnencdec/__init__.py | from encdec import RNNEncoderDecoder
from encdec import get_batch_iterator
from encdec import parse_input
from encdec import create_padded_batch
from state import prototype_state
from state import prototype_search_state
from state import prototype_sentence_state
from state import prototype_autoenc_state
from state import prototype_phrase_en_zn_state
from state import prototype_de2en_state
| from encdec import RNNEncoderDecoder
from encdec import get_batch_iterator
from encdec import parse_input
from encdec import create_padded_batch
from state import prototype_state
from state import prototype_search_state
from state import prototype_sentence_state
| Remove states that do not exist anymore | Remove states that do not exist anymore
| Python | bsd-3-clause | sebastien-j/LV_groundhog,ZenDevelopmentSystems/GroundHog,kyunghyuncho/GroundHog,dmitriy-serdyuk/EncDecASR,OlafLee/LV_groundhog,ZenDevelopmentSystems/GroundHog,sebastien-j/LV_groundhog,lisa-groundhog/GroundHog,vseledkin/LV_groundhog,vseledkin/LV_groundhog,zerkh/GroundHog,hezhenghao/GroundHog,zerkh/GroundHog,OlafLee/LV_groundhog,kyunghyuncho/GroundHog,ZhangAustin/GroundHog,vseledkin/LV_groundhog,ZhangAustin/GroundHog,zerkh/GroundHog,lisa-groundhog/GroundHog,hezhenghao/GroundHog,dmitriy-serdyuk/EncDecASR,OlafLee/LV_groundhog,kyunghyuncho/GroundHog,lisa-groundhog/GroundHog,ronuchit/GroundHog,ZhangAustin/GroundHog,ZenDevelopmentSystems/GroundHog,sebastien-j/LV_groundhog,kyunghyuncho/GroundHog,sebastien-j/LV_groundhog,OlafLee/LV_groundhog,zerkh/GroundHog,vseledkin/LV_groundhog,vseledkin/LV_groundhog,OlafLee/LV_groundhog,ronuchit/GroundHog,hezhenghao/GroundHog,vseledkin/LV_groundhog,dmitriy-serdyuk/EncDecASR,ronuchit/GroundHog,sebastien-j/LV_groundhog,ronuchit/GroundHog,ronuchit/GroundHog,dmitriy-serdyuk/EncDecASR,kyunghyuncho/GroundHog,hezhenghao/GroundHog,lisa-groundhog/GroundHog,OlafLee/LV_groundhog,zerkh/GroundHog,ZhangAustin/GroundHog,sebastien-j/LV_groundhog,hezhenghao/GroundHog,ZenDevelopmentSystems/GroundHog,lisa-groundhog/GroundHog,ZenDevelopmentSystems/GroundHog | from encdec import RNNEncoderDecoder
from encdec import get_batch_iterator
from encdec import parse_input
from encdec import create_padded_batch
from state import prototype_state
from state import prototype_search_state
from state import prototype_sentence_state
from state import prototype_autoenc_state
from state import prototype_phrase_en_zn_state
from state import prototype_de2en_state
Remove states that do not exist anymore | from encdec import RNNEncoderDecoder
from encdec import get_batch_iterator
from encdec import parse_input
from encdec import create_padded_batch
from state import prototype_state
from state import prototype_search_state
from state import prototype_sentence_state
| <commit_before>from encdec import RNNEncoderDecoder
from encdec import get_batch_iterator
from encdec import parse_input
from encdec import create_padded_batch
from state import prototype_state
from state import prototype_search_state
from state import prototype_sentence_state
from state import prototype_autoenc_state
from state import prototype_phrase_en_zn_state
from state import prototype_de2en_state
<commit_msg>Remove states that do not exist anymore<commit_after> | from encdec import RNNEncoderDecoder
from encdec import get_batch_iterator
from encdec import parse_input
from encdec import create_padded_batch
from state import prototype_state
from state import prototype_search_state
from state import prototype_sentence_state
| from encdec import RNNEncoderDecoder
from encdec import get_batch_iterator
from encdec import parse_input
from encdec import create_padded_batch
from state import prototype_state
from state import prototype_search_state
from state import prototype_sentence_state
from state import prototype_autoenc_state
from state import prototype_phrase_en_zn_state
from state import prototype_de2en_state
Remove states that do not exist anymorefrom encdec import RNNEncoderDecoder
from encdec import get_batch_iterator
from encdec import parse_input
from encdec import create_padded_batch
from state import prototype_state
from state import prototype_search_state
from state import prototype_sentence_state
| <commit_before>from encdec import RNNEncoderDecoder
from encdec import get_batch_iterator
from encdec import parse_input
from encdec import create_padded_batch
from state import prototype_state
from state import prototype_search_state
from state import prototype_sentence_state
from state import prototype_autoenc_state
from state import prototype_phrase_en_zn_state
from state import prototype_de2en_state
<commit_msg>Remove states that do not exist anymore<commit_after>from encdec import RNNEncoderDecoder
from encdec import get_batch_iterator
from encdec import parse_input
from encdec import create_padded_batch
from state import prototype_state
from state import prototype_search_state
from state import prototype_sentence_state
|
befe47c35c68e17231e21febbf52041f245b8985 | django_mailer/managers.py | django_mailer/managers.py | from django.db import models
from django_mailer import constants
class QueueManager(models.Manager):
use_for_related_fields = True
def high_priority(self):
"""
Return a QuerySet of high priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_HIGH)
def normal_priority(self):
"""
Return a QuerySet of normal priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_NORMAL)
def low_priority(self):
"""
Return a QuerySet of low priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_LOW)
def non_deferred(self):
"""
Return a QuerySet containing all non-deferred queued messages.
"""
return self.filter(deferred=False)
def deferred(self):
"""
Return a QuerySet of all deferred messages in the queue.
"""
return self.filter(deferred=True)
def retry_deferred(self, new_priority=None):
"""
Reset the deferred flag for all deferred messages so they will be
retried.
"""
count = self.deferred().count()
update_kwargs = dict(deferred=False)
if new_priority is not None:
update_kwargs['priority'] = new_priority
self.deferred().update(**update_kwargs)
return count
| from django.db import models
from django_mailer import constants
class QueueManager(models.Manager):
use_for_related_fields = True
def high_priority(self):
"""
Return a QuerySet of high priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_HIGH)
def normal_priority(self):
"""
Return a QuerySet of normal priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_NORMAL)
def low_priority(self):
"""
Return a QuerySet of low priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_LOW)
def non_deferred(self):
"""
Return a QuerySet containing all non-deferred queued messages.
"""
return self.filter(deferred=False)
def deferred(self):
"""
Return a QuerySet of all deferred messages in the queue.
"""
return self.filter(deferred=True)
def retry_deferred(self, new_priority=None):
"""
Reset the deferred flag for all deferred messages so they will be
retried.
"""
count = self.deferred().count()
update_kwargs = dict(deferred=False, retries=models.F('retries')+1)
if new_priority is not None:
update_kwargs['priority'] = new_priority
self.deferred().update(**update_kwargs)
return count
| Update the retries count of a queued message when it is changed back from deferred | Update the retries count of a queued message when it is changed back from deferred
| Python | mit | APSL/django-mailer-2,Giftovus/django-mailer-2,davidmarble/django-mailer-2,SmileyChris/django-mailer-2,kvh/django-mailer-2,maykinmedia/django-mailer-2,PSyton/django-mailer-2,APSL/django-mailer-2,colinhowe/django-mailer-2,rofrankel/django-mailer-2,maykinmedia/django-mailer-2,APSL/django-mailer-2,GreenLightGo/django-mailer-2,morenopc/django-mailer-2,shn/django-mailer-2,maykinmedia/django-mailer-2,mfwarren/django-mailer-2,mrbox/django-mailer-2,tachang/django-mailer-2,damkop/django-mailer-2,danfairs/django-mailer-2,tclancy/django-mailer-2,tsanders-kalloop/django-mailer-2,fenginx/django-mailer-2,victorfontes/django-mailer-2,k1000/django-mailer-2,pegler/django-mailer-2,torchbox/django-mailer-2 | from django.db import models
from django_mailer import constants
class QueueManager(models.Manager):
use_for_related_fields = True
def high_priority(self):
"""
Return a QuerySet of high priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_HIGH)
def normal_priority(self):
"""
Return a QuerySet of normal priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_NORMAL)
def low_priority(self):
"""
Return a QuerySet of low priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_LOW)
def non_deferred(self):
"""
Return a QuerySet containing all non-deferred queued messages.
"""
return self.filter(deferred=False)
def deferred(self):
"""
Return a QuerySet of all deferred messages in the queue.
"""
return self.filter(deferred=True)
def retry_deferred(self, new_priority=None):
"""
Reset the deferred flag for all deferred messages so they will be
retried.
"""
count = self.deferred().count()
update_kwargs = dict(deferred=False)
if new_priority is not None:
update_kwargs['priority'] = new_priority
self.deferred().update(**update_kwargs)
return count
Update the retries count of a queued message when it is changed back from deferred | from django.db import models
from django_mailer import constants
class QueueManager(models.Manager):
use_for_related_fields = True
def high_priority(self):
"""
Return a QuerySet of high priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_HIGH)
def normal_priority(self):
"""
Return a QuerySet of normal priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_NORMAL)
def low_priority(self):
"""
Return a QuerySet of low priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_LOW)
def non_deferred(self):
"""
Return a QuerySet containing all non-deferred queued messages.
"""
return self.filter(deferred=False)
def deferred(self):
"""
Return a QuerySet of all deferred messages in the queue.
"""
return self.filter(deferred=True)
def retry_deferred(self, new_priority=None):
"""
Reset the deferred flag for all deferred messages so they will be
retried.
"""
count = self.deferred().count()
update_kwargs = dict(deferred=False, retries=models.F('retries')+1)
if new_priority is not None:
update_kwargs['priority'] = new_priority
self.deferred().update(**update_kwargs)
return count
| <commit_before>from django.db import models
from django_mailer import constants
class QueueManager(models.Manager):
use_for_related_fields = True
def high_priority(self):
"""
Return a QuerySet of high priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_HIGH)
def normal_priority(self):
"""
Return a QuerySet of normal priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_NORMAL)
def low_priority(self):
"""
Return a QuerySet of low priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_LOW)
def non_deferred(self):
"""
Return a QuerySet containing all non-deferred queued messages.
"""
return self.filter(deferred=False)
def deferred(self):
"""
Return a QuerySet of all deferred messages in the queue.
"""
return self.filter(deferred=True)
def retry_deferred(self, new_priority=None):
"""
Reset the deferred flag for all deferred messages so they will be
retried.
"""
count = self.deferred().count()
update_kwargs = dict(deferred=False)
if new_priority is not None:
update_kwargs['priority'] = new_priority
self.deferred().update(**update_kwargs)
return count
<commit_msg>Update the retries count of a queued message when it is changed back from deferred<commit_after> | from django.db import models
from django_mailer import constants
class QueueManager(models.Manager):
use_for_related_fields = True
def high_priority(self):
"""
Return a QuerySet of high priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_HIGH)
def normal_priority(self):
"""
Return a QuerySet of normal priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_NORMAL)
def low_priority(self):
"""
Return a QuerySet of low priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_LOW)
def non_deferred(self):
"""
Return a QuerySet containing all non-deferred queued messages.
"""
return self.filter(deferred=False)
def deferred(self):
"""
Return a QuerySet of all deferred messages in the queue.
"""
return self.filter(deferred=True)
def retry_deferred(self, new_priority=None):
"""
Reset the deferred flag for all deferred messages so they will be
retried.
"""
count = self.deferred().count()
update_kwargs = dict(deferred=False, retries=models.F('retries')+1)
if new_priority is not None:
update_kwargs['priority'] = new_priority
self.deferred().update(**update_kwargs)
return count
| from django.db import models
from django_mailer import constants
class QueueManager(models.Manager):
use_for_related_fields = True
def high_priority(self):
"""
Return a QuerySet of high priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_HIGH)
def normal_priority(self):
"""
Return a QuerySet of normal priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_NORMAL)
def low_priority(self):
"""
Return a QuerySet of low priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_LOW)
def non_deferred(self):
"""
Return a QuerySet containing all non-deferred queued messages.
"""
return self.filter(deferred=False)
def deferred(self):
"""
Return a QuerySet of all deferred messages in the queue.
"""
return self.filter(deferred=True)
def retry_deferred(self, new_priority=None):
"""
Reset the deferred flag for all deferred messages so they will be
retried.
"""
count = self.deferred().count()
update_kwargs = dict(deferred=False)
if new_priority is not None:
update_kwargs['priority'] = new_priority
self.deferred().update(**update_kwargs)
return count
Update the retries count of a queued message when it is changed back from deferredfrom django.db import models
from django_mailer import constants
class QueueManager(models.Manager):
use_for_related_fields = True
def high_priority(self):
"""
Return a QuerySet of high priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_HIGH)
def normal_priority(self):
"""
Return a QuerySet of normal priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_NORMAL)
def low_priority(self):
"""
Return a QuerySet of low priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_LOW)
def non_deferred(self):
"""
Return a QuerySet containing all non-deferred queued messages.
"""
return self.filter(deferred=False)
def deferred(self):
"""
Return a QuerySet of all deferred messages in the queue.
"""
return self.filter(deferred=True)
def retry_deferred(self, new_priority=None):
"""
Reset the deferred flag for all deferred messages so they will be
retried.
"""
count = self.deferred().count()
update_kwargs = dict(deferred=False, retries=models.F('retries')+1)
if new_priority is not None:
update_kwargs['priority'] = new_priority
self.deferred().update(**update_kwargs)
return count
| <commit_before>from django.db import models
from django_mailer import constants
class QueueManager(models.Manager):
use_for_related_fields = True
def high_priority(self):
"""
Return a QuerySet of high priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_HIGH)
def normal_priority(self):
"""
Return a QuerySet of normal priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_NORMAL)
def low_priority(self):
"""
Return a QuerySet of low priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_LOW)
def non_deferred(self):
"""
Return a QuerySet containing all non-deferred queued messages.
"""
return self.filter(deferred=False)
def deferred(self):
"""
Return a QuerySet of all deferred messages in the queue.
"""
return self.filter(deferred=True)
def retry_deferred(self, new_priority=None):
"""
Reset the deferred flag for all deferred messages so they will be
retried.
"""
count = self.deferred().count()
update_kwargs = dict(deferred=False)
if new_priority is not None:
update_kwargs['priority'] = new_priority
self.deferred().update(**update_kwargs)
return count
<commit_msg>Update the retries count of a queued message when it is changed back from deferred<commit_after>from django.db import models
from django_mailer import constants
class QueueManager(models.Manager):
use_for_related_fields = True
def high_priority(self):
"""
Return a QuerySet of high priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_HIGH)
def normal_priority(self):
"""
Return a QuerySet of normal priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_NORMAL)
def low_priority(self):
"""
Return a QuerySet of low priority queued messages.
"""
return self.filter(priority=constants.PRIORITY_LOW)
def non_deferred(self):
"""
Return a QuerySet containing all non-deferred queued messages.
"""
return self.filter(deferred=False)
def deferred(self):
"""
Return a QuerySet of all deferred messages in the queue.
"""
return self.filter(deferred=True)
def retry_deferred(self, new_priority=None):
"""
Reset the deferred flag for all deferred messages so they will be
retried.
"""
count = self.deferred().count()
update_kwargs = dict(deferred=False, retries=models.F('retries')+1)
if new_priority is not None:
update_kwargs['priority'] = new_priority
self.deferred().update(**update_kwargs)
return count
|
53a38a716c01cfd15bc1aff89c6c7908a5218bfb | integration_tests/experiment_type.py | integration_tests/experiment_type.py | from enum import unique, IntEnum
@unique
class ExperimentType(IntEnum):
Detumbling = 1
EraseFlash = 2
SunS = 3
LEOP = 4
RadFET = 5
SADS = 6
Sail = 7
Fibo = 8
Payload = 9
Camera = 10
@unique
class StartResult(IntEnum):
Success = 0
Failure = 1
@unique
class IterationResult(IntEnum):
Finished = 0
LoopImmediately = 1
WaitForNextCycle = 2,
Failure = 3
| from enum import unique, IntEnum
@unique
class ExperimentType(IntEnum):
Detumbling = 1
EraseFlash = 2
SunS = 3
LEOP = 4
RadFET = 5
SADS = 6
Sail = 7
Fibo = 8
Payload = 9
Camera = 10
@unique
class StartResult(IntEnum):
Success = 0
Failure = 1
@unique
class IterationResult(IntEnum):
NoResult = 0
Finished = 1
LoopImmediately = 2
WaitForNextCycle = 3
Failure = 4
| Fix python projection of experiment iteration result enumeration. | Fix python projection of experiment iteration result enumeration.
| Python | agpl-3.0 | PW-Sat2/PWSat2OBC,PW-Sat2/PWSat2OBC,PW-Sat2/PWSat2OBC,PW-Sat2/PWSat2OBC | from enum import unique, IntEnum
@unique
class ExperimentType(IntEnum):
Detumbling = 1
EraseFlash = 2
SunS = 3
LEOP = 4
RadFET = 5
SADS = 6
Sail = 7
Fibo = 8
Payload = 9
Camera = 10
@unique
class StartResult(IntEnum):
Success = 0
Failure = 1
@unique
class IterationResult(IntEnum):
Finished = 0
LoopImmediately = 1
WaitForNextCycle = 2,
Failure = 3
Fix python projection of experiment iteration result enumeration. | from enum import unique, IntEnum
@unique
class ExperimentType(IntEnum):
Detumbling = 1
EraseFlash = 2
SunS = 3
LEOP = 4
RadFET = 5
SADS = 6
Sail = 7
Fibo = 8
Payload = 9
Camera = 10
@unique
class StartResult(IntEnum):
Success = 0
Failure = 1
@unique
class IterationResult(IntEnum):
NoResult = 0
Finished = 1
LoopImmediately = 2
WaitForNextCycle = 3
Failure = 4
| <commit_before>from enum import unique, IntEnum
@unique
class ExperimentType(IntEnum):
Detumbling = 1
EraseFlash = 2
SunS = 3
LEOP = 4
RadFET = 5
SADS = 6
Sail = 7
Fibo = 8
Payload = 9
Camera = 10
@unique
class StartResult(IntEnum):
Success = 0
Failure = 1
@unique
class IterationResult(IntEnum):
Finished = 0
LoopImmediately = 1
WaitForNextCycle = 2,
Failure = 3
<commit_msg>Fix python projection of experiment iteration result enumeration.<commit_after> | from enum import unique, IntEnum
@unique
class ExperimentType(IntEnum):
Detumbling = 1
EraseFlash = 2
SunS = 3
LEOP = 4
RadFET = 5
SADS = 6
Sail = 7
Fibo = 8
Payload = 9
Camera = 10
@unique
class StartResult(IntEnum):
Success = 0
Failure = 1
@unique
class IterationResult(IntEnum):
NoResult = 0
Finished = 1
LoopImmediately = 2
WaitForNextCycle = 3
Failure = 4
| from enum import unique, IntEnum
@unique
class ExperimentType(IntEnum):
Detumbling = 1
EraseFlash = 2
SunS = 3
LEOP = 4
RadFET = 5
SADS = 6
Sail = 7
Fibo = 8
Payload = 9
Camera = 10
@unique
class StartResult(IntEnum):
Success = 0
Failure = 1
@unique
class IterationResult(IntEnum):
Finished = 0
LoopImmediately = 1
WaitForNextCycle = 2,
Failure = 3
Fix python projection of experiment iteration result enumeration.from enum import unique, IntEnum
@unique
class ExperimentType(IntEnum):
Detumbling = 1
EraseFlash = 2
SunS = 3
LEOP = 4
RadFET = 5
SADS = 6
Sail = 7
Fibo = 8
Payload = 9
Camera = 10
@unique
class StartResult(IntEnum):
Success = 0
Failure = 1
@unique
class IterationResult(IntEnum):
NoResult = 0
Finished = 1
LoopImmediately = 2
WaitForNextCycle = 3
Failure = 4
| <commit_before>from enum import unique, IntEnum
@unique
class ExperimentType(IntEnum):
Detumbling = 1
EraseFlash = 2
SunS = 3
LEOP = 4
RadFET = 5
SADS = 6
Sail = 7
Fibo = 8
Payload = 9
Camera = 10
@unique
class StartResult(IntEnum):
Success = 0
Failure = 1
@unique
class IterationResult(IntEnum):
Finished = 0
LoopImmediately = 1
WaitForNextCycle = 2,
Failure = 3
<commit_msg>Fix python projection of experiment iteration result enumeration.<commit_after>from enum import unique, IntEnum
@unique
class ExperimentType(IntEnum):
Detumbling = 1
EraseFlash = 2
SunS = 3
LEOP = 4
RadFET = 5
SADS = 6
Sail = 7
Fibo = 8
Payload = 9
Camera = 10
@unique
class StartResult(IntEnum):
Success = 0
Failure = 1
@unique
class IterationResult(IntEnum):
NoResult = 0
Finished = 1
LoopImmediately = 2
WaitForNextCycle = 3
Failure = 4
|
31212104810f6c700ccc9561ac3d355b1894ef47 | glimpse/__init__.py | glimpse/__init__.py | """
Hierarchical visual models in C++ and Python
============================================
The Glimpse project is a library for implementing hierarchical visual models
in C++ and Python. The goal of this project is to allow a broad range of
feed-forward, hierarchical models to be encoded in a high-level declarative
manner, with low-level details of the implementation hidden from view. This
project combines an efficient implementation with the ability to leverage
parallel processing facilities and is designed to run on multiple operating
systems using only common, freely-available components.
See http://pythonhosted.org/glimpse/ for complete documentation.
"""
import sys
__version__ = '0.2.1'
| """
Hierarchical visual models in C++ and Python
============================================
The Glimpse project is a library for implementing hierarchical visual models
in C++ and Python. The goal of this project is to allow a broad range of
feed-forward, hierarchical models to be encoded in a high-level declarative
manner, with low-level details of the implementation hidden from view. This
project combines an efficient implementation with the ability to leverage
parallel processing facilities and is designed to run on multiple operating
systems using only common, freely-available components.
See http://pythonhosted.org/glimpse/ for complete documentation.
"""
import sys
__version__ = '0.2.1'
import glimpse.util.pil_fix # workaround PIL error on OS X
| Fix PIL bug on OS X. | Fix PIL bug on OS X.
| Python | mit | mthomure/glimpse-project,mthomure/glimpse-project,mthomure/glimpse-project | """
Hierarchical visual models in C++ and Python
============================================
The Glimpse project is a library for implementing hierarchical visual models
in C++ and Python. The goal of this project is to allow a broad range of
feed-forward, hierarchical models to be encoded in a high-level declarative
manner, with low-level details of the implementation hidden from view. This
project combines an efficient implementation with the ability to leverage
parallel processing facilities and is designed to run on multiple operating
systems using only common, freely-available components.
See http://pythonhosted.org/glimpse/ for complete documentation.
"""
import sys
__version__ = '0.2.1'
Fix PIL bug on OS X. | """
Hierarchical visual models in C++ and Python
============================================
The Glimpse project is a library for implementing hierarchical visual models
in C++ and Python. The goal of this project is to allow a broad range of
feed-forward, hierarchical models to be encoded in a high-level declarative
manner, with low-level details of the implementation hidden from view. This
project combines an efficient implementation with the ability to leverage
parallel processing facilities and is designed to run on multiple operating
systems using only common, freely-available components.
See http://pythonhosted.org/glimpse/ for complete documentation.
"""
import sys
__version__ = '0.2.1'
import glimpse.util.pil_fix # workaround PIL error on OS X
| <commit_before>"""
Hierarchical visual models in C++ and Python
============================================
The Glimpse project is a library for implementing hierarchical visual models
in C++ and Python. The goal of this project is to allow a broad range of
feed-forward, hierarchical models to be encoded in a high-level declarative
manner, with low-level details of the implementation hidden from view. This
project combines an efficient implementation with the ability to leverage
parallel processing facilities and is designed to run on multiple operating
systems using only common, freely-available components.
See http://pythonhosted.org/glimpse/ for complete documentation.
"""
import sys
__version__ = '0.2.1'
<commit_msg>Fix PIL bug on OS X.<commit_after> | """
Hierarchical visual models in C++ and Python
============================================
The Glimpse project is a library for implementing hierarchical visual models
in C++ and Python. The goal of this project is to allow a broad range of
feed-forward, hierarchical models to be encoded in a high-level declarative
manner, with low-level details of the implementation hidden from view. This
project combines an efficient implementation with the ability to leverage
parallel processing facilities and is designed to run on multiple operating
systems using only common, freely-available components.
See http://pythonhosted.org/glimpse/ for complete documentation.
"""
import sys
__version__ = '0.2.1'
import glimpse.util.pil_fix # workaround PIL error on OS X
| """
Hierarchical visual models in C++ and Python
============================================
The Glimpse project is a library for implementing hierarchical visual models
in C++ and Python. The goal of this project is to allow a broad range of
feed-forward, hierarchical models to be encoded in a high-level declarative
manner, with low-level details of the implementation hidden from view. This
project combines an efficient implementation with the ability to leverage
parallel processing facilities and is designed to run on multiple operating
systems using only common, freely-available components.
See http://pythonhosted.org/glimpse/ for complete documentation.
"""
import sys
__version__ = '0.2.1'
Fix PIL bug on OS X."""
Hierarchical visual models in C++ and Python
============================================
The Glimpse project is a library for implementing hierarchical visual models
in C++ and Python. The goal of this project is to allow a broad range of
feed-forward, hierarchical models to be encoded in a high-level declarative
manner, with low-level details of the implementation hidden from view. This
project combines an efficient implementation with the ability to leverage
parallel processing facilities and is designed to run on multiple operating
systems using only common, freely-available components.
See http://pythonhosted.org/glimpse/ for complete documentation.
"""
import sys
__version__ = '0.2.1'
import glimpse.util.pil_fix # workaround PIL error on OS X
| <commit_before>"""
Hierarchical visual models in C++ and Python
============================================
The Glimpse project is a library for implementing hierarchical visual models
in C++ and Python. The goal of this project is to allow a broad range of
feed-forward, hierarchical models to be encoded in a high-level declarative
manner, with low-level details of the implementation hidden from view. This
project combines an efficient implementation with the ability to leverage
parallel processing facilities and is designed to run on multiple operating
systems using only common, freely-available components.
See http://pythonhosted.org/glimpse/ for complete documentation.
"""
import sys
__version__ = '0.2.1'
<commit_msg>Fix PIL bug on OS X.<commit_after>"""
Hierarchical visual models in C++ and Python
============================================
The Glimpse project is a library for implementing hierarchical visual models
in C++ and Python. The goal of this project is to allow a broad range of
feed-forward, hierarchical models to be encoded in a high-level declarative
manner, with low-level details of the implementation hidden from view. This
project combines an efficient implementation with the ability to leverage
parallel processing facilities and is designed to run on multiple operating
systems using only common, freely-available components.
See http://pythonhosted.org/glimpse/ for complete documentation.
"""
import sys
__version__ = '0.2.1'
import glimpse.util.pil_fix # workaround PIL error on OS X
|
ce85550a4bf080e629fbf1443d31a5305c0e0ac3 | IPython/utils/tests/test_tempdir.py | IPython/utils/tests/test_tempdir.py | #-----------------------------------------------------------------------------
# Copyright (C) 2012- The IPython Development Team
#
# Distributed under the terms of the BSD License. The full license is in
# the file COPYING, distributed as part of this software.
#-----------------------------------------------------------------------------
import os
from IPython.utils.tempdir import NamedFileInTemporaryDirectory
def test_named_file_in_temporary_directory():
with NamedFileInTemporaryDirectory('filename') as file:
name = file.name
assert not file.closed
assert os.path.exists(name)
file.write('test')
assert file.closed
assert not os.path.exists(name)
| #-----------------------------------------------------------------------------
# Copyright (C) 2012- The IPython Development Team
#
# Distributed under the terms of the BSD License. The full license is in
# the file COPYING, distributed as part of this software.
#-----------------------------------------------------------------------------
import os
from IPython.utils.tempdir import NamedFileInTemporaryDirectory
def test_named_file_in_temporary_directory():
with NamedFileInTemporaryDirectory('filename') as file:
name = file.name
assert not file.closed
assert os.path.exists(name)
file.write(b'test')
assert file.closed
assert not os.path.exists(name)
| Fix failing test in Python 3 | Fix failing test in Python 3
| Python | bsd-3-clause | ipython/ipython,ipython/ipython | #-----------------------------------------------------------------------------
# Copyright (C) 2012- The IPython Development Team
#
# Distributed under the terms of the BSD License. The full license is in
# the file COPYING, distributed as part of this software.
#-----------------------------------------------------------------------------
import os
from IPython.utils.tempdir import NamedFileInTemporaryDirectory
def test_named_file_in_temporary_directory():
with NamedFileInTemporaryDirectory('filename') as file:
name = file.name
assert not file.closed
assert os.path.exists(name)
file.write('test')
assert file.closed
assert not os.path.exists(name)
Fix failing test in Python 3 | #-----------------------------------------------------------------------------
# Copyright (C) 2012- The IPython Development Team
#
# Distributed under the terms of the BSD License. The full license is in
# the file COPYING, distributed as part of this software.
#-----------------------------------------------------------------------------
import os
from IPython.utils.tempdir import NamedFileInTemporaryDirectory
def test_named_file_in_temporary_directory():
with NamedFileInTemporaryDirectory('filename') as file:
name = file.name
assert not file.closed
assert os.path.exists(name)
file.write(b'test')
assert file.closed
assert not os.path.exists(name)
| <commit_before>#-----------------------------------------------------------------------------
# Copyright (C) 2012- The IPython Development Team
#
# Distributed under the terms of the BSD License. The full license is in
# the file COPYING, distributed as part of this software.
#-----------------------------------------------------------------------------
import os
from IPython.utils.tempdir import NamedFileInTemporaryDirectory
def test_named_file_in_temporary_directory():
with NamedFileInTemporaryDirectory('filename') as file:
name = file.name
assert not file.closed
assert os.path.exists(name)
file.write('test')
assert file.closed
assert not os.path.exists(name)
<commit_msg>Fix failing test in Python 3<commit_after> | #-----------------------------------------------------------------------------
# Copyright (C) 2012- The IPython Development Team
#
# Distributed under the terms of the BSD License. The full license is in
# the file COPYING, distributed as part of this software.
#-----------------------------------------------------------------------------
import os
from IPython.utils.tempdir import NamedFileInTemporaryDirectory
def test_named_file_in_temporary_directory():
with NamedFileInTemporaryDirectory('filename') as file:
name = file.name
assert not file.closed
assert os.path.exists(name)
file.write(b'test')
assert file.closed
assert not os.path.exists(name)
| #-----------------------------------------------------------------------------
# Copyright (C) 2012- The IPython Development Team
#
# Distributed under the terms of the BSD License. The full license is in
# the file COPYING, distributed as part of this software.
#-----------------------------------------------------------------------------
import os
from IPython.utils.tempdir import NamedFileInTemporaryDirectory
def test_named_file_in_temporary_directory():
with NamedFileInTemporaryDirectory('filename') as file:
name = file.name
assert not file.closed
assert os.path.exists(name)
file.write('test')
assert file.closed
assert not os.path.exists(name)
Fix failing test in Python 3#-----------------------------------------------------------------------------
# Copyright (C) 2012- The IPython Development Team
#
# Distributed under the terms of the BSD License. The full license is in
# the file COPYING, distributed as part of this software.
#-----------------------------------------------------------------------------
import os
from IPython.utils.tempdir import NamedFileInTemporaryDirectory
def test_named_file_in_temporary_directory():
with NamedFileInTemporaryDirectory('filename') as file:
name = file.name
assert not file.closed
assert os.path.exists(name)
file.write(b'test')
assert file.closed
assert not os.path.exists(name)
| <commit_before>#-----------------------------------------------------------------------------
# Copyright (C) 2012- The IPython Development Team
#
# Distributed under the terms of the BSD License. The full license is in
# the file COPYING, distributed as part of this software.
#-----------------------------------------------------------------------------
import os
from IPython.utils.tempdir import NamedFileInTemporaryDirectory
def test_named_file_in_temporary_directory():
with NamedFileInTemporaryDirectory('filename') as file:
name = file.name
assert not file.closed
assert os.path.exists(name)
file.write('test')
assert file.closed
assert not os.path.exists(name)
<commit_msg>Fix failing test in Python 3<commit_after>#-----------------------------------------------------------------------------
# Copyright (C) 2012- The IPython Development Team
#
# Distributed under the terms of the BSD License. The full license is in
# the file COPYING, distributed as part of this software.
#-----------------------------------------------------------------------------
import os
from IPython.utils.tempdir import NamedFileInTemporaryDirectory
def test_named_file_in_temporary_directory():
with NamedFileInTemporaryDirectory('filename') as file:
name = file.name
assert not file.closed
assert os.path.exists(name)
file.write(b'test')
assert file.closed
assert not os.path.exists(name)
|
bf7f726821f2ac74e99fd5fd06729ea2becab0c9 | ModuleInterface.py | ModuleInterface.py | import GlobalVars
class ModuleInterface(object):
triggers = []
acceptedTypes = ['PRIVMSG']
help = '<no help defined (yet)>'
def __init__(self):
self.onStart()
def onStart(self):
pass
def hasAlias(self, message):
if message.Type in self.acceptedTypes and message.Command in GlobalVars.commandAliases.keys():
return True
def shouldTrigger(self, message):
if message.Type not in self.acceptedTypes:
return False
if message.Command not in self.triggers:
return False
return True
def onTrigger(self, Hubbot, message):
pass | import GlobalVars
class ModuleInterface(object):
triggers = []
acceptedTypes = ['PRIVMSG']
help = '<no help defined (yet)>'
def __init__(self):
self.onStart()
def onStart(self):
pass
def hasAlias(self, message):
if message.Type in self.acceptedTypes and message.Command in GlobalVars.commandAliases.keys():
return True
else:
return False
def shouldTrigger(self, message):
if message.Type not in self.acceptedTypes:
return False
if message.Command not in self.triggers:
return False
return True
def onTrigger(self, Hubbot, message):
pass | Return false if no alias. | [ModuleInteface] Return false if no alias.
| Python | mit | HubbeKing/Hubbot_Twisted | import GlobalVars
class ModuleInterface(object):
triggers = []
acceptedTypes = ['PRIVMSG']
help = '<no help defined (yet)>'
def __init__(self):
self.onStart()
def onStart(self):
pass
def hasAlias(self, message):
if message.Type in self.acceptedTypes and message.Command in GlobalVars.commandAliases.keys():
return True
def shouldTrigger(self, message):
if message.Type not in self.acceptedTypes:
return False
if message.Command not in self.triggers:
return False
return True
def onTrigger(self, Hubbot, message):
pass[ModuleInteface] Return false if no alias. | import GlobalVars
class ModuleInterface(object):
triggers = []
acceptedTypes = ['PRIVMSG']
help = '<no help defined (yet)>'
def __init__(self):
self.onStart()
def onStart(self):
pass
def hasAlias(self, message):
if message.Type in self.acceptedTypes and message.Command in GlobalVars.commandAliases.keys():
return True
else:
return False
def shouldTrigger(self, message):
if message.Type not in self.acceptedTypes:
return False
if message.Command not in self.triggers:
return False
return True
def onTrigger(self, Hubbot, message):
pass | <commit_before>import GlobalVars
class ModuleInterface(object):
triggers = []
acceptedTypes = ['PRIVMSG']
help = '<no help defined (yet)>'
def __init__(self):
self.onStart()
def onStart(self):
pass
def hasAlias(self, message):
if message.Type in self.acceptedTypes and message.Command in GlobalVars.commandAliases.keys():
return True
def shouldTrigger(self, message):
if message.Type not in self.acceptedTypes:
return False
if message.Command not in self.triggers:
return False
return True
def onTrigger(self, Hubbot, message):
pass<commit_msg>[ModuleInteface] Return false if no alias.<commit_after> | import GlobalVars
class ModuleInterface(object):
triggers = []
acceptedTypes = ['PRIVMSG']
help = '<no help defined (yet)>'
def __init__(self):
self.onStart()
def onStart(self):
pass
def hasAlias(self, message):
if message.Type in self.acceptedTypes and message.Command in GlobalVars.commandAliases.keys():
return True
else:
return False
def shouldTrigger(self, message):
if message.Type not in self.acceptedTypes:
return False
if message.Command not in self.triggers:
return False
return True
def onTrigger(self, Hubbot, message):
pass | import GlobalVars
class ModuleInterface(object):
triggers = []
acceptedTypes = ['PRIVMSG']
help = '<no help defined (yet)>'
def __init__(self):
self.onStart()
def onStart(self):
pass
def hasAlias(self, message):
if message.Type in self.acceptedTypes and message.Command in GlobalVars.commandAliases.keys():
return True
def shouldTrigger(self, message):
if message.Type not in self.acceptedTypes:
return False
if message.Command not in self.triggers:
return False
return True
def onTrigger(self, Hubbot, message):
pass[ModuleInteface] Return false if no alias.import GlobalVars
class ModuleInterface(object):
triggers = []
acceptedTypes = ['PRIVMSG']
help = '<no help defined (yet)>'
def __init__(self):
self.onStart()
def onStart(self):
pass
def hasAlias(self, message):
if message.Type in self.acceptedTypes and message.Command in GlobalVars.commandAliases.keys():
return True
else:
return False
def shouldTrigger(self, message):
if message.Type not in self.acceptedTypes:
return False
if message.Command not in self.triggers:
return False
return True
def onTrigger(self, Hubbot, message):
pass | <commit_before>import GlobalVars
class ModuleInterface(object):
triggers = []
acceptedTypes = ['PRIVMSG']
help = '<no help defined (yet)>'
def __init__(self):
self.onStart()
def onStart(self):
pass
def hasAlias(self, message):
if message.Type in self.acceptedTypes and message.Command in GlobalVars.commandAliases.keys():
return True
def shouldTrigger(self, message):
if message.Type not in self.acceptedTypes:
return False
if message.Command not in self.triggers:
return False
return True
def onTrigger(self, Hubbot, message):
pass<commit_msg>[ModuleInteface] Return false if no alias.<commit_after>import GlobalVars
class ModuleInterface(object):
triggers = []
acceptedTypes = ['PRIVMSG']
help = '<no help defined (yet)>'
def __init__(self):
self.onStart()
def onStart(self):
pass
def hasAlias(self, message):
if message.Type in self.acceptedTypes and message.Command in GlobalVars.commandAliases.keys():
return True
else:
return False
def shouldTrigger(self, message):
if message.Type not in self.acceptedTypes:
return False
if message.Command not in self.triggers:
return False
return True
def onTrigger(self, Hubbot, message):
pass |
2c4f32fbc407acb0b65ddc6fb71d192af74e740e | tcconfig/parser/_interface.py | tcconfig/parser/_interface.py | # encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com>
"""
from __future__ import absolute_import, unicode_literals
import abc
import six
@six.add_metaclass(abc.ABCMeta)
class ParserInterface(object):
@abc.abstractmethod
def parse(self, text): # pragma: no cover
pass
@six.add_metaclass(abc.ABCMeta)
class AbstractParser(ParserInterface):
def __init__(self):
self._clear()
@abc.abstractproperty
def _tc_subcommand(self): # pragma: no cover
pass
@abc.abstractmethod
def _clear(self): # pragma: no cover
pass
@staticmethod
def _to_unicode(text):
try:
return text.decode("ascii")
except AttributeError:
return text
| # encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com>
"""
from __future__ import absolute_import, unicode_literals
import abc
import six
@six.add_metaclass(abc.ABCMeta)
class ParserInterface(object):
@abc.abstractmethod
def parse(self, device, text): # pragma: no cover
pass
@six.add_metaclass(abc.ABCMeta)
class AbstractParser(ParserInterface):
def __init__(self):
self._clear()
@abc.abstractproperty
def _tc_subcommand(self): # pragma: no cover
pass
@abc.abstractmethod
def _clear(self): # pragma: no cover
pass
@staticmethod
def _to_unicode(text):
try:
return text.decode("ascii")
except AttributeError:
return text
| Align the method signature with subclasses | Align the method signature with subclasses
| Python | mit | thombashi/tcconfig,thombashi/tcconfig | # encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com>
"""
from __future__ import absolute_import, unicode_literals
import abc
import six
@six.add_metaclass(abc.ABCMeta)
class ParserInterface(object):
@abc.abstractmethod
def parse(self, text): # pragma: no cover
pass
@six.add_metaclass(abc.ABCMeta)
class AbstractParser(ParserInterface):
def __init__(self):
self._clear()
@abc.abstractproperty
def _tc_subcommand(self): # pragma: no cover
pass
@abc.abstractmethod
def _clear(self): # pragma: no cover
pass
@staticmethod
def _to_unicode(text):
try:
return text.decode("ascii")
except AttributeError:
return text
Align the method signature with subclasses | # encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com>
"""
from __future__ import absolute_import, unicode_literals
import abc
import six
@six.add_metaclass(abc.ABCMeta)
class ParserInterface(object):
@abc.abstractmethod
def parse(self, device, text): # pragma: no cover
pass
@six.add_metaclass(abc.ABCMeta)
class AbstractParser(ParserInterface):
def __init__(self):
self._clear()
@abc.abstractproperty
def _tc_subcommand(self): # pragma: no cover
pass
@abc.abstractmethod
def _clear(self): # pragma: no cover
pass
@staticmethod
def _to_unicode(text):
try:
return text.decode("ascii")
except AttributeError:
return text
| <commit_before># encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com>
"""
from __future__ import absolute_import, unicode_literals
import abc
import six
@six.add_metaclass(abc.ABCMeta)
class ParserInterface(object):
@abc.abstractmethod
def parse(self, text): # pragma: no cover
pass
@six.add_metaclass(abc.ABCMeta)
class AbstractParser(ParserInterface):
def __init__(self):
self._clear()
@abc.abstractproperty
def _tc_subcommand(self): # pragma: no cover
pass
@abc.abstractmethod
def _clear(self): # pragma: no cover
pass
@staticmethod
def _to_unicode(text):
try:
return text.decode("ascii")
except AttributeError:
return text
<commit_msg>Align the method signature with subclasses<commit_after> | # encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com>
"""
from __future__ import absolute_import, unicode_literals
import abc
import six
@six.add_metaclass(abc.ABCMeta)
class ParserInterface(object):
@abc.abstractmethod
def parse(self, device, text): # pragma: no cover
pass
@six.add_metaclass(abc.ABCMeta)
class AbstractParser(ParserInterface):
def __init__(self):
self._clear()
@abc.abstractproperty
def _tc_subcommand(self): # pragma: no cover
pass
@abc.abstractmethod
def _clear(self): # pragma: no cover
pass
@staticmethod
def _to_unicode(text):
try:
return text.decode("ascii")
except AttributeError:
return text
| # encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com>
"""
from __future__ import absolute_import, unicode_literals
import abc
import six
@six.add_metaclass(abc.ABCMeta)
class ParserInterface(object):
@abc.abstractmethod
def parse(self, text): # pragma: no cover
pass
@six.add_metaclass(abc.ABCMeta)
class AbstractParser(ParserInterface):
def __init__(self):
self._clear()
@abc.abstractproperty
def _tc_subcommand(self): # pragma: no cover
pass
@abc.abstractmethod
def _clear(self): # pragma: no cover
pass
@staticmethod
def _to_unicode(text):
try:
return text.decode("ascii")
except AttributeError:
return text
Align the method signature with subclasses# encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com>
"""
from __future__ import absolute_import, unicode_literals
import abc
import six
@six.add_metaclass(abc.ABCMeta)
class ParserInterface(object):
@abc.abstractmethod
def parse(self, device, text): # pragma: no cover
pass
@six.add_metaclass(abc.ABCMeta)
class AbstractParser(ParserInterface):
def __init__(self):
self._clear()
@abc.abstractproperty
def _tc_subcommand(self): # pragma: no cover
pass
@abc.abstractmethod
def _clear(self): # pragma: no cover
pass
@staticmethod
def _to_unicode(text):
try:
return text.decode("ascii")
except AttributeError:
return text
| <commit_before># encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com>
"""
from __future__ import absolute_import, unicode_literals
import abc
import six
@six.add_metaclass(abc.ABCMeta)
class ParserInterface(object):
@abc.abstractmethod
def parse(self, text): # pragma: no cover
pass
@six.add_metaclass(abc.ABCMeta)
class AbstractParser(ParserInterface):
def __init__(self):
self._clear()
@abc.abstractproperty
def _tc_subcommand(self): # pragma: no cover
pass
@abc.abstractmethod
def _clear(self): # pragma: no cover
pass
@staticmethod
def _to_unicode(text):
try:
return text.decode("ascii")
except AttributeError:
return text
<commit_msg>Align the method signature with subclasses<commit_after># encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com>
"""
from __future__ import absolute_import, unicode_literals
import abc
import six
@six.add_metaclass(abc.ABCMeta)
class ParserInterface(object):
@abc.abstractmethod
def parse(self, device, text): # pragma: no cover
pass
@six.add_metaclass(abc.ABCMeta)
class AbstractParser(ParserInterface):
def __init__(self):
self._clear()
@abc.abstractproperty
def _tc_subcommand(self): # pragma: no cover
pass
@abc.abstractmethod
def _clear(self): # pragma: no cover
pass
@staticmethod
def _to_unicode(text):
try:
return text.decode("ascii")
except AttributeError:
return text
|
bdc51bb7a71dba4a25fcded0d2758a9af5e15679 | pythran/tests/test_doc.py | pythran/tests/test_doc.py | import unittest
import doctest
class TestDoctest(unittest.TestCase):
modules = ('passes',)
def test_package(self):
import pythran
failed, _ = doctest.testmod(pythran)
self.assertEqual(failed, 0)
def test_passes(self):
from pythran import passes
failed, _ = doctest.testmod(passes)
self.assertEqual(failed, 0)
def test_optimizations(self):
from pythran import optimizations
failed, _ = doctest.testmod(optimizations)
self.assertEqual(failed, 0)
def test_backend(self):
from pythran import backend
failed, _ = doctest.testmod(backend)
self.assertEqual(failed, 0)
def test_cxxtypes(self):
from pythran import cxxtypes
failed, _ = doctest.testmod(cxxtypes)
self.assertEqual(failed, 0)
def test_openmp(self):
from pythran import openmp
failed, _ = doctest.testmod(openmp)
self.assertEqual(failed, 0)
#def test_typing(self):
# from pythran import typing
# failed, _ = doctest.testmod(typing)
# self.assertEqual(failed, 0)
if __name__ == '__main__':
unittest.main()
| import unittest
import doctest
import pythran
import inspect
class TestDoctest(unittest.TestCase):
'''
Enable automatic doctest integration to unittest
Every module in the pythran package is scanned for doctests
and one test per module is created
'''
pass
def generic_test_package(self, mod):
failed, _ = doctest.testmod(mod)
self.assertEqual(failed, 0)
def add_module_doctest(module_name):
module = getattr(pythran, module_name)
if inspect.ismodule(module):
setattr(TestDoctest, 'test_' + module_name,
lambda self: generic_test_package(self, module))
map(add_module_doctest, dir(pythran))
if __name__ == '__main__':
unittest.main()
| Make doctest integration in unittest generic. | Make doctest integration in unittest generic.
Rely on introspection rather than redundant typing.
| Python | bsd-3-clause | pbrunet/pythran,artas360/pythran,pbrunet/pythran,pbrunet/pythran,serge-sans-paille/pythran,artas360/pythran,hainm/pythran,serge-sans-paille/pythran,hainm/pythran,pombredanne/pythran,pombredanne/pythran,pombredanne/pythran,hainm/pythran,artas360/pythran | import unittest
import doctest
class TestDoctest(unittest.TestCase):
modules = ('passes',)
def test_package(self):
import pythran
failed, _ = doctest.testmod(pythran)
self.assertEqual(failed, 0)
def test_passes(self):
from pythran import passes
failed, _ = doctest.testmod(passes)
self.assertEqual(failed, 0)
def test_optimizations(self):
from pythran import optimizations
failed, _ = doctest.testmod(optimizations)
self.assertEqual(failed, 0)
def test_backend(self):
from pythran import backend
failed, _ = doctest.testmod(backend)
self.assertEqual(failed, 0)
def test_cxxtypes(self):
from pythran import cxxtypes
failed, _ = doctest.testmod(cxxtypes)
self.assertEqual(failed, 0)
def test_openmp(self):
from pythran import openmp
failed, _ = doctest.testmod(openmp)
self.assertEqual(failed, 0)
#def test_typing(self):
# from pythran import typing
# failed, _ = doctest.testmod(typing)
# self.assertEqual(failed, 0)
if __name__ == '__main__':
unittest.main()
Make doctest integration in unittest generic.
Rely on introspection rather than redundant typing. | import unittest
import doctest
import pythran
import inspect
class TestDoctest(unittest.TestCase):
'''
Enable automatic doctest integration to unittest
Every module in the pythran package is scanned for doctests
and one test per module is created
'''
pass
def generic_test_package(self, mod):
failed, _ = doctest.testmod(mod)
self.assertEqual(failed, 0)
def add_module_doctest(module_name):
module = getattr(pythran, module_name)
if inspect.ismodule(module):
setattr(TestDoctest, 'test_' + module_name,
lambda self: generic_test_package(self, module))
map(add_module_doctest, dir(pythran))
if __name__ == '__main__':
unittest.main()
| <commit_before>import unittest
import doctest
class TestDoctest(unittest.TestCase):
modules = ('passes',)
def test_package(self):
import pythran
failed, _ = doctest.testmod(pythran)
self.assertEqual(failed, 0)
def test_passes(self):
from pythran import passes
failed, _ = doctest.testmod(passes)
self.assertEqual(failed, 0)
def test_optimizations(self):
from pythran import optimizations
failed, _ = doctest.testmod(optimizations)
self.assertEqual(failed, 0)
def test_backend(self):
from pythran import backend
failed, _ = doctest.testmod(backend)
self.assertEqual(failed, 0)
def test_cxxtypes(self):
from pythran import cxxtypes
failed, _ = doctest.testmod(cxxtypes)
self.assertEqual(failed, 0)
def test_openmp(self):
from pythran import openmp
failed, _ = doctest.testmod(openmp)
self.assertEqual(failed, 0)
#def test_typing(self):
# from pythran import typing
# failed, _ = doctest.testmod(typing)
# self.assertEqual(failed, 0)
if __name__ == '__main__':
unittest.main()
<commit_msg>Make doctest integration in unittest generic.
Rely on introspection rather than redundant typing.<commit_after> | import unittest
import doctest
import pythran
import inspect
class TestDoctest(unittest.TestCase):
'''
Enable automatic doctest integration to unittest
Every module in the pythran package is scanned for doctests
and one test per module is created
'''
pass
def generic_test_package(self, mod):
failed, _ = doctest.testmod(mod)
self.assertEqual(failed, 0)
def add_module_doctest(module_name):
module = getattr(pythran, module_name)
if inspect.ismodule(module):
setattr(TestDoctest, 'test_' + module_name,
lambda self: generic_test_package(self, module))
map(add_module_doctest, dir(pythran))
if __name__ == '__main__':
unittest.main()
| import unittest
import doctest
class TestDoctest(unittest.TestCase):
modules = ('passes',)
def test_package(self):
import pythran
failed, _ = doctest.testmod(pythran)
self.assertEqual(failed, 0)
def test_passes(self):
from pythran import passes
failed, _ = doctest.testmod(passes)
self.assertEqual(failed, 0)
def test_optimizations(self):
from pythran import optimizations
failed, _ = doctest.testmod(optimizations)
self.assertEqual(failed, 0)
def test_backend(self):
from pythran import backend
failed, _ = doctest.testmod(backend)
self.assertEqual(failed, 0)
def test_cxxtypes(self):
from pythran import cxxtypes
failed, _ = doctest.testmod(cxxtypes)
self.assertEqual(failed, 0)
def test_openmp(self):
from pythran import openmp
failed, _ = doctest.testmod(openmp)
self.assertEqual(failed, 0)
#def test_typing(self):
# from pythran import typing
# failed, _ = doctest.testmod(typing)
# self.assertEqual(failed, 0)
if __name__ == '__main__':
unittest.main()
Make doctest integration in unittest generic.
Rely on introspection rather than redundant typing.import unittest
import doctest
import pythran
import inspect
class TestDoctest(unittest.TestCase):
'''
Enable automatic doctest integration to unittest
Every module in the pythran package is scanned for doctests
and one test per module is created
'''
pass
def generic_test_package(self, mod):
failed, _ = doctest.testmod(mod)
self.assertEqual(failed, 0)
def add_module_doctest(module_name):
module = getattr(pythran, module_name)
if inspect.ismodule(module):
setattr(TestDoctest, 'test_' + module_name,
lambda self: generic_test_package(self, module))
map(add_module_doctest, dir(pythran))
if __name__ == '__main__':
unittest.main()
| <commit_before>import unittest
import doctest
class TestDoctest(unittest.TestCase):
modules = ('passes',)
def test_package(self):
import pythran
failed, _ = doctest.testmod(pythran)
self.assertEqual(failed, 0)
def test_passes(self):
from pythran import passes
failed, _ = doctest.testmod(passes)
self.assertEqual(failed, 0)
def test_optimizations(self):
from pythran import optimizations
failed, _ = doctest.testmod(optimizations)
self.assertEqual(failed, 0)
def test_backend(self):
from pythran import backend
failed, _ = doctest.testmod(backend)
self.assertEqual(failed, 0)
def test_cxxtypes(self):
from pythran import cxxtypes
failed, _ = doctest.testmod(cxxtypes)
self.assertEqual(failed, 0)
def test_openmp(self):
from pythran import openmp
failed, _ = doctest.testmod(openmp)
self.assertEqual(failed, 0)
#def test_typing(self):
# from pythran import typing
# failed, _ = doctest.testmod(typing)
# self.assertEqual(failed, 0)
if __name__ == '__main__':
unittest.main()
<commit_msg>Make doctest integration in unittest generic.
Rely on introspection rather than redundant typing.<commit_after>import unittest
import doctest
import pythran
import inspect
class TestDoctest(unittest.TestCase):
'''
Enable automatic doctest integration to unittest
Every module in the pythran package is scanned for doctests
and one test per module is created
'''
pass
def generic_test_package(self, mod):
failed, _ = doctest.testmod(mod)
self.assertEqual(failed, 0)
def add_module_doctest(module_name):
module = getattr(pythran, module_name)
if inspect.ismodule(module):
setattr(TestDoctest, 'test_' + module_name,
lambda self: generic_test_package(self, module))
map(add_module_doctest, dir(pythran))
if __name__ == '__main__':
unittest.main()
|
9c7808b1f6571daaaf19dc1bfc57bf83cfb37bad | hybra/wordclouds.py | hybra/wordclouds.py | from __future__ import absolute_import, division, print_function, unicode_literals
from collections import Counter
import re
def create_wordcloud( data, stopwords = ["the", "a", "or", "tai", "and", "ja", "to", "on", "in", "of", "for", "is", "i", "this", "http", "www", "fi", "com"] ):
import types
if isinstance( data, types.GeneratorType ):
data = list( data )
if len(data) == 0:
print( "Dataset empty." )
return
from wordcloud import WordCloud
text = ''
for d in data:
text += d['text_content'].lower() + ' '
text = text.strip()
stopwords = map( lambda w: str(w), stopwords )
wc = WordCloud( background_color="white", width=800, height=400, stopwords = stopwords )
wc.generate( text )
plt.figure(figsize=(15,10))
plt.imshow(wc, interpolation="bilinear")
plt.axis("off")
plt.show()
| from __future__ import absolute_import, division, print_function, unicode_literals
from collections import Counter
import re
def create_wordcloud( data, plt, stopwords = ["the", "a", "or", "tai", "and", "ja", "to", "on", "in", "of", "for", "is", "i", "this", "http", "www", "fi", "com"] ):
import types
if isinstance( data, types.GeneratorType ):
data = list( data )
if len(data) == 0:
print( "Dataset empty." )
return
from wordcloud import WordCloud
text = ''
for d in data:
text += d['text_content'].lower() + ' '
text = text.strip()
stopwords = map( lambda w: str(w), stopwords )
wc = WordCloud( background_color="white", width=800, height=400, stopwords = stopwords )
wc.generate( text )
plt.figure(figsize=(15,10))
plt.imshow(wc, interpolation="bilinear")
plt.axis("off")
plt.show()
| Fix bug in wordcloud method | Fix bug in wordcloud method
| Python | mit | HIIT/hybra-core,HIIT/hybra-core,HIIT/hybra-core,HIIT/hybra-core,HIIT/hybra-core | from __future__ import absolute_import, division, print_function, unicode_literals
from collections import Counter
import re
def create_wordcloud( data, stopwords = ["the", "a", "or", "tai", "and", "ja", "to", "on", "in", "of", "for", "is", "i", "this", "http", "www", "fi", "com"] ):
import types
if isinstance( data, types.GeneratorType ):
data = list( data )
if len(data) == 0:
print( "Dataset empty." )
return
from wordcloud import WordCloud
text = ''
for d in data:
text += d['text_content'].lower() + ' '
text = text.strip()
stopwords = map( lambda w: str(w), stopwords )
wc = WordCloud( background_color="white", width=800, height=400, stopwords = stopwords )
wc.generate( text )
plt.figure(figsize=(15,10))
plt.imshow(wc, interpolation="bilinear")
plt.axis("off")
plt.show()
Fix bug in wordcloud method | from __future__ import absolute_import, division, print_function, unicode_literals
from collections import Counter
import re
def create_wordcloud( data, plt, stopwords = ["the", "a", "or", "tai", "and", "ja", "to", "on", "in", "of", "for", "is", "i", "this", "http", "www", "fi", "com"] ):
import types
if isinstance( data, types.GeneratorType ):
data = list( data )
if len(data) == 0:
print( "Dataset empty." )
return
from wordcloud import WordCloud
text = ''
for d in data:
text += d['text_content'].lower() + ' '
text = text.strip()
stopwords = map( lambda w: str(w), stopwords )
wc = WordCloud( background_color="white", width=800, height=400, stopwords = stopwords )
wc.generate( text )
plt.figure(figsize=(15,10))
plt.imshow(wc, interpolation="bilinear")
plt.axis("off")
plt.show()
| <commit_before>from __future__ import absolute_import, division, print_function, unicode_literals
from collections import Counter
import re
def create_wordcloud( data, stopwords = ["the", "a", "or", "tai", "and", "ja", "to", "on", "in", "of", "for", "is", "i", "this", "http", "www", "fi", "com"] ):
import types
if isinstance( data, types.GeneratorType ):
data = list( data )
if len(data) == 0:
print( "Dataset empty." )
return
from wordcloud import WordCloud
text = ''
for d in data:
text += d['text_content'].lower() + ' '
text = text.strip()
stopwords = map( lambda w: str(w), stopwords )
wc = WordCloud( background_color="white", width=800, height=400, stopwords = stopwords )
wc.generate( text )
plt.figure(figsize=(15,10))
plt.imshow(wc, interpolation="bilinear")
plt.axis("off")
plt.show()
<commit_msg>Fix bug in wordcloud method<commit_after> | from __future__ import absolute_import, division, print_function, unicode_literals
from collections import Counter
import re
def create_wordcloud( data, plt, stopwords = ["the", "a", "or", "tai", "and", "ja", "to", "on", "in", "of", "for", "is", "i", "this", "http", "www", "fi", "com"] ):
import types
if isinstance( data, types.GeneratorType ):
data = list( data )
if len(data) == 0:
print( "Dataset empty." )
return
from wordcloud import WordCloud
text = ''
for d in data:
text += d['text_content'].lower() + ' '
text = text.strip()
stopwords = map( lambda w: str(w), stopwords )
wc = WordCloud( background_color="white", width=800, height=400, stopwords = stopwords )
wc.generate( text )
plt.figure(figsize=(15,10))
plt.imshow(wc, interpolation="bilinear")
plt.axis("off")
plt.show()
| from __future__ import absolute_import, division, print_function, unicode_literals
from collections import Counter
import re
def create_wordcloud( data, stopwords = ["the", "a", "or", "tai", "and", "ja", "to", "on", "in", "of", "for", "is", "i", "this", "http", "www", "fi", "com"] ):
import types
if isinstance( data, types.GeneratorType ):
data = list( data )
if len(data) == 0:
print( "Dataset empty." )
return
from wordcloud import WordCloud
text = ''
for d in data:
text += d['text_content'].lower() + ' '
text = text.strip()
stopwords = map( lambda w: str(w), stopwords )
wc = WordCloud( background_color="white", width=800, height=400, stopwords = stopwords )
wc.generate( text )
plt.figure(figsize=(15,10))
plt.imshow(wc, interpolation="bilinear")
plt.axis("off")
plt.show()
Fix bug in wordcloud methodfrom __future__ import absolute_import, division, print_function, unicode_literals
from collections import Counter
import re
def create_wordcloud( data, plt, stopwords = ["the", "a", "or", "tai", "and", "ja", "to", "on", "in", "of", "for", "is", "i", "this", "http", "www", "fi", "com"] ):
import types
if isinstance( data, types.GeneratorType ):
data = list( data )
if len(data) == 0:
print( "Dataset empty." )
return
from wordcloud import WordCloud
text = ''
for d in data:
text += d['text_content'].lower() + ' '
text = text.strip()
stopwords = map( lambda w: str(w), stopwords )
wc = WordCloud( background_color="white", width=800, height=400, stopwords = stopwords )
wc.generate( text )
plt.figure(figsize=(15,10))
plt.imshow(wc, interpolation="bilinear")
plt.axis("off")
plt.show()
| <commit_before>from __future__ import absolute_import, division, print_function, unicode_literals
from collections import Counter
import re
def create_wordcloud( data, stopwords = ["the", "a", "or", "tai", "and", "ja", "to", "on", "in", "of", "for", "is", "i", "this", "http", "www", "fi", "com"] ):
import types
if isinstance( data, types.GeneratorType ):
data = list( data )
if len(data) == 0:
print( "Dataset empty." )
return
from wordcloud import WordCloud
text = ''
for d in data:
text += d['text_content'].lower() + ' '
text = text.strip()
stopwords = map( lambda w: str(w), stopwords )
wc = WordCloud( background_color="white", width=800, height=400, stopwords = stopwords )
wc.generate( text )
plt.figure(figsize=(15,10))
plt.imshow(wc, interpolation="bilinear")
plt.axis("off")
plt.show()
<commit_msg>Fix bug in wordcloud method<commit_after>from __future__ import absolute_import, division, print_function, unicode_literals
from collections import Counter
import re
def create_wordcloud( data, plt, stopwords = ["the", "a", "or", "tai", "and", "ja", "to", "on", "in", "of", "for", "is", "i", "this", "http", "www", "fi", "com"] ):
import types
if isinstance( data, types.GeneratorType ):
data = list( data )
if len(data) == 0:
print( "Dataset empty." )
return
from wordcloud import WordCloud
text = ''
for d in data:
text += d['text_content'].lower() + ' '
text = text.strip()
stopwords = map( lambda w: str(w), stopwords )
wc = WordCloud( background_color="white", width=800, height=400, stopwords = stopwords )
wc.generate( text )
plt.figure(figsize=(15,10))
plt.imshow(wc, interpolation="bilinear")
plt.axis("off")
plt.show()
|
0e2275c0d2623a7ec62e7109d3ffdd859118ed9d | external_tools/src/main/python/images/move_corrupt_images.py | external_tools/src/main/python/images/move_corrupt_images.py | """
Script to move corrupt images to 'dirty' directory
Reads list of images to move. Does not verify that images are corrupt -
Simply moves to 'dirty' directory of appropriate data-release creating
the required directory structure
"""
import os
import argparse
parser = argparse.ArgumentParser(
description="Move corrupt images to 'dirty' dir")
parser.add_argument('-i', dest='inputFiles', required=True,
help='File containing list of images to move'
)
parser.add_argument('-s', dest='splitString',
help='token to separate the basedir from input files'
)
parser.add_argument('-r', dest='replacementString',
help='String to replace the split string with'
)
parser.add_argument('-d', dest='destDirBase', required=True,
help='Path to the base of the destination dir'
)
args = parser.parse_args()
input_files = args.inputFiles
split_string = "" if args.splitString is None else args.splitString
replacement_string = "" if args.replacementString is None else args.replacementString
with open(input_files,'rt') as f:
fnames = [fname.strip('\n') for fname in f.readlines()]
for fname in fnames:
fname = fname.replace(' ','\ ')
fname2 = fname.replace(split_string, replacement_string)
if os.path.exists(fname2):
continue
out_dir = os.path.dirname(fname2)
if not os.path.isdir(out_dir):
os.makedirs(out_dir)
command = "mv " + fname + " " + fname2
print(command)
os.system(command)
| """
Script to move corrupt images to 'dirty' directory
Reads list of images to move. Does not verify that images are corrupt -
Simply moves to 'dirty' directory of appropriate data-release creating
the required directory structure
"""
import os
import argparse
parser = argparse.ArgumentParser(
description="Move corrupt images to 'dirty' dir")
parser.add_argument('-i', dest='inputFiles', required=True,
help='File containing list of images to move'
)
parser.add_argument('-s', dest='splitString',
help='token to separate the basedir from input files'
)
parser.add_argument('-r', dest='replacementString',
help='String to replace the split string with'
)
parser.add_argument('-d', dest='destDirBase', required=True,
help='Path to the base of the destination dir'
)
args = parser.parse_args()
input_files = args.inputFiles
split_string = "" if args.splitString is None else args.splitString
replacement_string = "" if args.replacementString is None else args.replacementString
with open(input_files,'rt') as f:
fnames = [fname.strip('\n') for fname in f.readlines()]
for fname in fnames:
fname2 = fname.replace(split_string, replacement_string)
os.rename(fname, fname2)
| Use os.rename instead of os.path in moving dirs | Use os.rename instead of os.path in moving dirs
| Python | apache-2.0 | mpi2/PhenotypeData,mpi2/PhenotypeData,mpi2/PhenotypeData,mpi2/PhenotypeData,mpi2/PhenotypeData,mpi2/PhenotypeData | """
Script to move corrupt images to 'dirty' directory
Reads list of images to move. Does not verify that images are corrupt -
Simply moves to 'dirty' directory of appropriate data-release creating
the required directory structure
"""
import os
import argparse
parser = argparse.ArgumentParser(
description="Move corrupt images to 'dirty' dir")
parser.add_argument('-i', dest='inputFiles', required=True,
help='File containing list of images to move'
)
parser.add_argument('-s', dest='splitString',
help='token to separate the basedir from input files'
)
parser.add_argument('-r', dest='replacementString',
help='String to replace the split string with'
)
parser.add_argument('-d', dest='destDirBase', required=True,
help='Path to the base of the destination dir'
)
args = parser.parse_args()
input_files = args.inputFiles
split_string = "" if args.splitString is None else args.splitString
replacement_string = "" if args.replacementString is None else args.replacementString
with open(input_files,'rt') as f:
fnames = [fname.strip('\n') for fname in f.readlines()]
for fname in fnames:
fname = fname.replace(' ','\ ')
fname2 = fname.replace(split_string, replacement_string)
if os.path.exists(fname2):
continue
out_dir = os.path.dirname(fname2)
if not os.path.isdir(out_dir):
os.makedirs(out_dir)
command = "mv " + fname + " " + fname2
print(command)
os.system(command)
Use os.rename instead of os.path in moving dirs | """
Script to move corrupt images to 'dirty' directory
Reads list of images to move. Does not verify that images are corrupt -
Simply moves to 'dirty' directory of appropriate data-release creating
the required directory structure
"""
import os
import argparse
parser = argparse.ArgumentParser(
description="Move corrupt images to 'dirty' dir")
parser.add_argument('-i', dest='inputFiles', required=True,
help='File containing list of images to move'
)
parser.add_argument('-s', dest='splitString',
help='token to separate the basedir from input files'
)
parser.add_argument('-r', dest='replacementString',
help='String to replace the split string with'
)
parser.add_argument('-d', dest='destDirBase', required=True,
help='Path to the base of the destination dir'
)
args = parser.parse_args()
input_files = args.inputFiles
split_string = "" if args.splitString is None else args.splitString
replacement_string = "" if args.replacementString is None else args.replacementString
with open(input_files,'rt') as f:
fnames = [fname.strip('\n') for fname in f.readlines()]
for fname in fnames:
fname2 = fname.replace(split_string, replacement_string)
os.rename(fname, fname2)
| <commit_before>"""
Script to move corrupt images to 'dirty' directory
Reads list of images to move. Does not verify that images are corrupt -
Simply moves to 'dirty' directory of appropriate data-release creating
the required directory structure
"""
import os
import argparse
parser = argparse.ArgumentParser(
description="Move corrupt images to 'dirty' dir")
parser.add_argument('-i', dest='inputFiles', required=True,
help='File containing list of images to move'
)
parser.add_argument('-s', dest='splitString',
help='token to separate the basedir from input files'
)
parser.add_argument('-r', dest='replacementString',
help='String to replace the split string with'
)
parser.add_argument('-d', dest='destDirBase', required=True,
help='Path to the base of the destination dir'
)
args = parser.parse_args()
input_files = args.inputFiles
split_string = "" if args.splitString is None else args.splitString
replacement_string = "" if args.replacementString is None else args.replacementString
with open(input_files,'rt') as f:
fnames = [fname.strip('\n') for fname in f.readlines()]
for fname in fnames:
fname = fname.replace(' ','\ ')
fname2 = fname.replace(split_string, replacement_string)
if os.path.exists(fname2):
continue
out_dir = os.path.dirname(fname2)
if not os.path.isdir(out_dir):
os.makedirs(out_dir)
command = "mv " + fname + " " + fname2
print(command)
os.system(command)
<commit_msg>Use os.rename instead of os.path in moving dirs<commit_after> | """
Script to move corrupt images to 'dirty' directory
Reads list of images to move. Does not verify that images are corrupt -
Simply moves to 'dirty' directory of appropriate data-release creating
the required directory structure
"""
import os
import argparse
parser = argparse.ArgumentParser(
description="Move corrupt images to 'dirty' dir")
parser.add_argument('-i', dest='inputFiles', required=True,
help='File containing list of images to move'
)
parser.add_argument('-s', dest='splitString',
help='token to separate the basedir from input files'
)
parser.add_argument('-r', dest='replacementString',
help='String to replace the split string with'
)
parser.add_argument('-d', dest='destDirBase', required=True,
help='Path to the base of the destination dir'
)
args = parser.parse_args()
input_files = args.inputFiles
split_string = "" if args.splitString is None else args.splitString
replacement_string = "" if args.replacementString is None else args.replacementString
with open(input_files,'rt') as f:
fnames = [fname.strip('\n') for fname in f.readlines()]
for fname in fnames:
fname2 = fname.replace(split_string, replacement_string)
os.rename(fname, fname2)
| """
Script to move corrupt images to 'dirty' directory
Reads list of images to move. Does not verify that images are corrupt -
Simply moves to 'dirty' directory of appropriate data-release creating
the required directory structure
"""
import os
import argparse
parser = argparse.ArgumentParser(
description="Move corrupt images to 'dirty' dir")
parser.add_argument('-i', dest='inputFiles', required=True,
help='File containing list of images to move'
)
parser.add_argument('-s', dest='splitString',
help='token to separate the basedir from input files'
)
parser.add_argument('-r', dest='replacementString',
help='String to replace the split string with'
)
parser.add_argument('-d', dest='destDirBase', required=True,
help='Path to the base of the destination dir'
)
args = parser.parse_args()
input_files = args.inputFiles
split_string = "" if args.splitString is None else args.splitString
replacement_string = "" if args.replacementString is None else args.replacementString
with open(input_files,'rt') as f:
fnames = [fname.strip('\n') for fname in f.readlines()]
for fname in fnames:
fname = fname.replace(' ','\ ')
fname2 = fname.replace(split_string, replacement_string)
if os.path.exists(fname2):
continue
out_dir = os.path.dirname(fname2)
if not os.path.isdir(out_dir):
os.makedirs(out_dir)
command = "mv " + fname + " " + fname2
print(command)
os.system(command)
Use os.rename instead of os.path in moving dirs"""
Script to move corrupt images to 'dirty' directory
Reads list of images to move. Does not verify that images are corrupt -
Simply moves to 'dirty' directory of appropriate data-release creating
the required directory structure
"""
import os
import argparse
parser = argparse.ArgumentParser(
description="Move corrupt images to 'dirty' dir")
parser.add_argument('-i', dest='inputFiles', required=True,
help='File containing list of images to move'
)
parser.add_argument('-s', dest='splitString',
help='token to separate the basedir from input files'
)
parser.add_argument('-r', dest='replacementString',
help='String to replace the split string with'
)
parser.add_argument('-d', dest='destDirBase', required=True,
help='Path to the base of the destination dir'
)
args = parser.parse_args()
input_files = args.inputFiles
split_string = "" if args.splitString is None else args.splitString
replacement_string = "" if args.replacementString is None else args.replacementString
with open(input_files,'rt') as f:
fnames = [fname.strip('\n') for fname in f.readlines()]
for fname in fnames:
fname2 = fname.replace(split_string, replacement_string)
os.rename(fname, fname2)
| <commit_before>"""
Script to move corrupt images to 'dirty' directory
Reads list of images to move. Does not verify that images are corrupt -
Simply moves to 'dirty' directory of appropriate data-release creating
the required directory structure
"""
import os
import argparse
parser = argparse.ArgumentParser(
description="Move corrupt images to 'dirty' dir")
parser.add_argument('-i', dest='inputFiles', required=True,
help='File containing list of images to move'
)
parser.add_argument('-s', dest='splitString',
help='token to separate the basedir from input files'
)
parser.add_argument('-r', dest='replacementString',
help='String to replace the split string with'
)
parser.add_argument('-d', dest='destDirBase', required=True,
help='Path to the base of the destination dir'
)
args = parser.parse_args()
input_files = args.inputFiles
split_string = "" if args.splitString is None else args.splitString
replacement_string = "" if args.replacementString is None else args.replacementString
with open(input_files,'rt') as f:
fnames = [fname.strip('\n') for fname in f.readlines()]
for fname in fnames:
fname = fname.replace(' ','\ ')
fname2 = fname.replace(split_string, replacement_string)
if os.path.exists(fname2):
continue
out_dir = os.path.dirname(fname2)
if not os.path.isdir(out_dir):
os.makedirs(out_dir)
command = "mv " + fname + " " + fname2
print(command)
os.system(command)
<commit_msg>Use os.rename instead of os.path in moving dirs<commit_after>"""
Script to move corrupt images to 'dirty' directory
Reads list of images to move. Does not verify that images are corrupt -
Simply moves to 'dirty' directory of appropriate data-release creating
the required directory structure
"""
import os
import argparse
parser = argparse.ArgumentParser(
description="Move corrupt images to 'dirty' dir")
parser.add_argument('-i', dest='inputFiles', required=True,
help='File containing list of images to move'
)
parser.add_argument('-s', dest='splitString',
help='token to separate the basedir from input files'
)
parser.add_argument('-r', dest='replacementString',
help='String to replace the split string with'
)
parser.add_argument('-d', dest='destDirBase', required=True,
help='Path to the base of the destination dir'
)
args = parser.parse_args()
input_files = args.inputFiles
split_string = "" if args.splitString is None else args.splitString
replacement_string = "" if args.replacementString is None else args.replacementString
with open(input_files,'rt') as f:
fnames = [fname.strip('\n') for fname in f.readlines()]
for fname in fnames:
fname2 = fname.replace(split_string, replacement_string)
os.rename(fname, fname2)
|
4b193d9f0c46f91c5a58446e6443d8779e7ca5ce | server/plugins/ardinfo/scripts/ard_info.py | server/plugins/ardinfo/scripts/ard_info.py | #!/usr/bin/python
import os
import sys
sys.path.append("/usr/local/munki/munkilib")
import FoundationPlist
sys.path.append("/usr/local/sal")
import utils
def main():
ard_path = "/Library/Preferences/com.apple.RemoteDesktop.plist"
if os.path.exists(ard_path):
ard_prefs = FoundationPlist.readPlist(ard_path)
else:
ard_prefs = {}
sal_result_key = "ARD_Info_{}"
prefs_key_prefix = "Text{}"
data = {
sal_result_key.format(i): ard_prefs.get(prefs_key_prefix.format(i), "")
for i in range(1, 5)}
utils.add_plugin_results('ARD_Info', data)
if __name__ == "__main__":
main()
| #!/usr/bin/python
import os
import sys
sys.path.append("/usr/local/munki")
from munkilib import FoundationPlist
sys.path.append("/usr/local/sal")
import utils
def main():
ard_path = "/Library/Preferences/com.apple.RemoteDesktop.plist"
if os.path.exists(ard_path):
ard_prefs = FoundationPlist.readPlist(ard_path)
else:
ard_prefs = {}
sal_result_key = "ARD_Info_{}"
prefs_key_prefix = "Text{}"
data = {
sal_result_key.format(i): ard_prefs.get(prefs_key_prefix.format(i), "")
for i in range(1, 5)}
utils.add_plugin_results('ARD_Info', data)
if __name__ == "__main__":
main()
| Fix name clash over "utils" in ardinfo plugin script. | Fix name clash over "utils" in ardinfo plugin script.
| Python | apache-2.0 | sheagcraig/sal,sheagcraig/sal,salopensource/sal,sheagcraig/sal,salopensource/sal,salopensource/sal,sheagcraig/sal,salopensource/sal | #!/usr/bin/python
import os
import sys
sys.path.append("/usr/local/munki/munkilib")
import FoundationPlist
sys.path.append("/usr/local/sal")
import utils
def main():
ard_path = "/Library/Preferences/com.apple.RemoteDesktop.plist"
if os.path.exists(ard_path):
ard_prefs = FoundationPlist.readPlist(ard_path)
else:
ard_prefs = {}
sal_result_key = "ARD_Info_{}"
prefs_key_prefix = "Text{}"
data = {
sal_result_key.format(i): ard_prefs.get(prefs_key_prefix.format(i), "")
for i in range(1, 5)}
utils.add_plugin_results('ARD_Info', data)
if __name__ == "__main__":
main()
Fix name clash over "utils" in ardinfo plugin script. | #!/usr/bin/python
import os
import sys
sys.path.append("/usr/local/munki")
from munkilib import FoundationPlist
sys.path.append("/usr/local/sal")
import utils
def main():
ard_path = "/Library/Preferences/com.apple.RemoteDesktop.plist"
if os.path.exists(ard_path):
ard_prefs = FoundationPlist.readPlist(ard_path)
else:
ard_prefs = {}
sal_result_key = "ARD_Info_{}"
prefs_key_prefix = "Text{}"
data = {
sal_result_key.format(i): ard_prefs.get(prefs_key_prefix.format(i), "")
for i in range(1, 5)}
utils.add_plugin_results('ARD_Info', data)
if __name__ == "__main__":
main()
| <commit_before>#!/usr/bin/python
import os
import sys
sys.path.append("/usr/local/munki/munkilib")
import FoundationPlist
sys.path.append("/usr/local/sal")
import utils
def main():
ard_path = "/Library/Preferences/com.apple.RemoteDesktop.plist"
if os.path.exists(ard_path):
ard_prefs = FoundationPlist.readPlist(ard_path)
else:
ard_prefs = {}
sal_result_key = "ARD_Info_{}"
prefs_key_prefix = "Text{}"
data = {
sal_result_key.format(i): ard_prefs.get(prefs_key_prefix.format(i), "")
for i in range(1, 5)}
utils.add_plugin_results('ARD_Info', data)
if __name__ == "__main__":
main()
<commit_msg>Fix name clash over "utils" in ardinfo plugin script.<commit_after> | #!/usr/bin/python
import os
import sys
sys.path.append("/usr/local/munki")
from munkilib import FoundationPlist
sys.path.append("/usr/local/sal")
import utils
def main():
ard_path = "/Library/Preferences/com.apple.RemoteDesktop.plist"
if os.path.exists(ard_path):
ard_prefs = FoundationPlist.readPlist(ard_path)
else:
ard_prefs = {}
sal_result_key = "ARD_Info_{}"
prefs_key_prefix = "Text{}"
data = {
sal_result_key.format(i): ard_prefs.get(prefs_key_prefix.format(i), "")
for i in range(1, 5)}
utils.add_plugin_results('ARD_Info', data)
if __name__ == "__main__":
main()
| #!/usr/bin/python
import os
import sys
sys.path.append("/usr/local/munki/munkilib")
import FoundationPlist
sys.path.append("/usr/local/sal")
import utils
def main():
ard_path = "/Library/Preferences/com.apple.RemoteDesktop.plist"
if os.path.exists(ard_path):
ard_prefs = FoundationPlist.readPlist(ard_path)
else:
ard_prefs = {}
sal_result_key = "ARD_Info_{}"
prefs_key_prefix = "Text{}"
data = {
sal_result_key.format(i): ard_prefs.get(prefs_key_prefix.format(i), "")
for i in range(1, 5)}
utils.add_plugin_results('ARD_Info', data)
if __name__ == "__main__":
main()
Fix name clash over "utils" in ardinfo plugin script.#!/usr/bin/python
import os
import sys
sys.path.append("/usr/local/munki")
from munkilib import FoundationPlist
sys.path.append("/usr/local/sal")
import utils
def main():
ard_path = "/Library/Preferences/com.apple.RemoteDesktop.plist"
if os.path.exists(ard_path):
ard_prefs = FoundationPlist.readPlist(ard_path)
else:
ard_prefs = {}
sal_result_key = "ARD_Info_{}"
prefs_key_prefix = "Text{}"
data = {
sal_result_key.format(i): ard_prefs.get(prefs_key_prefix.format(i), "")
for i in range(1, 5)}
utils.add_plugin_results('ARD_Info', data)
if __name__ == "__main__":
main()
| <commit_before>#!/usr/bin/python
import os
import sys
sys.path.append("/usr/local/munki/munkilib")
import FoundationPlist
sys.path.append("/usr/local/sal")
import utils
def main():
ard_path = "/Library/Preferences/com.apple.RemoteDesktop.plist"
if os.path.exists(ard_path):
ard_prefs = FoundationPlist.readPlist(ard_path)
else:
ard_prefs = {}
sal_result_key = "ARD_Info_{}"
prefs_key_prefix = "Text{}"
data = {
sal_result_key.format(i): ard_prefs.get(prefs_key_prefix.format(i), "")
for i in range(1, 5)}
utils.add_plugin_results('ARD_Info', data)
if __name__ == "__main__":
main()
<commit_msg>Fix name clash over "utils" in ardinfo plugin script.<commit_after>#!/usr/bin/python
import os
import sys
sys.path.append("/usr/local/munki")
from munkilib import FoundationPlist
sys.path.append("/usr/local/sal")
import utils
def main():
ard_path = "/Library/Preferences/com.apple.RemoteDesktop.plist"
if os.path.exists(ard_path):
ard_prefs = FoundationPlist.readPlist(ard_path)
else:
ard_prefs = {}
sal_result_key = "ARD_Info_{}"
prefs_key_prefix = "Text{}"
data = {
sal_result_key.format(i): ard_prefs.get(prefs_key_prefix.format(i), "")
for i in range(1, 5)}
utils.add_plugin_results('ARD_Info', data)
if __name__ == "__main__":
main()
|
b920b818f6b4a83ce39a34f4f1b3afe6f8002906 | integration-test/400-bay-water.py | integration-test/400-bay-water.py | # San Pablo Bay
# https://www.openstreetmap.org/way/43950409
assert_has_feature(
14, 2623, 6318, 'water',
{ 'kind': 'bay', 'label_placement': True })
# Sansum Narrows
# https://www.openstreetmap.org/relation/1019862
assert_has_feature(
11, 321, 705, 'water',
{ 'kind': 'strait', 'label_placement': True })
# Horsens Fjord
# https://www.openstreetmap.org/relation/1451065
assert_has_feature(
14, 8645, 5114, 'water',
{ 'kind': 'fjord', 'label_placement': True })
| # San Pablo Bay
# https://www.openstreetmap.org/way/43950409
assert_has_feature(
14, 2623, 6318, 'water',
{ 'kind': 'bay', 'label_placement': True })
# Sansum Narrows
# https://www.openstreetmap.org/relation/1019862
assert_has_feature(
11, 321, 705, 'water',
{ 'kind': 'strait', 'label_placement': True })
# Horsens Fjord
# https://www.openstreetmap.org/relation/1451065
assert_has_feature(
14, 8646, 5112, 'water',
{ 'kind': 'fjord', 'label_placement': True })
| Update tile coordinate for fjord label placement | Update tile coordinate for fjord label placement
| Python | mit | mapzen/vector-datasource,mapzen/vector-datasource,mapzen/vector-datasource | # San Pablo Bay
# https://www.openstreetmap.org/way/43950409
assert_has_feature(
14, 2623, 6318, 'water',
{ 'kind': 'bay', 'label_placement': True })
# Sansum Narrows
# https://www.openstreetmap.org/relation/1019862
assert_has_feature(
11, 321, 705, 'water',
{ 'kind': 'strait', 'label_placement': True })
# Horsens Fjord
# https://www.openstreetmap.org/relation/1451065
assert_has_feature(
14, 8645, 5114, 'water',
{ 'kind': 'fjord', 'label_placement': True })
Update tile coordinate for fjord label placement | # San Pablo Bay
# https://www.openstreetmap.org/way/43950409
assert_has_feature(
14, 2623, 6318, 'water',
{ 'kind': 'bay', 'label_placement': True })
# Sansum Narrows
# https://www.openstreetmap.org/relation/1019862
assert_has_feature(
11, 321, 705, 'water',
{ 'kind': 'strait', 'label_placement': True })
# Horsens Fjord
# https://www.openstreetmap.org/relation/1451065
assert_has_feature(
14, 8646, 5112, 'water',
{ 'kind': 'fjord', 'label_placement': True })
| <commit_before># San Pablo Bay
# https://www.openstreetmap.org/way/43950409
assert_has_feature(
14, 2623, 6318, 'water',
{ 'kind': 'bay', 'label_placement': True })
# Sansum Narrows
# https://www.openstreetmap.org/relation/1019862
assert_has_feature(
11, 321, 705, 'water',
{ 'kind': 'strait', 'label_placement': True })
# Horsens Fjord
# https://www.openstreetmap.org/relation/1451065
assert_has_feature(
14, 8645, 5114, 'water',
{ 'kind': 'fjord', 'label_placement': True })
<commit_msg>Update tile coordinate for fjord label placement<commit_after> | # San Pablo Bay
# https://www.openstreetmap.org/way/43950409
assert_has_feature(
14, 2623, 6318, 'water',
{ 'kind': 'bay', 'label_placement': True })
# Sansum Narrows
# https://www.openstreetmap.org/relation/1019862
assert_has_feature(
11, 321, 705, 'water',
{ 'kind': 'strait', 'label_placement': True })
# Horsens Fjord
# https://www.openstreetmap.org/relation/1451065
assert_has_feature(
14, 8646, 5112, 'water',
{ 'kind': 'fjord', 'label_placement': True })
| # San Pablo Bay
# https://www.openstreetmap.org/way/43950409
assert_has_feature(
14, 2623, 6318, 'water',
{ 'kind': 'bay', 'label_placement': True })
# Sansum Narrows
# https://www.openstreetmap.org/relation/1019862
assert_has_feature(
11, 321, 705, 'water',
{ 'kind': 'strait', 'label_placement': True })
# Horsens Fjord
# https://www.openstreetmap.org/relation/1451065
assert_has_feature(
14, 8645, 5114, 'water',
{ 'kind': 'fjord', 'label_placement': True })
Update tile coordinate for fjord label placement# San Pablo Bay
# https://www.openstreetmap.org/way/43950409
assert_has_feature(
14, 2623, 6318, 'water',
{ 'kind': 'bay', 'label_placement': True })
# Sansum Narrows
# https://www.openstreetmap.org/relation/1019862
assert_has_feature(
11, 321, 705, 'water',
{ 'kind': 'strait', 'label_placement': True })
# Horsens Fjord
# https://www.openstreetmap.org/relation/1451065
assert_has_feature(
14, 8646, 5112, 'water',
{ 'kind': 'fjord', 'label_placement': True })
| <commit_before># San Pablo Bay
# https://www.openstreetmap.org/way/43950409
assert_has_feature(
14, 2623, 6318, 'water',
{ 'kind': 'bay', 'label_placement': True })
# Sansum Narrows
# https://www.openstreetmap.org/relation/1019862
assert_has_feature(
11, 321, 705, 'water',
{ 'kind': 'strait', 'label_placement': True })
# Horsens Fjord
# https://www.openstreetmap.org/relation/1451065
assert_has_feature(
14, 8645, 5114, 'water',
{ 'kind': 'fjord', 'label_placement': True })
<commit_msg>Update tile coordinate for fjord label placement<commit_after># San Pablo Bay
# https://www.openstreetmap.org/way/43950409
assert_has_feature(
14, 2623, 6318, 'water',
{ 'kind': 'bay', 'label_placement': True })
# Sansum Narrows
# https://www.openstreetmap.org/relation/1019862
assert_has_feature(
11, 321, 705, 'water',
{ 'kind': 'strait', 'label_placement': True })
# Horsens Fjord
# https://www.openstreetmap.org/relation/1451065
assert_has_feature(
14, 8646, 5112, 'water',
{ 'kind': 'fjord', 'label_placement': True })
|
5bbd288c40e3a2bc1ee791545d704452699334f3 | cr8/aio.py | cr8/aio.py |
from tqdm import tqdm
import asyncio
try:
import uvloop
asyncio.set_event_loop_policy(uvloop.EventLoopPolicy())
except ImportError:
pass
async def map_async(q, corof, iterable):
for i in iterable:
task = asyncio.ensure_future(corof(*i))
await q.put(task)
await q.join()
await q.put(None)
async def consume(q):
with tqdm(unit=' requests') as t:
while True:
task = await q.get()
if task is None:
break
await task
t.update(1)
q.task_done()
def run(coro, iterable, concurrency, loop=None):
loop = loop or asyncio.get_event_loop()
q = asyncio.Queue(maxsize=concurrency)
loop.run_until_complete(asyncio.gather(
map_async(q, coro, iterable),
consume(q)))
|
from tqdm import tqdm
import asyncio
try:
import uvloop
asyncio.set_event_loop_policy(uvloop.EventLoopPolicy())
except ImportError:
pass
async def map_async(q, corof, iterable):
for i in iterable:
task = asyncio.ensure_future(corof(*i))
await q.put(task)
await q.put(None)
async def consume(q):
with tqdm(unit=' requests') as t:
while True:
task = await q.get()
if task is None:
break
await task
t.update(1)
def run(coro, iterable, concurrency, loop=None):
loop = loop or asyncio.get_event_loop()
q = asyncio.Queue(maxsize=concurrency)
loop.run_until_complete(asyncio.gather(
map_async(q, coro, iterable),
consume(q)))
| Remove q.join() / task_done() usage | Remove q.join() / task_done() usage
Don't have to block producer anymore - it will wait for the consumer to
finish anyway
| Python | mit | mikethebeer/cr8,mfussenegger/cr8 |
from tqdm import tqdm
import asyncio
try:
import uvloop
asyncio.set_event_loop_policy(uvloop.EventLoopPolicy())
except ImportError:
pass
async def map_async(q, corof, iterable):
for i in iterable:
task = asyncio.ensure_future(corof(*i))
await q.put(task)
await q.join()
await q.put(None)
async def consume(q):
with tqdm(unit=' requests') as t:
while True:
task = await q.get()
if task is None:
break
await task
t.update(1)
q.task_done()
def run(coro, iterable, concurrency, loop=None):
loop = loop or asyncio.get_event_loop()
q = asyncio.Queue(maxsize=concurrency)
loop.run_until_complete(asyncio.gather(
map_async(q, coro, iterable),
consume(q)))
Remove q.join() / task_done() usage
Don't have to block producer anymore - it will wait for the consumer to
finish anyway |
from tqdm import tqdm
import asyncio
try:
import uvloop
asyncio.set_event_loop_policy(uvloop.EventLoopPolicy())
except ImportError:
pass
async def map_async(q, corof, iterable):
for i in iterable:
task = asyncio.ensure_future(corof(*i))
await q.put(task)
await q.put(None)
async def consume(q):
with tqdm(unit=' requests') as t:
while True:
task = await q.get()
if task is None:
break
await task
t.update(1)
def run(coro, iterable, concurrency, loop=None):
loop = loop or asyncio.get_event_loop()
q = asyncio.Queue(maxsize=concurrency)
loop.run_until_complete(asyncio.gather(
map_async(q, coro, iterable),
consume(q)))
| <commit_before>
from tqdm import tqdm
import asyncio
try:
import uvloop
asyncio.set_event_loop_policy(uvloop.EventLoopPolicy())
except ImportError:
pass
async def map_async(q, corof, iterable):
for i in iterable:
task = asyncio.ensure_future(corof(*i))
await q.put(task)
await q.join()
await q.put(None)
async def consume(q):
with tqdm(unit=' requests') as t:
while True:
task = await q.get()
if task is None:
break
await task
t.update(1)
q.task_done()
def run(coro, iterable, concurrency, loop=None):
loop = loop or asyncio.get_event_loop()
q = asyncio.Queue(maxsize=concurrency)
loop.run_until_complete(asyncio.gather(
map_async(q, coro, iterable),
consume(q)))
<commit_msg>Remove q.join() / task_done() usage
Don't have to block producer anymore - it will wait for the consumer to
finish anyway<commit_after> |
from tqdm import tqdm
import asyncio
try:
import uvloop
asyncio.set_event_loop_policy(uvloop.EventLoopPolicy())
except ImportError:
pass
async def map_async(q, corof, iterable):
for i in iterable:
task = asyncio.ensure_future(corof(*i))
await q.put(task)
await q.put(None)
async def consume(q):
with tqdm(unit=' requests') as t:
while True:
task = await q.get()
if task is None:
break
await task
t.update(1)
def run(coro, iterable, concurrency, loop=None):
loop = loop or asyncio.get_event_loop()
q = asyncio.Queue(maxsize=concurrency)
loop.run_until_complete(asyncio.gather(
map_async(q, coro, iterable),
consume(q)))
|
from tqdm import tqdm
import asyncio
try:
import uvloop
asyncio.set_event_loop_policy(uvloop.EventLoopPolicy())
except ImportError:
pass
async def map_async(q, corof, iterable):
for i in iterable:
task = asyncio.ensure_future(corof(*i))
await q.put(task)
await q.join()
await q.put(None)
async def consume(q):
with tqdm(unit=' requests') as t:
while True:
task = await q.get()
if task is None:
break
await task
t.update(1)
q.task_done()
def run(coro, iterable, concurrency, loop=None):
loop = loop or asyncio.get_event_loop()
q = asyncio.Queue(maxsize=concurrency)
loop.run_until_complete(asyncio.gather(
map_async(q, coro, iterable),
consume(q)))
Remove q.join() / task_done() usage
Don't have to block producer anymore - it will wait for the consumer to
finish anyway
from tqdm import tqdm
import asyncio
try:
import uvloop
asyncio.set_event_loop_policy(uvloop.EventLoopPolicy())
except ImportError:
pass
async def map_async(q, corof, iterable):
for i in iterable:
task = asyncio.ensure_future(corof(*i))
await q.put(task)
await q.put(None)
async def consume(q):
with tqdm(unit=' requests') as t:
while True:
task = await q.get()
if task is None:
break
await task
t.update(1)
def run(coro, iterable, concurrency, loop=None):
loop = loop or asyncio.get_event_loop()
q = asyncio.Queue(maxsize=concurrency)
loop.run_until_complete(asyncio.gather(
map_async(q, coro, iterable),
consume(q)))
| <commit_before>
from tqdm import tqdm
import asyncio
try:
import uvloop
asyncio.set_event_loop_policy(uvloop.EventLoopPolicy())
except ImportError:
pass
async def map_async(q, corof, iterable):
for i in iterable:
task = asyncio.ensure_future(corof(*i))
await q.put(task)
await q.join()
await q.put(None)
async def consume(q):
with tqdm(unit=' requests') as t:
while True:
task = await q.get()
if task is None:
break
await task
t.update(1)
q.task_done()
def run(coro, iterable, concurrency, loop=None):
loop = loop or asyncio.get_event_loop()
q = asyncio.Queue(maxsize=concurrency)
loop.run_until_complete(asyncio.gather(
map_async(q, coro, iterable),
consume(q)))
<commit_msg>Remove q.join() / task_done() usage
Don't have to block producer anymore - it will wait for the consumer to
finish anyway<commit_after>
from tqdm import tqdm
import asyncio
try:
import uvloop
asyncio.set_event_loop_policy(uvloop.EventLoopPolicy())
except ImportError:
pass
async def map_async(q, corof, iterable):
for i in iterable:
task = asyncio.ensure_future(corof(*i))
await q.put(task)
await q.put(None)
async def consume(q):
with tqdm(unit=' requests') as t:
while True:
task = await q.get()
if task is None:
break
await task
t.update(1)
def run(coro, iterable, concurrency, loop=None):
loop = loop or asyncio.get_event_loop()
q = asyncio.Queue(maxsize=concurrency)
loop.run_until_complete(asyncio.gather(
map_async(q, coro, iterable),
consume(q)))
|
7698ec18abd25ed41b3104a382e7d8ca38d755ca | tests/unit/test_describe.py | tests/unit/test_describe.py | import pytest
from mock import Mock
from formica import cli
from tests.unit.constants import STACK
def test_describes_change_set(boto_client, change_set):
cli.main(['describe', '--stack', STACK])
boto_client.assert_called_with('cloudformation')
change_set.assert_called_with(stack=STACK)
change_set.return_value.describe.assert_called_once()
| import pytest
from mock import Mock
from formica import cli
from tests.unit.constants import STACK
def test_describes_change_set(boto_client, change_set):
cli.main(['describe', '--stack', STACK])
change_set.assert_called_with(stack=STACK)
change_set.return_value.describe.assert_called_once()
| Remove Assert Call not necessary anymore | Remove Assert Call not necessary anymore
| Python | mit | flomotlik/formica | import pytest
from mock import Mock
from formica import cli
from tests.unit.constants import STACK
def test_describes_change_set(boto_client, change_set):
cli.main(['describe', '--stack', STACK])
boto_client.assert_called_with('cloudformation')
change_set.assert_called_with(stack=STACK)
change_set.return_value.describe.assert_called_once()
Remove Assert Call not necessary anymore | import pytest
from mock import Mock
from formica import cli
from tests.unit.constants import STACK
def test_describes_change_set(boto_client, change_set):
cli.main(['describe', '--stack', STACK])
change_set.assert_called_with(stack=STACK)
change_set.return_value.describe.assert_called_once()
| <commit_before>import pytest
from mock import Mock
from formica import cli
from tests.unit.constants import STACK
def test_describes_change_set(boto_client, change_set):
cli.main(['describe', '--stack', STACK])
boto_client.assert_called_with('cloudformation')
change_set.assert_called_with(stack=STACK)
change_set.return_value.describe.assert_called_once()
<commit_msg>Remove Assert Call not necessary anymore<commit_after> | import pytest
from mock import Mock
from formica import cli
from tests.unit.constants import STACK
def test_describes_change_set(boto_client, change_set):
cli.main(['describe', '--stack', STACK])
change_set.assert_called_with(stack=STACK)
change_set.return_value.describe.assert_called_once()
| import pytest
from mock import Mock
from formica import cli
from tests.unit.constants import STACK
def test_describes_change_set(boto_client, change_set):
cli.main(['describe', '--stack', STACK])
boto_client.assert_called_with('cloudformation')
change_set.assert_called_with(stack=STACK)
change_set.return_value.describe.assert_called_once()
Remove Assert Call not necessary anymoreimport pytest
from mock import Mock
from formica import cli
from tests.unit.constants import STACK
def test_describes_change_set(boto_client, change_set):
cli.main(['describe', '--stack', STACK])
change_set.assert_called_with(stack=STACK)
change_set.return_value.describe.assert_called_once()
| <commit_before>import pytest
from mock import Mock
from formica import cli
from tests.unit.constants import STACK
def test_describes_change_set(boto_client, change_set):
cli.main(['describe', '--stack', STACK])
boto_client.assert_called_with('cloudformation')
change_set.assert_called_with(stack=STACK)
change_set.return_value.describe.assert_called_once()
<commit_msg>Remove Assert Call not necessary anymore<commit_after>import pytest
from mock import Mock
from formica import cli
from tests.unit.constants import STACK
def test_describes_change_set(boto_client, change_set):
cli.main(['describe', '--stack', STACK])
change_set.assert_called_with(stack=STACK)
change_set.return_value.describe.assert_called_once()
|
41b45872fae69e6c791aa79332981f12e33f7075 | numpy/distutils/fcompiler/nag.py | numpy/distutils/fcompiler/nag.py | import os
import sys
from numpy.distutils.cpuinfo import cpu
from numpy.distutils.fcompiler import FCompiler
class NAGFCompiler(FCompiler):
compiler_type = 'nag'
version_pattern = r'NAGWare Fortran 95 compiler Release (?P<version>[^\s]*)'
executables = {
'version_cmd' : ["f95", "-V"],
'compiler_f77' : ["f95", "-fixed"],
'compiler_fix' : ["f95", "-fixed"],
'compiler_f90' : ["f95"],
'linker_so' : ["f95"],
'archiver' : ["ar", "-cr"],
'ranlib' : ["ranlib"]
}
def get_flags_linker_so(self):
if sys.platform=='darwin':
return ['-unsharedf95','-Wl,-bundle,-flat_namespace,-undefined,suppress']
return ["-Wl,shared"]
def get_flags_opt(self):
return ['-O4']
def get_flags_arch(self):
return ['-target=native']
def get_flags_debug(self):
return ['-g','-gline','-g90','-nan','-C']
if __name__ == '__main__':
from distutils import log
log.set_verbosity(2)
from numpy.distutils.fcompiler import new_fcompiler
compiler = new_fcompiler(compiler='nag')
compiler.customize()
print compiler.get_version()
| import os
import sys
from numpy.distutils.cpuinfo import cpu
from numpy.distutils.fcompiler import FCompiler
class NAGFCompiler(FCompiler):
compiler_type = 'nag'
version_pattern = r'NAGWare Fortran 95 compiler Release (?P<version>[^\s]*)'
executables = {
'version_cmd' : ["f95", "-V"],
'compiler_f77' : ["f95", "-fixed"],
'compiler_fix' : ["f95", "-fixed"],
'compiler_f90' : ["f95"],
'linker_so' : ["f95"],
'archiver' : ["ar", "-cr"],
'ranlib' : ["ranlib"]
}
def get_flags_linker_so(self):
if sys.platform=='darwin':
return ['-unsharedf95','-Wl,-bundle,-flat_namespace,-undefined,suppress']
return ["-Wl,-shared"]
def get_flags_opt(self):
return ['-O4']
def get_flags_arch(self):
return ['-target=native']
def get_flags_debug(self):
return ['-g','-gline','-g90','-nan','-C']
if __name__ == '__main__':
from distutils import log
log.set_verbosity(2)
from numpy.distutils.fcompiler import new_fcompiler
compiler = new_fcompiler(compiler='nag')
compiler.customize()
print compiler.get_version()
| Fix for using NAG Fortran 95, due to James Graham <jg307@cam.ac.uk> | Fix for using NAG Fortran 95, due to James Graham <jg307@cam.ac.uk>
git-svn-id: 77a43f9646713b91fea7788fad5dfbf67e151ece@2515 94b884b6-d6fd-0310-90d3-974f1d3f35e1
| Python | bsd-3-clause | jasonmccampbell/numpy-refactor-sprint,teoliphant/numpy-refactor,teoliphant/numpy-refactor,efiring/numpy-work,Ademan/NumPy-GSoC,illume/numpy3k,illume/numpy3k,chadnetzer/numpy-gaurdro,illume/numpy3k,chadnetzer/numpy-gaurdro,teoliphant/numpy-refactor,jasonmccampbell/numpy-refactor-sprint,efiring/numpy-work,chadnetzer/numpy-gaurdro,teoliphant/numpy-refactor,Ademan/NumPy-GSoC,jasonmccampbell/numpy-refactor-sprint,efiring/numpy-work,teoliphant/numpy-refactor,Ademan/NumPy-GSoC,chadnetzer/numpy-gaurdro,efiring/numpy-work,illume/numpy3k,Ademan/NumPy-GSoC,jasonmccampbell/numpy-refactor-sprint | import os
import sys
from numpy.distutils.cpuinfo import cpu
from numpy.distutils.fcompiler import FCompiler
class NAGFCompiler(FCompiler):
compiler_type = 'nag'
version_pattern = r'NAGWare Fortran 95 compiler Release (?P<version>[^\s]*)'
executables = {
'version_cmd' : ["f95", "-V"],
'compiler_f77' : ["f95", "-fixed"],
'compiler_fix' : ["f95", "-fixed"],
'compiler_f90' : ["f95"],
'linker_so' : ["f95"],
'archiver' : ["ar", "-cr"],
'ranlib' : ["ranlib"]
}
def get_flags_linker_so(self):
if sys.platform=='darwin':
return ['-unsharedf95','-Wl,-bundle,-flat_namespace,-undefined,suppress']
return ["-Wl,shared"]
def get_flags_opt(self):
return ['-O4']
def get_flags_arch(self):
return ['-target=native']
def get_flags_debug(self):
return ['-g','-gline','-g90','-nan','-C']
if __name__ == '__main__':
from distutils import log
log.set_verbosity(2)
from numpy.distutils.fcompiler import new_fcompiler
compiler = new_fcompiler(compiler='nag')
compiler.customize()
print compiler.get_version()
Fix for using NAG Fortran 95, due to James Graham <jg307@cam.ac.uk>
git-svn-id: 77a43f9646713b91fea7788fad5dfbf67e151ece@2515 94b884b6-d6fd-0310-90d3-974f1d3f35e1 | import os
import sys
from numpy.distutils.cpuinfo import cpu
from numpy.distutils.fcompiler import FCompiler
class NAGFCompiler(FCompiler):
compiler_type = 'nag'
version_pattern = r'NAGWare Fortran 95 compiler Release (?P<version>[^\s]*)'
executables = {
'version_cmd' : ["f95", "-V"],
'compiler_f77' : ["f95", "-fixed"],
'compiler_fix' : ["f95", "-fixed"],
'compiler_f90' : ["f95"],
'linker_so' : ["f95"],
'archiver' : ["ar", "-cr"],
'ranlib' : ["ranlib"]
}
def get_flags_linker_so(self):
if sys.platform=='darwin':
return ['-unsharedf95','-Wl,-bundle,-flat_namespace,-undefined,suppress']
return ["-Wl,-shared"]
def get_flags_opt(self):
return ['-O4']
def get_flags_arch(self):
return ['-target=native']
def get_flags_debug(self):
return ['-g','-gline','-g90','-nan','-C']
if __name__ == '__main__':
from distutils import log
log.set_verbosity(2)
from numpy.distutils.fcompiler import new_fcompiler
compiler = new_fcompiler(compiler='nag')
compiler.customize()
print compiler.get_version()
| <commit_before>import os
import sys
from numpy.distutils.cpuinfo import cpu
from numpy.distutils.fcompiler import FCompiler
class NAGFCompiler(FCompiler):
compiler_type = 'nag'
version_pattern = r'NAGWare Fortran 95 compiler Release (?P<version>[^\s]*)'
executables = {
'version_cmd' : ["f95", "-V"],
'compiler_f77' : ["f95", "-fixed"],
'compiler_fix' : ["f95", "-fixed"],
'compiler_f90' : ["f95"],
'linker_so' : ["f95"],
'archiver' : ["ar", "-cr"],
'ranlib' : ["ranlib"]
}
def get_flags_linker_so(self):
if sys.platform=='darwin':
return ['-unsharedf95','-Wl,-bundle,-flat_namespace,-undefined,suppress']
return ["-Wl,shared"]
def get_flags_opt(self):
return ['-O4']
def get_flags_arch(self):
return ['-target=native']
def get_flags_debug(self):
return ['-g','-gline','-g90','-nan','-C']
if __name__ == '__main__':
from distutils import log
log.set_verbosity(2)
from numpy.distutils.fcompiler import new_fcompiler
compiler = new_fcompiler(compiler='nag')
compiler.customize()
print compiler.get_version()
<commit_msg>Fix for using NAG Fortran 95, due to James Graham <jg307@cam.ac.uk>
git-svn-id: 77a43f9646713b91fea7788fad5dfbf67e151ece@2515 94b884b6-d6fd-0310-90d3-974f1d3f35e1<commit_after> | import os
import sys
from numpy.distutils.cpuinfo import cpu
from numpy.distutils.fcompiler import FCompiler
class NAGFCompiler(FCompiler):
compiler_type = 'nag'
version_pattern = r'NAGWare Fortran 95 compiler Release (?P<version>[^\s]*)'
executables = {
'version_cmd' : ["f95", "-V"],
'compiler_f77' : ["f95", "-fixed"],
'compiler_fix' : ["f95", "-fixed"],
'compiler_f90' : ["f95"],
'linker_so' : ["f95"],
'archiver' : ["ar", "-cr"],
'ranlib' : ["ranlib"]
}
def get_flags_linker_so(self):
if sys.platform=='darwin':
return ['-unsharedf95','-Wl,-bundle,-flat_namespace,-undefined,suppress']
return ["-Wl,-shared"]
def get_flags_opt(self):
return ['-O4']
def get_flags_arch(self):
return ['-target=native']
def get_flags_debug(self):
return ['-g','-gline','-g90','-nan','-C']
if __name__ == '__main__':
from distutils import log
log.set_verbosity(2)
from numpy.distutils.fcompiler import new_fcompiler
compiler = new_fcompiler(compiler='nag')
compiler.customize()
print compiler.get_version()
| import os
import sys
from numpy.distutils.cpuinfo import cpu
from numpy.distutils.fcompiler import FCompiler
class NAGFCompiler(FCompiler):
compiler_type = 'nag'
version_pattern = r'NAGWare Fortran 95 compiler Release (?P<version>[^\s]*)'
executables = {
'version_cmd' : ["f95", "-V"],
'compiler_f77' : ["f95", "-fixed"],
'compiler_fix' : ["f95", "-fixed"],
'compiler_f90' : ["f95"],
'linker_so' : ["f95"],
'archiver' : ["ar", "-cr"],
'ranlib' : ["ranlib"]
}
def get_flags_linker_so(self):
if sys.platform=='darwin':
return ['-unsharedf95','-Wl,-bundle,-flat_namespace,-undefined,suppress']
return ["-Wl,shared"]
def get_flags_opt(self):
return ['-O4']
def get_flags_arch(self):
return ['-target=native']
def get_flags_debug(self):
return ['-g','-gline','-g90','-nan','-C']
if __name__ == '__main__':
from distutils import log
log.set_verbosity(2)
from numpy.distutils.fcompiler import new_fcompiler
compiler = new_fcompiler(compiler='nag')
compiler.customize()
print compiler.get_version()
Fix for using NAG Fortran 95, due to James Graham <jg307@cam.ac.uk>
git-svn-id: 77a43f9646713b91fea7788fad5dfbf67e151ece@2515 94b884b6-d6fd-0310-90d3-974f1d3f35e1import os
import sys
from numpy.distutils.cpuinfo import cpu
from numpy.distutils.fcompiler import FCompiler
class NAGFCompiler(FCompiler):
compiler_type = 'nag'
version_pattern = r'NAGWare Fortran 95 compiler Release (?P<version>[^\s]*)'
executables = {
'version_cmd' : ["f95", "-V"],
'compiler_f77' : ["f95", "-fixed"],
'compiler_fix' : ["f95", "-fixed"],
'compiler_f90' : ["f95"],
'linker_so' : ["f95"],
'archiver' : ["ar", "-cr"],
'ranlib' : ["ranlib"]
}
def get_flags_linker_so(self):
if sys.platform=='darwin':
return ['-unsharedf95','-Wl,-bundle,-flat_namespace,-undefined,suppress']
return ["-Wl,-shared"]
def get_flags_opt(self):
return ['-O4']
def get_flags_arch(self):
return ['-target=native']
def get_flags_debug(self):
return ['-g','-gline','-g90','-nan','-C']
if __name__ == '__main__':
from distutils import log
log.set_verbosity(2)
from numpy.distutils.fcompiler import new_fcompiler
compiler = new_fcompiler(compiler='nag')
compiler.customize()
print compiler.get_version()
| <commit_before>import os
import sys
from numpy.distutils.cpuinfo import cpu
from numpy.distutils.fcompiler import FCompiler
class NAGFCompiler(FCompiler):
compiler_type = 'nag'
version_pattern = r'NAGWare Fortran 95 compiler Release (?P<version>[^\s]*)'
executables = {
'version_cmd' : ["f95", "-V"],
'compiler_f77' : ["f95", "-fixed"],
'compiler_fix' : ["f95", "-fixed"],
'compiler_f90' : ["f95"],
'linker_so' : ["f95"],
'archiver' : ["ar", "-cr"],
'ranlib' : ["ranlib"]
}
def get_flags_linker_so(self):
if sys.platform=='darwin':
return ['-unsharedf95','-Wl,-bundle,-flat_namespace,-undefined,suppress']
return ["-Wl,shared"]
def get_flags_opt(self):
return ['-O4']
def get_flags_arch(self):
return ['-target=native']
def get_flags_debug(self):
return ['-g','-gline','-g90','-nan','-C']
if __name__ == '__main__':
from distutils import log
log.set_verbosity(2)
from numpy.distutils.fcompiler import new_fcompiler
compiler = new_fcompiler(compiler='nag')
compiler.customize()
print compiler.get_version()
<commit_msg>Fix for using NAG Fortran 95, due to James Graham <jg307@cam.ac.uk>
git-svn-id: 77a43f9646713b91fea7788fad5dfbf67e151ece@2515 94b884b6-d6fd-0310-90d3-974f1d3f35e1<commit_after>import os
import sys
from numpy.distutils.cpuinfo import cpu
from numpy.distutils.fcompiler import FCompiler
class NAGFCompiler(FCompiler):
compiler_type = 'nag'
version_pattern = r'NAGWare Fortran 95 compiler Release (?P<version>[^\s]*)'
executables = {
'version_cmd' : ["f95", "-V"],
'compiler_f77' : ["f95", "-fixed"],
'compiler_fix' : ["f95", "-fixed"],
'compiler_f90' : ["f95"],
'linker_so' : ["f95"],
'archiver' : ["ar", "-cr"],
'ranlib' : ["ranlib"]
}
def get_flags_linker_so(self):
if sys.platform=='darwin':
return ['-unsharedf95','-Wl,-bundle,-flat_namespace,-undefined,suppress']
return ["-Wl,-shared"]
def get_flags_opt(self):
return ['-O4']
def get_flags_arch(self):
return ['-target=native']
def get_flags_debug(self):
return ['-g','-gline','-g90','-nan','-C']
if __name__ == '__main__':
from distutils import log
log.set_verbosity(2)
from numpy.distutils.fcompiler import new_fcompiler
compiler = new_fcompiler(compiler='nag')
compiler.customize()
print compiler.get_version()
|
d6b3f4e0798f430761f51529ea61c368e1ce610a | utest/contrib/testrunner/test_pybot_arguments_validation.py | utest/contrib/testrunner/test_pybot_arguments_validation.py | import unittest
import robot.errors
from robotide.contrib.testrunner.runprofiles import PybotProfile
class TestPybotArgumentsValidation(unittest.TestCase):
def setUp(self):
self._profile = PybotProfile(lambda:0)
@unittest.expectedFailure # No more DataError, better argument detection
def test_invalid_argument(self):
try:
self.assertRaisesRegex(robot.errors.DataError,
'option --invalidargument not recognized',
self._profile._get_invalid_message,
'--invalidargument')
except AttributeError: # Python2
self.assertRaisesRegexp(robot.errors.DataError,
'option --invalidargument not recognized',
self._profile._get_invalid_message,
'--invalidargument')
def test_valid_argument_short(self):
self._working_arguments('-T')
def _working_arguments(self, args):
self.assertEqual(None, self._profile._get_invalid_message(args))
def test_valid_argument_long(self):
self._working_arguments('--timestampoutputs')
def test_valid_argument_with_value(self):
self._working_arguments('--log somelog.html')
def test_runfailed_argument_works(self):
self._working_arguments('--runfailed output.xml')
if __name__ == '__main__':
unittest.main()
| import unittest
import robotide.lib.robot.errors
from robotide.contrib.testrunner.runprofiles import PybotProfile
class TestPybotArgumentsValidation(unittest.TestCase):
def setUp(self):
self._profile = PybotProfile(lambda:0)
@unittest.expectedFailure # No more DataError, better argument detection
def test_invalid_argument(self):
try:
self.assertRaisesRegex(robotide.lib.robot.errors.DataError,
'option --invalidargument not recognized',
self._profile._get_invalid_message,
'--invalidargument')
except AttributeError: # Python2
self.assertRaisesRegexp(robotide.lib.robot.errors.DataError,
'option --invalidargument not recognized',
self._profile._get_invalid_message,
'--invalidargument')
def test_valid_argument_short(self):
self._working_arguments('-T')
def _working_arguments(self, args):
self.assertEqual(None, self._profile._get_invalid_message(args))
def test_valid_argument_long(self):
self._working_arguments('--timestampoutputs')
def test_valid_argument_with_value(self):
self._working_arguments('--log somelog.html')
def test_runfailed_argument_works(self):
self._working_arguments('--runfailed output.xml')
if __name__ == '__main__':
unittest.main()
| Fix unit test for when robotframework is not installed. | Fix unit test for when robotframework is not installed.
| Python | apache-2.0 | HelioGuilherme66/RIDE,robotframework/RIDE,robotframework/RIDE,HelioGuilherme66/RIDE,HelioGuilherme66/RIDE,robotframework/RIDE,HelioGuilherme66/RIDE,robotframework/RIDE | import unittest
import robot.errors
from robotide.contrib.testrunner.runprofiles import PybotProfile
class TestPybotArgumentsValidation(unittest.TestCase):
def setUp(self):
self._profile = PybotProfile(lambda:0)
@unittest.expectedFailure # No more DataError, better argument detection
def test_invalid_argument(self):
try:
self.assertRaisesRegex(robot.errors.DataError,
'option --invalidargument not recognized',
self._profile._get_invalid_message,
'--invalidargument')
except AttributeError: # Python2
self.assertRaisesRegexp(robot.errors.DataError,
'option --invalidargument not recognized',
self._profile._get_invalid_message,
'--invalidargument')
def test_valid_argument_short(self):
self._working_arguments('-T')
def _working_arguments(self, args):
self.assertEqual(None, self._profile._get_invalid_message(args))
def test_valid_argument_long(self):
self._working_arguments('--timestampoutputs')
def test_valid_argument_with_value(self):
self._working_arguments('--log somelog.html')
def test_runfailed_argument_works(self):
self._working_arguments('--runfailed output.xml')
if __name__ == '__main__':
unittest.main()
Fix unit test for when robotframework is not installed. | import unittest
import robotide.lib.robot.errors
from robotide.contrib.testrunner.runprofiles import PybotProfile
class TestPybotArgumentsValidation(unittest.TestCase):
def setUp(self):
self._profile = PybotProfile(lambda:0)
@unittest.expectedFailure # No more DataError, better argument detection
def test_invalid_argument(self):
try:
self.assertRaisesRegex(robotide.lib.robot.errors.DataError,
'option --invalidargument not recognized',
self._profile._get_invalid_message,
'--invalidargument')
except AttributeError: # Python2
self.assertRaisesRegexp(robotide.lib.robot.errors.DataError,
'option --invalidargument not recognized',
self._profile._get_invalid_message,
'--invalidargument')
def test_valid_argument_short(self):
self._working_arguments('-T')
def _working_arguments(self, args):
self.assertEqual(None, self._profile._get_invalid_message(args))
def test_valid_argument_long(self):
self._working_arguments('--timestampoutputs')
def test_valid_argument_with_value(self):
self._working_arguments('--log somelog.html')
def test_runfailed_argument_works(self):
self._working_arguments('--runfailed output.xml')
if __name__ == '__main__':
unittest.main()
| <commit_before>import unittest
import robot.errors
from robotide.contrib.testrunner.runprofiles import PybotProfile
class TestPybotArgumentsValidation(unittest.TestCase):
def setUp(self):
self._profile = PybotProfile(lambda:0)
@unittest.expectedFailure # No more DataError, better argument detection
def test_invalid_argument(self):
try:
self.assertRaisesRegex(robot.errors.DataError,
'option --invalidargument not recognized',
self._profile._get_invalid_message,
'--invalidargument')
except AttributeError: # Python2
self.assertRaisesRegexp(robot.errors.DataError,
'option --invalidargument not recognized',
self._profile._get_invalid_message,
'--invalidargument')
def test_valid_argument_short(self):
self._working_arguments('-T')
def _working_arguments(self, args):
self.assertEqual(None, self._profile._get_invalid_message(args))
def test_valid_argument_long(self):
self._working_arguments('--timestampoutputs')
def test_valid_argument_with_value(self):
self._working_arguments('--log somelog.html')
def test_runfailed_argument_works(self):
self._working_arguments('--runfailed output.xml')
if __name__ == '__main__':
unittest.main()
<commit_msg>Fix unit test for when robotframework is not installed.<commit_after> | import unittest
import robotide.lib.robot.errors
from robotide.contrib.testrunner.runprofiles import PybotProfile
class TestPybotArgumentsValidation(unittest.TestCase):
def setUp(self):
self._profile = PybotProfile(lambda:0)
@unittest.expectedFailure # No more DataError, better argument detection
def test_invalid_argument(self):
try:
self.assertRaisesRegex(robotide.lib.robot.errors.DataError,
'option --invalidargument not recognized',
self._profile._get_invalid_message,
'--invalidargument')
except AttributeError: # Python2
self.assertRaisesRegexp(robotide.lib.robot.errors.DataError,
'option --invalidargument not recognized',
self._profile._get_invalid_message,
'--invalidargument')
def test_valid_argument_short(self):
self._working_arguments('-T')
def _working_arguments(self, args):
self.assertEqual(None, self._profile._get_invalid_message(args))
def test_valid_argument_long(self):
self._working_arguments('--timestampoutputs')
def test_valid_argument_with_value(self):
self._working_arguments('--log somelog.html')
def test_runfailed_argument_works(self):
self._working_arguments('--runfailed output.xml')
if __name__ == '__main__':
unittest.main()
| import unittest
import robot.errors
from robotide.contrib.testrunner.runprofiles import PybotProfile
class TestPybotArgumentsValidation(unittest.TestCase):
def setUp(self):
self._profile = PybotProfile(lambda:0)
@unittest.expectedFailure # No more DataError, better argument detection
def test_invalid_argument(self):
try:
self.assertRaisesRegex(robot.errors.DataError,
'option --invalidargument not recognized',
self._profile._get_invalid_message,
'--invalidargument')
except AttributeError: # Python2
self.assertRaisesRegexp(robot.errors.DataError,
'option --invalidargument not recognized',
self._profile._get_invalid_message,
'--invalidargument')
def test_valid_argument_short(self):
self._working_arguments('-T')
def _working_arguments(self, args):
self.assertEqual(None, self._profile._get_invalid_message(args))
def test_valid_argument_long(self):
self._working_arguments('--timestampoutputs')
def test_valid_argument_with_value(self):
self._working_arguments('--log somelog.html')
def test_runfailed_argument_works(self):
self._working_arguments('--runfailed output.xml')
if __name__ == '__main__':
unittest.main()
Fix unit test for when robotframework is not installed.import unittest
import robotide.lib.robot.errors
from robotide.contrib.testrunner.runprofiles import PybotProfile
class TestPybotArgumentsValidation(unittest.TestCase):
def setUp(self):
self._profile = PybotProfile(lambda:0)
@unittest.expectedFailure # No more DataError, better argument detection
def test_invalid_argument(self):
try:
self.assertRaisesRegex(robotide.lib.robot.errors.DataError,
'option --invalidargument not recognized',
self._profile._get_invalid_message,
'--invalidargument')
except AttributeError: # Python2
self.assertRaisesRegexp(robotide.lib.robot.errors.DataError,
'option --invalidargument not recognized',
self._profile._get_invalid_message,
'--invalidargument')
def test_valid_argument_short(self):
self._working_arguments('-T')
def _working_arguments(self, args):
self.assertEqual(None, self._profile._get_invalid_message(args))
def test_valid_argument_long(self):
self._working_arguments('--timestampoutputs')
def test_valid_argument_with_value(self):
self._working_arguments('--log somelog.html')
def test_runfailed_argument_works(self):
self._working_arguments('--runfailed output.xml')
if __name__ == '__main__':
unittest.main()
| <commit_before>import unittest
import robot.errors
from robotide.contrib.testrunner.runprofiles import PybotProfile
class TestPybotArgumentsValidation(unittest.TestCase):
def setUp(self):
self._profile = PybotProfile(lambda:0)
@unittest.expectedFailure # No more DataError, better argument detection
def test_invalid_argument(self):
try:
self.assertRaisesRegex(robot.errors.DataError,
'option --invalidargument not recognized',
self._profile._get_invalid_message,
'--invalidargument')
except AttributeError: # Python2
self.assertRaisesRegexp(robot.errors.DataError,
'option --invalidargument not recognized',
self._profile._get_invalid_message,
'--invalidargument')
def test_valid_argument_short(self):
self._working_arguments('-T')
def _working_arguments(self, args):
self.assertEqual(None, self._profile._get_invalid_message(args))
def test_valid_argument_long(self):
self._working_arguments('--timestampoutputs')
def test_valid_argument_with_value(self):
self._working_arguments('--log somelog.html')
def test_runfailed_argument_works(self):
self._working_arguments('--runfailed output.xml')
if __name__ == '__main__':
unittest.main()
<commit_msg>Fix unit test for when robotframework is not installed.<commit_after>import unittest
import robotide.lib.robot.errors
from robotide.contrib.testrunner.runprofiles import PybotProfile
class TestPybotArgumentsValidation(unittest.TestCase):
def setUp(self):
self._profile = PybotProfile(lambda:0)
@unittest.expectedFailure # No more DataError, better argument detection
def test_invalid_argument(self):
try:
self.assertRaisesRegex(robotide.lib.robot.errors.DataError,
'option --invalidargument not recognized',
self._profile._get_invalid_message,
'--invalidargument')
except AttributeError: # Python2
self.assertRaisesRegexp(robotide.lib.robot.errors.DataError,
'option --invalidargument not recognized',
self._profile._get_invalid_message,
'--invalidargument')
def test_valid_argument_short(self):
self._working_arguments('-T')
def _working_arguments(self, args):
self.assertEqual(None, self._profile._get_invalid_message(args))
def test_valid_argument_long(self):
self._working_arguments('--timestampoutputs')
def test_valid_argument_with_value(self):
self._working_arguments('--log somelog.html')
def test_runfailed_argument_works(self):
self._working_arguments('--runfailed output.xml')
if __name__ == '__main__':
unittest.main()
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.