commit
stringlengths 40
40
| old_file
stringlengths 4
118
| new_file
stringlengths 4
118
| old_contents
stringlengths 0
2.94k
| new_contents
stringlengths 1
4.43k
| subject
stringlengths 15
444
| message
stringlengths 16
3.45k
| lang
stringclasses 1
value | license
stringclasses 13
values | repos
stringlengths 5
43.2k
| prompt
stringlengths 17
4.58k
| response
stringlengths 1
4.43k
| prompt_tagged
stringlengths 58
4.62k
| response_tagged
stringlengths 1
4.43k
| text
stringlengths 132
7.29k
| text_tagged
stringlengths 173
7.33k
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
67ea5109ddcdb19d77de882960d5eb791c1368ae
|
setup.py
|
setup.py
|
#!/usr/bin/python3
# SPDX-License-Identifier: LGPL-2.1+
from setuptools import setup, Command, find_packages
class BuildManpage(Command):
description = ('builds the manpage')
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def run(self):
self.spawn(['pandoc', '-t', 'man', '-s', '-o', 'man/mkosi.1', 'mkosi.md'])
setup(
name="mkosi",
version="13",
description="Build Bespoke OS Images",
url="https://github.com/systemd/mkosi",
maintainer="mkosi contributors",
maintainer_email="systemd-devel@lists.freedesktop.org",
license="LGPLv2+",
python_requires=">=3.7",
packages = find_packages(".", exclude=["tests"]),
package_data = {"": ["*.sh", "*.hook", "*.conf"]},
include_package_data = True,
scripts = ["bin/mkosi"],
cmdclass = { "man": BuildManpage },
data_files = [('share/man/man1', ["man/mkosi.1"])],
)
|
#!/usr/bin/python3
# SPDX-License-Identifier: LGPL-2.1+
from setuptools import setup, Command, find_packages
class BuildManpage(Command):
description = ('builds the manpage')
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def run(self):
self.spawn(['pandoc', '-t', 'man', '-s', '-o', 'man/mkosi.1', 'mkosi.md'])
setup(
name="mkosi",
version="13",
description="Build Bespoke OS Images",
url="https://github.com/systemd/mkosi",
maintainer="mkosi contributors",
maintainer_email="systemd-devel@lists.freedesktop.org",
license="LGPLv2+",
python_requires=">=3.7",
packages = find_packages(".", exclude=["tests"]),
package_data = {"": ["*.sh", "*.hook", "*.conf", "*.install"]},
include_package_data = True,
scripts = ["bin/mkosi"],
cmdclass = { "man": BuildManpage },
data_files = [('share/man/man1', ["man/mkosi.1"])],
)
|
Include any files ending in '.install' in package data
|
Include any files ending in '.install' in package data
This makes sure the new `dpkg-reconfigure-dracut.install` file under resources
gets included as package data.
|
Python
|
lgpl-2.1
|
systemd/mkosi,systemd/mkosi
|
#!/usr/bin/python3
# SPDX-License-Identifier: LGPL-2.1+
from setuptools import setup, Command, find_packages
class BuildManpage(Command):
description = ('builds the manpage')
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def run(self):
self.spawn(['pandoc', '-t', 'man', '-s', '-o', 'man/mkosi.1', 'mkosi.md'])
setup(
name="mkosi",
version="13",
description="Build Bespoke OS Images",
url="https://github.com/systemd/mkosi",
maintainer="mkosi contributors",
maintainer_email="systemd-devel@lists.freedesktop.org",
license="LGPLv2+",
python_requires=">=3.7",
packages = find_packages(".", exclude=["tests"]),
package_data = {"": ["*.sh", "*.hook", "*.conf"]},
include_package_data = True,
scripts = ["bin/mkosi"],
cmdclass = { "man": BuildManpage },
data_files = [('share/man/man1', ["man/mkosi.1"])],
)
Include any files ending in '.install' in package data
This makes sure the new `dpkg-reconfigure-dracut.install` file under resources
gets included as package data.
|
#!/usr/bin/python3
# SPDX-License-Identifier: LGPL-2.1+
from setuptools import setup, Command, find_packages
class BuildManpage(Command):
description = ('builds the manpage')
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def run(self):
self.spawn(['pandoc', '-t', 'man', '-s', '-o', 'man/mkosi.1', 'mkosi.md'])
setup(
name="mkosi",
version="13",
description="Build Bespoke OS Images",
url="https://github.com/systemd/mkosi",
maintainer="mkosi contributors",
maintainer_email="systemd-devel@lists.freedesktop.org",
license="LGPLv2+",
python_requires=">=3.7",
packages = find_packages(".", exclude=["tests"]),
package_data = {"": ["*.sh", "*.hook", "*.conf", "*.install"]},
include_package_data = True,
scripts = ["bin/mkosi"],
cmdclass = { "man": BuildManpage },
data_files = [('share/man/man1', ["man/mkosi.1"])],
)
|
<commit_before>#!/usr/bin/python3
# SPDX-License-Identifier: LGPL-2.1+
from setuptools import setup, Command, find_packages
class BuildManpage(Command):
description = ('builds the manpage')
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def run(self):
self.spawn(['pandoc', '-t', 'man', '-s', '-o', 'man/mkosi.1', 'mkosi.md'])
setup(
name="mkosi",
version="13",
description="Build Bespoke OS Images",
url="https://github.com/systemd/mkosi",
maintainer="mkosi contributors",
maintainer_email="systemd-devel@lists.freedesktop.org",
license="LGPLv2+",
python_requires=">=3.7",
packages = find_packages(".", exclude=["tests"]),
package_data = {"": ["*.sh", "*.hook", "*.conf"]},
include_package_data = True,
scripts = ["bin/mkosi"],
cmdclass = { "man": BuildManpage },
data_files = [('share/man/man1', ["man/mkosi.1"])],
)
<commit_msg>Include any files ending in '.install' in package data
This makes sure the new `dpkg-reconfigure-dracut.install` file under resources
gets included as package data.<commit_after>
|
#!/usr/bin/python3
# SPDX-License-Identifier: LGPL-2.1+
from setuptools import setup, Command, find_packages
class BuildManpage(Command):
description = ('builds the manpage')
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def run(self):
self.spawn(['pandoc', '-t', 'man', '-s', '-o', 'man/mkosi.1', 'mkosi.md'])
setup(
name="mkosi",
version="13",
description="Build Bespoke OS Images",
url="https://github.com/systemd/mkosi",
maintainer="mkosi contributors",
maintainer_email="systemd-devel@lists.freedesktop.org",
license="LGPLv2+",
python_requires=">=3.7",
packages = find_packages(".", exclude=["tests"]),
package_data = {"": ["*.sh", "*.hook", "*.conf", "*.install"]},
include_package_data = True,
scripts = ["bin/mkosi"],
cmdclass = { "man": BuildManpage },
data_files = [('share/man/man1', ["man/mkosi.1"])],
)
|
#!/usr/bin/python3
# SPDX-License-Identifier: LGPL-2.1+
from setuptools import setup, Command, find_packages
class BuildManpage(Command):
description = ('builds the manpage')
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def run(self):
self.spawn(['pandoc', '-t', 'man', '-s', '-o', 'man/mkosi.1', 'mkosi.md'])
setup(
name="mkosi",
version="13",
description="Build Bespoke OS Images",
url="https://github.com/systemd/mkosi",
maintainer="mkosi contributors",
maintainer_email="systemd-devel@lists.freedesktop.org",
license="LGPLv2+",
python_requires=">=3.7",
packages = find_packages(".", exclude=["tests"]),
package_data = {"": ["*.sh", "*.hook", "*.conf"]},
include_package_data = True,
scripts = ["bin/mkosi"],
cmdclass = { "man": BuildManpage },
data_files = [('share/man/man1', ["man/mkosi.1"])],
)
Include any files ending in '.install' in package data
This makes sure the new `dpkg-reconfigure-dracut.install` file under resources
gets included as package data.#!/usr/bin/python3
# SPDX-License-Identifier: LGPL-2.1+
from setuptools import setup, Command, find_packages
class BuildManpage(Command):
description = ('builds the manpage')
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def run(self):
self.spawn(['pandoc', '-t', 'man', '-s', '-o', 'man/mkosi.1', 'mkosi.md'])
setup(
name="mkosi",
version="13",
description="Build Bespoke OS Images",
url="https://github.com/systemd/mkosi",
maintainer="mkosi contributors",
maintainer_email="systemd-devel@lists.freedesktop.org",
license="LGPLv2+",
python_requires=">=3.7",
packages = find_packages(".", exclude=["tests"]),
package_data = {"": ["*.sh", "*.hook", "*.conf", "*.install"]},
include_package_data = True,
scripts = ["bin/mkosi"],
cmdclass = { "man": BuildManpage },
data_files = [('share/man/man1', ["man/mkosi.1"])],
)
|
<commit_before>#!/usr/bin/python3
# SPDX-License-Identifier: LGPL-2.1+
from setuptools import setup, Command, find_packages
class BuildManpage(Command):
description = ('builds the manpage')
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def run(self):
self.spawn(['pandoc', '-t', 'man', '-s', '-o', 'man/mkosi.1', 'mkosi.md'])
setup(
name="mkosi",
version="13",
description="Build Bespoke OS Images",
url="https://github.com/systemd/mkosi",
maintainer="mkosi contributors",
maintainer_email="systemd-devel@lists.freedesktop.org",
license="LGPLv2+",
python_requires=">=3.7",
packages = find_packages(".", exclude=["tests"]),
package_data = {"": ["*.sh", "*.hook", "*.conf"]},
include_package_data = True,
scripts = ["bin/mkosi"],
cmdclass = { "man": BuildManpage },
data_files = [('share/man/man1', ["man/mkosi.1"])],
)
<commit_msg>Include any files ending in '.install' in package data
This makes sure the new `dpkg-reconfigure-dracut.install` file under resources
gets included as package data.<commit_after>#!/usr/bin/python3
# SPDX-License-Identifier: LGPL-2.1+
from setuptools import setup, Command, find_packages
class BuildManpage(Command):
description = ('builds the manpage')
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def run(self):
self.spawn(['pandoc', '-t', 'man', '-s', '-o', 'man/mkosi.1', 'mkosi.md'])
setup(
name="mkosi",
version="13",
description="Build Bespoke OS Images",
url="https://github.com/systemd/mkosi",
maintainer="mkosi contributors",
maintainer_email="systemd-devel@lists.freedesktop.org",
license="LGPLv2+",
python_requires=">=3.7",
packages = find_packages(".", exclude=["tests"]),
package_data = {"": ["*.sh", "*.hook", "*.conf", "*.install"]},
include_package_data = True,
scripts = ["bin/mkosi"],
cmdclass = { "man": BuildManpage },
data_files = [('share/man/man1', ["man/mkosi.1"])],
)
|
b825ee12fd6abc91b80b8a62886b9c53b82cdeeb
|
test/task_test.py
|
test/task_test.py
|
import doctest
import unittest
import luigi.task
class TaskTest(unittest.TestCase):
def test_tasks_doctest(self):
doctest.testmod(luigi.task)
|
import doctest
import unittest
import luigi.task
import luigi
from datetime import datetime, timedelta
class DummyTask(luigi.Task):
param = luigi.Parameter()
bool_param = luigi.BooleanParameter()
int_param = luigi.IntParameter()
float_param = luigi.FloatParameter()
date_param = luigi.DateParameter()
datehour_param = luigi.DateHourParameter()
timedelta_param = luigi.TimeDeltaParameter()
list_param = luigi.Parameter(is_list=True)
class TaskTest(unittest.TestCase):
def test_tasks_doctest(self):
doctest.testmod(luigi.task)
def test_task_to_str_to_task(self):
params = dict(
param='test',
bool_param=True,
int_param=666,
float_param=123.456,
date_param=datetime(2014, 9, 13).date(),
datehour_param=datetime(2014, 9, 13, 9),
timedelta_param=timedelta(44), # doesn't support seconds
list_param=['in', 'flames'])
original = DummyTask(**params)
other = DummyTask.from_str_params(original.to_str_params(), {})
self.assertEqual(original, other)
if __name__ == '__main__':
unittest.main()
|
Add test for task to str params conversion
|
Add test for task to str params conversion
|
Python
|
apache-2.0
|
soxofaan/luigi,adaitche/luigi,alkemics/luigi,dlstadther/luigi,fw1121/luigi,lungetech/luigi,linsomniac/luigi,moandcompany/luigi,mbruggmann/luigi,vine/luigi,aeron15/luigi,leafjungle/luigi,Houzz/luigi,springcoil/luigi,ViaSat/luigi,Tarrasch/luigi,oldpa/luigi,linsomniac/luigi,dstandish/luigi,kalaidin/luigi,jw0201/luigi,h3biomed/luigi,aeron15/luigi,DomainGroupOSS/luigi,huiyi1990/luigi,alkemics/luigi,dstandish/luigi,laserson/luigi,jw0201/luigi,soxofaan/luigi,graingert/luigi,wakamori/luigi,dylanjbarth/luigi,dlstadther/luigi,alkemics/luigi,kevhill/luigi,Yoone/luigi,SeedScientific/luigi,spotify/luigi,PeteW/luigi,humanlongevity/luigi,JackDanger/luigi,sahitya-pavurala/luigi,glenndmello/luigi,moritzschaefer/luigi,Magnetic/luigi,LamCiuLoeng/luigi,walkers-mv/luigi,moritzschaefer/luigi,dhruvg/luigi,bmaggard/luigi,bowlofstew/luigi,samepage-labs/luigi,fabriziodemaria/luigi,pkexcellent/luigi,kevhill/luigi,dkroy/luigi,JackDanger/luigi,walkers-mv/luigi,ehdr/luigi,LamCiuLoeng/luigi,samepage-labs/luigi,spotify/luigi,ViaSat/luigi,tuulos/luigi,ThQ/luigi,wakamori/luigi,moandcompany/luigi,mbruggmann/luigi,belevtsoff/luigi,neilisaac/luigi,slvnperron/luigi,anyman/luigi,realgo/luigi,ThQ/luigi,ChrisBeaumont/luigi,ThQ/luigi,graingert/luigi,rizzatti/luigi,ZhenxingWu/luigi,torypages/luigi,jamesmcm/luigi,glenndmello/luigi,meyerson/luigi,ChrisBeaumont/luigi,kalaidin/luigi,rizzatti/luigi,mbruggmann/luigi,linearregression/luigi,springcoil/luigi,gpoulin/luigi,huiyi1990/luigi,aeron15/luigi,stephenpascoe/luigi,rayrrr/luigi,aeron15/luigi,lichia/luigi,linearregression/luigi,penelopy/luigi,hadesbox/luigi,ContextLogic/luigi,hellais/luigi,SkyTruth/luigi,theoryno3/luigi,fabriziodemaria/luigi,pkexcellent/luigi,Houzz/luigi,anyman/luigi,bowlofstew/luigi,jw0201/luigi,kalaidin/luigi,moritzschaefer/luigi,joeshaw/luigi,mfcabrera/luigi,ViaSat/luigi,rizzatti/luigi,dlstadther/luigi,thejens/luigi,gpoulin/luigi,mfcabrera/luigi,penelopy/luigi,thejens/luigi,lungetech/luigi,h3biomed/luigi,bmaggard/luigi,fabriziodemaria/luigi,ThQ/luigi,ivannotes/luigi,penelopy/luigi,casey-green/luigi,walkers-mv/luigi,hadesbox/luigi,ChrisBeaumont/luigi,kalaidin/luigi,sahitya-pavurala/luigi,hellais/luigi,laserson/luigi,gpoulin/luigi,altaf-ali/luigi,lichia/luigi,anyman/luigi,stephenpascoe/luigi,meyerson/luigi,lichia/luigi,ViaSat/luigi,Yoone/luigi,walkers-mv/luigi,joeshaw/luigi,realgo/luigi,republic-analytics/luigi,stroykova/luigi,Tarrasch/luigi,vine/luigi,samuell/luigi,foursquare/luigi,JackDanger/luigi,percyfal/luigi,tuulos/luigi,ContextLogic/luigi,upworthy/luigi,jw0201/luigi,stroykova/luigi,Yoone/luigi,lungetech/luigi,pkexcellent/luigi,humanlongevity/luigi,hellais/luigi,Houzz/luigi,lungetech/luigi,Dawny33/luigi,humanlongevity/luigi,hellais/luigi,ZhenxingWu/luigi,hadesbox/luigi,SkyTruth/luigi,dylanjbarth/luigi,edx/luigi,ehdr/luigi,glenndmello/luigi,PeteW/luigi,qpxu007/luigi,harveyxia/luigi,percyfal/luigi,drincruz/luigi,ehdr/luigi,sahitya-pavurala/luigi,SkyTruth/luigi,ZhenxingWu/luigi,huiyi1990/luigi,javrasya/luigi,penelopy/luigi,joeshaw/luigi,dstandish/luigi,riga/luigi,bowlofstew/luigi,foursquare/luigi,mfcabrera/luigi,dhruvg/luigi,glenndmello/luigi,bmaggard/luigi,Tarrasch/luigi,fw1121/luigi,samuell/luigi,moritzschaefer/luigi,javrasya/luigi,riga/luigi,fabriziodemaria/luigi,DomainGroupOSS/luigi,SkyTruth/luigi,neilisaac/luigi,leafjungle/luigi,dkroy/luigi,realgo/luigi,bmaggard/luigi,wakamori/luigi,qpxu007/luigi,dstandish/luigi,casey-green/luigi,dylanjbarth/luigi,republic-analytics/luigi,drincruz/luigi,vine/luigi,linsomniac/luigi,LamCiuLoeng/luigi,graingert/luigi,PeteW/luigi,riga/luigi,DomainGroupOSS/luigi,altaf-ali/luigi,Wattpad/luigi,JackDanger/luigi,javrasya/luigi,casey-green/luigi,fw1121/luigi,Wattpad/luigi,slvnperron/luigi,stephenpascoe/luigi,harveyxia/luigi,adaitche/luigi,springcoil/luigi,leafjungle/luigi,Houzz/luigi,ChrisBeaumont/luigi,dhruvg/luigi,jamesmcm/luigi,casey-green/luigi,stroykova/luigi,samepage-labs/luigi,moandcompany/luigi,torypages/luigi,h3biomed/luigi,samepage-labs/luigi,lichia/luigi,huiyi1990/luigi,dkroy/luigi,altaf-ali/luigi,tuulos/luigi,kevhill/luigi,javrasya/luigi,LamCiuLoeng/luigi,SeedScientific/luigi,meyerson/luigi,samuell/luigi,rizzatti/luigi,ContextLogic/luigi,foursquare/luigi,laserson/luigi,humanlongevity/luigi,theoryno3/luigi,Dawny33/luigi,neilisaac/luigi,SeedScientific/luigi,belevtsoff/luigi,belevtsoff/luigi,percyfal/luigi,mfcabrera/luigi,mbruggmann/luigi,realgo/luigi,vine/luigi,moandcompany/luigi,belevtsoff/luigi,wakamori/luigi,adaitche/luigi,17zuoye/luigi,linsomniac/luigi,meyerson/luigi,spotify/luigi,torypages/luigi,foursquare/luigi,dhruvg/luigi,DomainGroupOSS/luigi,laserson/luigi,jamesmcm/luigi,Magnetic/luigi,spotify/luigi,tuulos/luigi,fw1121/luigi,ivannotes/luigi,soxofaan/luigi,Yoone/luigi,qpxu007/luigi,gpoulin/luigi,17zuoye/luigi,harveyxia/luigi,hadesbox/luigi,h3biomed/luigi,rayrrr/luigi,17zuoye/luigi,alkemics/luigi,linearregression/luigi,ivannotes/luigi,riga/luigi,Magnetic/luigi,upworthy/luigi,theoryno3/luigi,torypages/luigi,harveyxia/luigi,SeedScientific/luigi,adaitche/luigi,slvnperron/luigi,dkroy/luigi,samuell/luigi,qpxu007/luigi,oldpa/luigi,edx/luigi,upworthy/luigi,linearregression/luigi,Dawny33/luigi,thejens/luigi,Wattpad/luigi,bowlofstew/luigi,republic-analytics/luigi,slvnperron/luigi,percyfal/luigi,edx/luigi,upworthy/luigi,dylanjbarth/luigi,pkexcellent/luigi,leafjungle/luigi,graingert/luigi,soxofaan/luigi,kevhill/luigi,springcoil/luigi,rayrrr/luigi,stephenpascoe/luigi,17zuoye/luigi,theoryno3/luigi,drincruz/luigi,thejens/luigi,rayrrr/luigi,jamesmcm/luigi,Magnetic/luigi,ehdr/luigi,Dawny33/luigi,stroykova/luigi,ivannotes/luigi,sahitya-pavurala/luigi,drincruz/luigi,republic-analytics/luigi,Tarrasch/luigi,ContextLogic/luigi,ZhenxingWu/luigi,neilisaac/luigi,dlstadther/luigi,oldpa/luigi,altaf-ali/luigi,oldpa/luigi,anyman/luigi,joeshaw/luigi,edx/luigi,PeteW/luigi
|
import doctest
import unittest
import luigi.task
class TaskTest(unittest.TestCase):
def test_tasks_doctest(self):
doctest.testmod(luigi.task)
Add test for task to str params conversion
|
import doctest
import unittest
import luigi.task
import luigi
from datetime import datetime, timedelta
class DummyTask(luigi.Task):
param = luigi.Parameter()
bool_param = luigi.BooleanParameter()
int_param = luigi.IntParameter()
float_param = luigi.FloatParameter()
date_param = luigi.DateParameter()
datehour_param = luigi.DateHourParameter()
timedelta_param = luigi.TimeDeltaParameter()
list_param = luigi.Parameter(is_list=True)
class TaskTest(unittest.TestCase):
def test_tasks_doctest(self):
doctest.testmod(luigi.task)
def test_task_to_str_to_task(self):
params = dict(
param='test',
bool_param=True,
int_param=666,
float_param=123.456,
date_param=datetime(2014, 9, 13).date(),
datehour_param=datetime(2014, 9, 13, 9),
timedelta_param=timedelta(44), # doesn't support seconds
list_param=['in', 'flames'])
original = DummyTask(**params)
other = DummyTask.from_str_params(original.to_str_params(), {})
self.assertEqual(original, other)
if __name__ == '__main__':
unittest.main()
|
<commit_before>import doctest
import unittest
import luigi.task
class TaskTest(unittest.TestCase):
def test_tasks_doctest(self):
doctest.testmod(luigi.task)
<commit_msg>Add test for task to str params conversion<commit_after>
|
import doctest
import unittest
import luigi.task
import luigi
from datetime import datetime, timedelta
class DummyTask(luigi.Task):
param = luigi.Parameter()
bool_param = luigi.BooleanParameter()
int_param = luigi.IntParameter()
float_param = luigi.FloatParameter()
date_param = luigi.DateParameter()
datehour_param = luigi.DateHourParameter()
timedelta_param = luigi.TimeDeltaParameter()
list_param = luigi.Parameter(is_list=True)
class TaskTest(unittest.TestCase):
def test_tasks_doctest(self):
doctest.testmod(luigi.task)
def test_task_to_str_to_task(self):
params = dict(
param='test',
bool_param=True,
int_param=666,
float_param=123.456,
date_param=datetime(2014, 9, 13).date(),
datehour_param=datetime(2014, 9, 13, 9),
timedelta_param=timedelta(44), # doesn't support seconds
list_param=['in', 'flames'])
original = DummyTask(**params)
other = DummyTask.from_str_params(original.to_str_params(), {})
self.assertEqual(original, other)
if __name__ == '__main__':
unittest.main()
|
import doctest
import unittest
import luigi.task
class TaskTest(unittest.TestCase):
def test_tasks_doctest(self):
doctest.testmod(luigi.task)
Add test for task to str params conversionimport doctest
import unittest
import luigi.task
import luigi
from datetime import datetime, timedelta
class DummyTask(luigi.Task):
param = luigi.Parameter()
bool_param = luigi.BooleanParameter()
int_param = luigi.IntParameter()
float_param = luigi.FloatParameter()
date_param = luigi.DateParameter()
datehour_param = luigi.DateHourParameter()
timedelta_param = luigi.TimeDeltaParameter()
list_param = luigi.Parameter(is_list=True)
class TaskTest(unittest.TestCase):
def test_tasks_doctest(self):
doctest.testmod(luigi.task)
def test_task_to_str_to_task(self):
params = dict(
param='test',
bool_param=True,
int_param=666,
float_param=123.456,
date_param=datetime(2014, 9, 13).date(),
datehour_param=datetime(2014, 9, 13, 9),
timedelta_param=timedelta(44), # doesn't support seconds
list_param=['in', 'flames'])
original = DummyTask(**params)
other = DummyTask.from_str_params(original.to_str_params(), {})
self.assertEqual(original, other)
if __name__ == '__main__':
unittest.main()
|
<commit_before>import doctest
import unittest
import luigi.task
class TaskTest(unittest.TestCase):
def test_tasks_doctest(self):
doctest.testmod(luigi.task)
<commit_msg>Add test for task to str params conversion<commit_after>import doctest
import unittest
import luigi.task
import luigi
from datetime import datetime, timedelta
class DummyTask(luigi.Task):
param = luigi.Parameter()
bool_param = luigi.BooleanParameter()
int_param = luigi.IntParameter()
float_param = luigi.FloatParameter()
date_param = luigi.DateParameter()
datehour_param = luigi.DateHourParameter()
timedelta_param = luigi.TimeDeltaParameter()
list_param = luigi.Parameter(is_list=True)
class TaskTest(unittest.TestCase):
def test_tasks_doctest(self):
doctest.testmod(luigi.task)
def test_task_to_str_to_task(self):
params = dict(
param='test',
bool_param=True,
int_param=666,
float_param=123.456,
date_param=datetime(2014, 9, 13).date(),
datehour_param=datetime(2014, 9, 13, 9),
timedelta_param=timedelta(44), # doesn't support seconds
list_param=['in', 'flames'])
original = DummyTask(**params)
other = DummyTask.from_str_params(original.to_str_params(), {})
self.assertEqual(original, other)
if __name__ == '__main__':
unittest.main()
|
b0202e8882f792feb041070baff7370cacf73751
|
tests/test_api.py
|
tests/test_api.py
|
# -*- coding: utf-8 -*-
import subprocess
import time
from unittest import TestCase
from nose.tools import assert_equal
class TestOldApi(TestCase):
def setUp(self):
self.process = subprocess.Popen("openfisca-serve")
def tearDown(self):
self.process.terminate()
def test_response(self):
try:
subprocess.check_call(['wget', '--quiet', '--retry-connrefused', '--waitretry=1', '--tries=10', 'http://localhost:2000', '--output-document=/dev/null'])
except subprocess.CalledProcessError:
raise subprocess.CalledProcessError("Could not reach OpenFisca Web API at localhost:2000 after 10s")
|
# -*- coding: utf-8 -*-
import subprocess
import time
from unittest import TestCase
from nose.tools import assert_equal
class TestOldApi(TestCase):
def setUp(self):
self.process = subprocess.Popen("openfisca-serve")
def tearDown(self):
self.process.terminate()
def test_response(self):
try:
subprocess.check_call(['wget', '--quiet', '--retry-connrefused', '--waitretry=1', '--tries=10', 'http://localhost:2000', '--output-document=/dev/null'])
except subprocess.CalledProcessError:
raise subprocess.CalledProcessError("Could not reach OpenFisca Web API at localhost:2000 after 10s")
class TestNewApi(TestCase):
def setUp(self):
self.process = subprocess.Popen(['openfisca', 'serve'])
def tearDown(self):
self.process.terminate()
def test_response(self):
try:
subprocess.check_call(['wget', '--quiet', '--retry-connrefused', '--waitretry=1', '--tries=10', 'http://localhost:6000/parameters', '--output-document=/dev/null'])
except subprocess.CalledProcessError:
raise subprocess.CalledProcessError("Could not reach OpenFisca Web API at localhost:6000 after 10s")
|
Test france compatibility with the new API
|
Test france compatibility with the new API
|
Python
|
agpl-3.0
|
antoinearnoud/openfisca-france,sgmap/openfisca-france,sgmap/openfisca-france,antoinearnoud/openfisca-france
|
# -*- coding: utf-8 -*-
import subprocess
import time
from unittest import TestCase
from nose.tools import assert_equal
class TestOldApi(TestCase):
def setUp(self):
self.process = subprocess.Popen("openfisca-serve")
def tearDown(self):
self.process.terminate()
def test_response(self):
try:
subprocess.check_call(['wget', '--quiet', '--retry-connrefused', '--waitretry=1', '--tries=10', 'http://localhost:2000', '--output-document=/dev/null'])
except subprocess.CalledProcessError:
raise subprocess.CalledProcessError("Could not reach OpenFisca Web API at localhost:2000 after 10s")
Test france compatibility with the new API
|
# -*- coding: utf-8 -*-
import subprocess
import time
from unittest import TestCase
from nose.tools import assert_equal
class TestOldApi(TestCase):
def setUp(self):
self.process = subprocess.Popen("openfisca-serve")
def tearDown(self):
self.process.terminate()
def test_response(self):
try:
subprocess.check_call(['wget', '--quiet', '--retry-connrefused', '--waitretry=1', '--tries=10', 'http://localhost:2000', '--output-document=/dev/null'])
except subprocess.CalledProcessError:
raise subprocess.CalledProcessError("Could not reach OpenFisca Web API at localhost:2000 after 10s")
class TestNewApi(TestCase):
def setUp(self):
self.process = subprocess.Popen(['openfisca', 'serve'])
def tearDown(self):
self.process.terminate()
def test_response(self):
try:
subprocess.check_call(['wget', '--quiet', '--retry-connrefused', '--waitretry=1', '--tries=10', 'http://localhost:6000/parameters', '--output-document=/dev/null'])
except subprocess.CalledProcessError:
raise subprocess.CalledProcessError("Could not reach OpenFisca Web API at localhost:6000 after 10s")
|
<commit_before># -*- coding: utf-8 -*-
import subprocess
import time
from unittest import TestCase
from nose.tools import assert_equal
class TestOldApi(TestCase):
def setUp(self):
self.process = subprocess.Popen("openfisca-serve")
def tearDown(self):
self.process.terminate()
def test_response(self):
try:
subprocess.check_call(['wget', '--quiet', '--retry-connrefused', '--waitretry=1', '--tries=10', 'http://localhost:2000', '--output-document=/dev/null'])
except subprocess.CalledProcessError:
raise subprocess.CalledProcessError("Could not reach OpenFisca Web API at localhost:2000 after 10s")
<commit_msg>Test france compatibility with the new API<commit_after>
|
# -*- coding: utf-8 -*-
import subprocess
import time
from unittest import TestCase
from nose.tools import assert_equal
class TestOldApi(TestCase):
def setUp(self):
self.process = subprocess.Popen("openfisca-serve")
def tearDown(self):
self.process.terminate()
def test_response(self):
try:
subprocess.check_call(['wget', '--quiet', '--retry-connrefused', '--waitretry=1', '--tries=10', 'http://localhost:2000', '--output-document=/dev/null'])
except subprocess.CalledProcessError:
raise subprocess.CalledProcessError("Could not reach OpenFisca Web API at localhost:2000 after 10s")
class TestNewApi(TestCase):
def setUp(self):
self.process = subprocess.Popen(['openfisca', 'serve'])
def tearDown(self):
self.process.terminate()
def test_response(self):
try:
subprocess.check_call(['wget', '--quiet', '--retry-connrefused', '--waitretry=1', '--tries=10', 'http://localhost:6000/parameters', '--output-document=/dev/null'])
except subprocess.CalledProcessError:
raise subprocess.CalledProcessError("Could not reach OpenFisca Web API at localhost:6000 after 10s")
|
# -*- coding: utf-8 -*-
import subprocess
import time
from unittest import TestCase
from nose.tools import assert_equal
class TestOldApi(TestCase):
def setUp(self):
self.process = subprocess.Popen("openfisca-serve")
def tearDown(self):
self.process.terminate()
def test_response(self):
try:
subprocess.check_call(['wget', '--quiet', '--retry-connrefused', '--waitretry=1', '--tries=10', 'http://localhost:2000', '--output-document=/dev/null'])
except subprocess.CalledProcessError:
raise subprocess.CalledProcessError("Could not reach OpenFisca Web API at localhost:2000 after 10s")
Test france compatibility with the new API# -*- coding: utf-8 -*-
import subprocess
import time
from unittest import TestCase
from nose.tools import assert_equal
class TestOldApi(TestCase):
def setUp(self):
self.process = subprocess.Popen("openfisca-serve")
def tearDown(self):
self.process.terminate()
def test_response(self):
try:
subprocess.check_call(['wget', '--quiet', '--retry-connrefused', '--waitretry=1', '--tries=10', 'http://localhost:2000', '--output-document=/dev/null'])
except subprocess.CalledProcessError:
raise subprocess.CalledProcessError("Could not reach OpenFisca Web API at localhost:2000 after 10s")
class TestNewApi(TestCase):
def setUp(self):
self.process = subprocess.Popen(['openfisca', 'serve'])
def tearDown(self):
self.process.terminate()
def test_response(self):
try:
subprocess.check_call(['wget', '--quiet', '--retry-connrefused', '--waitretry=1', '--tries=10', 'http://localhost:6000/parameters', '--output-document=/dev/null'])
except subprocess.CalledProcessError:
raise subprocess.CalledProcessError("Could not reach OpenFisca Web API at localhost:6000 after 10s")
|
<commit_before># -*- coding: utf-8 -*-
import subprocess
import time
from unittest import TestCase
from nose.tools import assert_equal
class TestOldApi(TestCase):
def setUp(self):
self.process = subprocess.Popen("openfisca-serve")
def tearDown(self):
self.process.terminate()
def test_response(self):
try:
subprocess.check_call(['wget', '--quiet', '--retry-connrefused', '--waitretry=1', '--tries=10', 'http://localhost:2000', '--output-document=/dev/null'])
except subprocess.CalledProcessError:
raise subprocess.CalledProcessError("Could not reach OpenFisca Web API at localhost:2000 after 10s")
<commit_msg>Test france compatibility with the new API<commit_after># -*- coding: utf-8 -*-
import subprocess
import time
from unittest import TestCase
from nose.tools import assert_equal
class TestOldApi(TestCase):
def setUp(self):
self.process = subprocess.Popen("openfisca-serve")
def tearDown(self):
self.process.terminate()
def test_response(self):
try:
subprocess.check_call(['wget', '--quiet', '--retry-connrefused', '--waitretry=1', '--tries=10', 'http://localhost:2000', '--output-document=/dev/null'])
except subprocess.CalledProcessError:
raise subprocess.CalledProcessError("Could not reach OpenFisca Web API at localhost:2000 after 10s")
class TestNewApi(TestCase):
def setUp(self):
self.process = subprocess.Popen(['openfisca', 'serve'])
def tearDown(self):
self.process.terminate()
def test_response(self):
try:
subprocess.check_call(['wget', '--quiet', '--retry-connrefused', '--waitretry=1', '--tries=10', 'http://localhost:6000/parameters', '--output-document=/dev/null'])
except subprocess.CalledProcessError:
raise subprocess.CalledProcessError("Could not reach OpenFisca Web API at localhost:6000 after 10s")
|
53f2e3e5b58b001743bdedb479697150a9205b3f
|
buffpy/tests/test_profiles_manager.py
|
buffpy/tests/test_profiles_manager.py
|
from nose.tools import eq_
from mock import MagicMock, patch
from buffpy.managers.profiles import Profiles
from buffpy.models.profile import Profile, PATHS
mocked_response = {
'name': 'me',
'service': 'twiter',
'id': 1
}
def test_profiles_manager_all_method():
'''
Test basic profiles retrieving
'''
mocked_api = MagicMock()
mocked_api.get.return_value = [{'a':'b'}]
with patch('buffpy.managers.profiles.Profile') as mocked_profile:
mocked_profile.return_value = 1
profiles = Profiles(api=mocked_api).all()
eq_(profiles, [1])
mocked_api.get.assert_called_once_with(url=PATHS['GET_PROFILES'])
mocked_profile.assert_called_once_with(mocked_api, {'a': 'b'})
def test_profiles_manager_filter_method():
'''
Test basic profiles filtering based on some minimal criteria
'''
mocked_api = MagicMock()
profiles = Profiles(mocked_api, [{'a':'b'}, {'a': 'c'}])
eq_(profiles.filter(a='b'), [{'a': 'b'}])
def test_profiles_manager_filter_method_empty():
'''
Test basic profiles filtering when the manager is empty
'''
mocked_api = MagicMock()
mocked_api.get.return_value = [{'a':'b'}, {'a': 'c'}]
profiles = Profiles(api=mocked_api)
eq_(profiles.filter(a='b'), [Profile(mocked_api, {'a': 'b'})])
|
from unittest.mock import MagicMock, patch
from buffpy.managers.profiles import Profiles
from buffpy.models.profile import Profile, PATHS
MOCKED_RESPONSE = {
"name": "me",
"service": "twiter",
"id": 1
}
def test_profiles_manager_all_method():
""" Should retrieve profile info. """
mocked_api = MagicMock()
mocked_api.get.return_value = [{"a": "b"}]
with patch("buffpy.managers.profiles.Profile", return_value=1) as mocked_profile:
profiles = Profiles(api=mocked_api).all()
assert profiles == [1]
mocked_api.get.assert_called_once_with(url=PATHS["GET_PROFILES"])
mocked_profile.assert_called_once_with(mocked_api, {"a": "b"})
def test_profiles_manager_filter_method():
""" Should filter based on criteria. """
mocked_api = MagicMock()
profiles = Profiles(mocked_api, [{"a": "b"}, {"a": "c"}])
assert profiles.filter(a="b") == [{"a": "b"}]
def test_profiles_manager_filter_method_empty():
""" Should filter if profile manager is None. """
mocked_api = MagicMock()
mocked_api.get.return_value = [{"a": "b"}, {"a": "c"}]
profiles = Profiles(api=mocked_api)
assert profiles.filter(a="b") == [Profile(mocked_api, {"a": "b"})]
|
Migrate profiles manager tests to pytest
|
Migrate profiles manager tests to pytest
|
Python
|
mit
|
vtemian/buffpy
|
from nose.tools import eq_
from mock import MagicMock, patch
from buffpy.managers.profiles import Profiles
from buffpy.models.profile import Profile, PATHS
mocked_response = {
'name': 'me',
'service': 'twiter',
'id': 1
}
def test_profiles_manager_all_method():
'''
Test basic profiles retrieving
'''
mocked_api = MagicMock()
mocked_api.get.return_value = [{'a':'b'}]
with patch('buffpy.managers.profiles.Profile') as mocked_profile:
mocked_profile.return_value = 1
profiles = Profiles(api=mocked_api).all()
eq_(profiles, [1])
mocked_api.get.assert_called_once_with(url=PATHS['GET_PROFILES'])
mocked_profile.assert_called_once_with(mocked_api, {'a': 'b'})
def test_profiles_manager_filter_method():
'''
Test basic profiles filtering based on some minimal criteria
'''
mocked_api = MagicMock()
profiles = Profiles(mocked_api, [{'a':'b'}, {'a': 'c'}])
eq_(profiles.filter(a='b'), [{'a': 'b'}])
def test_profiles_manager_filter_method_empty():
'''
Test basic profiles filtering when the manager is empty
'''
mocked_api = MagicMock()
mocked_api.get.return_value = [{'a':'b'}, {'a': 'c'}]
profiles = Profiles(api=mocked_api)
eq_(profiles.filter(a='b'), [Profile(mocked_api, {'a': 'b'})])
Migrate profiles manager tests to pytest
|
from unittest.mock import MagicMock, patch
from buffpy.managers.profiles import Profiles
from buffpy.models.profile import Profile, PATHS
MOCKED_RESPONSE = {
"name": "me",
"service": "twiter",
"id": 1
}
def test_profiles_manager_all_method():
""" Should retrieve profile info. """
mocked_api = MagicMock()
mocked_api.get.return_value = [{"a": "b"}]
with patch("buffpy.managers.profiles.Profile", return_value=1) as mocked_profile:
profiles = Profiles(api=mocked_api).all()
assert profiles == [1]
mocked_api.get.assert_called_once_with(url=PATHS["GET_PROFILES"])
mocked_profile.assert_called_once_with(mocked_api, {"a": "b"})
def test_profiles_manager_filter_method():
""" Should filter based on criteria. """
mocked_api = MagicMock()
profiles = Profiles(mocked_api, [{"a": "b"}, {"a": "c"}])
assert profiles.filter(a="b") == [{"a": "b"}]
def test_profiles_manager_filter_method_empty():
""" Should filter if profile manager is None. """
mocked_api = MagicMock()
mocked_api.get.return_value = [{"a": "b"}, {"a": "c"}]
profiles = Profiles(api=mocked_api)
assert profiles.filter(a="b") == [Profile(mocked_api, {"a": "b"})]
|
<commit_before>from nose.tools import eq_
from mock import MagicMock, patch
from buffpy.managers.profiles import Profiles
from buffpy.models.profile import Profile, PATHS
mocked_response = {
'name': 'me',
'service': 'twiter',
'id': 1
}
def test_profiles_manager_all_method():
'''
Test basic profiles retrieving
'''
mocked_api = MagicMock()
mocked_api.get.return_value = [{'a':'b'}]
with patch('buffpy.managers.profiles.Profile') as mocked_profile:
mocked_profile.return_value = 1
profiles = Profiles(api=mocked_api).all()
eq_(profiles, [1])
mocked_api.get.assert_called_once_with(url=PATHS['GET_PROFILES'])
mocked_profile.assert_called_once_with(mocked_api, {'a': 'b'})
def test_profiles_manager_filter_method():
'''
Test basic profiles filtering based on some minimal criteria
'''
mocked_api = MagicMock()
profiles = Profiles(mocked_api, [{'a':'b'}, {'a': 'c'}])
eq_(profiles.filter(a='b'), [{'a': 'b'}])
def test_profiles_manager_filter_method_empty():
'''
Test basic profiles filtering when the manager is empty
'''
mocked_api = MagicMock()
mocked_api.get.return_value = [{'a':'b'}, {'a': 'c'}]
profiles = Profiles(api=mocked_api)
eq_(profiles.filter(a='b'), [Profile(mocked_api, {'a': 'b'})])
<commit_msg>Migrate profiles manager tests to pytest<commit_after>
|
from unittest.mock import MagicMock, patch
from buffpy.managers.profiles import Profiles
from buffpy.models.profile import Profile, PATHS
MOCKED_RESPONSE = {
"name": "me",
"service": "twiter",
"id": 1
}
def test_profiles_manager_all_method():
""" Should retrieve profile info. """
mocked_api = MagicMock()
mocked_api.get.return_value = [{"a": "b"}]
with patch("buffpy.managers.profiles.Profile", return_value=1) as mocked_profile:
profiles = Profiles(api=mocked_api).all()
assert profiles == [1]
mocked_api.get.assert_called_once_with(url=PATHS["GET_PROFILES"])
mocked_profile.assert_called_once_with(mocked_api, {"a": "b"})
def test_profiles_manager_filter_method():
""" Should filter based on criteria. """
mocked_api = MagicMock()
profiles = Profiles(mocked_api, [{"a": "b"}, {"a": "c"}])
assert profiles.filter(a="b") == [{"a": "b"}]
def test_profiles_manager_filter_method_empty():
""" Should filter if profile manager is None. """
mocked_api = MagicMock()
mocked_api.get.return_value = [{"a": "b"}, {"a": "c"}]
profiles = Profiles(api=mocked_api)
assert profiles.filter(a="b") == [Profile(mocked_api, {"a": "b"})]
|
from nose.tools import eq_
from mock import MagicMock, patch
from buffpy.managers.profiles import Profiles
from buffpy.models.profile import Profile, PATHS
mocked_response = {
'name': 'me',
'service': 'twiter',
'id': 1
}
def test_profiles_manager_all_method():
'''
Test basic profiles retrieving
'''
mocked_api = MagicMock()
mocked_api.get.return_value = [{'a':'b'}]
with patch('buffpy.managers.profiles.Profile') as mocked_profile:
mocked_profile.return_value = 1
profiles = Profiles(api=mocked_api).all()
eq_(profiles, [1])
mocked_api.get.assert_called_once_with(url=PATHS['GET_PROFILES'])
mocked_profile.assert_called_once_with(mocked_api, {'a': 'b'})
def test_profiles_manager_filter_method():
'''
Test basic profiles filtering based on some minimal criteria
'''
mocked_api = MagicMock()
profiles = Profiles(mocked_api, [{'a':'b'}, {'a': 'c'}])
eq_(profiles.filter(a='b'), [{'a': 'b'}])
def test_profiles_manager_filter_method_empty():
'''
Test basic profiles filtering when the manager is empty
'''
mocked_api = MagicMock()
mocked_api.get.return_value = [{'a':'b'}, {'a': 'c'}]
profiles = Profiles(api=mocked_api)
eq_(profiles.filter(a='b'), [Profile(mocked_api, {'a': 'b'})])
Migrate profiles manager tests to pytestfrom unittest.mock import MagicMock, patch
from buffpy.managers.profiles import Profiles
from buffpy.models.profile import Profile, PATHS
MOCKED_RESPONSE = {
"name": "me",
"service": "twiter",
"id": 1
}
def test_profiles_manager_all_method():
""" Should retrieve profile info. """
mocked_api = MagicMock()
mocked_api.get.return_value = [{"a": "b"}]
with patch("buffpy.managers.profiles.Profile", return_value=1) as mocked_profile:
profiles = Profiles(api=mocked_api).all()
assert profiles == [1]
mocked_api.get.assert_called_once_with(url=PATHS["GET_PROFILES"])
mocked_profile.assert_called_once_with(mocked_api, {"a": "b"})
def test_profiles_manager_filter_method():
""" Should filter based on criteria. """
mocked_api = MagicMock()
profiles = Profiles(mocked_api, [{"a": "b"}, {"a": "c"}])
assert profiles.filter(a="b") == [{"a": "b"}]
def test_profiles_manager_filter_method_empty():
""" Should filter if profile manager is None. """
mocked_api = MagicMock()
mocked_api.get.return_value = [{"a": "b"}, {"a": "c"}]
profiles = Profiles(api=mocked_api)
assert profiles.filter(a="b") == [Profile(mocked_api, {"a": "b"})]
|
<commit_before>from nose.tools import eq_
from mock import MagicMock, patch
from buffpy.managers.profiles import Profiles
from buffpy.models.profile import Profile, PATHS
mocked_response = {
'name': 'me',
'service': 'twiter',
'id': 1
}
def test_profiles_manager_all_method():
'''
Test basic profiles retrieving
'''
mocked_api = MagicMock()
mocked_api.get.return_value = [{'a':'b'}]
with patch('buffpy.managers.profiles.Profile') as mocked_profile:
mocked_profile.return_value = 1
profiles = Profiles(api=mocked_api).all()
eq_(profiles, [1])
mocked_api.get.assert_called_once_with(url=PATHS['GET_PROFILES'])
mocked_profile.assert_called_once_with(mocked_api, {'a': 'b'})
def test_profiles_manager_filter_method():
'''
Test basic profiles filtering based on some minimal criteria
'''
mocked_api = MagicMock()
profiles = Profiles(mocked_api, [{'a':'b'}, {'a': 'c'}])
eq_(profiles.filter(a='b'), [{'a': 'b'}])
def test_profiles_manager_filter_method_empty():
'''
Test basic profiles filtering when the manager is empty
'''
mocked_api = MagicMock()
mocked_api.get.return_value = [{'a':'b'}, {'a': 'c'}]
profiles = Profiles(api=mocked_api)
eq_(profiles.filter(a='b'), [Profile(mocked_api, {'a': 'b'})])
<commit_msg>Migrate profiles manager tests to pytest<commit_after>from unittest.mock import MagicMock, patch
from buffpy.managers.profiles import Profiles
from buffpy.models.profile import Profile, PATHS
MOCKED_RESPONSE = {
"name": "me",
"service": "twiter",
"id": 1
}
def test_profiles_manager_all_method():
""" Should retrieve profile info. """
mocked_api = MagicMock()
mocked_api.get.return_value = [{"a": "b"}]
with patch("buffpy.managers.profiles.Profile", return_value=1) as mocked_profile:
profiles = Profiles(api=mocked_api).all()
assert profiles == [1]
mocked_api.get.assert_called_once_with(url=PATHS["GET_PROFILES"])
mocked_profile.assert_called_once_with(mocked_api, {"a": "b"})
def test_profiles_manager_filter_method():
""" Should filter based on criteria. """
mocked_api = MagicMock()
profiles = Profiles(mocked_api, [{"a": "b"}, {"a": "c"}])
assert profiles.filter(a="b") == [{"a": "b"}]
def test_profiles_manager_filter_method_empty():
""" Should filter if profile manager is None. """
mocked_api = MagicMock()
mocked_api.get.return_value = [{"a": "b"}, {"a": "c"}]
profiles = Profiles(api=mocked_api)
assert profiles.filter(a="b") == [Profile(mocked_api, {"a": "b"})]
|
bbc0b9cd9244079c14914763e60ec4ca9eb41b4e
|
byceps/blueprints/admin/site/forms.py
|
byceps/blueprints/admin/site/forms.py
|
"""
byceps.blueprints.admin.site.forms
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2019 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from wtforms import SelectField, StringField
from wtforms.validators import InputRequired, Length, Optional
from ....util.l10n import LocalizedForm
from ....services.email import service as email_service
from ....services.party import service as party_service
class UpdateForm(LocalizedForm):
title = StringField('Titel', validators=[Length(min=1, max=20)])
server_name = StringField('Servername', validators=[InputRequired()])
email_config_id = SelectField('E-Mail-Konfiguration', validators=[InputRequired()])
party_id = SelectField('Party-ID', validators=[Optional()])
def set_email_config_choices(self):
configs = email_service.get_all_configs()
configs.sort(key=lambda config: config.id)
self.email_config_id.choices = [(c.id, c.id) for c in configs]
def set_party_choices(self):
parties = party_service.get_all_parties()
parties.sort(key=lambda party: party.id)
choices = [(str(p.id), p.title) for p in parties]
choices.insert(0, ('', '<keine>'))
self.party_id.choices = choices
class CreateForm(UpdateForm):
id = StringField('ID', validators=[Length(min=1, max=40)])
|
"""
byceps.blueprints.admin.site.forms
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2019 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from wtforms import SelectField, StringField
from wtforms.validators import InputRequired, Length, Optional
from ....util.l10n import LocalizedForm
from ....services.email import service as email_service
from ....services.party import service as party_service
class UpdateForm(LocalizedForm):
title = StringField('Titel', validators=[Length(min=1, max=40)])
server_name = StringField('Servername', validators=[InputRequired()])
email_config_id = SelectField('E-Mail-Konfiguration', validators=[InputRequired()])
party_id = SelectField('Party-ID', validators=[Optional()])
def set_email_config_choices(self):
configs = email_service.get_all_configs()
configs.sort(key=lambda config: config.id)
self.email_config_id.choices = [(c.id, c.id) for c in configs]
def set_party_choices(self):
parties = party_service.get_all_parties()
parties.sort(key=lambda party: party.id)
choices = [(str(p.id), p.title) for p in parties]
choices.insert(0, ('', '<keine>'))
self.party_id.choices = choices
class CreateForm(UpdateForm):
id = StringField('ID', validators=[Length(min=1, max=40)])
|
Allow site names to be up to 40 chars long (instead of 20)
|
Allow site names to be up to 40 chars long (instead of 20)
|
Python
|
bsd-3-clause
|
homeworkprod/byceps,m-ober/byceps,homeworkprod/byceps,m-ober/byceps,m-ober/byceps,homeworkprod/byceps
|
"""
byceps.blueprints.admin.site.forms
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2019 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from wtforms import SelectField, StringField
from wtforms.validators import InputRequired, Length, Optional
from ....util.l10n import LocalizedForm
from ....services.email import service as email_service
from ....services.party import service as party_service
class UpdateForm(LocalizedForm):
title = StringField('Titel', validators=[Length(min=1, max=20)])
server_name = StringField('Servername', validators=[InputRequired()])
email_config_id = SelectField('E-Mail-Konfiguration', validators=[InputRequired()])
party_id = SelectField('Party-ID', validators=[Optional()])
def set_email_config_choices(self):
configs = email_service.get_all_configs()
configs.sort(key=lambda config: config.id)
self.email_config_id.choices = [(c.id, c.id) for c in configs]
def set_party_choices(self):
parties = party_service.get_all_parties()
parties.sort(key=lambda party: party.id)
choices = [(str(p.id), p.title) for p in parties]
choices.insert(0, ('', '<keine>'))
self.party_id.choices = choices
class CreateForm(UpdateForm):
id = StringField('ID', validators=[Length(min=1, max=40)])
Allow site names to be up to 40 chars long (instead of 20)
|
"""
byceps.blueprints.admin.site.forms
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2019 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from wtforms import SelectField, StringField
from wtforms.validators import InputRequired, Length, Optional
from ....util.l10n import LocalizedForm
from ....services.email import service as email_service
from ....services.party import service as party_service
class UpdateForm(LocalizedForm):
title = StringField('Titel', validators=[Length(min=1, max=40)])
server_name = StringField('Servername', validators=[InputRequired()])
email_config_id = SelectField('E-Mail-Konfiguration', validators=[InputRequired()])
party_id = SelectField('Party-ID', validators=[Optional()])
def set_email_config_choices(self):
configs = email_service.get_all_configs()
configs.sort(key=lambda config: config.id)
self.email_config_id.choices = [(c.id, c.id) for c in configs]
def set_party_choices(self):
parties = party_service.get_all_parties()
parties.sort(key=lambda party: party.id)
choices = [(str(p.id), p.title) for p in parties]
choices.insert(0, ('', '<keine>'))
self.party_id.choices = choices
class CreateForm(UpdateForm):
id = StringField('ID', validators=[Length(min=1, max=40)])
|
<commit_before>"""
byceps.blueprints.admin.site.forms
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2019 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from wtforms import SelectField, StringField
from wtforms.validators import InputRequired, Length, Optional
from ....util.l10n import LocalizedForm
from ....services.email import service as email_service
from ....services.party import service as party_service
class UpdateForm(LocalizedForm):
title = StringField('Titel', validators=[Length(min=1, max=20)])
server_name = StringField('Servername', validators=[InputRequired()])
email_config_id = SelectField('E-Mail-Konfiguration', validators=[InputRequired()])
party_id = SelectField('Party-ID', validators=[Optional()])
def set_email_config_choices(self):
configs = email_service.get_all_configs()
configs.sort(key=lambda config: config.id)
self.email_config_id.choices = [(c.id, c.id) for c in configs]
def set_party_choices(self):
parties = party_service.get_all_parties()
parties.sort(key=lambda party: party.id)
choices = [(str(p.id), p.title) for p in parties]
choices.insert(0, ('', '<keine>'))
self.party_id.choices = choices
class CreateForm(UpdateForm):
id = StringField('ID', validators=[Length(min=1, max=40)])
<commit_msg>Allow site names to be up to 40 chars long (instead of 20)<commit_after>
|
"""
byceps.blueprints.admin.site.forms
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2019 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from wtforms import SelectField, StringField
from wtforms.validators import InputRequired, Length, Optional
from ....util.l10n import LocalizedForm
from ....services.email import service as email_service
from ....services.party import service as party_service
class UpdateForm(LocalizedForm):
title = StringField('Titel', validators=[Length(min=1, max=40)])
server_name = StringField('Servername', validators=[InputRequired()])
email_config_id = SelectField('E-Mail-Konfiguration', validators=[InputRequired()])
party_id = SelectField('Party-ID', validators=[Optional()])
def set_email_config_choices(self):
configs = email_service.get_all_configs()
configs.sort(key=lambda config: config.id)
self.email_config_id.choices = [(c.id, c.id) for c in configs]
def set_party_choices(self):
parties = party_service.get_all_parties()
parties.sort(key=lambda party: party.id)
choices = [(str(p.id), p.title) for p in parties]
choices.insert(0, ('', '<keine>'))
self.party_id.choices = choices
class CreateForm(UpdateForm):
id = StringField('ID', validators=[Length(min=1, max=40)])
|
"""
byceps.blueprints.admin.site.forms
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2019 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from wtforms import SelectField, StringField
from wtforms.validators import InputRequired, Length, Optional
from ....util.l10n import LocalizedForm
from ....services.email import service as email_service
from ....services.party import service as party_service
class UpdateForm(LocalizedForm):
title = StringField('Titel', validators=[Length(min=1, max=20)])
server_name = StringField('Servername', validators=[InputRequired()])
email_config_id = SelectField('E-Mail-Konfiguration', validators=[InputRequired()])
party_id = SelectField('Party-ID', validators=[Optional()])
def set_email_config_choices(self):
configs = email_service.get_all_configs()
configs.sort(key=lambda config: config.id)
self.email_config_id.choices = [(c.id, c.id) for c in configs]
def set_party_choices(self):
parties = party_service.get_all_parties()
parties.sort(key=lambda party: party.id)
choices = [(str(p.id), p.title) for p in parties]
choices.insert(0, ('', '<keine>'))
self.party_id.choices = choices
class CreateForm(UpdateForm):
id = StringField('ID', validators=[Length(min=1, max=40)])
Allow site names to be up to 40 chars long (instead of 20)"""
byceps.blueprints.admin.site.forms
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2019 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from wtforms import SelectField, StringField
from wtforms.validators import InputRequired, Length, Optional
from ....util.l10n import LocalizedForm
from ....services.email import service as email_service
from ....services.party import service as party_service
class UpdateForm(LocalizedForm):
title = StringField('Titel', validators=[Length(min=1, max=40)])
server_name = StringField('Servername', validators=[InputRequired()])
email_config_id = SelectField('E-Mail-Konfiguration', validators=[InputRequired()])
party_id = SelectField('Party-ID', validators=[Optional()])
def set_email_config_choices(self):
configs = email_service.get_all_configs()
configs.sort(key=lambda config: config.id)
self.email_config_id.choices = [(c.id, c.id) for c in configs]
def set_party_choices(self):
parties = party_service.get_all_parties()
parties.sort(key=lambda party: party.id)
choices = [(str(p.id), p.title) for p in parties]
choices.insert(0, ('', '<keine>'))
self.party_id.choices = choices
class CreateForm(UpdateForm):
id = StringField('ID', validators=[Length(min=1, max=40)])
|
<commit_before>"""
byceps.blueprints.admin.site.forms
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2019 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from wtforms import SelectField, StringField
from wtforms.validators import InputRequired, Length, Optional
from ....util.l10n import LocalizedForm
from ....services.email import service as email_service
from ....services.party import service as party_service
class UpdateForm(LocalizedForm):
title = StringField('Titel', validators=[Length(min=1, max=20)])
server_name = StringField('Servername', validators=[InputRequired()])
email_config_id = SelectField('E-Mail-Konfiguration', validators=[InputRequired()])
party_id = SelectField('Party-ID', validators=[Optional()])
def set_email_config_choices(self):
configs = email_service.get_all_configs()
configs.sort(key=lambda config: config.id)
self.email_config_id.choices = [(c.id, c.id) for c in configs]
def set_party_choices(self):
parties = party_service.get_all_parties()
parties.sort(key=lambda party: party.id)
choices = [(str(p.id), p.title) for p in parties]
choices.insert(0, ('', '<keine>'))
self.party_id.choices = choices
class CreateForm(UpdateForm):
id = StringField('ID', validators=[Length(min=1, max=40)])
<commit_msg>Allow site names to be up to 40 chars long (instead of 20)<commit_after>"""
byceps.blueprints.admin.site.forms
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2019 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from wtforms import SelectField, StringField
from wtforms.validators import InputRequired, Length, Optional
from ....util.l10n import LocalizedForm
from ....services.email import service as email_service
from ....services.party import service as party_service
class UpdateForm(LocalizedForm):
title = StringField('Titel', validators=[Length(min=1, max=40)])
server_name = StringField('Servername', validators=[InputRequired()])
email_config_id = SelectField('E-Mail-Konfiguration', validators=[InputRequired()])
party_id = SelectField('Party-ID', validators=[Optional()])
def set_email_config_choices(self):
configs = email_service.get_all_configs()
configs.sort(key=lambda config: config.id)
self.email_config_id.choices = [(c.id, c.id) for c in configs]
def set_party_choices(self):
parties = party_service.get_all_parties()
parties.sort(key=lambda party: party.id)
choices = [(str(p.id), p.title) for p in parties]
choices.insert(0, ('', '<keine>'))
self.party_id.choices = choices
class CreateForm(UpdateForm):
id = StringField('ID', validators=[Length(min=1, max=40)])
|
1ab04355c0172682e9948847a01b073239d0ae64
|
words.py
|
words.py
|
"""Function to fetch words."""
import random
WORDLIST = 'wordlist.txt'
def get_random_word():
"""Get a random word from the wordlist."""
words = []
with open(WORDLIST, 'r') as f:
for word in f:
words.append(word)
return random.choice(words)
def get_random_word_scalable():
"""Get a random word from the wordlist using no extra memory."""
num_words_processed = 0
curr_word = None
with open(WORDLIST, 'r') as f:
for word in f:
num_words_processed += 1
if random.randint(1, num_words_processed) == 1:
curr_word = word
return curr_word
|
"""Function to fetch words."""
import random
WORDLIST = 'wordlist.txt'
def get_random_word(min_word_length):
"""Get a random word from the wordlist using no extra memory."""
num_words_processed = 0
curr_word = None
with open(WORDLIST, 'r') as f:
for word in f:
if len(word) < min_word_length:
continue
num_words_processed += 1
if random.randint(1, num_words_processed) == 1:
curr_word = word
return curr_word
|
Use scalable get_word by default
|
Use scalable get_word by default
|
Python
|
mit
|
andrewyang96/HangmanGame
|
"""Function to fetch words."""
import random
WORDLIST = 'wordlist.txt'
def get_random_word():
"""Get a random word from the wordlist."""
words = []
with open(WORDLIST, 'r') as f:
for word in f:
words.append(word)
return random.choice(words)
def get_random_word_scalable():
"""Get a random word from the wordlist using no extra memory."""
num_words_processed = 0
curr_word = None
with open(WORDLIST, 'r') as f:
for word in f:
num_words_processed += 1
if random.randint(1, num_words_processed) == 1:
curr_word = word
return curr_word
Use scalable get_word by default
|
"""Function to fetch words."""
import random
WORDLIST = 'wordlist.txt'
def get_random_word(min_word_length):
"""Get a random word from the wordlist using no extra memory."""
num_words_processed = 0
curr_word = None
with open(WORDLIST, 'r') as f:
for word in f:
if len(word) < min_word_length:
continue
num_words_processed += 1
if random.randint(1, num_words_processed) == 1:
curr_word = word
return curr_word
|
<commit_before>"""Function to fetch words."""
import random
WORDLIST = 'wordlist.txt'
def get_random_word():
"""Get a random word from the wordlist."""
words = []
with open(WORDLIST, 'r') as f:
for word in f:
words.append(word)
return random.choice(words)
def get_random_word_scalable():
"""Get a random word from the wordlist using no extra memory."""
num_words_processed = 0
curr_word = None
with open(WORDLIST, 'r') as f:
for word in f:
num_words_processed += 1
if random.randint(1, num_words_processed) == 1:
curr_word = word
return curr_word
<commit_msg>Use scalable get_word by default<commit_after>
|
"""Function to fetch words."""
import random
WORDLIST = 'wordlist.txt'
def get_random_word(min_word_length):
"""Get a random word from the wordlist using no extra memory."""
num_words_processed = 0
curr_word = None
with open(WORDLIST, 'r') as f:
for word in f:
if len(word) < min_word_length:
continue
num_words_processed += 1
if random.randint(1, num_words_processed) == 1:
curr_word = word
return curr_word
|
"""Function to fetch words."""
import random
WORDLIST = 'wordlist.txt'
def get_random_word():
"""Get a random word from the wordlist."""
words = []
with open(WORDLIST, 'r') as f:
for word in f:
words.append(word)
return random.choice(words)
def get_random_word_scalable():
"""Get a random word from the wordlist using no extra memory."""
num_words_processed = 0
curr_word = None
with open(WORDLIST, 'r') as f:
for word in f:
num_words_processed += 1
if random.randint(1, num_words_processed) == 1:
curr_word = word
return curr_word
Use scalable get_word by default"""Function to fetch words."""
import random
WORDLIST = 'wordlist.txt'
def get_random_word(min_word_length):
"""Get a random word from the wordlist using no extra memory."""
num_words_processed = 0
curr_word = None
with open(WORDLIST, 'r') as f:
for word in f:
if len(word) < min_word_length:
continue
num_words_processed += 1
if random.randint(1, num_words_processed) == 1:
curr_word = word
return curr_word
|
<commit_before>"""Function to fetch words."""
import random
WORDLIST = 'wordlist.txt'
def get_random_word():
"""Get a random word from the wordlist."""
words = []
with open(WORDLIST, 'r') as f:
for word in f:
words.append(word)
return random.choice(words)
def get_random_word_scalable():
"""Get a random word from the wordlist using no extra memory."""
num_words_processed = 0
curr_word = None
with open(WORDLIST, 'r') as f:
for word in f:
num_words_processed += 1
if random.randint(1, num_words_processed) == 1:
curr_word = word
return curr_word
<commit_msg>Use scalable get_word by default<commit_after>"""Function to fetch words."""
import random
WORDLIST = 'wordlist.txt'
def get_random_word(min_word_length):
"""Get a random word from the wordlist using no extra memory."""
num_words_processed = 0
curr_word = None
with open(WORDLIST, 'r') as f:
for word in f:
if len(word) < min_word_length:
continue
num_words_processed += 1
if random.randint(1, num_words_processed) == 1:
curr_word = word
return curr_word
|
ba4a68871ee326de94203bf401e4d325b87bec9c
|
docs/conf.py
|
docs/conf.py
|
import pymanopt
# Package information
project = "Pymanopt"
author = "Jamie Townsend, Niklas Koep, Sebastian Weichwald"
copyright = "2016-2020, {:s}".format(author)
release = version = pymanopt.__version__
# Build settings
extensions = [
"sphinx.ext.autodoc",
"sphinx.ext.coverage",
"sphinx.ext.mathjax",
"sphinx.ext.viewcode"
]
source_suffix = ".rst"
master_doc = "index"
language = None
exclude_patterns = ["build", "*.egg*"]
# Output options
html_theme = "sphinx_rtd_theme"
html_show_sphinx = False
html_baseurl = "www.pymanopt.org"
htmlhelp_basename = "pymanoptdoc"
html_last_updated_fmt = ""
# autodoc
autodoc_default_options = {
"member-order": "bysource",
"members": True,
"undoc-members": True,
"show-inheritance": True
}
|
import pymanopt
# Package information
project = "Pymanopt"
author = "Jamie Townsend, Niklas Koep, Sebastian Weichwald"
copyright = "2016-2021, {:s}".format(author)
release = version = pymanopt.__version__
# Build settings
extensions = [
"sphinx.ext.autodoc",
"sphinx.ext.coverage",
"sphinx.ext.mathjax",
"sphinx.ext.viewcode"
]
source_suffix = ".rst"
master_doc = "index"
language = None
exclude_patterns = ["build", "*.egg*"]
# Output options
html_theme = "sphinx_rtd_theme"
html_show_sphinx = False
html_baseurl = "www.pymanopt.org"
htmlhelp_basename = "pymanoptdoc"
html_last_updated_fmt = ""
# autodoc
autodoc_default_options = {
"member-order": "bysource",
"members": True,
"undoc-members": True,
"show-inheritance": True
}
|
Update copyright string in docs
|
Update copyright string in docs
Signed-off-by: Niklas Koep <342d5290239d9c5264c8f98185afedb99596601a@gmail.com>
|
Python
|
bsd-3-clause
|
pymanopt/pymanopt,pymanopt/pymanopt
|
import pymanopt
# Package information
project = "Pymanopt"
author = "Jamie Townsend, Niklas Koep, Sebastian Weichwald"
copyright = "2016-2020, {:s}".format(author)
release = version = pymanopt.__version__
# Build settings
extensions = [
"sphinx.ext.autodoc",
"sphinx.ext.coverage",
"sphinx.ext.mathjax",
"sphinx.ext.viewcode"
]
source_suffix = ".rst"
master_doc = "index"
language = None
exclude_patterns = ["build", "*.egg*"]
# Output options
html_theme = "sphinx_rtd_theme"
html_show_sphinx = False
html_baseurl = "www.pymanopt.org"
htmlhelp_basename = "pymanoptdoc"
html_last_updated_fmt = ""
# autodoc
autodoc_default_options = {
"member-order": "bysource",
"members": True,
"undoc-members": True,
"show-inheritance": True
}
Update copyright string in docs
Signed-off-by: Niklas Koep <342d5290239d9c5264c8f98185afedb99596601a@gmail.com>
|
import pymanopt
# Package information
project = "Pymanopt"
author = "Jamie Townsend, Niklas Koep, Sebastian Weichwald"
copyright = "2016-2021, {:s}".format(author)
release = version = pymanopt.__version__
# Build settings
extensions = [
"sphinx.ext.autodoc",
"sphinx.ext.coverage",
"sphinx.ext.mathjax",
"sphinx.ext.viewcode"
]
source_suffix = ".rst"
master_doc = "index"
language = None
exclude_patterns = ["build", "*.egg*"]
# Output options
html_theme = "sphinx_rtd_theme"
html_show_sphinx = False
html_baseurl = "www.pymanopt.org"
htmlhelp_basename = "pymanoptdoc"
html_last_updated_fmt = ""
# autodoc
autodoc_default_options = {
"member-order": "bysource",
"members": True,
"undoc-members": True,
"show-inheritance": True
}
|
<commit_before>import pymanopt
# Package information
project = "Pymanopt"
author = "Jamie Townsend, Niklas Koep, Sebastian Weichwald"
copyright = "2016-2020, {:s}".format(author)
release = version = pymanopt.__version__
# Build settings
extensions = [
"sphinx.ext.autodoc",
"sphinx.ext.coverage",
"sphinx.ext.mathjax",
"sphinx.ext.viewcode"
]
source_suffix = ".rst"
master_doc = "index"
language = None
exclude_patterns = ["build", "*.egg*"]
# Output options
html_theme = "sphinx_rtd_theme"
html_show_sphinx = False
html_baseurl = "www.pymanopt.org"
htmlhelp_basename = "pymanoptdoc"
html_last_updated_fmt = ""
# autodoc
autodoc_default_options = {
"member-order": "bysource",
"members": True,
"undoc-members": True,
"show-inheritance": True
}
<commit_msg>Update copyright string in docs
Signed-off-by: Niklas Koep <342d5290239d9c5264c8f98185afedb99596601a@gmail.com><commit_after>
|
import pymanopt
# Package information
project = "Pymanopt"
author = "Jamie Townsend, Niklas Koep, Sebastian Weichwald"
copyright = "2016-2021, {:s}".format(author)
release = version = pymanopt.__version__
# Build settings
extensions = [
"sphinx.ext.autodoc",
"sphinx.ext.coverage",
"sphinx.ext.mathjax",
"sphinx.ext.viewcode"
]
source_suffix = ".rst"
master_doc = "index"
language = None
exclude_patterns = ["build", "*.egg*"]
# Output options
html_theme = "sphinx_rtd_theme"
html_show_sphinx = False
html_baseurl = "www.pymanopt.org"
htmlhelp_basename = "pymanoptdoc"
html_last_updated_fmt = ""
# autodoc
autodoc_default_options = {
"member-order": "bysource",
"members": True,
"undoc-members": True,
"show-inheritance": True
}
|
import pymanopt
# Package information
project = "Pymanopt"
author = "Jamie Townsend, Niklas Koep, Sebastian Weichwald"
copyright = "2016-2020, {:s}".format(author)
release = version = pymanopt.__version__
# Build settings
extensions = [
"sphinx.ext.autodoc",
"sphinx.ext.coverage",
"sphinx.ext.mathjax",
"sphinx.ext.viewcode"
]
source_suffix = ".rst"
master_doc = "index"
language = None
exclude_patterns = ["build", "*.egg*"]
# Output options
html_theme = "sphinx_rtd_theme"
html_show_sphinx = False
html_baseurl = "www.pymanopt.org"
htmlhelp_basename = "pymanoptdoc"
html_last_updated_fmt = ""
# autodoc
autodoc_default_options = {
"member-order": "bysource",
"members": True,
"undoc-members": True,
"show-inheritance": True
}
Update copyright string in docs
Signed-off-by: Niklas Koep <342d5290239d9c5264c8f98185afedb99596601a@gmail.com>import pymanopt
# Package information
project = "Pymanopt"
author = "Jamie Townsend, Niklas Koep, Sebastian Weichwald"
copyright = "2016-2021, {:s}".format(author)
release = version = pymanopt.__version__
# Build settings
extensions = [
"sphinx.ext.autodoc",
"sphinx.ext.coverage",
"sphinx.ext.mathjax",
"sphinx.ext.viewcode"
]
source_suffix = ".rst"
master_doc = "index"
language = None
exclude_patterns = ["build", "*.egg*"]
# Output options
html_theme = "sphinx_rtd_theme"
html_show_sphinx = False
html_baseurl = "www.pymanopt.org"
htmlhelp_basename = "pymanoptdoc"
html_last_updated_fmt = ""
# autodoc
autodoc_default_options = {
"member-order": "bysource",
"members": True,
"undoc-members": True,
"show-inheritance": True
}
|
<commit_before>import pymanopt
# Package information
project = "Pymanopt"
author = "Jamie Townsend, Niklas Koep, Sebastian Weichwald"
copyright = "2016-2020, {:s}".format(author)
release = version = pymanopt.__version__
# Build settings
extensions = [
"sphinx.ext.autodoc",
"sphinx.ext.coverage",
"sphinx.ext.mathjax",
"sphinx.ext.viewcode"
]
source_suffix = ".rst"
master_doc = "index"
language = None
exclude_patterns = ["build", "*.egg*"]
# Output options
html_theme = "sphinx_rtd_theme"
html_show_sphinx = False
html_baseurl = "www.pymanopt.org"
htmlhelp_basename = "pymanoptdoc"
html_last_updated_fmt = ""
# autodoc
autodoc_default_options = {
"member-order": "bysource",
"members": True,
"undoc-members": True,
"show-inheritance": True
}
<commit_msg>Update copyright string in docs
Signed-off-by: Niklas Koep <342d5290239d9c5264c8f98185afedb99596601a@gmail.com><commit_after>import pymanopt
# Package information
project = "Pymanopt"
author = "Jamie Townsend, Niklas Koep, Sebastian Weichwald"
copyright = "2016-2021, {:s}".format(author)
release = version = pymanopt.__version__
# Build settings
extensions = [
"sphinx.ext.autodoc",
"sphinx.ext.coverage",
"sphinx.ext.mathjax",
"sphinx.ext.viewcode"
]
source_suffix = ".rst"
master_doc = "index"
language = None
exclude_patterns = ["build", "*.egg*"]
# Output options
html_theme = "sphinx_rtd_theme"
html_show_sphinx = False
html_baseurl = "www.pymanopt.org"
htmlhelp_basename = "pymanoptdoc"
html_last_updated_fmt = ""
# autodoc
autodoc_default_options = {
"member-order": "bysource",
"members": True,
"undoc-members": True,
"show-inheritance": True
}
|
e051ae3bdada17f31eb1c4ed68bcd41e6e20deab
|
cea/interfaces/dashboard/api/dashboard.py
|
cea/interfaces/dashboard/api/dashboard.py
|
from flask_restplus import Namespace, Resource, fields, abort
import cea.config
import cea.plots.cache
api = Namespace('Dashboard', description='Dashboard plots')
LAYOUTS = ['row', 'grid', 'map']
CATEGORIES = {c.name: {'label': c.label, 'plots': [{'id': p.id(), 'name': p.name} for p in c.plots]}
for c in cea.plots.categories.list_categories()}
@api.route('/')
class Dashboard(Resource):
def get(self):
"""
Get Dashboards from yaml file
"""
config = cea.config.Configuration()
plot_cache = cea.plots.cache.PlotCache(config)
dashboards = cea.plots.read_dashboards(config, plot_cache)
return [{'name': d.name, 'description': d.description, 'layout': d.layout if d.layout in LAYOUTS else 'row',
'plots': [{'title': plot.title, 'scenario':
plot.parameters['scenario-name'] if 'scenario-name' in plot.parameters.keys() else None}
for plot in d.plots]} for d in dashboards]
|
from flask_restplus import Namespace, Resource, fields, abort
import cea.config
import cea.plots.cache
api = Namespace('Dashboard', description='Dashboard plots')
LAYOUTS = ['row', 'grid', 'map']
CATEGORIES = {c.name: {'label': c.label, 'plots': [{'id': p.id(), 'name': p.name} for p in c.plots]}
for c in cea.plots.categories.list_categories()}
@api.route('/')
class Dashboard(Resource):
def get(self):
"""
Get Dashboards from yaml file
"""
config = cea.config.Configuration()
plot_cache = cea.plots.cache.PlotCache(config)
dashboards = cea.plots.read_dashboards(config, plot_cache)
out = []
for d in dashboards:
dashboard = d.to_dict()
for i, plot in enumerate(dashboard['plots']):
dashboard['plots'][i]['title'] = d.plots[i].title
out.append(dashboard)
return out
|
Include plot title to plots
|
Include plot title to plots
|
Python
|
mit
|
architecture-building-systems/CEAforArcGIS,architecture-building-systems/CEAforArcGIS
|
from flask_restplus import Namespace, Resource, fields, abort
import cea.config
import cea.plots.cache
api = Namespace('Dashboard', description='Dashboard plots')
LAYOUTS = ['row', 'grid', 'map']
CATEGORIES = {c.name: {'label': c.label, 'plots': [{'id': p.id(), 'name': p.name} for p in c.plots]}
for c in cea.plots.categories.list_categories()}
@api.route('/')
class Dashboard(Resource):
def get(self):
"""
Get Dashboards from yaml file
"""
config = cea.config.Configuration()
plot_cache = cea.plots.cache.PlotCache(config)
dashboards = cea.plots.read_dashboards(config, plot_cache)
return [{'name': d.name, 'description': d.description, 'layout': d.layout if d.layout in LAYOUTS else 'row',
'plots': [{'title': plot.title, 'scenario':
plot.parameters['scenario-name'] if 'scenario-name' in plot.parameters.keys() else None}
for plot in d.plots]} for d in dashboards]
Include plot title to plots
|
from flask_restplus import Namespace, Resource, fields, abort
import cea.config
import cea.plots.cache
api = Namespace('Dashboard', description='Dashboard plots')
LAYOUTS = ['row', 'grid', 'map']
CATEGORIES = {c.name: {'label': c.label, 'plots': [{'id': p.id(), 'name': p.name} for p in c.plots]}
for c in cea.plots.categories.list_categories()}
@api.route('/')
class Dashboard(Resource):
def get(self):
"""
Get Dashboards from yaml file
"""
config = cea.config.Configuration()
plot_cache = cea.plots.cache.PlotCache(config)
dashboards = cea.plots.read_dashboards(config, plot_cache)
out = []
for d in dashboards:
dashboard = d.to_dict()
for i, plot in enumerate(dashboard['plots']):
dashboard['plots'][i]['title'] = d.plots[i].title
out.append(dashboard)
return out
|
<commit_before>from flask_restplus import Namespace, Resource, fields, abort
import cea.config
import cea.plots.cache
api = Namespace('Dashboard', description='Dashboard plots')
LAYOUTS = ['row', 'grid', 'map']
CATEGORIES = {c.name: {'label': c.label, 'plots': [{'id': p.id(), 'name': p.name} for p in c.plots]}
for c in cea.plots.categories.list_categories()}
@api.route('/')
class Dashboard(Resource):
def get(self):
"""
Get Dashboards from yaml file
"""
config = cea.config.Configuration()
plot_cache = cea.plots.cache.PlotCache(config)
dashboards = cea.plots.read_dashboards(config, plot_cache)
return [{'name': d.name, 'description': d.description, 'layout': d.layout if d.layout in LAYOUTS else 'row',
'plots': [{'title': plot.title, 'scenario':
plot.parameters['scenario-name'] if 'scenario-name' in plot.parameters.keys() else None}
for plot in d.plots]} for d in dashboards]
<commit_msg>Include plot title to plots<commit_after>
|
from flask_restplus import Namespace, Resource, fields, abort
import cea.config
import cea.plots.cache
api = Namespace('Dashboard', description='Dashboard plots')
LAYOUTS = ['row', 'grid', 'map']
CATEGORIES = {c.name: {'label': c.label, 'plots': [{'id': p.id(), 'name': p.name} for p in c.plots]}
for c in cea.plots.categories.list_categories()}
@api.route('/')
class Dashboard(Resource):
def get(self):
"""
Get Dashboards from yaml file
"""
config = cea.config.Configuration()
plot_cache = cea.plots.cache.PlotCache(config)
dashboards = cea.plots.read_dashboards(config, plot_cache)
out = []
for d in dashboards:
dashboard = d.to_dict()
for i, plot in enumerate(dashboard['plots']):
dashboard['plots'][i]['title'] = d.plots[i].title
out.append(dashboard)
return out
|
from flask_restplus import Namespace, Resource, fields, abort
import cea.config
import cea.plots.cache
api = Namespace('Dashboard', description='Dashboard plots')
LAYOUTS = ['row', 'grid', 'map']
CATEGORIES = {c.name: {'label': c.label, 'plots': [{'id': p.id(), 'name': p.name} for p in c.plots]}
for c in cea.plots.categories.list_categories()}
@api.route('/')
class Dashboard(Resource):
def get(self):
"""
Get Dashboards from yaml file
"""
config = cea.config.Configuration()
plot_cache = cea.plots.cache.PlotCache(config)
dashboards = cea.plots.read_dashboards(config, plot_cache)
return [{'name': d.name, 'description': d.description, 'layout': d.layout if d.layout in LAYOUTS else 'row',
'plots': [{'title': plot.title, 'scenario':
plot.parameters['scenario-name'] if 'scenario-name' in plot.parameters.keys() else None}
for plot in d.plots]} for d in dashboards]
Include plot title to plotsfrom flask_restplus import Namespace, Resource, fields, abort
import cea.config
import cea.plots.cache
api = Namespace('Dashboard', description='Dashboard plots')
LAYOUTS = ['row', 'grid', 'map']
CATEGORIES = {c.name: {'label': c.label, 'plots': [{'id': p.id(), 'name': p.name} for p in c.plots]}
for c in cea.plots.categories.list_categories()}
@api.route('/')
class Dashboard(Resource):
def get(self):
"""
Get Dashboards from yaml file
"""
config = cea.config.Configuration()
plot_cache = cea.plots.cache.PlotCache(config)
dashboards = cea.plots.read_dashboards(config, plot_cache)
out = []
for d in dashboards:
dashboard = d.to_dict()
for i, plot in enumerate(dashboard['plots']):
dashboard['plots'][i]['title'] = d.plots[i].title
out.append(dashboard)
return out
|
<commit_before>from flask_restplus import Namespace, Resource, fields, abort
import cea.config
import cea.plots.cache
api = Namespace('Dashboard', description='Dashboard plots')
LAYOUTS = ['row', 'grid', 'map']
CATEGORIES = {c.name: {'label': c.label, 'plots': [{'id': p.id(), 'name': p.name} for p in c.plots]}
for c in cea.plots.categories.list_categories()}
@api.route('/')
class Dashboard(Resource):
def get(self):
"""
Get Dashboards from yaml file
"""
config = cea.config.Configuration()
plot_cache = cea.plots.cache.PlotCache(config)
dashboards = cea.plots.read_dashboards(config, plot_cache)
return [{'name': d.name, 'description': d.description, 'layout': d.layout if d.layout in LAYOUTS else 'row',
'plots': [{'title': plot.title, 'scenario':
plot.parameters['scenario-name'] if 'scenario-name' in plot.parameters.keys() else None}
for plot in d.plots]} for d in dashboards]
<commit_msg>Include plot title to plots<commit_after>from flask_restplus import Namespace, Resource, fields, abort
import cea.config
import cea.plots.cache
api = Namespace('Dashboard', description='Dashboard plots')
LAYOUTS = ['row', 'grid', 'map']
CATEGORIES = {c.name: {'label': c.label, 'plots': [{'id': p.id(), 'name': p.name} for p in c.plots]}
for c in cea.plots.categories.list_categories()}
@api.route('/')
class Dashboard(Resource):
def get(self):
"""
Get Dashboards from yaml file
"""
config = cea.config.Configuration()
plot_cache = cea.plots.cache.PlotCache(config)
dashboards = cea.plots.read_dashboards(config, plot_cache)
out = []
for d in dashboards:
dashboard = d.to_dict()
for i, plot in enumerate(dashboard['plots']):
dashboard['plots'][i]['title'] = d.plots[i].title
out.append(dashboard)
return out
|
7e9f8ca01b9cb1a70ee09dac9e0eecb8d370ad1f
|
acq4/devices/FalconTurret/falconturret.py
|
acq4/devices/FalconTurret/falconturret.py
|
import falconoptics
from ..FilterWheel import FilterWheel, FilterWheelFuture
class FalconTurret(FilterWheel):
def __init__(self, dm, config, name):
self.dev = falconoptics.Falcon(config_file=None, update_nonvolitile=True)
self.dev.home(block=False)
FilterWheel.__init__(self, dm, config, name)
def getPositionCount(self):
return self.dev._total_slides
def _getPosition(self):
return int(self.dev.current_slide)
def _setPosition(self, pos):
if pos == 'home':
self.dev.home(block=False)
else:
self.dev.move_to_slide(pos, block=False)
return FalconTurretFuture(self, pos)
def home(self):
self.setPosition('home')
def _stop(self):
self.dev.emergency_stop()
def isMoving(self):
return self.dev.is_moving
class FalconTurretFuture(FilterWheelFuture):
def _atTarget(self):
if self.position == 'home':
return self.dev.dev.is_homed
else:
return FilterWheelFuture._atTarget()
|
from acq4.pyqtgraph.Qt import QtGui
import falconoptics
from ..FilterWheel import FilterWheel, FilterWheelFuture
class FalconTurret(FilterWheel):
def __init__(self, dm, config, name):
self.dev = falconoptics.Falcon(config_file=None, update_nonvolitile=True)
self.dev.home(block=False)
FilterWheel.__init__(self, dm, config, name)
def getPositionCount(self):
return self.dev._total_slides
def _getPosition(self):
return int(self.dev.current_slide)
def _setPosition(self, pos):
if pos == 'home':
self.dev.home(block=False)
else:
self.dev.move_to_slide(pos, block=False)
return FalconTurretFuture(self, pos)
def home(self):
"""Search for home position on turret; used to recalibrate wheel location.
"""
self.setPosition('home')
def _stop(self):
self.dev.emergency_stop()
def isMoving(self):
return self.dev.is_moving
class FalconTurretFuture(FilterWheelFuture):
def _atTarget(self):
if self.position == 'home':
return self.dev.dev.is_homed
else:
return FilterWheelFuture._atTarget()
class FalconDevGui(FilterWheelDevGui):
def __init__(self, dev):
FilterWheelDevGui.__init__(self, dev)
self.homeBtn = QtGui.QPushButton("Find Home")
self.homeBtn.clicked.connect(self.dev.home)
|
Add home button to falcon turret dev gui
|
Add home button to falcon turret dev gui
|
Python
|
mit
|
campagnola/acq4,acq4/acq4,pbmanis/acq4,pbmanis/acq4,meganbkratz/acq4,acq4/acq4,pbmanis/acq4,campagnola/acq4,campagnola/acq4,meganbkratz/acq4,pbmanis/acq4,meganbkratz/acq4,campagnola/acq4,acq4/acq4,acq4/acq4,meganbkratz/acq4
|
import falconoptics
from ..FilterWheel import FilterWheel, FilterWheelFuture
class FalconTurret(FilterWheel):
def __init__(self, dm, config, name):
self.dev = falconoptics.Falcon(config_file=None, update_nonvolitile=True)
self.dev.home(block=False)
FilterWheel.__init__(self, dm, config, name)
def getPositionCount(self):
return self.dev._total_slides
def _getPosition(self):
return int(self.dev.current_slide)
def _setPosition(self, pos):
if pos == 'home':
self.dev.home(block=False)
else:
self.dev.move_to_slide(pos, block=False)
return FalconTurretFuture(self, pos)
def home(self):
self.setPosition('home')
def _stop(self):
self.dev.emergency_stop()
def isMoving(self):
return self.dev.is_moving
class FalconTurretFuture(FilterWheelFuture):
def _atTarget(self):
if self.position == 'home':
return self.dev.dev.is_homed
else:
return FilterWheelFuture._atTarget()
Add home button to falcon turret dev gui
|
from acq4.pyqtgraph.Qt import QtGui
import falconoptics
from ..FilterWheel import FilterWheel, FilterWheelFuture
class FalconTurret(FilterWheel):
def __init__(self, dm, config, name):
self.dev = falconoptics.Falcon(config_file=None, update_nonvolitile=True)
self.dev.home(block=False)
FilterWheel.__init__(self, dm, config, name)
def getPositionCount(self):
return self.dev._total_slides
def _getPosition(self):
return int(self.dev.current_slide)
def _setPosition(self, pos):
if pos == 'home':
self.dev.home(block=False)
else:
self.dev.move_to_slide(pos, block=False)
return FalconTurretFuture(self, pos)
def home(self):
"""Search for home position on turret; used to recalibrate wheel location.
"""
self.setPosition('home')
def _stop(self):
self.dev.emergency_stop()
def isMoving(self):
return self.dev.is_moving
class FalconTurretFuture(FilterWheelFuture):
def _atTarget(self):
if self.position == 'home':
return self.dev.dev.is_homed
else:
return FilterWheelFuture._atTarget()
class FalconDevGui(FilterWheelDevGui):
def __init__(self, dev):
FilterWheelDevGui.__init__(self, dev)
self.homeBtn = QtGui.QPushButton("Find Home")
self.homeBtn.clicked.connect(self.dev.home)
|
<commit_before>import falconoptics
from ..FilterWheel import FilterWheel, FilterWheelFuture
class FalconTurret(FilterWheel):
def __init__(self, dm, config, name):
self.dev = falconoptics.Falcon(config_file=None, update_nonvolitile=True)
self.dev.home(block=False)
FilterWheel.__init__(self, dm, config, name)
def getPositionCount(self):
return self.dev._total_slides
def _getPosition(self):
return int(self.dev.current_slide)
def _setPosition(self, pos):
if pos == 'home':
self.dev.home(block=False)
else:
self.dev.move_to_slide(pos, block=False)
return FalconTurretFuture(self, pos)
def home(self):
self.setPosition('home')
def _stop(self):
self.dev.emergency_stop()
def isMoving(self):
return self.dev.is_moving
class FalconTurretFuture(FilterWheelFuture):
def _atTarget(self):
if self.position == 'home':
return self.dev.dev.is_homed
else:
return FilterWheelFuture._atTarget()
<commit_msg>Add home button to falcon turret dev gui<commit_after>
|
from acq4.pyqtgraph.Qt import QtGui
import falconoptics
from ..FilterWheel import FilterWheel, FilterWheelFuture
class FalconTurret(FilterWheel):
def __init__(self, dm, config, name):
self.dev = falconoptics.Falcon(config_file=None, update_nonvolitile=True)
self.dev.home(block=False)
FilterWheel.__init__(self, dm, config, name)
def getPositionCount(self):
return self.dev._total_slides
def _getPosition(self):
return int(self.dev.current_slide)
def _setPosition(self, pos):
if pos == 'home':
self.dev.home(block=False)
else:
self.dev.move_to_slide(pos, block=False)
return FalconTurretFuture(self, pos)
def home(self):
"""Search for home position on turret; used to recalibrate wheel location.
"""
self.setPosition('home')
def _stop(self):
self.dev.emergency_stop()
def isMoving(self):
return self.dev.is_moving
class FalconTurretFuture(FilterWheelFuture):
def _atTarget(self):
if self.position == 'home':
return self.dev.dev.is_homed
else:
return FilterWheelFuture._atTarget()
class FalconDevGui(FilterWheelDevGui):
def __init__(self, dev):
FilterWheelDevGui.__init__(self, dev)
self.homeBtn = QtGui.QPushButton("Find Home")
self.homeBtn.clicked.connect(self.dev.home)
|
import falconoptics
from ..FilterWheel import FilterWheel, FilterWheelFuture
class FalconTurret(FilterWheel):
def __init__(self, dm, config, name):
self.dev = falconoptics.Falcon(config_file=None, update_nonvolitile=True)
self.dev.home(block=False)
FilterWheel.__init__(self, dm, config, name)
def getPositionCount(self):
return self.dev._total_slides
def _getPosition(self):
return int(self.dev.current_slide)
def _setPosition(self, pos):
if pos == 'home':
self.dev.home(block=False)
else:
self.dev.move_to_slide(pos, block=False)
return FalconTurretFuture(self, pos)
def home(self):
self.setPosition('home')
def _stop(self):
self.dev.emergency_stop()
def isMoving(self):
return self.dev.is_moving
class FalconTurretFuture(FilterWheelFuture):
def _atTarget(self):
if self.position == 'home':
return self.dev.dev.is_homed
else:
return FilterWheelFuture._atTarget()
Add home button to falcon turret dev guifrom acq4.pyqtgraph.Qt import QtGui
import falconoptics
from ..FilterWheel import FilterWheel, FilterWheelFuture
class FalconTurret(FilterWheel):
def __init__(self, dm, config, name):
self.dev = falconoptics.Falcon(config_file=None, update_nonvolitile=True)
self.dev.home(block=False)
FilterWheel.__init__(self, dm, config, name)
def getPositionCount(self):
return self.dev._total_slides
def _getPosition(self):
return int(self.dev.current_slide)
def _setPosition(self, pos):
if pos == 'home':
self.dev.home(block=False)
else:
self.dev.move_to_slide(pos, block=False)
return FalconTurretFuture(self, pos)
def home(self):
"""Search for home position on turret; used to recalibrate wheel location.
"""
self.setPosition('home')
def _stop(self):
self.dev.emergency_stop()
def isMoving(self):
return self.dev.is_moving
class FalconTurretFuture(FilterWheelFuture):
def _atTarget(self):
if self.position == 'home':
return self.dev.dev.is_homed
else:
return FilterWheelFuture._atTarget()
class FalconDevGui(FilterWheelDevGui):
def __init__(self, dev):
FilterWheelDevGui.__init__(self, dev)
self.homeBtn = QtGui.QPushButton("Find Home")
self.homeBtn.clicked.connect(self.dev.home)
|
<commit_before>import falconoptics
from ..FilterWheel import FilterWheel, FilterWheelFuture
class FalconTurret(FilterWheel):
def __init__(self, dm, config, name):
self.dev = falconoptics.Falcon(config_file=None, update_nonvolitile=True)
self.dev.home(block=False)
FilterWheel.__init__(self, dm, config, name)
def getPositionCount(self):
return self.dev._total_slides
def _getPosition(self):
return int(self.dev.current_slide)
def _setPosition(self, pos):
if pos == 'home':
self.dev.home(block=False)
else:
self.dev.move_to_slide(pos, block=False)
return FalconTurretFuture(self, pos)
def home(self):
self.setPosition('home')
def _stop(self):
self.dev.emergency_stop()
def isMoving(self):
return self.dev.is_moving
class FalconTurretFuture(FilterWheelFuture):
def _atTarget(self):
if self.position == 'home':
return self.dev.dev.is_homed
else:
return FilterWheelFuture._atTarget()
<commit_msg>Add home button to falcon turret dev gui<commit_after>from acq4.pyqtgraph.Qt import QtGui
import falconoptics
from ..FilterWheel import FilterWheel, FilterWheelFuture
class FalconTurret(FilterWheel):
def __init__(self, dm, config, name):
self.dev = falconoptics.Falcon(config_file=None, update_nonvolitile=True)
self.dev.home(block=False)
FilterWheel.__init__(self, dm, config, name)
def getPositionCount(self):
return self.dev._total_slides
def _getPosition(self):
return int(self.dev.current_slide)
def _setPosition(self, pos):
if pos == 'home':
self.dev.home(block=False)
else:
self.dev.move_to_slide(pos, block=False)
return FalconTurretFuture(self, pos)
def home(self):
"""Search for home position on turret; used to recalibrate wheel location.
"""
self.setPosition('home')
def _stop(self):
self.dev.emergency_stop()
def isMoving(self):
return self.dev.is_moving
class FalconTurretFuture(FilterWheelFuture):
def _atTarget(self):
if self.position == 'home':
return self.dev.dev.is_homed
else:
return FilterWheelFuture._atTarget()
class FalconDevGui(FilterWheelDevGui):
def __init__(self, dev):
FilterWheelDevGui.__init__(self, dev)
self.homeBtn = QtGui.QPushButton("Find Home")
self.homeBtn.clicked.connect(self.dev.home)
|
3e617e3ade1fa55562868c2e2bf8bc07f9b09a79
|
skflow/tests/test_io.py
|
skflow/tests/test_io.py
|
# Copyright 2015 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import random
from sklearn import datasets
from sklearn.metrics import accuracy_score, mean_squared_error, log_loss
import tensorflow as tf
from tensorflow.python.platform import googletest
from skflow.io import *
import skflow
class BaseTest(googletest.TestCase):
def test_pandas_data_classification(self):
if HAS_PANDAS:
random.seed(42)
iris = datasets.load_iris()
data = DataFrame(iris.data)
labels = DataFrame(iris.target)
classifier = skflow.TensorFlowLinearClassifier(n_classes=3)
classifier.fit(data, labels)
score = accuracy_score(classifier.predict(data), labels)
self.assertGreater(score, 0.5, "Failed with score = {0}".format(score))
if __name__ == '__main__':
tf.test.main()
|
# Copyright 2015 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import random
from sklearn import datasets
from sklearn.metrics import accuracy_score
import tensorflow as tf
from tensorflow.python.platform import googletest
from skflow.io import *
import skflow
class BaseTest(googletest.TestCase):
def test_pandas_data_classification(self):
if HAS_PANDAS:
random.seed(42)
iris = datasets.load_iris()
data = DataFrame(iris.data)
labels = DataFrame(iris.target)
classifier = skflow.TensorFlowLinearClassifier(n_classes=3)
classifier.fit(data, labels)
score = accuracy_score(classifier.predict(data), labels)
self.assertGreater(score, 0.5, "Failed with score = {0}".format(score))
else:
print("No pandas installed. test_pandas_data_classification skipped.")
if __name__ == '__main__':
tf.test.main()
|
Print when pandas not installed and removed unnecessary imports
|
Print when pandas not installed and removed unnecessary imports
|
Python
|
apache-2.0
|
anand-c-goog/tensorflow,alheinecke/tensorflow-xsmm,sandeepdsouza93/TensorFlow-15712,alisidd/tensorflow,paolodedios/tensorflow,kevin-coder/tensorflow-fork,sjperkins/tensorflow,yanchen036/tensorflow,with-git/tensorflow,benoitsteiner/tensorflow,kobejean/tensorflow,wangyum/tensorflow,Moriadry/tensorflow,rdipietro/tensorflow,Xeralux/tensorflow,anilmuthineni/tensorflow,ibmsoe/tensorflow,cancan101/tensorflow,xodus7/tensorflow,nikste/tensorflow,dhalleine/tensorflow,ZhangXinNan/tensorflow,jwlawson/tensorflow,thjashin/tensorflow,elingg/tensorflow,zasdfgbnm/tensorflow,ppwwyyxx/tensorflow,HKUST-SING/tensorflow,mavenlin/tensorflow,panmari/tensorflow,thesuperzapper/tensorflow,ppries/tensorflow,sandeepdsouza93/TensorFlow-15712,benoitsteiner/tensorflow,LUTAN/tensorflow,AndreasMadsen/tensorflow,nolanliou/tensorflow,pierreg/tensorflow,karllessard/tensorflow,ivano666/tensorflow,seaotterman/tensorflow,laszlocsomor/tensorflow,horance-liu/tensorflow,awni/tensorflow,sjperkins/tensorflow,haeusser/tensorflow,AnishShah/tensorflow,jhaux/tensorflow,seanli9jan/tensorflow,ArtsiomCh/tensorflow,freedomtan/tensorflow,EvenStrangest/tensorflow,annarev/tensorflow,jhaux/tensorflow,adamtiger/tensorflow,anilmuthineni/tensorflow,aldian/tensorflow,dendisuhubdy/tensorflow,jhaux/tensorflow,adamtiger/tensorflow,code-sauce/tensorflow,maciekcc/tensorflow,Intel-tensorflow/tensorflow,lakshayg/tensorflow,cg31/tensorflow,zycdragonball/tensorflow,gautam1858/tensorflow,mixturemodel-flow/tensorflow,calebfoss/tensorflow,tntnatbry/tensorflow,eaplatanios/tensorflow,taknevski/tensorflow-xsmm,av8ramit/tensorflow,jart/tensorflow,girving/tensorflow,nanditav/15712-TensorFlow,tensorflow/tensorflow,admcrae/tensorflow,markslwong/tensorflow,alistairlow/tensorflow,xzturn/tensorflow,TakayukiSakai/tensorflow,eaplatanios/tensorflow,kamcpp/tensorflow,dancingdan/tensorflow,alisidd/tensorflow,annarev/tensorflow,nburn42/tensorflow,DavidNorman/tensorflow,peterbraden/tensorflow,cg31/tensorflow,ppwwyyxx/tensorflow,unsiloai/syntaxnet-ops-hack,dancingdan/tensorflow,cancan101/tensorflow,aselle/tensorflow,ppwwyyxx/tensorflow,AnishShah/tensorflow,aldian/tensorflow,DavidNorman/tensorflow,dyoung418/tensorflow,Kongsea/tensorflow,andrewcmyers/tensorflow,kobejean/tensorflow,chris-chris/tensorflow,calebfoss/tensorflow,paolodedios/tensorflow,Kongsea/tensorflow,xzturn/tensorflow,nikste/tensorflow,gautam1858/tensorflow,gunan/tensorflow,ran5515/DeepDecision,kamcpp/tensorflow,zycdragonball/tensorflow,Bismarrck/tensorflow,xodus7/tensorflow,ArtsiomCh/tensorflow,dancingdan/tensorflow,ZhangXinNan/tensorflow,code-sauce/tensorflow,yanchen036/tensorflow,tongwang01/tensorflow,seaotterman/tensorflow,alistairlow/tensorflow,brchiu/tensorflow,HKUST-SING/tensorflow,neilhan/tensorflow,MoamerEncsConcordiaCa/tensorflow,lukeiwanski/tensorflow-opencl,calebfoss/tensorflow,apark263/tensorflow,nolanliou/tensorflow,nightjean/Deep-Learning,nburn42/tensorflow,laszlocsomor/tensorflow,eaplatanios/tensorflow,ychfan/tensorflow,scenarios/tensorflow,alsrgv/tensorflow,ibmsoe/tensorflow,thesuperzapper/tensorflow,jhaux/tensorflow,alistairlow/tensorflow,LUTAN/tensorflow,dancingdan/tensorflow,asadziach/tensorflow,aam-at/tensorflow,Bulochkin/tensorflow_pack,tornadozou/tensorflow,jeffzheng1/tensorflow,AndreasMadsen/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,thjashin/tensorflow,sjperkins/tensorflow,alsrgv/tensorflow,aselle/tensorflow,xzturn/tensorflow,mortada/tensorflow,JVillella/tensorflow,taknevski/tensorflow-xsmm,chemelnucfin/tensorflow,odejesush/tensorflow,thjashin/tensorflow,dongjoon-hyun/tensorflow,sandeepgupta2k4/tensorflow,wangyum/tensorflow,Mistobaan/tensorflow,hehongliang/tensorflow,ravindrapanda/tensorflow,brchiu/tensorflow,dyoung418/tensorflow,seaotterman/tensorflow,nikste/tensorflow,AndreasMadsen/tensorflow,JingJunYin/tensorflow,kobejean/tensorflow,hehongliang/tensorflow,av8ramit/tensorflow,HKUST-SING/tensorflow,vrv/tensorflow,benoitsteiner/tensorflow-xsmm,theflofly/tensorflow,JingJunYin/tensorflow,SnakeJenny/TensorFlow,benoitsteiner/tensorflow-xsmm,asimshankar/tensorflow,JVillella/tensorflow,yongtang/tensorflow,mortada/tensorflow,mengxn/tensorflow,tomasreimers/tensorflow-emscripten,aselle/tensorflow,snnn/tensorflow,tongwang01/tensorflow,peterbraden/tensorflow,allenlavoie/tensorflow,karllessard/tensorflow,Bulochkin/tensorflow_pack,thjashin/tensorflow,wchan/tensorflow,anand-c-goog/tensorflow,codrut3/tensorflow,MycChiu/tensorflow,Xeralux/tensorflow,jendap/tensorflow,laszlocsomor/tensorflow,ville-k/tensorflow,TakayukiSakai/tensorflow,tornadozou/tensorflow,tensorflow/tensorflow,karllessard/tensorflow,chenjun0210/tensorflow,manjunaths/tensorflow,tomasreimers/tensorflow-emscripten,kevin-coder/tensorflow-fork,eerwitt/tensorflow,jalexvig/tensorflow,pierreg/tensorflow,scenarios/tensorflow,alivecor/tensorflow,eaplatanios/tensorflow,alsrgv/tensorflow,arborh/tensorflow,ghchinoy/tensorflow,AndreasMadsen/tensorflow,jhseu/tensorflow,lukeiwanski/tensorflow,aam-at/tensorflow,anilmuthineni/tensorflow,cg31/tensorflow,nightjean/Deep-Learning,pavelchristof/gomoku-ai,elingg/tensorflow,alshedivat/tensorflow,gautam1858/tensorflow,sandeepgupta2k4/tensorflow,jhseu/tensorflow,ravindrapanda/tensorflow,ghchinoy/tensorflow,taknevski/tensorflow-xsmm,mdrumond/tensorflow,davidzchen/tensorflow,codrut3/tensorflow,nightjean/Deep-Learning,Moriadry/tensorflow,DavidNorman/tensorflow,martinwicke/tensorflow,xodus7/tensorflow,benoitsteiner/tensorflow-opencl,handroissuazo/tensorflow,DCSaunders/tensorflow,juharris/tensorflow,Intel-Corporation/tensorflow,theflofly/tensorflow,HaebinShin/tensorflow,xzturn/tensorflow,ppries/tensorflow,jostep/tensorflow,ibab/tensorflow,tensorflow/tensorflow-pywrap_saved_model,anand-c-goog/tensorflow,zasdfgbnm/tensorflow,LUTAN/tensorflow,ran5515/DeepDecision,benoitsteiner/tensorflow-opencl,nanditav/15712-TensorFlow,seanli9jan/tensorflow,zycdragonball/tensorflow,elingg/tensorflow,apark263/tensorflow,RapidApplicationDevelopment/tensorflow,bowang/tensorflow,adamtiger/tensorflow,ghchinoy/tensorflow,sandeepgupta2k4/tensorflow,krikru/tensorflow-opencl,theflofly/tensorflow,hehongliang/tensorflow,chris-chris/tensorflow,tensorflow/tensorflow-pywrap_saved_model,alheinecke/tensorflow-xsmm,hehongliang/tensorflow,ivano666/tensorflow,tillahoffmann/tensorflow,kamcpp/tensorflow,pierreg/tensorflow,apark263/tensorflow,Intel-tensorflow/tensorflow,Xeralux/tensorflow,bowang/tensorflow,DavidNorman/tensorflow,RapidApplicationDevelopment/tensorflow,brchiu/tensorflow,dansbecker/skflow,ninotoshi/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,juharris/tensorflow,laszlocsomor/tensorflow,ppwwyyxx/tensorflow,asadziach/tensorflow,alsrgv/tensorflow,nolanliou/tensorflow,ychfan/tensorflow,Intel-tensorflow/tensorflow,paolodedios/tensorflow,ZhangXinNan/tensorflow,cg31/tensorflow,tiagofrepereira2012/tensorflow,sandeepgupta2k4/tensorflow,ghchinoy/tensorflow,alisidd/tensorflow,XueqingLin/tensorflow,ychfan/tensorflow,Mistobaan/tensorflow,krikru/tensorflow-opencl,kevin-coder/tensorflow-fork,xzturn/tensorflow,jhaux/tensorflow,aam-at/tensorflow,pierreg/tensorflow,brchiu/tensorflow,lukeiwanski/tensorflow,guschmue/tensorflow,benoitsteiner/tensorflow-xsmm,jart/tensorflow,gunan/tensorflow,jart/tensorflow,krikru/tensorflow-opencl,guschmue/tensorflow,jhseu/tensorflow,a-doumoulakis/tensorflow,gibiansky/tensorflow,Moriadry/tensorflow,abhitopia/tensorflow,manipopopo/tensorflow,ivano666/tensorflow,sarvex/tensorflow,ibmsoe/tensorflow,anilmuthineni/tensorflow,zycdragonball/tensorflow,andrewcmyers/tensorflow,yufengg/tensorflow,suiyuan2009/tensorflow,petewarden/tensorflow,hfp/tensorflow-xsmm,alivecor/tensorflow,girving/tensorflow,mdrumond/tensorflow,eaplatanios/tensorflow,chris-chris/tensorflow,allenlavoie/tensorflow,admcrae/tensorflow,Bismarrck/tensorflow,renyi533/tensorflow,strint/tensorflow,drpngx/tensorflow,tensorflow/tensorflow,Intel-Corporation/tensorflow,manazhao/tf_recsys,tillahoffmann/tensorflow,TakayukiSakai/tensorflow,caisq/tensorflow,awni/tensorflow,av8ramit/tensorflow,ppwwyyxx/tensorflow,Mistobaan/tensorflow,ZhangXinNan/tensorflow,allenlavoie/tensorflow,arborh/tensorflow,DCSaunders/tensorflow,av8ramit/tensorflow,av8ramit/tensorflow,mdrumond/tensorflow,wchan/tensorflow,apark263/tensorflow,mavenlin/tensorflow,Bismarrck/tensorflow,andrewcmyers/tensorflow,pcm17/tensorflow,theflofly/tensorflow,kamcpp/tensorflow,tornadozou/tensorflow,lukeiwanski/tensorflow,theflofly/tensorflow,annarev/tensorflow,HaebinShin/tensorflow,sarvex/tensorflow,ivano666/tensorflow,ghchinoy/tensorflow,benoitsteiner/tensorflow-xsmm,ZhangXinNan/tensorflow,hehongliang/tensorflow,karllessard/tensorflow,ninotoshi/tensorflow,yongtang/tensorflow,eaplatanios/tensorflow,codrut3/tensorflow,drpngx/tensorflow,aselle/tensorflow,yaroslavvb/tensorflow,AndreasMadsen/tensorflow,xodus7/tensorflow,tensorflow/tensorflow,ppwwyyxx/tensorflow,frreiss/tensorflow-fred,ville-k/tensorflow,Bismarrck/tensorflow,cxxgtxy/tensorflow,rabipanda/tensorflow,Moriadry/tensorflow,kamcpp/tensorflow,kevin-coder/tensorflow-fork,maciekcc/tensorflow,hfp/tensorflow-xsmm,suiyuan2009/tensorflow,markslwong/tensorflow,krikru/tensorflow-opencl,jbedorf/tensorflow,meteorcloudy/tensorflow,lukeiwanski/tensorflow,MostafaGazar/tensorflow,lukeiwanski/tensorflow,seaotterman/tensorflow,elingg/tensorflow,laosiaudi/tensorflow,paolodedios/tensorflow,wangyum/tensorflow,mrry/tensorflow,benoitsteiner/tensorflow,kamcpp/tensorflow,jostep/tensorflow,admcrae/tensorflow,TakayukiSakai/tensorflow,ghchinoy/tensorflow,cxxgtxy/tensorflow,nburn42/tensorflow,lakshayg/tensorflow,nburn42/tensorflow,krikru/tensorflow-opencl,johndpope/tensorflow,ibmsoe/tensorflow,Bulochkin/tensorflow_pack,TakayukiSakai/tensorflow,manazhao/tf_recsys,drpngx/tensorflow,taknevski/tensorflow-xsmm,jendap/tensorflow,Xeralux/tensorflow,ville-k/tensorflow,jbedorf/tensorflow,yaroslavvb/tensorflow,ibab/tensorflow,ArtsiomCh/tensorflow,ville-k/tensorflow,petewarden/tensorflow,calebfoss/tensorflow,jart/tensorflow,scenarios/tensorflow,unsiloai/syntaxnet-ops-hack,laosiaudi/tensorflow,JingJunYin/tensorflow,sandeepgupta2k4/tensorflow,girving/tensorflow,Bulochkin/tensorflow_pack,ppwwyyxx/tensorflow,anand-c-goog/tensorflow,petewarden/tensorflow_makefile,sandeepgupta2k4/tensorflow,jalexvig/tensorflow,neilhan/tensorflow,aam-at/tensorflow,gnieboer/tensorflow,pavelchristof/gomoku-ai,ninotoshi/tensorflow,MostafaGazar/tensorflow,xzturn/tensorflow,guschmue/tensorflow,XueqingLin/tensorflow,kevin-coder/tensorflow-fork,tiagofrepereira2012/tensorflow,meteorcloudy/tensorflow,mrry/tensorflow,horance-liu/tensorflow,xzturn/tensorflow,chemelnucfin/tensorflow,DCSaunders/tensorflow,ninotoshi/tensorflow,vrv/tensorflow,strint/tensorflow,cxxgtxy/tensorflow,manjunaths/tensorflow,ArtsiomCh/tensorflow,snnn/tensorflow,scenarios/tensorflow,tomasreimers/tensorflow-emscripten,alivecor/tensorflow,neilhan/tensorflow,whn09/tensorflow,asimshankar/tensorflow,benoitsteiner/tensorflow-xsmm,gautam1858/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,asimshankar/tensorflow,adit-chandra/tensorflow,strint/tensorflow,gunan/tensorflow,chenjun0210/tensorflow,gojira/tensorflow,yanchen036/tensorflow,awni/tensorflow,Bulochkin/tensorflow_pack,wchan/tensorflow,yaroslavvb/tensorflow,benoitsteiner/tensorflow-opencl,tensorflow/tensorflow-pywrap_tf_optimizer,jhseu/tensorflow,whn09/tensorflow,jwlawson/tensorflow,cancan101/tensorflow,av8ramit/tensorflow,asimshankar/tensorflow,annarev/tensorflow,aam-at/tensorflow,brchiu/tensorflow,tornadozou/tensorflow,lukeiwanski/tensorflow-opencl,gunan/tensorflow,aldian/tensorflow,mdrumond/tensorflow,with-git/tensorflow,rabipanda/tensorflow,benoitsteiner/tensorflow-xsmm,ran5515/DeepDecision,aldian/tensorflow,nightjean/Deep-Learning,davidzchen/tensorflow,seanli9jan/tensorflow,ppries/tensorflow,brchiu/tensorflow,ibmsoe/tensorflow,ppwwyyxx/tensorflow,theflofly/tensorflow,zycdragonball/tensorflow,jbedorf/tensorflow,paolodedios/tensorflow,chemelnucfin/tensorflow,with-git/tensorflow,jhaux/tensorflow,pcm17/tensorflow,ibmsoe/tensorflow,lukeiwanski/tensorflow-opencl,tensorflow/tensorflow-experimental_link_static_libraries_once,yongtang/tensorflow,davidzchen/tensorflow,rdipietro/tensorflow,jwlawson/tensorflow,martinwicke/tensorflow,cxxgtxy/tensorflow,girving/tensorflow,anilmuthineni/tensorflow,allenlavoie/tensorflow,bowang/tensorflow,ravindrapanda/tensorflow,Intel-tensorflow/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,apark263/tensorflow,eerwitt/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,mrry/tensorflow,chemelnucfin/tensorflow,jendap/tensorflow,alisidd/tensorflow,meteorcloudy/tensorflow,av8ramit/tensorflow,Mazecreator/tensorflow,mixturemodel-flow/tensorflow,gautam1858/tensorflow,hfp/tensorflow-xsmm,johndpope/tensorflow,mixturemodel-flow/tensorflow,haeusser/tensorflow,EvenStrangest/tensorflow,yaroslavvb/tensorflow,lukeiwanski/tensorflow,Kongsea/tensorflow,gnieboer/tensorflow,juharris/tensorflow,guschmue/tensorflow,drpngx/tensorflow,mengxn/tensorflow,asimshankar/tensorflow,aselle/tensorflow,krikru/tensorflow-opencl,manipopopo/tensorflow,rdipietro/tensorflow,jwlawson/tensorflow,arborh/tensorflow,kamcpp/tensorflow,benoitsteiner/tensorflow-xsmm,benoitsteiner/tensorflow-xsmm,adit-chandra/tensorflow,lukeiwanski/tensorflow,renyi533/tensorflow,ageron/tensorflow,AnishShah/tensorflow,kobejean/tensorflow,cg31/tensorflow,nikste/tensorflow,davidzchen/tensorflow,mavenlin/tensorflow,nanditav/15712-TensorFlow,ychfan/tensorflow,JingJunYin/tensorflow,tntnatbry/tensorflow,manipopopo/tensorflow,tensorflow/tensorflow-pywrap_saved_model,eadgarchen/tensorflow,JingJunYin/tensorflow,freedomtan/tensorflow,unsiloai/syntaxnet-ops-hack,alisidd/tensorflow,MoamerEncsConcordiaCa/tensorflow,Mistobaan/tensorflow,benoitsteiner/tensorflow-opencl,alheinecke/tensorflow-xsmm,awni/tensorflow,chenjun0210/tensorflow,Bismarrck/tensorflow,ravindrapanda/tensorflow,Mistobaan/tensorflow,karllessard/tensorflow,alivecor/tensorflow,alsrgv/tensorflow,JVillella/tensorflow,horance-liu/tensorflow,jalexvig/tensorflow,ArtsiomCh/tensorflow,ivano666/tensorflow,kchodorow/tensorflow,eadgarchen/tensorflow,RapidApplicationDevelopment/tensorflow,haeusser/tensorflow,dongjoon-hyun/tensorflow,yanchen036/tensorflow,jostep/tensorflow,HKUST-SING/tensorflow,meteorcloudy/tensorflow,peterbraden/tensorflow,calebfoss/tensorflow,calebfoss/tensorflow,Intel-Corporation/tensorflow,ghchinoy/tensorflow,ville-k/tensorflow,codrut3/tensorflow,chemelnucfin/tensorflow,DavidNorman/tensorflow,adit-chandra/tensorflow,snnn/tensorflow,ZhangXinNan/tensorflow,hfp/tensorflow-xsmm,horance-liu/tensorflow,asimshankar/tensorflow,guschmue/tensorflow,memo/tensorflow,karllessard/tensorflow,neilhan/tensorflow,strint/tensorflow,abhitopia/tensorflow,alshedivat/tensorflow,kobejean/tensorflow,DCSaunders/tensorflow,theflofly/tensorflow,laszlocsomor/tensorflow,chenjun0210/tensorflow,gibiansky/tensorflow,nburn42/tensorflow,handroissuazo/tensorflow,dongjoon-hyun/tensorflow,alheinecke/tensorflow-xsmm,aselle/tensorflow,Intel-Corporation/tensorflow,admcrae/tensorflow,thjashin/tensorflow,sarvex/tensorflow,aselle/tensorflow,tillahoffmann/tensorflow,ageron/tensorflow,manjunaths/tensorflow,with-git/tensorflow,nolanliou/tensorflow,aselle/tensorflow,juharris/tensorflow,tongwang01/tensorflow,jeffzheng1/tensorflow,taknevski/tensorflow-xsmm,markslwong/tensorflow,dongjoon-hyun/tensorflow,paolodedios/tensorflow,Xeralux/tensorflow,dancingdan/tensorflow,seaotterman/tensorflow,nanditav/15712-TensorFlow,tntnatbry/tensorflow,adamtiger/tensorflow,jalexvig/tensorflow,ppwwyyxx/tensorflow,dancingdan/tensorflow,chris-chris/tensorflow,Mazecreator/tensorflow,arborh/tensorflow,laszlocsomor/tensorflow,code-sauce/tensorflow,tntnatbry/tensorflow,anilmuthineni/tensorflow,Bulochkin/tensorflow_pack,DavidNorman/tensorflow,Kongsea/tensorflow,haeusser/tensorflow,dendisuhubdy/tensorflow,anand-c-goog/tensorflow,alistairlow/tensorflow,dendisuhubdy/tensorflow,odejesush/tensorflow,ninotoshi/tensorflow,HKUST-SING/tensorflow,aam-at/tensorflow,strint/tensorflow,juharris/tensorflow,yanchen036/tensorflow,seanli9jan/tensorflow,peterbraden/tensorflow,wchan/tensorflow,MycChiu/tensorflow,naturali/tensorflow,anilmuthineni/tensorflow,MostafaGazar/tensorflow,Intel-tensorflow/tensorflow,nolanliou/tensorflow,Mistobaan/tensorflow,pavelchristof/gomoku-ai,mixturemodel-flow/tensorflow,theflofly/tensorflow,lakshayg/tensorflow,jhaux/tensorflow,lukeiwanski/tensorflow-opencl,Mistobaan/tensorflow,jostep/tensorflow,strint/tensorflow,arborh/tensorflow,zasdfgbnm/tensorflow,MostafaGazar/tensorflow,gojira/tensorflow,kevin-coder/tensorflow-fork,chemelnucfin/tensorflow,mavenlin/tensorflow,hehongliang/tensorflow,davidzchen/tensorflow,brchiu/tensorflow,kchodorow/tensorflow,nightjean/Deep-Learning,anilmuthineni/tensorflow,jalexvig/tensorflow,yaroslavvb/tensorflow,cxxgtxy/tensorflow,taknevski/tensorflow-xsmm,jbedorf/tensorflow,anand-c-goog/tensorflow,haeusser/tensorflow,MoamerEncsConcordiaCa/tensorflow,jwlawson/tensorflow,pavelchristof/gomoku-ai,petewarden/tensorflow,neilhan/tensorflow,lakshayg/tensorflow,hfp/tensorflow-xsmm,strint/tensorflow,Xeralux/tensorflow,manjunaths/tensorflow,ZhangXinNan/tensorflow,Bulochkin/tensorflow_pack,elingg/tensorflow,seanli9jan/tensorflow,tensorflow/tensorflow-pywrap_saved_model,sjperkins/tensorflow,Intel-Corporation/tensorflow,scenarios/tensorflow,guschmue/tensorflow,kevin-coder/tensorflow-fork,JVillella/tensorflow,paolodedios/tensorflow,jhseu/tensorflow,naturali/tensorflow,dendisuhubdy/tensorflow,TakayukiSakai/tensorflow,abhitopia/tensorflow,gnieboer/tensorflow,raymondxyang/tensorflow,drpngx/tensorflow,alivecor/tensorflow,haeusser/tensorflow,raymondxyang/tensorflow,jostep/tensorflow,hfp/tensorflow-xsmm,sarvex/tensorflow,petewarden/tensorflow_makefile,caisq/tensorflow,haeusser/tensorflow,MoamerEncsConcordiaCa/tensorflow,apark263/tensorflow,thesuperzapper/tensorflow,eerwitt/tensorflow,frreiss/tensorflow-fred,xzturn/tensorflow,ishay2b/tensorflow,ppries/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,abhitopia/tensorflow,alsrgv/tensorflow,meteorcloudy/tensorflow,tornadozou/tensorflow,lakshayg/tensorflow,pcm17/tensorflow,maciekcc/tensorflow,mrry/tensorflow,dancingdan/tensorflow,sarvex/tensorflow,yongtang/tensorflow,freedomtan/tensorflow,tensorflow/skflow,aldian/tensorflow,gojira/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,code-sauce/tensorflow,AnishShah/tensorflow,davidzchen/tensorflow,nanditav/15712-TensorFlow,elingg/tensorflow,eaplatanios/tensorflow,eadgarchen/tensorflow,renyi533/tensorflow,horance-liu/tensorflow,frreiss/tensorflow-fred,HaebinShin/tensorflow,panmari/tensorflow,chris-chris/tensorflow,bowang/tensorflow,cancan101/tensorflow,dyoung418/tensorflow,tongwang01/tensorflow,markslwong/tensorflow,zasdfgbnm/tensorflow,panmari/tensorflow,gautam1858/tensorflow,ychfan/tensorflow,brchiu/tensorflow,codrut3/tensorflow,codrut3/tensorflow,annarev/tensorflow,cxxgtxy/tensorflow,manazhao/tf_recsys,xodus7/tensorflow,kchodorow/tensorflow,petewarden/tensorflow,jbedorf/tensorflow,cancan101/tensorflow,nightjean/Deep-Learning,aselle/tensorflow,abhitopia/tensorflow,petewarden/tensorflow_makefile,nikste/tensorflow,ychfan/tensorflow,elingg/tensorflow,dongjoon-hyun/tensorflow,AndreasMadsen/tensorflow,calebfoss/tensorflow,pierreg/tensorflow,chemelnucfin/tensorflow,brchiu/tensorflow,adamtiger/tensorflow,lakshayg/tensorflow,alisidd/tensorflow,admcrae/tensorflow,cg31/tensorflow,arborh/tensorflow,wchan/tensorflow,Kongsea/tensorflow,Bulochkin/tensorflow_pack,Intel-tensorflow/tensorflow,adamtiger/tensorflow,yufengg/tensorflow,lukeiwanski/tensorflow-opencl,manipopopo/tensorflow,eaplatanios/tensorflow,neilhan/tensorflow,lukeiwanski/tensorflow,jhseu/tensorflow,dansbecker/skflow,snnn/tensorflow,nolanliou/tensorflow,chemelnucfin/tensorflow,naturali/tensorflow,alshedivat/tensorflow,adit-chandra/tensorflow,dancingdan/tensorflow,pavelchristof/gomoku-ai,xzturn/tensorflow,anand-c-goog/tensorflow,with-git/tensorflow,bowang/tensorflow,ArtsiomCh/tensorflow,jendap/tensorflow,panmari/tensorflow,thjashin/tensorflow,chris-chris/tensorflow,neilhan/tensorflow,a-doumoulakis/tensorflow,alisidd/tensorflow,paolodedios/tensorflow,petewarden/tensorflow,AndreasMadsen/tensorflow,tiagofrepereira2012/tensorflow,frreiss/tensorflow-fred,dhalleine/tensorflow,peterbraden/tensorflow,alistairlow/tensorflow,jostep/tensorflow,ran5515/DeepDecision,Intel-Corporation/tensorflow,vrv/tensorflow,kevin-coder/tensorflow-fork,awni/tensorflow,eadgarchen/tensorflow,eerwitt/tensorflow,Bulochkin/tensorflow_pack,jeffzheng1/tensorflow,jalexvig/tensorflow,dhalleine/tensorflow,martinwicke/tensorflow,benoitsteiner/tensorflow,Intel-tensorflow/tensorflow,ivano666/tensorflow,HaebinShin/tensorflow,code-sauce/tensorflow,seanli9jan/tensorflow,zasdfgbnm/tensorflow,annarev/tensorflow,hsaputra/tensorflow,ghchinoy/tensorflow,gnieboer/tensorflow,whn09/tensorflow,markslwong/tensorflow,yaroslavvb/tensorflow,xodus7/tensorflow,jwlawson/tensorflow,kobejean/tensorflow,dhalleine/tensorflow,ppries/tensorflow,elingg/tensorflow,alshedivat/tensorflow,mavenlin/tensorflow,johndpope/tensorflow,elingg/tensorflow,zasdfgbnm/tensorflow,MycChiu/tensorflow,codrut3/tensorflow,martinwicke/tensorflow,ninotoshi/tensorflow,johndpope/tensorflow,guschmue/tensorflow,rabipanda/tensorflow,martinwicke/tensorflow,martinwicke/tensorflow,AnishShah/tensorflow,ppries/tensorflow,MycChiu/tensorflow,jeffzheng1/tensorflow,scenarios/tensorflow,XueqingLin/tensorflow,dancingdan/tensorflow,yanchen036/tensorflow,memo/tensorflow,ArtsiomCh/tensorflow,tomasreimers/tensorflow-emscripten,gibiansky/tensorflow,lukeiwanski/tensorflow,a-doumoulakis/tensorflow,yongtang/tensorflow,tiagofrepereira2012/tensorflow,gnieboer/tensorflow,gautam1858/tensorflow,Carmezim/tensorflow,chris-chris/tensorflow,taknevski/tensorflow-xsmm,lukeiwanski/tensorflow-opencl,asimshankar/tensorflow,ageron/tensorflow,MycChiu/tensorflow,zasdfgbnm/tensorflow,wchan/tensorflow,ravindrapanda/tensorflow,gojira/tensorflow,ppwwyyxx/tensorflow,laosiaudi/tensorflow,RapidApplicationDevelopment/tensorflow,martinwicke/tensorflow,aselle/tensorflow,manipopopo/tensorflow,kchodorow/tensorflow,av8ramit/tensorflow,vrv/tensorflow,gautam1858/tensorflow,frreiss/tensorflow-fred,hsaputra/tensorflow,seanli9jan/tensorflow,zasdfgbnm/tensorflow,alistairlow/tensorflow,sandeepdsouza93/TensorFlow-15712,nikste/tensorflow,pcm17/tensorflow,HKUST-SING/tensorflow,laosiaudi/tensorflow,asimshankar/tensorflow,meteorcloudy/tensorflow,HKUST-SING/tensorflow,yongtang/tensorflow,andrewcmyers/tensorflow,sandeepdsouza93/TensorFlow-15712,adit-chandra/tensorflow,ychfan/tensorflow,renyi533/tensorflow,mavenlin/tensorflow,thesuperzapper/tensorflow,ppwwyyxx/tensorflow,rabipanda/tensorflow,DavidNorman/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,admcrae/tensorflow,zasdfgbnm/tensorflow,benoitsteiner/tensorflow-xsmm,dancingdan/tensorflow,apark263/tensorflow,ravindrapanda/tensorflow,jhseu/tensorflow,with-git/tensorflow,AnishShah/tensorflow,tensorflow/tensorflow-pywrap_saved_model,tensorflow/tensorflow-experimental_link_static_libraries_once,admcrae/tensorflow,LUTAN/tensorflow,ibab/tensorflow,MycChiu/tensorflow,panmari/tensorflow,renyi533/tensorflow,ravindrapanda/tensorflow,thesuperzapper/tensorflow,sandeepdsouza93/TensorFlow-15712,caisq/tensorflow,caisq/tensorflow,caisq/tensorflow,frreiss/tensorflow-fred,SnakeJenny/TensorFlow,jalexvig/tensorflow,odejesush/tensorflow,frreiss/tensorflow-fred,apark263/tensorflow,jhseu/tensorflow,tntnatbry/tensorflow,odejesush/tensorflow,johndpope/tensorflow,adit-chandra/tensorflow,apark263/tensorflow,thjashin/tensorflow,seaotterman/tensorflow,DCSaunders/tensorflow,yufengg/tensorflow,LUTAN/tensorflow,thjashin/tensorflow,guschmue/tensorflow,llhe/tensorflow,renyi533/tensorflow,MoamerEncsConcordiaCa/tensorflow,LUTAN/tensorflow,jwlawson/tensorflow,Intel-tensorflow/tensorflow,jwlawson/tensorflow,dendisuhubdy/tensorflow,manjunaths/tensorflow,bowang/tensorflow,alisidd/tensorflow,mortada/tensorflow,hsaputra/tensorflow,dendisuhubdy/tensorflow,ageron/tensorflow,Intel-tensorflow/tensorflow,Bulochkin/tensorflow_pack,HKUST-SING/tensorflow,jeffzheng1/tensorflow,Xeralux/tensorflow,xzturn/tensorflow,Mazecreator/tensorflow,EvenStrangest/tensorflow,mortada/tensorflow,jhaux/tensorflow,sandeepdsouza93/TensorFlow-15712,tensorflow/tensorflow-experimental_link_static_libraries_once,davidzchen/tensorflow,dhalleine/tensorflow,LUTAN/tensorflow,andrewcmyers/tensorflow,scenarios/tensorflow,drpngx/tensorflow,jart/tensorflow,hfp/tensorflow-xsmm,ppwwyyxx/tensorflow,tongwang01/tensorflow,gojira/tensorflow,alshedivat/tensorflow,arborh/tensorflow,eerwitt/tensorflow,haeusser/tensorflow,Mistobaan/tensorflow,Mazecreator/tensorflow,dongjoon-hyun/tensorflow,sjperkins/tensorflow,DavidNorman/tensorflow,dongjoon-hyun/tensorflow,theflofly/tensorflow,aam-at/tensorflow,petewarden/tensorflow_makefile,chenjun0210/tensorflow,sandeepgupta2k4/tensorflow,mengxn/tensorflow,tntnatbry/tensorflow,LUTAN/tensorflow,benoitsteiner/tensorflow-xsmm,maciekcc/tensorflow,raymondxyang/tensorflow,alsrgv/tensorflow,llhe/tensorflow,yufengg/tensorflow,adit-chandra/tensorflow,HKUST-SING/tensorflow,sandeepdsouza93/TensorFlow-15712,gautam1858/tensorflow,dongjoon-hyun/tensorflow,sarvex/tensorflow,Carmezim/tensorflow,asadziach/tensorflow,pcm17/tensorflow,nburn42/tensorflow,laszlocsomor/tensorflow,girving/tensorflow,XueqingLin/tensorflow,tornadozou/tensorflow,alsrgv/tensorflow,jbedorf/tensorflow,snnn/tensorflow,tensorflow/tensorflow-pywrap_saved_model,allenlavoie/tensorflow,Bulochkin/tensorflow_pack,apark263/tensorflow,cancan101/tensorflow,hsaputra/tensorflow,chemelnucfin/tensorflow,xodus7/tensorflow,tensorflow/tensorflow,odejesush/tensorflow,kobejean/tensorflow,arborh/tensorflow,dhalleine/tensorflow,hfp/tensorflow-xsmm,gautam1858/tensorflow,johndpope/tensorflow,jhseu/tensorflow,ageron/tensorflow,eadgarchen/tensorflow,girving/tensorflow,rabipanda/tensorflow,handroissuazo/tensorflow,allenlavoie/tensorflow,JVillella/tensorflow,mixturemodel-flow/tensorflow,dongjoon-hyun/tensorflow,gnieboer/tensorflow,jendap/tensorflow,nburn42/tensorflow,gibiansky/tensorflow,alshedivat/tensorflow,arborh/tensorflow,tillahoffmann/tensorflow,sjperkins/tensorflow,XueqingLin/tensorflow,pierreg/tensorflow,bowang/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,tensorflow/tensorflow,MostafaGazar/tensorflow,alivecor/tensorflow,rabipanda/tensorflow,benoitsteiner/tensorflow-opencl,zasdfgbnm/tensorflow,mortada/tensorflow,mrry/tensorflow,alheinecke/tensorflow-xsmm,tensorflow/tensorflow-experimental_link_static_libraries_once,alheinecke/tensorflow-xsmm,jwlawson/tensorflow,krikru/tensorflow-opencl,seaotterman/tensorflow,codrut3/tensorflow,RapidApplicationDevelopment/tensorflow,aam-at/tensorflow,Moriadry/tensorflow,TakayukiSakai/tensorflow,eadgarchen/tensorflow,ville-k/tensorflow,tensorflow/skflow,MoamerEncsConcordiaCa/tensorflow,tensorflow/tensorflow-pywrap_saved_model,peterbraden/tensorflow,ninotoshi/tensorflow,llhe/tensorflow,tensorflow/tensorflow-pywrap_saved_model,aldian/tensorflow,Mistobaan/tensorflow,hfp/tensorflow-xsmm,aam-at/tensorflow,wangyum/tensorflow,Carmezim/tensorflow,ishay2b/tensorflow,allenlavoie/tensorflow,AnishShah/tensorflow,juharris/tensorflow,memo/tensorflow,gunan/tensorflow,HaebinShin/tensorflow,a-doumoulakis/tensorflow,allenlavoie/tensorflow,rdipietro/tensorflow,mdrumond/tensorflow,rabipanda/tensorflow,gautam1858/tensorflow,alheinecke/tensorflow-xsmm,Mistobaan/tensorflow,jendap/tensorflow,benoitsteiner/tensorflow-opencl,Bismarrck/tensorflow,lukeiwanski/tensorflow-opencl,a-doumoulakis/tensorflow,dyoung418/tensorflow,dyoung418/tensorflow,tillahoffmann/tensorflow,panmari/tensorflow,laosiaudi/tensorflow,ibmsoe/tensorflow,alsrgv/tensorflow,dendisuhubdy/tensorflow,XueqingLin/tensorflow,ibmsoe/tensorflow,ychfan/tensorflow,benoitsteiner/tensorflow,llhe/tensorflow,petewarden/tensorflow,karllessard/tensorflow,pavelchristof/gomoku-ai,apark263/tensorflow,gibiansky/tensorflow,sandeepgupta2k4/tensorflow,llhe/tensorflow,snnn/tensorflow,suiyuan2009/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,Carmezim/tensorflow,markslwong/tensorflow,sjperkins/tensorflow,benoitsteiner/tensorflow-opencl,odejesush/tensorflow,hsaputra/tensorflow,MostafaGazar/tensorflow,suiyuan2009/tensorflow,aam-at/tensorflow,memo/tensorflow,Bismarrck/tensorflow,girving/tensorflow,gunan/tensorflow,nanditav/15712-TensorFlow,jeffzheng1/tensorflow,tensorflow/tensorflow,sandeepdsouza93/TensorFlow-15712,Mistobaan/tensorflow,DavidNorman/tensorflow,MoamerEncsConcordiaCa/tensorflow,jhaux/tensorflow,kamcpp/tensorflow,cancan101/tensorflow,jhaux/tensorflow,seanli9jan/tensorflow,ville-k/tensorflow,memo/tensorflow,llhe/tensorflow,pcm17/tensorflow,admcrae/tensorflow,snnn/tensorflow,tongwang01/tensorflow,ageron/tensorflow,asadziach/tensorflow,benoitsteiner/tensorflow,frreiss/tensorflow-fred,adit-chandra/tensorflow,asimshankar/tensorflow,AnishShah/tensorflow,manjunaths/tensorflow,Xeralux/tensorflow,yongtang/tensorflow,arborh/tensorflow,tiagofrepereira2012/tensorflow,meteorcloudy/tensorflow,abhitopia/tensorflow,cg31/tensorflow,ghchinoy/tensorflow,handroissuazo/tensorflow,vrv/tensorflow,mdrumond/tensorflow,nburn42/tensorflow,yongtang/tensorflow,eerwitt/tensorflow,frreiss/tensorflow-fred,markslwong/tensorflow,adit-chandra/tensorflow,andrewcmyers/tensorflow,calebfoss/tensorflow,theflofly/tensorflow,dongjoon-hyun/tensorflow,gautam1858/tensorflow,eadgarchen/tensorflow,aselle/tensorflow,jwlawson/tensorflow,ibab/tensorflow,mixturemodel-flow/tensorflow,SnakeJenny/TensorFlow,nanditav/15712-TensorFlow,mengxn/tensorflow,DCSaunders/tensorflow,nolanliou/tensorflow,caisq/tensorflow,yongtang/tensorflow,manazhao/tf_recsys,paolodedios/tensorflow,a-doumoulakis/tensorflow,krikru/tensorflow-opencl,with-git/tensorflow,vrv/tensorflow,Mazecreator/tensorflow,jart/tensorflow,alsrgv/tensorflow,seaotterman/tensorflow,andrewcmyers/tensorflow,tongwang01/tensorflow,DCSaunders/tensorflow,laosiaudi/tensorflow,whn09/tensorflow,DCSaunders/tensorflow,xzturn/tensorflow,EvenStrangest/tensorflow,handroissuazo/tensorflow,eadgarchen/tensorflow,Carmezim/tensorflow,girving/tensorflow,ville-k/tensorflow,kobejean/tensorflow,naturali/tensorflow,neilhan/tensorflow,ageron/tensorflow,mengxn/tensorflow,chemelnucfin/tensorflow,tornadozou/tensorflow,Mazecreator/tensorflow,ZhangXinNan/tensorflow,nightjean/Deep-Learning,abhitopia/tensorflow,ishay2b/tensorflow,abhitopia/tensorflow,ishay2b/tensorflow,chemelnucfin/tensorflow,dongjoon-hyun/tensorflow,wangyum/tensorflow,caisq/tensorflow,jhseu/tensorflow,manipopopo/tensorflow,kevin-coder/tensorflow-fork,benoitsteiner/tensorflow,alsrgv/tensorflow,suiyuan2009/tensorflow,alistairlow/tensorflow,yaroslavvb/tensorflow,nolanliou/tensorflow,suiyuan2009/tensorflow,kobejean/tensorflow,kobejean/tensorflow,unsiloai/syntaxnet-ops-hack,dhalleine/tensorflow,ravindrapanda/tensorflow,caisq/tensorflow,whn09/tensorflow,gnieboer/tensorflow,yongtang/tensorflow,petewarden/tensorflow_makefile,llhe/tensorflow,ishay2b/tensorflow,manjunaths/tensorflow,horance-liu/tensorflow,asadziach/tensorflow,freedomtan/tensorflow,tillahoffmann/tensorflow,gunan/tensorflow,with-git/tensorflow,mengxn/tensorflow,Moriadry/tensorflow,dyoung418/tensorflow,sjperkins/tensorflow,ArtsiomCh/tensorflow,thjashin/tensorflow,rabipanda/tensorflow,annarev/tensorflow,ibab/tensorflow,DCSaunders/tensorflow,awni/tensorflow,manipopopo/tensorflow,Xeralux/tensorflow,xodus7/tensorflow,chenjun0210/tensorflow,eaplatanios/tensorflow,jalexvig/tensorflow,chenjun0210/tensorflow,abhitopia/tensorflow,maciekcc/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,renyi533/tensorflow,jwlawson/tensorflow,ZhangXinNan/tensorflow,mavenlin/tensorflow,taknevski/tensorflow-xsmm,SnakeJenny/TensorFlow,RapidApplicationDevelopment/tensorflow,freedomtan/tensorflow,alistairlow/tensorflow,hehongliang/tensorflow,ibab/tensorflow,adamtiger/tensorflow,gojira/tensorflow,ishay2b/tensorflow,peterbraden/tensorflow,rdipietro/tensorflow,ville-k/tensorflow,Mazecreator/tensorflow,alistairlow/tensorflow,krikru/tensorflow-opencl,petewarden/tensorflow_makefile,handroissuazo/tensorflow,thesuperzapper/tensorflow,memo/tensorflow,aldian/tensorflow,nburn42/tensorflow,naturali/tensorflow,cxxgtxy/tensorflow,gibiansky/tensorflow,nanditav/15712-TensorFlow,manipopopo/tensorflow,vrv/tensorflow,lukeiwanski/tensorflow-opencl,mdrumond/tensorflow,memo/tensorflow,jart/tensorflow,jhseu/tensorflow,admcrae/tensorflow,eaplatanios/tensorflow,gojira/tensorflow,ageron/tensorflow,yufengg/tensorflow,ageron/tensorflow,pierreg/tensorflow,drpngx/tensorflow,jbedorf/tensorflow,wangyum/tensorflow,MoamerEncsConcordiaCa/tensorflow,nikste/tensorflow,jostep/tensorflow,thesuperzapper/tensorflow,mengxn/tensorflow,gunan/tensorflow,allenlavoie/tensorflow,karllessard/tensorflow,yufengg/tensorflow,mengxn/tensorflow,raymondxyang/tensorflow,freedomtan/tensorflow,tongwang01/tensorflow,lukeiwanski/tensorflow,manipopopo/tensorflow,johndpope/tensorflow,chenjun0210/tensorflow,RapidApplicationDevelopment/tensorflow,av8ramit/tensorflow,yufengg/tensorflow,ravindrapanda/tensorflow,Bismarrck/tensorflow,gnieboer/tensorflow,alheinecke/tensorflow-xsmm,mrry/tensorflow,tomasreimers/tensorflow-emscripten,alistairlow/tensorflow,brchiu/tensorflow,a-doumoulakis/tensorflow,code-sauce/tensorflow,freedomtan/tensorflow,MycChiu/tensorflow,Moriadry/tensorflow,Carmezim/tensorflow,chris-chris/tensorflow,xzturn/tensorflow,anand-c-goog/tensorflow,brchiu/tensorflow,laszlocsomor/tensorflow,tntnatbry/tensorflow,alheinecke/tensorflow-xsmm,kchodorow/tensorflow,rdipietro/tensorflow,kchodorow/tensorflow,arborh/tensorflow,wangyum/tensorflow,mortada/tensorflow,nikste/tensorflow,MostafaGazar/tensorflow,AndreasMadsen/tensorflow,awni/tensorflow,jart/tensorflow,pavelchristof/gomoku-ai,tiagofrepereira2012/tensorflow,benoitsteiner/tensorflow,snnn/tensorflow,XueqingLin/tensorflow,jendap/tensorflow,cg31/tensorflow,thesuperzapper/tensorflow,unsiloai/syntaxnet-ops-hack,gojira/tensorflow,SnakeJenny/TensorFlow,manipopopo/tensorflow,JVillella/tensorflow,jendap/tensorflow,benoitsteiner/tensorflow,raymondxyang/tensorflow,martinwicke/tensorflow,jbedorf/tensorflow,zycdragonball/tensorflow,eerwitt/tensorflow,sjperkins/tensorflow,haeusser/tensorflow,kevin-coder/tensorflow-fork,aldian/tensorflow,awni/tensorflow,cancan101/tensorflow,freedomtan/tensorflow,ghchinoy/tensorflow,alivecor/tensorflow,girving/tensorflow,seaotterman/tensorflow,horance-liu/tensorflow,horance-liu/tensorflow,guschmue/tensorflow,Bismarrck/tensorflow,zasdfgbnm/tensorflow,cancan101/tensorflow,naturali/tensorflow,JingJunYin/tensorflow,AnishShah/tensorflow,tiagofrepereira2012/tensorflow,ninotoshi/tensorflow,AnishShah/tensorflow,laszlocsomor/tensorflow,asadziach/tensorflow,JingJunYin/tensorflow,ppries/tensorflow,rabipanda/tensorflow,raymondxyang/tensorflow,mortada/tensorflow,zycdragonball/tensorflow,cg31/tensorflow,odejesush/tensorflow,freedomtan/tensorflow,maciekcc/tensorflow,markslwong/tensorflow,peterbraden/tensorflow,frreiss/tensorflow-fred,SnakeJenny/TensorFlow,mrry/tensorflow,ghchinoy/tensorflow,hfp/tensorflow-xsmm,neilhan/tensorflow,Mazecreator/tensorflow,ran5515/DeepDecision,davidzchen/tensorflow,yongtang/tensorflow,meteorcloudy/tensorflow,jeffzheng1/tensorflow,tntnatbry/tensorflow,XueqingLin/tensorflow,annarev/tensorflow,thesuperzapper/tensorflow,freedomtan/tensorflow,karllessard/tensorflow,caisq/tensorflow,vrv/tensorflow,xodus7/tensorflow,adit-chandra/tensorflow,ZhangXinNan/tensorflow,tomasreimers/tensorflow-emscripten,tensorflow/tensorflow,vrv/tensorflow,tomasreimers/tensorflow-emscripten,unsiloai/syntaxnet-ops-hack,AndreasMadsen/tensorflow,raymondxyang/tensorflow,MycChiu/tensorflow,xodus7/tensorflow,jbedorf/tensorflow,eaplatanios/tensorflow,sarvex/tensorflow,seanli9jan/tensorflow,johndpope/tensorflow,av8ramit/tensorflow,hsaputra/tensorflow,meteorcloudy/tensorflow,calebfoss/tensorflow,annarev/tensorflow,snnn/tensorflow,freedomtan/tensorflow,tensorflow/tensorflow,EvenStrangest/tensorflow,DavidNorman/tensorflow,codrut3/tensorflow,benoitsteiner/tensorflow-xsmm,nolanliou/tensorflow,tensorflow/tensorflow-pywrap_saved_model,petewarden/tensorflow,ychfan/tensorflow,JingJunYin/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,jhseu/tensorflow,dancingdan/tensorflow,unsiloai/syntaxnet-ops-hack,gunan/tensorflow,dendisuhubdy/tensorflow,jbedorf/tensorflow,arborh/tensorflow,scenarios/tensorflow,maciekcc/tensorflow,mavenlin/tensorflow,jart/tensorflow,kchodorow/tensorflow,mdrumond/tensorflow,rabipanda/tensorflow,LUTAN/tensorflow,frreiss/tensorflow-fred,tensorflow/tensorflow-experimental_link_static_libraries_once,gojira/tensorflow,yaroslavvb/tensorflow,a-doumoulakis/tensorflow,ran5515/DeepDecision,Xeralux/tensorflow,jendap/tensorflow,MycChiu/tensorflow,chenjun0210/tensorflow,laosiaudi/tensorflow,tornadozou/tensorflow,panmari/tensorflow,gojira/tensorflow,mrry/tensorflow,gunan/tensorflow,llhe/tensorflow,odejesush/tensorflow,Intel-Corporation/tensorflow,RapidApplicationDevelopment/tensorflow,suiyuan2009/tensorflow,ville-k/tensorflow,strint/tensorflow,handroissuazo/tensorflow,hsaputra/tensorflow,manjunaths/tensorflow,MoamerEncsConcordiaCa/tensorflow,cxxgtxy/tensorflow,gibiansky/tensorflow,markslwong/tensorflow,SnakeJenny/TensorFlow,HaebinShin/tensorflow,mdrumond/tensorflow,yanchen036/tensorflow,odejesush/tensorflow,alshedivat/tensorflow,benoitsteiner/tensorflow-opencl,alshedivat/tensorflow,ghchinoy/tensorflow,AnishShah/tensorflow,xodus7/tensorflow,aam-at/tensorflow,wangyum/tensorflow,martinwicke/tensorflow,ishay2b/tensorflow,pcm17/tensorflow,hsaputra/tensorflow,hsaputra/tensorflow,ppries/tensorflow,girving/tensorflow,allenlavoie/tensorflow,hfp/tensorflow-xsmm,pavelchristof/gomoku-ai,jalexvig/tensorflow,karllessard/tensorflow,manipopopo/tensorflow,rdipietro/tensorflow,andrewcmyers/tensorflow,hsaputra/tensorflow,ZhangXinNan/tensorflow,jeffzheng1/tensorflow,asadziach/tensorflow,petewarden/tensorflow,kchodorow/tensorflow,rabipanda/tensorflow,davidzchen/tensorflow,bowang/tensorflow,manazhao/tf_recsys,TakayukiSakai/tensorflow,wangyum/tensorflow,mengxn/tensorflow,av8ramit/tensorflow,benoitsteiner/tensorflow,asimshankar/tensorflow,ivano666/tensorflow,asadziach/tensorflow,maciekcc/tensorflow,karllessard/tensorflow,dendisuhubdy/tensorflow,jeffzheng1/tensorflow,guschmue/tensorflow,alshedivat/tensorflow,pierreg/tensorflow,caisq/tensorflow,theflofly/tensorflow,petewarden/tensorflow,freedomtan/tensorflow,ageron/tensorflow,taknevski/tensorflow-xsmm,jart/tensorflow,chris-chris/tensorflow,Intel-Corporation/tensorflow,whn09/tensorflow,unsiloai/syntaxnet-ops-hack,petewarden/tensorflow,rdipietro/tensorflow,lakshayg/tensorflow,ibab/tensorflow,renyi533/tensorflow,nburn42/tensorflow,gibiansky/tensorflow,theflofly/tensorflow,laosiaudi/tensorflow,tensorflow/tensorflow-pywrap_saved_model,tensorflow/tensorflow-experimental_link_static_libraries_once,ran5515/DeepDecision,chemelnucfin/tensorflow,gunan/tensorflow,drpngx/tensorflow,mixturemodel-flow/tensorflow,nightjean/Deep-Learning,tomasreimers/tensorflow-emscripten,tillahoffmann/tensorflow,code-sauce/tensorflow,naturali/tensorflow,code-sauce/tensorflow,mortada/tensorflow,RapidApplicationDevelopment/tensorflow,JVillella/tensorflow,Bulochkin/tensorflow_pack,sjperkins/tensorflow,pcm17/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,benoitsteiner/tensorflow-opencl,codrut3/tensorflow,tiagofrepereira2012/tensorflow,Mazecreator/tensorflow,anilmuthineni/tensorflow,kchodorow/tensorflow,raymondxyang/tensorflow,kamcpp/tensorflow,alshedivat/tensorflow,Carmezim/tensorflow,Xeralux/tensorflow,Intel-tensorflow/tensorflow,tomasreimers/tensorflow-emscripten,jostep/tensorflow,wchan/tensorflow,frreiss/tensorflow-fred,handroissuazo/tensorflow,jendap/tensorflow,sandeepdsouza93/TensorFlow-15712,nanditav/15712-TensorFlow,memo/tensorflow,jalexvig/tensorflow,manjunaths/tensorflow,annarev/tensorflow,sandeepgupta2k4/tensorflow,Moriadry/tensorflow,llhe/tensorflow,ageron/tensorflow,JingJunYin/tensorflow,HaebinShin/tensorflow,drpngx/tensorflow,MostafaGazar/tensorflow,DavidNorman/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,DavidNorman/tensorflow,sarvex/tensorflow,kobejean/tensorflow,ageron/tensorflow,ppries/tensorflow,manazhao/tf_recsys,eadgarchen/tensorflow,davidzchen/tensorflow,nburn42/tensorflow,whn09/tensorflow,girving/tensorflow,gnieboer/tensorflow,kevin-coder/tensorflow-fork,johndpope/tensorflow,dendisuhubdy/tensorflow,JingJunYin/tensorflow,adit-chandra/tensorflow,petewarden/tensorflow_makefile,gunan/tensorflow,renyi533/tensorflow,juharris/tensorflow,panmari/tensorflow,allenlavoie/tensorflow,seanli9jan/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,davidzchen/tensorflow,alshedivat/tensorflow,asadziach/tensorflow,Bismarrck/tensorflow,petewarden/tensorflow_makefile,paolodedios/tensorflow,XueqingLin/tensorflow,drpngx/tensorflow,lukeiwanski/tensorflow-opencl,alisidd/tensorflow,paolodedios/tensorflow,adit-chandra/tensorflow,rdipietro/tensorflow,gibiansky/tensorflow,yanchen036/tensorflow,mortada/tensorflow,petewarden/tensorflow,dyoung418/tensorflow,asimshankar/tensorflow,tensorflow/tensorflow,jalexvig/tensorflow,pcm17/tensorflow,llhe/tensorflow,davidzchen/tensorflow,seanli9jan/tensorflow,naturali/tensorflow,nolanliou/tensorflow,johndpope/tensorflow,nikste/tensorflow,EvenStrangest/tensorflow,renyi533/tensorflow,jbedorf/tensorflow,tensorflow/tensorflow,code-sauce/tensorflow,alsrgv/tensorflow,mrry/tensorflow,strint/tensorflow,jbedorf/tensorflow,jendap/tensorflow,ibab/tensorflow,Kongsea/tensorflow,wchan/tensorflow,manazhao/tf_recsys,laosiaudi/tensorflow,anand-c-goog/tensorflow,tillahoffmann/tensorflow,Bismarrck/tensorflow,scenarios/tensorflow,SnakeJenny/TensorFlow,snnn/tensorflow,mixturemodel-flow/tensorflow,annarev/tensorflow,Intel-tensorflow/tensorflow,eerwitt/tensorflow,meteorcloudy/tensorflow,Carmezim/tensorflow,laszlocsomor/tensorflow,alivecor/tensorflow,DCSaunders/tensorflow,gojira/tensorflow,renyi533/tensorflow,whn09/tensorflow,Carmezim/tensorflow,snnn/tensorflow,lakshayg/tensorflow,eadgarchen/tensorflow,EvenStrangest/tensorflow,eerwitt/tensorflow,tensorflow/tensorflow-pywrap_saved_model,petewarden/tensorflow,MostafaGazar/tensorflow,dyoung418/tensorflow,yaroslavvb/tensorflow,tntnatbry/tensorflow,renyi533/tensorflow,horance-liu/tensorflow,Kongsea/tensorflow,ibmsoe/tensorflow,dhalleine/tensorflow,ivano666/tensorflow,juharris/tensorflow,horance-liu/tensorflow,EvenStrangest/tensorflow,memo/tensorflow,whn09/tensorflow,handroissuazo/tensorflow,aam-at/tensorflow,Kongsea/tensorflow,HaebinShin/tensorflow
|
# Copyright 2015 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import random
from sklearn import datasets
from sklearn.metrics import accuracy_score, mean_squared_error, log_loss
import tensorflow as tf
from tensorflow.python.platform import googletest
from skflow.io import *
import skflow
class BaseTest(googletest.TestCase):
def test_pandas_data_classification(self):
if HAS_PANDAS:
random.seed(42)
iris = datasets.load_iris()
data = DataFrame(iris.data)
labels = DataFrame(iris.target)
classifier = skflow.TensorFlowLinearClassifier(n_classes=3)
classifier.fit(data, labels)
score = accuracy_score(classifier.predict(data), labels)
self.assertGreater(score, 0.5, "Failed with score = {0}".format(score))
if __name__ == '__main__':
tf.test.main()
Print when pandas not installed and removed unnecessary imports
|
# Copyright 2015 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import random
from sklearn import datasets
from sklearn.metrics import accuracy_score
import tensorflow as tf
from tensorflow.python.platform import googletest
from skflow.io import *
import skflow
class BaseTest(googletest.TestCase):
def test_pandas_data_classification(self):
if HAS_PANDAS:
random.seed(42)
iris = datasets.load_iris()
data = DataFrame(iris.data)
labels = DataFrame(iris.target)
classifier = skflow.TensorFlowLinearClassifier(n_classes=3)
classifier.fit(data, labels)
score = accuracy_score(classifier.predict(data), labels)
self.assertGreater(score, 0.5, "Failed with score = {0}".format(score))
else:
print("No pandas installed. test_pandas_data_classification skipped.")
if __name__ == '__main__':
tf.test.main()
|
<commit_before># Copyright 2015 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import random
from sklearn import datasets
from sklearn.metrics import accuracy_score, mean_squared_error, log_loss
import tensorflow as tf
from tensorflow.python.platform import googletest
from skflow.io import *
import skflow
class BaseTest(googletest.TestCase):
def test_pandas_data_classification(self):
if HAS_PANDAS:
random.seed(42)
iris = datasets.load_iris()
data = DataFrame(iris.data)
labels = DataFrame(iris.target)
classifier = skflow.TensorFlowLinearClassifier(n_classes=3)
classifier.fit(data, labels)
score = accuracy_score(classifier.predict(data), labels)
self.assertGreater(score, 0.5, "Failed with score = {0}".format(score))
if __name__ == '__main__':
tf.test.main()
<commit_msg>Print when pandas not installed and removed unnecessary imports<commit_after>
|
# Copyright 2015 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import random
from sklearn import datasets
from sklearn.metrics import accuracy_score
import tensorflow as tf
from tensorflow.python.platform import googletest
from skflow.io import *
import skflow
class BaseTest(googletest.TestCase):
def test_pandas_data_classification(self):
if HAS_PANDAS:
random.seed(42)
iris = datasets.load_iris()
data = DataFrame(iris.data)
labels = DataFrame(iris.target)
classifier = skflow.TensorFlowLinearClassifier(n_classes=3)
classifier.fit(data, labels)
score = accuracy_score(classifier.predict(data), labels)
self.assertGreater(score, 0.5, "Failed with score = {0}".format(score))
else:
print("No pandas installed. test_pandas_data_classification skipped.")
if __name__ == '__main__':
tf.test.main()
|
# Copyright 2015 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import random
from sklearn import datasets
from sklearn.metrics import accuracy_score, mean_squared_error, log_loss
import tensorflow as tf
from tensorflow.python.platform import googletest
from skflow.io import *
import skflow
class BaseTest(googletest.TestCase):
def test_pandas_data_classification(self):
if HAS_PANDAS:
random.seed(42)
iris = datasets.load_iris()
data = DataFrame(iris.data)
labels = DataFrame(iris.target)
classifier = skflow.TensorFlowLinearClassifier(n_classes=3)
classifier.fit(data, labels)
score = accuracy_score(classifier.predict(data), labels)
self.assertGreater(score, 0.5, "Failed with score = {0}".format(score))
if __name__ == '__main__':
tf.test.main()
Print when pandas not installed and removed unnecessary imports# Copyright 2015 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import random
from sklearn import datasets
from sklearn.metrics import accuracy_score
import tensorflow as tf
from tensorflow.python.platform import googletest
from skflow.io import *
import skflow
class BaseTest(googletest.TestCase):
def test_pandas_data_classification(self):
if HAS_PANDAS:
random.seed(42)
iris = datasets.load_iris()
data = DataFrame(iris.data)
labels = DataFrame(iris.target)
classifier = skflow.TensorFlowLinearClassifier(n_classes=3)
classifier.fit(data, labels)
score = accuracy_score(classifier.predict(data), labels)
self.assertGreater(score, 0.5, "Failed with score = {0}".format(score))
else:
print("No pandas installed. test_pandas_data_classification skipped.")
if __name__ == '__main__':
tf.test.main()
|
<commit_before># Copyright 2015 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import random
from sklearn import datasets
from sklearn.metrics import accuracy_score, mean_squared_error, log_loss
import tensorflow as tf
from tensorflow.python.platform import googletest
from skflow.io import *
import skflow
class BaseTest(googletest.TestCase):
def test_pandas_data_classification(self):
if HAS_PANDAS:
random.seed(42)
iris = datasets.load_iris()
data = DataFrame(iris.data)
labels = DataFrame(iris.target)
classifier = skflow.TensorFlowLinearClassifier(n_classes=3)
classifier.fit(data, labels)
score = accuracy_score(classifier.predict(data), labels)
self.assertGreater(score, 0.5, "Failed with score = {0}".format(score))
if __name__ == '__main__':
tf.test.main()
<commit_msg>Print when pandas not installed and removed unnecessary imports<commit_after># Copyright 2015 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import random
from sklearn import datasets
from sklearn.metrics import accuracy_score
import tensorflow as tf
from tensorflow.python.platform import googletest
from skflow.io import *
import skflow
class BaseTest(googletest.TestCase):
def test_pandas_data_classification(self):
if HAS_PANDAS:
random.seed(42)
iris = datasets.load_iris()
data = DataFrame(iris.data)
labels = DataFrame(iris.target)
classifier = skflow.TensorFlowLinearClassifier(n_classes=3)
classifier.fit(data, labels)
score = accuracy_score(classifier.predict(data), labels)
self.assertGreater(score, 0.5, "Failed with score = {0}".format(score))
else:
print("No pandas installed. test_pandas_data_classification skipped.")
if __name__ == '__main__':
tf.test.main()
|
411581b5b773daceee9d3e2d7751ca652d251c00
|
aiosmtplib/__init__.py
|
aiosmtplib/__init__.py
|
"""
aiosmtplib
==========
An asyncio SMTP client.
Roughly based (with API differences) on smtplib from the Python 3 standard
library by: The Dragon De Monsyne <dragondm@integral.org>
Author: Cole Maclean <hi@cole.io>
"""
__title__ = 'aiosmtplib'
__version__ = '0.1.7'
__author__ = 'Cole Maclean'
__license__ = 'MIT'
__copyright__ = 'Copyright 2016 Cole Maclean'
from .errors import * # NOQA
from .response import * # NOQA
from .smtp import * # NOQA
from .status import * # NOQA
__all__ = (
errors.__all__ + response.__all__ + smtp.__all__ + status.__all__ # NOQA
)
|
"""
aiosmtplib
==========
An asyncio SMTP client.
Roughly based (with API differences) on smtplib from the Python 3 standard
library by: The Dragon De Monsyne <dragondm@integral.org>
Author: Cole Maclean <hi@cole.io>
"""
__title__ = 'aiosmtplib'
__version__ = '1.0.0'
__author__ = 'Cole Maclean'
__license__ = 'MIT'
__copyright__ = 'Copyright 2016 Cole Maclean'
from .errors import * # NOQA
from .response import * # NOQA
from .smtp import * # NOQA
from .status import * # NOQA
__all__ = (
errors.__all__ + response.__all__ + smtp.__all__ + status.__all__ # NOQA
)
|
Move to 1.0.0; let's all use semver
|
Move to 1.0.0; let's all use semver
|
Python
|
mit
|
cole/aiosmtplib
|
"""
aiosmtplib
==========
An asyncio SMTP client.
Roughly based (with API differences) on smtplib from the Python 3 standard
library by: The Dragon De Monsyne <dragondm@integral.org>
Author: Cole Maclean <hi@cole.io>
"""
__title__ = 'aiosmtplib'
__version__ = '0.1.7'
__author__ = 'Cole Maclean'
__license__ = 'MIT'
__copyright__ = 'Copyright 2016 Cole Maclean'
from .errors import * # NOQA
from .response import * # NOQA
from .smtp import * # NOQA
from .status import * # NOQA
__all__ = (
errors.__all__ + response.__all__ + smtp.__all__ + status.__all__ # NOQA
)
Move to 1.0.0; let's all use semver
|
"""
aiosmtplib
==========
An asyncio SMTP client.
Roughly based (with API differences) on smtplib from the Python 3 standard
library by: The Dragon De Monsyne <dragondm@integral.org>
Author: Cole Maclean <hi@cole.io>
"""
__title__ = 'aiosmtplib'
__version__ = '1.0.0'
__author__ = 'Cole Maclean'
__license__ = 'MIT'
__copyright__ = 'Copyright 2016 Cole Maclean'
from .errors import * # NOQA
from .response import * # NOQA
from .smtp import * # NOQA
from .status import * # NOQA
__all__ = (
errors.__all__ + response.__all__ + smtp.__all__ + status.__all__ # NOQA
)
|
<commit_before>"""
aiosmtplib
==========
An asyncio SMTP client.
Roughly based (with API differences) on smtplib from the Python 3 standard
library by: The Dragon De Monsyne <dragondm@integral.org>
Author: Cole Maclean <hi@cole.io>
"""
__title__ = 'aiosmtplib'
__version__ = '0.1.7'
__author__ = 'Cole Maclean'
__license__ = 'MIT'
__copyright__ = 'Copyright 2016 Cole Maclean'
from .errors import * # NOQA
from .response import * # NOQA
from .smtp import * # NOQA
from .status import * # NOQA
__all__ = (
errors.__all__ + response.__all__ + smtp.__all__ + status.__all__ # NOQA
)
<commit_msg>Move to 1.0.0; let's all use semver<commit_after>
|
"""
aiosmtplib
==========
An asyncio SMTP client.
Roughly based (with API differences) on smtplib from the Python 3 standard
library by: The Dragon De Monsyne <dragondm@integral.org>
Author: Cole Maclean <hi@cole.io>
"""
__title__ = 'aiosmtplib'
__version__ = '1.0.0'
__author__ = 'Cole Maclean'
__license__ = 'MIT'
__copyright__ = 'Copyright 2016 Cole Maclean'
from .errors import * # NOQA
from .response import * # NOQA
from .smtp import * # NOQA
from .status import * # NOQA
__all__ = (
errors.__all__ + response.__all__ + smtp.__all__ + status.__all__ # NOQA
)
|
"""
aiosmtplib
==========
An asyncio SMTP client.
Roughly based (with API differences) on smtplib from the Python 3 standard
library by: The Dragon De Monsyne <dragondm@integral.org>
Author: Cole Maclean <hi@cole.io>
"""
__title__ = 'aiosmtplib'
__version__ = '0.1.7'
__author__ = 'Cole Maclean'
__license__ = 'MIT'
__copyright__ = 'Copyright 2016 Cole Maclean'
from .errors import * # NOQA
from .response import * # NOQA
from .smtp import * # NOQA
from .status import * # NOQA
__all__ = (
errors.__all__ + response.__all__ + smtp.__all__ + status.__all__ # NOQA
)
Move to 1.0.0; let's all use semver"""
aiosmtplib
==========
An asyncio SMTP client.
Roughly based (with API differences) on smtplib from the Python 3 standard
library by: The Dragon De Monsyne <dragondm@integral.org>
Author: Cole Maclean <hi@cole.io>
"""
__title__ = 'aiosmtplib'
__version__ = '1.0.0'
__author__ = 'Cole Maclean'
__license__ = 'MIT'
__copyright__ = 'Copyright 2016 Cole Maclean'
from .errors import * # NOQA
from .response import * # NOQA
from .smtp import * # NOQA
from .status import * # NOQA
__all__ = (
errors.__all__ + response.__all__ + smtp.__all__ + status.__all__ # NOQA
)
|
<commit_before>"""
aiosmtplib
==========
An asyncio SMTP client.
Roughly based (with API differences) on smtplib from the Python 3 standard
library by: The Dragon De Monsyne <dragondm@integral.org>
Author: Cole Maclean <hi@cole.io>
"""
__title__ = 'aiosmtplib'
__version__ = '0.1.7'
__author__ = 'Cole Maclean'
__license__ = 'MIT'
__copyright__ = 'Copyright 2016 Cole Maclean'
from .errors import * # NOQA
from .response import * # NOQA
from .smtp import * # NOQA
from .status import * # NOQA
__all__ = (
errors.__all__ + response.__all__ + smtp.__all__ + status.__all__ # NOQA
)
<commit_msg>Move to 1.0.0; let's all use semver<commit_after>"""
aiosmtplib
==========
An asyncio SMTP client.
Roughly based (with API differences) on smtplib from the Python 3 standard
library by: The Dragon De Monsyne <dragondm@integral.org>
Author: Cole Maclean <hi@cole.io>
"""
__title__ = 'aiosmtplib'
__version__ = '1.0.0'
__author__ = 'Cole Maclean'
__license__ = 'MIT'
__copyright__ = 'Copyright 2016 Cole Maclean'
from .errors import * # NOQA
from .response import * # NOQA
from .smtp import * # NOQA
from .status import * # NOQA
__all__ = (
errors.__all__ + response.__all__ + smtp.__all__ + status.__all__ # NOQA
)
|
f1cdc9d6d4736202480045aa633fb0ac5992e7e3
|
test/views.py
|
test/views.py
|
from django.core.urlresolvers import reverse_lazy, reverse
from django import forms
from .models import Author
from popupcrud.views import PopupCrudViewSet
# Create your views here.
class AuthorForm(forms.ModelForm):
sex = forms.ChoiceField(label="Sex", choices=(('M', 'Male'), ('F', 'Female')))
class Meta:
model = Author
fields = ('name', 'age')
class AuthorCrudViewset(PopupCrudViewSet):
model = Author
fields = ('name', 'age')
list_display = ('name', 'age', 'half_age', 'double_age')
list_url = reverse_lazy("authors")
new_url = reverse_lazy("new-author")
"""
form_class = AuthorForm
list_permission_required = ('tests.add_author',)
create_permission_required = ('tests.add_author',)
update_permission_required = ('tests.change_author',)
delete_permission_required = ('tests.delete_author',)
"""
def half_age(self, author):
return author.age/2
half_age.label = "Half Age"
def get_edit_url(self, obj):
return reverse("edit-author", kwargs={'pk': obj.pk})
def get_delete_url(self, obj):
return reverse("delete-author", kwargs={'pk': obj.pk})
|
from django.core.urlresolvers import reverse_lazy, reverse
from django import forms
from .models import Author
from popupcrud.views import PopupCrudViewSet
# Create your views here.
class AuthorForm(forms.ModelForm):
sex = forms.ChoiceField(label="Sex", choices=(('M', 'Male'), ('F', 'Female')))
class Meta:
model = Author
fields = ('name', 'age')
class AuthorCrudViewset(PopupCrudViewSet):
model = Author
fields = ('name', 'age')
list_display = ('name', 'age', 'half_age', 'double_age')
list_url = reverse_lazy("authors")
new_url = reverse_lazy("new-author")
"""
form_class = AuthorForm
list_permission_required = ('tests.add_author',)
create_permission_required = ('tests.add_author',)
update_permission_required = ('tests.change_author',)
delete_permission_required = ('tests.delete_author',)
"""
def half_age(self, author):
return int(author.age/2)
half_age.label = "Half Age"
def get_edit_url(self, obj):
return reverse("edit-author", kwargs={'pk': obj.pk})
def get_delete_url(self, obj):
return reverse("delete-author", kwargs={'pk': obj.pk})
|
Handle PY3 default data conversion vagaries in unit test
|
Handle PY3 default data conversion vagaries in unit test
|
Python
|
bsd-3-clause
|
harikvpy/django-popupcrud,harikvpy/django-popupcrud,harikvpy/django-popupcrud
|
from django.core.urlresolvers import reverse_lazy, reverse
from django import forms
from .models import Author
from popupcrud.views import PopupCrudViewSet
# Create your views here.
class AuthorForm(forms.ModelForm):
sex = forms.ChoiceField(label="Sex", choices=(('M', 'Male'), ('F', 'Female')))
class Meta:
model = Author
fields = ('name', 'age')
class AuthorCrudViewset(PopupCrudViewSet):
model = Author
fields = ('name', 'age')
list_display = ('name', 'age', 'half_age', 'double_age')
list_url = reverse_lazy("authors")
new_url = reverse_lazy("new-author")
"""
form_class = AuthorForm
list_permission_required = ('tests.add_author',)
create_permission_required = ('tests.add_author',)
update_permission_required = ('tests.change_author',)
delete_permission_required = ('tests.delete_author',)
"""
def half_age(self, author):
return author.age/2
half_age.label = "Half Age"
def get_edit_url(self, obj):
return reverse("edit-author", kwargs={'pk': obj.pk})
def get_delete_url(self, obj):
return reverse("delete-author", kwargs={'pk': obj.pk})
Handle PY3 default data conversion vagaries in unit test
|
from django.core.urlresolvers import reverse_lazy, reverse
from django import forms
from .models import Author
from popupcrud.views import PopupCrudViewSet
# Create your views here.
class AuthorForm(forms.ModelForm):
sex = forms.ChoiceField(label="Sex", choices=(('M', 'Male'), ('F', 'Female')))
class Meta:
model = Author
fields = ('name', 'age')
class AuthorCrudViewset(PopupCrudViewSet):
model = Author
fields = ('name', 'age')
list_display = ('name', 'age', 'half_age', 'double_age')
list_url = reverse_lazy("authors")
new_url = reverse_lazy("new-author")
"""
form_class = AuthorForm
list_permission_required = ('tests.add_author',)
create_permission_required = ('tests.add_author',)
update_permission_required = ('tests.change_author',)
delete_permission_required = ('tests.delete_author',)
"""
def half_age(self, author):
return int(author.age/2)
half_age.label = "Half Age"
def get_edit_url(self, obj):
return reverse("edit-author", kwargs={'pk': obj.pk})
def get_delete_url(self, obj):
return reverse("delete-author", kwargs={'pk': obj.pk})
|
<commit_before>from django.core.urlresolvers import reverse_lazy, reverse
from django import forms
from .models import Author
from popupcrud.views import PopupCrudViewSet
# Create your views here.
class AuthorForm(forms.ModelForm):
sex = forms.ChoiceField(label="Sex", choices=(('M', 'Male'), ('F', 'Female')))
class Meta:
model = Author
fields = ('name', 'age')
class AuthorCrudViewset(PopupCrudViewSet):
model = Author
fields = ('name', 'age')
list_display = ('name', 'age', 'half_age', 'double_age')
list_url = reverse_lazy("authors")
new_url = reverse_lazy("new-author")
"""
form_class = AuthorForm
list_permission_required = ('tests.add_author',)
create_permission_required = ('tests.add_author',)
update_permission_required = ('tests.change_author',)
delete_permission_required = ('tests.delete_author',)
"""
def half_age(self, author):
return author.age/2
half_age.label = "Half Age"
def get_edit_url(self, obj):
return reverse("edit-author", kwargs={'pk': obj.pk})
def get_delete_url(self, obj):
return reverse("delete-author", kwargs={'pk': obj.pk})
<commit_msg>Handle PY3 default data conversion vagaries in unit test<commit_after>
|
from django.core.urlresolvers import reverse_lazy, reverse
from django import forms
from .models import Author
from popupcrud.views import PopupCrudViewSet
# Create your views here.
class AuthorForm(forms.ModelForm):
sex = forms.ChoiceField(label="Sex", choices=(('M', 'Male'), ('F', 'Female')))
class Meta:
model = Author
fields = ('name', 'age')
class AuthorCrudViewset(PopupCrudViewSet):
model = Author
fields = ('name', 'age')
list_display = ('name', 'age', 'half_age', 'double_age')
list_url = reverse_lazy("authors")
new_url = reverse_lazy("new-author")
"""
form_class = AuthorForm
list_permission_required = ('tests.add_author',)
create_permission_required = ('tests.add_author',)
update_permission_required = ('tests.change_author',)
delete_permission_required = ('tests.delete_author',)
"""
def half_age(self, author):
return int(author.age/2)
half_age.label = "Half Age"
def get_edit_url(self, obj):
return reverse("edit-author", kwargs={'pk': obj.pk})
def get_delete_url(self, obj):
return reverse("delete-author", kwargs={'pk': obj.pk})
|
from django.core.urlresolvers import reverse_lazy, reverse
from django import forms
from .models import Author
from popupcrud.views import PopupCrudViewSet
# Create your views here.
class AuthorForm(forms.ModelForm):
sex = forms.ChoiceField(label="Sex", choices=(('M', 'Male'), ('F', 'Female')))
class Meta:
model = Author
fields = ('name', 'age')
class AuthorCrudViewset(PopupCrudViewSet):
model = Author
fields = ('name', 'age')
list_display = ('name', 'age', 'half_age', 'double_age')
list_url = reverse_lazy("authors")
new_url = reverse_lazy("new-author")
"""
form_class = AuthorForm
list_permission_required = ('tests.add_author',)
create_permission_required = ('tests.add_author',)
update_permission_required = ('tests.change_author',)
delete_permission_required = ('tests.delete_author',)
"""
def half_age(self, author):
return author.age/2
half_age.label = "Half Age"
def get_edit_url(self, obj):
return reverse("edit-author", kwargs={'pk': obj.pk})
def get_delete_url(self, obj):
return reverse("delete-author", kwargs={'pk': obj.pk})
Handle PY3 default data conversion vagaries in unit testfrom django.core.urlresolvers import reverse_lazy, reverse
from django import forms
from .models import Author
from popupcrud.views import PopupCrudViewSet
# Create your views here.
class AuthorForm(forms.ModelForm):
sex = forms.ChoiceField(label="Sex", choices=(('M', 'Male'), ('F', 'Female')))
class Meta:
model = Author
fields = ('name', 'age')
class AuthorCrudViewset(PopupCrudViewSet):
model = Author
fields = ('name', 'age')
list_display = ('name', 'age', 'half_age', 'double_age')
list_url = reverse_lazy("authors")
new_url = reverse_lazy("new-author")
"""
form_class = AuthorForm
list_permission_required = ('tests.add_author',)
create_permission_required = ('tests.add_author',)
update_permission_required = ('tests.change_author',)
delete_permission_required = ('tests.delete_author',)
"""
def half_age(self, author):
return int(author.age/2)
half_age.label = "Half Age"
def get_edit_url(self, obj):
return reverse("edit-author", kwargs={'pk': obj.pk})
def get_delete_url(self, obj):
return reverse("delete-author", kwargs={'pk': obj.pk})
|
<commit_before>from django.core.urlresolvers import reverse_lazy, reverse
from django import forms
from .models import Author
from popupcrud.views import PopupCrudViewSet
# Create your views here.
class AuthorForm(forms.ModelForm):
sex = forms.ChoiceField(label="Sex", choices=(('M', 'Male'), ('F', 'Female')))
class Meta:
model = Author
fields = ('name', 'age')
class AuthorCrudViewset(PopupCrudViewSet):
model = Author
fields = ('name', 'age')
list_display = ('name', 'age', 'half_age', 'double_age')
list_url = reverse_lazy("authors")
new_url = reverse_lazy("new-author")
"""
form_class = AuthorForm
list_permission_required = ('tests.add_author',)
create_permission_required = ('tests.add_author',)
update_permission_required = ('tests.change_author',)
delete_permission_required = ('tests.delete_author',)
"""
def half_age(self, author):
return author.age/2
half_age.label = "Half Age"
def get_edit_url(self, obj):
return reverse("edit-author", kwargs={'pk': obj.pk})
def get_delete_url(self, obj):
return reverse("delete-author", kwargs={'pk': obj.pk})
<commit_msg>Handle PY3 default data conversion vagaries in unit test<commit_after>from django.core.urlresolvers import reverse_lazy, reverse
from django import forms
from .models import Author
from popupcrud.views import PopupCrudViewSet
# Create your views here.
class AuthorForm(forms.ModelForm):
sex = forms.ChoiceField(label="Sex", choices=(('M', 'Male'), ('F', 'Female')))
class Meta:
model = Author
fields = ('name', 'age')
class AuthorCrudViewset(PopupCrudViewSet):
model = Author
fields = ('name', 'age')
list_display = ('name', 'age', 'half_age', 'double_age')
list_url = reverse_lazy("authors")
new_url = reverse_lazy("new-author")
"""
form_class = AuthorForm
list_permission_required = ('tests.add_author',)
create_permission_required = ('tests.add_author',)
update_permission_required = ('tests.change_author',)
delete_permission_required = ('tests.delete_author',)
"""
def half_age(self, author):
return int(author.age/2)
half_age.label = "Half Age"
def get_edit_url(self, obj):
return reverse("edit-author", kwargs={'pk': obj.pk})
def get_delete_url(self, obj):
return reverse("delete-author", kwargs={'pk': obj.pk})
|
e28c0bd84dc8814654850b607afbdaeb669956c8
|
tests/data.py
|
tests/data.py
|
# -*- coding: utf-8 -*-
import json as _json
from collections import OrderedDict as _OrderedDict
import os as _os
_thisdir = _os.path.dirname(__file__)
class MenuData(object):
_data_files = [
'data.json',
]
def __init__(self, *args, **kwargs):
dfiles = [_os.path.join(_thisdir, x) for x in self._data_files]
for fpath in dfiles:
with open(fpath, 'rb') as fdata:
dat = _json.load(fdata)
for k,v in dat.items():
setattr(self, k, v)
self.people = _OrderedDict(self.people)
|
# -*- coding: utf-8 -*-
import json as _json
from collections import OrderedDict as _OrderedDict
import os as _os
_thisdir = _os.path.dirname(__file__)
class MenuData(object):
_data_files = [
'data.json',
]
def __init__(self, *args, **kwargs):
dfiles = [_os.path.join(_thisdir, x) for x in self._data_files]
for fpath in dfiles:
with open(fpath, 'rt') as fdata:
dat = _json.load(fdata)
for k,v in dat.items():
setattr(self, k, v)
self.people = _OrderedDict(self.people)
|
Fix json bug for python35 in tests
|
Fix json bug for python35 in tests
|
Python
|
mit
|
frostidaho/dynmen
|
# -*- coding: utf-8 -*-
import json as _json
from collections import OrderedDict as _OrderedDict
import os as _os
_thisdir = _os.path.dirname(__file__)
class MenuData(object):
_data_files = [
'data.json',
]
def __init__(self, *args, **kwargs):
dfiles = [_os.path.join(_thisdir, x) for x in self._data_files]
for fpath in dfiles:
with open(fpath, 'rb') as fdata:
dat = _json.load(fdata)
for k,v in dat.items():
setattr(self, k, v)
self.people = _OrderedDict(self.people)
Fix json bug for python35 in tests
|
# -*- coding: utf-8 -*-
import json as _json
from collections import OrderedDict as _OrderedDict
import os as _os
_thisdir = _os.path.dirname(__file__)
class MenuData(object):
_data_files = [
'data.json',
]
def __init__(self, *args, **kwargs):
dfiles = [_os.path.join(_thisdir, x) for x in self._data_files]
for fpath in dfiles:
with open(fpath, 'rt') as fdata:
dat = _json.load(fdata)
for k,v in dat.items():
setattr(self, k, v)
self.people = _OrderedDict(self.people)
|
<commit_before># -*- coding: utf-8 -*-
import json as _json
from collections import OrderedDict as _OrderedDict
import os as _os
_thisdir = _os.path.dirname(__file__)
class MenuData(object):
_data_files = [
'data.json',
]
def __init__(self, *args, **kwargs):
dfiles = [_os.path.join(_thisdir, x) for x in self._data_files]
for fpath in dfiles:
with open(fpath, 'rb') as fdata:
dat = _json.load(fdata)
for k,v in dat.items():
setattr(self, k, v)
self.people = _OrderedDict(self.people)
<commit_msg>Fix json bug for python35 in tests<commit_after>
|
# -*- coding: utf-8 -*-
import json as _json
from collections import OrderedDict as _OrderedDict
import os as _os
_thisdir = _os.path.dirname(__file__)
class MenuData(object):
_data_files = [
'data.json',
]
def __init__(self, *args, **kwargs):
dfiles = [_os.path.join(_thisdir, x) for x in self._data_files]
for fpath in dfiles:
with open(fpath, 'rt') as fdata:
dat = _json.load(fdata)
for k,v in dat.items():
setattr(self, k, v)
self.people = _OrderedDict(self.people)
|
# -*- coding: utf-8 -*-
import json as _json
from collections import OrderedDict as _OrderedDict
import os as _os
_thisdir = _os.path.dirname(__file__)
class MenuData(object):
_data_files = [
'data.json',
]
def __init__(self, *args, **kwargs):
dfiles = [_os.path.join(_thisdir, x) for x in self._data_files]
for fpath in dfiles:
with open(fpath, 'rb') as fdata:
dat = _json.load(fdata)
for k,v in dat.items():
setattr(self, k, v)
self.people = _OrderedDict(self.people)
Fix json bug for python35 in tests# -*- coding: utf-8 -*-
import json as _json
from collections import OrderedDict as _OrderedDict
import os as _os
_thisdir = _os.path.dirname(__file__)
class MenuData(object):
_data_files = [
'data.json',
]
def __init__(self, *args, **kwargs):
dfiles = [_os.path.join(_thisdir, x) for x in self._data_files]
for fpath in dfiles:
with open(fpath, 'rt') as fdata:
dat = _json.load(fdata)
for k,v in dat.items():
setattr(self, k, v)
self.people = _OrderedDict(self.people)
|
<commit_before># -*- coding: utf-8 -*-
import json as _json
from collections import OrderedDict as _OrderedDict
import os as _os
_thisdir = _os.path.dirname(__file__)
class MenuData(object):
_data_files = [
'data.json',
]
def __init__(self, *args, **kwargs):
dfiles = [_os.path.join(_thisdir, x) for x in self._data_files]
for fpath in dfiles:
with open(fpath, 'rb') as fdata:
dat = _json.load(fdata)
for k,v in dat.items():
setattr(self, k, v)
self.people = _OrderedDict(self.people)
<commit_msg>Fix json bug for python35 in tests<commit_after># -*- coding: utf-8 -*-
import json as _json
from collections import OrderedDict as _OrderedDict
import os as _os
_thisdir = _os.path.dirname(__file__)
class MenuData(object):
_data_files = [
'data.json',
]
def __init__(self, *args, **kwargs):
dfiles = [_os.path.join(_thisdir, x) for x in self._data_files]
for fpath in dfiles:
with open(fpath, 'rt') as fdata:
dat = _json.load(fdata)
for k,v in dat.items():
setattr(self, k, v)
self.people = _OrderedDict(self.people)
|
52fdf6c0183233d34bc987e27a5e727b71ba09f0
|
src/main/translator-xml/PMLToXML.py
|
src/main/translator-xml/PMLToXML.py
|
#!/usr/bin/env/python
import sys
import os.path
import subprocess
# Read in a pml file and save to an xml file
def translate_pml_file(xml_file, pml_file):
pml_path = os.path.abspath(pml_file.name)
xml_path = os.path.abspath(xml_file.name)
# Call XML generator
# TODO: Remove abs-path
return_code = subprocess.call("/opt/pml-bnfc/xml/Pmlxml %s %s" % (xml_path, pml_path), shell=True)
if return_code != 0:
print "Error occured reading PML file, exiting."
sys.exit(1)
def main():
import argparse
parser = argparse.ArgumentParser(description="Program to output the ast of a PML program in XML format")
parser.add_argument('-x', '--xml', required=True, type=file, help="Output abstract syntax tree in XML format")
parser.add_argument('-p', '--pml', required=True, type=file, help="Input PML file")
try:
args = parser.parse_args()
translate_pml_file(args.xml, args.pml)
except IOError, msg:
parser.error(str(msg))
if __name__ == "__main__":
main()
|
#!/usr/bin/env/python
import sys
from os.path import isdir, split, abspath
import subprocess
# Read in a pml file and save to an xml file
def translate_pml_file(xml_file, pml_file):
pml_path = abspath(pml_file.name)
xml_path = abspath(xml_file)
# Call XML generator
# TODO: Remove abs-path
return_code = subprocess.call("/opt/pml-bnfc/xml/Pmlxml %s %s" % (xml_path, pml_path), shell=True)
if return_code != 0:
print "Error occured reading PML file, exiting."
sys.exit(1)
def valid_xml_path(xml_path):
path, _ = split(xml_path)
return isdir(path)
def main():
import argparse
parser = argparse.ArgumentParser(description="Program to output the ast of a PML program in XML format")
parser.add_argument('-x', '--xml', required=True, type=str, help="Output abstract syntax tree in XML format")
parser.add_argument('-p', '--pml', required=True, type=file, help="Input PML file")
try:
args = parser.parse_args()
if valid_xml_path(abspath(args.xml)):
translate_pml_file(args.xml, args.pml)
else:
print "XML path is invalid."
sys.exit(1)
except IOError, msg:
parser.error(str(msg))
if __name__ == "__main__":
main()
|
Change arg from file to path
|
Change arg from file to path
|
Python
|
mit
|
CS4098/GroupProject,CS4098/GroupProject,CS4098/GroupProject
|
#!/usr/bin/env/python
import sys
import os.path
import subprocess
# Read in a pml file and save to an xml file
def translate_pml_file(xml_file, pml_file):
pml_path = os.path.abspath(pml_file.name)
xml_path = os.path.abspath(xml_file.name)
# Call XML generator
# TODO: Remove abs-path
return_code = subprocess.call("/opt/pml-bnfc/xml/Pmlxml %s %s" % (xml_path, pml_path), shell=True)
if return_code != 0:
print "Error occured reading PML file, exiting."
sys.exit(1)
def main():
import argparse
parser = argparse.ArgumentParser(description="Program to output the ast of a PML program in XML format")
parser.add_argument('-x', '--xml', required=True, type=file, help="Output abstract syntax tree in XML format")
parser.add_argument('-p', '--pml', required=True, type=file, help="Input PML file")
try:
args = parser.parse_args()
translate_pml_file(args.xml, args.pml)
except IOError, msg:
parser.error(str(msg))
if __name__ == "__main__":
main()
Change arg from file to path
|
#!/usr/bin/env/python
import sys
from os.path import isdir, split, abspath
import subprocess
# Read in a pml file and save to an xml file
def translate_pml_file(xml_file, pml_file):
pml_path = abspath(pml_file.name)
xml_path = abspath(xml_file)
# Call XML generator
# TODO: Remove abs-path
return_code = subprocess.call("/opt/pml-bnfc/xml/Pmlxml %s %s" % (xml_path, pml_path), shell=True)
if return_code != 0:
print "Error occured reading PML file, exiting."
sys.exit(1)
def valid_xml_path(xml_path):
path, _ = split(xml_path)
return isdir(path)
def main():
import argparse
parser = argparse.ArgumentParser(description="Program to output the ast of a PML program in XML format")
parser.add_argument('-x', '--xml', required=True, type=str, help="Output abstract syntax tree in XML format")
parser.add_argument('-p', '--pml', required=True, type=file, help="Input PML file")
try:
args = parser.parse_args()
if valid_xml_path(abspath(args.xml)):
translate_pml_file(args.xml, args.pml)
else:
print "XML path is invalid."
sys.exit(1)
except IOError, msg:
parser.error(str(msg))
if __name__ == "__main__":
main()
|
<commit_before>#!/usr/bin/env/python
import sys
import os.path
import subprocess
# Read in a pml file and save to an xml file
def translate_pml_file(xml_file, pml_file):
pml_path = os.path.abspath(pml_file.name)
xml_path = os.path.abspath(xml_file.name)
# Call XML generator
# TODO: Remove abs-path
return_code = subprocess.call("/opt/pml-bnfc/xml/Pmlxml %s %s" % (xml_path, pml_path), shell=True)
if return_code != 0:
print "Error occured reading PML file, exiting."
sys.exit(1)
def main():
import argparse
parser = argparse.ArgumentParser(description="Program to output the ast of a PML program in XML format")
parser.add_argument('-x', '--xml', required=True, type=file, help="Output abstract syntax tree in XML format")
parser.add_argument('-p', '--pml', required=True, type=file, help="Input PML file")
try:
args = parser.parse_args()
translate_pml_file(args.xml, args.pml)
except IOError, msg:
parser.error(str(msg))
if __name__ == "__main__":
main()
<commit_msg>Change arg from file to path<commit_after>
|
#!/usr/bin/env/python
import sys
from os.path import isdir, split, abspath
import subprocess
# Read in a pml file and save to an xml file
def translate_pml_file(xml_file, pml_file):
pml_path = abspath(pml_file.name)
xml_path = abspath(xml_file)
# Call XML generator
# TODO: Remove abs-path
return_code = subprocess.call("/opt/pml-bnfc/xml/Pmlxml %s %s" % (xml_path, pml_path), shell=True)
if return_code != 0:
print "Error occured reading PML file, exiting."
sys.exit(1)
def valid_xml_path(xml_path):
path, _ = split(xml_path)
return isdir(path)
def main():
import argparse
parser = argparse.ArgumentParser(description="Program to output the ast of a PML program in XML format")
parser.add_argument('-x', '--xml', required=True, type=str, help="Output abstract syntax tree in XML format")
parser.add_argument('-p', '--pml', required=True, type=file, help="Input PML file")
try:
args = parser.parse_args()
if valid_xml_path(abspath(args.xml)):
translate_pml_file(args.xml, args.pml)
else:
print "XML path is invalid."
sys.exit(1)
except IOError, msg:
parser.error(str(msg))
if __name__ == "__main__":
main()
|
#!/usr/bin/env/python
import sys
import os.path
import subprocess
# Read in a pml file and save to an xml file
def translate_pml_file(xml_file, pml_file):
pml_path = os.path.abspath(pml_file.name)
xml_path = os.path.abspath(xml_file.name)
# Call XML generator
# TODO: Remove abs-path
return_code = subprocess.call("/opt/pml-bnfc/xml/Pmlxml %s %s" % (xml_path, pml_path), shell=True)
if return_code != 0:
print "Error occured reading PML file, exiting."
sys.exit(1)
def main():
import argparse
parser = argparse.ArgumentParser(description="Program to output the ast of a PML program in XML format")
parser.add_argument('-x', '--xml', required=True, type=file, help="Output abstract syntax tree in XML format")
parser.add_argument('-p', '--pml', required=True, type=file, help="Input PML file")
try:
args = parser.parse_args()
translate_pml_file(args.xml, args.pml)
except IOError, msg:
parser.error(str(msg))
if __name__ == "__main__":
main()
Change arg from file to path#!/usr/bin/env/python
import sys
from os.path import isdir, split, abspath
import subprocess
# Read in a pml file and save to an xml file
def translate_pml_file(xml_file, pml_file):
pml_path = abspath(pml_file.name)
xml_path = abspath(xml_file)
# Call XML generator
# TODO: Remove abs-path
return_code = subprocess.call("/opt/pml-bnfc/xml/Pmlxml %s %s" % (xml_path, pml_path), shell=True)
if return_code != 0:
print "Error occured reading PML file, exiting."
sys.exit(1)
def valid_xml_path(xml_path):
path, _ = split(xml_path)
return isdir(path)
def main():
import argparse
parser = argparse.ArgumentParser(description="Program to output the ast of a PML program in XML format")
parser.add_argument('-x', '--xml', required=True, type=str, help="Output abstract syntax tree in XML format")
parser.add_argument('-p', '--pml', required=True, type=file, help="Input PML file")
try:
args = parser.parse_args()
if valid_xml_path(abspath(args.xml)):
translate_pml_file(args.xml, args.pml)
else:
print "XML path is invalid."
sys.exit(1)
except IOError, msg:
parser.error(str(msg))
if __name__ == "__main__":
main()
|
<commit_before>#!/usr/bin/env/python
import sys
import os.path
import subprocess
# Read in a pml file and save to an xml file
def translate_pml_file(xml_file, pml_file):
pml_path = os.path.abspath(pml_file.name)
xml_path = os.path.abspath(xml_file.name)
# Call XML generator
# TODO: Remove abs-path
return_code = subprocess.call("/opt/pml-bnfc/xml/Pmlxml %s %s" % (xml_path, pml_path), shell=True)
if return_code != 0:
print "Error occured reading PML file, exiting."
sys.exit(1)
def main():
import argparse
parser = argparse.ArgumentParser(description="Program to output the ast of a PML program in XML format")
parser.add_argument('-x', '--xml', required=True, type=file, help="Output abstract syntax tree in XML format")
parser.add_argument('-p', '--pml', required=True, type=file, help="Input PML file")
try:
args = parser.parse_args()
translate_pml_file(args.xml, args.pml)
except IOError, msg:
parser.error(str(msg))
if __name__ == "__main__":
main()
<commit_msg>Change arg from file to path<commit_after>#!/usr/bin/env/python
import sys
from os.path import isdir, split, abspath
import subprocess
# Read in a pml file and save to an xml file
def translate_pml_file(xml_file, pml_file):
pml_path = abspath(pml_file.name)
xml_path = abspath(xml_file)
# Call XML generator
# TODO: Remove abs-path
return_code = subprocess.call("/opt/pml-bnfc/xml/Pmlxml %s %s" % (xml_path, pml_path), shell=True)
if return_code != 0:
print "Error occured reading PML file, exiting."
sys.exit(1)
def valid_xml_path(xml_path):
path, _ = split(xml_path)
return isdir(path)
def main():
import argparse
parser = argparse.ArgumentParser(description="Program to output the ast of a PML program in XML format")
parser.add_argument('-x', '--xml', required=True, type=str, help="Output abstract syntax tree in XML format")
parser.add_argument('-p', '--pml', required=True, type=file, help="Input PML file")
try:
args = parser.parse_args()
if valid_xml_path(abspath(args.xml)):
translate_pml_file(args.xml, args.pml)
else:
print "XML path is invalid."
sys.exit(1)
except IOError, msg:
parser.error(str(msg))
if __name__ == "__main__":
main()
|
a5a90924822754b483041ba29cefeba949e72f38
|
securesystemslib/gpg/exceptions.py
|
securesystemslib/gpg/exceptions.py
|
"""
<Program Name>
exceptions.py
<Author>
Santiago Torres-Arias <santiago@nyu.edu>
Lukas Puehringer <lukas.puehringer@nyu.edu>
<Started>
Dec 8, 2017
<Copyright>
See LICENSE for licensing information.
<Purpose>
Define Exceptions used in the gpg package. Following the practice from
securesystemslib the names chosen for exception classes should end in
'Error' (except where there is a good reason not to).
"""
class PacketParsingError(Exception):
pass
|
"""
<Program Name>
exceptions.py
<Author>
Santiago Torres-Arias <santiago@nyu.edu>
Lukas Puehringer <lukas.puehringer@nyu.edu>
<Started>
Dec 8, 2017
<Copyright>
See LICENSE for licensing information.
<Purpose>
Define Exceptions used in the gpg package. Following the practice from
securesystemslib the names chosen for exception classes should end in
'Error' (except where there is a good reason not to).
"""
class PacketParsingError(Exception):
pass
class KeyNotFoundError(Exception):
pass
|
Add custom KeyNotFoundError error to gpg module
|
Add custom KeyNotFoundError error to gpg module
|
Python
|
mit
|
secure-systems-lab/securesystemslib,secure-systems-lab/securesystemslib
|
"""
<Program Name>
exceptions.py
<Author>
Santiago Torres-Arias <santiago@nyu.edu>
Lukas Puehringer <lukas.puehringer@nyu.edu>
<Started>
Dec 8, 2017
<Copyright>
See LICENSE for licensing information.
<Purpose>
Define Exceptions used in the gpg package. Following the practice from
securesystemslib the names chosen for exception classes should end in
'Error' (except where there is a good reason not to).
"""
class PacketParsingError(Exception):
pass
Add custom KeyNotFoundError error to gpg module
|
"""
<Program Name>
exceptions.py
<Author>
Santiago Torres-Arias <santiago@nyu.edu>
Lukas Puehringer <lukas.puehringer@nyu.edu>
<Started>
Dec 8, 2017
<Copyright>
See LICENSE for licensing information.
<Purpose>
Define Exceptions used in the gpg package. Following the practice from
securesystemslib the names chosen for exception classes should end in
'Error' (except where there is a good reason not to).
"""
class PacketParsingError(Exception):
pass
class KeyNotFoundError(Exception):
pass
|
<commit_before>"""
<Program Name>
exceptions.py
<Author>
Santiago Torres-Arias <santiago@nyu.edu>
Lukas Puehringer <lukas.puehringer@nyu.edu>
<Started>
Dec 8, 2017
<Copyright>
See LICENSE for licensing information.
<Purpose>
Define Exceptions used in the gpg package. Following the practice from
securesystemslib the names chosen for exception classes should end in
'Error' (except where there is a good reason not to).
"""
class PacketParsingError(Exception):
pass
<commit_msg>Add custom KeyNotFoundError error to gpg module<commit_after>
|
"""
<Program Name>
exceptions.py
<Author>
Santiago Torres-Arias <santiago@nyu.edu>
Lukas Puehringer <lukas.puehringer@nyu.edu>
<Started>
Dec 8, 2017
<Copyright>
See LICENSE for licensing information.
<Purpose>
Define Exceptions used in the gpg package. Following the practice from
securesystemslib the names chosen for exception classes should end in
'Error' (except where there is a good reason not to).
"""
class PacketParsingError(Exception):
pass
class KeyNotFoundError(Exception):
pass
|
"""
<Program Name>
exceptions.py
<Author>
Santiago Torres-Arias <santiago@nyu.edu>
Lukas Puehringer <lukas.puehringer@nyu.edu>
<Started>
Dec 8, 2017
<Copyright>
See LICENSE for licensing information.
<Purpose>
Define Exceptions used in the gpg package. Following the practice from
securesystemslib the names chosen for exception classes should end in
'Error' (except where there is a good reason not to).
"""
class PacketParsingError(Exception):
pass
Add custom KeyNotFoundError error to gpg module"""
<Program Name>
exceptions.py
<Author>
Santiago Torres-Arias <santiago@nyu.edu>
Lukas Puehringer <lukas.puehringer@nyu.edu>
<Started>
Dec 8, 2017
<Copyright>
See LICENSE for licensing information.
<Purpose>
Define Exceptions used in the gpg package. Following the practice from
securesystemslib the names chosen for exception classes should end in
'Error' (except where there is a good reason not to).
"""
class PacketParsingError(Exception):
pass
class KeyNotFoundError(Exception):
pass
|
<commit_before>"""
<Program Name>
exceptions.py
<Author>
Santiago Torres-Arias <santiago@nyu.edu>
Lukas Puehringer <lukas.puehringer@nyu.edu>
<Started>
Dec 8, 2017
<Copyright>
See LICENSE for licensing information.
<Purpose>
Define Exceptions used in the gpg package. Following the practice from
securesystemslib the names chosen for exception classes should end in
'Error' (except where there is a good reason not to).
"""
class PacketParsingError(Exception):
pass
<commit_msg>Add custom KeyNotFoundError error to gpg module<commit_after>"""
<Program Name>
exceptions.py
<Author>
Santiago Torres-Arias <santiago@nyu.edu>
Lukas Puehringer <lukas.puehringer@nyu.edu>
<Started>
Dec 8, 2017
<Copyright>
See LICENSE for licensing information.
<Purpose>
Define Exceptions used in the gpg package. Following the practice from
securesystemslib the names chosen for exception classes should end in
'Error' (except where there is a good reason not to).
"""
class PacketParsingError(Exception):
pass
class KeyNotFoundError(Exception):
pass
|
dcbcd7434b8b4199242a479d187d2b833ca6ffcc
|
polling_stations/settings/constants/councils.py
|
polling_stations/settings/constants/councils.py
|
# settings for councils scraper
YVM_LA_URL = "https://www.yourvotematters.co.uk/_design/nested-content/results-page2/search-voting-locations-by-districtcode?queries_distcode_query=" # noqa
BOUNDARIES_URL = "https://ons-cache.s3.amazonaws.com/Local_Authority_Districts_April_2019_Boundaries_UK_BFE.geojson"
EC_COUNCIL_CONTACT_DETAILS_API_URL = ""
OLD_TO_NEW_MAP = {}
NEW_COUNCILS = []
|
# settings for councils scraper
YVM_LA_URL = "https://www.yourvotematters.co.uk/_design/nested-content/results-page2/search-voting-locations-by-districtcode?queries_distcode_query=" # noqa
BOUNDARIES_URL = "https://ons-cache.s3.amazonaws.com/Local_Authority_Districts_April_2019_Boundaries_UK_BFE.geojson"
EC_COUNCIL_CONTACT_DETAILS_API_URL = (
"https://electoralcommission.org.uk/api/v1/data/local-authorities.json"
)
OLD_TO_NEW_MAP = {}
NEW_COUNCILS = []
|
Set the EC API URL
|
Set the EC API URL
|
Python
|
bsd-3-clause
|
DemocracyClub/UK-Polling-Stations,DemocracyClub/UK-Polling-Stations,DemocracyClub/UK-Polling-Stations
|
# settings for councils scraper
YVM_LA_URL = "https://www.yourvotematters.co.uk/_design/nested-content/results-page2/search-voting-locations-by-districtcode?queries_distcode_query=" # noqa
BOUNDARIES_URL = "https://ons-cache.s3.amazonaws.com/Local_Authority_Districts_April_2019_Boundaries_UK_BFE.geojson"
EC_COUNCIL_CONTACT_DETAILS_API_URL = ""
OLD_TO_NEW_MAP = {}
NEW_COUNCILS = []
Set the EC API URL
|
# settings for councils scraper
YVM_LA_URL = "https://www.yourvotematters.co.uk/_design/nested-content/results-page2/search-voting-locations-by-districtcode?queries_distcode_query=" # noqa
BOUNDARIES_URL = "https://ons-cache.s3.amazonaws.com/Local_Authority_Districts_April_2019_Boundaries_UK_BFE.geojson"
EC_COUNCIL_CONTACT_DETAILS_API_URL = (
"https://electoralcommission.org.uk/api/v1/data/local-authorities.json"
)
OLD_TO_NEW_MAP = {}
NEW_COUNCILS = []
|
<commit_before># settings for councils scraper
YVM_LA_URL = "https://www.yourvotematters.co.uk/_design/nested-content/results-page2/search-voting-locations-by-districtcode?queries_distcode_query=" # noqa
BOUNDARIES_URL = "https://ons-cache.s3.amazonaws.com/Local_Authority_Districts_April_2019_Boundaries_UK_BFE.geojson"
EC_COUNCIL_CONTACT_DETAILS_API_URL = ""
OLD_TO_NEW_MAP = {}
NEW_COUNCILS = []
<commit_msg>Set the EC API URL<commit_after>
|
# settings for councils scraper
YVM_LA_URL = "https://www.yourvotematters.co.uk/_design/nested-content/results-page2/search-voting-locations-by-districtcode?queries_distcode_query=" # noqa
BOUNDARIES_URL = "https://ons-cache.s3.amazonaws.com/Local_Authority_Districts_April_2019_Boundaries_UK_BFE.geojson"
EC_COUNCIL_CONTACT_DETAILS_API_URL = (
"https://electoralcommission.org.uk/api/v1/data/local-authorities.json"
)
OLD_TO_NEW_MAP = {}
NEW_COUNCILS = []
|
# settings for councils scraper
YVM_LA_URL = "https://www.yourvotematters.co.uk/_design/nested-content/results-page2/search-voting-locations-by-districtcode?queries_distcode_query=" # noqa
BOUNDARIES_URL = "https://ons-cache.s3.amazonaws.com/Local_Authority_Districts_April_2019_Boundaries_UK_BFE.geojson"
EC_COUNCIL_CONTACT_DETAILS_API_URL = ""
OLD_TO_NEW_MAP = {}
NEW_COUNCILS = []
Set the EC API URL# settings for councils scraper
YVM_LA_URL = "https://www.yourvotematters.co.uk/_design/nested-content/results-page2/search-voting-locations-by-districtcode?queries_distcode_query=" # noqa
BOUNDARIES_URL = "https://ons-cache.s3.amazonaws.com/Local_Authority_Districts_April_2019_Boundaries_UK_BFE.geojson"
EC_COUNCIL_CONTACT_DETAILS_API_URL = (
"https://electoralcommission.org.uk/api/v1/data/local-authorities.json"
)
OLD_TO_NEW_MAP = {}
NEW_COUNCILS = []
|
<commit_before># settings for councils scraper
YVM_LA_URL = "https://www.yourvotematters.co.uk/_design/nested-content/results-page2/search-voting-locations-by-districtcode?queries_distcode_query=" # noqa
BOUNDARIES_URL = "https://ons-cache.s3.amazonaws.com/Local_Authority_Districts_April_2019_Boundaries_UK_BFE.geojson"
EC_COUNCIL_CONTACT_DETAILS_API_URL = ""
OLD_TO_NEW_MAP = {}
NEW_COUNCILS = []
<commit_msg>Set the EC API URL<commit_after># settings for councils scraper
YVM_LA_URL = "https://www.yourvotematters.co.uk/_design/nested-content/results-page2/search-voting-locations-by-districtcode?queries_distcode_query=" # noqa
BOUNDARIES_URL = "https://ons-cache.s3.amazonaws.com/Local_Authority_Districts_April_2019_Boundaries_UK_BFE.geojson"
EC_COUNCIL_CONTACT_DETAILS_API_URL = (
"https://electoralcommission.org.uk/api/v1/data/local-authorities.json"
)
OLD_TO_NEW_MAP = {}
NEW_COUNCILS = []
|
4e31e5c776c40997cccd76d4ce592d7f3d5de752
|
example/runner.py
|
example/runner.py
|
#!/usr/bin/python
import argparse
import sys
def args():
parser = argparse.ArgumentParser(description='Run the Furious Examples.')
parser.add_argument('--gae-sdk-path', metavar='S', dest="gae_lib_path",
default="/usr/local/google_appengine",
help='path to the GAE SDK')
parser.add_argument('--url', metavar='U', dest="url", default="",
help="the endpoint to run")
return parser.parse_args()
def setup(options):
sys.path.insert(0, options.gae_lib_path)
from dev_appserver import fix_sys_path
fix_sys_path()
def run(options):
from google.appengine.tools import appengine_rpc
from google.appengine.tools import appcfg
source = 'furious'
user_agent = appcfg.GetUserAgent()
server = appengine_rpc.HttpRpcServer(
'localhost:8080', lambda: ('test@example.com', 'password'), user_agent,
source, secure=False)
server._DevAppServerAuthenticate()
server.Send(options.url, content_type="text/html; charset=utf-8",
payload=None)
def main():
options = args()
setup(options)
run(options)
if __name__ == "__main__":
main()
|
#!/usr/bin/python
import argparse
import sys
def args():
parser = argparse.ArgumentParser(description='Run the Furious Examples.')
parser.add_argument('--gae-sdk-path', metavar='S', dest="gae_lib_path",
default="/usr/local/google_appengine",
help='path to the GAE SDK')
parser.add_argument('url', metavar='U', default="", nargs=1,
help="the endpoint to run")
return parser.parse_args()
def setup(options):
sys.path.insert(0, options.gae_lib_path)
from dev_appserver import fix_sys_path
fix_sys_path()
def run(options):
from google.appengine.tools import appengine_rpc
from google.appengine.tools import appcfg
source = 'furious'
user_agent = appcfg.GetUserAgent()
server = appengine_rpc.HttpRpcServer(
'localhost:8080', lambda: ('test@example.com', 'password'), user_agent,
source, secure=False)
url = "/"
if options.url:
url += options.url[0]
server._DevAppServerAuthenticate()
server.Send(url, content_type="text/html; charset=utf-8",
payload=None)
def main():
options = args()
setup(options)
run(options)
if __name__ == "__main__":
main()
|
Update the way the url is handled.
|
Update the way the url is handled.
|
Python
|
apache-2.0
|
andreleblanc-wf/furious,Workiva/furious,rosshendrickson-wf/furious,beaulyddon-wf/furious,mattsanders-wf/furious,rosshendrickson-wf/furious,mattsanders-wf/furious,beaulyddon-wf/furious,andreleblanc-wf/furious,Workiva/furious
|
#!/usr/bin/python
import argparse
import sys
def args():
parser = argparse.ArgumentParser(description='Run the Furious Examples.')
parser.add_argument('--gae-sdk-path', metavar='S', dest="gae_lib_path",
default="/usr/local/google_appengine",
help='path to the GAE SDK')
parser.add_argument('--url', metavar='U', dest="url", default="",
help="the endpoint to run")
return parser.parse_args()
def setup(options):
sys.path.insert(0, options.gae_lib_path)
from dev_appserver import fix_sys_path
fix_sys_path()
def run(options):
from google.appengine.tools import appengine_rpc
from google.appengine.tools import appcfg
source = 'furious'
user_agent = appcfg.GetUserAgent()
server = appengine_rpc.HttpRpcServer(
'localhost:8080', lambda: ('test@example.com', 'password'), user_agent,
source, secure=False)
server._DevAppServerAuthenticate()
server.Send(options.url, content_type="text/html; charset=utf-8",
payload=None)
def main():
options = args()
setup(options)
run(options)
if __name__ == "__main__":
main()
Update the way the url is handled.
|
#!/usr/bin/python
import argparse
import sys
def args():
parser = argparse.ArgumentParser(description='Run the Furious Examples.')
parser.add_argument('--gae-sdk-path', metavar='S', dest="gae_lib_path",
default="/usr/local/google_appengine",
help='path to the GAE SDK')
parser.add_argument('url', metavar='U', default="", nargs=1,
help="the endpoint to run")
return parser.parse_args()
def setup(options):
sys.path.insert(0, options.gae_lib_path)
from dev_appserver import fix_sys_path
fix_sys_path()
def run(options):
from google.appengine.tools import appengine_rpc
from google.appengine.tools import appcfg
source = 'furious'
user_agent = appcfg.GetUserAgent()
server = appengine_rpc.HttpRpcServer(
'localhost:8080', lambda: ('test@example.com', 'password'), user_agent,
source, secure=False)
url = "/"
if options.url:
url += options.url[0]
server._DevAppServerAuthenticate()
server.Send(url, content_type="text/html; charset=utf-8",
payload=None)
def main():
options = args()
setup(options)
run(options)
if __name__ == "__main__":
main()
|
<commit_before>#!/usr/bin/python
import argparse
import sys
def args():
parser = argparse.ArgumentParser(description='Run the Furious Examples.')
parser.add_argument('--gae-sdk-path', metavar='S', dest="gae_lib_path",
default="/usr/local/google_appengine",
help='path to the GAE SDK')
parser.add_argument('--url', metavar='U', dest="url", default="",
help="the endpoint to run")
return parser.parse_args()
def setup(options):
sys.path.insert(0, options.gae_lib_path)
from dev_appserver import fix_sys_path
fix_sys_path()
def run(options):
from google.appengine.tools import appengine_rpc
from google.appengine.tools import appcfg
source = 'furious'
user_agent = appcfg.GetUserAgent()
server = appengine_rpc.HttpRpcServer(
'localhost:8080', lambda: ('test@example.com', 'password'), user_agent,
source, secure=False)
server._DevAppServerAuthenticate()
server.Send(options.url, content_type="text/html; charset=utf-8",
payload=None)
def main():
options = args()
setup(options)
run(options)
if __name__ == "__main__":
main()
<commit_msg>Update the way the url is handled.<commit_after>
|
#!/usr/bin/python
import argparse
import sys
def args():
parser = argparse.ArgumentParser(description='Run the Furious Examples.')
parser.add_argument('--gae-sdk-path', metavar='S', dest="gae_lib_path",
default="/usr/local/google_appengine",
help='path to the GAE SDK')
parser.add_argument('url', metavar='U', default="", nargs=1,
help="the endpoint to run")
return parser.parse_args()
def setup(options):
sys.path.insert(0, options.gae_lib_path)
from dev_appserver import fix_sys_path
fix_sys_path()
def run(options):
from google.appengine.tools import appengine_rpc
from google.appengine.tools import appcfg
source = 'furious'
user_agent = appcfg.GetUserAgent()
server = appengine_rpc.HttpRpcServer(
'localhost:8080', lambda: ('test@example.com', 'password'), user_agent,
source, secure=False)
url = "/"
if options.url:
url += options.url[0]
server._DevAppServerAuthenticate()
server.Send(url, content_type="text/html; charset=utf-8",
payload=None)
def main():
options = args()
setup(options)
run(options)
if __name__ == "__main__":
main()
|
#!/usr/bin/python
import argparse
import sys
def args():
parser = argparse.ArgumentParser(description='Run the Furious Examples.')
parser.add_argument('--gae-sdk-path', metavar='S', dest="gae_lib_path",
default="/usr/local/google_appengine",
help='path to the GAE SDK')
parser.add_argument('--url', metavar='U', dest="url", default="",
help="the endpoint to run")
return parser.parse_args()
def setup(options):
sys.path.insert(0, options.gae_lib_path)
from dev_appserver import fix_sys_path
fix_sys_path()
def run(options):
from google.appengine.tools import appengine_rpc
from google.appengine.tools import appcfg
source = 'furious'
user_agent = appcfg.GetUserAgent()
server = appengine_rpc.HttpRpcServer(
'localhost:8080', lambda: ('test@example.com', 'password'), user_agent,
source, secure=False)
server._DevAppServerAuthenticate()
server.Send(options.url, content_type="text/html; charset=utf-8",
payload=None)
def main():
options = args()
setup(options)
run(options)
if __name__ == "__main__":
main()
Update the way the url is handled.#!/usr/bin/python
import argparse
import sys
def args():
parser = argparse.ArgumentParser(description='Run the Furious Examples.')
parser.add_argument('--gae-sdk-path', metavar='S', dest="gae_lib_path",
default="/usr/local/google_appengine",
help='path to the GAE SDK')
parser.add_argument('url', metavar='U', default="", nargs=1,
help="the endpoint to run")
return parser.parse_args()
def setup(options):
sys.path.insert(0, options.gae_lib_path)
from dev_appserver import fix_sys_path
fix_sys_path()
def run(options):
from google.appengine.tools import appengine_rpc
from google.appengine.tools import appcfg
source = 'furious'
user_agent = appcfg.GetUserAgent()
server = appengine_rpc.HttpRpcServer(
'localhost:8080', lambda: ('test@example.com', 'password'), user_agent,
source, secure=False)
url = "/"
if options.url:
url += options.url[0]
server._DevAppServerAuthenticate()
server.Send(url, content_type="text/html; charset=utf-8",
payload=None)
def main():
options = args()
setup(options)
run(options)
if __name__ == "__main__":
main()
|
<commit_before>#!/usr/bin/python
import argparse
import sys
def args():
parser = argparse.ArgumentParser(description='Run the Furious Examples.')
parser.add_argument('--gae-sdk-path', metavar='S', dest="gae_lib_path",
default="/usr/local/google_appengine",
help='path to the GAE SDK')
parser.add_argument('--url', metavar='U', dest="url", default="",
help="the endpoint to run")
return parser.parse_args()
def setup(options):
sys.path.insert(0, options.gae_lib_path)
from dev_appserver import fix_sys_path
fix_sys_path()
def run(options):
from google.appengine.tools import appengine_rpc
from google.appengine.tools import appcfg
source = 'furious'
user_agent = appcfg.GetUserAgent()
server = appengine_rpc.HttpRpcServer(
'localhost:8080', lambda: ('test@example.com', 'password'), user_agent,
source, secure=False)
server._DevAppServerAuthenticate()
server.Send(options.url, content_type="text/html; charset=utf-8",
payload=None)
def main():
options = args()
setup(options)
run(options)
if __name__ == "__main__":
main()
<commit_msg>Update the way the url is handled.<commit_after>#!/usr/bin/python
import argparse
import sys
def args():
parser = argparse.ArgumentParser(description='Run the Furious Examples.')
parser.add_argument('--gae-sdk-path', metavar='S', dest="gae_lib_path",
default="/usr/local/google_appengine",
help='path to the GAE SDK')
parser.add_argument('url', metavar='U', default="", nargs=1,
help="the endpoint to run")
return parser.parse_args()
def setup(options):
sys.path.insert(0, options.gae_lib_path)
from dev_appserver import fix_sys_path
fix_sys_path()
def run(options):
from google.appengine.tools import appengine_rpc
from google.appengine.tools import appcfg
source = 'furious'
user_agent = appcfg.GetUserAgent()
server = appengine_rpc.HttpRpcServer(
'localhost:8080', lambda: ('test@example.com', 'password'), user_agent,
source, secure=False)
url = "/"
if options.url:
url += options.url[0]
server._DevAppServerAuthenticate()
server.Send(url, content_type="text/html; charset=utf-8",
payload=None)
def main():
options = args()
setup(options)
run(options)
if __name__ == "__main__":
main()
|
24cbbd24e6398aa11956ac48282bd907806284c3
|
genderbot.py
|
genderbot.py
|
import re
from twitterbot import TwitterBot
import wikipedia
class Genderbot(TwitterBot):
boring_article_regex = (r"municipality|village|town|football|genus|family|"
"administrative|district|community|region|hamlet|"
"school|actor|mountain|basketball|city|species|film|"
"county|located|politician|professional|settlement")
def tweet(self):
article = self.__random_wikipedia_article()
match = re.search(r"\bis [^.?]+", article.content, re.UNICODE)
if match:
status = self.__format_status(match.group(0), article.url)
if self.__is_interesting(status):
self.post_tweet(status)
def __format_status(self, is_phrase, url):
status = 'gender %s' % (is_phrase)
if len(status) > 114: status = status[0:113] + '...'
return status + ' %s' % (url)
def __is_interesting(self, status):
boring_match = re.search(Genderbot.boring_article_regex, status, re.UNICODE)
return boring_match is None
def __random_wikipedia_article(self):
random_title = wikipedia.random(pages=1)
return wikipedia.page(title=random_title)
if __name__ == "__main__":
try:
Genderbot("CustomGender").tweet()
except:
pass
|
import re
from twitterbot import TwitterBot
import wikipedia
class Genderbot(TwitterBot):
boring_regex = (r"municipality|village|town|football|genus|family|"
"administrative|district|community|region|hamlet|"
"school|actor|mountain|basketball|city|species|film|"
"county|located|politician|professional|settlement|"
"river|lake|province|replaced|origin|band|park|song"
"approximately|north|south|east|west|business")
def tweet(self):
article = self.__random_wikipedia_article()
match = re.search(r"\bis [^.?]+", article.content, re.UNICODE)
if match:
status = self.__format_status(match.group(0), article.url)
if self.__is_interesting(status):
self.post_tweet(status)
def __format_status(self, is_phrase, url):
status = 'gender %s' % (is_phrase)
if len(status) > 114: status = status[0:113] + '...'
return status + ' %s' % (url)
def __is_interesting(self, status):
flags = re.UNICODE | re.IGNORECASE
boring = re.search(Genderbot.boring_regex, status, flags)
return boring is None
def __random_wikipedia_article(self):
random_title = wikipedia.random(pages=1)
return wikipedia.page(title=random_title)
if __name__ == "__main__":
try:
Genderbot("CustomGender").tweet()
except:
pass
|
Tweak boring regex to exclude more terms
|
Tweak boring regex to exclude more terms
|
Python
|
mit
|
DanielleSucher/genderbot
|
import re
from twitterbot import TwitterBot
import wikipedia
class Genderbot(TwitterBot):
boring_article_regex = (r"municipality|village|town|football|genus|family|"
"administrative|district|community|region|hamlet|"
"school|actor|mountain|basketball|city|species|film|"
"county|located|politician|professional|settlement")
def tweet(self):
article = self.__random_wikipedia_article()
match = re.search(r"\bis [^.?]+", article.content, re.UNICODE)
if match:
status = self.__format_status(match.group(0), article.url)
if self.__is_interesting(status):
self.post_tweet(status)
def __format_status(self, is_phrase, url):
status = 'gender %s' % (is_phrase)
if len(status) > 114: status = status[0:113] + '...'
return status + ' %s' % (url)
def __is_interesting(self, status):
boring_match = re.search(Genderbot.boring_article_regex, status, re.UNICODE)
return boring_match is None
def __random_wikipedia_article(self):
random_title = wikipedia.random(pages=1)
return wikipedia.page(title=random_title)
if __name__ == "__main__":
try:
Genderbot("CustomGender").tweet()
except:
pass
Tweak boring regex to exclude more terms
|
import re
from twitterbot import TwitterBot
import wikipedia
class Genderbot(TwitterBot):
boring_regex = (r"municipality|village|town|football|genus|family|"
"administrative|district|community|region|hamlet|"
"school|actor|mountain|basketball|city|species|film|"
"county|located|politician|professional|settlement|"
"river|lake|province|replaced|origin|band|park|song"
"approximately|north|south|east|west|business")
def tweet(self):
article = self.__random_wikipedia_article()
match = re.search(r"\bis [^.?]+", article.content, re.UNICODE)
if match:
status = self.__format_status(match.group(0), article.url)
if self.__is_interesting(status):
self.post_tweet(status)
def __format_status(self, is_phrase, url):
status = 'gender %s' % (is_phrase)
if len(status) > 114: status = status[0:113] + '...'
return status + ' %s' % (url)
def __is_interesting(self, status):
flags = re.UNICODE | re.IGNORECASE
boring = re.search(Genderbot.boring_regex, status, flags)
return boring is None
def __random_wikipedia_article(self):
random_title = wikipedia.random(pages=1)
return wikipedia.page(title=random_title)
if __name__ == "__main__":
try:
Genderbot("CustomGender").tweet()
except:
pass
|
<commit_before>import re
from twitterbot import TwitterBot
import wikipedia
class Genderbot(TwitterBot):
boring_article_regex = (r"municipality|village|town|football|genus|family|"
"administrative|district|community|region|hamlet|"
"school|actor|mountain|basketball|city|species|film|"
"county|located|politician|professional|settlement")
def tweet(self):
article = self.__random_wikipedia_article()
match = re.search(r"\bis [^.?]+", article.content, re.UNICODE)
if match:
status = self.__format_status(match.group(0), article.url)
if self.__is_interesting(status):
self.post_tweet(status)
def __format_status(self, is_phrase, url):
status = 'gender %s' % (is_phrase)
if len(status) > 114: status = status[0:113] + '...'
return status + ' %s' % (url)
def __is_interesting(self, status):
boring_match = re.search(Genderbot.boring_article_regex, status, re.UNICODE)
return boring_match is None
def __random_wikipedia_article(self):
random_title = wikipedia.random(pages=1)
return wikipedia.page(title=random_title)
if __name__ == "__main__":
try:
Genderbot("CustomGender").tweet()
except:
pass
<commit_msg>Tweak boring regex to exclude more terms<commit_after>
|
import re
from twitterbot import TwitterBot
import wikipedia
class Genderbot(TwitterBot):
boring_regex = (r"municipality|village|town|football|genus|family|"
"administrative|district|community|region|hamlet|"
"school|actor|mountain|basketball|city|species|film|"
"county|located|politician|professional|settlement|"
"river|lake|province|replaced|origin|band|park|song"
"approximately|north|south|east|west|business")
def tweet(self):
article = self.__random_wikipedia_article()
match = re.search(r"\bis [^.?]+", article.content, re.UNICODE)
if match:
status = self.__format_status(match.group(0), article.url)
if self.__is_interesting(status):
self.post_tweet(status)
def __format_status(self, is_phrase, url):
status = 'gender %s' % (is_phrase)
if len(status) > 114: status = status[0:113] + '...'
return status + ' %s' % (url)
def __is_interesting(self, status):
flags = re.UNICODE | re.IGNORECASE
boring = re.search(Genderbot.boring_regex, status, flags)
return boring is None
def __random_wikipedia_article(self):
random_title = wikipedia.random(pages=1)
return wikipedia.page(title=random_title)
if __name__ == "__main__":
try:
Genderbot("CustomGender").tweet()
except:
pass
|
import re
from twitterbot import TwitterBot
import wikipedia
class Genderbot(TwitterBot):
boring_article_regex = (r"municipality|village|town|football|genus|family|"
"administrative|district|community|region|hamlet|"
"school|actor|mountain|basketball|city|species|film|"
"county|located|politician|professional|settlement")
def tweet(self):
article = self.__random_wikipedia_article()
match = re.search(r"\bis [^.?]+", article.content, re.UNICODE)
if match:
status = self.__format_status(match.group(0), article.url)
if self.__is_interesting(status):
self.post_tweet(status)
def __format_status(self, is_phrase, url):
status = 'gender %s' % (is_phrase)
if len(status) > 114: status = status[0:113] + '...'
return status + ' %s' % (url)
def __is_interesting(self, status):
boring_match = re.search(Genderbot.boring_article_regex, status, re.UNICODE)
return boring_match is None
def __random_wikipedia_article(self):
random_title = wikipedia.random(pages=1)
return wikipedia.page(title=random_title)
if __name__ == "__main__":
try:
Genderbot("CustomGender").tweet()
except:
pass
Tweak boring regex to exclude more termsimport re
from twitterbot import TwitterBot
import wikipedia
class Genderbot(TwitterBot):
boring_regex = (r"municipality|village|town|football|genus|family|"
"administrative|district|community|region|hamlet|"
"school|actor|mountain|basketball|city|species|film|"
"county|located|politician|professional|settlement|"
"river|lake|province|replaced|origin|band|park|song"
"approximately|north|south|east|west|business")
def tweet(self):
article = self.__random_wikipedia_article()
match = re.search(r"\bis [^.?]+", article.content, re.UNICODE)
if match:
status = self.__format_status(match.group(0), article.url)
if self.__is_interesting(status):
self.post_tweet(status)
def __format_status(self, is_phrase, url):
status = 'gender %s' % (is_phrase)
if len(status) > 114: status = status[0:113] + '...'
return status + ' %s' % (url)
def __is_interesting(self, status):
flags = re.UNICODE | re.IGNORECASE
boring = re.search(Genderbot.boring_regex, status, flags)
return boring is None
def __random_wikipedia_article(self):
random_title = wikipedia.random(pages=1)
return wikipedia.page(title=random_title)
if __name__ == "__main__":
try:
Genderbot("CustomGender").tweet()
except:
pass
|
<commit_before>import re
from twitterbot import TwitterBot
import wikipedia
class Genderbot(TwitterBot):
boring_article_regex = (r"municipality|village|town|football|genus|family|"
"administrative|district|community|region|hamlet|"
"school|actor|mountain|basketball|city|species|film|"
"county|located|politician|professional|settlement")
def tweet(self):
article = self.__random_wikipedia_article()
match = re.search(r"\bis [^.?]+", article.content, re.UNICODE)
if match:
status = self.__format_status(match.group(0), article.url)
if self.__is_interesting(status):
self.post_tweet(status)
def __format_status(self, is_phrase, url):
status = 'gender %s' % (is_phrase)
if len(status) > 114: status = status[0:113] + '...'
return status + ' %s' % (url)
def __is_interesting(self, status):
boring_match = re.search(Genderbot.boring_article_regex, status, re.UNICODE)
return boring_match is None
def __random_wikipedia_article(self):
random_title = wikipedia.random(pages=1)
return wikipedia.page(title=random_title)
if __name__ == "__main__":
try:
Genderbot("CustomGender").tweet()
except:
pass
<commit_msg>Tweak boring regex to exclude more terms<commit_after>import re
from twitterbot import TwitterBot
import wikipedia
class Genderbot(TwitterBot):
boring_regex = (r"municipality|village|town|football|genus|family|"
"administrative|district|community|region|hamlet|"
"school|actor|mountain|basketball|city|species|film|"
"county|located|politician|professional|settlement|"
"river|lake|province|replaced|origin|band|park|song"
"approximately|north|south|east|west|business")
def tweet(self):
article = self.__random_wikipedia_article()
match = re.search(r"\bis [^.?]+", article.content, re.UNICODE)
if match:
status = self.__format_status(match.group(0), article.url)
if self.__is_interesting(status):
self.post_tweet(status)
def __format_status(self, is_phrase, url):
status = 'gender %s' % (is_phrase)
if len(status) > 114: status = status[0:113] + '...'
return status + ' %s' % (url)
def __is_interesting(self, status):
flags = re.UNICODE | re.IGNORECASE
boring = re.search(Genderbot.boring_regex, status, flags)
return boring is None
def __random_wikipedia_article(self):
random_title = wikipedia.random(pages=1)
return wikipedia.page(title=random_title)
if __name__ == "__main__":
try:
Genderbot("CustomGender").tweet()
except:
pass
|
f68a3874eb9b80898a6c1acfc74e493aad5817d8
|
source/services/rotten_tomatoes_service.py
|
source/services/rotten_tomatoes_service.py
|
import requests
from bs4 import BeautifulSoup
from source.models.rt_rating import RTRating
class RottenTomatoesService:
__URL = 'http://www.rottentomatoes.com/m/'
__SEPERATOR = '_'
def __init__(self, title):
self.title = title
def get_rt_rating(self):
search_url = self.__URL + self.format_title()
movie_page = requests.get(search_url)
contents = movie_page.text
soup = BeautifulSoup(contents, 'lxml')
ratings = self.get_ratings(soup)
ratings.link = search_url
return ratings
def format_title(self):
formatted_title = self.title
if formatted_title.startswith('The '):
formatted_title = formatted_title.replace('The ', '', 1)
if "'s" in formatted_title:
formatted_title = formatted_title.replace("'s", 's')
formatted_title = formatted_title.replace(' ', self.__SEPERATOR)
formatted_title = formatted_title.replace('-', '')
formatted_title = formatted_title.replace(':', '')
formatted_title = formatted_title.replace(',', '')
return formatted_title
def get_ratings(self, soup):
items = []
for item in soup.findAll(attrs={'itemprop': 'ratingValue'}):
items.append(item.get_text().strip('%'))
return RTRating(items)
|
import requests
from bs4 import BeautifulSoup
from source.models.rt_rating import RTRating
class RottenTomatoesService:
__URL = 'http://www.rottentomatoes.com/m/'
__SEPERATOR = '_'
def __init__(self, title):
self.title = title
def get_rt_rating(self):
search_url = self.__URL + self.format_title()
movie_page = requests.get(search_url)
contents = movie_page.text
soup = BeautifulSoup(contents, 'lxml')
ratings = self.get_ratings(soup)
ratings.link = search_url
return ratings
def format_title(self):
formatted_title = self.title
if formatted_title.startswith('The '):
formatted_title = formatted_title.replace('The ', '', 1)
if formatted_title.startswith('A '):
formatted_title = formatted_title.replace('A ', '', 1)
if "'s" in formatted_title:
formatted_title = formatted_title.replace("'s", 's')
formatted_title = formatted_title.replace(' ', self.__SEPERATOR)
formatted_title = formatted_title.replace('-', '')
formatted_title = formatted_title.replace(':', '')
formatted_title = formatted_title.replace(',', '')
return formatted_title
def get_ratings(self, soup):
items = []
for item in soup.findAll(attrs={'itemprop': 'ratingValue'}):
items.append(item.get_text().strip('%'))
return RTRating(items)
|
Remove "A" from start of title for RT search
|
Remove "A" from start of title for RT search
|
Python
|
mit
|
jeremyrea/caterblu,jeremyrea/caterblu,jeremyrea/caterblu,jeremyrea/caterblu
|
import requests
from bs4 import BeautifulSoup
from source.models.rt_rating import RTRating
class RottenTomatoesService:
__URL = 'http://www.rottentomatoes.com/m/'
__SEPERATOR = '_'
def __init__(self, title):
self.title = title
def get_rt_rating(self):
search_url = self.__URL + self.format_title()
movie_page = requests.get(search_url)
contents = movie_page.text
soup = BeautifulSoup(contents, 'lxml')
ratings = self.get_ratings(soup)
ratings.link = search_url
return ratings
def format_title(self):
formatted_title = self.title
if formatted_title.startswith('The '):
formatted_title = formatted_title.replace('The ', '', 1)
if "'s" in formatted_title:
formatted_title = formatted_title.replace("'s", 's')
formatted_title = formatted_title.replace(' ', self.__SEPERATOR)
formatted_title = formatted_title.replace('-', '')
formatted_title = formatted_title.replace(':', '')
formatted_title = formatted_title.replace(',', '')
return formatted_title
def get_ratings(self, soup):
items = []
for item in soup.findAll(attrs={'itemprop': 'ratingValue'}):
items.append(item.get_text().strip('%'))
return RTRating(items)
Remove "A" from start of title for RT search
|
import requests
from bs4 import BeautifulSoup
from source.models.rt_rating import RTRating
class RottenTomatoesService:
__URL = 'http://www.rottentomatoes.com/m/'
__SEPERATOR = '_'
def __init__(self, title):
self.title = title
def get_rt_rating(self):
search_url = self.__URL + self.format_title()
movie_page = requests.get(search_url)
contents = movie_page.text
soup = BeautifulSoup(contents, 'lxml')
ratings = self.get_ratings(soup)
ratings.link = search_url
return ratings
def format_title(self):
formatted_title = self.title
if formatted_title.startswith('The '):
formatted_title = formatted_title.replace('The ', '', 1)
if formatted_title.startswith('A '):
formatted_title = formatted_title.replace('A ', '', 1)
if "'s" in formatted_title:
formatted_title = formatted_title.replace("'s", 's')
formatted_title = formatted_title.replace(' ', self.__SEPERATOR)
formatted_title = formatted_title.replace('-', '')
formatted_title = formatted_title.replace(':', '')
formatted_title = formatted_title.replace(',', '')
return formatted_title
def get_ratings(self, soup):
items = []
for item in soup.findAll(attrs={'itemprop': 'ratingValue'}):
items.append(item.get_text().strip('%'))
return RTRating(items)
|
<commit_before>import requests
from bs4 import BeautifulSoup
from source.models.rt_rating import RTRating
class RottenTomatoesService:
__URL = 'http://www.rottentomatoes.com/m/'
__SEPERATOR = '_'
def __init__(self, title):
self.title = title
def get_rt_rating(self):
search_url = self.__URL + self.format_title()
movie_page = requests.get(search_url)
contents = movie_page.text
soup = BeautifulSoup(contents, 'lxml')
ratings = self.get_ratings(soup)
ratings.link = search_url
return ratings
def format_title(self):
formatted_title = self.title
if formatted_title.startswith('The '):
formatted_title = formatted_title.replace('The ', '', 1)
if "'s" in formatted_title:
formatted_title = formatted_title.replace("'s", 's')
formatted_title = formatted_title.replace(' ', self.__SEPERATOR)
formatted_title = formatted_title.replace('-', '')
formatted_title = formatted_title.replace(':', '')
formatted_title = formatted_title.replace(',', '')
return formatted_title
def get_ratings(self, soup):
items = []
for item in soup.findAll(attrs={'itemprop': 'ratingValue'}):
items.append(item.get_text().strip('%'))
return RTRating(items)
<commit_msg>Remove "A" from start of title for RT search<commit_after>
|
import requests
from bs4 import BeautifulSoup
from source.models.rt_rating import RTRating
class RottenTomatoesService:
__URL = 'http://www.rottentomatoes.com/m/'
__SEPERATOR = '_'
def __init__(self, title):
self.title = title
def get_rt_rating(self):
search_url = self.__URL + self.format_title()
movie_page = requests.get(search_url)
contents = movie_page.text
soup = BeautifulSoup(contents, 'lxml')
ratings = self.get_ratings(soup)
ratings.link = search_url
return ratings
def format_title(self):
formatted_title = self.title
if formatted_title.startswith('The '):
formatted_title = formatted_title.replace('The ', '', 1)
if formatted_title.startswith('A '):
formatted_title = formatted_title.replace('A ', '', 1)
if "'s" in formatted_title:
formatted_title = formatted_title.replace("'s", 's')
formatted_title = formatted_title.replace(' ', self.__SEPERATOR)
formatted_title = formatted_title.replace('-', '')
formatted_title = formatted_title.replace(':', '')
formatted_title = formatted_title.replace(',', '')
return formatted_title
def get_ratings(self, soup):
items = []
for item in soup.findAll(attrs={'itemprop': 'ratingValue'}):
items.append(item.get_text().strip('%'))
return RTRating(items)
|
import requests
from bs4 import BeautifulSoup
from source.models.rt_rating import RTRating
class RottenTomatoesService:
__URL = 'http://www.rottentomatoes.com/m/'
__SEPERATOR = '_'
def __init__(self, title):
self.title = title
def get_rt_rating(self):
search_url = self.__URL + self.format_title()
movie_page = requests.get(search_url)
contents = movie_page.text
soup = BeautifulSoup(contents, 'lxml')
ratings = self.get_ratings(soup)
ratings.link = search_url
return ratings
def format_title(self):
formatted_title = self.title
if formatted_title.startswith('The '):
formatted_title = formatted_title.replace('The ', '', 1)
if "'s" in formatted_title:
formatted_title = formatted_title.replace("'s", 's')
formatted_title = formatted_title.replace(' ', self.__SEPERATOR)
formatted_title = formatted_title.replace('-', '')
formatted_title = formatted_title.replace(':', '')
formatted_title = formatted_title.replace(',', '')
return formatted_title
def get_ratings(self, soup):
items = []
for item in soup.findAll(attrs={'itemprop': 'ratingValue'}):
items.append(item.get_text().strip('%'))
return RTRating(items)
Remove "A" from start of title for RT searchimport requests
from bs4 import BeautifulSoup
from source.models.rt_rating import RTRating
class RottenTomatoesService:
__URL = 'http://www.rottentomatoes.com/m/'
__SEPERATOR = '_'
def __init__(self, title):
self.title = title
def get_rt_rating(self):
search_url = self.__URL + self.format_title()
movie_page = requests.get(search_url)
contents = movie_page.text
soup = BeautifulSoup(contents, 'lxml')
ratings = self.get_ratings(soup)
ratings.link = search_url
return ratings
def format_title(self):
formatted_title = self.title
if formatted_title.startswith('The '):
formatted_title = formatted_title.replace('The ', '', 1)
if formatted_title.startswith('A '):
formatted_title = formatted_title.replace('A ', '', 1)
if "'s" in formatted_title:
formatted_title = formatted_title.replace("'s", 's')
formatted_title = formatted_title.replace(' ', self.__SEPERATOR)
formatted_title = formatted_title.replace('-', '')
formatted_title = formatted_title.replace(':', '')
formatted_title = formatted_title.replace(',', '')
return formatted_title
def get_ratings(self, soup):
items = []
for item in soup.findAll(attrs={'itemprop': 'ratingValue'}):
items.append(item.get_text().strip('%'))
return RTRating(items)
|
<commit_before>import requests
from bs4 import BeautifulSoup
from source.models.rt_rating import RTRating
class RottenTomatoesService:
__URL = 'http://www.rottentomatoes.com/m/'
__SEPERATOR = '_'
def __init__(self, title):
self.title = title
def get_rt_rating(self):
search_url = self.__URL + self.format_title()
movie_page = requests.get(search_url)
contents = movie_page.text
soup = BeautifulSoup(contents, 'lxml')
ratings = self.get_ratings(soup)
ratings.link = search_url
return ratings
def format_title(self):
formatted_title = self.title
if formatted_title.startswith('The '):
formatted_title = formatted_title.replace('The ', '', 1)
if "'s" in formatted_title:
formatted_title = formatted_title.replace("'s", 's')
formatted_title = formatted_title.replace(' ', self.__SEPERATOR)
formatted_title = formatted_title.replace('-', '')
formatted_title = formatted_title.replace(':', '')
formatted_title = formatted_title.replace(',', '')
return formatted_title
def get_ratings(self, soup):
items = []
for item in soup.findAll(attrs={'itemprop': 'ratingValue'}):
items.append(item.get_text().strip('%'))
return RTRating(items)
<commit_msg>Remove "A" from start of title for RT search<commit_after>import requests
from bs4 import BeautifulSoup
from source.models.rt_rating import RTRating
class RottenTomatoesService:
__URL = 'http://www.rottentomatoes.com/m/'
__SEPERATOR = '_'
def __init__(self, title):
self.title = title
def get_rt_rating(self):
search_url = self.__URL + self.format_title()
movie_page = requests.get(search_url)
contents = movie_page.text
soup = BeautifulSoup(contents, 'lxml')
ratings = self.get_ratings(soup)
ratings.link = search_url
return ratings
def format_title(self):
formatted_title = self.title
if formatted_title.startswith('The '):
formatted_title = formatted_title.replace('The ', '', 1)
if formatted_title.startswith('A '):
formatted_title = formatted_title.replace('A ', '', 1)
if "'s" in formatted_title:
formatted_title = formatted_title.replace("'s", 's')
formatted_title = formatted_title.replace(' ', self.__SEPERATOR)
formatted_title = formatted_title.replace('-', '')
formatted_title = formatted_title.replace(':', '')
formatted_title = formatted_title.replace(',', '')
return formatted_title
def get_ratings(self, soup):
items = []
for item in soup.findAll(attrs={'itemprop': 'ratingValue'}):
items.append(item.get_text().strip('%'))
return RTRating(items)
|
033b17a8be5be32188ca9b5f286fe023fc07d34a
|
frappe/utils/pdf.py
|
frappe/utils/pdf.py
|
# Copyright (c) 2013, Web Notes Technologies Pvt. Ltd. and Contributors
# MIT License. See license.txt
from __future__ import unicode_literals
import pdfkit, os, frappe
from frappe.utils import scrub_urls
def get_pdf(html, options=None):
if not options:
options = {}
options.update({
"print-media-type": None,
"background": None,
"images": None,
'margin-top': '15mm',
'margin-right': '15mm',
'margin-bottom': '15mm',
'margin-left': '15mm',
'encoding': "UTF-8",
'no-outline': None
})
if not options.get("page-size"):
options['page-size'] = frappe.db.get_single_value("Print Settings", "pdf_page_size") or "A4"
html = scrub_urls(html)
fname = os.path.join("/tmp", frappe.generate_hash() + ".pdf")
pdfkit.from_string(html, fname, options=options or {})
with open(fname, "rb") as fileobj:
filedata = fileobj.read()
os.remove(fname)
return filedata
|
# Copyright (c) 2013, Web Notes Technologies Pvt. Ltd. and Contributors
# MIT License. See license.txt
from __future__ import unicode_literals
import pdfkit, os, frappe
from frappe.utils import scrub_urls
def get_pdf(html, options=None):
if not options:
options = {}
options.update({
"print-media-type": None,
"background": None,
"images": None,
# 'margin-top': '10mm',
# 'margin-right': '1mm',
# 'margin-bottom': '10mm',
# 'margin-left': '1mm',
'encoding': "UTF-8",
'no-outline': None
})
if not options.get("page-size"):
options['page-size'] = frappe.db.get_single_value("Print Settings", "pdf_page_size") or "A4"
html = scrub_urls(html)
fname = os.path.join("/tmp", frappe.generate_hash() + ".pdf")
pdfkit.from_string(html, fname, options=options or {})
with open(fname, "rb") as fileobj:
filedata = fileobj.read()
os.remove(fname)
return filedata
|
Remove margin constrains from PDF printing
|
Remove margin constrains from PDF printing
|
Python
|
mit
|
BhupeshGupta/frappe,BhupeshGupta/frappe,BhupeshGupta/frappe,BhupeshGupta/frappe
|
# Copyright (c) 2013, Web Notes Technologies Pvt. Ltd. and Contributors
# MIT License. See license.txt
from __future__ import unicode_literals
import pdfkit, os, frappe
from frappe.utils import scrub_urls
def get_pdf(html, options=None):
if not options:
options = {}
options.update({
"print-media-type": None,
"background": None,
"images": None,
'margin-top': '15mm',
'margin-right': '15mm',
'margin-bottom': '15mm',
'margin-left': '15mm',
'encoding': "UTF-8",
'no-outline': None
})
if not options.get("page-size"):
options['page-size'] = frappe.db.get_single_value("Print Settings", "pdf_page_size") or "A4"
html = scrub_urls(html)
fname = os.path.join("/tmp", frappe.generate_hash() + ".pdf")
pdfkit.from_string(html, fname, options=options or {})
with open(fname, "rb") as fileobj:
filedata = fileobj.read()
os.remove(fname)
return filedata
Remove margin constrains from PDF printing
|
# Copyright (c) 2013, Web Notes Technologies Pvt. Ltd. and Contributors
# MIT License. See license.txt
from __future__ import unicode_literals
import pdfkit, os, frappe
from frappe.utils import scrub_urls
def get_pdf(html, options=None):
if not options:
options = {}
options.update({
"print-media-type": None,
"background": None,
"images": None,
# 'margin-top': '10mm',
# 'margin-right': '1mm',
# 'margin-bottom': '10mm',
# 'margin-left': '1mm',
'encoding': "UTF-8",
'no-outline': None
})
if not options.get("page-size"):
options['page-size'] = frappe.db.get_single_value("Print Settings", "pdf_page_size") or "A4"
html = scrub_urls(html)
fname = os.path.join("/tmp", frappe.generate_hash() + ".pdf")
pdfkit.from_string(html, fname, options=options or {})
with open(fname, "rb") as fileobj:
filedata = fileobj.read()
os.remove(fname)
return filedata
|
<commit_before># Copyright (c) 2013, Web Notes Technologies Pvt. Ltd. and Contributors
# MIT License. See license.txt
from __future__ import unicode_literals
import pdfkit, os, frappe
from frappe.utils import scrub_urls
def get_pdf(html, options=None):
if not options:
options = {}
options.update({
"print-media-type": None,
"background": None,
"images": None,
'margin-top': '15mm',
'margin-right': '15mm',
'margin-bottom': '15mm',
'margin-left': '15mm',
'encoding': "UTF-8",
'no-outline': None
})
if not options.get("page-size"):
options['page-size'] = frappe.db.get_single_value("Print Settings", "pdf_page_size") or "A4"
html = scrub_urls(html)
fname = os.path.join("/tmp", frappe.generate_hash() + ".pdf")
pdfkit.from_string(html, fname, options=options or {})
with open(fname, "rb") as fileobj:
filedata = fileobj.read()
os.remove(fname)
return filedata
<commit_msg>Remove margin constrains from PDF printing<commit_after>
|
# Copyright (c) 2013, Web Notes Technologies Pvt. Ltd. and Contributors
# MIT License. See license.txt
from __future__ import unicode_literals
import pdfkit, os, frappe
from frappe.utils import scrub_urls
def get_pdf(html, options=None):
if not options:
options = {}
options.update({
"print-media-type": None,
"background": None,
"images": None,
# 'margin-top': '10mm',
# 'margin-right': '1mm',
# 'margin-bottom': '10mm',
# 'margin-left': '1mm',
'encoding': "UTF-8",
'no-outline': None
})
if not options.get("page-size"):
options['page-size'] = frappe.db.get_single_value("Print Settings", "pdf_page_size") or "A4"
html = scrub_urls(html)
fname = os.path.join("/tmp", frappe.generate_hash() + ".pdf")
pdfkit.from_string(html, fname, options=options or {})
with open(fname, "rb") as fileobj:
filedata = fileobj.read()
os.remove(fname)
return filedata
|
# Copyright (c) 2013, Web Notes Technologies Pvt. Ltd. and Contributors
# MIT License. See license.txt
from __future__ import unicode_literals
import pdfkit, os, frappe
from frappe.utils import scrub_urls
def get_pdf(html, options=None):
if not options:
options = {}
options.update({
"print-media-type": None,
"background": None,
"images": None,
'margin-top': '15mm',
'margin-right': '15mm',
'margin-bottom': '15mm',
'margin-left': '15mm',
'encoding': "UTF-8",
'no-outline': None
})
if not options.get("page-size"):
options['page-size'] = frappe.db.get_single_value("Print Settings", "pdf_page_size") or "A4"
html = scrub_urls(html)
fname = os.path.join("/tmp", frappe.generate_hash() + ".pdf")
pdfkit.from_string(html, fname, options=options or {})
with open(fname, "rb") as fileobj:
filedata = fileobj.read()
os.remove(fname)
return filedata
Remove margin constrains from PDF printing# Copyright (c) 2013, Web Notes Technologies Pvt. Ltd. and Contributors
# MIT License. See license.txt
from __future__ import unicode_literals
import pdfkit, os, frappe
from frappe.utils import scrub_urls
def get_pdf(html, options=None):
if not options:
options = {}
options.update({
"print-media-type": None,
"background": None,
"images": None,
# 'margin-top': '10mm',
# 'margin-right': '1mm',
# 'margin-bottom': '10mm',
# 'margin-left': '1mm',
'encoding': "UTF-8",
'no-outline': None
})
if not options.get("page-size"):
options['page-size'] = frappe.db.get_single_value("Print Settings", "pdf_page_size") or "A4"
html = scrub_urls(html)
fname = os.path.join("/tmp", frappe.generate_hash() + ".pdf")
pdfkit.from_string(html, fname, options=options or {})
with open(fname, "rb") as fileobj:
filedata = fileobj.read()
os.remove(fname)
return filedata
|
<commit_before># Copyright (c) 2013, Web Notes Technologies Pvt. Ltd. and Contributors
# MIT License. See license.txt
from __future__ import unicode_literals
import pdfkit, os, frappe
from frappe.utils import scrub_urls
def get_pdf(html, options=None):
if not options:
options = {}
options.update({
"print-media-type": None,
"background": None,
"images": None,
'margin-top': '15mm',
'margin-right': '15mm',
'margin-bottom': '15mm',
'margin-left': '15mm',
'encoding': "UTF-8",
'no-outline': None
})
if not options.get("page-size"):
options['page-size'] = frappe.db.get_single_value("Print Settings", "pdf_page_size") or "A4"
html = scrub_urls(html)
fname = os.path.join("/tmp", frappe.generate_hash() + ".pdf")
pdfkit.from_string(html, fname, options=options or {})
with open(fname, "rb") as fileobj:
filedata = fileobj.read()
os.remove(fname)
return filedata
<commit_msg>Remove margin constrains from PDF printing<commit_after># Copyright (c) 2013, Web Notes Technologies Pvt. Ltd. and Contributors
# MIT License. See license.txt
from __future__ import unicode_literals
import pdfkit, os, frappe
from frappe.utils import scrub_urls
def get_pdf(html, options=None):
if not options:
options = {}
options.update({
"print-media-type": None,
"background": None,
"images": None,
# 'margin-top': '10mm',
# 'margin-right': '1mm',
# 'margin-bottom': '10mm',
# 'margin-left': '1mm',
'encoding': "UTF-8",
'no-outline': None
})
if not options.get("page-size"):
options['page-size'] = frappe.db.get_single_value("Print Settings", "pdf_page_size") or "A4"
html = scrub_urls(html)
fname = os.path.join("/tmp", frappe.generate_hash() + ".pdf")
pdfkit.from_string(html, fname, options=options or {})
with open(fname, "rb") as fileobj:
filedata = fileobj.read()
os.remove(fname)
return filedata
|
026c22ef25b28889e24f20f96ca6285289bcd46d
|
seven23/models/stats/middleware.py
|
seven23/models/stats/middleware.py
|
from datetime import datetime
from seven23.models.profile.models import Profile
from rest_framework.authentication import TokenAuthentication
from seven23.models.stats.models import MonthlyActiveUser, DailyActiveUser
def active_user_middleware(get_response):
def middleware(request):
user = request.user
user_auth_tuple = TokenAuthentication().authenticate(request)
if user_auth_tuple is not None:
(user, token) = user_auth_tuple
if user.is_authenticated and not user.is_superuser:
# If user has no profile, we create on.
if not hasattr(user, 'profile'):
Profile.objects.create(user=user)
now = datetime.now()
last_api_call = user.profile.last_api_call
udpate_user = False
if now.year != last_api_call.year or now.month != last_api_call.month :
MonthlyActiveUser.objects.update_or_create(year=now.year, month=now.month)
udpate_user = True
if now.year != last_api_call.year or now.month != last_api_call.month or now.day != last_api_call.day :
DailyActiveUser.objects.update_or_create(year=now.year, month=now.month, day=now.day)
udpate_user = True
if udpate_user:
user.profile.last_api_call = now
user.save()
# Perform actual request
response = get_response(request)
return response
return middleware
|
from datetime import datetime
from seven23.models.profile.models import Profile
from rest_framework.authentication import TokenAuthentication
# from seven23.models.stats.models import MonthlyActiveUser, DailyActiveUser
def active_user_middleware(get_response):
def middleware(request):
user = request.user
user_auth_tuple = TokenAuthentication().authenticate(request)
if user_auth_tuple is not None:
(user, token) = user_auth_tuple
if user.is_authenticated and not user.is_superuser:
# If user has no profile, we create on.
if not hasattr(user, 'profile'):
Profile.objects.create(user=user)
now = datetime.now()
last_api_call = user.profile.last_api_call
udpate_user = False
if now.year != last_api_call.year or now.month != last_api_call.month :
# MonthlyActiveUser.objects.update_or_create(year=now.year, month=now.month)
udpate_user = True
if now.year != last_api_call.year or now.month != last_api_call.month or now.day != last_api_call.day :
# DailyActiveUser.objects.update_or_create(year=now.year, month=now.month, day=now.day)
udpate_user = True
if udpate_user:
user.profile.last_api_call = now
user.save()
# Perform actual request
response = get_response(request)
return response
return middleware
|
Disable stats after crash report until we find a alternative solution
|
Disable stats after crash report until we find a alternative solution
|
Python
|
mit
|
sebastienbarbier/723e,sebastienbarbier/723e_server,sebastienbarbier/723e_server,sebastienbarbier/723e
|
from datetime import datetime
from seven23.models.profile.models import Profile
from rest_framework.authentication import TokenAuthentication
from seven23.models.stats.models import MonthlyActiveUser, DailyActiveUser
def active_user_middleware(get_response):
def middleware(request):
user = request.user
user_auth_tuple = TokenAuthentication().authenticate(request)
if user_auth_tuple is not None:
(user, token) = user_auth_tuple
if user.is_authenticated and not user.is_superuser:
# If user has no profile, we create on.
if not hasattr(user, 'profile'):
Profile.objects.create(user=user)
now = datetime.now()
last_api_call = user.profile.last_api_call
udpate_user = False
if now.year != last_api_call.year or now.month != last_api_call.month :
MonthlyActiveUser.objects.update_or_create(year=now.year, month=now.month)
udpate_user = True
if now.year != last_api_call.year or now.month != last_api_call.month or now.day != last_api_call.day :
DailyActiveUser.objects.update_or_create(year=now.year, month=now.month, day=now.day)
udpate_user = True
if udpate_user:
user.profile.last_api_call = now
user.save()
# Perform actual request
response = get_response(request)
return response
return middlewareDisable stats after crash report until we find a alternative solution
|
from datetime import datetime
from seven23.models.profile.models import Profile
from rest_framework.authentication import TokenAuthentication
# from seven23.models.stats.models import MonthlyActiveUser, DailyActiveUser
def active_user_middleware(get_response):
def middleware(request):
user = request.user
user_auth_tuple = TokenAuthentication().authenticate(request)
if user_auth_tuple is not None:
(user, token) = user_auth_tuple
if user.is_authenticated and not user.is_superuser:
# If user has no profile, we create on.
if not hasattr(user, 'profile'):
Profile.objects.create(user=user)
now = datetime.now()
last_api_call = user.profile.last_api_call
udpate_user = False
if now.year != last_api_call.year or now.month != last_api_call.month :
# MonthlyActiveUser.objects.update_or_create(year=now.year, month=now.month)
udpate_user = True
if now.year != last_api_call.year or now.month != last_api_call.month or now.day != last_api_call.day :
# DailyActiveUser.objects.update_or_create(year=now.year, month=now.month, day=now.day)
udpate_user = True
if udpate_user:
user.profile.last_api_call = now
user.save()
# Perform actual request
response = get_response(request)
return response
return middleware
|
<commit_before>
from datetime import datetime
from seven23.models.profile.models import Profile
from rest_framework.authentication import TokenAuthentication
from seven23.models.stats.models import MonthlyActiveUser, DailyActiveUser
def active_user_middleware(get_response):
def middleware(request):
user = request.user
user_auth_tuple = TokenAuthentication().authenticate(request)
if user_auth_tuple is not None:
(user, token) = user_auth_tuple
if user.is_authenticated and not user.is_superuser:
# If user has no profile, we create on.
if not hasattr(user, 'profile'):
Profile.objects.create(user=user)
now = datetime.now()
last_api_call = user.profile.last_api_call
udpate_user = False
if now.year != last_api_call.year or now.month != last_api_call.month :
MonthlyActiveUser.objects.update_or_create(year=now.year, month=now.month)
udpate_user = True
if now.year != last_api_call.year or now.month != last_api_call.month or now.day != last_api_call.day :
DailyActiveUser.objects.update_or_create(year=now.year, month=now.month, day=now.day)
udpate_user = True
if udpate_user:
user.profile.last_api_call = now
user.save()
# Perform actual request
response = get_response(request)
return response
return middleware<commit_msg>Disable stats after crash report until we find a alternative solution<commit_after>
|
from datetime import datetime
from seven23.models.profile.models import Profile
from rest_framework.authentication import TokenAuthentication
# from seven23.models.stats.models import MonthlyActiveUser, DailyActiveUser
def active_user_middleware(get_response):
def middleware(request):
user = request.user
user_auth_tuple = TokenAuthentication().authenticate(request)
if user_auth_tuple is not None:
(user, token) = user_auth_tuple
if user.is_authenticated and not user.is_superuser:
# If user has no profile, we create on.
if not hasattr(user, 'profile'):
Profile.objects.create(user=user)
now = datetime.now()
last_api_call = user.profile.last_api_call
udpate_user = False
if now.year != last_api_call.year or now.month != last_api_call.month :
# MonthlyActiveUser.objects.update_or_create(year=now.year, month=now.month)
udpate_user = True
if now.year != last_api_call.year or now.month != last_api_call.month or now.day != last_api_call.day :
# DailyActiveUser.objects.update_or_create(year=now.year, month=now.month, day=now.day)
udpate_user = True
if udpate_user:
user.profile.last_api_call = now
user.save()
# Perform actual request
response = get_response(request)
return response
return middleware
|
from datetime import datetime
from seven23.models.profile.models import Profile
from rest_framework.authentication import TokenAuthentication
from seven23.models.stats.models import MonthlyActiveUser, DailyActiveUser
def active_user_middleware(get_response):
def middleware(request):
user = request.user
user_auth_tuple = TokenAuthentication().authenticate(request)
if user_auth_tuple is not None:
(user, token) = user_auth_tuple
if user.is_authenticated and not user.is_superuser:
# If user has no profile, we create on.
if not hasattr(user, 'profile'):
Profile.objects.create(user=user)
now = datetime.now()
last_api_call = user.profile.last_api_call
udpate_user = False
if now.year != last_api_call.year or now.month != last_api_call.month :
MonthlyActiveUser.objects.update_or_create(year=now.year, month=now.month)
udpate_user = True
if now.year != last_api_call.year or now.month != last_api_call.month or now.day != last_api_call.day :
DailyActiveUser.objects.update_or_create(year=now.year, month=now.month, day=now.day)
udpate_user = True
if udpate_user:
user.profile.last_api_call = now
user.save()
# Perform actual request
response = get_response(request)
return response
return middlewareDisable stats after crash report until we find a alternative solution
from datetime import datetime
from seven23.models.profile.models import Profile
from rest_framework.authentication import TokenAuthentication
# from seven23.models.stats.models import MonthlyActiveUser, DailyActiveUser
def active_user_middleware(get_response):
def middleware(request):
user = request.user
user_auth_tuple = TokenAuthentication().authenticate(request)
if user_auth_tuple is not None:
(user, token) = user_auth_tuple
if user.is_authenticated and not user.is_superuser:
# If user has no profile, we create on.
if not hasattr(user, 'profile'):
Profile.objects.create(user=user)
now = datetime.now()
last_api_call = user.profile.last_api_call
udpate_user = False
if now.year != last_api_call.year or now.month != last_api_call.month :
# MonthlyActiveUser.objects.update_or_create(year=now.year, month=now.month)
udpate_user = True
if now.year != last_api_call.year or now.month != last_api_call.month or now.day != last_api_call.day :
# DailyActiveUser.objects.update_or_create(year=now.year, month=now.month, day=now.day)
udpate_user = True
if udpate_user:
user.profile.last_api_call = now
user.save()
# Perform actual request
response = get_response(request)
return response
return middleware
|
<commit_before>
from datetime import datetime
from seven23.models.profile.models import Profile
from rest_framework.authentication import TokenAuthentication
from seven23.models.stats.models import MonthlyActiveUser, DailyActiveUser
def active_user_middleware(get_response):
def middleware(request):
user = request.user
user_auth_tuple = TokenAuthentication().authenticate(request)
if user_auth_tuple is not None:
(user, token) = user_auth_tuple
if user.is_authenticated and not user.is_superuser:
# If user has no profile, we create on.
if not hasattr(user, 'profile'):
Profile.objects.create(user=user)
now = datetime.now()
last_api_call = user.profile.last_api_call
udpate_user = False
if now.year != last_api_call.year or now.month != last_api_call.month :
MonthlyActiveUser.objects.update_or_create(year=now.year, month=now.month)
udpate_user = True
if now.year != last_api_call.year or now.month != last_api_call.month or now.day != last_api_call.day :
DailyActiveUser.objects.update_or_create(year=now.year, month=now.month, day=now.day)
udpate_user = True
if udpate_user:
user.profile.last_api_call = now
user.save()
# Perform actual request
response = get_response(request)
return response
return middleware<commit_msg>Disable stats after crash report until we find a alternative solution<commit_after>
from datetime import datetime
from seven23.models.profile.models import Profile
from rest_framework.authentication import TokenAuthentication
# from seven23.models.stats.models import MonthlyActiveUser, DailyActiveUser
def active_user_middleware(get_response):
def middleware(request):
user = request.user
user_auth_tuple = TokenAuthentication().authenticate(request)
if user_auth_tuple is not None:
(user, token) = user_auth_tuple
if user.is_authenticated and not user.is_superuser:
# If user has no profile, we create on.
if not hasattr(user, 'profile'):
Profile.objects.create(user=user)
now = datetime.now()
last_api_call = user.profile.last_api_call
udpate_user = False
if now.year != last_api_call.year or now.month != last_api_call.month :
# MonthlyActiveUser.objects.update_or_create(year=now.year, month=now.month)
udpate_user = True
if now.year != last_api_call.year or now.month != last_api_call.month or now.day != last_api_call.day :
# DailyActiveUser.objects.update_or_create(year=now.year, month=now.month, day=now.day)
udpate_user = True
if udpate_user:
user.profile.last_api_call = now
user.save()
# Perform actual request
response = get_response(request)
return response
return middleware
|
500859e22bd4fda1fe55f4375642ccd5c1186d44
|
d_parser/helpers/parser_extender.py
|
d_parser/helpers/parser_extender.py
|
import logging
from d_parser.helpers import url_lib
from d_parser.helpers.get_body import get_body
from helpers.config import Config
logger = logging.getLogger('ddd_site_parse')
def check_body_errors(self, grab, task):
try:
self.status_counter[str(grab.doc.code)] += 1
except KeyError:
self.status_counter[str(grab.doc.code)] = 1
if grab.doc.body == '' or grab.doc.code != 200:
err = '[{}] Code is {}, url is {}, body is {}'.format(task.name, grab.doc.code, task.url, grab.doc.body)
logger.error(err)
return True
return False
def process_error(self, grab, task, exception):
try:
self.status_counter['EXC'] += 1
except KeyError:
self.status_counter['EXC'] = 1
html = get_body(grab)
err = '[{}] Url {} parse failed (e: {}), debug: {}'.format(task.name, task.url, exception, html)
self.logger.error(err)
def common_init(self, writer, try_limit):
self.logger = logger
self.result = writer
self.status_counter = {}
self.cookie_jar = {}
self.err_limit = try_limit
self.domain = url_lib.get_host_from_url(Config.get_seq('SITE_URL')[0])
self.logger.info('Init parser ok...')
|
import logging
from d_parser.helpers import url_lib
from d_parser.helpers.get_body import get_body
from helpers.config import Config
logger = logging.getLogger('ddd_site_parse')
def check_body_errors(self, grab, task):
try:
self.status_counter[str(grab.doc.code)] += 1
except KeyError:
self.status_counter[str(grab.doc.code)] = 1
if grab.doc.body == '' or grab.doc.code != 200:
err = '[{}] Code is {}, url is {}, body is {}'.format(task.name, grab.doc.code, task.url, grab.doc.body)
logger.error(err)
return True
return False
def process_error(self, grab, task, exception):
try:
self.status_counter['EXC'] += 1
except KeyError:
self.status_counter['EXC'] = 1
if Config.get('APP_LOG_HTML_ERR', '') == 'True':
html = get_body(grab)
else:
html = '(skipped by config)'
self.logger.error('[{}] Url {} parse failed (e: {}), debug: {}'.format(task.name, task.url, exception, html))
def common_init(self, try_limit):
self.result = []
self.logger = logger
self.status_counter = {}
self.cookie_jar = {}
self.err_limit = try_limit
self.domain = url_lib.get_host_from_url(Config.get_seq('SITE_URL')[0])
self.logger.info('Init parser ok...')
|
Add skip html log output config key
|
Add skip html log output config key
|
Python
|
mit
|
Holovin/D_GrabDemo
|
import logging
from d_parser.helpers import url_lib
from d_parser.helpers.get_body import get_body
from helpers.config import Config
logger = logging.getLogger('ddd_site_parse')
def check_body_errors(self, grab, task):
try:
self.status_counter[str(grab.doc.code)] += 1
except KeyError:
self.status_counter[str(grab.doc.code)] = 1
if grab.doc.body == '' or grab.doc.code != 200:
err = '[{}] Code is {}, url is {}, body is {}'.format(task.name, grab.doc.code, task.url, grab.doc.body)
logger.error(err)
return True
return False
def process_error(self, grab, task, exception):
try:
self.status_counter['EXC'] += 1
except KeyError:
self.status_counter['EXC'] = 1
html = get_body(grab)
err = '[{}] Url {} parse failed (e: {}), debug: {}'.format(task.name, task.url, exception, html)
self.logger.error(err)
def common_init(self, writer, try_limit):
self.logger = logger
self.result = writer
self.status_counter = {}
self.cookie_jar = {}
self.err_limit = try_limit
self.domain = url_lib.get_host_from_url(Config.get_seq('SITE_URL')[0])
self.logger.info('Init parser ok...')Add skip html log output config key
|
import logging
from d_parser.helpers import url_lib
from d_parser.helpers.get_body import get_body
from helpers.config import Config
logger = logging.getLogger('ddd_site_parse')
def check_body_errors(self, grab, task):
try:
self.status_counter[str(grab.doc.code)] += 1
except KeyError:
self.status_counter[str(grab.doc.code)] = 1
if grab.doc.body == '' or grab.doc.code != 200:
err = '[{}] Code is {}, url is {}, body is {}'.format(task.name, grab.doc.code, task.url, grab.doc.body)
logger.error(err)
return True
return False
def process_error(self, grab, task, exception):
try:
self.status_counter['EXC'] += 1
except KeyError:
self.status_counter['EXC'] = 1
if Config.get('APP_LOG_HTML_ERR', '') == 'True':
html = get_body(grab)
else:
html = '(skipped by config)'
self.logger.error('[{}] Url {} parse failed (e: {}), debug: {}'.format(task.name, task.url, exception, html))
def common_init(self, try_limit):
self.result = []
self.logger = logger
self.status_counter = {}
self.cookie_jar = {}
self.err_limit = try_limit
self.domain = url_lib.get_host_from_url(Config.get_seq('SITE_URL')[0])
self.logger.info('Init parser ok...')
|
<commit_before>import logging
from d_parser.helpers import url_lib
from d_parser.helpers.get_body import get_body
from helpers.config import Config
logger = logging.getLogger('ddd_site_parse')
def check_body_errors(self, grab, task):
try:
self.status_counter[str(grab.doc.code)] += 1
except KeyError:
self.status_counter[str(grab.doc.code)] = 1
if grab.doc.body == '' or grab.doc.code != 200:
err = '[{}] Code is {}, url is {}, body is {}'.format(task.name, grab.doc.code, task.url, grab.doc.body)
logger.error(err)
return True
return False
def process_error(self, grab, task, exception):
try:
self.status_counter['EXC'] += 1
except KeyError:
self.status_counter['EXC'] = 1
html = get_body(grab)
err = '[{}] Url {} parse failed (e: {}), debug: {}'.format(task.name, task.url, exception, html)
self.logger.error(err)
def common_init(self, writer, try_limit):
self.logger = logger
self.result = writer
self.status_counter = {}
self.cookie_jar = {}
self.err_limit = try_limit
self.domain = url_lib.get_host_from_url(Config.get_seq('SITE_URL')[0])
self.logger.info('Init parser ok...')<commit_msg>Add skip html log output config key<commit_after>
|
import logging
from d_parser.helpers import url_lib
from d_parser.helpers.get_body import get_body
from helpers.config import Config
logger = logging.getLogger('ddd_site_parse')
def check_body_errors(self, grab, task):
try:
self.status_counter[str(grab.doc.code)] += 1
except KeyError:
self.status_counter[str(grab.doc.code)] = 1
if grab.doc.body == '' or grab.doc.code != 200:
err = '[{}] Code is {}, url is {}, body is {}'.format(task.name, grab.doc.code, task.url, grab.doc.body)
logger.error(err)
return True
return False
def process_error(self, grab, task, exception):
try:
self.status_counter['EXC'] += 1
except KeyError:
self.status_counter['EXC'] = 1
if Config.get('APP_LOG_HTML_ERR', '') == 'True':
html = get_body(grab)
else:
html = '(skipped by config)'
self.logger.error('[{}] Url {} parse failed (e: {}), debug: {}'.format(task.name, task.url, exception, html))
def common_init(self, try_limit):
self.result = []
self.logger = logger
self.status_counter = {}
self.cookie_jar = {}
self.err_limit = try_limit
self.domain = url_lib.get_host_from_url(Config.get_seq('SITE_URL')[0])
self.logger.info('Init parser ok...')
|
import logging
from d_parser.helpers import url_lib
from d_parser.helpers.get_body import get_body
from helpers.config import Config
logger = logging.getLogger('ddd_site_parse')
def check_body_errors(self, grab, task):
try:
self.status_counter[str(grab.doc.code)] += 1
except KeyError:
self.status_counter[str(grab.doc.code)] = 1
if grab.doc.body == '' or grab.doc.code != 200:
err = '[{}] Code is {}, url is {}, body is {}'.format(task.name, grab.doc.code, task.url, grab.doc.body)
logger.error(err)
return True
return False
def process_error(self, grab, task, exception):
try:
self.status_counter['EXC'] += 1
except KeyError:
self.status_counter['EXC'] = 1
html = get_body(grab)
err = '[{}] Url {} parse failed (e: {}), debug: {}'.format(task.name, task.url, exception, html)
self.logger.error(err)
def common_init(self, writer, try_limit):
self.logger = logger
self.result = writer
self.status_counter = {}
self.cookie_jar = {}
self.err_limit = try_limit
self.domain = url_lib.get_host_from_url(Config.get_seq('SITE_URL')[0])
self.logger.info('Init parser ok...')Add skip html log output config keyimport logging
from d_parser.helpers import url_lib
from d_parser.helpers.get_body import get_body
from helpers.config import Config
logger = logging.getLogger('ddd_site_parse')
def check_body_errors(self, grab, task):
try:
self.status_counter[str(grab.doc.code)] += 1
except KeyError:
self.status_counter[str(grab.doc.code)] = 1
if grab.doc.body == '' or grab.doc.code != 200:
err = '[{}] Code is {}, url is {}, body is {}'.format(task.name, grab.doc.code, task.url, grab.doc.body)
logger.error(err)
return True
return False
def process_error(self, grab, task, exception):
try:
self.status_counter['EXC'] += 1
except KeyError:
self.status_counter['EXC'] = 1
if Config.get('APP_LOG_HTML_ERR', '') == 'True':
html = get_body(grab)
else:
html = '(skipped by config)'
self.logger.error('[{}] Url {} parse failed (e: {}), debug: {}'.format(task.name, task.url, exception, html))
def common_init(self, try_limit):
self.result = []
self.logger = logger
self.status_counter = {}
self.cookie_jar = {}
self.err_limit = try_limit
self.domain = url_lib.get_host_from_url(Config.get_seq('SITE_URL')[0])
self.logger.info('Init parser ok...')
|
<commit_before>import logging
from d_parser.helpers import url_lib
from d_parser.helpers.get_body import get_body
from helpers.config import Config
logger = logging.getLogger('ddd_site_parse')
def check_body_errors(self, grab, task):
try:
self.status_counter[str(grab.doc.code)] += 1
except KeyError:
self.status_counter[str(grab.doc.code)] = 1
if grab.doc.body == '' or grab.doc.code != 200:
err = '[{}] Code is {}, url is {}, body is {}'.format(task.name, grab.doc.code, task.url, grab.doc.body)
logger.error(err)
return True
return False
def process_error(self, grab, task, exception):
try:
self.status_counter['EXC'] += 1
except KeyError:
self.status_counter['EXC'] = 1
html = get_body(grab)
err = '[{}] Url {} parse failed (e: {}), debug: {}'.format(task.name, task.url, exception, html)
self.logger.error(err)
def common_init(self, writer, try_limit):
self.logger = logger
self.result = writer
self.status_counter = {}
self.cookie_jar = {}
self.err_limit = try_limit
self.domain = url_lib.get_host_from_url(Config.get_seq('SITE_URL')[0])
self.logger.info('Init parser ok...')<commit_msg>Add skip html log output config key<commit_after>import logging
from d_parser.helpers import url_lib
from d_parser.helpers.get_body import get_body
from helpers.config import Config
logger = logging.getLogger('ddd_site_parse')
def check_body_errors(self, grab, task):
try:
self.status_counter[str(grab.doc.code)] += 1
except KeyError:
self.status_counter[str(grab.doc.code)] = 1
if grab.doc.body == '' or grab.doc.code != 200:
err = '[{}] Code is {}, url is {}, body is {}'.format(task.name, grab.doc.code, task.url, grab.doc.body)
logger.error(err)
return True
return False
def process_error(self, grab, task, exception):
try:
self.status_counter['EXC'] += 1
except KeyError:
self.status_counter['EXC'] = 1
if Config.get('APP_LOG_HTML_ERR', '') == 'True':
html = get_body(grab)
else:
html = '(skipped by config)'
self.logger.error('[{}] Url {} parse failed (e: {}), debug: {}'.format(task.name, task.url, exception, html))
def common_init(self, try_limit):
self.result = []
self.logger = logger
self.status_counter = {}
self.cookie_jar = {}
self.err_limit = try_limit
self.domain = url_lib.get_host_from_url(Config.get_seq('SITE_URL')[0])
self.logger.info('Init parser ok...')
|
2f2cef54a98e2328a638d9bbdfd2e0312606d906
|
plugins/GCodeWriter/__init__.py
|
plugins/GCodeWriter/__init__.py
|
# Copyright (c) 2015 Ultimaker B.V.
# Cura is released under the terms of the AGPLv3 or higher.
from . import GCodeWriter
from UM.i18n import i18nCatalog
catalog = i18nCatalog("cura")
def getMetaData():
return {
"type": "mesh_writer",
"plugin": {
"name": "GCode Writer",
"author": "Ultimaker",
"version": "1.0",
"description": catalog.i18nc("GCode Writer Plugin Description", "Writes GCode to a file")
},
"mesh_writer": {
"extension": "gcode",
"description": catalog.i18nc("GCode Writer File Description", "GCode File")
}
}
def register(app):
return { "mesh_writer": GCodeWriter.GCodeWriter() }
|
# Copyright (c) 2015 Ultimaker B.V.
# Cura is released under the terms of the AGPLv3 or higher.
from . import GCodeWriter
from UM.i18n import i18nCatalog
catalog = i18nCatalog("cura")
def getMetaData():
return {
"type": "mesh_writer",
"plugin": {
"name": "GCode Writer",
"author": "Ultimaker",
"version": "1.0",
"description": catalog.i18nc("GCode Writer Plugin Description", "Writes GCode to a file")
},
"mesh_writer": {
"extension": "gcode",
"description": catalog.i18nc("GCode Writer File Description", "GCode File"),
"mime_types": [
"text/x-gcode"
]
}
}
def register(app):
return { "mesh_writer": GCodeWriter.GCodeWriter() }
|
Add mime types to GCodeWriter plugin
|
Add mime types to GCodeWriter plugin
|
Python
|
agpl-3.0
|
Curahelper/Cura,senttech/Cura,fieldOfView/Cura,hmflash/Cura,lo0ol/Ultimaker-Cura,Curahelper/Cura,markwal/Cura,ad1217/Cura,senttech/Cura,ad1217/Cura,lo0ol/Ultimaker-Cura,totalretribution/Cura,hmflash/Cura,bq/Ultimaker-Cura,ynotstartups/Wanhao,fieldOfView/Cura,totalretribution/Cura,fxtentacle/Cura,fxtentacle/Cura,ynotstartups/Wanhao,markwal/Cura,bq/Ultimaker-Cura
|
# Copyright (c) 2015 Ultimaker B.V.
# Cura is released under the terms of the AGPLv3 or higher.
from . import GCodeWriter
from UM.i18n import i18nCatalog
catalog = i18nCatalog("cura")
def getMetaData():
return {
"type": "mesh_writer",
"plugin": {
"name": "GCode Writer",
"author": "Ultimaker",
"version": "1.0",
"description": catalog.i18nc("GCode Writer Plugin Description", "Writes GCode to a file")
},
"mesh_writer": {
"extension": "gcode",
"description": catalog.i18nc("GCode Writer File Description", "GCode File")
}
}
def register(app):
return { "mesh_writer": GCodeWriter.GCodeWriter() }
Add mime types to GCodeWriter plugin
|
# Copyright (c) 2015 Ultimaker B.V.
# Cura is released under the terms of the AGPLv3 or higher.
from . import GCodeWriter
from UM.i18n import i18nCatalog
catalog = i18nCatalog("cura")
def getMetaData():
return {
"type": "mesh_writer",
"plugin": {
"name": "GCode Writer",
"author": "Ultimaker",
"version": "1.0",
"description": catalog.i18nc("GCode Writer Plugin Description", "Writes GCode to a file")
},
"mesh_writer": {
"extension": "gcode",
"description": catalog.i18nc("GCode Writer File Description", "GCode File"),
"mime_types": [
"text/x-gcode"
]
}
}
def register(app):
return { "mesh_writer": GCodeWriter.GCodeWriter() }
|
<commit_before># Copyright (c) 2015 Ultimaker B.V.
# Cura is released under the terms of the AGPLv3 or higher.
from . import GCodeWriter
from UM.i18n import i18nCatalog
catalog = i18nCatalog("cura")
def getMetaData():
return {
"type": "mesh_writer",
"plugin": {
"name": "GCode Writer",
"author": "Ultimaker",
"version": "1.0",
"description": catalog.i18nc("GCode Writer Plugin Description", "Writes GCode to a file")
},
"mesh_writer": {
"extension": "gcode",
"description": catalog.i18nc("GCode Writer File Description", "GCode File")
}
}
def register(app):
return { "mesh_writer": GCodeWriter.GCodeWriter() }
<commit_msg>Add mime types to GCodeWriter plugin<commit_after>
|
# Copyright (c) 2015 Ultimaker B.V.
# Cura is released under the terms of the AGPLv3 or higher.
from . import GCodeWriter
from UM.i18n import i18nCatalog
catalog = i18nCatalog("cura")
def getMetaData():
return {
"type": "mesh_writer",
"plugin": {
"name": "GCode Writer",
"author": "Ultimaker",
"version": "1.0",
"description": catalog.i18nc("GCode Writer Plugin Description", "Writes GCode to a file")
},
"mesh_writer": {
"extension": "gcode",
"description": catalog.i18nc("GCode Writer File Description", "GCode File"),
"mime_types": [
"text/x-gcode"
]
}
}
def register(app):
return { "mesh_writer": GCodeWriter.GCodeWriter() }
|
# Copyright (c) 2015 Ultimaker B.V.
# Cura is released under the terms of the AGPLv3 or higher.
from . import GCodeWriter
from UM.i18n import i18nCatalog
catalog = i18nCatalog("cura")
def getMetaData():
return {
"type": "mesh_writer",
"plugin": {
"name": "GCode Writer",
"author": "Ultimaker",
"version": "1.0",
"description": catalog.i18nc("GCode Writer Plugin Description", "Writes GCode to a file")
},
"mesh_writer": {
"extension": "gcode",
"description": catalog.i18nc("GCode Writer File Description", "GCode File")
}
}
def register(app):
return { "mesh_writer": GCodeWriter.GCodeWriter() }
Add mime types to GCodeWriter plugin# Copyright (c) 2015 Ultimaker B.V.
# Cura is released under the terms of the AGPLv3 or higher.
from . import GCodeWriter
from UM.i18n import i18nCatalog
catalog = i18nCatalog("cura")
def getMetaData():
return {
"type": "mesh_writer",
"plugin": {
"name": "GCode Writer",
"author": "Ultimaker",
"version": "1.0",
"description": catalog.i18nc("GCode Writer Plugin Description", "Writes GCode to a file")
},
"mesh_writer": {
"extension": "gcode",
"description": catalog.i18nc("GCode Writer File Description", "GCode File"),
"mime_types": [
"text/x-gcode"
]
}
}
def register(app):
return { "mesh_writer": GCodeWriter.GCodeWriter() }
|
<commit_before># Copyright (c) 2015 Ultimaker B.V.
# Cura is released under the terms of the AGPLv3 or higher.
from . import GCodeWriter
from UM.i18n import i18nCatalog
catalog = i18nCatalog("cura")
def getMetaData():
return {
"type": "mesh_writer",
"plugin": {
"name": "GCode Writer",
"author": "Ultimaker",
"version": "1.0",
"description": catalog.i18nc("GCode Writer Plugin Description", "Writes GCode to a file")
},
"mesh_writer": {
"extension": "gcode",
"description": catalog.i18nc("GCode Writer File Description", "GCode File")
}
}
def register(app):
return { "mesh_writer": GCodeWriter.GCodeWriter() }
<commit_msg>Add mime types to GCodeWriter plugin<commit_after># Copyright (c) 2015 Ultimaker B.V.
# Cura is released under the terms of the AGPLv3 or higher.
from . import GCodeWriter
from UM.i18n import i18nCatalog
catalog = i18nCatalog("cura")
def getMetaData():
return {
"type": "mesh_writer",
"plugin": {
"name": "GCode Writer",
"author": "Ultimaker",
"version": "1.0",
"description": catalog.i18nc("GCode Writer Plugin Description", "Writes GCode to a file")
},
"mesh_writer": {
"extension": "gcode",
"description": catalog.i18nc("GCode Writer File Description", "GCode File"),
"mime_types": [
"text/x-gcode"
]
}
}
def register(app):
return { "mesh_writer": GCodeWriter.GCodeWriter() }
|
87a72b219c2699e6bbb4354ae4b4f4ee356fd2c5
|
plumeria/plugins/bing_images.py
|
plumeria/plugins/bing_images.py
|
from aiohttp import BasicAuth
from plumeria import config
from plumeria.command import commands, CommandError
from plumeria.message import Response
from plumeria.util import http
from plumeria.util.ratelimit import rate_limit
SEARCH_URL = "https://api.datamarket.azure.com/Bing/Search/v1/Image"
api_key = config.create("bing", "key",
fallback="unset",
comment="An API key from Bing")
@commands.register("image", "i", category="Search")
@rate_limit()
async def image(message):
"""
Search Bing for an image.
"""
q = message.content.strip()
if not q:
raise CommandError("Search term required!")
r = await http.get(SEARCH_URL, params=[
('$format', 'json'),
('$top', '10'),
('Query', "'{}'".format(q)),
], auth=BasicAuth("", password=api_key()))
data = r.json()['d']
if len(data['results']):
return Response(data['results'][0]['MediaUrl'])
else:
raise CommandError("no results found")
|
from aiohttp import BasicAuth
from plumeria import config
from plumeria.command import commands, CommandError
from plumeria.message import Response
from plumeria.util import http
from plumeria.util.ratelimit import rate_limit
SEARCH_URL = "https://api.datamarket.azure.com/Bing/Search/v1/Image"
api_key = config.create("bing", "key",
fallback="unset",
comment="An API key from Bing")
@commands.register("image", "images", ""i", category="Search")
@rate_limit()
async def image(message):
"""
Search Bing for an image.
"""
q = message.content.strip()
if not q:
raise CommandError("Search term required!")
r = await http.get(SEARCH_URL, params=[
('$format', 'json'),
('$top', '10'),
('Query', "'{}'".format(q)),
], auth=BasicAuth("", password=api_key()))
data = r.json()['d']
if len(data['results']):
return Response(data['results'][0]['MediaUrl'])
else:
raise CommandError("no results found")
|
Add !images as Bing !image alias.
|
Add !images as Bing !image alias.
|
Python
|
mit
|
sk89q/Plumeria,sk89q/Plumeria,sk89q/Plumeria
|
from aiohttp import BasicAuth
from plumeria import config
from plumeria.command import commands, CommandError
from plumeria.message import Response
from plumeria.util import http
from plumeria.util.ratelimit import rate_limit
SEARCH_URL = "https://api.datamarket.azure.com/Bing/Search/v1/Image"
api_key = config.create("bing", "key",
fallback="unset",
comment="An API key from Bing")
@commands.register("image", "i", category="Search")
@rate_limit()
async def image(message):
"""
Search Bing for an image.
"""
q = message.content.strip()
if not q:
raise CommandError("Search term required!")
r = await http.get(SEARCH_URL, params=[
('$format', 'json'),
('$top', '10'),
('Query', "'{}'".format(q)),
], auth=BasicAuth("", password=api_key()))
data = r.json()['d']
if len(data['results']):
return Response(data['results'][0]['MediaUrl'])
else:
raise CommandError("no results found")
Add !images as Bing !image alias.
|
from aiohttp import BasicAuth
from plumeria import config
from plumeria.command import commands, CommandError
from plumeria.message import Response
from plumeria.util import http
from plumeria.util.ratelimit import rate_limit
SEARCH_URL = "https://api.datamarket.azure.com/Bing/Search/v1/Image"
api_key = config.create("bing", "key",
fallback="unset",
comment="An API key from Bing")
@commands.register("image", "images", ""i", category="Search")
@rate_limit()
async def image(message):
"""
Search Bing for an image.
"""
q = message.content.strip()
if not q:
raise CommandError("Search term required!")
r = await http.get(SEARCH_URL, params=[
('$format', 'json'),
('$top', '10'),
('Query', "'{}'".format(q)),
], auth=BasicAuth("", password=api_key()))
data = r.json()['d']
if len(data['results']):
return Response(data['results'][0]['MediaUrl'])
else:
raise CommandError("no results found")
|
<commit_before>from aiohttp import BasicAuth
from plumeria import config
from plumeria.command import commands, CommandError
from plumeria.message import Response
from plumeria.util import http
from plumeria.util.ratelimit import rate_limit
SEARCH_URL = "https://api.datamarket.azure.com/Bing/Search/v1/Image"
api_key = config.create("bing", "key",
fallback="unset",
comment="An API key from Bing")
@commands.register("image", "i", category="Search")
@rate_limit()
async def image(message):
"""
Search Bing for an image.
"""
q = message.content.strip()
if not q:
raise CommandError("Search term required!")
r = await http.get(SEARCH_URL, params=[
('$format', 'json'),
('$top', '10'),
('Query', "'{}'".format(q)),
], auth=BasicAuth("", password=api_key()))
data = r.json()['d']
if len(data['results']):
return Response(data['results'][0]['MediaUrl'])
else:
raise CommandError("no results found")
<commit_msg>Add !images as Bing !image alias.<commit_after>
|
from aiohttp import BasicAuth
from plumeria import config
from plumeria.command import commands, CommandError
from plumeria.message import Response
from plumeria.util import http
from plumeria.util.ratelimit import rate_limit
SEARCH_URL = "https://api.datamarket.azure.com/Bing/Search/v1/Image"
api_key = config.create("bing", "key",
fallback="unset",
comment="An API key from Bing")
@commands.register("image", "images", ""i", category="Search")
@rate_limit()
async def image(message):
"""
Search Bing for an image.
"""
q = message.content.strip()
if not q:
raise CommandError("Search term required!")
r = await http.get(SEARCH_URL, params=[
('$format', 'json'),
('$top', '10'),
('Query', "'{}'".format(q)),
], auth=BasicAuth("", password=api_key()))
data = r.json()['d']
if len(data['results']):
return Response(data['results'][0]['MediaUrl'])
else:
raise CommandError("no results found")
|
from aiohttp import BasicAuth
from plumeria import config
from plumeria.command import commands, CommandError
from plumeria.message import Response
from plumeria.util import http
from plumeria.util.ratelimit import rate_limit
SEARCH_URL = "https://api.datamarket.azure.com/Bing/Search/v1/Image"
api_key = config.create("bing", "key",
fallback="unset",
comment="An API key from Bing")
@commands.register("image", "i", category="Search")
@rate_limit()
async def image(message):
"""
Search Bing for an image.
"""
q = message.content.strip()
if not q:
raise CommandError("Search term required!")
r = await http.get(SEARCH_URL, params=[
('$format', 'json'),
('$top', '10'),
('Query', "'{}'".format(q)),
], auth=BasicAuth("", password=api_key()))
data = r.json()['d']
if len(data['results']):
return Response(data['results'][0]['MediaUrl'])
else:
raise CommandError("no results found")
Add !images as Bing !image alias.from aiohttp import BasicAuth
from plumeria import config
from plumeria.command import commands, CommandError
from plumeria.message import Response
from plumeria.util import http
from plumeria.util.ratelimit import rate_limit
SEARCH_URL = "https://api.datamarket.azure.com/Bing/Search/v1/Image"
api_key = config.create("bing", "key",
fallback="unset",
comment="An API key from Bing")
@commands.register("image", "images", ""i", category="Search")
@rate_limit()
async def image(message):
"""
Search Bing for an image.
"""
q = message.content.strip()
if not q:
raise CommandError("Search term required!")
r = await http.get(SEARCH_URL, params=[
('$format', 'json'),
('$top', '10'),
('Query', "'{}'".format(q)),
], auth=BasicAuth("", password=api_key()))
data = r.json()['d']
if len(data['results']):
return Response(data['results'][0]['MediaUrl'])
else:
raise CommandError("no results found")
|
<commit_before>from aiohttp import BasicAuth
from plumeria import config
from plumeria.command import commands, CommandError
from plumeria.message import Response
from plumeria.util import http
from plumeria.util.ratelimit import rate_limit
SEARCH_URL = "https://api.datamarket.azure.com/Bing/Search/v1/Image"
api_key = config.create("bing", "key",
fallback="unset",
comment="An API key from Bing")
@commands.register("image", "i", category="Search")
@rate_limit()
async def image(message):
"""
Search Bing for an image.
"""
q = message.content.strip()
if not q:
raise CommandError("Search term required!")
r = await http.get(SEARCH_URL, params=[
('$format', 'json'),
('$top', '10'),
('Query', "'{}'".format(q)),
], auth=BasicAuth("", password=api_key()))
data = r.json()['d']
if len(data['results']):
return Response(data['results'][0]['MediaUrl'])
else:
raise CommandError("no results found")
<commit_msg>Add !images as Bing !image alias.<commit_after>from aiohttp import BasicAuth
from plumeria import config
from plumeria.command import commands, CommandError
from plumeria.message import Response
from plumeria.util import http
from plumeria.util.ratelimit import rate_limit
SEARCH_URL = "https://api.datamarket.azure.com/Bing/Search/v1/Image"
api_key = config.create("bing", "key",
fallback="unset",
comment="An API key from Bing")
@commands.register("image", "images", ""i", category="Search")
@rate_limit()
async def image(message):
"""
Search Bing for an image.
"""
q = message.content.strip()
if not q:
raise CommandError("Search term required!")
r = await http.get(SEARCH_URL, params=[
('$format', 'json'),
('$top', '10'),
('Query', "'{}'".format(q)),
], auth=BasicAuth("", password=api_key()))
data = r.json()['d']
if len(data['results']):
return Response(data['results'][0]['MediaUrl'])
else:
raise CommandError("no results found")
|
2fbc2522f473c5255e678d435689aead9116d2d3
|
sieve/sieve.py
|
sieve/sieve.py
|
def sieve(n):
if n < 2:
return []
not_prime = set()
prime = [2]
for i in range(3, n+1, 2):
if i not in not_prime:
prime.append(i)
not_prime.update(range(i*i, n, i))
return prime
|
def sieve(n):
if n < 2:
return []
not_prime = set()
prime = [2]
for i in range(3, n+1, 2):
if i not in not_prime:
prime.append(i)
not_prime.update(range(i*i, n+1, i))
return prime
|
Fix bug where n is the square of a prime
|
Fix bug where n is the square of a prime
|
Python
|
agpl-3.0
|
CubicComet/exercism-python-solutions
|
def sieve(n):
if n < 2:
return []
not_prime = set()
prime = [2]
for i in range(3, n+1, 2):
if i not in not_prime:
prime.append(i)
not_prime.update(range(i*i, n, i))
return prime
Fix bug where n is the square of a prime
|
def sieve(n):
if n < 2:
return []
not_prime = set()
prime = [2]
for i in range(3, n+1, 2):
if i not in not_prime:
prime.append(i)
not_prime.update(range(i*i, n+1, i))
return prime
|
<commit_before>def sieve(n):
if n < 2:
return []
not_prime = set()
prime = [2]
for i in range(3, n+1, 2):
if i not in not_prime:
prime.append(i)
not_prime.update(range(i*i, n, i))
return prime
<commit_msg>Fix bug where n is the square of a prime<commit_after>
|
def sieve(n):
if n < 2:
return []
not_prime = set()
prime = [2]
for i in range(3, n+1, 2):
if i not in not_prime:
prime.append(i)
not_prime.update(range(i*i, n+1, i))
return prime
|
def sieve(n):
if n < 2:
return []
not_prime = set()
prime = [2]
for i in range(3, n+1, 2):
if i not in not_prime:
prime.append(i)
not_prime.update(range(i*i, n, i))
return prime
Fix bug where n is the square of a primedef sieve(n):
if n < 2:
return []
not_prime = set()
prime = [2]
for i in range(3, n+1, 2):
if i not in not_prime:
prime.append(i)
not_prime.update(range(i*i, n+1, i))
return prime
|
<commit_before>def sieve(n):
if n < 2:
return []
not_prime = set()
prime = [2]
for i in range(3, n+1, 2):
if i not in not_prime:
prime.append(i)
not_prime.update(range(i*i, n, i))
return prime
<commit_msg>Fix bug where n is the square of a prime<commit_after>def sieve(n):
if n < 2:
return []
not_prime = set()
prime = [2]
for i in range(3, n+1, 2):
if i not in not_prime:
prime.append(i)
not_prime.update(range(i*i, n+1, i))
return prime
|
4f7371dad85843c42b9cb427edebe5020586b61e
|
server/core/management/commands/poll_urls.py
|
server/core/management/commands/poll_urls.py
|
import datetime
import multiprocessing
from django.conf import settings
from django.core.management.base import BaseCommand, CommandError
from core.models import WebPage, PageScrapeResult
from core.views import scrape_url
class Command(BaseCommand):
help = "Poll all the urls and scrape the results"
can_import_settings = True
def handle(self, *args, **options):
pool = multiprocessing.Pool(multiprocessing.cpu_count() + 2)
today = int(datetime.date.today().strftime("%s"))
now = datetime.datetime.now()
curr_time = float(int(now.strftime("%s")) - now.second)
mins_passed = curr_time - today
for page in WebPage.objects.all():
if mins_passed % page.interval == 0 or settings.DEBUG:
pool.apply_async(scrape_url, (page, ))
pool.close()
pool.join()
|
import datetime
import multiprocessing
from django.conf import settings
from django.core.management.base import BaseCommand, CommandError
from core.models import WebPage, PageScrapeResult
from core.views import scrape_url
class Command(BaseCommand):
help = "Poll all the urls and scrape the results"
can_import_settings = True
def handle(self, *args, **options):
pool = multiprocessing.Pool(multiprocessing.cpu_count())
today = int(datetime.date.today().strftime("%s"))
now = datetime.datetime.now()
curr_time = float(int(now.strftime("%s")) - now.second)
mins_passed = curr_time - today
for page in WebPage.objects.all():
if mins_passed % page.interval == 0 or settings.DEBUG:
pool.apply_async(scrape_url, (page, ))
pool.close()
pool.join()
|
Reduce the number of processes used.
|
Reduce the number of processes used.
|
Python
|
mit
|
theju/atifier,theju/atifier
|
import datetime
import multiprocessing
from django.conf import settings
from django.core.management.base import BaseCommand, CommandError
from core.models import WebPage, PageScrapeResult
from core.views import scrape_url
class Command(BaseCommand):
help = "Poll all the urls and scrape the results"
can_import_settings = True
def handle(self, *args, **options):
pool = multiprocessing.Pool(multiprocessing.cpu_count() + 2)
today = int(datetime.date.today().strftime("%s"))
now = datetime.datetime.now()
curr_time = float(int(now.strftime("%s")) - now.second)
mins_passed = curr_time - today
for page in WebPage.objects.all():
if mins_passed % page.interval == 0 or settings.DEBUG:
pool.apply_async(scrape_url, (page, ))
pool.close()
pool.join()
Reduce the number of processes used.
|
import datetime
import multiprocessing
from django.conf import settings
from django.core.management.base import BaseCommand, CommandError
from core.models import WebPage, PageScrapeResult
from core.views import scrape_url
class Command(BaseCommand):
help = "Poll all the urls and scrape the results"
can_import_settings = True
def handle(self, *args, **options):
pool = multiprocessing.Pool(multiprocessing.cpu_count())
today = int(datetime.date.today().strftime("%s"))
now = datetime.datetime.now()
curr_time = float(int(now.strftime("%s")) - now.second)
mins_passed = curr_time - today
for page in WebPage.objects.all():
if mins_passed % page.interval == 0 or settings.DEBUG:
pool.apply_async(scrape_url, (page, ))
pool.close()
pool.join()
|
<commit_before>import datetime
import multiprocessing
from django.conf import settings
from django.core.management.base import BaseCommand, CommandError
from core.models import WebPage, PageScrapeResult
from core.views import scrape_url
class Command(BaseCommand):
help = "Poll all the urls and scrape the results"
can_import_settings = True
def handle(self, *args, **options):
pool = multiprocessing.Pool(multiprocessing.cpu_count() + 2)
today = int(datetime.date.today().strftime("%s"))
now = datetime.datetime.now()
curr_time = float(int(now.strftime("%s")) - now.second)
mins_passed = curr_time - today
for page in WebPage.objects.all():
if mins_passed % page.interval == 0 or settings.DEBUG:
pool.apply_async(scrape_url, (page, ))
pool.close()
pool.join()
<commit_msg>Reduce the number of processes used.<commit_after>
|
import datetime
import multiprocessing
from django.conf import settings
from django.core.management.base import BaseCommand, CommandError
from core.models import WebPage, PageScrapeResult
from core.views import scrape_url
class Command(BaseCommand):
help = "Poll all the urls and scrape the results"
can_import_settings = True
def handle(self, *args, **options):
pool = multiprocessing.Pool(multiprocessing.cpu_count())
today = int(datetime.date.today().strftime("%s"))
now = datetime.datetime.now()
curr_time = float(int(now.strftime("%s")) - now.second)
mins_passed = curr_time - today
for page in WebPage.objects.all():
if mins_passed % page.interval == 0 or settings.DEBUG:
pool.apply_async(scrape_url, (page, ))
pool.close()
pool.join()
|
import datetime
import multiprocessing
from django.conf import settings
from django.core.management.base import BaseCommand, CommandError
from core.models import WebPage, PageScrapeResult
from core.views import scrape_url
class Command(BaseCommand):
help = "Poll all the urls and scrape the results"
can_import_settings = True
def handle(self, *args, **options):
pool = multiprocessing.Pool(multiprocessing.cpu_count() + 2)
today = int(datetime.date.today().strftime("%s"))
now = datetime.datetime.now()
curr_time = float(int(now.strftime("%s")) - now.second)
mins_passed = curr_time - today
for page in WebPage.objects.all():
if mins_passed % page.interval == 0 or settings.DEBUG:
pool.apply_async(scrape_url, (page, ))
pool.close()
pool.join()
Reduce the number of processes used.import datetime
import multiprocessing
from django.conf import settings
from django.core.management.base import BaseCommand, CommandError
from core.models import WebPage, PageScrapeResult
from core.views import scrape_url
class Command(BaseCommand):
help = "Poll all the urls and scrape the results"
can_import_settings = True
def handle(self, *args, **options):
pool = multiprocessing.Pool(multiprocessing.cpu_count())
today = int(datetime.date.today().strftime("%s"))
now = datetime.datetime.now()
curr_time = float(int(now.strftime("%s")) - now.second)
mins_passed = curr_time - today
for page in WebPage.objects.all():
if mins_passed % page.interval == 0 or settings.DEBUG:
pool.apply_async(scrape_url, (page, ))
pool.close()
pool.join()
|
<commit_before>import datetime
import multiprocessing
from django.conf import settings
from django.core.management.base import BaseCommand, CommandError
from core.models import WebPage, PageScrapeResult
from core.views import scrape_url
class Command(BaseCommand):
help = "Poll all the urls and scrape the results"
can_import_settings = True
def handle(self, *args, **options):
pool = multiprocessing.Pool(multiprocessing.cpu_count() + 2)
today = int(datetime.date.today().strftime("%s"))
now = datetime.datetime.now()
curr_time = float(int(now.strftime("%s")) - now.second)
mins_passed = curr_time - today
for page in WebPage.objects.all():
if mins_passed % page.interval == 0 or settings.DEBUG:
pool.apply_async(scrape_url, (page, ))
pool.close()
pool.join()
<commit_msg>Reduce the number of processes used.<commit_after>import datetime
import multiprocessing
from django.conf import settings
from django.core.management.base import BaseCommand, CommandError
from core.models import WebPage, PageScrapeResult
from core.views import scrape_url
class Command(BaseCommand):
help = "Poll all the urls and scrape the results"
can_import_settings = True
def handle(self, *args, **options):
pool = multiprocessing.Pool(multiprocessing.cpu_count())
today = int(datetime.date.today().strftime("%s"))
now = datetime.datetime.now()
curr_time = float(int(now.strftime("%s")) - now.second)
mins_passed = curr_time - today
for page in WebPage.objects.all():
if mins_passed % page.interval == 0 or settings.DEBUG:
pool.apply_async(scrape_url, (page, ))
pool.close()
pool.join()
|
e8c6be3565bd8b33dfb7a01dfb77938534ce9d09
|
pysswords/crypt.py
|
pysswords/crypt.py
|
import os
import gnupg
import logging
from .utils import which
def create_key_input(gpg, passphrase, testing=False):
key_input = gpg.gen_key_input(
name_real='Pysswords',
name_email='pysswords@pysswords',
name_comment='Autogenerated by Pysswords',
passphrase=passphrase,
testing=testing
)
return key_input
def create_gpg(binary, database_path, passphrase):
gnupg_path = os.path.join(database_path, ".gnupg")
gpg = gnupg.GPG(which(binary), homedir=gnupg_path)
gpg.gen_key(create_key_input(gpg, passphrase))
return gpg
|
import os
import gnupg
import logging
from .utils import which
def create_key_input(gpg, passphrase, testing=False):
key_input = gpg.gen_key_input(
name_real='Pysswords',
name_email='pysswords@pysswords',
name_comment='Autogenerated by Pysswords',
passphrase=passphrase,
testing=testing
)
return key_input
def create_gpg(binary, database_path, passphrase):
gnupg_path = os.path.join(database_path, ".gnupg")
gpg = gnupg.GPG(which(binary), homedir=gnupg_path)
gpg.gen_key(create_key_input(gpg, passphrase))
return gpg
def load_gpg(binary, database_path):
gnupg_path = os.path.join(database_path, ".gnupg")
gpg = gnupg.GPG(which(binary), homedir=gnupg_path)
return gpg
|
Add load gpg to get an instance of gpg
|
Add load gpg to get an instance of gpg
|
Python
|
mit
|
scorphus/passpie,scorphus/passpie,marcwebbie/passpie,eiginn/passpie,marcwebbie/pysswords,marcwebbie/passpie,eiginn/passpie
|
import os
import gnupg
import logging
from .utils import which
def create_key_input(gpg, passphrase, testing=False):
key_input = gpg.gen_key_input(
name_real='Pysswords',
name_email='pysswords@pysswords',
name_comment='Autogenerated by Pysswords',
passphrase=passphrase,
testing=testing
)
return key_input
def create_gpg(binary, database_path, passphrase):
gnupg_path = os.path.join(database_path, ".gnupg")
gpg = gnupg.GPG(which(binary), homedir=gnupg_path)
gpg.gen_key(create_key_input(gpg, passphrase))
return gpg
Add load gpg to get an instance of gpg
|
import os
import gnupg
import logging
from .utils import which
def create_key_input(gpg, passphrase, testing=False):
key_input = gpg.gen_key_input(
name_real='Pysswords',
name_email='pysswords@pysswords',
name_comment='Autogenerated by Pysswords',
passphrase=passphrase,
testing=testing
)
return key_input
def create_gpg(binary, database_path, passphrase):
gnupg_path = os.path.join(database_path, ".gnupg")
gpg = gnupg.GPG(which(binary), homedir=gnupg_path)
gpg.gen_key(create_key_input(gpg, passphrase))
return gpg
def load_gpg(binary, database_path):
gnupg_path = os.path.join(database_path, ".gnupg")
gpg = gnupg.GPG(which(binary), homedir=gnupg_path)
return gpg
|
<commit_before>import os
import gnupg
import logging
from .utils import which
def create_key_input(gpg, passphrase, testing=False):
key_input = gpg.gen_key_input(
name_real='Pysswords',
name_email='pysswords@pysswords',
name_comment='Autogenerated by Pysswords',
passphrase=passphrase,
testing=testing
)
return key_input
def create_gpg(binary, database_path, passphrase):
gnupg_path = os.path.join(database_path, ".gnupg")
gpg = gnupg.GPG(which(binary), homedir=gnupg_path)
gpg.gen_key(create_key_input(gpg, passphrase))
return gpg
<commit_msg>Add load gpg to get an instance of gpg<commit_after>
|
import os
import gnupg
import logging
from .utils import which
def create_key_input(gpg, passphrase, testing=False):
key_input = gpg.gen_key_input(
name_real='Pysswords',
name_email='pysswords@pysswords',
name_comment='Autogenerated by Pysswords',
passphrase=passphrase,
testing=testing
)
return key_input
def create_gpg(binary, database_path, passphrase):
gnupg_path = os.path.join(database_path, ".gnupg")
gpg = gnupg.GPG(which(binary), homedir=gnupg_path)
gpg.gen_key(create_key_input(gpg, passphrase))
return gpg
def load_gpg(binary, database_path):
gnupg_path = os.path.join(database_path, ".gnupg")
gpg = gnupg.GPG(which(binary), homedir=gnupg_path)
return gpg
|
import os
import gnupg
import logging
from .utils import which
def create_key_input(gpg, passphrase, testing=False):
key_input = gpg.gen_key_input(
name_real='Pysswords',
name_email='pysswords@pysswords',
name_comment='Autogenerated by Pysswords',
passphrase=passphrase,
testing=testing
)
return key_input
def create_gpg(binary, database_path, passphrase):
gnupg_path = os.path.join(database_path, ".gnupg")
gpg = gnupg.GPG(which(binary), homedir=gnupg_path)
gpg.gen_key(create_key_input(gpg, passphrase))
return gpg
Add load gpg to get an instance of gpgimport os
import gnupg
import logging
from .utils import which
def create_key_input(gpg, passphrase, testing=False):
key_input = gpg.gen_key_input(
name_real='Pysswords',
name_email='pysswords@pysswords',
name_comment='Autogenerated by Pysswords',
passphrase=passphrase,
testing=testing
)
return key_input
def create_gpg(binary, database_path, passphrase):
gnupg_path = os.path.join(database_path, ".gnupg")
gpg = gnupg.GPG(which(binary), homedir=gnupg_path)
gpg.gen_key(create_key_input(gpg, passphrase))
return gpg
def load_gpg(binary, database_path):
gnupg_path = os.path.join(database_path, ".gnupg")
gpg = gnupg.GPG(which(binary), homedir=gnupg_path)
return gpg
|
<commit_before>import os
import gnupg
import logging
from .utils import which
def create_key_input(gpg, passphrase, testing=False):
key_input = gpg.gen_key_input(
name_real='Pysswords',
name_email='pysswords@pysswords',
name_comment='Autogenerated by Pysswords',
passphrase=passphrase,
testing=testing
)
return key_input
def create_gpg(binary, database_path, passphrase):
gnupg_path = os.path.join(database_path, ".gnupg")
gpg = gnupg.GPG(which(binary), homedir=gnupg_path)
gpg.gen_key(create_key_input(gpg, passphrase))
return gpg
<commit_msg>Add load gpg to get an instance of gpg<commit_after>import os
import gnupg
import logging
from .utils import which
def create_key_input(gpg, passphrase, testing=False):
key_input = gpg.gen_key_input(
name_real='Pysswords',
name_email='pysswords@pysswords',
name_comment='Autogenerated by Pysswords',
passphrase=passphrase,
testing=testing
)
return key_input
def create_gpg(binary, database_path, passphrase):
gnupg_path = os.path.join(database_path, ".gnupg")
gpg = gnupg.GPG(which(binary), homedir=gnupg_path)
gpg.gen_key(create_key_input(gpg, passphrase))
return gpg
def load_gpg(binary, database_path):
gnupg_path = os.path.join(database_path, ".gnupg")
gpg = gnupg.GPG(which(binary), homedir=gnupg_path)
return gpg
|
044d2e1761a6330dad326470728ea4fadceef8d8
|
PropertyVerification/ContractDebugger.py
|
PropertyVerification/ContractDebugger.py
|
class ContractDebugger:
def __init__(self, pathCondGen):
self.pathCondGen = pathCondGen
def explain_failures(self, contract_name, contract, success_pcs, failed_pcs):
print("Explaining why contract fails: " + contract_name)
print(success_pcs)
print(failed_pcs)
self.get_rule_differences(success_pcs, failed_pcs)
def get_rule_differences(self, success_pcs, failed_pcs):
rules_in_success = self.get_rules(success_pcs)
rules_in_failed = self.get_rules(failed_pcs)
good_rules = [rule for rule in rules_in_success if not rule in rules_in_failed]
bad_rules = [rule for rule in rules_in_failed if not rule in rules_in_success]
print("Good rules: (Rules in success set and not failure set)")
print(good_rules)
print("Bad rules: (Rules in failure set and not success set)")
print(bad_rules)
def get_rules(self, pcs):
rules = []
for pc in pcs:
r = self.pathCondGen.rules_in_pc_real_name(pc)
print(r)
return rules
|
class ContractDebugger:
def __init__(self, pathCondGen):
self.pathCondGen = pathCondGen
def explain_failures(self, contract_name, contract, success_pcs, failed_pcs):
print("Explaining why contract fails: " + contract_name)
# print("Success PCs: ")
# print(success_pcs)
# print("Failed PCs: ")
# print(failed_pcs)
self.get_rule_differences(success_pcs, failed_pcs)
def get_rule_differences(self, success_pcs, failed_pcs):
rules_in_success = self.get_rules(success_pcs)
rules_in_failed = self.get_rules(failed_pcs)
good_rules = sorted([rule for rule in rules_in_success if not rule in rules_in_failed])
bad_rules = sorted([rule for rule in rules_in_failed if not rule in rules_in_success])
print("Good rules: (Rules in success set and not failure set)")
print(good_rules)
print("Bad rules: (Rules in failure set and not success set)")
print(bad_rules)
def get_rules(self, pcs):
rules = []
for pc in pcs:
rules += self.pathCondGen.rules_in_pc_real_name(pc)
return list(set(rules))
|
Print rule difference between good and bad sets.
|
Print rule difference between good and bad sets.
|
Python
|
mit
|
levilucio/SyVOLT,levilucio/SyVOLT
|
class ContractDebugger:
def __init__(self, pathCondGen):
self.pathCondGen = pathCondGen
def explain_failures(self, contract_name, contract, success_pcs, failed_pcs):
print("Explaining why contract fails: " + contract_name)
print(success_pcs)
print(failed_pcs)
self.get_rule_differences(success_pcs, failed_pcs)
def get_rule_differences(self, success_pcs, failed_pcs):
rules_in_success = self.get_rules(success_pcs)
rules_in_failed = self.get_rules(failed_pcs)
good_rules = [rule for rule in rules_in_success if not rule in rules_in_failed]
bad_rules = [rule for rule in rules_in_failed if not rule in rules_in_success]
print("Good rules: (Rules in success set and not failure set)")
print(good_rules)
print("Bad rules: (Rules in failure set and not success set)")
print(bad_rules)
def get_rules(self, pcs):
rules = []
for pc in pcs:
r = self.pathCondGen.rules_in_pc_real_name(pc)
print(r)
return rules
Print rule difference between good and bad sets.
|
class ContractDebugger:
def __init__(self, pathCondGen):
self.pathCondGen = pathCondGen
def explain_failures(self, contract_name, contract, success_pcs, failed_pcs):
print("Explaining why contract fails: " + contract_name)
# print("Success PCs: ")
# print(success_pcs)
# print("Failed PCs: ")
# print(failed_pcs)
self.get_rule_differences(success_pcs, failed_pcs)
def get_rule_differences(self, success_pcs, failed_pcs):
rules_in_success = self.get_rules(success_pcs)
rules_in_failed = self.get_rules(failed_pcs)
good_rules = sorted([rule for rule in rules_in_success if not rule in rules_in_failed])
bad_rules = sorted([rule for rule in rules_in_failed if not rule in rules_in_success])
print("Good rules: (Rules in success set and not failure set)")
print(good_rules)
print("Bad rules: (Rules in failure set and not success set)")
print(bad_rules)
def get_rules(self, pcs):
rules = []
for pc in pcs:
rules += self.pathCondGen.rules_in_pc_real_name(pc)
return list(set(rules))
|
<commit_before>class ContractDebugger:
def __init__(self, pathCondGen):
self.pathCondGen = pathCondGen
def explain_failures(self, contract_name, contract, success_pcs, failed_pcs):
print("Explaining why contract fails: " + contract_name)
print(success_pcs)
print(failed_pcs)
self.get_rule_differences(success_pcs, failed_pcs)
def get_rule_differences(self, success_pcs, failed_pcs):
rules_in_success = self.get_rules(success_pcs)
rules_in_failed = self.get_rules(failed_pcs)
good_rules = [rule for rule in rules_in_success if not rule in rules_in_failed]
bad_rules = [rule for rule in rules_in_failed if not rule in rules_in_success]
print("Good rules: (Rules in success set and not failure set)")
print(good_rules)
print("Bad rules: (Rules in failure set and not success set)")
print(bad_rules)
def get_rules(self, pcs):
rules = []
for pc in pcs:
r = self.pathCondGen.rules_in_pc_real_name(pc)
print(r)
return rules
<commit_msg>Print rule difference between good and bad sets.<commit_after>
|
class ContractDebugger:
def __init__(self, pathCondGen):
self.pathCondGen = pathCondGen
def explain_failures(self, contract_name, contract, success_pcs, failed_pcs):
print("Explaining why contract fails: " + contract_name)
# print("Success PCs: ")
# print(success_pcs)
# print("Failed PCs: ")
# print(failed_pcs)
self.get_rule_differences(success_pcs, failed_pcs)
def get_rule_differences(self, success_pcs, failed_pcs):
rules_in_success = self.get_rules(success_pcs)
rules_in_failed = self.get_rules(failed_pcs)
good_rules = sorted([rule for rule in rules_in_success if not rule in rules_in_failed])
bad_rules = sorted([rule for rule in rules_in_failed if not rule in rules_in_success])
print("Good rules: (Rules in success set and not failure set)")
print(good_rules)
print("Bad rules: (Rules in failure set and not success set)")
print(bad_rules)
def get_rules(self, pcs):
rules = []
for pc in pcs:
rules += self.pathCondGen.rules_in_pc_real_name(pc)
return list(set(rules))
|
class ContractDebugger:
def __init__(self, pathCondGen):
self.pathCondGen = pathCondGen
def explain_failures(self, contract_name, contract, success_pcs, failed_pcs):
print("Explaining why contract fails: " + contract_name)
print(success_pcs)
print(failed_pcs)
self.get_rule_differences(success_pcs, failed_pcs)
def get_rule_differences(self, success_pcs, failed_pcs):
rules_in_success = self.get_rules(success_pcs)
rules_in_failed = self.get_rules(failed_pcs)
good_rules = [rule for rule in rules_in_success if not rule in rules_in_failed]
bad_rules = [rule for rule in rules_in_failed if not rule in rules_in_success]
print("Good rules: (Rules in success set and not failure set)")
print(good_rules)
print("Bad rules: (Rules in failure set and not success set)")
print(bad_rules)
def get_rules(self, pcs):
rules = []
for pc in pcs:
r = self.pathCondGen.rules_in_pc_real_name(pc)
print(r)
return rules
Print rule difference between good and bad sets.class ContractDebugger:
def __init__(self, pathCondGen):
self.pathCondGen = pathCondGen
def explain_failures(self, contract_name, contract, success_pcs, failed_pcs):
print("Explaining why contract fails: " + contract_name)
# print("Success PCs: ")
# print(success_pcs)
# print("Failed PCs: ")
# print(failed_pcs)
self.get_rule_differences(success_pcs, failed_pcs)
def get_rule_differences(self, success_pcs, failed_pcs):
rules_in_success = self.get_rules(success_pcs)
rules_in_failed = self.get_rules(failed_pcs)
good_rules = sorted([rule for rule in rules_in_success if not rule in rules_in_failed])
bad_rules = sorted([rule for rule in rules_in_failed if not rule in rules_in_success])
print("Good rules: (Rules in success set and not failure set)")
print(good_rules)
print("Bad rules: (Rules in failure set and not success set)")
print(bad_rules)
def get_rules(self, pcs):
rules = []
for pc in pcs:
rules += self.pathCondGen.rules_in_pc_real_name(pc)
return list(set(rules))
|
<commit_before>class ContractDebugger:
def __init__(self, pathCondGen):
self.pathCondGen = pathCondGen
def explain_failures(self, contract_name, contract, success_pcs, failed_pcs):
print("Explaining why contract fails: " + contract_name)
print(success_pcs)
print(failed_pcs)
self.get_rule_differences(success_pcs, failed_pcs)
def get_rule_differences(self, success_pcs, failed_pcs):
rules_in_success = self.get_rules(success_pcs)
rules_in_failed = self.get_rules(failed_pcs)
good_rules = [rule for rule in rules_in_success if not rule in rules_in_failed]
bad_rules = [rule for rule in rules_in_failed if not rule in rules_in_success]
print("Good rules: (Rules in success set and not failure set)")
print(good_rules)
print("Bad rules: (Rules in failure set and not success set)")
print(bad_rules)
def get_rules(self, pcs):
rules = []
for pc in pcs:
r = self.pathCondGen.rules_in_pc_real_name(pc)
print(r)
return rules
<commit_msg>Print rule difference between good and bad sets.<commit_after>class ContractDebugger:
def __init__(self, pathCondGen):
self.pathCondGen = pathCondGen
def explain_failures(self, contract_name, contract, success_pcs, failed_pcs):
print("Explaining why contract fails: " + contract_name)
# print("Success PCs: ")
# print(success_pcs)
# print("Failed PCs: ")
# print(failed_pcs)
self.get_rule_differences(success_pcs, failed_pcs)
def get_rule_differences(self, success_pcs, failed_pcs):
rules_in_success = self.get_rules(success_pcs)
rules_in_failed = self.get_rules(failed_pcs)
good_rules = sorted([rule for rule in rules_in_success if not rule in rules_in_failed])
bad_rules = sorted([rule for rule in rules_in_failed if not rule in rules_in_success])
print("Good rules: (Rules in success set and not failure set)")
print(good_rules)
print("Bad rules: (Rules in failure set and not success set)")
print(bad_rules)
def get_rules(self, pcs):
rules = []
for pc in pcs:
rules += self.pathCondGen.rules_in_pc_real_name(pc)
return list(set(rules))
|
04bbe400396a5ef5b930b9db9d8d8e30ff6bf678
|
medical_patient_ethnicity/models/medical_patient_ethnicity.py
|
medical_patient_ethnicity/models/medical_patient_ethnicity.py
|
# -*- coding: utf-8 -*-
# #############################################################################
#
# Tech-Receptives Solutions Pvt. Ltd.
# Copyright (C) 2004-TODAY Tech-Receptives(<http://www.techreceptives.com>)
# Special Credit and Thanks to Thymbra Latinoamericana S.A.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
# #############################################################################
from openerp import models, fields
class MedicalPatientEthnicity(models.Model):
_name = 'medical.patient.ethnicity'
notes = fields.Char()
code = fields.Char(required=True, )
name = fields.Char(required=True, translate=True)
_sql_constraints = [
('name_uniq', 'UNIQUE(name)', 'Ethnicity name must be unique!'),
('code_uniq', 'UNIQUE(code)', 'Ethnicity code must be unique!'),
]
|
# -*- coding: utf-8 -*-
# #############################################################################
#
# Tech-Receptives Solutions Pvt. Ltd.
# Copyright (C) 2004-TODAY Tech-Receptives(<http://www.techreceptives.com>)
# Special Credit and Thanks to Thymbra Latinoamericana S.A.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
# #############################################################################
from openerp import models, fields
class MedicalPatientEthnicity(models.Model):
_name = 'medical.patient.ethnicity'
_description = 'Medical Patient Ethnicity'
notes = fields.Char()
code = fields.Char(required=True, )
name = fields.Char(required=True, translate=True)
_sql_constraints = [
('name_uniq', 'UNIQUE(name)', 'Ethnicity name must be unique!'),
('code_uniq', 'UNIQUE(code)', 'Ethnicity code must be unique!'),
]
|
Add description to ethnicity model
|
Add description to ethnicity model
|
Python
|
agpl-3.0
|
ShaheenHossain/eagle-medical,laslabs/vertical-medical,laslabs/vertical-medical,ShaheenHossain/eagle-medical
|
# -*- coding: utf-8 -*-
# #############################################################################
#
# Tech-Receptives Solutions Pvt. Ltd.
# Copyright (C) 2004-TODAY Tech-Receptives(<http://www.techreceptives.com>)
# Special Credit and Thanks to Thymbra Latinoamericana S.A.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
# #############################################################################
from openerp import models, fields
class MedicalPatientEthnicity(models.Model):
_name = 'medical.patient.ethnicity'
notes = fields.Char()
code = fields.Char(required=True, )
name = fields.Char(required=True, translate=True)
_sql_constraints = [
('name_uniq', 'UNIQUE(name)', 'Ethnicity name must be unique!'),
('code_uniq', 'UNIQUE(code)', 'Ethnicity code must be unique!'),
]
Add description to ethnicity model
|
# -*- coding: utf-8 -*-
# #############################################################################
#
# Tech-Receptives Solutions Pvt. Ltd.
# Copyright (C) 2004-TODAY Tech-Receptives(<http://www.techreceptives.com>)
# Special Credit and Thanks to Thymbra Latinoamericana S.A.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
# #############################################################################
from openerp import models, fields
class MedicalPatientEthnicity(models.Model):
_name = 'medical.patient.ethnicity'
_description = 'Medical Patient Ethnicity'
notes = fields.Char()
code = fields.Char(required=True, )
name = fields.Char(required=True, translate=True)
_sql_constraints = [
('name_uniq', 'UNIQUE(name)', 'Ethnicity name must be unique!'),
('code_uniq', 'UNIQUE(code)', 'Ethnicity code must be unique!'),
]
|
<commit_before># -*- coding: utf-8 -*-
# #############################################################################
#
# Tech-Receptives Solutions Pvt. Ltd.
# Copyright (C) 2004-TODAY Tech-Receptives(<http://www.techreceptives.com>)
# Special Credit and Thanks to Thymbra Latinoamericana S.A.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
# #############################################################################
from openerp import models, fields
class MedicalPatientEthnicity(models.Model):
_name = 'medical.patient.ethnicity'
notes = fields.Char()
code = fields.Char(required=True, )
name = fields.Char(required=True, translate=True)
_sql_constraints = [
('name_uniq', 'UNIQUE(name)', 'Ethnicity name must be unique!'),
('code_uniq', 'UNIQUE(code)', 'Ethnicity code must be unique!'),
]
<commit_msg>Add description to ethnicity model<commit_after>
|
# -*- coding: utf-8 -*-
# #############################################################################
#
# Tech-Receptives Solutions Pvt. Ltd.
# Copyright (C) 2004-TODAY Tech-Receptives(<http://www.techreceptives.com>)
# Special Credit and Thanks to Thymbra Latinoamericana S.A.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
# #############################################################################
from openerp import models, fields
class MedicalPatientEthnicity(models.Model):
_name = 'medical.patient.ethnicity'
_description = 'Medical Patient Ethnicity'
notes = fields.Char()
code = fields.Char(required=True, )
name = fields.Char(required=True, translate=True)
_sql_constraints = [
('name_uniq', 'UNIQUE(name)', 'Ethnicity name must be unique!'),
('code_uniq', 'UNIQUE(code)', 'Ethnicity code must be unique!'),
]
|
# -*- coding: utf-8 -*-
# #############################################################################
#
# Tech-Receptives Solutions Pvt. Ltd.
# Copyright (C) 2004-TODAY Tech-Receptives(<http://www.techreceptives.com>)
# Special Credit and Thanks to Thymbra Latinoamericana S.A.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
# #############################################################################
from openerp import models, fields
class MedicalPatientEthnicity(models.Model):
_name = 'medical.patient.ethnicity'
notes = fields.Char()
code = fields.Char(required=True, )
name = fields.Char(required=True, translate=True)
_sql_constraints = [
('name_uniq', 'UNIQUE(name)', 'Ethnicity name must be unique!'),
('code_uniq', 'UNIQUE(code)', 'Ethnicity code must be unique!'),
]
Add description to ethnicity model# -*- coding: utf-8 -*-
# #############################################################################
#
# Tech-Receptives Solutions Pvt. Ltd.
# Copyright (C) 2004-TODAY Tech-Receptives(<http://www.techreceptives.com>)
# Special Credit and Thanks to Thymbra Latinoamericana S.A.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
# #############################################################################
from openerp import models, fields
class MedicalPatientEthnicity(models.Model):
_name = 'medical.patient.ethnicity'
_description = 'Medical Patient Ethnicity'
notes = fields.Char()
code = fields.Char(required=True, )
name = fields.Char(required=True, translate=True)
_sql_constraints = [
('name_uniq', 'UNIQUE(name)', 'Ethnicity name must be unique!'),
('code_uniq', 'UNIQUE(code)', 'Ethnicity code must be unique!'),
]
|
<commit_before># -*- coding: utf-8 -*-
# #############################################################################
#
# Tech-Receptives Solutions Pvt. Ltd.
# Copyright (C) 2004-TODAY Tech-Receptives(<http://www.techreceptives.com>)
# Special Credit and Thanks to Thymbra Latinoamericana S.A.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
# #############################################################################
from openerp import models, fields
class MedicalPatientEthnicity(models.Model):
_name = 'medical.patient.ethnicity'
notes = fields.Char()
code = fields.Char(required=True, )
name = fields.Char(required=True, translate=True)
_sql_constraints = [
('name_uniq', 'UNIQUE(name)', 'Ethnicity name must be unique!'),
('code_uniq', 'UNIQUE(code)', 'Ethnicity code must be unique!'),
]
<commit_msg>Add description to ethnicity model<commit_after># -*- coding: utf-8 -*-
# #############################################################################
#
# Tech-Receptives Solutions Pvt. Ltd.
# Copyright (C) 2004-TODAY Tech-Receptives(<http://www.techreceptives.com>)
# Special Credit and Thanks to Thymbra Latinoamericana S.A.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
# #############################################################################
from openerp import models, fields
class MedicalPatientEthnicity(models.Model):
_name = 'medical.patient.ethnicity'
_description = 'Medical Patient Ethnicity'
notes = fields.Char()
code = fields.Char(required=True, )
name = fields.Char(required=True, translate=True)
_sql_constraints = [
('name_uniq', 'UNIQUE(name)', 'Ethnicity name must be unique!'),
('code_uniq', 'UNIQUE(code)', 'Ethnicity code must be unique!'),
]
|
10ddda3e230aa72889c81cd69792122b265010fe
|
rental/views/rental_state_view.py
|
rental/views/rental_state_view.py
|
from django.http import HttpResponseForbidden
from django.shortcuts import redirect, get_object_or_404
from django.views import View
from rental.state_transitions import allowed_transitions
from rental.models import Rental
class RentalStateView(View):
"""
Change the state of a given rental
If given an invalid state, this shows a 403 Forbidden response.
:author: Florian Stamer
"""
def post(self, request, rental_uuid):
rental = get_object_or_404(Rental, pk=rental_uuid)
managed_by_user = rental.depot.managed_by(request.user)
data = request.POST
state = data.get('state')
old_state = data.get('old_state')
# message = data.get('message')
if old_state != rental.state:
return HttpResponseForbidden('The state of the rental request has changed')
if state not in allowed_transitions(managed_by_user, rental.state):
return HttpResponseForbidden('Invalid state transition')
rental.state = state
rental.save()
return redirect('rental:detail', rental_uuid=rental.uuid)
|
from django.http import HttpResponseForbidden
from django.shortcuts import redirect, get_object_or_404
from django.views import View
from rental.availability import Availability
from rental.state_transitions import allowed_transitions
from rental.models import Rental
class RentalStateView(View):
"""
Change the state of a given rental
If given an invalid state, this shows a 403 Forbidden response.
:author: Florian Stamer
"""
def check_availability(self, rental):
availability = Availability(rental.start_date, rental.return_date, rental.depot_id)
for item_rental in rental.itemrental_set:
intervals = availability.get_availability_intervals(item_rental.item)
available = availability.get_minimum_availability(intervals)
if item_rental.quantity > available:
raise ValidationError({
'quantity': 'The quantity must not exceed the availability '
'of the item in the requested time frame.'
})
def post(self, request, rental_uuid):
rental = get_object_or_404(Rental, pk=rental_uuid)
managed_by_user = rental.depot.managed_by(request.user)
data = request.POST
state = data.get('state')
old_state = data.get('old_state')
# message = data.get('message')
if old_state != rental.state:
return HttpResponseForbidden('The state of the rental request has changed')
if state not in allowed_transitions(managed_by_user, rental.state):
return HttpResponseForbidden('Invalid state transition')
if state == Rental.STATE_APPROVED:
self.check_availability(rental)
rental.state = state
rental.save()
return redirect('rental:detail', rental_uuid=rental.uuid)
|
Check availability when approving rental request
|
Check availability when approving rental request
|
Python
|
agpl-3.0
|
verleihtool/verleihtool,verleihtool/verleihtool,verleihtool/verleihtool,verleihtool/verleihtool
|
from django.http import HttpResponseForbidden
from django.shortcuts import redirect, get_object_or_404
from django.views import View
from rental.state_transitions import allowed_transitions
from rental.models import Rental
class RentalStateView(View):
"""
Change the state of a given rental
If given an invalid state, this shows a 403 Forbidden response.
:author: Florian Stamer
"""
def post(self, request, rental_uuid):
rental = get_object_or_404(Rental, pk=rental_uuid)
managed_by_user = rental.depot.managed_by(request.user)
data = request.POST
state = data.get('state')
old_state = data.get('old_state')
# message = data.get('message')
if old_state != rental.state:
return HttpResponseForbidden('The state of the rental request has changed')
if state not in allowed_transitions(managed_by_user, rental.state):
return HttpResponseForbidden('Invalid state transition')
rental.state = state
rental.save()
return redirect('rental:detail', rental_uuid=rental.uuid)
Check availability when approving rental request
|
from django.http import HttpResponseForbidden
from django.shortcuts import redirect, get_object_or_404
from django.views import View
from rental.availability import Availability
from rental.state_transitions import allowed_transitions
from rental.models import Rental
class RentalStateView(View):
"""
Change the state of a given rental
If given an invalid state, this shows a 403 Forbidden response.
:author: Florian Stamer
"""
def check_availability(self, rental):
availability = Availability(rental.start_date, rental.return_date, rental.depot_id)
for item_rental in rental.itemrental_set:
intervals = availability.get_availability_intervals(item_rental.item)
available = availability.get_minimum_availability(intervals)
if item_rental.quantity > available:
raise ValidationError({
'quantity': 'The quantity must not exceed the availability '
'of the item in the requested time frame.'
})
def post(self, request, rental_uuid):
rental = get_object_or_404(Rental, pk=rental_uuid)
managed_by_user = rental.depot.managed_by(request.user)
data = request.POST
state = data.get('state')
old_state = data.get('old_state')
# message = data.get('message')
if old_state != rental.state:
return HttpResponseForbidden('The state of the rental request has changed')
if state not in allowed_transitions(managed_by_user, rental.state):
return HttpResponseForbidden('Invalid state transition')
if state == Rental.STATE_APPROVED:
self.check_availability(rental)
rental.state = state
rental.save()
return redirect('rental:detail', rental_uuid=rental.uuid)
|
<commit_before>from django.http import HttpResponseForbidden
from django.shortcuts import redirect, get_object_or_404
from django.views import View
from rental.state_transitions import allowed_transitions
from rental.models import Rental
class RentalStateView(View):
"""
Change the state of a given rental
If given an invalid state, this shows a 403 Forbidden response.
:author: Florian Stamer
"""
def post(self, request, rental_uuid):
rental = get_object_or_404(Rental, pk=rental_uuid)
managed_by_user = rental.depot.managed_by(request.user)
data = request.POST
state = data.get('state')
old_state = data.get('old_state')
# message = data.get('message')
if old_state != rental.state:
return HttpResponseForbidden('The state of the rental request has changed')
if state not in allowed_transitions(managed_by_user, rental.state):
return HttpResponseForbidden('Invalid state transition')
rental.state = state
rental.save()
return redirect('rental:detail', rental_uuid=rental.uuid)
<commit_msg>Check availability when approving rental request<commit_after>
|
from django.http import HttpResponseForbidden
from django.shortcuts import redirect, get_object_or_404
from django.views import View
from rental.availability import Availability
from rental.state_transitions import allowed_transitions
from rental.models import Rental
class RentalStateView(View):
"""
Change the state of a given rental
If given an invalid state, this shows a 403 Forbidden response.
:author: Florian Stamer
"""
def check_availability(self, rental):
availability = Availability(rental.start_date, rental.return_date, rental.depot_id)
for item_rental in rental.itemrental_set:
intervals = availability.get_availability_intervals(item_rental.item)
available = availability.get_minimum_availability(intervals)
if item_rental.quantity > available:
raise ValidationError({
'quantity': 'The quantity must not exceed the availability '
'of the item in the requested time frame.'
})
def post(self, request, rental_uuid):
rental = get_object_or_404(Rental, pk=rental_uuid)
managed_by_user = rental.depot.managed_by(request.user)
data = request.POST
state = data.get('state')
old_state = data.get('old_state')
# message = data.get('message')
if old_state != rental.state:
return HttpResponseForbidden('The state of the rental request has changed')
if state not in allowed_transitions(managed_by_user, rental.state):
return HttpResponseForbidden('Invalid state transition')
if state == Rental.STATE_APPROVED:
self.check_availability(rental)
rental.state = state
rental.save()
return redirect('rental:detail', rental_uuid=rental.uuid)
|
from django.http import HttpResponseForbidden
from django.shortcuts import redirect, get_object_or_404
from django.views import View
from rental.state_transitions import allowed_transitions
from rental.models import Rental
class RentalStateView(View):
"""
Change the state of a given rental
If given an invalid state, this shows a 403 Forbidden response.
:author: Florian Stamer
"""
def post(self, request, rental_uuid):
rental = get_object_or_404(Rental, pk=rental_uuid)
managed_by_user = rental.depot.managed_by(request.user)
data = request.POST
state = data.get('state')
old_state = data.get('old_state')
# message = data.get('message')
if old_state != rental.state:
return HttpResponseForbidden('The state of the rental request has changed')
if state not in allowed_transitions(managed_by_user, rental.state):
return HttpResponseForbidden('Invalid state transition')
rental.state = state
rental.save()
return redirect('rental:detail', rental_uuid=rental.uuid)
Check availability when approving rental requestfrom django.http import HttpResponseForbidden
from django.shortcuts import redirect, get_object_or_404
from django.views import View
from rental.availability import Availability
from rental.state_transitions import allowed_transitions
from rental.models import Rental
class RentalStateView(View):
"""
Change the state of a given rental
If given an invalid state, this shows a 403 Forbidden response.
:author: Florian Stamer
"""
def check_availability(self, rental):
availability = Availability(rental.start_date, rental.return_date, rental.depot_id)
for item_rental in rental.itemrental_set:
intervals = availability.get_availability_intervals(item_rental.item)
available = availability.get_minimum_availability(intervals)
if item_rental.quantity > available:
raise ValidationError({
'quantity': 'The quantity must not exceed the availability '
'of the item in the requested time frame.'
})
def post(self, request, rental_uuid):
rental = get_object_or_404(Rental, pk=rental_uuid)
managed_by_user = rental.depot.managed_by(request.user)
data = request.POST
state = data.get('state')
old_state = data.get('old_state')
# message = data.get('message')
if old_state != rental.state:
return HttpResponseForbidden('The state of the rental request has changed')
if state not in allowed_transitions(managed_by_user, rental.state):
return HttpResponseForbidden('Invalid state transition')
if state == Rental.STATE_APPROVED:
self.check_availability(rental)
rental.state = state
rental.save()
return redirect('rental:detail', rental_uuid=rental.uuid)
|
<commit_before>from django.http import HttpResponseForbidden
from django.shortcuts import redirect, get_object_or_404
from django.views import View
from rental.state_transitions import allowed_transitions
from rental.models import Rental
class RentalStateView(View):
"""
Change the state of a given rental
If given an invalid state, this shows a 403 Forbidden response.
:author: Florian Stamer
"""
def post(self, request, rental_uuid):
rental = get_object_or_404(Rental, pk=rental_uuid)
managed_by_user = rental.depot.managed_by(request.user)
data = request.POST
state = data.get('state')
old_state = data.get('old_state')
# message = data.get('message')
if old_state != rental.state:
return HttpResponseForbidden('The state of the rental request has changed')
if state not in allowed_transitions(managed_by_user, rental.state):
return HttpResponseForbidden('Invalid state transition')
rental.state = state
rental.save()
return redirect('rental:detail', rental_uuid=rental.uuid)
<commit_msg>Check availability when approving rental request<commit_after>from django.http import HttpResponseForbidden
from django.shortcuts import redirect, get_object_or_404
from django.views import View
from rental.availability import Availability
from rental.state_transitions import allowed_transitions
from rental.models import Rental
class RentalStateView(View):
"""
Change the state of a given rental
If given an invalid state, this shows a 403 Forbidden response.
:author: Florian Stamer
"""
def check_availability(self, rental):
availability = Availability(rental.start_date, rental.return_date, rental.depot_id)
for item_rental in rental.itemrental_set:
intervals = availability.get_availability_intervals(item_rental.item)
available = availability.get_minimum_availability(intervals)
if item_rental.quantity > available:
raise ValidationError({
'quantity': 'The quantity must not exceed the availability '
'of the item in the requested time frame.'
})
def post(self, request, rental_uuid):
rental = get_object_or_404(Rental, pk=rental_uuid)
managed_by_user = rental.depot.managed_by(request.user)
data = request.POST
state = data.get('state')
old_state = data.get('old_state')
# message = data.get('message')
if old_state != rental.state:
return HttpResponseForbidden('The state of the rental request has changed')
if state not in allowed_transitions(managed_by_user, rental.state):
return HttpResponseForbidden('Invalid state transition')
if state == Rental.STATE_APPROVED:
self.check_availability(rental)
rental.state = state
rental.save()
return redirect('rental:detail', rental_uuid=rental.uuid)
|
dc5235afec231454594201a54039869da26db576
|
enactiveagents/model/perceptionhandler.py
|
enactiveagents/model/perceptionhandler.py
|
"""
Module that holds classes that represent an agent's perception handler.
"""
import abc
import world
import structure
class PerceptionHandler(object):
@abc.abstractmethod
def perceive(self, agent, world):
"""
Generates a percept given an agent and a world.
:param agent: The agent to generate the percept for.
:param world: The world to generate the percept for.
:return: The percept.
"""
raise NotImplementedError("Should be implemented by child")
class EmptyPerceptionHandler(PerceptionHandler):
def perceive(self, agent, world):
return ""
class BasicPerceptionHandler(PerceptionHandler):
def perceive(self, agent_, world_):
for delta in range(1, 10):
pos = world.Position(agent_.get_position())
pos.add(agent_.get_move_delta(delta))
entities = world_.get_entities_at(pos)
if len(entities) > 0:
if isinstance(entities[0], structure.Wall):
return "w%s" % delta
return ""
|
"""
Module that holds classes that represent an agent's perception handler.
"""
import abc
import world
import structure
class PerceptionHandler(object):
@abc.abstractmethod
def perceive(self, agent, world):
"""
Generates a percept given an agent and a world.
:param agent: The agent to generate the percept for.
:param world: The world to generate the percept for.
:return: The percept.
"""
raise NotImplementedError("Should be implemented by child")
class EmptyPerceptionHandler(PerceptionHandler):
def perceive(self, agent, world):
return ""
class BasicPerceptionHandler(PerceptionHandler):
def perceive(self, agent_, world_):
for delta in range(0, 10):
pos = world.Position(agent_.get_position())
pos.add(agent_.get_move_delta(delta))
entities = world_.get_entities_at(pos)
for entity in entities:
if entity == agent_:
continue
if isinstance(entity, structure.Wall):
return "w%s" % delta
elif isinstance(entity, structure.Block):
return "b%s" % delta
return ""
|
Add block structure to perception handler. Slightly change perception handler logic.
|
Add block structure to perception handler. Slightly change perception handler logic.
|
Python
|
mit
|
Beskhue/enactive-agents,Beskhue/enactive-agents,Beskhue/enactive-agents
|
"""
Module that holds classes that represent an agent's perception handler.
"""
import abc
import world
import structure
class PerceptionHandler(object):
@abc.abstractmethod
def perceive(self, agent, world):
"""
Generates a percept given an agent and a world.
:param agent: The agent to generate the percept for.
:param world: The world to generate the percept for.
:return: The percept.
"""
raise NotImplementedError("Should be implemented by child")
class EmptyPerceptionHandler(PerceptionHandler):
def perceive(self, agent, world):
return ""
class BasicPerceptionHandler(PerceptionHandler):
def perceive(self, agent_, world_):
for delta in range(1, 10):
pos = world.Position(agent_.get_position())
pos.add(agent_.get_move_delta(delta))
entities = world_.get_entities_at(pos)
if len(entities) > 0:
if isinstance(entities[0], structure.Wall):
return "w%s" % delta
return ""Add block structure to perception handler. Slightly change perception handler logic.
|
"""
Module that holds classes that represent an agent's perception handler.
"""
import abc
import world
import structure
class PerceptionHandler(object):
@abc.abstractmethod
def perceive(self, agent, world):
"""
Generates a percept given an agent and a world.
:param agent: The agent to generate the percept for.
:param world: The world to generate the percept for.
:return: The percept.
"""
raise NotImplementedError("Should be implemented by child")
class EmptyPerceptionHandler(PerceptionHandler):
def perceive(self, agent, world):
return ""
class BasicPerceptionHandler(PerceptionHandler):
def perceive(self, agent_, world_):
for delta in range(0, 10):
pos = world.Position(agent_.get_position())
pos.add(agent_.get_move_delta(delta))
entities = world_.get_entities_at(pos)
for entity in entities:
if entity == agent_:
continue
if isinstance(entity, structure.Wall):
return "w%s" % delta
elif isinstance(entity, structure.Block):
return "b%s" % delta
return ""
|
<commit_before>"""
Module that holds classes that represent an agent's perception handler.
"""
import abc
import world
import structure
class PerceptionHandler(object):
@abc.abstractmethod
def perceive(self, agent, world):
"""
Generates a percept given an agent and a world.
:param agent: The agent to generate the percept for.
:param world: The world to generate the percept for.
:return: The percept.
"""
raise NotImplementedError("Should be implemented by child")
class EmptyPerceptionHandler(PerceptionHandler):
def perceive(self, agent, world):
return ""
class BasicPerceptionHandler(PerceptionHandler):
def perceive(self, agent_, world_):
for delta in range(1, 10):
pos = world.Position(agent_.get_position())
pos.add(agent_.get_move_delta(delta))
entities = world_.get_entities_at(pos)
if len(entities) > 0:
if isinstance(entities[0], structure.Wall):
return "w%s" % delta
return ""<commit_msg>Add block structure to perception handler. Slightly change perception handler logic.<commit_after>
|
"""
Module that holds classes that represent an agent's perception handler.
"""
import abc
import world
import structure
class PerceptionHandler(object):
@abc.abstractmethod
def perceive(self, agent, world):
"""
Generates a percept given an agent and a world.
:param agent: The agent to generate the percept for.
:param world: The world to generate the percept for.
:return: The percept.
"""
raise NotImplementedError("Should be implemented by child")
class EmptyPerceptionHandler(PerceptionHandler):
def perceive(self, agent, world):
return ""
class BasicPerceptionHandler(PerceptionHandler):
def perceive(self, agent_, world_):
for delta in range(0, 10):
pos = world.Position(agent_.get_position())
pos.add(agent_.get_move_delta(delta))
entities = world_.get_entities_at(pos)
for entity in entities:
if entity == agent_:
continue
if isinstance(entity, structure.Wall):
return "w%s" % delta
elif isinstance(entity, structure.Block):
return "b%s" % delta
return ""
|
"""
Module that holds classes that represent an agent's perception handler.
"""
import abc
import world
import structure
class PerceptionHandler(object):
@abc.abstractmethod
def perceive(self, agent, world):
"""
Generates a percept given an agent and a world.
:param agent: The agent to generate the percept for.
:param world: The world to generate the percept for.
:return: The percept.
"""
raise NotImplementedError("Should be implemented by child")
class EmptyPerceptionHandler(PerceptionHandler):
def perceive(self, agent, world):
return ""
class BasicPerceptionHandler(PerceptionHandler):
def perceive(self, agent_, world_):
for delta in range(1, 10):
pos = world.Position(agent_.get_position())
pos.add(agent_.get_move_delta(delta))
entities = world_.get_entities_at(pos)
if len(entities) > 0:
if isinstance(entities[0], structure.Wall):
return "w%s" % delta
return ""Add block structure to perception handler. Slightly change perception handler logic."""
Module that holds classes that represent an agent's perception handler.
"""
import abc
import world
import structure
class PerceptionHandler(object):
@abc.abstractmethod
def perceive(self, agent, world):
"""
Generates a percept given an agent and a world.
:param agent: The agent to generate the percept for.
:param world: The world to generate the percept for.
:return: The percept.
"""
raise NotImplementedError("Should be implemented by child")
class EmptyPerceptionHandler(PerceptionHandler):
def perceive(self, agent, world):
return ""
class BasicPerceptionHandler(PerceptionHandler):
def perceive(self, agent_, world_):
for delta in range(0, 10):
pos = world.Position(agent_.get_position())
pos.add(agent_.get_move_delta(delta))
entities = world_.get_entities_at(pos)
for entity in entities:
if entity == agent_:
continue
if isinstance(entity, structure.Wall):
return "w%s" % delta
elif isinstance(entity, structure.Block):
return "b%s" % delta
return ""
|
<commit_before>"""
Module that holds classes that represent an agent's perception handler.
"""
import abc
import world
import structure
class PerceptionHandler(object):
@abc.abstractmethod
def perceive(self, agent, world):
"""
Generates a percept given an agent and a world.
:param agent: The agent to generate the percept for.
:param world: The world to generate the percept for.
:return: The percept.
"""
raise NotImplementedError("Should be implemented by child")
class EmptyPerceptionHandler(PerceptionHandler):
def perceive(self, agent, world):
return ""
class BasicPerceptionHandler(PerceptionHandler):
def perceive(self, agent_, world_):
for delta in range(1, 10):
pos = world.Position(agent_.get_position())
pos.add(agent_.get_move_delta(delta))
entities = world_.get_entities_at(pos)
if len(entities) > 0:
if isinstance(entities[0], structure.Wall):
return "w%s" % delta
return ""<commit_msg>Add block structure to perception handler. Slightly change perception handler logic.<commit_after>"""
Module that holds classes that represent an agent's perception handler.
"""
import abc
import world
import structure
class PerceptionHandler(object):
@abc.abstractmethod
def perceive(self, agent, world):
"""
Generates a percept given an agent and a world.
:param agent: The agent to generate the percept for.
:param world: The world to generate the percept for.
:return: The percept.
"""
raise NotImplementedError("Should be implemented by child")
class EmptyPerceptionHandler(PerceptionHandler):
def perceive(self, agent, world):
return ""
class BasicPerceptionHandler(PerceptionHandler):
def perceive(self, agent_, world_):
for delta in range(0, 10):
pos = world.Position(agent_.get_position())
pos.add(agent_.get_move_delta(delta))
entities = world_.get_entities_at(pos)
for entity in entities:
if entity == agent_:
continue
if isinstance(entity, structure.Wall):
return "w%s" % delta
elif isinstance(entity, structure.Block):
return "b%s" % delta
return ""
|
f6abaa5dba68bf010b847de0c6d37b87e5732eea
|
github2/commits.py
|
github2/commits.py
|
from github2.core import BaseData, GithubCommand, Attribute, DateAttribute
class Commit(BaseData):
message = Attribute("Commit message.")
parents = Attribute("List of parents for this commit.")
url = Attribute("Canonical URL for this commit.")
author = Attribute("Author metadata (dict with name/email.)")
id = Attribute("Commit ID.")
committed_date = DateAttribute("Date committed.", format="commit")
authored_data = DateAttribute("Date authored.", format="commit")
tree = Attribute("Tree SHA for this commit.")
committer = Attribute("Comitter metadata (dict with name/email.)")
added = Attribute("(If present) Datastructure representing what's been "
"added since last commit.")
removed = Attribute("(if present) Datastructure representing what's been "
"removed since last commit.")
modified = Attribute("(If present) Datastructure representing what's "
"been modified since last commit.")
class Commits(GithubCommand):
domain = "commits"
def list(self, project, branch="master", file=None):
return self.get_values("list", project, branch, file,
filter="commits", datatype=Commit)
def show(self, project, sha):
return self.get_value("show", project, sha,
filter="commit", datatype=Commit)
|
from github2.core import BaseData, GithubCommand, Attribute, DateAttribute
class Commit(BaseData):
message = Attribute("Commit message.")
parents = Attribute("List of parents for this commit.")
url = Attribute("Canonical URL for this commit.")
author = Attribute("Author metadata (dict with name/email.)")
id = Attribute("Commit ID.")
committed_date = DateAttribute("Date committed.", format="commit")
authored_date = DateAttribute("Date authored.", format="commit")
tree = Attribute("Tree SHA for this commit.")
committer = Attribute("Comitter metadata (dict with name/email.)")
added = Attribute("(If present) Datastructure representing what's been "
"added since last commit.")
removed = Attribute("(if present) Datastructure representing what's been "
"removed since last commit.")
modified = Attribute("(If present) Datastructure representing what's "
"been modified since last commit.")
class Commits(GithubCommand):
domain = "commits"
def list(self, project, branch="master", file=None):
return self.get_values("list", project, branch, file,
filter="commits", datatype=Commit)
def show(self, project, sha):
return self.get_value("show", project, sha,
filter="commit", datatype=Commit)
|
Fix typo authored_data -> authored_date
|
Fix typo authored_data -> authored_date
|
Python
|
bsd-3-clause
|
ask/python-github2
|
from github2.core import BaseData, GithubCommand, Attribute, DateAttribute
class Commit(BaseData):
message = Attribute("Commit message.")
parents = Attribute("List of parents for this commit.")
url = Attribute("Canonical URL for this commit.")
author = Attribute("Author metadata (dict with name/email.)")
id = Attribute("Commit ID.")
committed_date = DateAttribute("Date committed.", format="commit")
authored_data = DateAttribute("Date authored.", format="commit")
tree = Attribute("Tree SHA for this commit.")
committer = Attribute("Comitter metadata (dict with name/email.)")
added = Attribute("(If present) Datastructure representing what's been "
"added since last commit.")
removed = Attribute("(if present) Datastructure representing what's been "
"removed since last commit.")
modified = Attribute("(If present) Datastructure representing what's "
"been modified since last commit.")
class Commits(GithubCommand):
domain = "commits"
def list(self, project, branch="master", file=None):
return self.get_values("list", project, branch, file,
filter="commits", datatype=Commit)
def show(self, project, sha):
return self.get_value("show", project, sha,
filter="commit", datatype=Commit)
Fix typo authored_data -> authored_date
|
from github2.core import BaseData, GithubCommand, Attribute, DateAttribute
class Commit(BaseData):
message = Attribute("Commit message.")
parents = Attribute("List of parents for this commit.")
url = Attribute("Canonical URL for this commit.")
author = Attribute("Author metadata (dict with name/email.)")
id = Attribute("Commit ID.")
committed_date = DateAttribute("Date committed.", format="commit")
authored_date = DateAttribute("Date authored.", format="commit")
tree = Attribute("Tree SHA for this commit.")
committer = Attribute("Comitter metadata (dict with name/email.)")
added = Attribute("(If present) Datastructure representing what's been "
"added since last commit.")
removed = Attribute("(if present) Datastructure representing what's been "
"removed since last commit.")
modified = Attribute("(If present) Datastructure representing what's "
"been modified since last commit.")
class Commits(GithubCommand):
domain = "commits"
def list(self, project, branch="master", file=None):
return self.get_values("list", project, branch, file,
filter="commits", datatype=Commit)
def show(self, project, sha):
return self.get_value("show", project, sha,
filter="commit", datatype=Commit)
|
<commit_before>from github2.core import BaseData, GithubCommand, Attribute, DateAttribute
class Commit(BaseData):
message = Attribute("Commit message.")
parents = Attribute("List of parents for this commit.")
url = Attribute("Canonical URL for this commit.")
author = Attribute("Author metadata (dict with name/email.)")
id = Attribute("Commit ID.")
committed_date = DateAttribute("Date committed.", format="commit")
authored_data = DateAttribute("Date authored.", format="commit")
tree = Attribute("Tree SHA for this commit.")
committer = Attribute("Comitter metadata (dict with name/email.)")
added = Attribute("(If present) Datastructure representing what's been "
"added since last commit.")
removed = Attribute("(if present) Datastructure representing what's been "
"removed since last commit.")
modified = Attribute("(If present) Datastructure representing what's "
"been modified since last commit.")
class Commits(GithubCommand):
domain = "commits"
def list(self, project, branch="master", file=None):
return self.get_values("list", project, branch, file,
filter="commits", datatype=Commit)
def show(self, project, sha):
return self.get_value("show", project, sha,
filter="commit", datatype=Commit)
<commit_msg>Fix typo authored_data -> authored_date<commit_after>
|
from github2.core import BaseData, GithubCommand, Attribute, DateAttribute
class Commit(BaseData):
message = Attribute("Commit message.")
parents = Attribute("List of parents for this commit.")
url = Attribute("Canonical URL for this commit.")
author = Attribute("Author metadata (dict with name/email.)")
id = Attribute("Commit ID.")
committed_date = DateAttribute("Date committed.", format="commit")
authored_date = DateAttribute("Date authored.", format="commit")
tree = Attribute("Tree SHA for this commit.")
committer = Attribute("Comitter metadata (dict with name/email.)")
added = Attribute("(If present) Datastructure representing what's been "
"added since last commit.")
removed = Attribute("(if present) Datastructure representing what's been "
"removed since last commit.")
modified = Attribute("(If present) Datastructure representing what's "
"been modified since last commit.")
class Commits(GithubCommand):
domain = "commits"
def list(self, project, branch="master", file=None):
return self.get_values("list", project, branch, file,
filter="commits", datatype=Commit)
def show(self, project, sha):
return self.get_value("show", project, sha,
filter="commit", datatype=Commit)
|
from github2.core import BaseData, GithubCommand, Attribute, DateAttribute
class Commit(BaseData):
message = Attribute("Commit message.")
parents = Attribute("List of parents for this commit.")
url = Attribute("Canonical URL for this commit.")
author = Attribute("Author metadata (dict with name/email.)")
id = Attribute("Commit ID.")
committed_date = DateAttribute("Date committed.", format="commit")
authored_data = DateAttribute("Date authored.", format="commit")
tree = Attribute("Tree SHA for this commit.")
committer = Attribute("Comitter metadata (dict with name/email.)")
added = Attribute("(If present) Datastructure representing what's been "
"added since last commit.")
removed = Attribute("(if present) Datastructure representing what's been "
"removed since last commit.")
modified = Attribute("(If present) Datastructure representing what's "
"been modified since last commit.")
class Commits(GithubCommand):
domain = "commits"
def list(self, project, branch="master", file=None):
return self.get_values("list", project, branch, file,
filter="commits", datatype=Commit)
def show(self, project, sha):
return self.get_value("show", project, sha,
filter="commit", datatype=Commit)
Fix typo authored_data -> authored_datefrom github2.core import BaseData, GithubCommand, Attribute, DateAttribute
class Commit(BaseData):
message = Attribute("Commit message.")
parents = Attribute("List of parents for this commit.")
url = Attribute("Canonical URL for this commit.")
author = Attribute("Author metadata (dict with name/email.)")
id = Attribute("Commit ID.")
committed_date = DateAttribute("Date committed.", format="commit")
authored_date = DateAttribute("Date authored.", format="commit")
tree = Attribute("Tree SHA for this commit.")
committer = Attribute("Comitter metadata (dict with name/email.)")
added = Attribute("(If present) Datastructure representing what's been "
"added since last commit.")
removed = Attribute("(if present) Datastructure representing what's been "
"removed since last commit.")
modified = Attribute("(If present) Datastructure representing what's "
"been modified since last commit.")
class Commits(GithubCommand):
domain = "commits"
def list(self, project, branch="master", file=None):
return self.get_values("list", project, branch, file,
filter="commits", datatype=Commit)
def show(self, project, sha):
return self.get_value("show", project, sha,
filter="commit", datatype=Commit)
|
<commit_before>from github2.core import BaseData, GithubCommand, Attribute, DateAttribute
class Commit(BaseData):
message = Attribute("Commit message.")
parents = Attribute("List of parents for this commit.")
url = Attribute("Canonical URL for this commit.")
author = Attribute("Author metadata (dict with name/email.)")
id = Attribute("Commit ID.")
committed_date = DateAttribute("Date committed.", format="commit")
authored_data = DateAttribute("Date authored.", format="commit")
tree = Attribute("Tree SHA for this commit.")
committer = Attribute("Comitter metadata (dict with name/email.)")
added = Attribute("(If present) Datastructure representing what's been "
"added since last commit.")
removed = Attribute("(if present) Datastructure representing what's been "
"removed since last commit.")
modified = Attribute("(If present) Datastructure representing what's "
"been modified since last commit.")
class Commits(GithubCommand):
domain = "commits"
def list(self, project, branch="master", file=None):
return self.get_values("list", project, branch, file,
filter="commits", datatype=Commit)
def show(self, project, sha):
return self.get_value("show", project, sha,
filter="commit", datatype=Commit)
<commit_msg>Fix typo authored_data -> authored_date<commit_after>from github2.core import BaseData, GithubCommand, Attribute, DateAttribute
class Commit(BaseData):
message = Attribute("Commit message.")
parents = Attribute("List of parents for this commit.")
url = Attribute("Canonical URL for this commit.")
author = Attribute("Author metadata (dict with name/email.)")
id = Attribute("Commit ID.")
committed_date = DateAttribute("Date committed.", format="commit")
authored_date = DateAttribute("Date authored.", format="commit")
tree = Attribute("Tree SHA for this commit.")
committer = Attribute("Comitter metadata (dict with name/email.)")
added = Attribute("(If present) Datastructure representing what's been "
"added since last commit.")
removed = Attribute("(if present) Datastructure representing what's been "
"removed since last commit.")
modified = Attribute("(If present) Datastructure representing what's "
"been modified since last commit.")
class Commits(GithubCommand):
domain = "commits"
def list(self, project, branch="master", file=None):
return self.get_values("list", project, branch, file,
filter="commits", datatype=Commit)
def show(self, project, sha):
return self.get_value("show", project, sha,
filter="commit", datatype=Commit)
|
b5a21c39c37c02ea7077ce92596d68e496473af0
|
grako/rendering.py
|
grako/rendering.py
|
# -*- coding: utf-8 -*-
from __future__ import print_function, division, absolute_import, unicode_literals
import itertools
from .util import trim
def render(item, **fields):
""" Render the given item
"""
if item is None:
return ''
elif isinstance(item, Renderer):
return item.render(**fields)
elif isinstance(item, list):
return ''.join(render(e) for e in item)
else:
return str(item)
class Renderer(object):
template = ''
_counter = itertools.count()
def __init__(self, template=None):
if template is not None:
self.template = template
def counter(self):
return next(self._counter)
def render_fields(self, fields):
pass
def render(self, template=None, **fields):
fields.update({k:v for k, v in vars(self).items() if not k.startswith('_')})
self.render_fields(fields)
if template is None:
template = self.template
fields = {k:render(v) for k, v in fields.items()}
try:
return trim(template).format(**fields)
except KeyError as e:
raise KeyError(str(e), type(self))
|
# -*- coding: utf-8 -*-
from __future__ import print_function, division, absolute_import, unicode_literals
import itertools
from .util import trim
def render(item, **fields):
""" Render the given item
"""
if item is None:
return ''
elif isinstance(item, Renderer):
return item.render(**fields)
elif isinstance(item, list):
return ''.join(render(e) for e in item)
else:
return str(item)
class Renderer(object):
template = ''
_counter = itertools.count()
def __init__(self, template=None):
if template is not None:
self.template = template
def counter(self):
return next(self._counter)
def render_fields(self, fields):
pass
def render(self, template=None, **kwargs):
fields = ({k:v for k, v in vars(self).items() if not k.startswith('_')})
override = self.render_fields(fields)
if template is None:
if override is not None:
template = override
else:
template = self.template
fields.update(kwargs)
fields = {k:render(v) for k, v in fields.items()}
try:
return trim(template).format(**fields)
except KeyError as e:
raise KeyError(str(e), type(self))
|
Allow override of template through return value of render_fields.
|
Allow override of template through return value of render_fields.
|
Python
|
bsd-2-clause
|
frnknglrt/grako,vmuriart/grako
|
# -*- coding: utf-8 -*-
from __future__ import print_function, division, absolute_import, unicode_literals
import itertools
from .util import trim
def render(item, **fields):
""" Render the given item
"""
if item is None:
return ''
elif isinstance(item, Renderer):
return item.render(**fields)
elif isinstance(item, list):
return ''.join(render(e) for e in item)
else:
return str(item)
class Renderer(object):
template = ''
_counter = itertools.count()
def __init__(self, template=None):
if template is not None:
self.template = template
def counter(self):
return next(self._counter)
def render_fields(self, fields):
pass
def render(self, template=None, **fields):
fields.update({k:v for k, v in vars(self).items() if not k.startswith('_')})
self.render_fields(fields)
if template is None:
template = self.template
fields = {k:render(v) for k, v in fields.items()}
try:
return trim(template).format(**fields)
except KeyError as e:
raise KeyError(str(e), type(self))
Allow override of template through return value of render_fields.
|
# -*- coding: utf-8 -*-
from __future__ import print_function, division, absolute_import, unicode_literals
import itertools
from .util import trim
def render(item, **fields):
""" Render the given item
"""
if item is None:
return ''
elif isinstance(item, Renderer):
return item.render(**fields)
elif isinstance(item, list):
return ''.join(render(e) for e in item)
else:
return str(item)
class Renderer(object):
template = ''
_counter = itertools.count()
def __init__(self, template=None):
if template is not None:
self.template = template
def counter(self):
return next(self._counter)
def render_fields(self, fields):
pass
def render(self, template=None, **kwargs):
fields = ({k:v for k, v in vars(self).items() if not k.startswith('_')})
override = self.render_fields(fields)
if template is None:
if override is not None:
template = override
else:
template = self.template
fields.update(kwargs)
fields = {k:render(v) for k, v in fields.items()}
try:
return trim(template).format(**fields)
except KeyError as e:
raise KeyError(str(e), type(self))
|
<commit_before># -*- coding: utf-8 -*-
from __future__ import print_function, division, absolute_import, unicode_literals
import itertools
from .util import trim
def render(item, **fields):
""" Render the given item
"""
if item is None:
return ''
elif isinstance(item, Renderer):
return item.render(**fields)
elif isinstance(item, list):
return ''.join(render(e) for e in item)
else:
return str(item)
class Renderer(object):
template = ''
_counter = itertools.count()
def __init__(self, template=None):
if template is not None:
self.template = template
def counter(self):
return next(self._counter)
def render_fields(self, fields):
pass
def render(self, template=None, **fields):
fields.update({k:v for k, v in vars(self).items() if not k.startswith('_')})
self.render_fields(fields)
if template is None:
template = self.template
fields = {k:render(v) for k, v in fields.items()}
try:
return trim(template).format(**fields)
except KeyError as e:
raise KeyError(str(e), type(self))
<commit_msg>Allow override of template through return value of render_fields.<commit_after>
|
# -*- coding: utf-8 -*-
from __future__ import print_function, division, absolute_import, unicode_literals
import itertools
from .util import trim
def render(item, **fields):
""" Render the given item
"""
if item is None:
return ''
elif isinstance(item, Renderer):
return item.render(**fields)
elif isinstance(item, list):
return ''.join(render(e) for e in item)
else:
return str(item)
class Renderer(object):
template = ''
_counter = itertools.count()
def __init__(self, template=None):
if template is not None:
self.template = template
def counter(self):
return next(self._counter)
def render_fields(self, fields):
pass
def render(self, template=None, **kwargs):
fields = ({k:v for k, v in vars(self).items() if not k.startswith('_')})
override = self.render_fields(fields)
if template is None:
if override is not None:
template = override
else:
template = self.template
fields.update(kwargs)
fields = {k:render(v) for k, v in fields.items()}
try:
return trim(template).format(**fields)
except KeyError as e:
raise KeyError(str(e), type(self))
|
# -*- coding: utf-8 -*-
from __future__ import print_function, division, absolute_import, unicode_literals
import itertools
from .util import trim
def render(item, **fields):
""" Render the given item
"""
if item is None:
return ''
elif isinstance(item, Renderer):
return item.render(**fields)
elif isinstance(item, list):
return ''.join(render(e) for e in item)
else:
return str(item)
class Renderer(object):
template = ''
_counter = itertools.count()
def __init__(self, template=None):
if template is not None:
self.template = template
def counter(self):
return next(self._counter)
def render_fields(self, fields):
pass
def render(self, template=None, **fields):
fields.update({k:v for k, v in vars(self).items() if not k.startswith('_')})
self.render_fields(fields)
if template is None:
template = self.template
fields = {k:render(v) for k, v in fields.items()}
try:
return trim(template).format(**fields)
except KeyError as e:
raise KeyError(str(e), type(self))
Allow override of template through return value of render_fields.# -*- coding: utf-8 -*-
from __future__ import print_function, division, absolute_import, unicode_literals
import itertools
from .util import trim
def render(item, **fields):
""" Render the given item
"""
if item is None:
return ''
elif isinstance(item, Renderer):
return item.render(**fields)
elif isinstance(item, list):
return ''.join(render(e) for e in item)
else:
return str(item)
class Renderer(object):
template = ''
_counter = itertools.count()
def __init__(self, template=None):
if template is not None:
self.template = template
def counter(self):
return next(self._counter)
def render_fields(self, fields):
pass
def render(self, template=None, **kwargs):
fields = ({k:v for k, v in vars(self).items() if not k.startswith('_')})
override = self.render_fields(fields)
if template is None:
if override is not None:
template = override
else:
template = self.template
fields.update(kwargs)
fields = {k:render(v) for k, v in fields.items()}
try:
return trim(template).format(**fields)
except KeyError as e:
raise KeyError(str(e), type(self))
|
<commit_before># -*- coding: utf-8 -*-
from __future__ import print_function, division, absolute_import, unicode_literals
import itertools
from .util import trim
def render(item, **fields):
""" Render the given item
"""
if item is None:
return ''
elif isinstance(item, Renderer):
return item.render(**fields)
elif isinstance(item, list):
return ''.join(render(e) for e in item)
else:
return str(item)
class Renderer(object):
template = ''
_counter = itertools.count()
def __init__(self, template=None):
if template is not None:
self.template = template
def counter(self):
return next(self._counter)
def render_fields(self, fields):
pass
def render(self, template=None, **fields):
fields.update({k:v for k, v in vars(self).items() if not k.startswith('_')})
self.render_fields(fields)
if template is None:
template = self.template
fields = {k:render(v) for k, v in fields.items()}
try:
return trim(template).format(**fields)
except KeyError as e:
raise KeyError(str(e), type(self))
<commit_msg>Allow override of template through return value of render_fields.<commit_after># -*- coding: utf-8 -*-
from __future__ import print_function, division, absolute_import, unicode_literals
import itertools
from .util import trim
def render(item, **fields):
""" Render the given item
"""
if item is None:
return ''
elif isinstance(item, Renderer):
return item.render(**fields)
elif isinstance(item, list):
return ''.join(render(e) for e in item)
else:
return str(item)
class Renderer(object):
template = ''
_counter = itertools.count()
def __init__(self, template=None):
if template is not None:
self.template = template
def counter(self):
return next(self._counter)
def render_fields(self, fields):
pass
def render(self, template=None, **kwargs):
fields = ({k:v for k, v in vars(self).items() if not k.startswith('_')})
override = self.render_fields(fields)
if template is None:
if override is not None:
template = override
else:
template = self.template
fields.update(kwargs)
fields = {k:render(v) for k, v in fields.items()}
try:
return trim(template).format(**fields)
except KeyError as e:
raise KeyError(str(e), type(self))
|
17a42110978d2fc38daaf0e09e25da760ccdc339
|
adhocracy4/emails/mixins.py
|
adhocracy4/emails/mixins.py
|
from email.mime.image import MIMEImage
from django.contrib.staticfiles import finders
from .base import EmailBase
class PlatformEmailMixin:
"""
Attaches the static file images/logo.png so it can be used in an html
email.
"""
def get_attachments(self):
attachments = super().get_attachments()
filename = finders.find('images/email_logo.png')
if filename:
f = open(filename, 'rb')
logo = MIMEImage(f.read())
logo.add_header('Content-ID', '<{}>'.format('logo'))
return attachments + [logo]
return attachments
class SyncEmailMixin(EmailBase):
"""Send Emails synchronously."""
@classmethod
def send(cls, object, *args, **kwargs):
"""Call dispatch immediately"""
return cls().dispatch(object, *args, **kwargs)
|
from email.mime.image import MIMEImage
from django.contrib.staticfiles import finders
from .base import EmailBase
class PlatformEmailMixin:
"""
Attaches the static file images/logo.png so it can be used in an html
email.
"""
def get_attachments(self):
attachments = super().get_attachments()
filename = (
finders.find('images/email_logo.png')
or finders.find('images/email_logo.svg')
)
if filename:
f = open(filename, 'rb')
logo = MIMEImage(f.read())
logo.add_header('Content-ID', '<{}>'.format('logo'))
return attachments + [logo]
return attachments
class SyncEmailMixin(EmailBase):
"""Send Emails synchronously."""
@classmethod
def send(cls, object, *args, **kwargs):
"""Call dispatch immediately"""
return cls().dispatch(object, *args, **kwargs)
|
Allow svg as email logo attachment
|
Allow svg as email logo attachment
|
Python
|
agpl-3.0
|
liqd/adhocracy4,liqd/adhocracy4,liqd/adhocracy4,liqd/adhocracy4
|
from email.mime.image import MIMEImage
from django.contrib.staticfiles import finders
from .base import EmailBase
class PlatformEmailMixin:
"""
Attaches the static file images/logo.png so it can be used in an html
email.
"""
def get_attachments(self):
attachments = super().get_attachments()
filename = finders.find('images/email_logo.png')
if filename:
f = open(filename, 'rb')
logo = MIMEImage(f.read())
logo.add_header('Content-ID', '<{}>'.format('logo'))
return attachments + [logo]
return attachments
class SyncEmailMixin(EmailBase):
"""Send Emails synchronously."""
@classmethod
def send(cls, object, *args, **kwargs):
"""Call dispatch immediately"""
return cls().dispatch(object, *args, **kwargs)
Allow svg as email logo attachment
|
from email.mime.image import MIMEImage
from django.contrib.staticfiles import finders
from .base import EmailBase
class PlatformEmailMixin:
"""
Attaches the static file images/logo.png so it can be used in an html
email.
"""
def get_attachments(self):
attachments = super().get_attachments()
filename = (
finders.find('images/email_logo.png')
or finders.find('images/email_logo.svg')
)
if filename:
f = open(filename, 'rb')
logo = MIMEImage(f.read())
logo.add_header('Content-ID', '<{}>'.format('logo'))
return attachments + [logo]
return attachments
class SyncEmailMixin(EmailBase):
"""Send Emails synchronously."""
@classmethod
def send(cls, object, *args, **kwargs):
"""Call dispatch immediately"""
return cls().dispatch(object, *args, **kwargs)
|
<commit_before>from email.mime.image import MIMEImage
from django.contrib.staticfiles import finders
from .base import EmailBase
class PlatformEmailMixin:
"""
Attaches the static file images/logo.png so it can be used in an html
email.
"""
def get_attachments(self):
attachments = super().get_attachments()
filename = finders.find('images/email_logo.png')
if filename:
f = open(filename, 'rb')
logo = MIMEImage(f.read())
logo.add_header('Content-ID', '<{}>'.format('logo'))
return attachments + [logo]
return attachments
class SyncEmailMixin(EmailBase):
"""Send Emails synchronously."""
@classmethod
def send(cls, object, *args, **kwargs):
"""Call dispatch immediately"""
return cls().dispatch(object, *args, **kwargs)
<commit_msg>Allow svg as email logo attachment<commit_after>
|
from email.mime.image import MIMEImage
from django.contrib.staticfiles import finders
from .base import EmailBase
class PlatformEmailMixin:
"""
Attaches the static file images/logo.png so it can be used in an html
email.
"""
def get_attachments(self):
attachments = super().get_attachments()
filename = (
finders.find('images/email_logo.png')
or finders.find('images/email_logo.svg')
)
if filename:
f = open(filename, 'rb')
logo = MIMEImage(f.read())
logo.add_header('Content-ID', '<{}>'.format('logo'))
return attachments + [logo]
return attachments
class SyncEmailMixin(EmailBase):
"""Send Emails synchronously."""
@classmethod
def send(cls, object, *args, **kwargs):
"""Call dispatch immediately"""
return cls().dispatch(object, *args, **kwargs)
|
from email.mime.image import MIMEImage
from django.contrib.staticfiles import finders
from .base import EmailBase
class PlatformEmailMixin:
"""
Attaches the static file images/logo.png so it can be used in an html
email.
"""
def get_attachments(self):
attachments = super().get_attachments()
filename = finders.find('images/email_logo.png')
if filename:
f = open(filename, 'rb')
logo = MIMEImage(f.read())
logo.add_header('Content-ID', '<{}>'.format('logo'))
return attachments + [logo]
return attachments
class SyncEmailMixin(EmailBase):
"""Send Emails synchronously."""
@classmethod
def send(cls, object, *args, **kwargs):
"""Call dispatch immediately"""
return cls().dispatch(object, *args, **kwargs)
Allow svg as email logo attachmentfrom email.mime.image import MIMEImage
from django.contrib.staticfiles import finders
from .base import EmailBase
class PlatformEmailMixin:
"""
Attaches the static file images/logo.png so it can be used in an html
email.
"""
def get_attachments(self):
attachments = super().get_attachments()
filename = (
finders.find('images/email_logo.png')
or finders.find('images/email_logo.svg')
)
if filename:
f = open(filename, 'rb')
logo = MIMEImage(f.read())
logo.add_header('Content-ID', '<{}>'.format('logo'))
return attachments + [logo]
return attachments
class SyncEmailMixin(EmailBase):
"""Send Emails synchronously."""
@classmethod
def send(cls, object, *args, **kwargs):
"""Call dispatch immediately"""
return cls().dispatch(object, *args, **kwargs)
|
<commit_before>from email.mime.image import MIMEImage
from django.contrib.staticfiles import finders
from .base import EmailBase
class PlatformEmailMixin:
"""
Attaches the static file images/logo.png so it can be used in an html
email.
"""
def get_attachments(self):
attachments = super().get_attachments()
filename = finders.find('images/email_logo.png')
if filename:
f = open(filename, 'rb')
logo = MIMEImage(f.read())
logo.add_header('Content-ID', '<{}>'.format('logo'))
return attachments + [logo]
return attachments
class SyncEmailMixin(EmailBase):
"""Send Emails synchronously."""
@classmethod
def send(cls, object, *args, **kwargs):
"""Call dispatch immediately"""
return cls().dispatch(object, *args, **kwargs)
<commit_msg>Allow svg as email logo attachment<commit_after>from email.mime.image import MIMEImage
from django.contrib.staticfiles import finders
from .base import EmailBase
class PlatformEmailMixin:
"""
Attaches the static file images/logo.png so it can be used in an html
email.
"""
def get_attachments(self):
attachments = super().get_attachments()
filename = (
finders.find('images/email_logo.png')
or finders.find('images/email_logo.svg')
)
if filename:
f = open(filename, 'rb')
logo = MIMEImage(f.read())
logo.add_header('Content-ID', '<{}>'.format('logo'))
return attachments + [logo]
return attachments
class SyncEmailMixin(EmailBase):
"""Send Emails synchronously."""
@classmethod
def send(cls, object, *args, **kwargs):
"""Call dispatch immediately"""
return cls().dispatch(object, *args, **kwargs)
|
9e7c7c17f17553d010bf61b4d8014b5f0c064aca
|
examples/freesolv/generate_experiments.py
|
examples/freesolv/generate_experiments.py
|
import yaml
import copy
vacuum_switching_lengths = [0, 100, 500, 1000, 5000, 10000]
solvent_switching_lengths = [500, 1000, 5000, 10000, 20000, 50000]
use_sterics = [True, False]
geometry_divisions = [90, 180, 360, 720]
# Load in template yaml:
with open("rj_hydration.yaml", "r") as templatefile:
template_yaml = yaml.load(templatefile)
# Set up vacuum simulations:
for phase in ['vacuum', 'explicit']:
if phase == "vacuum":
switching_lengths = vacuum_switching_lengths
else:
switching_lengths = solvent_switching_lengths
for switching_length in switching_lengths:
for sterics in use_sterics:
for geometry_division in geometry_divisions:
yaml_dict = copy.deepcopy(template_yaml)
specification_file_prefix = "{}_{}ncmc_{}sterics_{}geometry".format(phase, switching_length, sterics, geometry_division)
specification_filename = specification_file_prefix + ".yaml"
yaml_dict['geometry_divisions'][phase] = geometry_division
yaml_dict['use_sterics']['vacuum'] = sterics
yaml_dict['ncmc_switching_times'][phase] = switching_length
yaml_dict['phase'] = phase
yaml_dict['output_filename'] = specification_file_prefix + ".nc"
with open(specification_filename, 'w') as yam_outfile:
yaml.dump(yaml_dict, yam_outfile)
|
Add script to generate the yaml files for experiments
|
Add script to generate the yaml files for experiments
|
Python
|
mit
|
choderalab/perses,choderalab/perses
|
Add script to generate the yaml files for experiments
|
import yaml
import copy
vacuum_switching_lengths = [0, 100, 500, 1000, 5000, 10000]
solvent_switching_lengths = [500, 1000, 5000, 10000, 20000, 50000]
use_sterics = [True, False]
geometry_divisions = [90, 180, 360, 720]
# Load in template yaml:
with open("rj_hydration.yaml", "r") as templatefile:
template_yaml = yaml.load(templatefile)
# Set up vacuum simulations:
for phase in ['vacuum', 'explicit']:
if phase == "vacuum":
switching_lengths = vacuum_switching_lengths
else:
switching_lengths = solvent_switching_lengths
for switching_length in switching_lengths:
for sterics in use_sterics:
for geometry_division in geometry_divisions:
yaml_dict = copy.deepcopy(template_yaml)
specification_file_prefix = "{}_{}ncmc_{}sterics_{}geometry".format(phase, switching_length, sterics, geometry_division)
specification_filename = specification_file_prefix + ".yaml"
yaml_dict['geometry_divisions'][phase] = geometry_division
yaml_dict['use_sterics']['vacuum'] = sterics
yaml_dict['ncmc_switching_times'][phase] = switching_length
yaml_dict['phase'] = phase
yaml_dict['output_filename'] = specification_file_prefix + ".nc"
with open(specification_filename, 'w') as yam_outfile:
yaml.dump(yaml_dict, yam_outfile)
|
<commit_before><commit_msg>Add script to generate the yaml files for experiments<commit_after>
|
import yaml
import copy
vacuum_switching_lengths = [0, 100, 500, 1000, 5000, 10000]
solvent_switching_lengths = [500, 1000, 5000, 10000, 20000, 50000]
use_sterics = [True, False]
geometry_divisions = [90, 180, 360, 720]
# Load in template yaml:
with open("rj_hydration.yaml", "r") as templatefile:
template_yaml = yaml.load(templatefile)
# Set up vacuum simulations:
for phase in ['vacuum', 'explicit']:
if phase == "vacuum":
switching_lengths = vacuum_switching_lengths
else:
switching_lengths = solvent_switching_lengths
for switching_length in switching_lengths:
for sterics in use_sterics:
for geometry_division in geometry_divisions:
yaml_dict = copy.deepcopy(template_yaml)
specification_file_prefix = "{}_{}ncmc_{}sterics_{}geometry".format(phase, switching_length, sterics, geometry_division)
specification_filename = specification_file_prefix + ".yaml"
yaml_dict['geometry_divisions'][phase] = geometry_division
yaml_dict['use_sterics']['vacuum'] = sterics
yaml_dict['ncmc_switching_times'][phase] = switching_length
yaml_dict['phase'] = phase
yaml_dict['output_filename'] = specification_file_prefix + ".nc"
with open(specification_filename, 'w') as yam_outfile:
yaml.dump(yaml_dict, yam_outfile)
|
Add script to generate the yaml files for experimentsimport yaml
import copy
vacuum_switching_lengths = [0, 100, 500, 1000, 5000, 10000]
solvent_switching_lengths = [500, 1000, 5000, 10000, 20000, 50000]
use_sterics = [True, False]
geometry_divisions = [90, 180, 360, 720]
# Load in template yaml:
with open("rj_hydration.yaml", "r") as templatefile:
template_yaml = yaml.load(templatefile)
# Set up vacuum simulations:
for phase in ['vacuum', 'explicit']:
if phase == "vacuum":
switching_lengths = vacuum_switching_lengths
else:
switching_lengths = solvent_switching_lengths
for switching_length in switching_lengths:
for sterics in use_sterics:
for geometry_division in geometry_divisions:
yaml_dict = copy.deepcopy(template_yaml)
specification_file_prefix = "{}_{}ncmc_{}sterics_{}geometry".format(phase, switching_length, sterics, geometry_division)
specification_filename = specification_file_prefix + ".yaml"
yaml_dict['geometry_divisions'][phase] = geometry_division
yaml_dict['use_sterics']['vacuum'] = sterics
yaml_dict['ncmc_switching_times'][phase] = switching_length
yaml_dict['phase'] = phase
yaml_dict['output_filename'] = specification_file_prefix + ".nc"
with open(specification_filename, 'w') as yam_outfile:
yaml.dump(yaml_dict, yam_outfile)
|
<commit_before><commit_msg>Add script to generate the yaml files for experiments<commit_after>import yaml
import copy
vacuum_switching_lengths = [0, 100, 500, 1000, 5000, 10000]
solvent_switching_lengths = [500, 1000, 5000, 10000, 20000, 50000]
use_sterics = [True, False]
geometry_divisions = [90, 180, 360, 720]
# Load in template yaml:
with open("rj_hydration.yaml", "r") as templatefile:
template_yaml = yaml.load(templatefile)
# Set up vacuum simulations:
for phase in ['vacuum', 'explicit']:
if phase == "vacuum":
switching_lengths = vacuum_switching_lengths
else:
switching_lengths = solvent_switching_lengths
for switching_length in switching_lengths:
for sterics in use_sterics:
for geometry_division in geometry_divisions:
yaml_dict = copy.deepcopy(template_yaml)
specification_file_prefix = "{}_{}ncmc_{}sterics_{}geometry".format(phase, switching_length, sterics, geometry_division)
specification_filename = specification_file_prefix + ".yaml"
yaml_dict['geometry_divisions'][phase] = geometry_division
yaml_dict['use_sterics']['vacuum'] = sterics
yaml_dict['ncmc_switching_times'][phase] = switching_length
yaml_dict['phase'] = phase
yaml_dict['output_filename'] = specification_file_prefix + ".nc"
with open(specification_filename, 'w') as yam_outfile:
yaml.dump(yaml_dict, yam_outfile)
|
|
9ce7ee71b5eddc0ceff578b45c1324f8eb09ffe1
|
artbot_scraper/pipelines.py
|
artbot_scraper/pipelines.py
|
# -*- coding: utf-8 -*-
from django.db import IntegrityError
from scrapy.exceptions import DropItem
from titlecase import titlecase
from dateutil import parser, relativedelta
class EventPipeline(object):
def process_item(self, item, spider):
item['titleRaw'] = item['title']
item['title'] = titlecase(item['title'])
if (item['end'] < item['start']):
item['end'] = item['end'] + relativedelta.relativedelta(years =+ 1)
try:
item.save()
except IntegrityError:
raise DropItem('Duplicate: ' + item['venue'] + ' - ' + item['title'])
return item
|
# -*- coding: utf-8 -*-
from django.db import IntegrityError
from scrapy.exceptions import DropItem
from titlecase import titlecase
from dateutil import parser, relativedelta
class EventPipeline(object):
def process_item(self, item, spider):
item['titleRaw'] = item['title']
item['title'] = titlecase(item['title'])
if 'end' in item and 'start' in item:
if (item['end'] < item['start']):
item['end'] = item['end'] + relativedelta.relativedelta(years =+ 1)
try:
item.save()
except IntegrityError:
raise DropItem('Duplicate: ' + item['venue'] + ' - ' + item['title'])
return item
|
Verify that event end and start keys exist before accessing.
|
Verify that event end and start keys exist before accessing.
|
Python
|
mit
|
coreymcdermott/artbot,coreymcdermott/artbot
|
# -*- coding: utf-8 -*-
from django.db import IntegrityError
from scrapy.exceptions import DropItem
from titlecase import titlecase
from dateutil import parser, relativedelta
class EventPipeline(object):
def process_item(self, item, spider):
item['titleRaw'] = item['title']
item['title'] = titlecase(item['title'])
if (item['end'] < item['start']):
item['end'] = item['end'] + relativedelta.relativedelta(years =+ 1)
try:
item.save()
except IntegrityError:
raise DropItem('Duplicate: ' + item['venue'] + ' - ' + item['title'])
return item
Verify that event end and start keys exist before accessing.
|
# -*- coding: utf-8 -*-
from django.db import IntegrityError
from scrapy.exceptions import DropItem
from titlecase import titlecase
from dateutil import parser, relativedelta
class EventPipeline(object):
def process_item(self, item, spider):
item['titleRaw'] = item['title']
item['title'] = titlecase(item['title'])
if 'end' in item and 'start' in item:
if (item['end'] < item['start']):
item['end'] = item['end'] + relativedelta.relativedelta(years =+ 1)
try:
item.save()
except IntegrityError:
raise DropItem('Duplicate: ' + item['venue'] + ' - ' + item['title'])
return item
|
<commit_before># -*- coding: utf-8 -*-
from django.db import IntegrityError
from scrapy.exceptions import DropItem
from titlecase import titlecase
from dateutil import parser, relativedelta
class EventPipeline(object):
def process_item(self, item, spider):
item['titleRaw'] = item['title']
item['title'] = titlecase(item['title'])
if (item['end'] < item['start']):
item['end'] = item['end'] + relativedelta.relativedelta(years =+ 1)
try:
item.save()
except IntegrityError:
raise DropItem('Duplicate: ' + item['venue'] + ' - ' + item['title'])
return item
<commit_msg>Verify that event end and start keys exist before accessing.<commit_after>
|
# -*- coding: utf-8 -*-
from django.db import IntegrityError
from scrapy.exceptions import DropItem
from titlecase import titlecase
from dateutil import parser, relativedelta
class EventPipeline(object):
def process_item(self, item, spider):
item['titleRaw'] = item['title']
item['title'] = titlecase(item['title'])
if 'end' in item and 'start' in item:
if (item['end'] < item['start']):
item['end'] = item['end'] + relativedelta.relativedelta(years =+ 1)
try:
item.save()
except IntegrityError:
raise DropItem('Duplicate: ' + item['venue'] + ' - ' + item['title'])
return item
|
# -*- coding: utf-8 -*-
from django.db import IntegrityError
from scrapy.exceptions import DropItem
from titlecase import titlecase
from dateutil import parser, relativedelta
class EventPipeline(object):
def process_item(self, item, spider):
item['titleRaw'] = item['title']
item['title'] = titlecase(item['title'])
if (item['end'] < item['start']):
item['end'] = item['end'] + relativedelta.relativedelta(years =+ 1)
try:
item.save()
except IntegrityError:
raise DropItem('Duplicate: ' + item['venue'] + ' - ' + item['title'])
return item
Verify that event end and start keys exist before accessing.# -*- coding: utf-8 -*-
from django.db import IntegrityError
from scrapy.exceptions import DropItem
from titlecase import titlecase
from dateutil import parser, relativedelta
class EventPipeline(object):
def process_item(self, item, spider):
item['titleRaw'] = item['title']
item['title'] = titlecase(item['title'])
if 'end' in item and 'start' in item:
if (item['end'] < item['start']):
item['end'] = item['end'] + relativedelta.relativedelta(years =+ 1)
try:
item.save()
except IntegrityError:
raise DropItem('Duplicate: ' + item['venue'] + ' - ' + item['title'])
return item
|
<commit_before># -*- coding: utf-8 -*-
from django.db import IntegrityError
from scrapy.exceptions import DropItem
from titlecase import titlecase
from dateutil import parser, relativedelta
class EventPipeline(object):
def process_item(self, item, spider):
item['titleRaw'] = item['title']
item['title'] = titlecase(item['title'])
if (item['end'] < item['start']):
item['end'] = item['end'] + relativedelta.relativedelta(years =+ 1)
try:
item.save()
except IntegrityError:
raise DropItem('Duplicate: ' + item['venue'] + ' - ' + item['title'])
return item
<commit_msg>Verify that event end and start keys exist before accessing.<commit_after># -*- coding: utf-8 -*-
from django.db import IntegrityError
from scrapy.exceptions import DropItem
from titlecase import titlecase
from dateutil import parser, relativedelta
class EventPipeline(object):
def process_item(self, item, spider):
item['titleRaw'] = item['title']
item['title'] = titlecase(item['title'])
if 'end' in item and 'start' in item:
if (item['end'] < item['start']):
item['end'] = item['end'] + relativedelta.relativedelta(years =+ 1)
try:
item.save()
except IntegrityError:
raise DropItem('Duplicate: ' + item['venue'] + ' - ' + item['title'])
return item
|
e2a4262035e0d99c83a1bef8fdd594745a66b011
|
tests/app/views/test_application.py
|
tests/app/views/test_application.py
|
from nose.tools import assert_equal, assert_true
from ...helpers import BaseApplicationTest
class TestApplication(BaseApplicationTest):
def setup(self):
super(TestApplication, self).setup()
def test_analytics_code_should_be_in_javascript(self):
res = self.client.get('/static/javascripts/application.js')
assert_equal(200, res.status_code)
assert_true(
'GOVUK.analytics.trackPageview'
in res.get_data(as_text=True))
|
from nose.tools import assert_equal, assert_true
from ...helpers import BaseApplicationTest
class TestApplication(BaseApplicationTest):
def setup(self):
super(TestApplication, self).setup()
def test_analytics_code_should_be_in_javascript(self):
res = self.client.get('/static/javascripts/application.js')
assert_equal(200, res.status_code)
assert_true(
'trackPageview'
in res.get_data(as_text=True))
|
Test for presence of analytics in minified JS
|
Test for presence of analytics in minified JS
Minifying the JS shortens variable names wherever their scope makes it possible.
This means that `GOVUK.analytics.trackPageView` gets shortened to something like
`e.t.trackPageView`, and not in a predicatable way. Because the `trackPageView`
method could be used by indeterminable other scripts the compiler can't minify
it, so testing for just the presence of its name is safe.
|
Python
|
mit
|
AusDTO/dto-digitalmarketplace-buyer-frontend,mtekel/digitalmarketplace-buyer-frontend,alphagov/digitalmarketplace-buyer-frontend,alphagov/digitalmarketplace-buyer-frontend,alphagov/digitalmarketplace-buyer-frontend,mtekel/digitalmarketplace-buyer-frontend,AusDTO/dto-digitalmarketplace-buyer-frontend,mtekel/digitalmarketplace-buyer-frontend,AusDTO/dto-digitalmarketplace-buyer-frontend,alphagov/digitalmarketplace-buyer-frontend,mtekel/digitalmarketplace-buyer-frontend,AusDTO/dto-digitalmarketplace-buyer-frontend
|
from nose.tools import assert_equal, assert_true
from ...helpers import BaseApplicationTest
class TestApplication(BaseApplicationTest):
def setup(self):
super(TestApplication, self).setup()
def test_analytics_code_should_be_in_javascript(self):
res = self.client.get('/static/javascripts/application.js')
assert_equal(200, res.status_code)
assert_true(
'GOVUK.analytics.trackPageview'
in res.get_data(as_text=True))
Test for presence of analytics in minified JS
Minifying the JS shortens variable names wherever their scope makes it possible.
This means that `GOVUK.analytics.trackPageView` gets shortened to something like
`e.t.trackPageView`, and not in a predicatable way. Because the `trackPageView`
method could be used by indeterminable other scripts the compiler can't minify
it, so testing for just the presence of its name is safe.
|
from nose.tools import assert_equal, assert_true
from ...helpers import BaseApplicationTest
class TestApplication(BaseApplicationTest):
def setup(self):
super(TestApplication, self).setup()
def test_analytics_code_should_be_in_javascript(self):
res = self.client.get('/static/javascripts/application.js')
assert_equal(200, res.status_code)
assert_true(
'trackPageview'
in res.get_data(as_text=True))
|
<commit_before>from nose.tools import assert_equal, assert_true
from ...helpers import BaseApplicationTest
class TestApplication(BaseApplicationTest):
def setup(self):
super(TestApplication, self).setup()
def test_analytics_code_should_be_in_javascript(self):
res = self.client.get('/static/javascripts/application.js')
assert_equal(200, res.status_code)
assert_true(
'GOVUK.analytics.trackPageview'
in res.get_data(as_text=True))
<commit_msg>Test for presence of analytics in minified JS
Minifying the JS shortens variable names wherever their scope makes it possible.
This means that `GOVUK.analytics.trackPageView` gets shortened to something like
`e.t.trackPageView`, and not in a predicatable way. Because the `trackPageView`
method could be used by indeterminable other scripts the compiler can't minify
it, so testing for just the presence of its name is safe.<commit_after>
|
from nose.tools import assert_equal, assert_true
from ...helpers import BaseApplicationTest
class TestApplication(BaseApplicationTest):
def setup(self):
super(TestApplication, self).setup()
def test_analytics_code_should_be_in_javascript(self):
res = self.client.get('/static/javascripts/application.js')
assert_equal(200, res.status_code)
assert_true(
'trackPageview'
in res.get_data(as_text=True))
|
from nose.tools import assert_equal, assert_true
from ...helpers import BaseApplicationTest
class TestApplication(BaseApplicationTest):
def setup(self):
super(TestApplication, self).setup()
def test_analytics_code_should_be_in_javascript(self):
res = self.client.get('/static/javascripts/application.js')
assert_equal(200, res.status_code)
assert_true(
'GOVUK.analytics.trackPageview'
in res.get_data(as_text=True))
Test for presence of analytics in minified JS
Minifying the JS shortens variable names wherever their scope makes it possible.
This means that `GOVUK.analytics.trackPageView` gets shortened to something like
`e.t.trackPageView`, and not in a predicatable way. Because the `trackPageView`
method could be used by indeterminable other scripts the compiler can't minify
it, so testing for just the presence of its name is safe.from nose.tools import assert_equal, assert_true
from ...helpers import BaseApplicationTest
class TestApplication(BaseApplicationTest):
def setup(self):
super(TestApplication, self).setup()
def test_analytics_code_should_be_in_javascript(self):
res = self.client.get('/static/javascripts/application.js')
assert_equal(200, res.status_code)
assert_true(
'trackPageview'
in res.get_data(as_text=True))
|
<commit_before>from nose.tools import assert_equal, assert_true
from ...helpers import BaseApplicationTest
class TestApplication(BaseApplicationTest):
def setup(self):
super(TestApplication, self).setup()
def test_analytics_code_should_be_in_javascript(self):
res = self.client.get('/static/javascripts/application.js')
assert_equal(200, res.status_code)
assert_true(
'GOVUK.analytics.trackPageview'
in res.get_data(as_text=True))
<commit_msg>Test for presence of analytics in minified JS
Minifying the JS shortens variable names wherever their scope makes it possible.
This means that `GOVUK.analytics.trackPageView` gets shortened to something like
`e.t.trackPageView`, and not in a predicatable way. Because the `trackPageView`
method could be used by indeterminable other scripts the compiler can't minify
it, so testing for just the presence of its name is safe.<commit_after>from nose.tools import assert_equal, assert_true
from ...helpers import BaseApplicationTest
class TestApplication(BaseApplicationTest):
def setup(self):
super(TestApplication, self).setup()
def test_analytics_code_should_be_in_javascript(self):
res = self.client.get('/static/javascripts/application.js')
assert_equal(200, res.status_code)
assert_true(
'trackPageview'
in res.get_data(as_text=True))
|
4dce72f60d5575212448a7432eecc118bfd3c845
|
apps/reactions/permissions.py
|
apps/reactions/permissions.py
|
from rest_framework import permissions
# TODO Add write permission for 1%CREW / Assitants.
class IsAuthorOrReadOnly(permissions.BasePermission):
"""
Custom permission to only allow author of an object to edit it.
"""
def has_permission(self, request, view, obj=None):
# Skip the check unless this is an object-level test.
if obj is None:
return True
# Read permissions are allowed to any request.
if request.method in permissions.SAFE_METHODS:
return True
# Write permissions are only allowed to the author of the reaction.
return obj.author == request.user
|
from rest_framework import permissions
# TODO Add write permission for 1%CREW / Assistants.
class IsAuthorOrReadOnly(permissions.BasePermission):
"""
Custom permission to only allow author of an object to edit it.
"""
def has_permission(self, request, view, obj=None):
# Skip the check unless this is an object-level test.
if obj is None:
return True
# Read permissions are allowed to any request.
if request.method in permissions.SAFE_METHODS:
return True
# Write permissions are only allowed to the author of the reaction.
return obj.author == request.user
|
Fix spelling mistake in comment.
|
Fix spelling mistake in comment.
|
Python
|
bsd-3-clause
|
onepercentclub/onepercentclub-site,onepercentclub/onepercentclub-site,onepercentclub/onepercentclub-site,onepercentclub/onepercentclub-site,onepercentclub/onepercentclub-site
|
from rest_framework import permissions
# TODO Add write permission for 1%CREW / Assitants.
class IsAuthorOrReadOnly(permissions.BasePermission):
"""
Custom permission to only allow author of an object to edit it.
"""
def has_permission(self, request, view, obj=None):
# Skip the check unless this is an object-level test.
if obj is None:
return True
# Read permissions are allowed to any request.
if request.method in permissions.SAFE_METHODS:
return True
# Write permissions are only allowed to the author of the reaction.
return obj.author == request.user
Fix spelling mistake in comment.
|
from rest_framework import permissions
# TODO Add write permission for 1%CREW / Assistants.
class IsAuthorOrReadOnly(permissions.BasePermission):
"""
Custom permission to only allow author of an object to edit it.
"""
def has_permission(self, request, view, obj=None):
# Skip the check unless this is an object-level test.
if obj is None:
return True
# Read permissions are allowed to any request.
if request.method in permissions.SAFE_METHODS:
return True
# Write permissions are only allowed to the author of the reaction.
return obj.author == request.user
|
<commit_before>from rest_framework import permissions
# TODO Add write permission for 1%CREW / Assitants.
class IsAuthorOrReadOnly(permissions.BasePermission):
"""
Custom permission to only allow author of an object to edit it.
"""
def has_permission(self, request, view, obj=None):
# Skip the check unless this is an object-level test.
if obj is None:
return True
# Read permissions are allowed to any request.
if request.method in permissions.SAFE_METHODS:
return True
# Write permissions are only allowed to the author of the reaction.
return obj.author == request.user
<commit_msg>Fix spelling mistake in comment.<commit_after>
|
from rest_framework import permissions
# TODO Add write permission for 1%CREW / Assistants.
class IsAuthorOrReadOnly(permissions.BasePermission):
"""
Custom permission to only allow author of an object to edit it.
"""
def has_permission(self, request, view, obj=None):
# Skip the check unless this is an object-level test.
if obj is None:
return True
# Read permissions are allowed to any request.
if request.method in permissions.SAFE_METHODS:
return True
# Write permissions are only allowed to the author of the reaction.
return obj.author == request.user
|
from rest_framework import permissions
# TODO Add write permission for 1%CREW / Assitants.
class IsAuthorOrReadOnly(permissions.BasePermission):
"""
Custom permission to only allow author of an object to edit it.
"""
def has_permission(self, request, view, obj=None):
# Skip the check unless this is an object-level test.
if obj is None:
return True
# Read permissions are allowed to any request.
if request.method in permissions.SAFE_METHODS:
return True
# Write permissions are only allowed to the author of the reaction.
return obj.author == request.user
Fix spelling mistake in comment.from rest_framework import permissions
# TODO Add write permission for 1%CREW / Assistants.
class IsAuthorOrReadOnly(permissions.BasePermission):
"""
Custom permission to only allow author of an object to edit it.
"""
def has_permission(self, request, view, obj=None):
# Skip the check unless this is an object-level test.
if obj is None:
return True
# Read permissions are allowed to any request.
if request.method in permissions.SAFE_METHODS:
return True
# Write permissions are only allowed to the author of the reaction.
return obj.author == request.user
|
<commit_before>from rest_framework import permissions
# TODO Add write permission for 1%CREW / Assitants.
class IsAuthorOrReadOnly(permissions.BasePermission):
"""
Custom permission to only allow author of an object to edit it.
"""
def has_permission(self, request, view, obj=None):
# Skip the check unless this is an object-level test.
if obj is None:
return True
# Read permissions are allowed to any request.
if request.method in permissions.SAFE_METHODS:
return True
# Write permissions are only allowed to the author of the reaction.
return obj.author == request.user
<commit_msg>Fix spelling mistake in comment.<commit_after>from rest_framework import permissions
# TODO Add write permission for 1%CREW / Assistants.
class IsAuthorOrReadOnly(permissions.BasePermission):
"""
Custom permission to only allow author of an object to edit it.
"""
def has_permission(self, request, view, obj=None):
# Skip the check unless this is an object-level test.
if obj is None:
return True
# Read permissions are allowed to any request.
if request.method in permissions.SAFE_METHODS:
return True
# Write permissions are only allowed to the author of the reaction.
return obj.author == request.user
|
50f2acfcfe482c5452a80243b186ec411f672afc
|
boundaryservice/urls.py
|
boundaryservice/urls.py
|
from django.conf.urls.defaults import patterns, include, url
from boundaryservice.views import *
urlpatterns = patterns('',
url(r'^boundary-set/$', BoundarySetListView.as_view(), name='boundaryservice_set_list'),
url(r'^boundary-set/(?P<slug>[\w_-]+)/$', BoundarySetDetailView.as_view(), name='boundaryservice_set_detail'),
url(r'^boundary/$', BoundaryListView.as_view(), name='boundaryservice_boundary_list'),
url(r'^boundary/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryListView.as_view()),
url(r'^boundary/(?P<set_slug>[\w_-]+)/$', BoundaryListView.as_view(), name='boundaryservice_boundary_list'),
url(r'^boundary/(?P<set_slug>[\w_-]+)/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryListView.as_view()),
url(r'^boundary/(?P<set_slug>[\w_-]+)/(?P<slug>[\w_-]+)/$', BoundaryDetailView.as_view(), name='boundaryservice_boundary_detail'),
url(r'^boundary/(?P<set_slug>[\w_-]+)/(?P<slug>[\w_-]+)/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryGeoDetailView.as_view()),
)
|
from django.conf.urls.defaults import patterns, include, url
from boundaryservice.views import *
urlpatterns = patterns('',
url(r'^boundary-sets/$', BoundarySetListView.as_view(), name='boundaryservice_set_list'),
url(r'^boundary-sets/(?P<slug>[\w_-]+)/$', BoundarySetDetailView.as_view(), name='boundaryservice_set_detail'),
url(r'^boundaries/$', BoundaryListView.as_view(), name='boundaryservice_boundary_list'),
url(r'^boundaries/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryListView.as_view()),
url(r'^boundaries/(?P<set_slug>[\w_-]+)/$', BoundaryListView.as_view(), name='boundaryservice_boundary_list'),
url(r'^boundaries/(?P<set_slug>[\w_-]+)/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryListView.as_view()),
url(r'^boundaries/(?P<set_slug>[\w_-]+)/(?P<slug>[\w_-]+)/$', BoundaryDetailView.as_view(), name='boundaryservice_boundary_detail'),
url(r'^boundaries/(?P<set_slug>[\w_-]+)/(?P<slug>[\w_-]+)/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryGeoDetailView.as_view()),
)
|
Use plural names for resource types in URLs
|
Use plural names for resource types in URLs
|
Python
|
mit
|
datamade/represent-boundaries,opencorato/represent-boundaries,opencorato/represent-boundaries,datamade/represent-boundaries,datamade/represent-boundaries,opencorato/represent-boundaries
|
from django.conf.urls.defaults import patterns, include, url
from boundaryservice.views import *
urlpatterns = patterns('',
url(r'^boundary-set/$', BoundarySetListView.as_view(), name='boundaryservice_set_list'),
url(r'^boundary-set/(?P<slug>[\w_-]+)/$', BoundarySetDetailView.as_view(), name='boundaryservice_set_detail'),
url(r'^boundary/$', BoundaryListView.as_view(), name='boundaryservice_boundary_list'),
url(r'^boundary/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryListView.as_view()),
url(r'^boundary/(?P<set_slug>[\w_-]+)/$', BoundaryListView.as_view(), name='boundaryservice_boundary_list'),
url(r'^boundary/(?P<set_slug>[\w_-]+)/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryListView.as_view()),
url(r'^boundary/(?P<set_slug>[\w_-]+)/(?P<slug>[\w_-]+)/$', BoundaryDetailView.as_view(), name='boundaryservice_boundary_detail'),
url(r'^boundary/(?P<set_slug>[\w_-]+)/(?P<slug>[\w_-]+)/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryGeoDetailView.as_view()),
)
Use plural names for resource types in URLs
|
from django.conf.urls.defaults import patterns, include, url
from boundaryservice.views import *
urlpatterns = patterns('',
url(r'^boundary-sets/$', BoundarySetListView.as_view(), name='boundaryservice_set_list'),
url(r'^boundary-sets/(?P<slug>[\w_-]+)/$', BoundarySetDetailView.as_view(), name='boundaryservice_set_detail'),
url(r'^boundaries/$', BoundaryListView.as_view(), name='boundaryservice_boundary_list'),
url(r'^boundaries/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryListView.as_view()),
url(r'^boundaries/(?P<set_slug>[\w_-]+)/$', BoundaryListView.as_view(), name='boundaryservice_boundary_list'),
url(r'^boundaries/(?P<set_slug>[\w_-]+)/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryListView.as_view()),
url(r'^boundaries/(?P<set_slug>[\w_-]+)/(?P<slug>[\w_-]+)/$', BoundaryDetailView.as_view(), name='boundaryservice_boundary_detail'),
url(r'^boundaries/(?P<set_slug>[\w_-]+)/(?P<slug>[\w_-]+)/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryGeoDetailView.as_view()),
)
|
<commit_before>from django.conf.urls.defaults import patterns, include, url
from boundaryservice.views import *
urlpatterns = patterns('',
url(r'^boundary-set/$', BoundarySetListView.as_view(), name='boundaryservice_set_list'),
url(r'^boundary-set/(?P<slug>[\w_-]+)/$', BoundarySetDetailView.as_view(), name='boundaryservice_set_detail'),
url(r'^boundary/$', BoundaryListView.as_view(), name='boundaryservice_boundary_list'),
url(r'^boundary/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryListView.as_view()),
url(r'^boundary/(?P<set_slug>[\w_-]+)/$', BoundaryListView.as_view(), name='boundaryservice_boundary_list'),
url(r'^boundary/(?P<set_slug>[\w_-]+)/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryListView.as_view()),
url(r'^boundary/(?P<set_slug>[\w_-]+)/(?P<slug>[\w_-]+)/$', BoundaryDetailView.as_view(), name='boundaryservice_boundary_detail'),
url(r'^boundary/(?P<set_slug>[\w_-]+)/(?P<slug>[\w_-]+)/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryGeoDetailView.as_view()),
)
<commit_msg>Use plural names for resource types in URLs<commit_after>
|
from django.conf.urls.defaults import patterns, include, url
from boundaryservice.views import *
urlpatterns = patterns('',
url(r'^boundary-sets/$', BoundarySetListView.as_view(), name='boundaryservice_set_list'),
url(r'^boundary-sets/(?P<slug>[\w_-]+)/$', BoundarySetDetailView.as_view(), name='boundaryservice_set_detail'),
url(r'^boundaries/$', BoundaryListView.as_view(), name='boundaryservice_boundary_list'),
url(r'^boundaries/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryListView.as_view()),
url(r'^boundaries/(?P<set_slug>[\w_-]+)/$', BoundaryListView.as_view(), name='boundaryservice_boundary_list'),
url(r'^boundaries/(?P<set_slug>[\w_-]+)/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryListView.as_view()),
url(r'^boundaries/(?P<set_slug>[\w_-]+)/(?P<slug>[\w_-]+)/$', BoundaryDetailView.as_view(), name='boundaryservice_boundary_detail'),
url(r'^boundaries/(?P<set_slug>[\w_-]+)/(?P<slug>[\w_-]+)/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryGeoDetailView.as_view()),
)
|
from django.conf.urls.defaults import patterns, include, url
from boundaryservice.views import *
urlpatterns = patterns('',
url(r'^boundary-set/$', BoundarySetListView.as_view(), name='boundaryservice_set_list'),
url(r'^boundary-set/(?P<slug>[\w_-]+)/$', BoundarySetDetailView.as_view(), name='boundaryservice_set_detail'),
url(r'^boundary/$', BoundaryListView.as_view(), name='boundaryservice_boundary_list'),
url(r'^boundary/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryListView.as_view()),
url(r'^boundary/(?P<set_slug>[\w_-]+)/$', BoundaryListView.as_view(), name='boundaryservice_boundary_list'),
url(r'^boundary/(?P<set_slug>[\w_-]+)/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryListView.as_view()),
url(r'^boundary/(?P<set_slug>[\w_-]+)/(?P<slug>[\w_-]+)/$', BoundaryDetailView.as_view(), name='boundaryservice_boundary_detail'),
url(r'^boundary/(?P<set_slug>[\w_-]+)/(?P<slug>[\w_-]+)/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryGeoDetailView.as_view()),
)
Use plural names for resource types in URLsfrom django.conf.urls.defaults import patterns, include, url
from boundaryservice.views import *
urlpatterns = patterns('',
url(r'^boundary-sets/$', BoundarySetListView.as_view(), name='boundaryservice_set_list'),
url(r'^boundary-sets/(?P<slug>[\w_-]+)/$', BoundarySetDetailView.as_view(), name='boundaryservice_set_detail'),
url(r'^boundaries/$', BoundaryListView.as_view(), name='boundaryservice_boundary_list'),
url(r'^boundaries/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryListView.as_view()),
url(r'^boundaries/(?P<set_slug>[\w_-]+)/$', BoundaryListView.as_view(), name='boundaryservice_boundary_list'),
url(r'^boundaries/(?P<set_slug>[\w_-]+)/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryListView.as_view()),
url(r'^boundaries/(?P<set_slug>[\w_-]+)/(?P<slug>[\w_-]+)/$', BoundaryDetailView.as_view(), name='boundaryservice_boundary_detail'),
url(r'^boundaries/(?P<set_slug>[\w_-]+)/(?P<slug>[\w_-]+)/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryGeoDetailView.as_view()),
)
|
<commit_before>from django.conf.urls.defaults import patterns, include, url
from boundaryservice.views import *
urlpatterns = patterns('',
url(r'^boundary-set/$', BoundarySetListView.as_view(), name='boundaryservice_set_list'),
url(r'^boundary-set/(?P<slug>[\w_-]+)/$', BoundarySetDetailView.as_view(), name='boundaryservice_set_detail'),
url(r'^boundary/$', BoundaryListView.as_view(), name='boundaryservice_boundary_list'),
url(r'^boundary/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryListView.as_view()),
url(r'^boundary/(?P<set_slug>[\w_-]+)/$', BoundaryListView.as_view(), name='boundaryservice_boundary_list'),
url(r'^boundary/(?P<set_slug>[\w_-]+)/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryListView.as_view()),
url(r'^boundary/(?P<set_slug>[\w_-]+)/(?P<slug>[\w_-]+)/$', BoundaryDetailView.as_view(), name='boundaryservice_boundary_detail'),
url(r'^boundary/(?P<set_slug>[\w_-]+)/(?P<slug>[\w_-]+)/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryGeoDetailView.as_view()),
)
<commit_msg>Use plural names for resource types in URLs<commit_after>from django.conf.urls.defaults import patterns, include, url
from boundaryservice.views import *
urlpatterns = patterns('',
url(r'^boundary-sets/$', BoundarySetListView.as_view(), name='boundaryservice_set_list'),
url(r'^boundary-sets/(?P<slug>[\w_-]+)/$', BoundarySetDetailView.as_view(), name='boundaryservice_set_detail'),
url(r'^boundaries/$', BoundaryListView.as_view(), name='boundaryservice_boundary_list'),
url(r'^boundaries/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryListView.as_view()),
url(r'^boundaries/(?P<set_slug>[\w_-]+)/$', BoundaryListView.as_view(), name='boundaryservice_boundary_list'),
url(r'^boundaries/(?P<set_slug>[\w_-]+)/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryListView.as_view()),
url(r'^boundaries/(?P<set_slug>[\w_-]+)/(?P<slug>[\w_-]+)/$', BoundaryDetailView.as_view(), name='boundaryservice_boundary_detail'),
url(r'^boundaries/(?P<set_slug>[\w_-]+)/(?P<slug>[\w_-]+)/(?P<geo_field>shape|simple_shape|centroid)$', BoundaryGeoDetailView.as_view()),
)
|
1631eb5d1e009c236bdc3db1d2d44da9e9e9102a
|
kokki/cookbooks/busket/recipes/default.py
|
kokki/cookbooks/busket/recipes/default.py
|
import os
from kokki import *
Package("erlang")
Package("mercurial",
provider = "kokki.providers.package.easy_install.EasyInstallProvider")
Script("install-busket",
not_if = lambda:os.path.exists(env.config.busket.path),
cwd = "/usr/local/src",
code = (
"git clone git://github.com/samuel/busket.git busket\n"
"cd busket\n"
"make release\n"
"mv rel/busket {install_path}\n"
).format(install_path=env.config.busket.path)
)
|
import os
from kokki import *
Package("erlang")
# ubuntu's erlang is a bit messed up.. remove the man link
File("/usr/lib/erlang/man",
action = "delete")
Package("mercurial",
provider = "kokki.providers.package.easy_install.EasyInstallProvider")
Script("install-busket",
not_if = lambda:os.path.exists(env.config.busket.path),
cwd = "/usr/local/src",
code = (
"git clone git://github.com/samuel/busket.git busket\n"
"cd busket\n"
"make release\n"
"mv rel/busket {install_path}\n"
).format(install_path=env.config.busket.path)
)
|
Remove man link for erlang in ubuntu
|
Remove man link for erlang in ubuntu
|
Python
|
bsd-3-clause
|
samuel/kokki
|
import os
from kokki import *
Package("erlang")
Package("mercurial",
provider = "kokki.providers.package.easy_install.EasyInstallProvider")
Script("install-busket",
not_if = lambda:os.path.exists(env.config.busket.path),
cwd = "/usr/local/src",
code = (
"git clone git://github.com/samuel/busket.git busket\n"
"cd busket\n"
"make release\n"
"mv rel/busket {install_path}\n"
).format(install_path=env.config.busket.path)
)
Remove man link for erlang in ubuntu
|
import os
from kokki import *
Package("erlang")
# ubuntu's erlang is a bit messed up.. remove the man link
File("/usr/lib/erlang/man",
action = "delete")
Package("mercurial",
provider = "kokki.providers.package.easy_install.EasyInstallProvider")
Script("install-busket",
not_if = lambda:os.path.exists(env.config.busket.path),
cwd = "/usr/local/src",
code = (
"git clone git://github.com/samuel/busket.git busket\n"
"cd busket\n"
"make release\n"
"mv rel/busket {install_path}\n"
).format(install_path=env.config.busket.path)
)
|
<commit_before>
import os
from kokki import *
Package("erlang")
Package("mercurial",
provider = "kokki.providers.package.easy_install.EasyInstallProvider")
Script("install-busket",
not_if = lambda:os.path.exists(env.config.busket.path),
cwd = "/usr/local/src",
code = (
"git clone git://github.com/samuel/busket.git busket\n"
"cd busket\n"
"make release\n"
"mv rel/busket {install_path}\n"
).format(install_path=env.config.busket.path)
)
<commit_msg>Remove man link for erlang in ubuntu<commit_after>
|
import os
from kokki import *
Package("erlang")
# ubuntu's erlang is a bit messed up.. remove the man link
File("/usr/lib/erlang/man",
action = "delete")
Package("mercurial",
provider = "kokki.providers.package.easy_install.EasyInstallProvider")
Script("install-busket",
not_if = lambda:os.path.exists(env.config.busket.path),
cwd = "/usr/local/src",
code = (
"git clone git://github.com/samuel/busket.git busket\n"
"cd busket\n"
"make release\n"
"mv rel/busket {install_path}\n"
).format(install_path=env.config.busket.path)
)
|
import os
from kokki import *
Package("erlang")
Package("mercurial",
provider = "kokki.providers.package.easy_install.EasyInstallProvider")
Script("install-busket",
not_if = lambda:os.path.exists(env.config.busket.path),
cwd = "/usr/local/src",
code = (
"git clone git://github.com/samuel/busket.git busket\n"
"cd busket\n"
"make release\n"
"mv rel/busket {install_path}\n"
).format(install_path=env.config.busket.path)
)
Remove man link for erlang in ubuntu
import os
from kokki import *
Package("erlang")
# ubuntu's erlang is a bit messed up.. remove the man link
File("/usr/lib/erlang/man",
action = "delete")
Package("mercurial",
provider = "kokki.providers.package.easy_install.EasyInstallProvider")
Script("install-busket",
not_if = lambda:os.path.exists(env.config.busket.path),
cwd = "/usr/local/src",
code = (
"git clone git://github.com/samuel/busket.git busket\n"
"cd busket\n"
"make release\n"
"mv rel/busket {install_path}\n"
).format(install_path=env.config.busket.path)
)
|
<commit_before>
import os
from kokki import *
Package("erlang")
Package("mercurial",
provider = "kokki.providers.package.easy_install.EasyInstallProvider")
Script("install-busket",
not_if = lambda:os.path.exists(env.config.busket.path),
cwd = "/usr/local/src",
code = (
"git clone git://github.com/samuel/busket.git busket\n"
"cd busket\n"
"make release\n"
"mv rel/busket {install_path}\n"
).format(install_path=env.config.busket.path)
)
<commit_msg>Remove man link for erlang in ubuntu<commit_after>
import os
from kokki import *
Package("erlang")
# ubuntu's erlang is a bit messed up.. remove the man link
File("/usr/lib/erlang/man",
action = "delete")
Package("mercurial",
provider = "kokki.providers.package.easy_install.EasyInstallProvider")
Script("install-busket",
not_if = lambda:os.path.exists(env.config.busket.path),
cwd = "/usr/local/src",
code = (
"git clone git://github.com/samuel/busket.git busket\n"
"cd busket\n"
"make release\n"
"mv rel/busket {install_path}\n"
).format(install_path=env.config.busket.path)
)
|
963f9ed01b400cd95e14aecdee7c265fe48a4d41
|
mopidy_nad/__init__.py
|
mopidy_nad/__init__.py
|
import os
import pkg_resources
from mopidy import config, ext
__version__ = pkg_resources.get_distribution("Mopidy-NAD").version
class Extension(ext.Extension):
dist_name = "Mopidy-NAD"
ext_name = "nad"
version = __version__
def get_default_config(self):
conf_file = os.path.join(os.path.dirname(__file__), "ext.conf")
return config.read(conf_file)
def get_config_schema(self):
schema = super().get_config_schema()
schema["port"] = config.String()
schema["source"] = config.String(optional=True)
schema["speakers-a"] = config.Boolean(optional=True)
schema["speakers-b"] = config.Boolean(optional=True)
return schema
def setup(self, registry):
from mopidy_nad.mixer import NadMixer
registry.add("mixer", NadMixer)
|
import pathlib
import pkg_resources
from mopidy import config, ext
__version__ = pkg_resources.get_distribution("Mopidy-NAD").version
class Extension(ext.Extension):
dist_name = "Mopidy-NAD"
ext_name = "nad"
version = __version__
def get_default_config(self):
return config.read(pathlib.Path(__file__).parent / "ext.conf")
def get_config_schema(self):
schema = super().get_config_schema()
schema["port"] = config.String()
schema["source"] = config.String(optional=True)
schema["speakers-a"] = config.Boolean(optional=True)
schema["speakers-b"] = config.Boolean(optional=True)
return schema
def setup(self, registry):
from mopidy_nad.mixer import NadMixer
registry.add("mixer", NadMixer)
|
Use pathlib to read ext.conf
|
Use pathlib to read ext.conf
|
Python
|
apache-2.0
|
mopidy/mopidy-nad
|
import os
import pkg_resources
from mopidy import config, ext
__version__ = pkg_resources.get_distribution("Mopidy-NAD").version
class Extension(ext.Extension):
dist_name = "Mopidy-NAD"
ext_name = "nad"
version = __version__
def get_default_config(self):
conf_file = os.path.join(os.path.dirname(__file__), "ext.conf")
return config.read(conf_file)
def get_config_schema(self):
schema = super().get_config_schema()
schema["port"] = config.String()
schema["source"] = config.String(optional=True)
schema["speakers-a"] = config.Boolean(optional=True)
schema["speakers-b"] = config.Boolean(optional=True)
return schema
def setup(self, registry):
from mopidy_nad.mixer import NadMixer
registry.add("mixer", NadMixer)
Use pathlib to read ext.conf
|
import pathlib
import pkg_resources
from mopidy import config, ext
__version__ = pkg_resources.get_distribution("Mopidy-NAD").version
class Extension(ext.Extension):
dist_name = "Mopidy-NAD"
ext_name = "nad"
version = __version__
def get_default_config(self):
return config.read(pathlib.Path(__file__).parent / "ext.conf")
def get_config_schema(self):
schema = super().get_config_schema()
schema["port"] = config.String()
schema["source"] = config.String(optional=True)
schema["speakers-a"] = config.Boolean(optional=True)
schema["speakers-b"] = config.Boolean(optional=True)
return schema
def setup(self, registry):
from mopidy_nad.mixer import NadMixer
registry.add("mixer", NadMixer)
|
<commit_before>import os
import pkg_resources
from mopidy import config, ext
__version__ = pkg_resources.get_distribution("Mopidy-NAD").version
class Extension(ext.Extension):
dist_name = "Mopidy-NAD"
ext_name = "nad"
version = __version__
def get_default_config(self):
conf_file = os.path.join(os.path.dirname(__file__), "ext.conf")
return config.read(conf_file)
def get_config_schema(self):
schema = super().get_config_schema()
schema["port"] = config.String()
schema["source"] = config.String(optional=True)
schema["speakers-a"] = config.Boolean(optional=True)
schema["speakers-b"] = config.Boolean(optional=True)
return schema
def setup(self, registry):
from mopidy_nad.mixer import NadMixer
registry.add("mixer", NadMixer)
<commit_msg>Use pathlib to read ext.conf<commit_after>
|
import pathlib
import pkg_resources
from mopidy import config, ext
__version__ = pkg_resources.get_distribution("Mopidy-NAD").version
class Extension(ext.Extension):
dist_name = "Mopidy-NAD"
ext_name = "nad"
version = __version__
def get_default_config(self):
return config.read(pathlib.Path(__file__).parent / "ext.conf")
def get_config_schema(self):
schema = super().get_config_schema()
schema["port"] = config.String()
schema["source"] = config.String(optional=True)
schema["speakers-a"] = config.Boolean(optional=True)
schema["speakers-b"] = config.Boolean(optional=True)
return schema
def setup(self, registry):
from mopidy_nad.mixer import NadMixer
registry.add("mixer", NadMixer)
|
import os
import pkg_resources
from mopidy import config, ext
__version__ = pkg_resources.get_distribution("Mopidy-NAD").version
class Extension(ext.Extension):
dist_name = "Mopidy-NAD"
ext_name = "nad"
version = __version__
def get_default_config(self):
conf_file = os.path.join(os.path.dirname(__file__), "ext.conf")
return config.read(conf_file)
def get_config_schema(self):
schema = super().get_config_schema()
schema["port"] = config.String()
schema["source"] = config.String(optional=True)
schema["speakers-a"] = config.Boolean(optional=True)
schema["speakers-b"] = config.Boolean(optional=True)
return schema
def setup(self, registry):
from mopidy_nad.mixer import NadMixer
registry.add("mixer", NadMixer)
Use pathlib to read ext.confimport pathlib
import pkg_resources
from mopidy import config, ext
__version__ = pkg_resources.get_distribution("Mopidy-NAD").version
class Extension(ext.Extension):
dist_name = "Mopidy-NAD"
ext_name = "nad"
version = __version__
def get_default_config(self):
return config.read(pathlib.Path(__file__).parent / "ext.conf")
def get_config_schema(self):
schema = super().get_config_schema()
schema["port"] = config.String()
schema["source"] = config.String(optional=True)
schema["speakers-a"] = config.Boolean(optional=True)
schema["speakers-b"] = config.Boolean(optional=True)
return schema
def setup(self, registry):
from mopidy_nad.mixer import NadMixer
registry.add("mixer", NadMixer)
|
<commit_before>import os
import pkg_resources
from mopidy import config, ext
__version__ = pkg_resources.get_distribution("Mopidy-NAD").version
class Extension(ext.Extension):
dist_name = "Mopidy-NAD"
ext_name = "nad"
version = __version__
def get_default_config(self):
conf_file = os.path.join(os.path.dirname(__file__), "ext.conf")
return config.read(conf_file)
def get_config_schema(self):
schema = super().get_config_schema()
schema["port"] = config.String()
schema["source"] = config.String(optional=True)
schema["speakers-a"] = config.Boolean(optional=True)
schema["speakers-b"] = config.Boolean(optional=True)
return schema
def setup(self, registry):
from mopidy_nad.mixer import NadMixer
registry.add("mixer", NadMixer)
<commit_msg>Use pathlib to read ext.conf<commit_after>import pathlib
import pkg_resources
from mopidy import config, ext
__version__ = pkg_resources.get_distribution("Mopidy-NAD").version
class Extension(ext.Extension):
dist_name = "Mopidy-NAD"
ext_name = "nad"
version = __version__
def get_default_config(self):
return config.read(pathlib.Path(__file__).parent / "ext.conf")
def get_config_schema(self):
schema = super().get_config_schema()
schema["port"] = config.String()
schema["source"] = config.String(optional=True)
schema["speakers-a"] = config.Boolean(optional=True)
schema["speakers-b"] = config.Boolean(optional=True)
return schema
def setup(self, registry):
from mopidy_nad.mixer import NadMixer
registry.add("mixer", NadMixer)
|
334e13b39945dab3e2c03752f32af0de5e382e9d
|
base/consensus.py
|
base/consensus.py
|
from abc import ABCMeta, abstractmethod
class Consensus(metaclass=ABCMeta):
"""
An interface for defining a consensus protocol.
The 'propose' and 'decide' methods need to be defined
"""
@abstractmethod
def propose(self, message):
#raise NotImplementedError("Method 'propose' needs to be implemented")
pass
@abstractmethod
def decide(self):
#raise NotImplementedError("Method 'decide' needs to be implemented")
pass
|
Add abstract Consensus protocol class
|
Add abstract Consensus protocol class
|
Python
|
mit
|
koevskinikola/ByzantineRandomizedConsensus
|
Add abstract Consensus protocol class
|
from abc import ABCMeta, abstractmethod
class Consensus(metaclass=ABCMeta):
"""
An interface for defining a consensus protocol.
The 'propose' and 'decide' methods need to be defined
"""
@abstractmethod
def propose(self, message):
#raise NotImplementedError("Method 'propose' needs to be implemented")
pass
@abstractmethod
def decide(self):
#raise NotImplementedError("Method 'decide' needs to be implemented")
pass
|
<commit_before><commit_msg>Add abstract Consensus protocol class<commit_after>
|
from abc import ABCMeta, abstractmethod
class Consensus(metaclass=ABCMeta):
"""
An interface for defining a consensus protocol.
The 'propose' and 'decide' methods need to be defined
"""
@abstractmethod
def propose(self, message):
#raise NotImplementedError("Method 'propose' needs to be implemented")
pass
@abstractmethod
def decide(self):
#raise NotImplementedError("Method 'decide' needs to be implemented")
pass
|
Add abstract Consensus protocol classfrom abc import ABCMeta, abstractmethod
class Consensus(metaclass=ABCMeta):
"""
An interface for defining a consensus protocol.
The 'propose' and 'decide' methods need to be defined
"""
@abstractmethod
def propose(self, message):
#raise NotImplementedError("Method 'propose' needs to be implemented")
pass
@abstractmethod
def decide(self):
#raise NotImplementedError("Method 'decide' needs to be implemented")
pass
|
<commit_before><commit_msg>Add abstract Consensus protocol class<commit_after>from abc import ABCMeta, abstractmethod
class Consensus(metaclass=ABCMeta):
"""
An interface for defining a consensus protocol.
The 'propose' and 'decide' methods need to be defined
"""
@abstractmethod
def propose(self, message):
#raise NotImplementedError("Method 'propose' needs to be implemented")
pass
@abstractmethod
def decide(self):
#raise NotImplementedError("Method 'decide' needs to be implemented")
pass
|
|
cd9a32c9c6ff2adc9e85fe471c30cf555b8871b0
|
tests/abm/test_pops.py
|
tests/abm/test_pops.py
|
# -*- coding: utf-8 -*-
"""
test_pops
~~~~~~~~~
tests for population code
"""
from abm import pops
from abm.entities import Task
import pytest
from scipy.stats.distributions import uniform
import numpy as np
@pytest.fixture
def basicenv():
return pops.Environment()
@pytest.mark.unit
def test_distribute_awards(basicenv):
class MockEntity(object):
total_award = 0
def award(self, amount):
self.total_award += amount
basicenv.population = []
for i in range(6):
basicenv.population.append(MockEntity())
basicenv.path = [1, 2, 3, 2, 4]
task = Task(4)
basicenv._distribute_awards(task)
observed = [
x.total_award for x in basicenv.population
]
expected = ([0] + [.25] * 3 + [0, 0])
assert observed == expected
assert np.isclose(sum(observed), .75)
basicenv.path = [0, 4]
basicenv._distribute_awards(task)
assert basicenv.population[0].total_award == 1.
basicenv.path = [5] * (basicenv.path_cutoff + 1)
basicenv._distribute_awards(task)
assert basicenv.population[5].total_award == -.05
|
# -*- coding: utf-8 -*-
"""
test_pops
~~~~~~~~~
tests for population code
"""
from abm import pops
from abm.entities import Task
import pytest
from scipy.stats.distributions import uniform
import numpy as np
@pytest.fixture
def basicenv():
return pops.Environment()
@pytest.mark.unit
def test_distribute_awards(basicenv):
class MockEntity(object):
total_award = 0
def award(self, amount):
self.total_award += amount
basicenv.population = []
for i in range(6):
basicenv.population.append(MockEntity())
basicenv.path = [1, 2, 3, 2, 4]
task = Task(4)
basicenv._distribute_awards(task)
observed = [
x.total_award for x in basicenv.population
]
expected = ([0] + [.25] * 3 + [0, 0])
assert observed == expected
assert np.isclose(sum(observed), .75)
basicenv.path = [0, 4]
basicenv._distribute_awards(task)
assert basicenv.population[0].total_award == 1.
basicenv.path = [5] * (basicenv.path_cutoff + 1)
basicenv._distribute_awards(task)
assert basicenv.population[5].total_award == -.05
|
Add newline at the end of the file for my sanity.
|
Add newline at the end of the file for my sanity.
|
Python
|
mit
|
bhtucker/agents
|
# -*- coding: utf-8 -*-
"""
test_pops
~~~~~~~~~
tests for population code
"""
from abm import pops
from abm.entities import Task
import pytest
from scipy.stats.distributions import uniform
import numpy as np
@pytest.fixture
def basicenv():
return pops.Environment()
@pytest.mark.unit
def test_distribute_awards(basicenv):
class MockEntity(object):
total_award = 0
def award(self, amount):
self.total_award += amount
basicenv.population = []
for i in range(6):
basicenv.population.append(MockEntity())
basicenv.path = [1, 2, 3, 2, 4]
task = Task(4)
basicenv._distribute_awards(task)
observed = [
x.total_award for x in basicenv.population
]
expected = ([0] + [.25] * 3 + [0, 0])
assert observed == expected
assert np.isclose(sum(observed), .75)
basicenv.path = [0, 4]
basicenv._distribute_awards(task)
assert basicenv.population[0].total_award == 1.
basicenv.path = [5] * (basicenv.path_cutoff + 1)
basicenv._distribute_awards(task)
assert basicenv.population[5].total_award == -.05
Add newline at the end of the file for my sanity.
|
# -*- coding: utf-8 -*-
"""
test_pops
~~~~~~~~~
tests for population code
"""
from abm import pops
from abm.entities import Task
import pytest
from scipy.stats.distributions import uniform
import numpy as np
@pytest.fixture
def basicenv():
return pops.Environment()
@pytest.mark.unit
def test_distribute_awards(basicenv):
class MockEntity(object):
total_award = 0
def award(self, amount):
self.total_award += amount
basicenv.population = []
for i in range(6):
basicenv.population.append(MockEntity())
basicenv.path = [1, 2, 3, 2, 4]
task = Task(4)
basicenv._distribute_awards(task)
observed = [
x.total_award for x in basicenv.population
]
expected = ([0] + [.25] * 3 + [0, 0])
assert observed == expected
assert np.isclose(sum(observed), .75)
basicenv.path = [0, 4]
basicenv._distribute_awards(task)
assert basicenv.population[0].total_award == 1.
basicenv.path = [5] * (basicenv.path_cutoff + 1)
basicenv._distribute_awards(task)
assert basicenv.population[5].total_award == -.05
|
<commit_before># -*- coding: utf-8 -*-
"""
test_pops
~~~~~~~~~
tests for population code
"""
from abm import pops
from abm.entities import Task
import pytest
from scipy.stats.distributions import uniform
import numpy as np
@pytest.fixture
def basicenv():
return pops.Environment()
@pytest.mark.unit
def test_distribute_awards(basicenv):
class MockEntity(object):
total_award = 0
def award(self, amount):
self.total_award += amount
basicenv.population = []
for i in range(6):
basicenv.population.append(MockEntity())
basicenv.path = [1, 2, 3, 2, 4]
task = Task(4)
basicenv._distribute_awards(task)
observed = [
x.total_award for x in basicenv.population
]
expected = ([0] + [.25] * 3 + [0, 0])
assert observed == expected
assert np.isclose(sum(observed), .75)
basicenv.path = [0, 4]
basicenv._distribute_awards(task)
assert basicenv.population[0].total_award == 1.
basicenv.path = [5] * (basicenv.path_cutoff + 1)
basicenv._distribute_awards(task)
assert basicenv.population[5].total_award == -.05
<commit_msg>Add newline at the end of the file for my sanity.<commit_after>
|
# -*- coding: utf-8 -*-
"""
test_pops
~~~~~~~~~
tests for population code
"""
from abm import pops
from abm.entities import Task
import pytest
from scipy.stats.distributions import uniform
import numpy as np
@pytest.fixture
def basicenv():
return pops.Environment()
@pytest.mark.unit
def test_distribute_awards(basicenv):
class MockEntity(object):
total_award = 0
def award(self, amount):
self.total_award += amount
basicenv.population = []
for i in range(6):
basicenv.population.append(MockEntity())
basicenv.path = [1, 2, 3, 2, 4]
task = Task(4)
basicenv._distribute_awards(task)
observed = [
x.total_award for x in basicenv.population
]
expected = ([0] + [.25] * 3 + [0, 0])
assert observed == expected
assert np.isclose(sum(observed), .75)
basicenv.path = [0, 4]
basicenv._distribute_awards(task)
assert basicenv.population[0].total_award == 1.
basicenv.path = [5] * (basicenv.path_cutoff + 1)
basicenv._distribute_awards(task)
assert basicenv.population[5].total_award == -.05
|
# -*- coding: utf-8 -*-
"""
test_pops
~~~~~~~~~
tests for population code
"""
from abm import pops
from abm.entities import Task
import pytest
from scipy.stats.distributions import uniform
import numpy as np
@pytest.fixture
def basicenv():
return pops.Environment()
@pytest.mark.unit
def test_distribute_awards(basicenv):
class MockEntity(object):
total_award = 0
def award(self, amount):
self.total_award += amount
basicenv.population = []
for i in range(6):
basicenv.population.append(MockEntity())
basicenv.path = [1, 2, 3, 2, 4]
task = Task(4)
basicenv._distribute_awards(task)
observed = [
x.total_award for x in basicenv.population
]
expected = ([0] + [.25] * 3 + [0, 0])
assert observed == expected
assert np.isclose(sum(observed), .75)
basicenv.path = [0, 4]
basicenv._distribute_awards(task)
assert basicenv.population[0].total_award == 1.
basicenv.path = [5] * (basicenv.path_cutoff + 1)
basicenv._distribute_awards(task)
assert basicenv.population[5].total_award == -.05
Add newline at the end of the file for my sanity.# -*- coding: utf-8 -*-
"""
test_pops
~~~~~~~~~
tests for population code
"""
from abm import pops
from abm.entities import Task
import pytest
from scipy.stats.distributions import uniform
import numpy as np
@pytest.fixture
def basicenv():
return pops.Environment()
@pytest.mark.unit
def test_distribute_awards(basicenv):
class MockEntity(object):
total_award = 0
def award(self, amount):
self.total_award += amount
basicenv.population = []
for i in range(6):
basicenv.population.append(MockEntity())
basicenv.path = [1, 2, 3, 2, 4]
task = Task(4)
basicenv._distribute_awards(task)
observed = [
x.total_award for x in basicenv.population
]
expected = ([0] + [.25] * 3 + [0, 0])
assert observed == expected
assert np.isclose(sum(observed), .75)
basicenv.path = [0, 4]
basicenv._distribute_awards(task)
assert basicenv.population[0].total_award == 1.
basicenv.path = [5] * (basicenv.path_cutoff + 1)
basicenv._distribute_awards(task)
assert basicenv.population[5].total_award == -.05
|
<commit_before># -*- coding: utf-8 -*-
"""
test_pops
~~~~~~~~~
tests for population code
"""
from abm import pops
from abm.entities import Task
import pytest
from scipy.stats.distributions import uniform
import numpy as np
@pytest.fixture
def basicenv():
return pops.Environment()
@pytest.mark.unit
def test_distribute_awards(basicenv):
class MockEntity(object):
total_award = 0
def award(self, amount):
self.total_award += amount
basicenv.population = []
for i in range(6):
basicenv.population.append(MockEntity())
basicenv.path = [1, 2, 3, 2, 4]
task = Task(4)
basicenv._distribute_awards(task)
observed = [
x.total_award for x in basicenv.population
]
expected = ([0] + [.25] * 3 + [0, 0])
assert observed == expected
assert np.isclose(sum(observed), .75)
basicenv.path = [0, 4]
basicenv._distribute_awards(task)
assert basicenv.population[0].total_award == 1.
basicenv.path = [5] * (basicenv.path_cutoff + 1)
basicenv._distribute_awards(task)
assert basicenv.population[5].total_award == -.05
<commit_msg>Add newline at the end of the file for my sanity.<commit_after># -*- coding: utf-8 -*-
"""
test_pops
~~~~~~~~~
tests for population code
"""
from abm import pops
from abm.entities import Task
import pytest
from scipy.stats.distributions import uniform
import numpy as np
@pytest.fixture
def basicenv():
return pops.Environment()
@pytest.mark.unit
def test_distribute_awards(basicenv):
class MockEntity(object):
total_award = 0
def award(self, amount):
self.total_award += amount
basicenv.population = []
for i in range(6):
basicenv.population.append(MockEntity())
basicenv.path = [1, 2, 3, 2, 4]
task = Task(4)
basicenv._distribute_awards(task)
observed = [
x.total_award for x in basicenv.population
]
expected = ([0] + [.25] * 3 + [0, 0])
assert observed == expected
assert np.isclose(sum(observed), .75)
basicenv.path = [0, 4]
basicenv._distribute_awards(task)
assert basicenv.population[0].total_award == 1.
basicenv.path = [5] * (basicenv.path_cutoff + 1)
basicenv._distribute_awards(task)
assert basicenv.population[5].total_award == -.05
|
d3130c380c4b0621bbb9d9a990df91850bccb16b
|
comics/comics/crfh.py
|
comics/comics/crfh.py
|
from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = "Colleges Roomies from Hell"
language = "en"
url = "http://www.crfh.net/"
start_date = "1999-01-01"
rights = "Maritza Campos"
class Crawler(CrawlerBase):
history_capable_date = "1999-01-01"
time_zone = "America/Merida"
def crawl(self, pub_date):
page_url = "http://www.crfh.net/d2/%s.html" % (
pub_date.strftime("%Y%m%d"),
)
page = self.parse_page(page_url)
url = page.src('img[src*="crfh%s"]' % pub_date.strftime("%Y%m%d"))
url = url.replace("\n", "")
return CrawlerImage(url)
|
from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = "Colleges Roomies from Hell"
language = "en"
url = "http://www.crfh.net/"
start_date = "1999-01-01"
rights = "Maritza Campos"
class Crawler(CrawlerBase):
history_capable_date = "1999-01-01"
time_zone = "America/Merida"
def crawl(self, pub_date):
page_url = "http://www.crfh.net/d/%s.html" % (
pub_date.strftime("%Y%m%d"),
)
page = self.parse_page(page_url)
url = page.src('img[src*="crfh%s"]' % pub_date.strftime("%Y%m%d"))
url = url.replace("\n", "")
return CrawlerImage(url)
|
Update "Colleges Roomies from Hell" after site change
|
Update "Colleges Roomies from Hell" after site change
|
Python
|
agpl-3.0
|
datagutten/comics,datagutten/comics,datagutten/comics,jodal/comics,jodal/comics,jodal/comics,jodal/comics,datagutten/comics
|
from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = "Colleges Roomies from Hell"
language = "en"
url = "http://www.crfh.net/"
start_date = "1999-01-01"
rights = "Maritza Campos"
class Crawler(CrawlerBase):
history_capable_date = "1999-01-01"
time_zone = "America/Merida"
def crawl(self, pub_date):
page_url = "http://www.crfh.net/d2/%s.html" % (
pub_date.strftime("%Y%m%d"),
)
page = self.parse_page(page_url)
url = page.src('img[src*="crfh%s"]' % pub_date.strftime("%Y%m%d"))
url = url.replace("\n", "")
return CrawlerImage(url)
Update "Colleges Roomies from Hell" after site change
|
from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = "Colleges Roomies from Hell"
language = "en"
url = "http://www.crfh.net/"
start_date = "1999-01-01"
rights = "Maritza Campos"
class Crawler(CrawlerBase):
history_capable_date = "1999-01-01"
time_zone = "America/Merida"
def crawl(self, pub_date):
page_url = "http://www.crfh.net/d/%s.html" % (
pub_date.strftime("%Y%m%d"),
)
page = self.parse_page(page_url)
url = page.src('img[src*="crfh%s"]' % pub_date.strftime("%Y%m%d"))
url = url.replace("\n", "")
return CrawlerImage(url)
|
<commit_before>from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = "Colleges Roomies from Hell"
language = "en"
url = "http://www.crfh.net/"
start_date = "1999-01-01"
rights = "Maritza Campos"
class Crawler(CrawlerBase):
history_capable_date = "1999-01-01"
time_zone = "America/Merida"
def crawl(self, pub_date):
page_url = "http://www.crfh.net/d2/%s.html" % (
pub_date.strftime("%Y%m%d"),
)
page = self.parse_page(page_url)
url = page.src('img[src*="crfh%s"]' % pub_date.strftime("%Y%m%d"))
url = url.replace("\n", "")
return CrawlerImage(url)
<commit_msg>Update "Colleges Roomies from Hell" after site change<commit_after>
|
from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = "Colleges Roomies from Hell"
language = "en"
url = "http://www.crfh.net/"
start_date = "1999-01-01"
rights = "Maritza Campos"
class Crawler(CrawlerBase):
history_capable_date = "1999-01-01"
time_zone = "America/Merida"
def crawl(self, pub_date):
page_url = "http://www.crfh.net/d/%s.html" % (
pub_date.strftime("%Y%m%d"),
)
page = self.parse_page(page_url)
url = page.src('img[src*="crfh%s"]' % pub_date.strftime("%Y%m%d"))
url = url.replace("\n", "")
return CrawlerImage(url)
|
from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = "Colleges Roomies from Hell"
language = "en"
url = "http://www.crfh.net/"
start_date = "1999-01-01"
rights = "Maritza Campos"
class Crawler(CrawlerBase):
history_capable_date = "1999-01-01"
time_zone = "America/Merida"
def crawl(self, pub_date):
page_url = "http://www.crfh.net/d2/%s.html" % (
pub_date.strftime("%Y%m%d"),
)
page = self.parse_page(page_url)
url = page.src('img[src*="crfh%s"]' % pub_date.strftime("%Y%m%d"))
url = url.replace("\n", "")
return CrawlerImage(url)
Update "Colleges Roomies from Hell" after site changefrom comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = "Colleges Roomies from Hell"
language = "en"
url = "http://www.crfh.net/"
start_date = "1999-01-01"
rights = "Maritza Campos"
class Crawler(CrawlerBase):
history_capable_date = "1999-01-01"
time_zone = "America/Merida"
def crawl(self, pub_date):
page_url = "http://www.crfh.net/d/%s.html" % (
pub_date.strftime("%Y%m%d"),
)
page = self.parse_page(page_url)
url = page.src('img[src*="crfh%s"]' % pub_date.strftime("%Y%m%d"))
url = url.replace("\n", "")
return CrawlerImage(url)
|
<commit_before>from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = "Colleges Roomies from Hell"
language = "en"
url = "http://www.crfh.net/"
start_date = "1999-01-01"
rights = "Maritza Campos"
class Crawler(CrawlerBase):
history_capable_date = "1999-01-01"
time_zone = "America/Merida"
def crawl(self, pub_date):
page_url = "http://www.crfh.net/d2/%s.html" % (
pub_date.strftime("%Y%m%d"),
)
page = self.parse_page(page_url)
url = page.src('img[src*="crfh%s"]' % pub_date.strftime("%Y%m%d"))
url = url.replace("\n", "")
return CrawlerImage(url)
<commit_msg>Update "Colleges Roomies from Hell" after site change<commit_after>from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.core.comic_data import ComicDataBase
class ComicData(ComicDataBase):
name = "Colleges Roomies from Hell"
language = "en"
url = "http://www.crfh.net/"
start_date = "1999-01-01"
rights = "Maritza Campos"
class Crawler(CrawlerBase):
history_capable_date = "1999-01-01"
time_zone = "America/Merida"
def crawl(self, pub_date):
page_url = "http://www.crfh.net/d/%s.html" % (
pub_date.strftime("%Y%m%d"),
)
page = self.parse_page(page_url)
url = page.src('img[src*="crfh%s"]' % pub_date.strftime("%Y%m%d"))
url = url.replace("\n", "")
return CrawlerImage(url)
|
33f7e94385a8d4fbba797fc81b2565906604c9a4
|
src/zeit/content/cp/browser/area.py
|
src/zeit/content/cp/browser/area.py
|
# Copyright (c) 2009-2010 gocept gmbh & co. kg
# See also LICENSE.txt
import zeit.content.cp.browser.blocks.teaser
import zeit.content.cp.interfaces
import zeit.edit.browser.block
import zeit.edit.browser.view
import zope.formlib.form
class ViewletManager(zeit.edit.browser.block.BlockViewletManager):
@property
def css_class(self):
classes = super(ViewletManager, self).css_class
return ' '.join(['editable-area', classes])
class EditProperties(zeit.content.cp.browser.blocks.teaser.EditLayout):
interface = zeit.content.cp.interfaces.IArea
layout_prefix = 'teaserbar' # XXX should be area
layouts = ()
form_fields = zope.formlib.form.Fields() # XXX implement me
class EditCommon(zeit.edit.browser.view.EditBox):
form_fields = zope.formlib.form.Fields(
zeit.content.cp.interfaces.IArea).select(
'supertitle', 'teaserText', 'background_color')
form_fields['background_color'].custom_widget = (
zeit.cms.browser.widget.ColorpickerWidget)
|
# Copyright (c) 2009-2010 gocept gmbh & co. kg
# See also LICENSE.txt
import zeit.content.cp.browser.blocks.teaser
import zeit.content.cp.interfaces
import zeit.edit.browser.block
import zeit.edit.browser.view
import zope.formlib.form
class ViewletManager(zeit.edit.browser.block.BlockViewletManager):
@property
def css_class(self):
classes = super(ViewletManager, self).css_class
return ' '.join(['editable-area', classes])
class EditProperties(zeit.content.cp.browser.blocks.teaser.EditLayout):
interface = zeit.content.cp.interfaces.IArea
layout_prefix = 'teaserbar' # XXX should be area
layouts = ()
class EditCommon(zeit.edit.browser.view.EditBox):
form_fields = zope.formlib.form.Fields(
zeit.content.cp.interfaces.IArea).select(
'supertitle', 'teaserText', 'background_color')
form_fields['background_color'].custom_widget = (
zeit.cms.browser.widget.ColorpickerWidget)
|
Remove field that has now the same default implementation on it's super class.
|
Remove field that has now the same default implementation on it's super class.
|
Python
|
bsd-3-clause
|
ZeitOnline/zeit.content.cp,ZeitOnline/zeit.content.cp
|
# Copyright (c) 2009-2010 gocept gmbh & co. kg
# See also LICENSE.txt
import zeit.content.cp.browser.blocks.teaser
import zeit.content.cp.interfaces
import zeit.edit.browser.block
import zeit.edit.browser.view
import zope.formlib.form
class ViewletManager(zeit.edit.browser.block.BlockViewletManager):
@property
def css_class(self):
classes = super(ViewletManager, self).css_class
return ' '.join(['editable-area', classes])
class EditProperties(zeit.content.cp.browser.blocks.teaser.EditLayout):
interface = zeit.content.cp.interfaces.IArea
layout_prefix = 'teaserbar' # XXX should be area
layouts = ()
form_fields = zope.formlib.form.Fields() # XXX implement me
class EditCommon(zeit.edit.browser.view.EditBox):
form_fields = zope.formlib.form.Fields(
zeit.content.cp.interfaces.IArea).select(
'supertitle', 'teaserText', 'background_color')
form_fields['background_color'].custom_widget = (
zeit.cms.browser.widget.ColorpickerWidget)
Remove field that has now the same default implementation on it's super class.
|
# Copyright (c) 2009-2010 gocept gmbh & co. kg
# See also LICENSE.txt
import zeit.content.cp.browser.blocks.teaser
import zeit.content.cp.interfaces
import zeit.edit.browser.block
import zeit.edit.browser.view
import zope.formlib.form
class ViewletManager(zeit.edit.browser.block.BlockViewletManager):
@property
def css_class(self):
classes = super(ViewletManager, self).css_class
return ' '.join(['editable-area', classes])
class EditProperties(zeit.content.cp.browser.blocks.teaser.EditLayout):
interface = zeit.content.cp.interfaces.IArea
layout_prefix = 'teaserbar' # XXX should be area
layouts = ()
class EditCommon(zeit.edit.browser.view.EditBox):
form_fields = zope.formlib.form.Fields(
zeit.content.cp.interfaces.IArea).select(
'supertitle', 'teaserText', 'background_color')
form_fields['background_color'].custom_widget = (
zeit.cms.browser.widget.ColorpickerWidget)
|
<commit_before># Copyright (c) 2009-2010 gocept gmbh & co. kg
# See also LICENSE.txt
import zeit.content.cp.browser.blocks.teaser
import zeit.content.cp.interfaces
import zeit.edit.browser.block
import zeit.edit.browser.view
import zope.formlib.form
class ViewletManager(zeit.edit.browser.block.BlockViewletManager):
@property
def css_class(self):
classes = super(ViewletManager, self).css_class
return ' '.join(['editable-area', classes])
class EditProperties(zeit.content.cp.browser.blocks.teaser.EditLayout):
interface = zeit.content.cp.interfaces.IArea
layout_prefix = 'teaserbar' # XXX should be area
layouts = ()
form_fields = zope.formlib.form.Fields() # XXX implement me
class EditCommon(zeit.edit.browser.view.EditBox):
form_fields = zope.formlib.form.Fields(
zeit.content.cp.interfaces.IArea).select(
'supertitle', 'teaserText', 'background_color')
form_fields['background_color'].custom_widget = (
zeit.cms.browser.widget.ColorpickerWidget)
<commit_msg>Remove field that has now the same default implementation on it's super class.<commit_after>
|
# Copyright (c) 2009-2010 gocept gmbh & co. kg
# See also LICENSE.txt
import zeit.content.cp.browser.blocks.teaser
import zeit.content.cp.interfaces
import zeit.edit.browser.block
import zeit.edit.browser.view
import zope.formlib.form
class ViewletManager(zeit.edit.browser.block.BlockViewletManager):
@property
def css_class(self):
classes = super(ViewletManager, self).css_class
return ' '.join(['editable-area', classes])
class EditProperties(zeit.content.cp.browser.blocks.teaser.EditLayout):
interface = zeit.content.cp.interfaces.IArea
layout_prefix = 'teaserbar' # XXX should be area
layouts = ()
class EditCommon(zeit.edit.browser.view.EditBox):
form_fields = zope.formlib.form.Fields(
zeit.content.cp.interfaces.IArea).select(
'supertitle', 'teaserText', 'background_color')
form_fields['background_color'].custom_widget = (
zeit.cms.browser.widget.ColorpickerWidget)
|
# Copyright (c) 2009-2010 gocept gmbh & co. kg
# See also LICENSE.txt
import zeit.content.cp.browser.blocks.teaser
import zeit.content.cp.interfaces
import zeit.edit.browser.block
import zeit.edit.browser.view
import zope.formlib.form
class ViewletManager(zeit.edit.browser.block.BlockViewletManager):
@property
def css_class(self):
classes = super(ViewletManager, self).css_class
return ' '.join(['editable-area', classes])
class EditProperties(zeit.content.cp.browser.blocks.teaser.EditLayout):
interface = zeit.content.cp.interfaces.IArea
layout_prefix = 'teaserbar' # XXX should be area
layouts = ()
form_fields = zope.formlib.form.Fields() # XXX implement me
class EditCommon(zeit.edit.browser.view.EditBox):
form_fields = zope.formlib.form.Fields(
zeit.content.cp.interfaces.IArea).select(
'supertitle', 'teaserText', 'background_color')
form_fields['background_color'].custom_widget = (
zeit.cms.browser.widget.ColorpickerWidget)
Remove field that has now the same default implementation on it's super class.# Copyright (c) 2009-2010 gocept gmbh & co. kg
# See also LICENSE.txt
import zeit.content.cp.browser.blocks.teaser
import zeit.content.cp.interfaces
import zeit.edit.browser.block
import zeit.edit.browser.view
import zope.formlib.form
class ViewletManager(zeit.edit.browser.block.BlockViewletManager):
@property
def css_class(self):
classes = super(ViewletManager, self).css_class
return ' '.join(['editable-area', classes])
class EditProperties(zeit.content.cp.browser.blocks.teaser.EditLayout):
interface = zeit.content.cp.interfaces.IArea
layout_prefix = 'teaserbar' # XXX should be area
layouts = ()
class EditCommon(zeit.edit.browser.view.EditBox):
form_fields = zope.formlib.form.Fields(
zeit.content.cp.interfaces.IArea).select(
'supertitle', 'teaserText', 'background_color')
form_fields['background_color'].custom_widget = (
zeit.cms.browser.widget.ColorpickerWidget)
|
<commit_before># Copyright (c) 2009-2010 gocept gmbh & co. kg
# See also LICENSE.txt
import zeit.content.cp.browser.blocks.teaser
import zeit.content.cp.interfaces
import zeit.edit.browser.block
import zeit.edit.browser.view
import zope.formlib.form
class ViewletManager(zeit.edit.browser.block.BlockViewletManager):
@property
def css_class(self):
classes = super(ViewletManager, self).css_class
return ' '.join(['editable-area', classes])
class EditProperties(zeit.content.cp.browser.blocks.teaser.EditLayout):
interface = zeit.content.cp.interfaces.IArea
layout_prefix = 'teaserbar' # XXX should be area
layouts = ()
form_fields = zope.formlib.form.Fields() # XXX implement me
class EditCommon(zeit.edit.browser.view.EditBox):
form_fields = zope.formlib.form.Fields(
zeit.content.cp.interfaces.IArea).select(
'supertitle', 'teaserText', 'background_color')
form_fields['background_color'].custom_widget = (
zeit.cms.browser.widget.ColorpickerWidget)
<commit_msg>Remove field that has now the same default implementation on it's super class.<commit_after># Copyright (c) 2009-2010 gocept gmbh & co. kg
# See also LICENSE.txt
import zeit.content.cp.browser.blocks.teaser
import zeit.content.cp.interfaces
import zeit.edit.browser.block
import zeit.edit.browser.view
import zope.formlib.form
class ViewletManager(zeit.edit.browser.block.BlockViewletManager):
@property
def css_class(self):
classes = super(ViewletManager, self).css_class
return ' '.join(['editable-area', classes])
class EditProperties(zeit.content.cp.browser.blocks.teaser.EditLayout):
interface = zeit.content.cp.interfaces.IArea
layout_prefix = 'teaserbar' # XXX should be area
layouts = ()
class EditCommon(zeit.edit.browser.view.EditBox):
form_fields = zope.formlib.form.Fields(
zeit.content.cp.interfaces.IArea).select(
'supertitle', 'teaserText', 'background_color')
form_fields['background_color'].custom_widget = (
zeit.cms.browser.widget.ColorpickerWidget)
|
a802501943757dd85ce66a11fcd7ae40c0239462
|
datastructures.py
|
datastructures.py
|
#!/usr/bin/env python3
class Triangle:
"""A class structure for storing and minipulating a triangle.
The trianlge is represented as a 3-tuple of points. Each point is
represented as a 2-tuple of floats, the first element being the
x-coordinate and the second element being the y-coordinate.
Several useful operations can be applied to a triangle such as, rotate,
translate, split across altitude, and rectanglify.
The Triangle (and underlying tuple) should be treated as an immutable
data structure. All methods return a new triangle and do not modify the
existing one."""
def __init__(self, tpl):
"""tpl is a 3-tuple of coordinates"""
self.points = tpl
class Shape:
"""A class structure for representing and minipulating arbitary shapes.
A shape is defines as a list of triangles (see Triangle). Several
operations can be applied to a shape such as rotation, translation and
splitting the shape into two.
This object should be treated as an immutable data structure. All methods
return new shapes and do not modify the existing one."""
def __init__(self, triangle_list):
"""triangle_list is a list of triangles"""
self.triangles = triangle_list
|
#!/usr/bin/env python3
import math
class Triangle:
"""A class structure for storing and minipulating a triangle.
The trianlge is represented as a 3-tuple of points. Each point is
represented as a 2-tuple of floats, the first element being the
x-coordinate and the second element being the y-coordinate.
Several useful operations can be applied to a triangle such as, rotate,
translate, split across altitude, and rectanglify.
The Triangle (and underlying tuple) should be treated as an immutable
data structure. All methods return a new triangle and do not modify the
existing one."""
def __init__(self, tpl):
"""tpl is a 3-tuple of coordinates"""
self.points = tpl
def rotate(self, pivot, rangle):
"""Return a new triangle rotate clockwise (by angle) around pivot.
pivot -- A coordinate pair
rangle -- The angle to rotate by in radians"""
new_points = list()
px, py = pivot
for x, y in self.points:
dx, dy = x - px, y - py
current_angle = math.atan2(dy, dx)
total_angle = current_angle + rangle
r = math.hypot(dx, dy)
nx = r*math.cos(total_angle) + px
ny = r*math.sin(total_angle) + py
new_points.append((nx, ny))
return Triangle(tuple(new_points))
class Shape:
"""A class structure for representing and minipulating arbitary shapes.
A shape is defines as a list of triangles (see Triangle). Several
operations can be applied to a shape such as rotation, translation and
splitting the shape into two.
This object should be treated as an immutable data structure. All methods
return new shapes and do not modify the existing one."""
def __init__(self, triangle_list):
"""triangle_list is a list of triangles"""
self.triangles = triangle_list
|
Add method to rotate triangle
|
Add method to rotate triangle
|
Python
|
mit
|
moyamo/polygon2square
|
#!/usr/bin/env python3
class Triangle:
"""A class structure for storing and minipulating a triangle.
The trianlge is represented as a 3-tuple of points. Each point is
represented as a 2-tuple of floats, the first element being the
x-coordinate and the second element being the y-coordinate.
Several useful operations can be applied to a triangle such as, rotate,
translate, split across altitude, and rectanglify.
The Triangle (and underlying tuple) should be treated as an immutable
data structure. All methods return a new triangle and do not modify the
existing one."""
def __init__(self, tpl):
"""tpl is a 3-tuple of coordinates"""
self.points = tpl
class Shape:
"""A class structure for representing and minipulating arbitary shapes.
A shape is defines as a list of triangles (see Triangle). Several
operations can be applied to a shape such as rotation, translation and
splitting the shape into two.
This object should be treated as an immutable data structure. All methods
return new shapes and do not modify the existing one."""
def __init__(self, triangle_list):
"""triangle_list is a list of triangles"""
self.triangles = triangle_list
Add method to rotate triangle
|
#!/usr/bin/env python3
import math
class Triangle:
"""A class structure for storing and minipulating a triangle.
The trianlge is represented as a 3-tuple of points. Each point is
represented as a 2-tuple of floats, the first element being the
x-coordinate and the second element being the y-coordinate.
Several useful operations can be applied to a triangle such as, rotate,
translate, split across altitude, and rectanglify.
The Triangle (and underlying tuple) should be treated as an immutable
data structure. All methods return a new triangle and do not modify the
existing one."""
def __init__(self, tpl):
"""tpl is a 3-tuple of coordinates"""
self.points = tpl
def rotate(self, pivot, rangle):
"""Return a new triangle rotate clockwise (by angle) around pivot.
pivot -- A coordinate pair
rangle -- The angle to rotate by in radians"""
new_points = list()
px, py = pivot
for x, y in self.points:
dx, dy = x - px, y - py
current_angle = math.atan2(dy, dx)
total_angle = current_angle + rangle
r = math.hypot(dx, dy)
nx = r*math.cos(total_angle) + px
ny = r*math.sin(total_angle) + py
new_points.append((nx, ny))
return Triangle(tuple(new_points))
class Shape:
"""A class structure for representing and minipulating arbitary shapes.
A shape is defines as a list of triangles (see Triangle). Several
operations can be applied to a shape such as rotation, translation and
splitting the shape into two.
This object should be treated as an immutable data structure. All methods
return new shapes and do not modify the existing one."""
def __init__(self, triangle_list):
"""triangle_list is a list of triangles"""
self.triangles = triangle_list
|
<commit_before>#!/usr/bin/env python3
class Triangle:
"""A class structure for storing and minipulating a triangle.
The trianlge is represented as a 3-tuple of points. Each point is
represented as a 2-tuple of floats, the first element being the
x-coordinate and the second element being the y-coordinate.
Several useful operations can be applied to a triangle such as, rotate,
translate, split across altitude, and rectanglify.
The Triangle (and underlying tuple) should be treated as an immutable
data structure. All methods return a new triangle and do not modify the
existing one."""
def __init__(self, tpl):
"""tpl is a 3-tuple of coordinates"""
self.points = tpl
class Shape:
"""A class structure for representing and minipulating arbitary shapes.
A shape is defines as a list of triangles (see Triangle). Several
operations can be applied to a shape such as rotation, translation and
splitting the shape into two.
This object should be treated as an immutable data structure. All methods
return new shapes and do not modify the existing one."""
def __init__(self, triangle_list):
"""triangle_list is a list of triangles"""
self.triangles = triangle_list
<commit_msg>Add method to rotate triangle<commit_after>
|
#!/usr/bin/env python3
import math
class Triangle:
"""A class structure for storing and minipulating a triangle.
The trianlge is represented as a 3-tuple of points. Each point is
represented as a 2-tuple of floats, the first element being the
x-coordinate and the second element being the y-coordinate.
Several useful operations can be applied to a triangle such as, rotate,
translate, split across altitude, and rectanglify.
The Triangle (and underlying tuple) should be treated as an immutable
data structure. All methods return a new triangle and do not modify the
existing one."""
def __init__(self, tpl):
"""tpl is a 3-tuple of coordinates"""
self.points = tpl
def rotate(self, pivot, rangle):
"""Return a new triangle rotate clockwise (by angle) around pivot.
pivot -- A coordinate pair
rangle -- The angle to rotate by in radians"""
new_points = list()
px, py = pivot
for x, y in self.points:
dx, dy = x - px, y - py
current_angle = math.atan2(dy, dx)
total_angle = current_angle + rangle
r = math.hypot(dx, dy)
nx = r*math.cos(total_angle) + px
ny = r*math.sin(total_angle) + py
new_points.append((nx, ny))
return Triangle(tuple(new_points))
class Shape:
"""A class structure for representing and minipulating arbitary shapes.
A shape is defines as a list of triangles (see Triangle). Several
operations can be applied to a shape such as rotation, translation and
splitting the shape into two.
This object should be treated as an immutable data structure. All methods
return new shapes and do not modify the existing one."""
def __init__(self, triangle_list):
"""triangle_list is a list of triangles"""
self.triangles = triangle_list
|
#!/usr/bin/env python3
class Triangle:
"""A class structure for storing and minipulating a triangle.
The trianlge is represented as a 3-tuple of points. Each point is
represented as a 2-tuple of floats, the first element being the
x-coordinate and the second element being the y-coordinate.
Several useful operations can be applied to a triangle such as, rotate,
translate, split across altitude, and rectanglify.
The Triangle (and underlying tuple) should be treated as an immutable
data structure. All methods return a new triangle and do not modify the
existing one."""
def __init__(self, tpl):
"""tpl is a 3-tuple of coordinates"""
self.points = tpl
class Shape:
"""A class structure for representing and minipulating arbitary shapes.
A shape is defines as a list of triangles (see Triangle). Several
operations can be applied to a shape such as rotation, translation and
splitting the shape into two.
This object should be treated as an immutable data structure. All methods
return new shapes and do not modify the existing one."""
def __init__(self, triangle_list):
"""triangle_list is a list of triangles"""
self.triangles = triangle_list
Add method to rotate triangle#!/usr/bin/env python3
import math
class Triangle:
"""A class structure for storing and minipulating a triangle.
The trianlge is represented as a 3-tuple of points. Each point is
represented as a 2-tuple of floats, the first element being the
x-coordinate and the second element being the y-coordinate.
Several useful operations can be applied to a triangle such as, rotate,
translate, split across altitude, and rectanglify.
The Triangle (and underlying tuple) should be treated as an immutable
data structure. All methods return a new triangle and do not modify the
existing one."""
def __init__(self, tpl):
"""tpl is a 3-tuple of coordinates"""
self.points = tpl
def rotate(self, pivot, rangle):
"""Return a new triangle rotate clockwise (by angle) around pivot.
pivot -- A coordinate pair
rangle -- The angle to rotate by in radians"""
new_points = list()
px, py = pivot
for x, y in self.points:
dx, dy = x - px, y - py
current_angle = math.atan2(dy, dx)
total_angle = current_angle + rangle
r = math.hypot(dx, dy)
nx = r*math.cos(total_angle) + px
ny = r*math.sin(total_angle) + py
new_points.append((nx, ny))
return Triangle(tuple(new_points))
class Shape:
"""A class structure for representing and minipulating arbitary shapes.
A shape is defines as a list of triangles (see Triangle). Several
operations can be applied to a shape such as rotation, translation and
splitting the shape into two.
This object should be treated as an immutable data structure. All methods
return new shapes and do not modify the existing one."""
def __init__(self, triangle_list):
"""triangle_list is a list of triangles"""
self.triangles = triangle_list
|
<commit_before>#!/usr/bin/env python3
class Triangle:
"""A class structure for storing and minipulating a triangle.
The trianlge is represented as a 3-tuple of points. Each point is
represented as a 2-tuple of floats, the first element being the
x-coordinate and the second element being the y-coordinate.
Several useful operations can be applied to a triangle such as, rotate,
translate, split across altitude, and rectanglify.
The Triangle (and underlying tuple) should be treated as an immutable
data structure. All methods return a new triangle and do not modify the
existing one."""
def __init__(self, tpl):
"""tpl is a 3-tuple of coordinates"""
self.points = tpl
class Shape:
"""A class structure for representing and minipulating arbitary shapes.
A shape is defines as a list of triangles (see Triangle). Several
operations can be applied to a shape such as rotation, translation and
splitting the shape into two.
This object should be treated as an immutable data structure. All methods
return new shapes and do not modify the existing one."""
def __init__(self, triangle_list):
"""triangle_list is a list of triangles"""
self.triangles = triangle_list
<commit_msg>Add method to rotate triangle<commit_after>#!/usr/bin/env python3
import math
class Triangle:
"""A class structure for storing and minipulating a triangle.
The trianlge is represented as a 3-tuple of points. Each point is
represented as a 2-tuple of floats, the first element being the
x-coordinate and the second element being the y-coordinate.
Several useful operations can be applied to a triangle such as, rotate,
translate, split across altitude, and rectanglify.
The Triangle (and underlying tuple) should be treated as an immutable
data structure. All methods return a new triangle and do not modify the
existing one."""
def __init__(self, tpl):
"""tpl is a 3-tuple of coordinates"""
self.points = tpl
def rotate(self, pivot, rangle):
"""Return a new triangle rotate clockwise (by angle) around pivot.
pivot -- A coordinate pair
rangle -- The angle to rotate by in radians"""
new_points = list()
px, py = pivot
for x, y in self.points:
dx, dy = x - px, y - py
current_angle = math.atan2(dy, dx)
total_angle = current_angle + rangle
r = math.hypot(dx, dy)
nx = r*math.cos(total_angle) + px
ny = r*math.sin(total_angle) + py
new_points.append((nx, ny))
return Triangle(tuple(new_points))
class Shape:
"""A class structure for representing and minipulating arbitary shapes.
A shape is defines as a list of triangles (see Triangle). Several
operations can be applied to a shape such as rotation, translation and
splitting the shape into two.
This object should be treated as an immutable data structure. All methods
return new shapes and do not modify the existing one."""
def __init__(self, triangle_list):
"""triangle_list is a list of triangles"""
self.triangles = triangle_list
|
dfcb61ef1187f9d3cf80ffc55ad8aceafb0b29b3
|
djoauth2/helpers.py
|
djoauth2/helpers.py
|
# coding: utf-8
import random
import urlparse
from string import ascii_letters, digits
from urllib import urlencode
# From http://tools.ietf.org/html/rfc6750#section-2.1
BEARER_TOKEN_CHARSET = ascii_letters + digits + '-._~+/'
def random_hash(length):
return ''.join(random.sample(BEARER_TOKEN_CHARSET, length))
def random_hash_generator(length):
return lambda: random_hash(length)
def update_parameters(url, parameters):
""" Updates a URL's existing GET parameters.
@url: a URL string.
@parameters: a dictionary of parameters, {string:string}.
"""
parsed_url = urlparse(url)
query_parameters = urlparse.parse_qsl(parsed_url.query)
parsed_url.query = urlencode(query_parameters + parameters.items())
return urlparse.urlunparse(parsed_url)
|
# coding: utf-8
import random
import urlparse
from string import ascii_letters, digits
from urllib import urlencode
# From http://tools.ietf.org/html/rfc6750#section-2.1
BEARER_TOKEN_CHARSET = ascii_letters + digits + '-._~+/'
def random_hash(length):
return ''.join(random.sample(BEARER_TOKEN_CHARSET, length))
def random_hash_generator(length):
return lambda: random_hash(length)
def update_parameters(url, parameters):
""" Updates a URL's existing GET parameters.
@url: a URL string.
@parameters: a dictionary of parameters, {string:string}.
"""
parsed_url = urlparse.urlparse(url)
existing_query_parameters = urlparse.parse_qsl(parsed_url.query)
# Read http://docs.python.org/2/library/urlparse.html#urlparse.urlparse
# if this is confusing.
return urlparse.urlunparse((
parsed_url.scheme,
parsed_url.netloc,
parsed_url.path,
parsed_url.params,
urlencode(existing_query_parameters + parameters.items()),
parsed_url.fragment
))
|
Fix query string update helper.
|
Fix query string update helper.
|
Python
|
mit
|
seler/djoauth2,vden/djoauth2-ng,Locu/djoauth2,Locu/djoauth2,seler/djoauth2,vden/djoauth2-ng
|
# coding: utf-8
import random
import urlparse
from string import ascii_letters, digits
from urllib import urlencode
# From http://tools.ietf.org/html/rfc6750#section-2.1
BEARER_TOKEN_CHARSET = ascii_letters + digits + '-._~+/'
def random_hash(length):
return ''.join(random.sample(BEARER_TOKEN_CHARSET, length))
def random_hash_generator(length):
return lambda: random_hash(length)
def update_parameters(url, parameters):
""" Updates a URL's existing GET parameters.
@url: a URL string.
@parameters: a dictionary of parameters, {string:string}.
"""
parsed_url = urlparse(url)
query_parameters = urlparse.parse_qsl(parsed_url.query)
parsed_url.query = urlencode(query_parameters + parameters.items())
return urlparse.urlunparse(parsed_url)
Fix query string update helper.
|
# coding: utf-8
import random
import urlparse
from string import ascii_letters, digits
from urllib import urlencode
# From http://tools.ietf.org/html/rfc6750#section-2.1
BEARER_TOKEN_CHARSET = ascii_letters + digits + '-._~+/'
def random_hash(length):
return ''.join(random.sample(BEARER_TOKEN_CHARSET, length))
def random_hash_generator(length):
return lambda: random_hash(length)
def update_parameters(url, parameters):
""" Updates a URL's existing GET parameters.
@url: a URL string.
@parameters: a dictionary of parameters, {string:string}.
"""
parsed_url = urlparse.urlparse(url)
existing_query_parameters = urlparse.parse_qsl(parsed_url.query)
# Read http://docs.python.org/2/library/urlparse.html#urlparse.urlparse
# if this is confusing.
return urlparse.urlunparse((
parsed_url.scheme,
parsed_url.netloc,
parsed_url.path,
parsed_url.params,
urlencode(existing_query_parameters + parameters.items()),
parsed_url.fragment
))
|
<commit_before># coding: utf-8
import random
import urlparse
from string import ascii_letters, digits
from urllib import urlencode
# From http://tools.ietf.org/html/rfc6750#section-2.1
BEARER_TOKEN_CHARSET = ascii_letters + digits + '-._~+/'
def random_hash(length):
return ''.join(random.sample(BEARER_TOKEN_CHARSET, length))
def random_hash_generator(length):
return lambda: random_hash(length)
def update_parameters(url, parameters):
""" Updates a URL's existing GET parameters.
@url: a URL string.
@parameters: a dictionary of parameters, {string:string}.
"""
parsed_url = urlparse(url)
query_parameters = urlparse.parse_qsl(parsed_url.query)
parsed_url.query = urlencode(query_parameters + parameters.items())
return urlparse.urlunparse(parsed_url)
<commit_msg>Fix query string update helper.<commit_after>
|
# coding: utf-8
import random
import urlparse
from string import ascii_letters, digits
from urllib import urlencode
# From http://tools.ietf.org/html/rfc6750#section-2.1
BEARER_TOKEN_CHARSET = ascii_letters + digits + '-._~+/'
def random_hash(length):
return ''.join(random.sample(BEARER_TOKEN_CHARSET, length))
def random_hash_generator(length):
return lambda: random_hash(length)
def update_parameters(url, parameters):
""" Updates a URL's existing GET parameters.
@url: a URL string.
@parameters: a dictionary of parameters, {string:string}.
"""
parsed_url = urlparse.urlparse(url)
existing_query_parameters = urlparse.parse_qsl(parsed_url.query)
# Read http://docs.python.org/2/library/urlparse.html#urlparse.urlparse
# if this is confusing.
return urlparse.urlunparse((
parsed_url.scheme,
parsed_url.netloc,
parsed_url.path,
parsed_url.params,
urlencode(existing_query_parameters + parameters.items()),
parsed_url.fragment
))
|
# coding: utf-8
import random
import urlparse
from string import ascii_letters, digits
from urllib import urlencode
# From http://tools.ietf.org/html/rfc6750#section-2.1
BEARER_TOKEN_CHARSET = ascii_letters + digits + '-._~+/'
def random_hash(length):
return ''.join(random.sample(BEARER_TOKEN_CHARSET, length))
def random_hash_generator(length):
return lambda: random_hash(length)
def update_parameters(url, parameters):
""" Updates a URL's existing GET parameters.
@url: a URL string.
@parameters: a dictionary of parameters, {string:string}.
"""
parsed_url = urlparse(url)
query_parameters = urlparse.parse_qsl(parsed_url.query)
parsed_url.query = urlencode(query_parameters + parameters.items())
return urlparse.urlunparse(parsed_url)
Fix query string update helper.# coding: utf-8
import random
import urlparse
from string import ascii_letters, digits
from urllib import urlencode
# From http://tools.ietf.org/html/rfc6750#section-2.1
BEARER_TOKEN_CHARSET = ascii_letters + digits + '-._~+/'
def random_hash(length):
return ''.join(random.sample(BEARER_TOKEN_CHARSET, length))
def random_hash_generator(length):
return lambda: random_hash(length)
def update_parameters(url, parameters):
""" Updates a URL's existing GET parameters.
@url: a URL string.
@parameters: a dictionary of parameters, {string:string}.
"""
parsed_url = urlparse.urlparse(url)
existing_query_parameters = urlparse.parse_qsl(parsed_url.query)
# Read http://docs.python.org/2/library/urlparse.html#urlparse.urlparse
# if this is confusing.
return urlparse.urlunparse((
parsed_url.scheme,
parsed_url.netloc,
parsed_url.path,
parsed_url.params,
urlencode(existing_query_parameters + parameters.items()),
parsed_url.fragment
))
|
<commit_before># coding: utf-8
import random
import urlparse
from string import ascii_letters, digits
from urllib import urlencode
# From http://tools.ietf.org/html/rfc6750#section-2.1
BEARER_TOKEN_CHARSET = ascii_letters + digits + '-._~+/'
def random_hash(length):
return ''.join(random.sample(BEARER_TOKEN_CHARSET, length))
def random_hash_generator(length):
return lambda: random_hash(length)
def update_parameters(url, parameters):
""" Updates a URL's existing GET parameters.
@url: a URL string.
@parameters: a dictionary of parameters, {string:string}.
"""
parsed_url = urlparse(url)
query_parameters = urlparse.parse_qsl(parsed_url.query)
parsed_url.query = urlencode(query_parameters + parameters.items())
return urlparse.urlunparse(parsed_url)
<commit_msg>Fix query string update helper.<commit_after># coding: utf-8
import random
import urlparse
from string import ascii_letters, digits
from urllib import urlencode
# From http://tools.ietf.org/html/rfc6750#section-2.1
BEARER_TOKEN_CHARSET = ascii_letters + digits + '-._~+/'
def random_hash(length):
return ''.join(random.sample(BEARER_TOKEN_CHARSET, length))
def random_hash_generator(length):
return lambda: random_hash(length)
def update_parameters(url, parameters):
""" Updates a URL's existing GET parameters.
@url: a URL string.
@parameters: a dictionary of parameters, {string:string}.
"""
parsed_url = urlparse.urlparse(url)
existing_query_parameters = urlparse.parse_qsl(parsed_url.query)
# Read http://docs.python.org/2/library/urlparse.html#urlparse.urlparse
# if this is confusing.
return urlparse.urlunparse((
parsed_url.scheme,
parsed_url.netloc,
parsed_url.path,
parsed_url.params,
urlencode(existing_query_parameters + parameters.items()),
parsed_url.fragment
))
|
986901c9e91d44758200fb8d3264b88c0977be37
|
lvsr/configs/timit_bothgru_hybrid2.py
|
lvsr/configs/timit_bothgru_hybrid2.py
|
Config(
net=Config(attention_type='hybrid2',
shift_predictor_dims=[100],
max_left=10,
max_right=100),
initialization=[
("/recognizer", "rec_weights_init", "IsotropicGaussian(0.1)"),
("/recognizer/generator/att_trans/hybrid_att/loc_att",
"weights_init", "IsotropicGaussian(0.01)"),
("/recognizer/generator/att_trans/hybrid_att/loc_att",
"biases_init", "IsotropicGaussian(5.0)")],
data=Config(normalization="norm.pkl"))
|
Config(
net=Config(dec_transition='GatedRecurrent',
enc_transition='GatedRecurrent',
attention_type='hybrid2',
shift_predictor_dims=[100],
max_left=10,
max_right=100),
initialization=[
("/recognizer", "rec_weights_init", "IsotropicGaussian(0.1)"),
("/recognizer/generator/att_trans/hybrid_att/loc_att",
"weights_init", "IsotropicGaussian(0.001)"),
("/recognizer/generator/att_trans/hybrid_att/loc_att",
"biases_init", "IsotropicGaussian(5.0)")],
data=Config(normalization="norm.pkl"))
|
Fix hybrid2, but it is still no use
|
Fix hybrid2, but it is still no use
|
Python
|
mit
|
nke001/attention-lvcsr,rizar/attention-lvcsr,rizar/attention-lvcsr,nke001/attention-lvcsr,nke001/attention-lvcsr,rizar/attention-lvcsr,nke001/attention-lvcsr,rizar/attention-lvcsr,rizar/attention-lvcsr,nke001/attention-lvcsr
|
Config(
net=Config(attention_type='hybrid2',
shift_predictor_dims=[100],
max_left=10,
max_right=100),
initialization=[
("/recognizer", "rec_weights_init", "IsotropicGaussian(0.1)"),
("/recognizer/generator/att_trans/hybrid_att/loc_att",
"weights_init", "IsotropicGaussian(0.01)"),
("/recognizer/generator/att_trans/hybrid_att/loc_att",
"biases_init", "IsotropicGaussian(5.0)")],
data=Config(normalization="norm.pkl"))
Fix hybrid2, but it is still no use
|
Config(
net=Config(dec_transition='GatedRecurrent',
enc_transition='GatedRecurrent',
attention_type='hybrid2',
shift_predictor_dims=[100],
max_left=10,
max_right=100),
initialization=[
("/recognizer", "rec_weights_init", "IsotropicGaussian(0.1)"),
("/recognizer/generator/att_trans/hybrid_att/loc_att",
"weights_init", "IsotropicGaussian(0.001)"),
("/recognizer/generator/att_trans/hybrid_att/loc_att",
"biases_init", "IsotropicGaussian(5.0)")],
data=Config(normalization="norm.pkl"))
|
<commit_before>Config(
net=Config(attention_type='hybrid2',
shift_predictor_dims=[100],
max_left=10,
max_right=100),
initialization=[
("/recognizer", "rec_weights_init", "IsotropicGaussian(0.1)"),
("/recognizer/generator/att_trans/hybrid_att/loc_att",
"weights_init", "IsotropicGaussian(0.01)"),
("/recognizer/generator/att_trans/hybrid_att/loc_att",
"biases_init", "IsotropicGaussian(5.0)")],
data=Config(normalization="norm.pkl"))
<commit_msg>Fix hybrid2, but it is still no use<commit_after>
|
Config(
net=Config(dec_transition='GatedRecurrent',
enc_transition='GatedRecurrent',
attention_type='hybrid2',
shift_predictor_dims=[100],
max_left=10,
max_right=100),
initialization=[
("/recognizer", "rec_weights_init", "IsotropicGaussian(0.1)"),
("/recognizer/generator/att_trans/hybrid_att/loc_att",
"weights_init", "IsotropicGaussian(0.001)"),
("/recognizer/generator/att_trans/hybrid_att/loc_att",
"biases_init", "IsotropicGaussian(5.0)")],
data=Config(normalization="norm.pkl"))
|
Config(
net=Config(attention_type='hybrid2',
shift_predictor_dims=[100],
max_left=10,
max_right=100),
initialization=[
("/recognizer", "rec_weights_init", "IsotropicGaussian(0.1)"),
("/recognizer/generator/att_trans/hybrid_att/loc_att",
"weights_init", "IsotropicGaussian(0.01)"),
("/recognizer/generator/att_trans/hybrid_att/loc_att",
"biases_init", "IsotropicGaussian(5.0)")],
data=Config(normalization="norm.pkl"))
Fix hybrid2, but it is still no useConfig(
net=Config(dec_transition='GatedRecurrent',
enc_transition='GatedRecurrent',
attention_type='hybrid2',
shift_predictor_dims=[100],
max_left=10,
max_right=100),
initialization=[
("/recognizer", "rec_weights_init", "IsotropicGaussian(0.1)"),
("/recognizer/generator/att_trans/hybrid_att/loc_att",
"weights_init", "IsotropicGaussian(0.001)"),
("/recognizer/generator/att_trans/hybrid_att/loc_att",
"biases_init", "IsotropicGaussian(5.0)")],
data=Config(normalization="norm.pkl"))
|
<commit_before>Config(
net=Config(attention_type='hybrid2',
shift_predictor_dims=[100],
max_left=10,
max_right=100),
initialization=[
("/recognizer", "rec_weights_init", "IsotropicGaussian(0.1)"),
("/recognizer/generator/att_trans/hybrid_att/loc_att",
"weights_init", "IsotropicGaussian(0.01)"),
("/recognizer/generator/att_trans/hybrid_att/loc_att",
"biases_init", "IsotropicGaussian(5.0)")],
data=Config(normalization="norm.pkl"))
<commit_msg>Fix hybrid2, but it is still no use<commit_after>Config(
net=Config(dec_transition='GatedRecurrent',
enc_transition='GatedRecurrent',
attention_type='hybrid2',
shift_predictor_dims=[100],
max_left=10,
max_right=100),
initialization=[
("/recognizer", "rec_weights_init", "IsotropicGaussian(0.1)"),
("/recognizer/generator/att_trans/hybrid_att/loc_att",
"weights_init", "IsotropicGaussian(0.001)"),
("/recognizer/generator/att_trans/hybrid_att/loc_att",
"biases_init", "IsotropicGaussian(5.0)")],
data=Config(normalization="norm.pkl"))
|
d48946c89b4436fad97fdee65e34d7ca77f58d95
|
modules/base.py
|
modules/base.py
|
#-*- coding: utf-8 -*-
import pandas as pd
import pandas_datareader.data as web
import datetime
import config
import os
import re
import pickle
def get_file_path(code):
return os.path.join(config.DATA_PATH, 'data', code + '.pkl')
def download(code, year1, month1, day1, year2, month2, day2):
start = datetime.datetime(year1, month1, day1)
end = datetime.datetime(year2, month2, day2)
df = web.DataReader('%s.KS' % code, 'yahoo', start, end)
save(code, df)
return df
def load(code):
try:
return pd.read_pickle(code)
except:
pass
return None
def save(code, df):
df.to_pickle(code)
def dump(code, df):
with open(get_file_path(code), 'wb') as handle:
pickle.dump(df, handle)
|
#-*- coding: utf-8 -*-
import pandas as pd
import pandas_datareader.data as web
import datetime
import config
import os
import re
import pickle
def get_file_path(code):
if not os.path.exists(config.DATA_PATH):
try:
os.makedirs(config.DATA_PATH)
except:
pass
return os.path.join(config.DATA_PATH, 'data', code + '.pkl')
def download(code, year1, month1, day1, year2, month2, day2):
start = datetime.datetime(year1, month1, day1)
end = datetime.datetime(year2, month2, day2)
df = web.DataReader('%s.KS' % code, 'yahoo', start, end)
save(code, df)
return df
def load(code):
try:
return pd.read_pickle(code)
except:
pass
return None
def save(code, df):
df.to_pickle(code)
def dump(code, df):
with open(get_file_path(code), 'wb') as handle:
pickle.dump(df, handle)
|
Fix the FileNotFoundError when data director is not exist
|
Fix the FileNotFoundError when data director is not exist
|
Python
|
mit
|
jongha/stock-ai,jongha/stock-ai,jongha/stock-ai,jongha/stock-ai
|
#-*- coding: utf-8 -*-
import pandas as pd
import pandas_datareader.data as web
import datetime
import config
import os
import re
import pickle
def get_file_path(code):
return os.path.join(config.DATA_PATH, 'data', code + '.pkl')
def download(code, year1, month1, day1, year2, month2, day2):
start = datetime.datetime(year1, month1, day1)
end = datetime.datetime(year2, month2, day2)
df = web.DataReader('%s.KS' % code, 'yahoo', start, end)
save(code, df)
return df
def load(code):
try:
return pd.read_pickle(code)
except:
pass
return None
def save(code, df):
df.to_pickle(code)
def dump(code, df):
with open(get_file_path(code), 'wb') as handle:
pickle.dump(df, handle)
Fix the FileNotFoundError when data director is not exist
|
#-*- coding: utf-8 -*-
import pandas as pd
import pandas_datareader.data as web
import datetime
import config
import os
import re
import pickle
def get_file_path(code):
if not os.path.exists(config.DATA_PATH):
try:
os.makedirs(config.DATA_PATH)
except:
pass
return os.path.join(config.DATA_PATH, 'data', code + '.pkl')
def download(code, year1, month1, day1, year2, month2, day2):
start = datetime.datetime(year1, month1, day1)
end = datetime.datetime(year2, month2, day2)
df = web.DataReader('%s.KS' % code, 'yahoo', start, end)
save(code, df)
return df
def load(code):
try:
return pd.read_pickle(code)
except:
pass
return None
def save(code, df):
df.to_pickle(code)
def dump(code, df):
with open(get_file_path(code), 'wb') as handle:
pickle.dump(df, handle)
|
<commit_before>#-*- coding: utf-8 -*-
import pandas as pd
import pandas_datareader.data as web
import datetime
import config
import os
import re
import pickle
def get_file_path(code):
return os.path.join(config.DATA_PATH, 'data', code + '.pkl')
def download(code, year1, month1, day1, year2, month2, day2):
start = datetime.datetime(year1, month1, day1)
end = datetime.datetime(year2, month2, day2)
df = web.DataReader('%s.KS' % code, 'yahoo', start, end)
save(code, df)
return df
def load(code):
try:
return pd.read_pickle(code)
except:
pass
return None
def save(code, df):
df.to_pickle(code)
def dump(code, df):
with open(get_file_path(code), 'wb') as handle:
pickle.dump(df, handle)
<commit_msg>Fix the FileNotFoundError when data director is not exist<commit_after>
|
#-*- coding: utf-8 -*-
import pandas as pd
import pandas_datareader.data as web
import datetime
import config
import os
import re
import pickle
def get_file_path(code):
if not os.path.exists(config.DATA_PATH):
try:
os.makedirs(config.DATA_PATH)
except:
pass
return os.path.join(config.DATA_PATH, 'data', code + '.pkl')
def download(code, year1, month1, day1, year2, month2, day2):
start = datetime.datetime(year1, month1, day1)
end = datetime.datetime(year2, month2, day2)
df = web.DataReader('%s.KS' % code, 'yahoo', start, end)
save(code, df)
return df
def load(code):
try:
return pd.read_pickle(code)
except:
pass
return None
def save(code, df):
df.to_pickle(code)
def dump(code, df):
with open(get_file_path(code), 'wb') as handle:
pickle.dump(df, handle)
|
#-*- coding: utf-8 -*-
import pandas as pd
import pandas_datareader.data as web
import datetime
import config
import os
import re
import pickle
def get_file_path(code):
return os.path.join(config.DATA_PATH, 'data', code + '.pkl')
def download(code, year1, month1, day1, year2, month2, day2):
start = datetime.datetime(year1, month1, day1)
end = datetime.datetime(year2, month2, day2)
df = web.DataReader('%s.KS' % code, 'yahoo', start, end)
save(code, df)
return df
def load(code):
try:
return pd.read_pickle(code)
except:
pass
return None
def save(code, df):
df.to_pickle(code)
def dump(code, df):
with open(get_file_path(code), 'wb') as handle:
pickle.dump(df, handle)
Fix the FileNotFoundError when data director is not exist#-*- coding: utf-8 -*-
import pandas as pd
import pandas_datareader.data as web
import datetime
import config
import os
import re
import pickle
def get_file_path(code):
if not os.path.exists(config.DATA_PATH):
try:
os.makedirs(config.DATA_PATH)
except:
pass
return os.path.join(config.DATA_PATH, 'data', code + '.pkl')
def download(code, year1, month1, day1, year2, month2, day2):
start = datetime.datetime(year1, month1, day1)
end = datetime.datetime(year2, month2, day2)
df = web.DataReader('%s.KS' % code, 'yahoo', start, end)
save(code, df)
return df
def load(code):
try:
return pd.read_pickle(code)
except:
pass
return None
def save(code, df):
df.to_pickle(code)
def dump(code, df):
with open(get_file_path(code), 'wb') as handle:
pickle.dump(df, handle)
|
<commit_before>#-*- coding: utf-8 -*-
import pandas as pd
import pandas_datareader.data as web
import datetime
import config
import os
import re
import pickle
def get_file_path(code):
return os.path.join(config.DATA_PATH, 'data', code + '.pkl')
def download(code, year1, month1, day1, year2, month2, day2):
start = datetime.datetime(year1, month1, day1)
end = datetime.datetime(year2, month2, day2)
df = web.DataReader('%s.KS' % code, 'yahoo', start, end)
save(code, df)
return df
def load(code):
try:
return pd.read_pickle(code)
except:
pass
return None
def save(code, df):
df.to_pickle(code)
def dump(code, df):
with open(get_file_path(code), 'wb') as handle:
pickle.dump(df, handle)
<commit_msg>Fix the FileNotFoundError when data director is not exist<commit_after>#-*- coding: utf-8 -*-
import pandas as pd
import pandas_datareader.data as web
import datetime
import config
import os
import re
import pickle
def get_file_path(code):
if not os.path.exists(config.DATA_PATH):
try:
os.makedirs(config.DATA_PATH)
except:
pass
return os.path.join(config.DATA_PATH, 'data', code + '.pkl')
def download(code, year1, month1, day1, year2, month2, day2):
start = datetime.datetime(year1, month1, day1)
end = datetime.datetime(year2, month2, day2)
df = web.DataReader('%s.KS' % code, 'yahoo', start, end)
save(code, df)
return df
def load(code):
try:
return pd.read_pickle(code)
except:
pass
return None
def save(code, df):
df.to_pickle(code)
def dump(code, df):
with open(get_file_path(code), 'wb') as handle:
pickle.dump(df, handle)
|
785c154cb97dcf8bbdc9c3ad5d4da6049bf7155c
|
web_blog.py
|
web_blog.py
|
# -*- coding: utf-8 -*-
from flask import Flask
DB_SCHEMA = """
DROP TABLE IF EXISTS entries;
CREATE TABLE entries (
id serial PRIMARY KEY,
title VARCHAR(127) NOT NULL,
text VARCHAR(10000) NOT NULL,
created TIMESTAMP NOT NULL,
)
"""
app = FLask(__name__)
@app.route('/')
def hello():
return u'Hello World!'
if __name__ == '__main__':
app.run(debug=True)
|
# -*- coding: utf-8 -*-
from flask import Flask
import os
import psycopg2
from contextlib import closing
DB_SCHEMA = """
DROP TABLE IF EXISTS entries;
CREATE TABLE entries (
id serial PRIMARY KEY,
title VARCHAR(127) NOT NULL,
text VARCHAR(10000) NOT NULL,
created TIMESTAMP NOT NULL
)
"""
app = Flask(__name__)
app.config['DATABASE'] = os.environ.get(
'DATABASE_URL', 'dbname=web_blog user=store'
)
def connect_db():
"""Return a connection to the database"""
return psycopg2.connect(app.config['DATABASE'])
def init_db():
"""Initialize the database
WARNING: executing this function will drop existing tables.
"""
with closing(connect_db()) as db:
db.cursor().execute(DB_SCHEMA)
db.commit()
@app.route('/')
def hello():
return u'Hello World!'
if __name__ == '__main__':
app.run(debug=True)
|
Add functionality to connect to DB server and initialize our blog's database
|
Add functionality to connect to DB server and initialize our blog's database
|
Python
|
mit
|
charlieRode/web_blog
|
# -*- coding: utf-8 -*-
from flask import Flask
DB_SCHEMA = """
DROP TABLE IF EXISTS entries;
CREATE TABLE entries (
id serial PRIMARY KEY,
title VARCHAR(127) NOT NULL,
text VARCHAR(10000) NOT NULL,
created TIMESTAMP NOT NULL,
)
"""
app = FLask(__name__)
@app.route('/')
def hello():
return u'Hello World!'
if __name__ == '__main__':
app.run(debug=True)
Add functionality to connect to DB server and initialize our blog's database
|
# -*- coding: utf-8 -*-
from flask import Flask
import os
import psycopg2
from contextlib import closing
DB_SCHEMA = """
DROP TABLE IF EXISTS entries;
CREATE TABLE entries (
id serial PRIMARY KEY,
title VARCHAR(127) NOT NULL,
text VARCHAR(10000) NOT NULL,
created TIMESTAMP NOT NULL
)
"""
app = Flask(__name__)
app.config['DATABASE'] = os.environ.get(
'DATABASE_URL', 'dbname=web_blog user=store'
)
def connect_db():
"""Return a connection to the database"""
return psycopg2.connect(app.config['DATABASE'])
def init_db():
"""Initialize the database
WARNING: executing this function will drop existing tables.
"""
with closing(connect_db()) as db:
db.cursor().execute(DB_SCHEMA)
db.commit()
@app.route('/')
def hello():
return u'Hello World!'
if __name__ == '__main__':
app.run(debug=True)
|
<commit_before># -*- coding: utf-8 -*-
from flask import Flask
DB_SCHEMA = """
DROP TABLE IF EXISTS entries;
CREATE TABLE entries (
id serial PRIMARY KEY,
title VARCHAR(127) NOT NULL,
text VARCHAR(10000) NOT NULL,
created TIMESTAMP NOT NULL,
)
"""
app = FLask(__name__)
@app.route('/')
def hello():
return u'Hello World!'
if __name__ == '__main__':
app.run(debug=True)
<commit_msg>Add functionality to connect to DB server and initialize our blog's database<commit_after>
|
# -*- coding: utf-8 -*-
from flask import Flask
import os
import psycopg2
from contextlib import closing
DB_SCHEMA = """
DROP TABLE IF EXISTS entries;
CREATE TABLE entries (
id serial PRIMARY KEY,
title VARCHAR(127) NOT NULL,
text VARCHAR(10000) NOT NULL,
created TIMESTAMP NOT NULL
)
"""
app = Flask(__name__)
app.config['DATABASE'] = os.environ.get(
'DATABASE_URL', 'dbname=web_blog user=store'
)
def connect_db():
"""Return a connection to the database"""
return psycopg2.connect(app.config['DATABASE'])
def init_db():
"""Initialize the database
WARNING: executing this function will drop existing tables.
"""
with closing(connect_db()) as db:
db.cursor().execute(DB_SCHEMA)
db.commit()
@app.route('/')
def hello():
return u'Hello World!'
if __name__ == '__main__':
app.run(debug=True)
|
# -*- coding: utf-8 -*-
from flask import Flask
DB_SCHEMA = """
DROP TABLE IF EXISTS entries;
CREATE TABLE entries (
id serial PRIMARY KEY,
title VARCHAR(127) NOT NULL,
text VARCHAR(10000) NOT NULL,
created TIMESTAMP NOT NULL,
)
"""
app = FLask(__name__)
@app.route('/')
def hello():
return u'Hello World!'
if __name__ == '__main__':
app.run(debug=True)
Add functionality to connect to DB server and initialize our blog's database# -*- coding: utf-8 -*-
from flask import Flask
import os
import psycopg2
from contextlib import closing
DB_SCHEMA = """
DROP TABLE IF EXISTS entries;
CREATE TABLE entries (
id serial PRIMARY KEY,
title VARCHAR(127) NOT NULL,
text VARCHAR(10000) NOT NULL,
created TIMESTAMP NOT NULL
)
"""
app = Flask(__name__)
app.config['DATABASE'] = os.environ.get(
'DATABASE_URL', 'dbname=web_blog user=store'
)
def connect_db():
"""Return a connection to the database"""
return psycopg2.connect(app.config['DATABASE'])
def init_db():
"""Initialize the database
WARNING: executing this function will drop existing tables.
"""
with closing(connect_db()) as db:
db.cursor().execute(DB_SCHEMA)
db.commit()
@app.route('/')
def hello():
return u'Hello World!'
if __name__ == '__main__':
app.run(debug=True)
|
<commit_before># -*- coding: utf-8 -*-
from flask import Flask
DB_SCHEMA = """
DROP TABLE IF EXISTS entries;
CREATE TABLE entries (
id serial PRIMARY KEY,
title VARCHAR(127) NOT NULL,
text VARCHAR(10000) NOT NULL,
created TIMESTAMP NOT NULL,
)
"""
app = FLask(__name__)
@app.route('/')
def hello():
return u'Hello World!'
if __name__ == '__main__':
app.run(debug=True)
<commit_msg>Add functionality to connect to DB server and initialize our blog's database<commit_after># -*- coding: utf-8 -*-
from flask import Flask
import os
import psycopg2
from contextlib import closing
DB_SCHEMA = """
DROP TABLE IF EXISTS entries;
CREATE TABLE entries (
id serial PRIMARY KEY,
title VARCHAR(127) NOT NULL,
text VARCHAR(10000) NOT NULL,
created TIMESTAMP NOT NULL
)
"""
app = Flask(__name__)
app.config['DATABASE'] = os.environ.get(
'DATABASE_URL', 'dbname=web_blog user=store'
)
def connect_db():
"""Return a connection to the database"""
return psycopg2.connect(app.config['DATABASE'])
def init_db():
"""Initialize the database
WARNING: executing this function will drop existing tables.
"""
with closing(connect_db()) as db:
db.cursor().execute(DB_SCHEMA)
db.commit()
@app.route('/')
def hello():
return u'Hello World!'
if __name__ == '__main__':
app.run(debug=True)
|
ae5c29e06ce110de1c44ffc4c466a4c611007d22
|
spyder_unittest/widgets/tests/__init__.py
|
spyder_unittest/widgets/tests/__init__.py
|
# -*- coding: utf-8 -*-
#
# Copyright © 2017 Spyder Project Contributors
# Licensed under the terms of the MIT License
# (see LICENSE.txt for details)
# noqa: D104
|
Add copyright notice, skip docstring checks
|
Add copyright notice, skip docstring checks
|
Python
|
mit
|
jitseniesen/spyder-unittest
|
Add copyright notice, skip docstring checks
|
# -*- coding: utf-8 -*-
#
# Copyright © 2017 Spyder Project Contributors
# Licensed under the terms of the MIT License
# (see LICENSE.txt for details)
# noqa: D104
|
<commit_before><commit_msg>Add copyright notice, skip docstring checks<commit_after>
|
# -*- coding: utf-8 -*-
#
# Copyright © 2017 Spyder Project Contributors
# Licensed under the terms of the MIT License
# (see LICENSE.txt for details)
# noqa: D104
|
Add copyright notice, skip docstring checks# -*- coding: utf-8 -*-
#
# Copyright © 2017 Spyder Project Contributors
# Licensed under the terms of the MIT License
# (see LICENSE.txt for details)
# noqa: D104
|
<commit_before><commit_msg>Add copyright notice, skip docstring checks<commit_after># -*- coding: utf-8 -*-
#
# Copyright © 2017 Spyder Project Contributors
# Licensed under the terms of the MIT License
# (see LICENSE.txt for details)
# noqa: D104
|
|
9492441a3587f7257d6295ebcb93a3e20c16a1d2
|
opensrs/models.py
|
opensrs/models.py
|
from dateutil.parser import parse
class Domain(object):
def __init__(self, data):
self.name = data['name']
self.auto_renew = (data['f_auto_renew'] == 'Y')
self.expiry_date = parse(data['expiredate']).date()
@property
def tld(self):
return self.name.rsplit('.', 1)[1]
|
from dateutil.parser import parse
class Domain(object):
def __init__(self, data):
self.name = data['name']
self.auto_renew = (data['f_auto_renew'] == 'Y')
self.expiry_date = parse(data['expiredate']).date()
@property
def tld(self):
return self.name.split('.')[-1]
|
Use more concise way to get tld
|
Use more concise way to get tld
|
Python
|
mit
|
yola/opensrs,yola/opensrs
|
from dateutil.parser import parse
class Domain(object):
def __init__(self, data):
self.name = data['name']
self.auto_renew = (data['f_auto_renew'] == 'Y')
self.expiry_date = parse(data['expiredate']).date()
@property
def tld(self):
return self.name.rsplit('.', 1)[1]
Use more concise way to get tld
|
from dateutil.parser import parse
class Domain(object):
def __init__(self, data):
self.name = data['name']
self.auto_renew = (data['f_auto_renew'] == 'Y')
self.expiry_date = parse(data['expiredate']).date()
@property
def tld(self):
return self.name.split('.')[-1]
|
<commit_before>from dateutil.parser import parse
class Domain(object):
def __init__(self, data):
self.name = data['name']
self.auto_renew = (data['f_auto_renew'] == 'Y')
self.expiry_date = parse(data['expiredate']).date()
@property
def tld(self):
return self.name.rsplit('.', 1)[1]
<commit_msg>Use more concise way to get tld<commit_after>
|
from dateutil.parser import parse
class Domain(object):
def __init__(self, data):
self.name = data['name']
self.auto_renew = (data['f_auto_renew'] == 'Y')
self.expiry_date = parse(data['expiredate']).date()
@property
def tld(self):
return self.name.split('.')[-1]
|
from dateutil.parser import parse
class Domain(object):
def __init__(self, data):
self.name = data['name']
self.auto_renew = (data['f_auto_renew'] == 'Y')
self.expiry_date = parse(data['expiredate']).date()
@property
def tld(self):
return self.name.rsplit('.', 1)[1]
Use more concise way to get tldfrom dateutil.parser import parse
class Domain(object):
def __init__(self, data):
self.name = data['name']
self.auto_renew = (data['f_auto_renew'] == 'Y')
self.expiry_date = parse(data['expiredate']).date()
@property
def tld(self):
return self.name.split('.')[-1]
|
<commit_before>from dateutil.parser import parse
class Domain(object):
def __init__(self, data):
self.name = data['name']
self.auto_renew = (data['f_auto_renew'] == 'Y')
self.expiry_date = parse(data['expiredate']).date()
@property
def tld(self):
return self.name.rsplit('.', 1)[1]
<commit_msg>Use more concise way to get tld<commit_after>from dateutil.parser import parse
class Domain(object):
def __init__(self, data):
self.name = data['name']
self.auto_renew = (data['f_auto_renew'] == 'Y')
self.expiry_date = parse(data['expiredate']).date()
@property
def tld(self):
return self.name.split('.')[-1]
|
c1bafcaa2c826ab450bd7a5e77a48fd742098e19
|
trex/serializers.py
|
trex/serializers.py
|
# -*- coding: utf-8 -*-
#
# (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com>
#
# See LICENSE comming with the source of 'trex' for details.
#
from rest_framework.serializers import (
HyperlinkedModelSerializer, HyperlinkedIdentityField,
)
from trex.models.project import Project, Entry
class ProjectSerializer(HyperlinkedModelSerializer):
class Meta:
model = Project
fields = ("url", "id", "name", "description", "active", "created")
class ProjectDetailSerializer(HyperlinkedModelSerializer):
entries = HyperlinkedIdentityField(view_name="project-entries-list")
class Meta:
model = Project
fields = ("id", "name", "description", "active", "created", "entries")
class EntryProjectSerializer(HyperlinkedModelSerializer):
class Meta:
model = Project
fields = ("url", "id", "name")
class EntryDetailSerializer(HyperlinkedModelSerializer):
project = EntryProjectSerializer()
class Meta:
model = Entry
fields = ("url", "id", "date", "duration", "description", "state",
"user_abbr", "user", "created", "project")
|
# -*- coding: utf-8 -*-
#
# (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com>
#
# See LICENSE comming with the source of 'trex' for details.
#
from rest_framework.serializers import (
HyperlinkedModelSerializer, HyperlinkedIdentityField,
)
from trex.models.project import Project, Entry, Tag
class ProjectSerializer(HyperlinkedModelSerializer):
class Meta:
model = Project
fields = ("url", "id", "name", "description", "active", "created")
class ProjectDetailSerializer(HyperlinkedModelSerializer):
entries = HyperlinkedIdentityField(view_name="project-entries-list")
class Meta:
model = Project
fields = ("id", "name", "description", "active", "created", "entries")
class EntryTagsSerializer(HyperlinkedModelSerializer):
class Meta:
model = Tag
fields = ("url", "id", "name")
class EntryProjectSerializer(HyperlinkedModelSerializer):
class Meta:
model = Project
fields = ("url", "id", "name")
class EntryDetailSerializer(HyperlinkedModelSerializer):
tags = EntryTagsSerializer(many=True)
project = EntryProjectSerializer()
class Meta:
model = Entry
fields = ("url", "id", "date", "duration", "description", "state",
"user_abbr", "user", "created", "project", "tags")
|
Add EntryTagsSerializer for returning tags of an Entry
|
Add EntryTagsSerializer for returning tags of an Entry
|
Python
|
mit
|
bjoernricks/trex,bjoernricks/trex
|
# -*- coding: utf-8 -*-
#
# (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com>
#
# See LICENSE comming with the source of 'trex' for details.
#
from rest_framework.serializers import (
HyperlinkedModelSerializer, HyperlinkedIdentityField,
)
from trex.models.project import Project, Entry
class ProjectSerializer(HyperlinkedModelSerializer):
class Meta:
model = Project
fields = ("url", "id", "name", "description", "active", "created")
class ProjectDetailSerializer(HyperlinkedModelSerializer):
entries = HyperlinkedIdentityField(view_name="project-entries-list")
class Meta:
model = Project
fields = ("id", "name", "description", "active", "created", "entries")
class EntryProjectSerializer(HyperlinkedModelSerializer):
class Meta:
model = Project
fields = ("url", "id", "name")
class EntryDetailSerializer(HyperlinkedModelSerializer):
project = EntryProjectSerializer()
class Meta:
model = Entry
fields = ("url", "id", "date", "duration", "description", "state",
"user_abbr", "user", "created", "project")
Add EntryTagsSerializer for returning tags of an Entry
|
# -*- coding: utf-8 -*-
#
# (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com>
#
# See LICENSE comming with the source of 'trex' for details.
#
from rest_framework.serializers import (
HyperlinkedModelSerializer, HyperlinkedIdentityField,
)
from trex.models.project import Project, Entry, Tag
class ProjectSerializer(HyperlinkedModelSerializer):
class Meta:
model = Project
fields = ("url", "id", "name", "description", "active", "created")
class ProjectDetailSerializer(HyperlinkedModelSerializer):
entries = HyperlinkedIdentityField(view_name="project-entries-list")
class Meta:
model = Project
fields = ("id", "name", "description", "active", "created", "entries")
class EntryTagsSerializer(HyperlinkedModelSerializer):
class Meta:
model = Tag
fields = ("url", "id", "name")
class EntryProjectSerializer(HyperlinkedModelSerializer):
class Meta:
model = Project
fields = ("url", "id", "name")
class EntryDetailSerializer(HyperlinkedModelSerializer):
tags = EntryTagsSerializer(many=True)
project = EntryProjectSerializer()
class Meta:
model = Entry
fields = ("url", "id", "date", "duration", "description", "state",
"user_abbr", "user", "created", "project", "tags")
|
<commit_before># -*- coding: utf-8 -*-
#
# (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com>
#
# See LICENSE comming with the source of 'trex' for details.
#
from rest_framework.serializers import (
HyperlinkedModelSerializer, HyperlinkedIdentityField,
)
from trex.models.project import Project, Entry
class ProjectSerializer(HyperlinkedModelSerializer):
class Meta:
model = Project
fields = ("url", "id", "name", "description", "active", "created")
class ProjectDetailSerializer(HyperlinkedModelSerializer):
entries = HyperlinkedIdentityField(view_name="project-entries-list")
class Meta:
model = Project
fields = ("id", "name", "description", "active", "created", "entries")
class EntryProjectSerializer(HyperlinkedModelSerializer):
class Meta:
model = Project
fields = ("url", "id", "name")
class EntryDetailSerializer(HyperlinkedModelSerializer):
project = EntryProjectSerializer()
class Meta:
model = Entry
fields = ("url", "id", "date", "duration", "description", "state",
"user_abbr", "user", "created", "project")
<commit_msg>Add EntryTagsSerializer for returning tags of an Entry<commit_after>
|
# -*- coding: utf-8 -*-
#
# (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com>
#
# See LICENSE comming with the source of 'trex' for details.
#
from rest_framework.serializers import (
HyperlinkedModelSerializer, HyperlinkedIdentityField,
)
from trex.models.project import Project, Entry, Tag
class ProjectSerializer(HyperlinkedModelSerializer):
class Meta:
model = Project
fields = ("url", "id", "name", "description", "active", "created")
class ProjectDetailSerializer(HyperlinkedModelSerializer):
entries = HyperlinkedIdentityField(view_name="project-entries-list")
class Meta:
model = Project
fields = ("id", "name", "description", "active", "created", "entries")
class EntryTagsSerializer(HyperlinkedModelSerializer):
class Meta:
model = Tag
fields = ("url", "id", "name")
class EntryProjectSerializer(HyperlinkedModelSerializer):
class Meta:
model = Project
fields = ("url", "id", "name")
class EntryDetailSerializer(HyperlinkedModelSerializer):
tags = EntryTagsSerializer(many=True)
project = EntryProjectSerializer()
class Meta:
model = Entry
fields = ("url", "id", "date", "duration", "description", "state",
"user_abbr", "user", "created", "project", "tags")
|
# -*- coding: utf-8 -*-
#
# (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com>
#
# See LICENSE comming with the source of 'trex' for details.
#
from rest_framework.serializers import (
HyperlinkedModelSerializer, HyperlinkedIdentityField,
)
from trex.models.project import Project, Entry
class ProjectSerializer(HyperlinkedModelSerializer):
class Meta:
model = Project
fields = ("url", "id", "name", "description", "active", "created")
class ProjectDetailSerializer(HyperlinkedModelSerializer):
entries = HyperlinkedIdentityField(view_name="project-entries-list")
class Meta:
model = Project
fields = ("id", "name", "description", "active", "created", "entries")
class EntryProjectSerializer(HyperlinkedModelSerializer):
class Meta:
model = Project
fields = ("url", "id", "name")
class EntryDetailSerializer(HyperlinkedModelSerializer):
project = EntryProjectSerializer()
class Meta:
model = Entry
fields = ("url", "id", "date", "duration", "description", "state",
"user_abbr", "user", "created", "project")
Add EntryTagsSerializer for returning tags of an Entry# -*- coding: utf-8 -*-
#
# (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com>
#
# See LICENSE comming with the source of 'trex' for details.
#
from rest_framework.serializers import (
HyperlinkedModelSerializer, HyperlinkedIdentityField,
)
from trex.models.project import Project, Entry, Tag
class ProjectSerializer(HyperlinkedModelSerializer):
class Meta:
model = Project
fields = ("url", "id", "name", "description", "active", "created")
class ProjectDetailSerializer(HyperlinkedModelSerializer):
entries = HyperlinkedIdentityField(view_name="project-entries-list")
class Meta:
model = Project
fields = ("id", "name", "description", "active", "created", "entries")
class EntryTagsSerializer(HyperlinkedModelSerializer):
class Meta:
model = Tag
fields = ("url", "id", "name")
class EntryProjectSerializer(HyperlinkedModelSerializer):
class Meta:
model = Project
fields = ("url", "id", "name")
class EntryDetailSerializer(HyperlinkedModelSerializer):
tags = EntryTagsSerializer(many=True)
project = EntryProjectSerializer()
class Meta:
model = Entry
fields = ("url", "id", "date", "duration", "description", "state",
"user_abbr", "user", "created", "project", "tags")
|
<commit_before># -*- coding: utf-8 -*-
#
# (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com>
#
# See LICENSE comming with the source of 'trex' for details.
#
from rest_framework.serializers import (
HyperlinkedModelSerializer, HyperlinkedIdentityField,
)
from trex.models.project import Project, Entry
class ProjectSerializer(HyperlinkedModelSerializer):
class Meta:
model = Project
fields = ("url", "id", "name", "description", "active", "created")
class ProjectDetailSerializer(HyperlinkedModelSerializer):
entries = HyperlinkedIdentityField(view_name="project-entries-list")
class Meta:
model = Project
fields = ("id", "name", "description", "active", "created", "entries")
class EntryProjectSerializer(HyperlinkedModelSerializer):
class Meta:
model = Project
fields = ("url", "id", "name")
class EntryDetailSerializer(HyperlinkedModelSerializer):
project = EntryProjectSerializer()
class Meta:
model = Entry
fields = ("url", "id", "date", "duration", "description", "state",
"user_abbr", "user", "created", "project")
<commit_msg>Add EntryTagsSerializer for returning tags of an Entry<commit_after># -*- coding: utf-8 -*-
#
# (c) 2014 Bjoern Ricks <bjoern.ricks@gmail.com>
#
# See LICENSE comming with the source of 'trex' for details.
#
from rest_framework.serializers import (
HyperlinkedModelSerializer, HyperlinkedIdentityField,
)
from trex.models.project import Project, Entry, Tag
class ProjectSerializer(HyperlinkedModelSerializer):
class Meta:
model = Project
fields = ("url", "id", "name", "description", "active", "created")
class ProjectDetailSerializer(HyperlinkedModelSerializer):
entries = HyperlinkedIdentityField(view_name="project-entries-list")
class Meta:
model = Project
fields = ("id", "name", "description", "active", "created", "entries")
class EntryTagsSerializer(HyperlinkedModelSerializer):
class Meta:
model = Tag
fields = ("url", "id", "name")
class EntryProjectSerializer(HyperlinkedModelSerializer):
class Meta:
model = Project
fields = ("url", "id", "name")
class EntryDetailSerializer(HyperlinkedModelSerializer):
tags = EntryTagsSerializer(many=True)
project = EntryProjectSerializer()
class Meta:
model = Entry
fields = ("url", "id", "date", "duration", "description", "state",
"user_abbr", "user", "created", "project", "tags")
|
1add854d3855dade0d6eea9a2740e8233d02cf6b
|
amitgroup/features/edge_descriptor.py
|
amitgroup/features/edge_descriptor.py
|
from binary_descriptor import BinaryDescriptor
import amitgroup as ag
import amitgroup.features
# TODO: This is temporarily moved
#@BinaryDescriptor.register('edges')
class EdgeDescriptor(BinaryDescriptor):
"""
Binary descriptor based on edges.
The parameters are similar to :func:`amitgroup.features.bedges`.
Parameters
----------
polarity_sensitive : bool
If True, the polarity of the edges will matter. If False, then the direction of edges will not matter.
k : int
See :func:`amitgroup.features.bedges`.
radius : int
Radius of edge spreading. See :func:`amitgroup.features.bedges`.
min_contrast : float
See :func:`amitgroup.features.bedges`.
"""
def __init__(self, polarity_sensitive=True, k=5, radius=1, min_contrast=0.1):
self.settings = {}
# Change this
self.settings['contrast_insensitive'] = polarity_sensitive
self.settings['k'] = k
self.settings['radius'] = radius
self.settings['min_contrast'] = min_contrast
self.settings.update(settings)
def extract_features(self, img):
#return ag.features.bedges_from_image(img, **self.settings)
return ag.features.bedges(img, **self.settings)
def save_to_dict(self):
return self.settings
@classmethod
def load_from_dict(cls, d):
return cls(d)
EdgeDescriptor = BinaryDescriptor.register('edges')(EdgeDescriptor)
|
from binary_descriptor import BinaryDescriptor
import amitgroup as ag
import amitgroup.features
# TODO: This is temporarily moved
#@BinaryDescriptor.register('edges')
class EdgeDescriptor(BinaryDescriptor):
"""
Binary descriptor based on edges.
The parameters are similar to :func:`amitgroup.features.bedges`.
Parameters
----------
polarity_sensitive : bool
If True, the polarity of the edges will matter. If False, then the direction of edges will not matter.
k : int
See :func:`amitgroup.features.bedges`.
radius : int
Radius of edge spreading. See :func:`amitgroup.features.bedges`.
min_contrast : float
See :func:`amitgroup.features.bedges`.
"""
def __init__(self, polarity_sensitive=True, k=5, radius=1, min_contrast=0.1):
self.settings = {}
# Change this
self.settings['contrast_insensitive'] = not polarity_sensitive
self.settings['k'] = k
self.settings['radius'] = radius
self.settings['min_contrast'] = min_contrast
self.settings.update(settings)
def extract_features(self, img):
#return ag.features.bedges_from_image(img, **self.settings)
return ag.features.bedges(img, **self.settings)
def save_to_dict(self):
return self.settings
@classmethod
def load_from_dict(cls, d):
return cls(d)
EdgeDescriptor = BinaryDescriptor.register('edges')(EdgeDescriptor)
|
Fix bug in EdgeDescriptor setting.
|
Fix bug in EdgeDescriptor setting.
|
Python
|
bsd-3-clause
|
amitgroup/amitgroup
|
from binary_descriptor import BinaryDescriptor
import amitgroup as ag
import amitgroup.features
# TODO: This is temporarily moved
#@BinaryDescriptor.register('edges')
class EdgeDescriptor(BinaryDescriptor):
"""
Binary descriptor based on edges.
The parameters are similar to :func:`amitgroup.features.bedges`.
Parameters
----------
polarity_sensitive : bool
If True, the polarity of the edges will matter. If False, then the direction of edges will not matter.
k : int
See :func:`amitgroup.features.bedges`.
radius : int
Radius of edge spreading. See :func:`amitgroup.features.bedges`.
min_contrast : float
See :func:`amitgroup.features.bedges`.
"""
def __init__(self, polarity_sensitive=True, k=5, radius=1, min_contrast=0.1):
self.settings = {}
# Change this
self.settings['contrast_insensitive'] = polarity_sensitive
self.settings['k'] = k
self.settings['radius'] = radius
self.settings['min_contrast'] = min_contrast
self.settings.update(settings)
def extract_features(self, img):
#return ag.features.bedges_from_image(img, **self.settings)
return ag.features.bedges(img, **self.settings)
def save_to_dict(self):
return self.settings
@classmethod
def load_from_dict(cls, d):
return cls(d)
EdgeDescriptor = BinaryDescriptor.register('edges')(EdgeDescriptor)
Fix bug in EdgeDescriptor setting.
|
from binary_descriptor import BinaryDescriptor
import amitgroup as ag
import amitgroup.features
# TODO: This is temporarily moved
#@BinaryDescriptor.register('edges')
class EdgeDescriptor(BinaryDescriptor):
"""
Binary descriptor based on edges.
The parameters are similar to :func:`amitgroup.features.bedges`.
Parameters
----------
polarity_sensitive : bool
If True, the polarity of the edges will matter. If False, then the direction of edges will not matter.
k : int
See :func:`amitgroup.features.bedges`.
radius : int
Radius of edge spreading. See :func:`amitgroup.features.bedges`.
min_contrast : float
See :func:`amitgroup.features.bedges`.
"""
def __init__(self, polarity_sensitive=True, k=5, radius=1, min_contrast=0.1):
self.settings = {}
# Change this
self.settings['contrast_insensitive'] = not polarity_sensitive
self.settings['k'] = k
self.settings['radius'] = radius
self.settings['min_contrast'] = min_contrast
self.settings.update(settings)
def extract_features(self, img):
#return ag.features.bedges_from_image(img, **self.settings)
return ag.features.bedges(img, **self.settings)
def save_to_dict(self):
return self.settings
@classmethod
def load_from_dict(cls, d):
return cls(d)
EdgeDescriptor = BinaryDescriptor.register('edges')(EdgeDescriptor)
|
<commit_before>
from binary_descriptor import BinaryDescriptor
import amitgroup as ag
import amitgroup.features
# TODO: This is temporarily moved
#@BinaryDescriptor.register('edges')
class EdgeDescriptor(BinaryDescriptor):
"""
Binary descriptor based on edges.
The parameters are similar to :func:`amitgroup.features.bedges`.
Parameters
----------
polarity_sensitive : bool
If True, the polarity of the edges will matter. If False, then the direction of edges will not matter.
k : int
See :func:`amitgroup.features.bedges`.
radius : int
Radius of edge spreading. See :func:`amitgroup.features.bedges`.
min_contrast : float
See :func:`amitgroup.features.bedges`.
"""
def __init__(self, polarity_sensitive=True, k=5, radius=1, min_contrast=0.1):
self.settings = {}
# Change this
self.settings['contrast_insensitive'] = polarity_sensitive
self.settings['k'] = k
self.settings['radius'] = radius
self.settings['min_contrast'] = min_contrast
self.settings.update(settings)
def extract_features(self, img):
#return ag.features.bedges_from_image(img, **self.settings)
return ag.features.bedges(img, **self.settings)
def save_to_dict(self):
return self.settings
@classmethod
def load_from_dict(cls, d):
return cls(d)
EdgeDescriptor = BinaryDescriptor.register('edges')(EdgeDescriptor)
<commit_msg>Fix bug in EdgeDescriptor setting.<commit_after>
|
from binary_descriptor import BinaryDescriptor
import amitgroup as ag
import amitgroup.features
# TODO: This is temporarily moved
#@BinaryDescriptor.register('edges')
class EdgeDescriptor(BinaryDescriptor):
"""
Binary descriptor based on edges.
The parameters are similar to :func:`amitgroup.features.bedges`.
Parameters
----------
polarity_sensitive : bool
If True, the polarity of the edges will matter. If False, then the direction of edges will not matter.
k : int
See :func:`amitgroup.features.bedges`.
radius : int
Radius of edge spreading. See :func:`amitgroup.features.bedges`.
min_contrast : float
See :func:`amitgroup.features.bedges`.
"""
def __init__(self, polarity_sensitive=True, k=5, radius=1, min_contrast=0.1):
self.settings = {}
# Change this
self.settings['contrast_insensitive'] = not polarity_sensitive
self.settings['k'] = k
self.settings['radius'] = radius
self.settings['min_contrast'] = min_contrast
self.settings.update(settings)
def extract_features(self, img):
#return ag.features.bedges_from_image(img, **self.settings)
return ag.features.bedges(img, **self.settings)
def save_to_dict(self):
return self.settings
@classmethod
def load_from_dict(cls, d):
return cls(d)
EdgeDescriptor = BinaryDescriptor.register('edges')(EdgeDescriptor)
|
from binary_descriptor import BinaryDescriptor
import amitgroup as ag
import amitgroup.features
# TODO: This is temporarily moved
#@BinaryDescriptor.register('edges')
class EdgeDescriptor(BinaryDescriptor):
"""
Binary descriptor based on edges.
The parameters are similar to :func:`amitgroup.features.bedges`.
Parameters
----------
polarity_sensitive : bool
If True, the polarity of the edges will matter. If False, then the direction of edges will not matter.
k : int
See :func:`amitgroup.features.bedges`.
radius : int
Radius of edge spreading. See :func:`amitgroup.features.bedges`.
min_contrast : float
See :func:`amitgroup.features.bedges`.
"""
def __init__(self, polarity_sensitive=True, k=5, radius=1, min_contrast=0.1):
self.settings = {}
# Change this
self.settings['contrast_insensitive'] = polarity_sensitive
self.settings['k'] = k
self.settings['radius'] = radius
self.settings['min_contrast'] = min_contrast
self.settings.update(settings)
def extract_features(self, img):
#return ag.features.bedges_from_image(img, **self.settings)
return ag.features.bedges(img, **self.settings)
def save_to_dict(self):
return self.settings
@classmethod
def load_from_dict(cls, d):
return cls(d)
EdgeDescriptor = BinaryDescriptor.register('edges')(EdgeDescriptor)
Fix bug in EdgeDescriptor setting.
from binary_descriptor import BinaryDescriptor
import amitgroup as ag
import amitgroup.features
# TODO: This is temporarily moved
#@BinaryDescriptor.register('edges')
class EdgeDescriptor(BinaryDescriptor):
"""
Binary descriptor based on edges.
The parameters are similar to :func:`amitgroup.features.bedges`.
Parameters
----------
polarity_sensitive : bool
If True, the polarity of the edges will matter. If False, then the direction of edges will not matter.
k : int
See :func:`amitgroup.features.bedges`.
radius : int
Radius of edge spreading. See :func:`amitgroup.features.bedges`.
min_contrast : float
See :func:`amitgroup.features.bedges`.
"""
def __init__(self, polarity_sensitive=True, k=5, radius=1, min_contrast=0.1):
self.settings = {}
# Change this
self.settings['contrast_insensitive'] = not polarity_sensitive
self.settings['k'] = k
self.settings['radius'] = radius
self.settings['min_contrast'] = min_contrast
self.settings.update(settings)
def extract_features(self, img):
#return ag.features.bedges_from_image(img, **self.settings)
return ag.features.bedges(img, **self.settings)
def save_to_dict(self):
return self.settings
@classmethod
def load_from_dict(cls, d):
return cls(d)
EdgeDescriptor = BinaryDescriptor.register('edges')(EdgeDescriptor)
|
<commit_before>
from binary_descriptor import BinaryDescriptor
import amitgroup as ag
import amitgroup.features
# TODO: This is temporarily moved
#@BinaryDescriptor.register('edges')
class EdgeDescriptor(BinaryDescriptor):
"""
Binary descriptor based on edges.
The parameters are similar to :func:`amitgroup.features.bedges`.
Parameters
----------
polarity_sensitive : bool
If True, the polarity of the edges will matter. If False, then the direction of edges will not matter.
k : int
See :func:`amitgroup.features.bedges`.
radius : int
Radius of edge spreading. See :func:`amitgroup.features.bedges`.
min_contrast : float
See :func:`amitgroup.features.bedges`.
"""
def __init__(self, polarity_sensitive=True, k=5, radius=1, min_contrast=0.1):
self.settings = {}
# Change this
self.settings['contrast_insensitive'] = polarity_sensitive
self.settings['k'] = k
self.settings['radius'] = radius
self.settings['min_contrast'] = min_contrast
self.settings.update(settings)
def extract_features(self, img):
#return ag.features.bedges_from_image(img, **self.settings)
return ag.features.bedges(img, **self.settings)
def save_to_dict(self):
return self.settings
@classmethod
def load_from_dict(cls, d):
return cls(d)
EdgeDescriptor = BinaryDescriptor.register('edges')(EdgeDescriptor)
<commit_msg>Fix bug in EdgeDescriptor setting.<commit_after>
from binary_descriptor import BinaryDescriptor
import amitgroup as ag
import amitgroup.features
# TODO: This is temporarily moved
#@BinaryDescriptor.register('edges')
class EdgeDescriptor(BinaryDescriptor):
"""
Binary descriptor based on edges.
The parameters are similar to :func:`amitgroup.features.bedges`.
Parameters
----------
polarity_sensitive : bool
If True, the polarity of the edges will matter. If False, then the direction of edges will not matter.
k : int
See :func:`amitgroup.features.bedges`.
radius : int
Radius of edge spreading. See :func:`amitgroup.features.bedges`.
min_contrast : float
See :func:`amitgroup.features.bedges`.
"""
def __init__(self, polarity_sensitive=True, k=5, radius=1, min_contrast=0.1):
self.settings = {}
# Change this
self.settings['contrast_insensitive'] = not polarity_sensitive
self.settings['k'] = k
self.settings['radius'] = radius
self.settings['min_contrast'] = min_contrast
self.settings.update(settings)
def extract_features(self, img):
#return ag.features.bedges_from_image(img, **self.settings)
return ag.features.bedges(img, **self.settings)
def save_to_dict(self):
return self.settings
@classmethod
def load_from_dict(cls, d):
return cls(d)
EdgeDescriptor = BinaryDescriptor.register('edges')(EdgeDescriptor)
|
a2589c5203c90b3b8b5cc504da36708038e0eb58
|
links/maker/urls.py
|
links/maker/urls.py
|
from django.conf.urls import patterns, url
from maker.views import (RegsitrationView,
AuthenticationView,
MakerSelfView,
MakerProfileView,
ResetPasswordRequestView,
ResetPasswordProcessView,
ChangePasswordView,
EmailChangeRequestView,
EmailChangeProcessView)
urlpatterns = patterns(
'',
url(
r'^/?$',
MakerSelfView.as_view(),
name='maker-self'
),
url(
r'^register/?$',
RegsitrationView.as_view(),
name='registration'
),
url(
r'^authenticate/?$',
AuthenticationView.as_view(),
name='authentication'
),
url(
r'^password/?$',
ChangePasswordView.as_view(),
name='change-password'
),
url(
r'^password/reset/?$',
ResetPasswordRequestView.as_view(),
name='password-reset'
),
url(
r'^password/reset/update?$',
ResetPasswordProcessView.as_view(),
name='password-reset-process'
),
url(
r'^email/?$',
EmailChangeRequestView.as_view(),
name='email-change-request'
),
url(
r'^email/update?$',
EmailChangeProcessView.as_view(),
name='email-change-process'
),
url(
r'^(?P<pk>[0-9]+)/?$',
MakerProfileView.as_view(),
name='maker-profile-view'
)
)
|
from django.conf.urls import patterns, url
from maker.views import (RegsitrationView,
AuthenticationView,
MakerSelfView,
MakerProfileView,
ResetPasswordRequestView,
ResetPasswordProcessView,
ChangePasswordView,
EmailChangeRequestView,
EmailChangeProcessView)
urlpatterns = patterns(
'',
url(
r'^self/?$',
MakerSelfView.as_view(),
name='maker-self'
),
url(
r'^register/?$',
RegsitrationView.as_view(),
name='registration'
),
url(
r'^authenticate/?$',
AuthenticationView.as_view(),
name='authentication'
),
url(
r'^password/?$',
ChangePasswordView.as_view(),
name='change-password'
),
url(
r'^password/reset/?$',
ResetPasswordRequestView.as_view(),
name='password-reset'
),
url(
r'^password/reset/update?$',
ResetPasswordProcessView.as_view(),
name='password-reset-process'
),
url(
r'^email/?$',
EmailChangeRequestView.as_view(),
name='email-change-request'
),
url(
r'^email/update?$',
EmailChangeProcessView.as_view(),
name='email-change-process'
),
url(
r'^(?P<pk>[0-9]+)/?$',
MakerProfileView.as_view(),
name='maker-profile-view'
)
)
|
Change maker self URL name
|
Change maker self URL name
|
Python
|
mit
|
projectweekend/Links-API,projectweekend/Links-API
|
from django.conf.urls import patterns, url
from maker.views import (RegsitrationView,
AuthenticationView,
MakerSelfView,
MakerProfileView,
ResetPasswordRequestView,
ResetPasswordProcessView,
ChangePasswordView,
EmailChangeRequestView,
EmailChangeProcessView)
urlpatterns = patterns(
'',
url(
r'^/?$',
MakerSelfView.as_view(),
name='maker-self'
),
url(
r'^register/?$',
RegsitrationView.as_view(),
name='registration'
),
url(
r'^authenticate/?$',
AuthenticationView.as_view(),
name='authentication'
),
url(
r'^password/?$',
ChangePasswordView.as_view(),
name='change-password'
),
url(
r'^password/reset/?$',
ResetPasswordRequestView.as_view(),
name='password-reset'
),
url(
r'^password/reset/update?$',
ResetPasswordProcessView.as_view(),
name='password-reset-process'
),
url(
r'^email/?$',
EmailChangeRequestView.as_view(),
name='email-change-request'
),
url(
r'^email/update?$',
EmailChangeProcessView.as_view(),
name='email-change-process'
),
url(
r'^(?P<pk>[0-9]+)/?$',
MakerProfileView.as_view(),
name='maker-profile-view'
)
)
Change maker self URL name
|
from django.conf.urls import patterns, url
from maker.views import (RegsitrationView,
AuthenticationView,
MakerSelfView,
MakerProfileView,
ResetPasswordRequestView,
ResetPasswordProcessView,
ChangePasswordView,
EmailChangeRequestView,
EmailChangeProcessView)
urlpatterns = patterns(
'',
url(
r'^self/?$',
MakerSelfView.as_view(),
name='maker-self'
),
url(
r'^register/?$',
RegsitrationView.as_view(),
name='registration'
),
url(
r'^authenticate/?$',
AuthenticationView.as_view(),
name='authentication'
),
url(
r'^password/?$',
ChangePasswordView.as_view(),
name='change-password'
),
url(
r'^password/reset/?$',
ResetPasswordRequestView.as_view(),
name='password-reset'
),
url(
r'^password/reset/update?$',
ResetPasswordProcessView.as_view(),
name='password-reset-process'
),
url(
r'^email/?$',
EmailChangeRequestView.as_view(),
name='email-change-request'
),
url(
r'^email/update?$',
EmailChangeProcessView.as_view(),
name='email-change-process'
),
url(
r'^(?P<pk>[0-9]+)/?$',
MakerProfileView.as_view(),
name='maker-profile-view'
)
)
|
<commit_before>from django.conf.urls import patterns, url
from maker.views import (RegsitrationView,
AuthenticationView,
MakerSelfView,
MakerProfileView,
ResetPasswordRequestView,
ResetPasswordProcessView,
ChangePasswordView,
EmailChangeRequestView,
EmailChangeProcessView)
urlpatterns = patterns(
'',
url(
r'^/?$',
MakerSelfView.as_view(),
name='maker-self'
),
url(
r'^register/?$',
RegsitrationView.as_view(),
name='registration'
),
url(
r'^authenticate/?$',
AuthenticationView.as_view(),
name='authentication'
),
url(
r'^password/?$',
ChangePasswordView.as_view(),
name='change-password'
),
url(
r'^password/reset/?$',
ResetPasswordRequestView.as_view(),
name='password-reset'
),
url(
r'^password/reset/update?$',
ResetPasswordProcessView.as_view(),
name='password-reset-process'
),
url(
r'^email/?$',
EmailChangeRequestView.as_view(),
name='email-change-request'
),
url(
r'^email/update?$',
EmailChangeProcessView.as_view(),
name='email-change-process'
),
url(
r'^(?P<pk>[0-9]+)/?$',
MakerProfileView.as_view(),
name='maker-profile-view'
)
)
<commit_msg>Change maker self URL name<commit_after>
|
from django.conf.urls import patterns, url
from maker.views import (RegsitrationView,
AuthenticationView,
MakerSelfView,
MakerProfileView,
ResetPasswordRequestView,
ResetPasswordProcessView,
ChangePasswordView,
EmailChangeRequestView,
EmailChangeProcessView)
urlpatterns = patterns(
'',
url(
r'^self/?$',
MakerSelfView.as_view(),
name='maker-self'
),
url(
r'^register/?$',
RegsitrationView.as_view(),
name='registration'
),
url(
r'^authenticate/?$',
AuthenticationView.as_view(),
name='authentication'
),
url(
r'^password/?$',
ChangePasswordView.as_view(),
name='change-password'
),
url(
r'^password/reset/?$',
ResetPasswordRequestView.as_view(),
name='password-reset'
),
url(
r'^password/reset/update?$',
ResetPasswordProcessView.as_view(),
name='password-reset-process'
),
url(
r'^email/?$',
EmailChangeRequestView.as_view(),
name='email-change-request'
),
url(
r'^email/update?$',
EmailChangeProcessView.as_view(),
name='email-change-process'
),
url(
r'^(?P<pk>[0-9]+)/?$',
MakerProfileView.as_view(),
name='maker-profile-view'
)
)
|
from django.conf.urls import patterns, url
from maker.views import (RegsitrationView,
AuthenticationView,
MakerSelfView,
MakerProfileView,
ResetPasswordRequestView,
ResetPasswordProcessView,
ChangePasswordView,
EmailChangeRequestView,
EmailChangeProcessView)
urlpatterns = patterns(
'',
url(
r'^/?$',
MakerSelfView.as_view(),
name='maker-self'
),
url(
r'^register/?$',
RegsitrationView.as_view(),
name='registration'
),
url(
r'^authenticate/?$',
AuthenticationView.as_view(),
name='authentication'
),
url(
r'^password/?$',
ChangePasswordView.as_view(),
name='change-password'
),
url(
r'^password/reset/?$',
ResetPasswordRequestView.as_view(),
name='password-reset'
),
url(
r'^password/reset/update?$',
ResetPasswordProcessView.as_view(),
name='password-reset-process'
),
url(
r'^email/?$',
EmailChangeRequestView.as_view(),
name='email-change-request'
),
url(
r'^email/update?$',
EmailChangeProcessView.as_view(),
name='email-change-process'
),
url(
r'^(?P<pk>[0-9]+)/?$',
MakerProfileView.as_view(),
name='maker-profile-view'
)
)
Change maker self URL namefrom django.conf.urls import patterns, url
from maker.views import (RegsitrationView,
AuthenticationView,
MakerSelfView,
MakerProfileView,
ResetPasswordRequestView,
ResetPasswordProcessView,
ChangePasswordView,
EmailChangeRequestView,
EmailChangeProcessView)
urlpatterns = patterns(
'',
url(
r'^self/?$',
MakerSelfView.as_view(),
name='maker-self'
),
url(
r'^register/?$',
RegsitrationView.as_view(),
name='registration'
),
url(
r'^authenticate/?$',
AuthenticationView.as_view(),
name='authentication'
),
url(
r'^password/?$',
ChangePasswordView.as_view(),
name='change-password'
),
url(
r'^password/reset/?$',
ResetPasswordRequestView.as_view(),
name='password-reset'
),
url(
r'^password/reset/update?$',
ResetPasswordProcessView.as_view(),
name='password-reset-process'
),
url(
r'^email/?$',
EmailChangeRequestView.as_view(),
name='email-change-request'
),
url(
r'^email/update?$',
EmailChangeProcessView.as_view(),
name='email-change-process'
),
url(
r'^(?P<pk>[0-9]+)/?$',
MakerProfileView.as_view(),
name='maker-profile-view'
)
)
|
<commit_before>from django.conf.urls import patterns, url
from maker.views import (RegsitrationView,
AuthenticationView,
MakerSelfView,
MakerProfileView,
ResetPasswordRequestView,
ResetPasswordProcessView,
ChangePasswordView,
EmailChangeRequestView,
EmailChangeProcessView)
urlpatterns = patterns(
'',
url(
r'^/?$',
MakerSelfView.as_view(),
name='maker-self'
),
url(
r'^register/?$',
RegsitrationView.as_view(),
name='registration'
),
url(
r'^authenticate/?$',
AuthenticationView.as_view(),
name='authentication'
),
url(
r'^password/?$',
ChangePasswordView.as_view(),
name='change-password'
),
url(
r'^password/reset/?$',
ResetPasswordRequestView.as_view(),
name='password-reset'
),
url(
r'^password/reset/update?$',
ResetPasswordProcessView.as_view(),
name='password-reset-process'
),
url(
r'^email/?$',
EmailChangeRequestView.as_view(),
name='email-change-request'
),
url(
r'^email/update?$',
EmailChangeProcessView.as_view(),
name='email-change-process'
),
url(
r'^(?P<pk>[0-9]+)/?$',
MakerProfileView.as_view(),
name='maker-profile-view'
)
)
<commit_msg>Change maker self URL name<commit_after>from django.conf.urls import patterns, url
from maker.views import (RegsitrationView,
AuthenticationView,
MakerSelfView,
MakerProfileView,
ResetPasswordRequestView,
ResetPasswordProcessView,
ChangePasswordView,
EmailChangeRequestView,
EmailChangeProcessView)
urlpatterns = patterns(
'',
url(
r'^self/?$',
MakerSelfView.as_view(),
name='maker-self'
),
url(
r'^register/?$',
RegsitrationView.as_view(),
name='registration'
),
url(
r'^authenticate/?$',
AuthenticationView.as_view(),
name='authentication'
),
url(
r'^password/?$',
ChangePasswordView.as_view(),
name='change-password'
),
url(
r'^password/reset/?$',
ResetPasswordRequestView.as_view(),
name='password-reset'
),
url(
r'^password/reset/update?$',
ResetPasswordProcessView.as_view(),
name='password-reset-process'
),
url(
r'^email/?$',
EmailChangeRequestView.as_view(),
name='email-change-request'
),
url(
r'^email/update?$',
EmailChangeProcessView.as_view(),
name='email-change-process'
),
url(
r'^(?P<pk>[0-9]+)/?$',
MakerProfileView.as_view(),
name='maker-profile-view'
)
)
|
332ed6c26830bf2ac8e154948c4c58b745d5b5ae
|
cosmo_tester/test_suites/snapshots/conftest.py
|
cosmo_tester/test_suites/snapshots/conftest.py
|
import pytest
from cosmo_tester.framework.test_hosts import Hosts, get_image
from cosmo_tester.test_suites.snapshots import get_multi_tenant_versions_list
@pytest.fixture(scope='function', params=get_multi_tenant_versions_list())
def hosts(request, ssh_key, module_tmpdir, test_config, logger):
hosts = Hosts(
ssh_key, module_tmpdir,
test_config, logger, request,
number_of_instances=4,
)
hosts.instances[0] = get_image(request.param, test_config)
hosts.instances[1] = get_image('master', test_config)
hosts.instances[2] = get_image('centos', test_config)
hosts.instances[3] = get_image('centos', test_config)
win_vm = hosts.instances[2]
win_vm.prepare_for_windows('windows_2012')
lin_vm = hosts.instances[3]
lin_vm.image_name = test_config.platform['centos_7_image']
lin_vm.username = test_config['test_os_usernames']['centos_7']
hosts.create()
try:
yield hosts
finally:
hosts.destroy()
|
import pytest
from cosmo_tester.framework.test_hosts import Hosts, get_image
from cosmo_tester.test_suites.snapshots import get_multi_tenant_versions_list
@pytest.fixture(scope='function', params=get_multi_tenant_versions_list())
def hosts(request, ssh_key, module_tmpdir, test_config, logger):
hosts = Hosts(
ssh_key, module_tmpdir,
test_config, logger, request,
number_of_instances=4,
)
hosts.instances[0] = get_image(request.param, test_config)
hosts.instances[1] = get_image('master', test_config)
hosts.instances[2] = get_image('centos', test_config)
hosts.instances[3] = get_image('centos', test_config)
win_vm = hosts.instances[2]
win_vm.prepare_for_windows('windows_2012')
lin_vm = hosts.instances[3]
lin_vm.image_name = test_config.platform['centos_7_image']
lin_vm.username = test_config['test_os_usernames']['centos_7']
hosts.create()
try:
if request.param in ['5.0.5', '5.1.0']:
old_mgr = hosts.instances[0]
old_mgr.wait_for_manager()
old_mgr.run_command('mv /etc/cloudify/ssl/rabbitmq{_,-}cert.pem',
use_sudo=True)
old_mgr.run_command('mv /etc/cloudify/ssl/rabbitmq{_,-}key.pem',
use_sudo=True)
old_mgr.run_command(
'chown rabbitmq. /etc/cloudify/ssl/rabbitmq-*', use_sudo=True)
old_mgr.run_command('systemctl restart cloudify-rabbitmq',
use_sudo=True)
yield hosts
finally:
hosts.destroy()
|
Use correct rabbit certs on old IP setter
|
Use correct rabbit certs on old IP setter
|
Python
|
apache-2.0
|
cloudify-cosmo/cloudify-system-tests,cloudify-cosmo/cloudify-system-tests
|
import pytest
from cosmo_tester.framework.test_hosts import Hosts, get_image
from cosmo_tester.test_suites.snapshots import get_multi_tenant_versions_list
@pytest.fixture(scope='function', params=get_multi_tenant_versions_list())
def hosts(request, ssh_key, module_tmpdir, test_config, logger):
hosts = Hosts(
ssh_key, module_tmpdir,
test_config, logger, request,
number_of_instances=4,
)
hosts.instances[0] = get_image(request.param, test_config)
hosts.instances[1] = get_image('master', test_config)
hosts.instances[2] = get_image('centos', test_config)
hosts.instances[3] = get_image('centos', test_config)
win_vm = hosts.instances[2]
win_vm.prepare_for_windows('windows_2012')
lin_vm = hosts.instances[3]
lin_vm.image_name = test_config.platform['centos_7_image']
lin_vm.username = test_config['test_os_usernames']['centos_7']
hosts.create()
try:
yield hosts
finally:
hosts.destroy()
Use correct rabbit certs on old IP setter
|
import pytest
from cosmo_tester.framework.test_hosts import Hosts, get_image
from cosmo_tester.test_suites.snapshots import get_multi_tenant_versions_list
@pytest.fixture(scope='function', params=get_multi_tenant_versions_list())
def hosts(request, ssh_key, module_tmpdir, test_config, logger):
hosts = Hosts(
ssh_key, module_tmpdir,
test_config, logger, request,
number_of_instances=4,
)
hosts.instances[0] = get_image(request.param, test_config)
hosts.instances[1] = get_image('master', test_config)
hosts.instances[2] = get_image('centos', test_config)
hosts.instances[3] = get_image('centos', test_config)
win_vm = hosts.instances[2]
win_vm.prepare_for_windows('windows_2012')
lin_vm = hosts.instances[3]
lin_vm.image_name = test_config.platform['centos_7_image']
lin_vm.username = test_config['test_os_usernames']['centos_7']
hosts.create()
try:
if request.param in ['5.0.5', '5.1.0']:
old_mgr = hosts.instances[0]
old_mgr.wait_for_manager()
old_mgr.run_command('mv /etc/cloudify/ssl/rabbitmq{_,-}cert.pem',
use_sudo=True)
old_mgr.run_command('mv /etc/cloudify/ssl/rabbitmq{_,-}key.pem',
use_sudo=True)
old_mgr.run_command(
'chown rabbitmq. /etc/cloudify/ssl/rabbitmq-*', use_sudo=True)
old_mgr.run_command('systemctl restart cloudify-rabbitmq',
use_sudo=True)
yield hosts
finally:
hosts.destroy()
|
<commit_before>import pytest
from cosmo_tester.framework.test_hosts import Hosts, get_image
from cosmo_tester.test_suites.snapshots import get_multi_tenant_versions_list
@pytest.fixture(scope='function', params=get_multi_tenant_versions_list())
def hosts(request, ssh_key, module_tmpdir, test_config, logger):
hosts = Hosts(
ssh_key, module_tmpdir,
test_config, logger, request,
number_of_instances=4,
)
hosts.instances[0] = get_image(request.param, test_config)
hosts.instances[1] = get_image('master', test_config)
hosts.instances[2] = get_image('centos', test_config)
hosts.instances[3] = get_image('centos', test_config)
win_vm = hosts.instances[2]
win_vm.prepare_for_windows('windows_2012')
lin_vm = hosts.instances[3]
lin_vm.image_name = test_config.platform['centos_7_image']
lin_vm.username = test_config['test_os_usernames']['centos_7']
hosts.create()
try:
yield hosts
finally:
hosts.destroy()
<commit_msg>Use correct rabbit certs on old IP setter<commit_after>
|
import pytest
from cosmo_tester.framework.test_hosts import Hosts, get_image
from cosmo_tester.test_suites.snapshots import get_multi_tenant_versions_list
@pytest.fixture(scope='function', params=get_multi_tenant_versions_list())
def hosts(request, ssh_key, module_tmpdir, test_config, logger):
hosts = Hosts(
ssh_key, module_tmpdir,
test_config, logger, request,
number_of_instances=4,
)
hosts.instances[0] = get_image(request.param, test_config)
hosts.instances[1] = get_image('master', test_config)
hosts.instances[2] = get_image('centos', test_config)
hosts.instances[3] = get_image('centos', test_config)
win_vm = hosts.instances[2]
win_vm.prepare_for_windows('windows_2012')
lin_vm = hosts.instances[3]
lin_vm.image_name = test_config.platform['centos_7_image']
lin_vm.username = test_config['test_os_usernames']['centos_7']
hosts.create()
try:
if request.param in ['5.0.5', '5.1.0']:
old_mgr = hosts.instances[0]
old_mgr.wait_for_manager()
old_mgr.run_command('mv /etc/cloudify/ssl/rabbitmq{_,-}cert.pem',
use_sudo=True)
old_mgr.run_command('mv /etc/cloudify/ssl/rabbitmq{_,-}key.pem',
use_sudo=True)
old_mgr.run_command(
'chown rabbitmq. /etc/cloudify/ssl/rabbitmq-*', use_sudo=True)
old_mgr.run_command('systemctl restart cloudify-rabbitmq',
use_sudo=True)
yield hosts
finally:
hosts.destroy()
|
import pytest
from cosmo_tester.framework.test_hosts import Hosts, get_image
from cosmo_tester.test_suites.snapshots import get_multi_tenant_versions_list
@pytest.fixture(scope='function', params=get_multi_tenant_versions_list())
def hosts(request, ssh_key, module_tmpdir, test_config, logger):
hosts = Hosts(
ssh_key, module_tmpdir,
test_config, logger, request,
number_of_instances=4,
)
hosts.instances[0] = get_image(request.param, test_config)
hosts.instances[1] = get_image('master', test_config)
hosts.instances[2] = get_image('centos', test_config)
hosts.instances[3] = get_image('centos', test_config)
win_vm = hosts.instances[2]
win_vm.prepare_for_windows('windows_2012')
lin_vm = hosts.instances[3]
lin_vm.image_name = test_config.platform['centos_7_image']
lin_vm.username = test_config['test_os_usernames']['centos_7']
hosts.create()
try:
yield hosts
finally:
hosts.destroy()
Use correct rabbit certs on old IP setterimport pytest
from cosmo_tester.framework.test_hosts import Hosts, get_image
from cosmo_tester.test_suites.snapshots import get_multi_tenant_versions_list
@pytest.fixture(scope='function', params=get_multi_tenant_versions_list())
def hosts(request, ssh_key, module_tmpdir, test_config, logger):
hosts = Hosts(
ssh_key, module_tmpdir,
test_config, logger, request,
number_of_instances=4,
)
hosts.instances[0] = get_image(request.param, test_config)
hosts.instances[1] = get_image('master', test_config)
hosts.instances[2] = get_image('centos', test_config)
hosts.instances[3] = get_image('centos', test_config)
win_vm = hosts.instances[2]
win_vm.prepare_for_windows('windows_2012')
lin_vm = hosts.instances[3]
lin_vm.image_name = test_config.platform['centos_7_image']
lin_vm.username = test_config['test_os_usernames']['centos_7']
hosts.create()
try:
if request.param in ['5.0.5', '5.1.0']:
old_mgr = hosts.instances[0]
old_mgr.wait_for_manager()
old_mgr.run_command('mv /etc/cloudify/ssl/rabbitmq{_,-}cert.pem',
use_sudo=True)
old_mgr.run_command('mv /etc/cloudify/ssl/rabbitmq{_,-}key.pem',
use_sudo=True)
old_mgr.run_command(
'chown rabbitmq. /etc/cloudify/ssl/rabbitmq-*', use_sudo=True)
old_mgr.run_command('systemctl restart cloudify-rabbitmq',
use_sudo=True)
yield hosts
finally:
hosts.destroy()
|
<commit_before>import pytest
from cosmo_tester.framework.test_hosts import Hosts, get_image
from cosmo_tester.test_suites.snapshots import get_multi_tenant_versions_list
@pytest.fixture(scope='function', params=get_multi_tenant_versions_list())
def hosts(request, ssh_key, module_tmpdir, test_config, logger):
hosts = Hosts(
ssh_key, module_tmpdir,
test_config, logger, request,
number_of_instances=4,
)
hosts.instances[0] = get_image(request.param, test_config)
hosts.instances[1] = get_image('master', test_config)
hosts.instances[2] = get_image('centos', test_config)
hosts.instances[3] = get_image('centos', test_config)
win_vm = hosts.instances[2]
win_vm.prepare_for_windows('windows_2012')
lin_vm = hosts.instances[3]
lin_vm.image_name = test_config.platform['centos_7_image']
lin_vm.username = test_config['test_os_usernames']['centos_7']
hosts.create()
try:
yield hosts
finally:
hosts.destroy()
<commit_msg>Use correct rabbit certs on old IP setter<commit_after>import pytest
from cosmo_tester.framework.test_hosts import Hosts, get_image
from cosmo_tester.test_suites.snapshots import get_multi_tenant_versions_list
@pytest.fixture(scope='function', params=get_multi_tenant_versions_list())
def hosts(request, ssh_key, module_tmpdir, test_config, logger):
hosts = Hosts(
ssh_key, module_tmpdir,
test_config, logger, request,
number_of_instances=4,
)
hosts.instances[0] = get_image(request.param, test_config)
hosts.instances[1] = get_image('master', test_config)
hosts.instances[2] = get_image('centos', test_config)
hosts.instances[3] = get_image('centos', test_config)
win_vm = hosts.instances[2]
win_vm.prepare_for_windows('windows_2012')
lin_vm = hosts.instances[3]
lin_vm.image_name = test_config.platform['centos_7_image']
lin_vm.username = test_config['test_os_usernames']['centos_7']
hosts.create()
try:
if request.param in ['5.0.5', '5.1.0']:
old_mgr = hosts.instances[0]
old_mgr.wait_for_manager()
old_mgr.run_command('mv /etc/cloudify/ssl/rabbitmq{_,-}cert.pem',
use_sudo=True)
old_mgr.run_command('mv /etc/cloudify/ssl/rabbitmq{_,-}key.pem',
use_sudo=True)
old_mgr.run_command(
'chown rabbitmq. /etc/cloudify/ssl/rabbitmq-*', use_sudo=True)
old_mgr.run_command('systemctl restart cloudify-rabbitmq',
use_sudo=True)
yield hosts
finally:
hosts.destroy()
|
428ff018ccda3862446ebaadf61db1a03470c18f
|
tests/mltils/test_infrequent_value_encoder.py
|
tests/mltils/test_infrequent_value_encoder.py
|
# pylint: disable=missing-docstring, invalid-name, import-error
import pandas as pd
from mltils.encoders import InfrequentValueEncoder
def test_infrequent_value_encoder_1():
ive = InfrequentValueEncoder()
assert ive is not None
def test_infrequent_value_encoder_2():
df = pd.DataFrame({'A': ['a', 'a', 'b', 'b', 'c']})
ive = InfrequentValueEncoder(thrshld=1, str_rpl='ifq')
encoded = ive.fit_transform(df)
expected = pd.DataFrame({'A': ['a', 'a', 'b', 'b', 'ifq']})
assert expected.equals(encoded)
|
Add unit tests for InfrequentValueEncoder
|
Add unit tests for InfrequentValueEncoder
|
Python
|
mit
|
rladeira/mltils
|
Add unit tests for InfrequentValueEncoder
|
# pylint: disable=missing-docstring, invalid-name, import-error
import pandas as pd
from mltils.encoders import InfrequentValueEncoder
def test_infrequent_value_encoder_1():
ive = InfrequentValueEncoder()
assert ive is not None
def test_infrequent_value_encoder_2():
df = pd.DataFrame({'A': ['a', 'a', 'b', 'b', 'c']})
ive = InfrequentValueEncoder(thrshld=1, str_rpl='ifq')
encoded = ive.fit_transform(df)
expected = pd.DataFrame({'A': ['a', 'a', 'b', 'b', 'ifq']})
assert expected.equals(encoded)
|
<commit_before><commit_msg>Add unit tests for InfrequentValueEncoder<commit_after>
|
# pylint: disable=missing-docstring, invalid-name, import-error
import pandas as pd
from mltils.encoders import InfrequentValueEncoder
def test_infrequent_value_encoder_1():
ive = InfrequentValueEncoder()
assert ive is not None
def test_infrequent_value_encoder_2():
df = pd.DataFrame({'A': ['a', 'a', 'b', 'b', 'c']})
ive = InfrequentValueEncoder(thrshld=1, str_rpl='ifq')
encoded = ive.fit_transform(df)
expected = pd.DataFrame({'A': ['a', 'a', 'b', 'b', 'ifq']})
assert expected.equals(encoded)
|
Add unit tests for InfrequentValueEncoder# pylint: disable=missing-docstring, invalid-name, import-error
import pandas as pd
from mltils.encoders import InfrequentValueEncoder
def test_infrequent_value_encoder_1():
ive = InfrequentValueEncoder()
assert ive is not None
def test_infrequent_value_encoder_2():
df = pd.DataFrame({'A': ['a', 'a', 'b', 'b', 'c']})
ive = InfrequentValueEncoder(thrshld=1, str_rpl='ifq')
encoded = ive.fit_transform(df)
expected = pd.DataFrame({'A': ['a', 'a', 'b', 'b', 'ifq']})
assert expected.equals(encoded)
|
<commit_before><commit_msg>Add unit tests for InfrequentValueEncoder<commit_after># pylint: disable=missing-docstring, invalid-name, import-error
import pandas as pd
from mltils.encoders import InfrequentValueEncoder
def test_infrequent_value_encoder_1():
ive = InfrequentValueEncoder()
assert ive is not None
def test_infrequent_value_encoder_2():
df = pd.DataFrame({'A': ['a', 'a', 'b', 'b', 'c']})
ive = InfrequentValueEncoder(thrshld=1, str_rpl='ifq')
encoded = ive.fit_transform(df)
expected = pd.DataFrame({'A': ['a', 'a', 'b', 'b', 'ifq']})
assert expected.equals(encoded)
|
|
0fd7b771823b97cb5fb7789c981d4ab3befcd28e
|
bluebottle/homepage/models.py
|
bluebottle/homepage/models.py
|
from bluebottle.quotes.models import Quote
from bluebottle.slides.models import Slide
from bluebottle.statistics.models import Statistic
from bluebottle.projects.models import Project
class HomePage(object):
"""
Instead of serving all the objects separately we combine
Slide, Quote and Stats into a dummy object
"""
def get(self, language):
self.id = language
self.quotes = Quote.objects.published().filter(language=language)
self.slides = Slide.objects.published().filter(language=language)
self.statistics = Statistic.objects.filter(active=True, language=language).all()
projects = Project.objects.filter(is_campaign=True, status__viewable=True)
if language == 'en':
projects = projects.filter(language__code=language)
projects = projects.order_by('?')
if len(projects) > 4:
self.projects = projects[0:4]
elif len(projects) > 0:
self.projects = projects[0:len(projects)]
else:
self.projects = None
return self
|
from bluebottle.quotes.models import Quote
from bluebottle.slides.models import Slide
from bluebottle.statistics.models import Statistic
from bluebottle.projects.models import Project
class HomePage(object):
"""
Instead of serving all the objects separately we combine
Slide, Quote and Stats into a dummy object
"""
def get(self, language):
self.id = language
self.quotes = Quote.objects.published().filter(language=language)
self.slides = Slide.objects.published().filter(language=language)
self.statistics = Statistic.objects.filter(active=True, language=language).all()
projects = Project.objects.filter(is_campaign=True, status__viewable=True)
if language == 'en':
projects = projects.filter(language__code=language)
projects = projects.order_by('?')
if len(projects) > 4:
self.projects = projects[0:4]
elif len(projects) > 0:
self.projects = projects[0:len(projects)]
else:
self.projects = Project.objects.none()
return self
|
Send an empty list instead of None if no projects
|
Send an empty list instead of None if no projects
selected for homepage.
|
Python
|
bsd-3-clause
|
onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle
|
from bluebottle.quotes.models import Quote
from bluebottle.slides.models import Slide
from bluebottle.statistics.models import Statistic
from bluebottle.projects.models import Project
class HomePage(object):
"""
Instead of serving all the objects separately we combine
Slide, Quote and Stats into a dummy object
"""
def get(self, language):
self.id = language
self.quotes = Quote.objects.published().filter(language=language)
self.slides = Slide.objects.published().filter(language=language)
self.statistics = Statistic.objects.filter(active=True, language=language).all()
projects = Project.objects.filter(is_campaign=True, status__viewable=True)
if language == 'en':
projects = projects.filter(language__code=language)
projects = projects.order_by('?')
if len(projects) > 4:
self.projects = projects[0:4]
elif len(projects) > 0:
self.projects = projects[0:len(projects)]
else:
self.projects = None
return self
Send an empty list instead of None if no projects
selected for homepage.
|
from bluebottle.quotes.models import Quote
from bluebottle.slides.models import Slide
from bluebottle.statistics.models import Statistic
from bluebottle.projects.models import Project
class HomePage(object):
"""
Instead of serving all the objects separately we combine
Slide, Quote and Stats into a dummy object
"""
def get(self, language):
self.id = language
self.quotes = Quote.objects.published().filter(language=language)
self.slides = Slide.objects.published().filter(language=language)
self.statistics = Statistic.objects.filter(active=True, language=language).all()
projects = Project.objects.filter(is_campaign=True, status__viewable=True)
if language == 'en':
projects = projects.filter(language__code=language)
projects = projects.order_by('?')
if len(projects) > 4:
self.projects = projects[0:4]
elif len(projects) > 0:
self.projects = projects[0:len(projects)]
else:
self.projects = Project.objects.none()
return self
|
<commit_before>from bluebottle.quotes.models import Quote
from bluebottle.slides.models import Slide
from bluebottle.statistics.models import Statistic
from bluebottle.projects.models import Project
class HomePage(object):
"""
Instead of serving all the objects separately we combine
Slide, Quote and Stats into a dummy object
"""
def get(self, language):
self.id = language
self.quotes = Quote.objects.published().filter(language=language)
self.slides = Slide.objects.published().filter(language=language)
self.statistics = Statistic.objects.filter(active=True, language=language).all()
projects = Project.objects.filter(is_campaign=True, status__viewable=True)
if language == 'en':
projects = projects.filter(language__code=language)
projects = projects.order_by('?')
if len(projects) > 4:
self.projects = projects[0:4]
elif len(projects) > 0:
self.projects = projects[0:len(projects)]
else:
self.projects = None
return self
<commit_msg>Send an empty list instead of None if no projects
selected for homepage.<commit_after>
|
from bluebottle.quotes.models import Quote
from bluebottle.slides.models import Slide
from bluebottle.statistics.models import Statistic
from bluebottle.projects.models import Project
class HomePage(object):
"""
Instead of serving all the objects separately we combine
Slide, Quote and Stats into a dummy object
"""
def get(self, language):
self.id = language
self.quotes = Quote.objects.published().filter(language=language)
self.slides = Slide.objects.published().filter(language=language)
self.statistics = Statistic.objects.filter(active=True, language=language).all()
projects = Project.objects.filter(is_campaign=True, status__viewable=True)
if language == 'en':
projects = projects.filter(language__code=language)
projects = projects.order_by('?')
if len(projects) > 4:
self.projects = projects[0:4]
elif len(projects) > 0:
self.projects = projects[0:len(projects)]
else:
self.projects = Project.objects.none()
return self
|
from bluebottle.quotes.models import Quote
from bluebottle.slides.models import Slide
from bluebottle.statistics.models import Statistic
from bluebottle.projects.models import Project
class HomePage(object):
"""
Instead of serving all the objects separately we combine
Slide, Quote and Stats into a dummy object
"""
def get(self, language):
self.id = language
self.quotes = Quote.objects.published().filter(language=language)
self.slides = Slide.objects.published().filter(language=language)
self.statistics = Statistic.objects.filter(active=True, language=language).all()
projects = Project.objects.filter(is_campaign=True, status__viewable=True)
if language == 'en':
projects = projects.filter(language__code=language)
projects = projects.order_by('?')
if len(projects) > 4:
self.projects = projects[0:4]
elif len(projects) > 0:
self.projects = projects[0:len(projects)]
else:
self.projects = None
return self
Send an empty list instead of None if no projects
selected for homepage.from bluebottle.quotes.models import Quote
from bluebottle.slides.models import Slide
from bluebottle.statistics.models import Statistic
from bluebottle.projects.models import Project
class HomePage(object):
"""
Instead of serving all the objects separately we combine
Slide, Quote and Stats into a dummy object
"""
def get(self, language):
self.id = language
self.quotes = Quote.objects.published().filter(language=language)
self.slides = Slide.objects.published().filter(language=language)
self.statistics = Statistic.objects.filter(active=True, language=language).all()
projects = Project.objects.filter(is_campaign=True, status__viewable=True)
if language == 'en':
projects = projects.filter(language__code=language)
projects = projects.order_by('?')
if len(projects) > 4:
self.projects = projects[0:4]
elif len(projects) > 0:
self.projects = projects[0:len(projects)]
else:
self.projects = Project.objects.none()
return self
|
<commit_before>from bluebottle.quotes.models import Quote
from bluebottle.slides.models import Slide
from bluebottle.statistics.models import Statistic
from bluebottle.projects.models import Project
class HomePage(object):
"""
Instead of serving all the objects separately we combine
Slide, Quote and Stats into a dummy object
"""
def get(self, language):
self.id = language
self.quotes = Quote.objects.published().filter(language=language)
self.slides = Slide.objects.published().filter(language=language)
self.statistics = Statistic.objects.filter(active=True, language=language).all()
projects = Project.objects.filter(is_campaign=True, status__viewable=True)
if language == 'en':
projects = projects.filter(language__code=language)
projects = projects.order_by('?')
if len(projects) > 4:
self.projects = projects[0:4]
elif len(projects) > 0:
self.projects = projects[0:len(projects)]
else:
self.projects = None
return self
<commit_msg>Send an empty list instead of None if no projects
selected for homepage.<commit_after>from bluebottle.quotes.models import Quote
from bluebottle.slides.models import Slide
from bluebottle.statistics.models import Statistic
from bluebottle.projects.models import Project
class HomePage(object):
"""
Instead of serving all the objects separately we combine
Slide, Quote and Stats into a dummy object
"""
def get(self, language):
self.id = language
self.quotes = Quote.objects.published().filter(language=language)
self.slides = Slide.objects.published().filter(language=language)
self.statistics = Statistic.objects.filter(active=True, language=language).all()
projects = Project.objects.filter(is_campaign=True, status__viewable=True)
if language == 'en':
projects = projects.filter(language__code=language)
projects = projects.order_by('?')
if len(projects) > 4:
self.projects = projects[0:4]
elif len(projects) > 0:
self.projects = projects[0:len(projects)]
else:
self.projects = Project.objects.none()
return self
|
b46dc26e5e1b4c0388c330017dc52393417c3323
|
tests/test_init.py
|
tests/test_init.py
|
from disco.test import TestCase, TestJob
class InitJob(TestJob):
sort = False
@staticmethod
def map_reader(stream, size, url, params):
params.x = 10
return (stream, size, url)
@staticmethod
def map_init(iter, params):
assert hasattr(params, 'x')
iter.next()
params['x'] += 100
@staticmethod
def map(e, params):
yield e, int(e) + params['x']
@staticmethod
def reduce_init(iter, params):
params['y'] = 1000
@staticmethod
def reduce(iter, params):
for k, v in iter:
yield k, int(v) + params['y']
class InitTestCase(TestCase):
def serve(self, path):
return 'skipthis\n' + ('%s\n' % path) * 10
def runTest(self):
self.job = InitJob().run(input=self.test_server.urls(range(10)))
results = list(self.results(self.job))
for k, v in results:
self.assertEquals(int(k) + 1110, int(v))
self.assertEquals(len(results), 100)
|
from disco.test import TestCase, TestJob
class InitJob(TestJob):
params = {'x': 10}
sort = False
@staticmethod
def map_init(iter, params):
iter.next()
params['x'] += 100
@staticmethod
def map(e, params):
yield e, int(e) + params['x']
@staticmethod
def reduce_init(iter, params):
params['y'] = 1000
@staticmethod
def reduce(iter, params):
for k, v in iter:
yield k, int(v) + params['y']
class InitTestCase(TestCase):
def serve(self, path):
return 'skipthis\n' + ('%s\n' % path) * 10
def runTest(self):
self.job = InitJob().run(input=self.test_server.urls(range(10)))
results = list(self.results(self.job))
for k, v in results:
self.assertEquals(int(k) + 1110, int(v))
self.assertEquals(len(results), 100)
|
Revert "added a test for the map_reader before map_init -case which fails currently" (deprecate init functions instead)
|
Revert "added a test for the map_reader before map_init -case which fails currently"
(deprecate init functions instead)
This reverts commit 88551bf444b7b358fea8e7eb4475df2c5d87ceeb.
|
Python
|
bsd-3-clause
|
ErikDubbelboer/disco,pombredanne/disco,mwilliams3/disco,simudream/disco,pombredanne/disco,simudream/disco,mozilla/disco,beni55/disco,ErikDubbelboer/disco,pavlobaron/disco_playground,pooya/disco,ktkt2009/disco,scrapinghub/disco,seabirdzh/disco,pombredanne/disco,pooya/disco,pombredanne/disco,mwilliams3/disco,ktkt2009/disco,oldmantaiter/disco,simudream/disco,discoproject/disco,seabirdzh/disco,discoproject/disco,mwilliams3/disco,pavlobaron/disco_playground,oldmantaiter/disco,discoproject/disco,simudream/disco,beni55/disco,mozilla/disco,scrapinghub/disco,oldmantaiter/disco,pombredanne/disco,mwilliams3/disco,beni55/disco,mozilla/disco,scrapinghub/disco,ktkt2009/disco,scrapinghub/disco,seabirdzh/disco,discoproject/disco,seabirdzh/disco,beni55/disco,ErikDubbelboer/disco,mozilla/disco,ErikDubbelboer/disco,mwilliams3/disco,pooya/disco,seabirdzh/disco,pavlobaron/disco_playground,pavlobaron/disco_playground,ktkt2009/disco,ktkt2009/disco,discoproject/disco,simudream/disco,pooya/disco,beni55/disco,oldmantaiter/disco,oldmantaiter/disco,ErikDubbelboer/disco
|
from disco.test import TestCase, TestJob
class InitJob(TestJob):
sort = False
@staticmethod
def map_reader(stream, size, url, params):
params.x = 10
return (stream, size, url)
@staticmethod
def map_init(iter, params):
assert hasattr(params, 'x')
iter.next()
params['x'] += 100
@staticmethod
def map(e, params):
yield e, int(e) + params['x']
@staticmethod
def reduce_init(iter, params):
params['y'] = 1000
@staticmethod
def reduce(iter, params):
for k, v in iter:
yield k, int(v) + params['y']
class InitTestCase(TestCase):
def serve(self, path):
return 'skipthis\n' + ('%s\n' % path) * 10
def runTest(self):
self.job = InitJob().run(input=self.test_server.urls(range(10)))
results = list(self.results(self.job))
for k, v in results:
self.assertEquals(int(k) + 1110, int(v))
self.assertEquals(len(results), 100)
Revert "added a test for the map_reader before map_init -case which fails currently"
(deprecate init functions instead)
This reverts commit 88551bf444b7b358fea8e7eb4475df2c5d87ceeb.
|
from disco.test import TestCase, TestJob
class InitJob(TestJob):
params = {'x': 10}
sort = False
@staticmethod
def map_init(iter, params):
iter.next()
params['x'] += 100
@staticmethod
def map(e, params):
yield e, int(e) + params['x']
@staticmethod
def reduce_init(iter, params):
params['y'] = 1000
@staticmethod
def reduce(iter, params):
for k, v in iter:
yield k, int(v) + params['y']
class InitTestCase(TestCase):
def serve(self, path):
return 'skipthis\n' + ('%s\n' % path) * 10
def runTest(self):
self.job = InitJob().run(input=self.test_server.urls(range(10)))
results = list(self.results(self.job))
for k, v in results:
self.assertEquals(int(k) + 1110, int(v))
self.assertEquals(len(results), 100)
|
<commit_before>from disco.test import TestCase, TestJob
class InitJob(TestJob):
sort = False
@staticmethod
def map_reader(stream, size, url, params):
params.x = 10
return (stream, size, url)
@staticmethod
def map_init(iter, params):
assert hasattr(params, 'x')
iter.next()
params['x'] += 100
@staticmethod
def map(e, params):
yield e, int(e) + params['x']
@staticmethod
def reduce_init(iter, params):
params['y'] = 1000
@staticmethod
def reduce(iter, params):
for k, v in iter:
yield k, int(v) + params['y']
class InitTestCase(TestCase):
def serve(self, path):
return 'skipthis\n' + ('%s\n' % path) * 10
def runTest(self):
self.job = InitJob().run(input=self.test_server.urls(range(10)))
results = list(self.results(self.job))
for k, v in results:
self.assertEquals(int(k) + 1110, int(v))
self.assertEquals(len(results), 100)
<commit_msg>Revert "added a test for the map_reader before map_init -case which fails currently"
(deprecate init functions instead)
This reverts commit 88551bf444b7b358fea8e7eb4475df2c5d87ceeb.<commit_after>
|
from disco.test import TestCase, TestJob
class InitJob(TestJob):
params = {'x': 10}
sort = False
@staticmethod
def map_init(iter, params):
iter.next()
params['x'] += 100
@staticmethod
def map(e, params):
yield e, int(e) + params['x']
@staticmethod
def reduce_init(iter, params):
params['y'] = 1000
@staticmethod
def reduce(iter, params):
for k, v in iter:
yield k, int(v) + params['y']
class InitTestCase(TestCase):
def serve(self, path):
return 'skipthis\n' + ('%s\n' % path) * 10
def runTest(self):
self.job = InitJob().run(input=self.test_server.urls(range(10)))
results = list(self.results(self.job))
for k, v in results:
self.assertEquals(int(k) + 1110, int(v))
self.assertEquals(len(results), 100)
|
from disco.test import TestCase, TestJob
class InitJob(TestJob):
sort = False
@staticmethod
def map_reader(stream, size, url, params):
params.x = 10
return (stream, size, url)
@staticmethod
def map_init(iter, params):
assert hasattr(params, 'x')
iter.next()
params['x'] += 100
@staticmethod
def map(e, params):
yield e, int(e) + params['x']
@staticmethod
def reduce_init(iter, params):
params['y'] = 1000
@staticmethod
def reduce(iter, params):
for k, v in iter:
yield k, int(v) + params['y']
class InitTestCase(TestCase):
def serve(self, path):
return 'skipthis\n' + ('%s\n' % path) * 10
def runTest(self):
self.job = InitJob().run(input=self.test_server.urls(range(10)))
results = list(self.results(self.job))
for k, v in results:
self.assertEquals(int(k) + 1110, int(v))
self.assertEquals(len(results), 100)
Revert "added a test for the map_reader before map_init -case which fails currently"
(deprecate init functions instead)
This reverts commit 88551bf444b7b358fea8e7eb4475df2c5d87ceeb.from disco.test import TestCase, TestJob
class InitJob(TestJob):
params = {'x': 10}
sort = False
@staticmethod
def map_init(iter, params):
iter.next()
params['x'] += 100
@staticmethod
def map(e, params):
yield e, int(e) + params['x']
@staticmethod
def reduce_init(iter, params):
params['y'] = 1000
@staticmethod
def reduce(iter, params):
for k, v in iter:
yield k, int(v) + params['y']
class InitTestCase(TestCase):
def serve(self, path):
return 'skipthis\n' + ('%s\n' % path) * 10
def runTest(self):
self.job = InitJob().run(input=self.test_server.urls(range(10)))
results = list(self.results(self.job))
for k, v in results:
self.assertEquals(int(k) + 1110, int(v))
self.assertEquals(len(results), 100)
|
<commit_before>from disco.test import TestCase, TestJob
class InitJob(TestJob):
sort = False
@staticmethod
def map_reader(stream, size, url, params):
params.x = 10
return (stream, size, url)
@staticmethod
def map_init(iter, params):
assert hasattr(params, 'x')
iter.next()
params['x'] += 100
@staticmethod
def map(e, params):
yield e, int(e) + params['x']
@staticmethod
def reduce_init(iter, params):
params['y'] = 1000
@staticmethod
def reduce(iter, params):
for k, v in iter:
yield k, int(v) + params['y']
class InitTestCase(TestCase):
def serve(self, path):
return 'skipthis\n' + ('%s\n' % path) * 10
def runTest(self):
self.job = InitJob().run(input=self.test_server.urls(range(10)))
results = list(self.results(self.job))
for k, v in results:
self.assertEquals(int(k) + 1110, int(v))
self.assertEquals(len(results), 100)
<commit_msg>Revert "added a test for the map_reader before map_init -case which fails currently"
(deprecate init functions instead)
This reverts commit 88551bf444b7b358fea8e7eb4475df2c5d87ceeb.<commit_after>from disco.test import TestCase, TestJob
class InitJob(TestJob):
params = {'x': 10}
sort = False
@staticmethod
def map_init(iter, params):
iter.next()
params['x'] += 100
@staticmethod
def map(e, params):
yield e, int(e) + params['x']
@staticmethod
def reduce_init(iter, params):
params['y'] = 1000
@staticmethod
def reduce(iter, params):
for k, v in iter:
yield k, int(v) + params['y']
class InitTestCase(TestCase):
def serve(self, path):
return 'skipthis\n' + ('%s\n' % path) * 10
def runTest(self):
self.job = InitJob().run(input=self.test_server.urls(range(10)))
results = list(self.results(self.job))
for k, v in results:
self.assertEquals(int(k) + 1110, int(v))
self.assertEquals(len(results), 100)
|
fefa46e21724fcd87cda0fa58101e1a74a31adec
|
molly/apps/places/importers/naptan.py
|
molly/apps/places/importers/naptan.py
|
from datetime import timedelta
import httplib
from tempfile import TemporaryFile
from zipfile import ZipFile
from celery.schedules import schedule
from molly.apps.places.parsers.naptan import NaptanParser
class NaptanImporter(object):
IMPORTER_NAME = 'naptan'
IMPORT_SCHEDULE = schedule(run_every=timedelta(weeks=1))
HTTP_HOST = "www.dft.gov.uk"
REMOTE_PATH = "/NaPTAN/snapshot/NaPTANxml.zip"
def __init__(self, config):
self._http_connection = httplib.HTTPConnection(self.HTTP_HOST)
self._url = "http://%s%s" % (self.HTTP_HOST, self.REMOTE_PATH)
def _get_file_from_url(self):
temporary = TemporaryFile()
self._http_connection.request('GET', self._url)
temporary.write(self._http_connection.getresponse().read())
return ZipFile(temporary).open('NaPTAN.xml')
def load(self):
parser = NaptanParser()
for stop in parser.import_from_file(self._get_file_from_url(), self._url):
self.stop_service.insert_and_merge(stop)
Provider = NaptanImporter
|
from datetime import timedelta
import httplib
from tempfile import TemporaryFile
from zipfile import ZipFile
from celery.schedules import schedule
from molly.apps.places.parsers.naptan import NaptanParser
class NaptanImporter(object):
IMPORTER_NAME = 'naptan'
IMPORT_SCHEDULE = schedule(run_every=timedelta(weeks=1))
HTTP_HOST = "www.dft.gov.uk"
REMOTE_PATH = "/NaPTAN/snapshot/NaPTANxml.zip"
def __init__(self, config):
self._http_connection = httplib.HTTPConnection(self.HTTP_HOST)
self._url = "http://%s%s" % (self.HTTP_HOST, self.REMOTE_PATH)
def _get_file_from_url(self):
temporary = TemporaryFile()
self._http_connection.request('GET', self._url)
temporary.write(self._http_connection.getresponse().read())
return ZipFile(temporary).open('NaPTAN.xml')
def load(self):
parser = NaptanParser()
for stop in parser.import_from_file(self._get_file_from_url(), self._url):
self.poi_service.add_or_update(stop)
Provider = NaptanImporter
|
Update the importer to use the places service
|
Update the importer to use the places service
|
Python
|
apache-2.0
|
ManchesterIO/mollyproject-next,ManchesterIO/mollyproject-next,ManchesterIO/mollyproject-next
|
from datetime import timedelta
import httplib
from tempfile import TemporaryFile
from zipfile import ZipFile
from celery.schedules import schedule
from molly.apps.places.parsers.naptan import NaptanParser
class NaptanImporter(object):
IMPORTER_NAME = 'naptan'
IMPORT_SCHEDULE = schedule(run_every=timedelta(weeks=1))
HTTP_HOST = "www.dft.gov.uk"
REMOTE_PATH = "/NaPTAN/snapshot/NaPTANxml.zip"
def __init__(self, config):
self._http_connection = httplib.HTTPConnection(self.HTTP_HOST)
self._url = "http://%s%s" % (self.HTTP_HOST, self.REMOTE_PATH)
def _get_file_from_url(self):
temporary = TemporaryFile()
self._http_connection.request('GET', self._url)
temporary.write(self._http_connection.getresponse().read())
return ZipFile(temporary).open('NaPTAN.xml')
def load(self):
parser = NaptanParser()
for stop in parser.import_from_file(self._get_file_from_url(), self._url):
self.stop_service.insert_and_merge(stop)
Provider = NaptanImporter
Update the importer to use the places service
|
from datetime import timedelta
import httplib
from tempfile import TemporaryFile
from zipfile import ZipFile
from celery.schedules import schedule
from molly.apps.places.parsers.naptan import NaptanParser
class NaptanImporter(object):
IMPORTER_NAME = 'naptan'
IMPORT_SCHEDULE = schedule(run_every=timedelta(weeks=1))
HTTP_HOST = "www.dft.gov.uk"
REMOTE_PATH = "/NaPTAN/snapshot/NaPTANxml.zip"
def __init__(self, config):
self._http_connection = httplib.HTTPConnection(self.HTTP_HOST)
self._url = "http://%s%s" % (self.HTTP_HOST, self.REMOTE_PATH)
def _get_file_from_url(self):
temporary = TemporaryFile()
self._http_connection.request('GET', self._url)
temporary.write(self._http_connection.getresponse().read())
return ZipFile(temporary).open('NaPTAN.xml')
def load(self):
parser = NaptanParser()
for stop in parser.import_from_file(self._get_file_from_url(), self._url):
self.poi_service.add_or_update(stop)
Provider = NaptanImporter
|
<commit_before>from datetime import timedelta
import httplib
from tempfile import TemporaryFile
from zipfile import ZipFile
from celery.schedules import schedule
from molly.apps.places.parsers.naptan import NaptanParser
class NaptanImporter(object):
IMPORTER_NAME = 'naptan'
IMPORT_SCHEDULE = schedule(run_every=timedelta(weeks=1))
HTTP_HOST = "www.dft.gov.uk"
REMOTE_PATH = "/NaPTAN/snapshot/NaPTANxml.zip"
def __init__(self, config):
self._http_connection = httplib.HTTPConnection(self.HTTP_HOST)
self._url = "http://%s%s" % (self.HTTP_HOST, self.REMOTE_PATH)
def _get_file_from_url(self):
temporary = TemporaryFile()
self._http_connection.request('GET', self._url)
temporary.write(self._http_connection.getresponse().read())
return ZipFile(temporary).open('NaPTAN.xml')
def load(self):
parser = NaptanParser()
for stop in parser.import_from_file(self._get_file_from_url(), self._url):
self.stop_service.insert_and_merge(stop)
Provider = NaptanImporter
<commit_msg>Update the importer to use the places service<commit_after>
|
from datetime import timedelta
import httplib
from tempfile import TemporaryFile
from zipfile import ZipFile
from celery.schedules import schedule
from molly.apps.places.parsers.naptan import NaptanParser
class NaptanImporter(object):
IMPORTER_NAME = 'naptan'
IMPORT_SCHEDULE = schedule(run_every=timedelta(weeks=1))
HTTP_HOST = "www.dft.gov.uk"
REMOTE_PATH = "/NaPTAN/snapshot/NaPTANxml.zip"
def __init__(self, config):
self._http_connection = httplib.HTTPConnection(self.HTTP_HOST)
self._url = "http://%s%s" % (self.HTTP_HOST, self.REMOTE_PATH)
def _get_file_from_url(self):
temporary = TemporaryFile()
self._http_connection.request('GET', self._url)
temporary.write(self._http_connection.getresponse().read())
return ZipFile(temporary).open('NaPTAN.xml')
def load(self):
parser = NaptanParser()
for stop in parser.import_from_file(self._get_file_from_url(), self._url):
self.poi_service.add_or_update(stop)
Provider = NaptanImporter
|
from datetime import timedelta
import httplib
from tempfile import TemporaryFile
from zipfile import ZipFile
from celery.schedules import schedule
from molly.apps.places.parsers.naptan import NaptanParser
class NaptanImporter(object):
IMPORTER_NAME = 'naptan'
IMPORT_SCHEDULE = schedule(run_every=timedelta(weeks=1))
HTTP_HOST = "www.dft.gov.uk"
REMOTE_PATH = "/NaPTAN/snapshot/NaPTANxml.zip"
def __init__(self, config):
self._http_connection = httplib.HTTPConnection(self.HTTP_HOST)
self._url = "http://%s%s" % (self.HTTP_HOST, self.REMOTE_PATH)
def _get_file_from_url(self):
temporary = TemporaryFile()
self._http_connection.request('GET', self._url)
temporary.write(self._http_connection.getresponse().read())
return ZipFile(temporary).open('NaPTAN.xml')
def load(self):
parser = NaptanParser()
for stop in parser.import_from_file(self._get_file_from_url(), self._url):
self.stop_service.insert_and_merge(stop)
Provider = NaptanImporter
Update the importer to use the places servicefrom datetime import timedelta
import httplib
from tempfile import TemporaryFile
from zipfile import ZipFile
from celery.schedules import schedule
from molly.apps.places.parsers.naptan import NaptanParser
class NaptanImporter(object):
IMPORTER_NAME = 'naptan'
IMPORT_SCHEDULE = schedule(run_every=timedelta(weeks=1))
HTTP_HOST = "www.dft.gov.uk"
REMOTE_PATH = "/NaPTAN/snapshot/NaPTANxml.zip"
def __init__(self, config):
self._http_connection = httplib.HTTPConnection(self.HTTP_HOST)
self._url = "http://%s%s" % (self.HTTP_HOST, self.REMOTE_PATH)
def _get_file_from_url(self):
temporary = TemporaryFile()
self._http_connection.request('GET', self._url)
temporary.write(self._http_connection.getresponse().read())
return ZipFile(temporary).open('NaPTAN.xml')
def load(self):
parser = NaptanParser()
for stop in parser.import_from_file(self._get_file_from_url(), self._url):
self.poi_service.add_or_update(stop)
Provider = NaptanImporter
|
<commit_before>from datetime import timedelta
import httplib
from tempfile import TemporaryFile
from zipfile import ZipFile
from celery.schedules import schedule
from molly.apps.places.parsers.naptan import NaptanParser
class NaptanImporter(object):
IMPORTER_NAME = 'naptan'
IMPORT_SCHEDULE = schedule(run_every=timedelta(weeks=1))
HTTP_HOST = "www.dft.gov.uk"
REMOTE_PATH = "/NaPTAN/snapshot/NaPTANxml.zip"
def __init__(self, config):
self._http_connection = httplib.HTTPConnection(self.HTTP_HOST)
self._url = "http://%s%s" % (self.HTTP_HOST, self.REMOTE_PATH)
def _get_file_from_url(self):
temporary = TemporaryFile()
self._http_connection.request('GET', self._url)
temporary.write(self._http_connection.getresponse().read())
return ZipFile(temporary).open('NaPTAN.xml')
def load(self):
parser = NaptanParser()
for stop in parser.import_from_file(self._get_file_from_url(), self._url):
self.stop_service.insert_and_merge(stop)
Provider = NaptanImporter
<commit_msg>Update the importer to use the places service<commit_after>from datetime import timedelta
import httplib
from tempfile import TemporaryFile
from zipfile import ZipFile
from celery.schedules import schedule
from molly.apps.places.parsers.naptan import NaptanParser
class NaptanImporter(object):
IMPORTER_NAME = 'naptan'
IMPORT_SCHEDULE = schedule(run_every=timedelta(weeks=1))
HTTP_HOST = "www.dft.gov.uk"
REMOTE_PATH = "/NaPTAN/snapshot/NaPTANxml.zip"
def __init__(self, config):
self._http_connection = httplib.HTTPConnection(self.HTTP_HOST)
self._url = "http://%s%s" % (self.HTTP_HOST, self.REMOTE_PATH)
def _get_file_from_url(self):
temporary = TemporaryFile()
self._http_connection.request('GET', self._url)
temporary.write(self._http_connection.getresponse().read())
return ZipFile(temporary).open('NaPTAN.xml')
def load(self):
parser = NaptanParser()
for stop in parser.import_from_file(self._get_file_from_url(), self._url):
self.poi_service.add_or_update(stop)
Provider = NaptanImporter
|
ff0da634e1fa0f8b190a3ba2cac3a03f7df75f91
|
memegen/test/test_routes__common.py
|
memegen/test/test_routes__common.py
|
# pylint: disable=unused-variable
from unittest.mock import patch, Mock
from memegen.app import create_app
from memegen.settings import get_config
from memegen.routes._common import display
def describe_display():
app = create_app(get_config('test'))
app.config['GOOGLE_ANALYTICS_TID'] = 'my_tid'
request_html = Mock()
request_html.headers.get = Mock(return_value="text/html")
request_html.path = "it's a path"
@patch('memegen.routes._common.request', request_html)
def it_returns_html_for_browsers():
with app.test_request_context():
html = display("my_title", "my_path", raw=True)
print(html)
assert "<title>my_title</title>" in html
assert 'url("it\'s a path")' in html
assert "ga('create', 'my_tid', 'auto');" in html
|
# pylint: disable=unused-variable,expression-not-assigned
from unittest.mock import patch, call, Mock
import pytest
from expecter import expect
from memegen.app import create_app
from memegen.settings import get_config
from memegen.routes._common import display
def describe_display():
@pytest.fixture
def app():
app = create_app(get_config('test'))
app.config['GOOGLE_ANALYTICS_TID'] = 'my_tid'
return app
request_html = Mock(path="it's a path")
request_html.headers.get = Mock(return_value="text/html")
request_image = Mock(path="it's a path")
request_image.headers.get = Mock(return_value="(not a browser)")
@patch('memegen.routes._common.request', request_html)
def it_returns_html_for_browsers(app):
with app.test_request_context():
html = display("my_title", "my_path", raw=True)
print(html)
assert "<title>my_title</title>" in html
assert 'url("it\'s a path")' in html
assert "ga('create', 'my_tid', 'auto');" in html
@patch('memegen.routes._common._track')
@patch('memegen.routes._common.send_file')
@patch('memegen.routes._common.request', request_image)
def it_returns_an_image_otherwise(mock_send_file, mock_track):
display("my_title", "my_path")
expect(mock_track.mock_calls) == [
call("my_title"),
]
expect(mock_send_file.mock_calls) == [
call("my_path", mimetype='image/jpeg'),
]
|
Test that a request defaults to sending an image
|
Test that a request defaults to sending an image
|
Python
|
mit
|
joshfriend/memegen,joshfriend/memegen,DanLindeman/memegen,DanLindeman/memegen,DanLindeman/memegen,DanLindeman/memegen,joshfriend/memegen,joshfriend/memegen
|
# pylint: disable=unused-variable
from unittest.mock import patch, Mock
from memegen.app import create_app
from memegen.settings import get_config
from memegen.routes._common import display
def describe_display():
app = create_app(get_config('test'))
app.config['GOOGLE_ANALYTICS_TID'] = 'my_tid'
request_html = Mock()
request_html.headers.get = Mock(return_value="text/html")
request_html.path = "it's a path"
@patch('memegen.routes._common.request', request_html)
def it_returns_html_for_browsers():
with app.test_request_context():
html = display("my_title", "my_path", raw=True)
print(html)
assert "<title>my_title</title>" in html
assert 'url("it\'s a path")' in html
assert "ga('create', 'my_tid', 'auto');" in html
Test that a request defaults to sending an image
|
# pylint: disable=unused-variable,expression-not-assigned
from unittest.mock import patch, call, Mock
import pytest
from expecter import expect
from memegen.app import create_app
from memegen.settings import get_config
from memegen.routes._common import display
def describe_display():
@pytest.fixture
def app():
app = create_app(get_config('test'))
app.config['GOOGLE_ANALYTICS_TID'] = 'my_tid'
return app
request_html = Mock(path="it's a path")
request_html.headers.get = Mock(return_value="text/html")
request_image = Mock(path="it's a path")
request_image.headers.get = Mock(return_value="(not a browser)")
@patch('memegen.routes._common.request', request_html)
def it_returns_html_for_browsers(app):
with app.test_request_context():
html = display("my_title", "my_path", raw=True)
print(html)
assert "<title>my_title</title>" in html
assert 'url("it\'s a path")' in html
assert "ga('create', 'my_tid', 'auto');" in html
@patch('memegen.routes._common._track')
@patch('memegen.routes._common.send_file')
@patch('memegen.routes._common.request', request_image)
def it_returns_an_image_otherwise(mock_send_file, mock_track):
display("my_title", "my_path")
expect(mock_track.mock_calls) == [
call("my_title"),
]
expect(mock_send_file.mock_calls) == [
call("my_path", mimetype='image/jpeg'),
]
|
<commit_before># pylint: disable=unused-variable
from unittest.mock import patch, Mock
from memegen.app import create_app
from memegen.settings import get_config
from memegen.routes._common import display
def describe_display():
app = create_app(get_config('test'))
app.config['GOOGLE_ANALYTICS_TID'] = 'my_tid'
request_html = Mock()
request_html.headers.get = Mock(return_value="text/html")
request_html.path = "it's a path"
@patch('memegen.routes._common.request', request_html)
def it_returns_html_for_browsers():
with app.test_request_context():
html = display("my_title", "my_path", raw=True)
print(html)
assert "<title>my_title</title>" in html
assert 'url("it\'s a path")' in html
assert "ga('create', 'my_tid', 'auto');" in html
<commit_msg>Test that a request defaults to sending an image<commit_after>
|
# pylint: disable=unused-variable,expression-not-assigned
from unittest.mock import patch, call, Mock
import pytest
from expecter import expect
from memegen.app import create_app
from memegen.settings import get_config
from memegen.routes._common import display
def describe_display():
@pytest.fixture
def app():
app = create_app(get_config('test'))
app.config['GOOGLE_ANALYTICS_TID'] = 'my_tid'
return app
request_html = Mock(path="it's a path")
request_html.headers.get = Mock(return_value="text/html")
request_image = Mock(path="it's a path")
request_image.headers.get = Mock(return_value="(not a browser)")
@patch('memegen.routes._common.request', request_html)
def it_returns_html_for_browsers(app):
with app.test_request_context():
html = display("my_title", "my_path", raw=True)
print(html)
assert "<title>my_title</title>" in html
assert 'url("it\'s a path")' in html
assert "ga('create', 'my_tid', 'auto');" in html
@patch('memegen.routes._common._track')
@patch('memegen.routes._common.send_file')
@patch('memegen.routes._common.request', request_image)
def it_returns_an_image_otherwise(mock_send_file, mock_track):
display("my_title", "my_path")
expect(mock_track.mock_calls) == [
call("my_title"),
]
expect(mock_send_file.mock_calls) == [
call("my_path", mimetype='image/jpeg'),
]
|
# pylint: disable=unused-variable
from unittest.mock import patch, Mock
from memegen.app import create_app
from memegen.settings import get_config
from memegen.routes._common import display
def describe_display():
app = create_app(get_config('test'))
app.config['GOOGLE_ANALYTICS_TID'] = 'my_tid'
request_html = Mock()
request_html.headers.get = Mock(return_value="text/html")
request_html.path = "it's a path"
@patch('memegen.routes._common.request', request_html)
def it_returns_html_for_browsers():
with app.test_request_context():
html = display("my_title", "my_path", raw=True)
print(html)
assert "<title>my_title</title>" in html
assert 'url("it\'s a path")' in html
assert "ga('create', 'my_tid', 'auto');" in html
Test that a request defaults to sending an image# pylint: disable=unused-variable,expression-not-assigned
from unittest.mock import patch, call, Mock
import pytest
from expecter import expect
from memegen.app import create_app
from memegen.settings import get_config
from memegen.routes._common import display
def describe_display():
@pytest.fixture
def app():
app = create_app(get_config('test'))
app.config['GOOGLE_ANALYTICS_TID'] = 'my_tid'
return app
request_html = Mock(path="it's a path")
request_html.headers.get = Mock(return_value="text/html")
request_image = Mock(path="it's a path")
request_image.headers.get = Mock(return_value="(not a browser)")
@patch('memegen.routes._common.request', request_html)
def it_returns_html_for_browsers(app):
with app.test_request_context():
html = display("my_title", "my_path", raw=True)
print(html)
assert "<title>my_title</title>" in html
assert 'url("it\'s a path")' in html
assert "ga('create', 'my_tid', 'auto');" in html
@patch('memegen.routes._common._track')
@patch('memegen.routes._common.send_file')
@patch('memegen.routes._common.request', request_image)
def it_returns_an_image_otherwise(mock_send_file, mock_track):
display("my_title", "my_path")
expect(mock_track.mock_calls) == [
call("my_title"),
]
expect(mock_send_file.mock_calls) == [
call("my_path", mimetype='image/jpeg'),
]
|
<commit_before># pylint: disable=unused-variable
from unittest.mock import patch, Mock
from memegen.app import create_app
from memegen.settings import get_config
from memegen.routes._common import display
def describe_display():
app = create_app(get_config('test'))
app.config['GOOGLE_ANALYTICS_TID'] = 'my_tid'
request_html = Mock()
request_html.headers.get = Mock(return_value="text/html")
request_html.path = "it's a path"
@patch('memegen.routes._common.request', request_html)
def it_returns_html_for_browsers():
with app.test_request_context():
html = display("my_title", "my_path", raw=True)
print(html)
assert "<title>my_title</title>" in html
assert 'url("it\'s a path")' in html
assert "ga('create', 'my_tid', 'auto');" in html
<commit_msg>Test that a request defaults to sending an image<commit_after># pylint: disable=unused-variable,expression-not-assigned
from unittest.mock import patch, call, Mock
import pytest
from expecter import expect
from memegen.app import create_app
from memegen.settings import get_config
from memegen.routes._common import display
def describe_display():
@pytest.fixture
def app():
app = create_app(get_config('test'))
app.config['GOOGLE_ANALYTICS_TID'] = 'my_tid'
return app
request_html = Mock(path="it's a path")
request_html.headers.get = Mock(return_value="text/html")
request_image = Mock(path="it's a path")
request_image.headers.get = Mock(return_value="(not a browser)")
@patch('memegen.routes._common.request', request_html)
def it_returns_html_for_browsers(app):
with app.test_request_context():
html = display("my_title", "my_path", raw=True)
print(html)
assert "<title>my_title</title>" in html
assert 'url("it\'s a path")' in html
assert "ga('create', 'my_tid', 'auto');" in html
@patch('memegen.routes._common._track')
@patch('memegen.routes._common.send_file')
@patch('memegen.routes._common.request', request_image)
def it_returns_an_image_otherwise(mock_send_file, mock_track):
display("my_title", "my_path")
expect(mock_track.mock_calls) == [
call("my_title"),
]
expect(mock_send_file.mock_calls) == [
call("my_path", mimetype='image/jpeg'),
]
|
ccf60e9e79b8b2db8cbf7918caf23314e8790134
|
lib/reporter.py
|
lib/reporter.py
|
#!/usr/bin/python
import sys
import os
name = sys.argv[1]
status = sys.stdin.readline()
status = status.rstrip(os.linesep)
print("<%s>" % name)
print("\t<status=\"%s\" />" % status)
if status != "SKIP":
print("\t<outcome>")
for line in sys.stdin:
# Escaping, ... !
print(line.rstrip(os.linesep))
print("\t</outcome>")
print("</%s>" % name)
|
#!/usr/bin/python
import sys
import os
name = sys.argv[1]
status = sys.stdin.readline()
status = status.rstrip(os.linesep)
print("<%s status=\"%s\">" % (name, status))
print("\t<outcome>")
for line in sys.stdin:
# Escaping, ... !
print(line.rstrip(os.linesep))
print("\t</outcome>")
print("</%s>" % name)
|
Fix the XML format produced
|
Fix the XML format produced
|
Python
|
apache-2.0
|
CESNET/secant,CESNET/secant
|
#!/usr/bin/python
import sys
import os
name = sys.argv[1]
status = sys.stdin.readline()
status = status.rstrip(os.linesep)
print("<%s>" % name)
print("\t<status=\"%s\" />" % status)
if status != "SKIP":
print("\t<outcome>")
for line in sys.stdin:
# Escaping, ... !
print(line.rstrip(os.linesep))
print("\t</outcome>")
print("</%s>" % name)
Fix the XML format produced
|
#!/usr/bin/python
import sys
import os
name = sys.argv[1]
status = sys.stdin.readline()
status = status.rstrip(os.linesep)
print("<%s status=\"%s\">" % (name, status))
print("\t<outcome>")
for line in sys.stdin:
# Escaping, ... !
print(line.rstrip(os.linesep))
print("\t</outcome>")
print("</%s>" % name)
|
<commit_before>#!/usr/bin/python
import sys
import os
name = sys.argv[1]
status = sys.stdin.readline()
status = status.rstrip(os.linesep)
print("<%s>" % name)
print("\t<status=\"%s\" />" % status)
if status != "SKIP":
print("\t<outcome>")
for line in sys.stdin:
# Escaping, ... !
print(line.rstrip(os.linesep))
print("\t</outcome>")
print("</%s>" % name)
<commit_msg>Fix the XML format produced<commit_after>
|
#!/usr/bin/python
import sys
import os
name = sys.argv[1]
status = sys.stdin.readline()
status = status.rstrip(os.linesep)
print("<%s status=\"%s\">" % (name, status))
print("\t<outcome>")
for line in sys.stdin:
# Escaping, ... !
print(line.rstrip(os.linesep))
print("\t</outcome>")
print("</%s>" % name)
|
#!/usr/bin/python
import sys
import os
name = sys.argv[1]
status = sys.stdin.readline()
status = status.rstrip(os.linesep)
print("<%s>" % name)
print("\t<status=\"%s\" />" % status)
if status != "SKIP":
print("\t<outcome>")
for line in sys.stdin:
# Escaping, ... !
print(line.rstrip(os.linesep))
print("\t</outcome>")
print("</%s>" % name)
Fix the XML format produced#!/usr/bin/python
import sys
import os
name = sys.argv[1]
status = sys.stdin.readline()
status = status.rstrip(os.linesep)
print("<%s status=\"%s\">" % (name, status))
print("\t<outcome>")
for line in sys.stdin:
# Escaping, ... !
print(line.rstrip(os.linesep))
print("\t</outcome>")
print("</%s>" % name)
|
<commit_before>#!/usr/bin/python
import sys
import os
name = sys.argv[1]
status = sys.stdin.readline()
status = status.rstrip(os.linesep)
print("<%s>" % name)
print("\t<status=\"%s\" />" % status)
if status != "SKIP":
print("\t<outcome>")
for line in sys.stdin:
# Escaping, ... !
print(line.rstrip(os.linesep))
print("\t</outcome>")
print("</%s>" % name)
<commit_msg>Fix the XML format produced<commit_after>#!/usr/bin/python
import sys
import os
name = sys.argv[1]
status = sys.stdin.readline()
status = status.rstrip(os.linesep)
print("<%s status=\"%s\">" % (name, status))
print("\t<outcome>")
for line in sys.stdin:
# Escaping, ... !
print(line.rstrip(os.linesep))
print("\t</outcome>")
print("</%s>" % name)
|
d2bac1fe8dc6d90d0d680a97aec0646ad9674bae
|
qrl/core/ntp.py
|
qrl/core/ntp.py
|
# coding=utf-8
# Distributed under the MIT software license, see the accompanying
# file LICENSE or http://www.opensource.org/licenses/mit-license.php.
import sys
from time import time
from ntplib import NTPClient
from qrl.core import logger
ntp_server = 'pool.ntp.org'
version = 3
times = 5
drift = None
def get_ntp_response():
try:
ntp_client = NTPClient()
response = ntp_client.request(ntp_server, version=version)
except Exception as e:
logger.exception(e)
sys.exit(0)
return response
def getNTP():
ntp_timestamp = 0
response = get_ntp_response()
if response:
ntp_timestamp = int(response.tx_time)
return ntp_timestamp
def setDrift():
global drift
response = get_ntp_response()
if not response:
return response
drift = response.offset
def getTime():
global drift
curr_time = drift + int(time())
return curr_time
|
# coding=utf-8
# Distributed under the MIT software license, see the accompanying
# file LICENSE or http://www.opensource.org/licenses/mit-license.php.
import sys
from time import time
from ntplib import NTPClient
from qrl.core import logger
ntp_servers = ['pool.ntp.org', 'ntp.ubuntu.com']
NTP_VERSION = 3
NTP_RETRIES = 6
drift = None
def get_ntp_response():
for retry in range(NTP_RETRIES):
ntp_server = ntp_servers[retry % len(ntp_servers)]
try:
ntp_client = NTPClient()
response = ntp_client.request(ntp_server, version=NTP_VERSION)
except Exception as e:
logger.warning(e)
continue
return response
# FIXME: Provide some proper clean before exiting
logger.fatal("Could not contact NTP servers after %d retries", NTP_RETRIES)
sys.exit(-1)
def getNTP():
ntp_timestamp = 0
response = get_ntp_response()
if response:
ntp_timestamp = int(response.tx_time)
return ntp_timestamp
def setDrift():
global drift
response = get_ntp_response()
if not response:
return response
drift = response.offset
def getTime():
global drift
curr_time = drift + int(time())
return curr_time
|
Support multiple servers and retries
|
Support multiple servers and retries
|
Python
|
mit
|
jleni/QRL,elliottdehn/QRL,randomshinichi/QRL,elliottdehn/QRL,theQRL/QRL,theQRL/QRL,cyyber/QRL,randomshinichi/QRL,jleni/QRL,cyyber/QRL,elliottdehn/QRL,elliottdehn/QRL
|
# coding=utf-8
# Distributed under the MIT software license, see the accompanying
# file LICENSE or http://www.opensource.org/licenses/mit-license.php.
import sys
from time import time
from ntplib import NTPClient
from qrl.core import logger
ntp_server = 'pool.ntp.org'
version = 3
times = 5
drift = None
def get_ntp_response():
try:
ntp_client = NTPClient()
response = ntp_client.request(ntp_server, version=version)
except Exception as e:
logger.exception(e)
sys.exit(0)
return response
def getNTP():
ntp_timestamp = 0
response = get_ntp_response()
if response:
ntp_timestamp = int(response.tx_time)
return ntp_timestamp
def setDrift():
global drift
response = get_ntp_response()
if not response:
return response
drift = response.offset
def getTime():
global drift
curr_time = drift + int(time())
return curr_time
Support multiple servers and retries
|
# coding=utf-8
# Distributed under the MIT software license, see the accompanying
# file LICENSE or http://www.opensource.org/licenses/mit-license.php.
import sys
from time import time
from ntplib import NTPClient
from qrl.core import logger
ntp_servers = ['pool.ntp.org', 'ntp.ubuntu.com']
NTP_VERSION = 3
NTP_RETRIES = 6
drift = None
def get_ntp_response():
for retry in range(NTP_RETRIES):
ntp_server = ntp_servers[retry % len(ntp_servers)]
try:
ntp_client = NTPClient()
response = ntp_client.request(ntp_server, version=NTP_VERSION)
except Exception as e:
logger.warning(e)
continue
return response
# FIXME: Provide some proper clean before exiting
logger.fatal("Could not contact NTP servers after %d retries", NTP_RETRIES)
sys.exit(-1)
def getNTP():
ntp_timestamp = 0
response = get_ntp_response()
if response:
ntp_timestamp = int(response.tx_time)
return ntp_timestamp
def setDrift():
global drift
response = get_ntp_response()
if not response:
return response
drift = response.offset
def getTime():
global drift
curr_time = drift + int(time())
return curr_time
|
<commit_before># coding=utf-8
# Distributed under the MIT software license, see the accompanying
# file LICENSE or http://www.opensource.org/licenses/mit-license.php.
import sys
from time import time
from ntplib import NTPClient
from qrl.core import logger
ntp_server = 'pool.ntp.org'
version = 3
times = 5
drift = None
def get_ntp_response():
try:
ntp_client = NTPClient()
response = ntp_client.request(ntp_server, version=version)
except Exception as e:
logger.exception(e)
sys.exit(0)
return response
def getNTP():
ntp_timestamp = 0
response = get_ntp_response()
if response:
ntp_timestamp = int(response.tx_time)
return ntp_timestamp
def setDrift():
global drift
response = get_ntp_response()
if not response:
return response
drift = response.offset
def getTime():
global drift
curr_time = drift + int(time())
return curr_time
<commit_msg>Support multiple servers and retries<commit_after>
|
# coding=utf-8
# Distributed under the MIT software license, see the accompanying
# file LICENSE or http://www.opensource.org/licenses/mit-license.php.
import sys
from time import time
from ntplib import NTPClient
from qrl.core import logger
ntp_servers = ['pool.ntp.org', 'ntp.ubuntu.com']
NTP_VERSION = 3
NTP_RETRIES = 6
drift = None
def get_ntp_response():
for retry in range(NTP_RETRIES):
ntp_server = ntp_servers[retry % len(ntp_servers)]
try:
ntp_client = NTPClient()
response = ntp_client.request(ntp_server, version=NTP_VERSION)
except Exception as e:
logger.warning(e)
continue
return response
# FIXME: Provide some proper clean before exiting
logger.fatal("Could not contact NTP servers after %d retries", NTP_RETRIES)
sys.exit(-1)
def getNTP():
ntp_timestamp = 0
response = get_ntp_response()
if response:
ntp_timestamp = int(response.tx_time)
return ntp_timestamp
def setDrift():
global drift
response = get_ntp_response()
if not response:
return response
drift = response.offset
def getTime():
global drift
curr_time = drift + int(time())
return curr_time
|
# coding=utf-8
# Distributed under the MIT software license, see the accompanying
# file LICENSE or http://www.opensource.org/licenses/mit-license.php.
import sys
from time import time
from ntplib import NTPClient
from qrl.core import logger
ntp_server = 'pool.ntp.org'
version = 3
times = 5
drift = None
def get_ntp_response():
try:
ntp_client = NTPClient()
response = ntp_client.request(ntp_server, version=version)
except Exception as e:
logger.exception(e)
sys.exit(0)
return response
def getNTP():
ntp_timestamp = 0
response = get_ntp_response()
if response:
ntp_timestamp = int(response.tx_time)
return ntp_timestamp
def setDrift():
global drift
response = get_ntp_response()
if not response:
return response
drift = response.offset
def getTime():
global drift
curr_time = drift + int(time())
return curr_time
Support multiple servers and retries# coding=utf-8
# Distributed under the MIT software license, see the accompanying
# file LICENSE or http://www.opensource.org/licenses/mit-license.php.
import sys
from time import time
from ntplib import NTPClient
from qrl.core import logger
ntp_servers = ['pool.ntp.org', 'ntp.ubuntu.com']
NTP_VERSION = 3
NTP_RETRIES = 6
drift = None
def get_ntp_response():
for retry in range(NTP_RETRIES):
ntp_server = ntp_servers[retry % len(ntp_servers)]
try:
ntp_client = NTPClient()
response = ntp_client.request(ntp_server, version=NTP_VERSION)
except Exception as e:
logger.warning(e)
continue
return response
# FIXME: Provide some proper clean before exiting
logger.fatal("Could not contact NTP servers after %d retries", NTP_RETRIES)
sys.exit(-1)
def getNTP():
ntp_timestamp = 0
response = get_ntp_response()
if response:
ntp_timestamp = int(response.tx_time)
return ntp_timestamp
def setDrift():
global drift
response = get_ntp_response()
if not response:
return response
drift = response.offset
def getTime():
global drift
curr_time = drift + int(time())
return curr_time
|
<commit_before># coding=utf-8
# Distributed under the MIT software license, see the accompanying
# file LICENSE or http://www.opensource.org/licenses/mit-license.php.
import sys
from time import time
from ntplib import NTPClient
from qrl.core import logger
ntp_server = 'pool.ntp.org'
version = 3
times = 5
drift = None
def get_ntp_response():
try:
ntp_client = NTPClient()
response = ntp_client.request(ntp_server, version=version)
except Exception as e:
logger.exception(e)
sys.exit(0)
return response
def getNTP():
ntp_timestamp = 0
response = get_ntp_response()
if response:
ntp_timestamp = int(response.tx_time)
return ntp_timestamp
def setDrift():
global drift
response = get_ntp_response()
if not response:
return response
drift = response.offset
def getTime():
global drift
curr_time = drift + int(time())
return curr_time
<commit_msg>Support multiple servers and retries<commit_after># coding=utf-8
# Distributed under the MIT software license, see the accompanying
# file LICENSE or http://www.opensource.org/licenses/mit-license.php.
import sys
from time import time
from ntplib import NTPClient
from qrl.core import logger
ntp_servers = ['pool.ntp.org', 'ntp.ubuntu.com']
NTP_VERSION = 3
NTP_RETRIES = 6
drift = None
def get_ntp_response():
for retry in range(NTP_RETRIES):
ntp_server = ntp_servers[retry % len(ntp_servers)]
try:
ntp_client = NTPClient()
response = ntp_client.request(ntp_server, version=NTP_VERSION)
except Exception as e:
logger.warning(e)
continue
return response
# FIXME: Provide some proper clean before exiting
logger.fatal("Could not contact NTP servers after %d retries", NTP_RETRIES)
sys.exit(-1)
def getNTP():
ntp_timestamp = 0
response = get_ntp_response()
if response:
ntp_timestamp = int(response.tx_time)
return ntp_timestamp
def setDrift():
global drift
response = get_ntp_response()
if not response:
return response
drift = response.offset
def getTime():
global drift
curr_time = drift + int(time())
return curr_time
|
83ed5ca9bc388dbe9b2d82510842a99b3a2e5ce7
|
src/personalisation/middleware.py
|
src/personalisation/middleware.py
|
from personalisation.models import AbstractBaseRule, Segment
class SegmentMiddleware(object):
"""Middleware for testing and putting a user in a segment"""
def __init__(self, get_response=None):
self.get_response = get_response
def __call__(self, request):
segments = Segment.objects.all().filter(status="enabled")
chosen_segments = []
for segment in segments:
rules = AbstractBaseRule.objects.filter(segment=segment).select_subclasses()
result = self.test_rules(rules, request)
if result:
self.add_segment_to_user(segment, request)
response = self.get_response(request)
print(request.session['segments'])
return response
def test_rules(self, rules, request):
for rule in rules:
result = rule.test_user(request)
if result is False:
return False
return True
def add_segment_to_user(self, segment, request):
if 'segments' not in request.session:
request.session['segments'] = []
if segment not in request.session['segments']:
request.session['segments'].append(segment.encoded_name())
|
from personalisation.models import AbstractBaseRule, Segment
class SegmentMiddleware(object):
"""Middleware for testing and putting a user in a segment"""
def __init__(self, get_response=None):
self.get_response = get_response
def __call__(self, request):
segments = Segment.objects.all().filter(status="enabled")
chosen_segments = []
for segment in segments:
rules = AbstractBaseRule.objects.filter(segment=segment).select_subclasses()
result = self.test_rules(rules, request)
if result:
self.add_segment_to_user(segment, request)
response = self.get_response(request)
if not request.session.get('segments'):
request.session['segments'] = []
print(request.session['segments'])
return response
def test_rules(self, rules, request):
for rule in rules:
result = rule.test_user(request)
if result is False:
return False
return True
def add_segment_to_user(self, segment, request):
if 'segments' not in request.session:
request.session['segments'] = []
if segment not in request.session['segments']:
request.session['segments'].append(segment.encoded_name())
|
Create empty 'segments' object in session if none exists
|
Create empty 'segments' object in session if none exists
|
Python
|
mit
|
LabD/wagtail-personalisation,LabD/wagtail-personalisation,LabD/wagtail-personalisation
|
from personalisation.models import AbstractBaseRule, Segment
class SegmentMiddleware(object):
"""Middleware for testing and putting a user in a segment"""
def __init__(self, get_response=None):
self.get_response = get_response
def __call__(self, request):
segments = Segment.objects.all().filter(status="enabled")
chosen_segments = []
for segment in segments:
rules = AbstractBaseRule.objects.filter(segment=segment).select_subclasses()
result = self.test_rules(rules, request)
if result:
self.add_segment_to_user(segment, request)
response = self.get_response(request)
print(request.session['segments'])
return response
def test_rules(self, rules, request):
for rule in rules:
result = rule.test_user(request)
if result is False:
return False
return True
def add_segment_to_user(self, segment, request):
if 'segments' not in request.session:
request.session['segments'] = []
if segment not in request.session['segments']:
request.session['segments'].append(segment.encoded_name())
Create empty 'segments' object in session if none exists
|
from personalisation.models import AbstractBaseRule, Segment
class SegmentMiddleware(object):
"""Middleware for testing and putting a user in a segment"""
def __init__(self, get_response=None):
self.get_response = get_response
def __call__(self, request):
segments = Segment.objects.all().filter(status="enabled")
chosen_segments = []
for segment in segments:
rules = AbstractBaseRule.objects.filter(segment=segment).select_subclasses()
result = self.test_rules(rules, request)
if result:
self.add_segment_to_user(segment, request)
response = self.get_response(request)
if not request.session.get('segments'):
request.session['segments'] = []
print(request.session['segments'])
return response
def test_rules(self, rules, request):
for rule in rules:
result = rule.test_user(request)
if result is False:
return False
return True
def add_segment_to_user(self, segment, request):
if 'segments' not in request.session:
request.session['segments'] = []
if segment not in request.session['segments']:
request.session['segments'].append(segment.encoded_name())
|
<commit_before>from personalisation.models import AbstractBaseRule, Segment
class SegmentMiddleware(object):
"""Middleware for testing and putting a user in a segment"""
def __init__(self, get_response=None):
self.get_response = get_response
def __call__(self, request):
segments = Segment.objects.all().filter(status="enabled")
chosen_segments = []
for segment in segments:
rules = AbstractBaseRule.objects.filter(segment=segment).select_subclasses()
result = self.test_rules(rules, request)
if result:
self.add_segment_to_user(segment, request)
response = self.get_response(request)
print(request.session['segments'])
return response
def test_rules(self, rules, request):
for rule in rules:
result = rule.test_user(request)
if result is False:
return False
return True
def add_segment_to_user(self, segment, request):
if 'segments' not in request.session:
request.session['segments'] = []
if segment not in request.session['segments']:
request.session['segments'].append(segment.encoded_name())
<commit_msg>Create empty 'segments' object in session if none exists<commit_after>
|
from personalisation.models import AbstractBaseRule, Segment
class SegmentMiddleware(object):
"""Middleware for testing and putting a user in a segment"""
def __init__(self, get_response=None):
self.get_response = get_response
def __call__(self, request):
segments = Segment.objects.all().filter(status="enabled")
chosen_segments = []
for segment in segments:
rules = AbstractBaseRule.objects.filter(segment=segment).select_subclasses()
result = self.test_rules(rules, request)
if result:
self.add_segment_to_user(segment, request)
response = self.get_response(request)
if not request.session.get('segments'):
request.session['segments'] = []
print(request.session['segments'])
return response
def test_rules(self, rules, request):
for rule in rules:
result = rule.test_user(request)
if result is False:
return False
return True
def add_segment_to_user(self, segment, request):
if 'segments' not in request.session:
request.session['segments'] = []
if segment not in request.session['segments']:
request.session['segments'].append(segment.encoded_name())
|
from personalisation.models import AbstractBaseRule, Segment
class SegmentMiddleware(object):
"""Middleware for testing and putting a user in a segment"""
def __init__(self, get_response=None):
self.get_response = get_response
def __call__(self, request):
segments = Segment.objects.all().filter(status="enabled")
chosen_segments = []
for segment in segments:
rules = AbstractBaseRule.objects.filter(segment=segment).select_subclasses()
result = self.test_rules(rules, request)
if result:
self.add_segment_to_user(segment, request)
response = self.get_response(request)
print(request.session['segments'])
return response
def test_rules(self, rules, request):
for rule in rules:
result = rule.test_user(request)
if result is False:
return False
return True
def add_segment_to_user(self, segment, request):
if 'segments' not in request.session:
request.session['segments'] = []
if segment not in request.session['segments']:
request.session['segments'].append(segment.encoded_name())
Create empty 'segments' object in session if none existsfrom personalisation.models import AbstractBaseRule, Segment
class SegmentMiddleware(object):
"""Middleware for testing and putting a user in a segment"""
def __init__(self, get_response=None):
self.get_response = get_response
def __call__(self, request):
segments = Segment.objects.all().filter(status="enabled")
chosen_segments = []
for segment in segments:
rules = AbstractBaseRule.objects.filter(segment=segment).select_subclasses()
result = self.test_rules(rules, request)
if result:
self.add_segment_to_user(segment, request)
response = self.get_response(request)
if not request.session.get('segments'):
request.session['segments'] = []
print(request.session['segments'])
return response
def test_rules(self, rules, request):
for rule in rules:
result = rule.test_user(request)
if result is False:
return False
return True
def add_segment_to_user(self, segment, request):
if 'segments' not in request.session:
request.session['segments'] = []
if segment not in request.session['segments']:
request.session['segments'].append(segment.encoded_name())
|
<commit_before>from personalisation.models import AbstractBaseRule, Segment
class SegmentMiddleware(object):
"""Middleware for testing and putting a user in a segment"""
def __init__(self, get_response=None):
self.get_response = get_response
def __call__(self, request):
segments = Segment.objects.all().filter(status="enabled")
chosen_segments = []
for segment in segments:
rules = AbstractBaseRule.objects.filter(segment=segment).select_subclasses()
result = self.test_rules(rules, request)
if result:
self.add_segment_to_user(segment, request)
response = self.get_response(request)
print(request.session['segments'])
return response
def test_rules(self, rules, request):
for rule in rules:
result = rule.test_user(request)
if result is False:
return False
return True
def add_segment_to_user(self, segment, request):
if 'segments' not in request.session:
request.session['segments'] = []
if segment not in request.session['segments']:
request.session['segments'].append(segment.encoded_name())
<commit_msg>Create empty 'segments' object in session if none exists<commit_after>from personalisation.models import AbstractBaseRule, Segment
class SegmentMiddleware(object):
"""Middleware for testing and putting a user in a segment"""
def __init__(self, get_response=None):
self.get_response = get_response
def __call__(self, request):
segments = Segment.objects.all().filter(status="enabled")
chosen_segments = []
for segment in segments:
rules = AbstractBaseRule.objects.filter(segment=segment).select_subclasses()
result = self.test_rules(rules, request)
if result:
self.add_segment_to_user(segment, request)
response = self.get_response(request)
if not request.session.get('segments'):
request.session['segments'] = []
print(request.session['segments'])
return response
def test_rules(self, rules, request):
for rule in rules:
result = rule.test_user(request)
if result is False:
return False
return True
def add_segment_to_user(self, segment, request):
if 'segments' not in request.session:
request.session['segments'] = []
if segment not in request.session['segments']:
request.session['segments'].append(segment.encoded_name())
|
c0e98c14813c966ecd9e6b47395cb336a244f090
|
discussion/forms.py
|
discussion/forms.py
|
from django import forms
from discussion.models import Comment, Post, Discussion
from notification.models import NoticeSetting
class CommentForm(forms.ModelForm):
class Meta:
exclude = ('user', 'post')
model = Comment
widgets = {
'body': forms.Textarea(attrs={'placeholder': 'Reply to this conversation'}),
}
class PostForm(forms.ModelForm):
class Meta:
exclude = ('user', 'discussion')
model = Post
widgets = {
'body': forms.Textarea(attrs={'placeholder': 'Start a conversation'}),
}
class SearchForm(forms.Form):
search = forms.CharField()
discussion = forms.ModelChoiceField(
required=False,
queryset=Discussion.objects,
empty_label='All discussions')
class SubscribeForm(forms.Form):
send = forms.ModelMultipleChoiceField(
NoticeSetting.objects,
required=False,
label=u'Notify me with updates from this discussion by',
widget=forms.CheckboxSelectMultiple)
def __init__(self, *args, **kwargs):
qs = kwargs.pop('qs')
super(SubscribeForm, self).__init__(*args, **kwargs)
self.fields['send'].queryset = qs
|
from django import forms
from django.utils.translation import ugettext_lazy as _
from discussion.models import Comment, Post, Discussion
from notification.models import NoticeSetting
class CommentForm(forms.ModelForm):
class Meta:
exclude = ('user', 'post')
model = Comment
widgets = {
'body': forms.Textarea(attrs={'placeholder': _('Reply to this conversation')}),
}
class PostForm(forms.ModelForm):
class Meta:
exclude = ('user', 'discussion')
model = Post
widgets = {
'body': forms.Textarea(attrs={'placeholder': _('Start a conversation')}),
}
class SearchForm(forms.Form):
search = forms.CharField()
discussion = forms.ModelChoiceField(
required=False,
queryset=Discussion.objects,
empty_label=_('All discussions'))
class SubscribeForm(forms.Form):
send = forms.ModelMultipleChoiceField(
NoticeSetting.objects,
required=False,
label=_('Notify me with updates from this discussion by'),
widget=forms.CheckboxSelectMultiple)
def __init__(self, *args, **kwargs):
qs = kwargs.pop('qs')
super(SubscribeForm, self).__init__(*args, **kwargs)
self.fields['send'].queryset = qs
|
Make the odd string translatable.
|
Make the odd string translatable.
|
Python
|
bsd-2-clause
|
incuna/django-discussion,lehins/lehins-discussion,lehins/lehins-discussion,incuna/django-discussion,lehins/lehins-discussion
|
from django import forms
from discussion.models import Comment, Post, Discussion
from notification.models import NoticeSetting
class CommentForm(forms.ModelForm):
class Meta:
exclude = ('user', 'post')
model = Comment
widgets = {
'body': forms.Textarea(attrs={'placeholder': 'Reply to this conversation'}),
}
class PostForm(forms.ModelForm):
class Meta:
exclude = ('user', 'discussion')
model = Post
widgets = {
'body': forms.Textarea(attrs={'placeholder': 'Start a conversation'}),
}
class SearchForm(forms.Form):
search = forms.CharField()
discussion = forms.ModelChoiceField(
required=False,
queryset=Discussion.objects,
empty_label='All discussions')
class SubscribeForm(forms.Form):
send = forms.ModelMultipleChoiceField(
NoticeSetting.objects,
required=False,
label=u'Notify me with updates from this discussion by',
widget=forms.CheckboxSelectMultiple)
def __init__(self, *args, **kwargs):
qs = kwargs.pop('qs')
super(SubscribeForm, self).__init__(*args, **kwargs)
self.fields['send'].queryset = qs
Make the odd string translatable.
|
from django import forms
from django.utils.translation import ugettext_lazy as _
from discussion.models import Comment, Post, Discussion
from notification.models import NoticeSetting
class CommentForm(forms.ModelForm):
class Meta:
exclude = ('user', 'post')
model = Comment
widgets = {
'body': forms.Textarea(attrs={'placeholder': _('Reply to this conversation')}),
}
class PostForm(forms.ModelForm):
class Meta:
exclude = ('user', 'discussion')
model = Post
widgets = {
'body': forms.Textarea(attrs={'placeholder': _('Start a conversation')}),
}
class SearchForm(forms.Form):
search = forms.CharField()
discussion = forms.ModelChoiceField(
required=False,
queryset=Discussion.objects,
empty_label=_('All discussions'))
class SubscribeForm(forms.Form):
send = forms.ModelMultipleChoiceField(
NoticeSetting.objects,
required=False,
label=_('Notify me with updates from this discussion by'),
widget=forms.CheckboxSelectMultiple)
def __init__(self, *args, **kwargs):
qs = kwargs.pop('qs')
super(SubscribeForm, self).__init__(*args, **kwargs)
self.fields['send'].queryset = qs
|
<commit_before>from django import forms
from discussion.models import Comment, Post, Discussion
from notification.models import NoticeSetting
class CommentForm(forms.ModelForm):
class Meta:
exclude = ('user', 'post')
model = Comment
widgets = {
'body': forms.Textarea(attrs={'placeholder': 'Reply to this conversation'}),
}
class PostForm(forms.ModelForm):
class Meta:
exclude = ('user', 'discussion')
model = Post
widgets = {
'body': forms.Textarea(attrs={'placeholder': 'Start a conversation'}),
}
class SearchForm(forms.Form):
search = forms.CharField()
discussion = forms.ModelChoiceField(
required=False,
queryset=Discussion.objects,
empty_label='All discussions')
class SubscribeForm(forms.Form):
send = forms.ModelMultipleChoiceField(
NoticeSetting.objects,
required=False,
label=u'Notify me with updates from this discussion by',
widget=forms.CheckboxSelectMultiple)
def __init__(self, *args, **kwargs):
qs = kwargs.pop('qs')
super(SubscribeForm, self).__init__(*args, **kwargs)
self.fields['send'].queryset = qs
<commit_msg>Make the odd string translatable.<commit_after>
|
from django import forms
from django.utils.translation import ugettext_lazy as _
from discussion.models import Comment, Post, Discussion
from notification.models import NoticeSetting
class CommentForm(forms.ModelForm):
class Meta:
exclude = ('user', 'post')
model = Comment
widgets = {
'body': forms.Textarea(attrs={'placeholder': _('Reply to this conversation')}),
}
class PostForm(forms.ModelForm):
class Meta:
exclude = ('user', 'discussion')
model = Post
widgets = {
'body': forms.Textarea(attrs={'placeholder': _('Start a conversation')}),
}
class SearchForm(forms.Form):
search = forms.CharField()
discussion = forms.ModelChoiceField(
required=False,
queryset=Discussion.objects,
empty_label=_('All discussions'))
class SubscribeForm(forms.Form):
send = forms.ModelMultipleChoiceField(
NoticeSetting.objects,
required=False,
label=_('Notify me with updates from this discussion by'),
widget=forms.CheckboxSelectMultiple)
def __init__(self, *args, **kwargs):
qs = kwargs.pop('qs')
super(SubscribeForm, self).__init__(*args, **kwargs)
self.fields['send'].queryset = qs
|
from django import forms
from discussion.models import Comment, Post, Discussion
from notification.models import NoticeSetting
class CommentForm(forms.ModelForm):
class Meta:
exclude = ('user', 'post')
model = Comment
widgets = {
'body': forms.Textarea(attrs={'placeholder': 'Reply to this conversation'}),
}
class PostForm(forms.ModelForm):
class Meta:
exclude = ('user', 'discussion')
model = Post
widgets = {
'body': forms.Textarea(attrs={'placeholder': 'Start a conversation'}),
}
class SearchForm(forms.Form):
search = forms.CharField()
discussion = forms.ModelChoiceField(
required=False,
queryset=Discussion.objects,
empty_label='All discussions')
class SubscribeForm(forms.Form):
send = forms.ModelMultipleChoiceField(
NoticeSetting.objects,
required=False,
label=u'Notify me with updates from this discussion by',
widget=forms.CheckboxSelectMultiple)
def __init__(self, *args, **kwargs):
qs = kwargs.pop('qs')
super(SubscribeForm, self).__init__(*args, **kwargs)
self.fields['send'].queryset = qs
Make the odd string translatable.from django import forms
from django.utils.translation import ugettext_lazy as _
from discussion.models import Comment, Post, Discussion
from notification.models import NoticeSetting
class CommentForm(forms.ModelForm):
class Meta:
exclude = ('user', 'post')
model = Comment
widgets = {
'body': forms.Textarea(attrs={'placeholder': _('Reply to this conversation')}),
}
class PostForm(forms.ModelForm):
class Meta:
exclude = ('user', 'discussion')
model = Post
widgets = {
'body': forms.Textarea(attrs={'placeholder': _('Start a conversation')}),
}
class SearchForm(forms.Form):
search = forms.CharField()
discussion = forms.ModelChoiceField(
required=False,
queryset=Discussion.objects,
empty_label=_('All discussions'))
class SubscribeForm(forms.Form):
send = forms.ModelMultipleChoiceField(
NoticeSetting.objects,
required=False,
label=_('Notify me with updates from this discussion by'),
widget=forms.CheckboxSelectMultiple)
def __init__(self, *args, **kwargs):
qs = kwargs.pop('qs')
super(SubscribeForm, self).__init__(*args, **kwargs)
self.fields['send'].queryset = qs
|
<commit_before>from django import forms
from discussion.models import Comment, Post, Discussion
from notification.models import NoticeSetting
class CommentForm(forms.ModelForm):
class Meta:
exclude = ('user', 'post')
model = Comment
widgets = {
'body': forms.Textarea(attrs={'placeholder': 'Reply to this conversation'}),
}
class PostForm(forms.ModelForm):
class Meta:
exclude = ('user', 'discussion')
model = Post
widgets = {
'body': forms.Textarea(attrs={'placeholder': 'Start a conversation'}),
}
class SearchForm(forms.Form):
search = forms.CharField()
discussion = forms.ModelChoiceField(
required=False,
queryset=Discussion.objects,
empty_label='All discussions')
class SubscribeForm(forms.Form):
send = forms.ModelMultipleChoiceField(
NoticeSetting.objects,
required=False,
label=u'Notify me with updates from this discussion by',
widget=forms.CheckboxSelectMultiple)
def __init__(self, *args, **kwargs):
qs = kwargs.pop('qs')
super(SubscribeForm, self).__init__(*args, **kwargs)
self.fields['send'].queryset = qs
<commit_msg>Make the odd string translatable.<commit_after>from django import forms
from django.utils.translation import ugettext_lazy as _
from discussion.models import Comment, Post, Discussion
from notification.models import NoticeSetting
class CommentForm(forms.ModelForm):
class Meta:
exclude = ('user', 'post')
model = Comment
widgets = {
'body': forms.Textarea(attrs={'placeholder': _('Reply to this conversation')}),
}
class PostForm(forms.ModelForm):
class Meta:
exclude = ('user', 'discussion')
model = Post
widgets = {
'body': forms.Textarea(attrs={'placeholder': _('Start a conversation')}),
}
class SearchForm(forms.Form):
search = forms.CharField()
discussion = forms.ModelChoiceField(
required=False,
queryset=Discussion.objects,
empty_label=_('All discussions'))
class SubscribeForm(forms.Form):
send = forms.ModelMultipleChoiceField(
NoticeSetting.objects,
required=False,
label=_('Notify me with updates from this discussion by'),
widget=forms.CheckboxSelectMultiple)
def __init__(self, *args, **kwargs):
qs = kwargs.pop('qs')
super(SubscribeForm, self).__init__(*args, **kwargs)
self.fields['send'].queryset = qs
|
18e310680f7dfd8f5a5186baf37cab9968f19012
|
django_base/urls.py
|
django_base/urls.py
|
from django.conf.urls import patterns, include, url
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
url(r'^admin/doc/', include('django.contrib.admindocs.urls')),
url(r'^admin/', include(admin.site.urls)),
)
|
from django.conf.urls import patterns, include, url
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
url(r'^admin/', include(admin.site.urls)),
)
|
Remove admin docs by default since they are never used.
|
Remove admin docs by default since they are never used.
|
Python
|
bsd-3-clause
|
SheepDogInc/django-base,SheepDogInc/django-base
|
from django.conf.urls import patterns, include, url
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
url(r'^admin/doc/', include('django.contrib.admindocs.urls')),
url(r'^admin/', include(admin.site.urls)),
)
Remove admin docs by default since they are never used.
|
from django.conf.urls import patterns, include, url
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
url(r'^admin/', include(admin.site.urls)),
)
|
<commit_before>from django.conf.urls import patterns, include, url
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
url(r'^admin/doc/', include('django.contrib.admindocs.urls')),
url(r'^admin/', include(admin.site.urls)),
)
<commit_msg>Remove admin docs by default since they are never used.<commit_after>
|
from django.conf.urls import patterns, include, url
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
url(r'^admin/', include(admin.site.urls)),
)
|
from django.conf.urls import patterns, include, url
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
url(r'^admin/doc/', include('django.contrib.admindocs.urls')),
url(r'^admin/', include(admin.site.urls)),
)
Remove admin docs by default since they are never used.from django.conf.urls import patterns, include, url
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
url(r'^admin/', include(admin.site.urls)),
)
|
<commit_before>from django.conf.urls import patterns, include, url
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
url(r'^admin/doc/', include('django.contrib.admindocs.urls')),
url(r'^admin/', include(admin.site.urls)),
)
<commit_msg>Remove admin docs by default since they are never used.<commit_after>from django.conf.urls import patterns, include, url
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
url(r'^admin/', include(admin.site.urls)),
)
|
d6601b9d7bdbf81d89f3d165f11845384d09797c
|
dsub/_dsub_version.py
|
dsub/_dsub_version.py
|
# Copyright 2017 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Single source of truth for dsub's version.
This must remain small and dependency-free so that any dsub module may
import it without creating circular dependencies. Note that this module
is parsed as a text file by setup.py and changes to the format of this
file could break setup.py.
The version should follow formatting requirements specified in PEP-440.
- https://www.python.org/dev/peps/pep-0440
A typical release sequence will be versioned as:
0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...
"""
DSUB_VERSION = '0.4.2.dev0'
|
# Copyright 2017 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Single source of truth for dsub's version.
This must remain small and dependency-free so that any dsub module may
import it without creating circular dependencies. Note that this module
is parsed as a text file by setup.py and changes to the format of this
file could break setup.py.
The version should follow formatting requirements specified in PEP-440.
- https://www.python.org/dev/peps/pep-0440
A typical release sequence will be versioned as:
0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...
"""
DSUB_VERSION = '0.4.2'
|
Update dsub version to 0.4.2
|
Update dsub version to 0.4.2
PiperOrigin-RevId: 337172014
|
Python
|
apache-2.0
|
DataBiosphere/dsub,DataBiosphere/dsub
|
# Copyright 2017 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Single source of truth for dsub's version.
This must remain small and dependency-free so that any dsub module may
import it without creating circular dependencies. Note that this module
is parsed as a text file by setup.py and changes to the format of this
file could break setup.py.
The version should follow formatting requirements specified in PEP-440.
- https://www.python.org/dev/peps/pep-0440
A typical release sequence will be versioned as:
0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...
"""
DSUB_VERSION = '0.4.2.dev0'
Update dsub version to 0.4.2
PiperOrigin-RevId: 337172014
|
# Copyright 2017 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Single source of truth for dsub's version.
This must remain small and dependency-free so that any dsub module may
import it without creating circular dependencies. Note that this module
is parsed as a text file by setup.py and changes to the format of this
file could break setup.py.
The version should follow formatting requirements specified in PEP-440.
- https://www.python.org/dev/peps/pep-0440
A typical release sequence will be versioned as:
0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...
"""
DSUB_VERSION = '0.4.2'
|
<commit_before># Copyright 2017 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Single source of truth for dsub's version.
This must remain small and dependency-free so that any dsub module may
import it without creating circular dependencies. Note that this module
is parsed as a text file by setup.py and changes to the format of this
file could break setup.py.
The version should follow formatting requirements specified in PEP-440.
- https://www.python.org/dev/peps/pep-0440
A typical release sequence will be versioned as:
0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...
"""
DSUB_VERSION = '0.4.2.dev0'
<commit_msg>Update dsub version to 0.4.2
PiperOrigin-RevId: 337172014<commit_after>
|
# Copyright 2017 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Single source of truth for dsub's version.
This must remain small and dependency-free so that any dsub module may
import it without creating circular dependencies. Note that this module
is parsed as a text file by setup.py and changes to the format of this
file could break setup.py.
The version should follow formatting requirements specified in PEP-440.
- https://www.python.org/dev/peps/pep-0440
A typical release sequence will be versioned as:
0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...
"""
DSUB_VERSION = '0.4.2'
|
# Copyright 2017 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Single source of truth for dsub's version.
This must remain small and dependency-free so that any dsub module may
import it without creating circular dependencies. Note that this module
is parsed as a text file by setup.py and changes to the format of this
file could break setup.py.
The version should follow formatting requirements specified in PEP-440.
- https://www.python.org/dev/peps/pep-0440
A typical release sequence will be versioned as:
0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...
"""
DSUB_VERSION = '0.4.2.dev0'
Update dsub version to 0.4.2
PiperOrigin-RevId: 337172014# Copyright 2017 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Single source of truth for dsub's version.
This must remain small and dependency-free so that any dsub module may
import it without creating circular dependencies. Note that this module
is parsed as a text file by setup.py and changes to the format of this
file could break setup.py.
The version should follow formatting requirements specified in PEP-440.
- https://www.python.org/dev/peps/pep-0440
A typical release sequence will be versioned as:
0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...
"""
DSUB_VERSION = '0.4.2'
|
<commit_before># Copyright 2017 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Single source of truth for dsub's version.
This must remain small and dependency-free so that any dsub module may
import it without creating circular dependencies. Note that this module
is parsed as a text file by setup.py and changes to the format of this
file could break setup.py.
The version should follow formatting requirements specified in PEP-440.
- https://www.python.org/dev/peps/pep-0440
A typical release sequence will be versioned as:
0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...
"""
DSUB_VERSION = '0.4.2.dev0'
<commit_msg>Update dsub version to 0.4.2
PiperOrigin-RevId: 337172014<commit_after># Copyright 2017 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Single source of truth for dsub's version.
This must remain small and dependency-free so that any dsub module may
import it without creating circular dependencies. Note that this module
is parsed as a text file by setup.py and changes to the format of this
file could break setup.py.
The version should follow formatting requirements specified in PEP-440.
- https://www.python.org/dev/peps/pep-0440
A typical release sequence will be versioned as:
0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...
"""
DSUB_VERSION = '0.4.2'
|
24045cd16a862ebd31f4a88a733a05bf2aff03a5
|
easygeoip/urls_api.py
|
easygeoip/urls_api.py
|
from django.conf.urls import patterns, url
# API URLs
from .views import LocationFromIpView
urlpatterns = patterns('',
url(r'^location/(?P<ip_address>(\d{1,3}\.\d{1,3}\.\d{1,3}\.\d{1,3}))/$', LocationFromIpView.as_view(),
name='geoip-explicit-ip-view'),
url(r'^location/$', LocationFromIpView.as_view(), name='geoip-implicit-ip-view') # Take IP addr from request
)
|
from django.conf.urls import url
# API URLs
from .views import LocationFromIpView
urlpatterns = [
url(r'^location/(?P<ip_address>(\d{1,3}\.\d{1,3}\.\d{1,3}\.\d{1,3}))/$', LocationFromIpView.as_view(),
name='geoip-explicit-ip-view'),
url(r'^location/$', LocationFromIpView.as_view(), name='geoip-implicit-ip-view') # Take IP addr from request
]
|
Upgrade to new urlpatterns format
|
Upgrade to new urlpatterns format
|
Python
|
mit
|
lambdacomplete/django-easygeoip
|
from django.conf.urls import patterns, url
# API URLs
from .views import LocationFromIpView
urlpatterns = patterns('',
url(r'^location/(?P<ip_address>(\d{1,3}\.\d{1,3}\.\d{1,3}\.\d{1,3}))/$', LocationFromIpView.as_view(),
name='geoip-explicit-ip-view'),
url(r'^location/$', LocationFromIpView.as_view(), name='geoip-implicit-ip-view') # Take IP addr from request
)
Upgrade to new urlpatterns format
|
from django.conf.urls import url
# API URLs
from .views import LocationFromIpView
urlpatterns = [
url(r'^location/(?P<ip_address>(\d{1,3}\.\d{1,3}\.\d{1,3}\.\d{1,3}))/$', LocationFromIpView.as_view(),
name='geoip-explicit-ip-view'),
url(r'^location/$', LocationFromIpView.as_view(), name='geoip-implicit-ip-view') # Take IP addr from request
]
|
<commit_before>from django.conf.urls import patterns, url
# API URLs
from .views import LocationFromIpView
urlpatterns = patterns('',
url(r'^location/(?P<ip_address>(\d{1,3}\.\d{1,3}\.\d{1,3}\.\d{1,3}))/$', LocationFromIpView.as_view(),
name='geoip-explicit-ip-view'),
url(r'^location/$', LocationFromIpView.as_view(), name='geoip-implicit-ip-view') # Take IP addr from request
)
<commit_msg>Upgrade to new urlpatterns format<commit_after>
|
from django.conf.urls import url
# API URLs
from .views import LocationFromIpView
urlpatterns = [
url(r'^location/(?P<ip_address>(\d{1,3}\.\d{1,3}\.\d{1,3}\.\d{1,3}))/$', LocationFromIpView.as_view(),
name='geoip-explicit-ip-view'),
url(r'^location/$', LocationFromIpView.as_view(), name='geoip-implicit-ip-view') # Take IP addr from request
]
|
from django.conf.urls import patterns, url
# API URLs
from .views import LocationFromIpView
urlpatterns = patterns('',
url(r'^location/(?P<ip_address>(\d{1,3}\.\d{1,3}\.\d{1,3}\.\d{1,3}))/$', LocationFromIpView.as_view(),
name='geoip-explicit-ip-view'),
url(r'^location/$', LocationFromIpView.as_view(), name='geoip-implicit-ip-view') # Take IP addr from request
)
Upgrade to new urlpatterns formatfrom django.conf.urls import url
# API URLs
from .views import LocationFromIpView
urlpatterns = [
url(r'^location/(?P<ip_address>(\d{1,3}\.\d{1,3}\.\d{1,3}\.\d{1,3}))/$', LocationFromIpView.as_view(),
name='geoip-explicit-ip-view'),
url(r'^location/$', LocationFromIpView.as_view(), name='geoip-implicit-ip-view') # Take IP addr from request
]
|
<commit_before>from django.conf.urls import patterns, url
# API URLs
from .views import LocationFromIpView
urlpatterns = patterns('',
url(r'^location/(?P<ip_address>(\d{1,3}\.\d{1,3}\.\d{1,3}\.\d{1,3}))/$', LocationFromIpView.as_view(),
name='geoip-explicit-ip-view'),
url(r'^location/$', LocationFromIpView.as_view(), name='geoip-implicit-ip-view') # Take IP addr from request
)
<commit_msg>Upgrade to new urlpatterns format<commit_after>from django.conf.urls import url
# API URLs
from .views import LocationFromIpView
urlpatterns = [
url(r'^location/(?P<ip_address>(\d{1,3}\.\d{1,3}\.\d{1,3}\.\d{1,3}))/$', LocationFromIpView.as_view(),
name='geoip-explicit-ip-view'),
url(r'^location/$', LocationFromIpView.as_view(), name='geoip-implicit-ip-view') # Take IP addr from request
]
|
cb4973909ea662abdf718e5a831806dcb0ecc821
|
14B-088/HI/HI_correct_mask_model.py
|
14B-088/HI/HI_correct_mask_model.py
|
'''
Swap the spatial axes. Swap the spectral and stokes axes.
'''
import sys
from astropy.io import fits
hdu = fits.open(sys.argv[1], mode='update')
hdu[0].data = hdu[0].data.swapaxes(0, 1)
hdu[0].data = hdu[0].data[:, :, :, ::-1]
hdu[0].data = hdu[0].data[:, :, ::-1, :]
hdu.flush()
execfile("~/Dropbox/code_development/ewky_scripts/header_swap_axis.py")
hdu[0].header = header_swapaxes(hdu[0].header, 2, 3)
hdu.flush()
hdu.close()
|
'''
\Swap the spectral and stokes axes. Needed due to issue in regridding function
'''
import sys
from astropy.io import fits
hdu = fits.open(sys.argv[1], mode='update')
hdu[0].data = hdu[0].data.swapaxes(0, 1)
execfile("/home/eric/Dropbox/code_development/ewky_scripts/header_swap_axis.py")
hdu[0].header = header_swapaxes(hdu[0].header, 2, 3)
hdu.flush()
hdu.close()
|
Update what's needed to correct mask and model
|
Update what's needed to correct mask and model
|
Python
|
mit
|
e-koch/VLA_Lband,e-koch/VLA_Lband
|
'''
Swap the spatial axes. Swap the spectral and stokes axes.
'''
import sys
from astropy.io import fits
hdu = fits.open(sys.argv[1], mode='update')
hdu[0].data = hdu[0].data.swapaxes(0, 1)
hdu[0].data = hdu[0].data[:, :, :, ::-1]
hdu[0].data = hdu[0].data[:, :, ::-1, :]
hdu.flush()
execfile("~/Dropbox/code_development/ewky_scripts/header_swap_axis.py")
hdu[0].header = header_swapaxes(hdu[0].header, 2, 3)
hdu.flush()
hdu.close()
Update what's needed to correct mask and model
|
'''
\Swap the spectral and stokes axes. Needed due to issue in regridding function
'''
import sys
from astropy.io import fits
hdu = fits.open(sys.argv[1], mode='update')
hdu[0].data = hdu[0].data.swapaxes(0, 1)
execfile("/home/eric/Dropbox/code_development/ewky_scripts/header_swap_axis.py")
hdu[0].header = header_swapaxes(hdu[0].header, 2, 3)
hdu.flush()
hdu.close()
|
<commit_before>
'''
Swap the spatial axes. Swap the spectral and stokes axes.
'''
import sys
from astropy.io import fits
hdu = fits.open(sys.argv[1], mode='update')
hdu[0].data = hdu[0].data.swapaxes(0, 1)
hdu[0].data = hdu[0].data[:, :, :, ::-1]
hdu[0].data = hdu[0].data[:, :, ::-1, :]
hdu.flush()
execfile("~/Dropbox/code_development/ewky_scripts/header_swap_axis.py")
hdu[0].header = header_swapaxes(hdu[0].header, 2, 3)
hdu.flush()
hdu.close()
<commit_msg>Update what's needed to correct mask and model<commit_after>
|
'''
\Swap the spectral and stokes axes. Needed due to issue in regridding function
'''
import sys
from astropy.io import fits
hdu = fits.open(sys.argv[1], mode='update')
hdu[0].data = hdu[0].data.swapaxes(0, 1)
execfile("/home/eric/Dropbox/code_development/ewky_scripts/header_swap_axis.py")
hdu[0].header = header_swapaxes(hdu[0].header, 2, 3)
hdu.flush()
hdu.close()
|
'''
Swap the spatial axes. Swap the spectral and stokes axes.
'''
import sys
from astropy.io import fits
hdu = fits.open(sys.argv[1], mode='update')
hdu[0].data = hdu[0].data.swapaxes(0, 1)
hdu[0].data = hdu[0].data[:, :, :, ::-1]
hdu[0].data = hdu[0].data[:, :, ::-1, :]
hdu.flush()
execfile("~/Dropbox/code_development/ewky_scripts/header_swap_axis.py")
hdu[0].header = header_swapaxes(hdu[0].header, 2, 3)
hdu.flush()
hdu.close()
Update what's needed to correct mask and model
'''
\Swap the spectral and stokes axes. Needed due to issue in regridding function
'''
import sys
from astropy.io import fits
hdu = fits.open(sys.argv[1], mode='update')
hdu[0].data = hdu[0].data.swapaxes(0, 1)
execfile("/home/eric/Dropbox/code_development/ewky_scripts/header_swap_axis.py")
hdu[0].header = header_swapaxes(hdu[0].header, 2, 3)
hdu.flush()
hdu.close()
|
<commit_before>
'''
Swap the spatial axes. Swap the spectral and stokes axes.
'''
import sys
from astropy.io import fits
hdu = fits.open(sys.argv[1], mode='update')
hdu[0].data = hdu[0].data.swapaxes(0, 1)
hdu[0].data = hdu[0].data[:, :, :, ::-1]
hdu[0].data = hdu[0].data[:, :, ::-1, :]
hdu.flush()
execfile("~/Dropbox/code_development/ewky_scripts/header_swap_axis.py")
hdu[0].header = header_swapaxes(hdu[0].header, 2, 3)
hdu.flush()
hdu.close()
<commit_msg>Update what's needed to correct mask and model<commit_after>
'''
\Swap the spectral and stokes axes. Needed due to issue in regridding function
'''
import sys
from astropy.io import fits
hdu = fits.open(sys.argv[1], mode='update')
hdu[0].data = hdu[0].data.swapaxes(0, 1)
execfile("/home/eric/Dropbox/code_development/ewky_scripts/header_swap_axis.py")
hdu[0].header = header_swapaxes(hdu[0].header, 2, 3)
hdu.flush()
hdu.close()
|
f16add1160e5a76f94be30ea54cea27045c32705
|
tests/test_blacklist.py
|
tests/test_blacklist.py
|
import unittest
import config
from .. import ntokloapi
class BlacklistTest(unittest.TestCase):
def setUp(self):
self.blacklist = ntokloapi.Blacklist(config.TEST_KEY, config.TEST_SECRET)
def test_blacklist_add_singleitem(self):
response = self.blacklist.add(productid=['10201', ])
assert response == "204"
def test_blacklist_add_multipleitems(self):
response = self.blacklist.add(productid=['10202', '10203'])
assert response == "204"
def test_blacklist_add_empty_elements(self):
response = self.blacklist.add(productid=['10204', '10205', '', ''])
assert response == "204"
def test_blacklist_remove_singleitem(self):
response = self.blacklist.remove(productid=['10201', ])
assert response == "204"
def test_blacklist_remove_multipleitems(self):
response = self.blacklist.remove(productid=['10202', '10203'])
assert response == "204"
def test_blacklist_remove_empty_elements(self):
response = self.blacklist.remove(productid=['10204', '10205', '', ''])
assert response == "204"
def test_blacklist_show_items(self):
response = self.blacklist.list()
assert not response
|
import unittest
import config
import ntokloapi
class BlacklistTest(unittest.TestCase):
def setUp(self):
self.blacklist = ntokloapi.Blacklist(config.TEST_KEY, config.TEST_SECRET)
def test_blacklist_add_singleitem(self):
response = self.blacklist.add(productid=['10201', ])
assert response == 204
def test_blacklist_add_multipleitems(self):
response = self.blacklist.add(productid=['10202', '10203'])
assert response == 204
def test_blacklist_add_empty_elements(self):
response = self.blacklist.add(productid=['10204', '10205', '', ''])
assert response == 204
def test_blacklist_remove_singleitem(self):
response = self.blacklist.remove(productid=['10201', ])
assert response == 204
def test_blacklist_remove_multipleitems(self):
response = self.blacklist.remove(productid=['10202', '10203'])
assert response == 204
def test_blacklist_remove_empty_elements(self):
response = self.blacklist.remove(productid=['10204', '10205', '', ''])
assert response == 204
def test_blacklist_show_items(self):
response = self.blacklist.list()
assert not response
|
Fix unit tests for the blacklist
|
Fix unit tests for the blacklist
|
Python
|
apache-2.0
|
nToklo/ntokloapi-python
|
import unittest
import config
from .. import ntokloapi
class BlacklistTest(unittest.TestCase):
def setUp(self):
self.blacklist = ntokloapi.Blacklist(config.TEST_KEY, config.TEST_SECRET)
def test_blacklist_add_singleitem(self):
response = self.blacklist.add(productid=['10201', ])
assert response == "204"
def test_blacklist_add_multipleitems(self):
response = self.blacklist.add(productid=['10202', '10203'])
assert response == "204"
def test_blacklist_add_empty_elements(self):
response = self.blacklist.add(productid=['10204', '10205', '', ''])
assert response == "204"
def test_blacklist_remove_singleitem(self):
response = self.blacklist.remove(productid=['10201', ])
assert response == "204"
def test_blacklist_remove_multipleitems(self):
response = self.blacklist.remove(productid=['10202', '10203'])
assert response == "204"
def test_blacklist_remove_empty_elements(self):
response = self.blacklist.remove(productid=['10204', '10205', '', ''])
assert response == "204"
def test_blacklist_show_items(self):
response = self.blacklist.list()
assert not response
Fix unit tests for the blacklist
|
import unittest
import config
import ntokloapi
class BlacklistTest(unittest.TestCase):
def setUp(self):
self.blacklist = ntokloapi.Blacklist(config.TEST_KEY, config.TEST_SECRET)
def test_blacklist_add_singleitem(self):
response = self.blacklist.add(productid=['10201', ])
assert response == 204
def test_blacklist_add_multipleitems(self):
response = self.blacklist.add(productid=['10202', '10203'])
assert response == 204
def test_blacklist_add_empty_elements(self):
response = self.blacklist.add(productid=['10204', '10205', '', ''])
assert response == 204
def test_blacklist_remove_singleitem(self):
response = self.blacklist.remove(productid=['10201', ])
assert response == 204
def test_blacklist_remove_multipleitems(self):
response = self.blacklist.remove(productid=['10202', '10203'])
assert response == 204
def test_blacklist_remove_empty_elements(self):
response = self.blacklist.remove(productid=['10204', '10205', '', ''])
assert response == 204
def test_blacklist_show_items(self):
response = self.blacklist.list()
assert not response
|
<commit_before>import unittest
import config
from .. import ntokloapi
class BlacklistTest(unittest.TestCase):
def setUp(self):
self.blacklist = ntokloapi.Blacklist(config.TEST_KEY, config.TEST_SECRET)
def test_blacklist_add_singleitem(self):
response = self.blacklist.add(productid=['10201', ])
assert response == "204"
def test_blacklist_add_multipleitems(self):
response = self.blacklist.add(productid=['10202', '10203'])
assert response == "204"
def test_blacklist_add_empty_elements(self):
response = self.blacklist.add(productid=['10204', '10205', '', ''])
assert response == "204"
def test_blacklist_remove_singleitem(self):
response = self.blacklist.remove(productid=['10201', ])
assert response == "204"
def test_blacklist_remove_multipleitems(self):
response = self.blacklist.remove(productid=['10202', '10203'])
assert response == "204"
def test_blacklist_remove_empty_elements(self):
response = self.blacklist.remove(productid=['10204', '10205', '', ''])
assert response == "204"
def test_blacklist_show_items(self):
response = self.blacklist.list()
assert not response
<commit_msg>Fix unit tests for the blacklist<commit_after>
|
import unittest
import config
import ntokloapi
class BlacklistTest(unittest.TestCase):
def setUp(self):
self.blacklist = ntokloapi.Blacklist(config.TEST_KEY, config.TEST_SECRET)
def test_blacklist_add_singleitem(self):
response = self.blacklist.add(productid=['10201', ])
assert response == 204
def test_blacklist_add_multipleitems(self):
response = self.blacklist.add(productid=['10202', '10203'])
assert response == 204
def test_blacklist_add_empty_elements(self):
response = self.blacklist.add(productid=['10204', '10205', '', ''])
assert response == 204
def test_blacklist_remove_singleitem(self):
response = self.blacklist.remove(productid=['10201', ])
assert response == 204
def test_blacklist_remove_multipleitems(self):
response = self.blacklist.remove(productid=['10202', '10203'])
assert response == 204
def test_blacklist_remove_empty_elements(self):
response = self.blacklist.remove(productid=['10204', '10205', '', ''])
assert response == 204
def test_blacklist_show_items(self):
response = self.blacklist.list()
assert not response
|
import unittest
import config
from .. import ntokloapi
class BlacklistTest(unittest.TestCase):
def setUp(self):
self.blacklist = ntokloapi.Blacklist(config.TEST_KEY, config.TEST_SECRET)
def test_blacklist_add_singleitem(self):
response = self.blacklist.add(productid=['10201', ])
assert response == "204"
def test_blacklist_add_multipleitems(self):
response = self.blacklist.add(productid=['10202', '10203'])
assert response == "204"
def test_blacklist_add_empty_elements(self):
response = self.blacklist.add(productid=['10204', '10205', '', ''])
assert response == "204"
def test_blacklist_remove_singleitem(self):
response = self.blacklist.remove(productid=['10201', ])
assert response == "204"
def test_blacklist_remove_multipleitems(self):
response = self.blacklist.remove(productid=['10202', '10203'])
assert response == "204"
def test_blacklist_remove_empty_elements(self):
response = self.blacklist.remove(productid=['10204', '10205', '', ''])
assert response == "204"
def test_blacklist_show_items(self):
response = self.blacklist.list()
assert not response
Fix unit tests for the blacklistimport unittest
import config
import ntokloapi
class BlacklistTest(unittest.TestCase):
def setUp(self):
self.blacklist = ntokloapi.Blacklist(config.TEST_KEY, config.TEST_SECRET)
def test_blacklist_add_singleitem(self):
response = self.blacklist.add(productid=['10201', ])
assert response == 204
def test_blacklist_add_multipleitems(self):
response = self.blacklist.add(productid=['10202', '10203'])
assert response == 204
def test_blacklist_add_empty_elements(self):
response = self.blacklist.add(productid=['10204', '10205', '', ''])
assert response == 204
def test_blacklist_remove_singleitem(self):
response = self.blacklist.remove(productid=['10201', ])
assert response == 204
def test_blacklist_remove_multipleitems(self):
response = self.blacklist.remove(productid=['10202', '10203'])
assert response == 204
def test_blacklist_remove_empty_elements(self):
response = self.blacklist.remove(productid=['10204', '10205', '', ''])
assert response == 204
def test_blacklist_show_items(self):
response = self.blacklist.list()
assert not response
|
<commit_before>import unittest
import config
from .. import ntokloapi
class BlacklistTest(unittest.TestCase):
def setUp(self):
self.blacklist = ntokloapi.Blacklist(config.TEST_KEY, config.TEST_SECRET)
def test_blacklist_add_singleitem(self):
response = self.blacklist.add(productid=['10201', ])
assert response == "204"
def test_blacklist_add_multipleitems(self):
response = self.blacklist.add(productid=['10202', '10203'])
assert response == "204"
def test_blacklist_add_empty_elements(self):
response = self.blacklist.add(productid=['10204', '10205', '', ''])
assert response == "204"
def test_blacklist_remove_singleitem(self):
response = self.blacklist.remove(productid=['10201', ])
assert response == "204"
def test_blacklist_remove_multipleitems(self):
response = self.blacklist.remove(productid=['10202', '10203'])
assert response == "204"
def test_blacklist_remove_empty_elements(self):
response = self.blacklist.remove(productid=['10204', '10205', '', ''])
assert response == "204"
def test_blacklist_show_items(self):
response = self.blacklist.list()
assert not response
<commit_msg>Fix unit tests for the blacklist<commit_after>import unittest
import config
import ntokloapi
class BlacklistTest(unittest.TestCase):
def setUp(self):
self.blacklist = ntokloapi.Blacklist(config.TEST_KEY, config.TEST_SECRET)
def test_blacklist_add_singleitem(self):
response = self.blacklist.add(productid=['10201', ])
assert response == 204
def test_blacklist_add_multipleitems(self):
response = self.blacklist.add(productid=['10202', '10203'])
assert response == 204
def test_blacklist_add_empty_elements(self):
response = self.blacklist.add(productid=['10204', '10205', '', ''])
assert response == 204
def test_blacklist_remove_singleitem(self):
response = self.blacklist.remove(productid=['10201', ])
assert response == 204
def test_blacklist_remove_multipleitems(self):
response = self.blacklist.remove(productid=['10202', '10203'])
assert response == 204
def test_blacklist_remove_empty_elements(self):
response = self.blacklist.remove(productid=['10204', '10205', '', ''])
assert response == 204
def test_blacklist_show_items(self):
response = self.blacklist.list()
assert not response
|
f8f0335a1a790b1ef8163a2be968b29769be80a2
|
arim/models.py
|
arim/models.py
|
from django.db import models
class Lease(models.Model):
class Meta:
db_table = 'autoreg'
mac = models.CharField(max_length=17, db_index=True)
ip = models.IntegerField(primary_key=True)
date = models.IntegerField()
|
from django.db import models
from ipaddr import IPv4Address
class Lease(models.Model):
class Meta:
db_table = 'autoreg'
mac = models.CharField(max_length=17, db_index=True)
ip = models.IntegerField(primary_key=True)
date = models.IntegerField()
def __str__(self):
return unicode(self).encode('ascii', 'replace')
def __unicode__(self):
return unicode(IPv4Address(self.ip)) + u' = ' + unicode(self.mac)
def __repr__(self):
return u'<Lease: ' + unicode(self) + u'>'
|
Add __str__, __unicode__, and __repr__
|
Add __str__, __unicode__, and __repr__
|
Python
|
bsd-3-clause
|
drkitty/arim,OSU-Net/arim,OSU-Net/arim,drkitty/arim,drkitty/arim,OSU-Net/arim
|
from django.db import models
class Lease(models.Model):
class Meta:
db_table = 'autoreg'
mac = models.CharField(max_length=17, db_index=True)
ip = models.IntegerField(primary_key=True)
date = models.IntegerField()
Add __str__, __unicode__, and __repr__
|
from django.db import models
from ipaddr import IPv4Address
class Lease(models.Model):
class Meta:
db_table = 'autoreg'
mac = models.CharField(max_length=17, db_index=True)
ip = models.IntegerField(primary_key=True)
date = models.IntegerField()
def __str__(self):
return unicode(self).encode('ascii', 'replace')
def __unicode__(self):
return unicode(IPv4Address(self.ip)) + u' = ' + unicode(self.mac)
def __repr__(self):
return u'<Lease: ' + unicode(self) + u'>'
|
<commit_before>from django.db import models
class Lease(models.Model):
class Meta:
db_table = 'autoreg'
mac = models.CharField(max_length=17, db_index=True)
ip = models.IntegerField(primary_key=True)
date = models.IntegerField()
<commit_msg>Add __str__, __unicode__, and __repr__<commit_after>
|
from django.db import models
from ipaddr import IPv4Address
class Lease(models.Model):
class Meta:
db_table = 'autoreg'
mac = models.CharField(max_length=17, db_index=True)
ip = models.IntegerField(primary_key=True)
date = models.IntegerField()
def __str__(self):
return unicode(self).encode('ascii', 'replace')
def __unicode__(self):
return unicode(IPv4Address(self.ip)) + u' = ' + unicode(self.mac)
def __repr__(self):
return u'<Lease: ' + unicode(self) + u'>'
|
from django.db import models
class Lease(models.Model):
class Meta:
db_table = 'autoreg'
mac = models.CharField(max_length=17, db_index=True)
ip = models.IntegerField(primary_key=True)
date = models.IntegerField()
Add __str__, __unicode__, and __repr__from django.db import models
from ipaddr import IPv4Address
class Lease(models.Model):
class Meta:
db_table = 'autoreg'
mac = models.CharField(max_length=17, db_index=True)
ip = models.IntegerField(primary_key=True)
date = models.IntegerField()
def __str__(self):
return unicode(self).encode('ascii', 'replace')
def __unicode__(self):
return unicode(IPv4Address(self.ip)) + u' = ' + unicode(self.mac)
def __repr__(self):
return u'<Lease: ' + unicode(self) + u'>'
|
<commit_before>from django.db import models
class Lease(models.Model):
class Meta:
db_table = 'autoreg'
mac = models.CharField(max_length=17, db_index=True)
ip = models.IntegerField(primary_key=True)
date = models.IntegerField()
<commit_msg>Add __str__, __unicode__, and __repr__<commit_after>from django.db import models
from ipaddr import IPv4Address
class Lease(models.Model):
class Meta:
db_table = 'autoreg'
mac = models.CharField(max_length=17, db_index=True)
ip = models.IntegerField(primary_key=True)
date = models.IntegerField()
def __str__(self):
return unicode(self).encode('ascii', 'replace')
def __unicode__(self):
return unicode(IPv4Address(self.ip)) + u' = ' + unicode(self.mac)
def __repr__(self):
return u'<Lease: ' + unicode(self) + u'>'
|
f21ae3ffb99c5b90cb329317b2c6282e4992f6cc
|
safety/utils.py
|
safety/utils.py
|
# -*- coding: utf-8 -*-
import importlib
import re
import warnings
from django.conf import settings
from django.utils.translation import ugettext_lazy as _, ugettext
BROWSERS = (
(re.compile('Chrome'), _('Chrome')),
(re.compile('Safari'), _('Safari')),
(re.compile('Firefox'), _('Firefox')),
(re.compile('Opera'), _('Opera')),
(re.compile('IE'), _('Internet Explorer')),
)
DEVICES = (
(re.compile('Android'), _('Android')),
(re.compile('Linux'), _('Linux')),
(re.compile('iPhone'), _('iPhone')),
(re.compile('iPad'), _('iPad')),
(re.compile('(Mac OS X)'), _('OS X')),
(re.compile('NT 5.1'), _('Windows XP')),
(re.compile('NT 6.0'), _('Windows Vista')),
(re.compile('NT 6.1'), _('Windows 7')),
(re.compile('NT 6.2'), _('Windows 8')),
(re.compile('NT 6.3'), _('Windows 8.1')),
(re.compile('Windows'), _('Windows')),
)
def get_device(user_agent):
"""
Transform a User Agent into a human readable text.
"""
infos = []
for regex, name in BROWSERS:
if regex.search(user_agent):
infos.append('%s' % name)
break
for regex, name in DEVICES:
if regex.search(user_agent):
infos.append('%s' % name)
break
return ', '.join(infos)
def get_session_store():
mod = getattr(settings, 'SESSION_ENGINE', 'django.contrib.sessions.backends.db')
engine = importlib.import_module(mod)
store = engine.SessionStore()
return store
|
# -*- coding: utf-8 -*-
try:
from django.utils.importlib import import_module
except ImportError:
from importlib import import_module
from django.conf import settings
from django.core.exceptions import ImproperlyConfigured
def get_session_store():
mod = getattr(settings, 'SESSION_ENGINE', 'django.contrib.sessions.backends.db')
engine = import_module(mod)
store = engine.SessionStore()
return store
def get_resolver(request, setting):
module_path = getattr(app_settings, setting)
try:
module, attribute = module_path.rsplit('.', 1)
resolver_module = import_module(module)
resolver = getattr(resolver_module, attribute)
except ImportError:
raise ImproperlyConfigured(
"Please specify a valid %s module. "
"Could not find %s " % (setting, module))
except AttributeError:
raise ImproperlyConfigured(
"Please specify a valid %s "
"function. Could not find %s function in module %s" %
(setting, attribute, module))
return resolver(request)
|
Add get_resolver() util and remove get_device() (now use ua-parser).
|
Add get_resolver() util and remove get_device() (now use ua-parser).
|
Python
|
mit
|
ulule/django-safety,ulule/django-safety
|
# -*- coding: utf-8 -*-
import importlib
import re
import warnings
from django.conf import settings
from django.utils.translation import ugettext_lazy as _, ugettext
BROWSERS = (
(re.compile('Chrome'), _('Chrome')),
(re.compile('Safari'), _('Safari')),
(re.compile('Firefox'), _('Firefox')),
(re.compile('Opera'), _('Opera')),
(re.compile('IE'), _('Internet Explorer')),
)
DEVICES = (
(re.compile('Android'), _('Android')),
(re.compile('Linux'), _('Linux')),
(re.compile('iPhone'), _('iPhone')),
(re.compile('iPad'), _('iPad')),
(re.compile('(Mac OS X)'), _('OS X')),
(re.compile('NT 5.1'), _('Windows XP')),
(re.compile('NT 6.0'), _('Windows Vista')),
(re.compile('NT 6.1'), _('Windows 7')),
(re.compile('NT 6.2'), _('Windows 8')),
(re.compile('NT 6.3'), _('Windows 8.1')),
(re.compile('Windows'), _('Windows')),
)
def get_device(user_agent):
"""
Transform a User Agent into a human readable text.
"""
infos = []
for regex, name in BROWSERS:
if regex.search(user_agent):
infos.append('%s' % name)
break
for regex, name in DEVICES:
if regex.search(user_agent):
infos.append('%s' % name)
break
return ', '.join(infos)
def get_session_store():
mod = getattr(settings, 'SESSION_ENGINE', 'django.contrib.sessions.backends.db')
engine = importlib.import_module(mod)
store = engine.SessionStore()
return store
Add get_resolver() util and remove get_device() (now use ua-parser).
|
# -*- coding: utf-8 -*-
try:
from django.utils.importlib import import_module
except ImportError:
from importlib import import_module
from django.conf import settings
from django.core.exceptions import ImproperlyConfigured
def get_session_store():
mod = getattr(settings, 'SESSION_ENGINE', 'django.contrib.sessions.backends.db')
engine = import_module(mod)
store = engine.SessionStore()
return store
def get_resolver(request, setting):
module_path = getattr(app_settings, setting)
try:
module, attribute = module_path.rsplit('.', 1)
resolver_module = import_module(module)
resolver = getattr(resolver_module, attribute)
except ImportError:
raise ImproperlyConfigured(
"Please specify a valid %s module. "
"Could not find %s " % (setting, module))
except AttributeError:
raise ImproperlyConfigured(
"Please specify a valid %s "
"function. Could not find %s function in module %s" %
(setting, attribute, module))
return resolver(request)
|
<commit_before># -*- coding: utf-8 -*-
import importlib
import re
import warnings
from django.conf import settings
from django.utils.translation import ugettext_lazy as _, ugettext
BROWSERS = (
(re.compile('Chrome'), _('Chrome')),
(re.compile('Safari'), _('Safari')),
(re.compile('Firefox'), _('Firefox')),
(re.compile('Opera'), _('Opera')),
(re.compile('IE'), _('Internet Explorer')),
)
DEVICES = (
(re.compile('Android'), _('Android')),
(re.compile('Linux'), _('Linux')),
(re.compile('iPhone'), _('iPhone')),
(re.compile('iPad'), _('iPad')),
(re.compile('(Mac OS X)'), _('OS X')),
(re.compile('NT 5.1'), _('Windows XP')),
(re.compile('NT 6.0'), _('Windows Vista')),
(re.compile('NT 6.1'), _('Windows 7')),
(re.compile('NT 6.2'), _('Windows 8')),
(re.compile('NT 6.3'), _('Windows 8.1')),
(re.compile('Windows'), _('Windows')),
)
def get_device(user_agent):
"""
Transform a User Agent into a human readable text.
"""
infos = []
for regex, name in BROWSERS:
if regex.search(user_agent):
infos.append('%s' % name)
break
for regex, name in DEVICES:
if regex.search(user_agent):
infos.append('%s' % name)
break
return ', '.join(infos)
def get_session_store():
mod = getattr(settings, 'SESSION_ENGINE', 'django.contrib.sessions.backends.db')
engine = importlib.import_module(mod)
store = engine.SessionStore()
return store
<commit_msg>Add get_resolver() util and remove get_device() (now use ua-parser).<commit_after>
|
# -*- coding: utf-8 -*-
try:
from django.utils.importlib import import_module
except ImportError:
from importlib import import_module
from django.conf import settings
from django.core.exceptions import ImproperlyConfigured
def get_session_store():
mod = getattr(settings, 'SESSION_ENGINE', 'django.contrib.sessions.backends.db')
engine = import_module(mod)
store = engine.SessionStore()
return store
def get_resolver(request, setting):
module_path = getattr(app_settings, setting)
try:
module, attribute = module_path.rsplit('.', 1)
resolver_module = import_module(module)
resolver = getattr(resolver_module, attribute)
except ImportError:
raise ImproperlyConfigured(
"Please specify a valid %s module. "
"Could not find %s " % (setting, module))
except AttributeError:
raise ImproperlyConfigured(
"Please specify a valid %s "
"function. Could not find %s function in module %s" %
(setting, attribute, module))
return resolver(request)
|
# -*- coding: utf-8 -*-
import importlib
import re
import warnings
from django.conf import settings
from django.utils.translation import ugettext_lazy as _, ugettext
BROWSERS = (
(re.compile('Chrome'), _('Chrome')),
(re.compile('Safari'), _('Safari')),
(re.compile('Firefox'), _('Firefox')),
(re.compile('Opera'), _('Opera')),
(re.compile('IE'), _('Internet Explorer')),
)
DEVICES = (
(re.compile('Android'), _('Android')),
(re.compile('Linux'), _('Linux')),
(re.compile('iPhone'), _('iPhone')),
(re.compile('iPad'), _('iPad')),
(re.compile('(Mac OS X)'), _('OS X')),
(re.compile('NT 5.1'), _('Windows XP')),
(re.compile('NT 6.0'), _('Windows Vista')),
(re.compile('NT 6.1'), _('Windows 7')),
(re.compile('NT 6.2'), _('Windows 8')),
(re.compile('NT 6.3'), _('Windows 8.1')),
(re.compile('Windows'), _('Windows')),
)
def get_device(user_agent):
"""
Transform a User Agent into a human readable text.
"""
infos = []
for regex, name in BROWSERS:
if regex.search(user_agent):
infos.append('%s' % name)
break
for regex, name in DEVICES:
if regex.search(user_agent):
infos.append('%s' % name)
break
return ', '.join(infos)
def get_session_store():
mod = getattr(settings, 'SESSION_ENGINE', 'django.contrib.sessions.backends.db')
engine = importlib.import_module(mod)
store = engine.SessionStore()
return store
Add get_resolver() util and remove get_device() (now use ua-parser).# -*- coding: utf-8 -*-
try:
from django.utils.importlib import import_module
except ImportError:
from importlib import import_module
from django.conf import settings
from django.core.exceptions import ImproperlyConfigured
def get_session_store():
mod = getattr(settings, 'SESSION_ENGINE', 'django.contrib.sessions.backends.db')
engine = import_module(mod)
store = engine.SessionStore()
return store
def get_resolver(request, setting):
module_path = getattr(app_settings, setting)
try:
module, attribute = module_path.rsplit('.', 1)
resolver_module = import_module(module)
resolver = getattr(resolver_module, attribute)
except ImportError:
raise ImproperlyConfigured(
"Please specify a valid %s module. "
"Could not find %s " % (setting, module))
except AttributeError:
raise ImproperlyConfigured(
"Please specify a valid %s "
"function. Could not find %s function in module %s" %
(setting, attribute, module))
return resolver(request)
|
<commit_before># -*- coding: utf-8 -*-
import importlib
import re
import warnings
from django.conf import settings
from django.utils.translation import ugettext_lazy as _, ugettext
BROWSERS = (
(re.compile('Chrome'), _('Chrome')),
(re.compile('Safari'), _('Safari')),
(re.compile('Firefox'), _('Firefox')),
(re.compile('Opera'), _('Opera')),
(re.compile('IE'), _('Internet Explorer')),
)
DEVICES = (
(re.compile('Android'), _('Android')),
(re.compile('Linux'), _('Linux')),
(re.compile('iPhone'), _('iPhone')),
(re.compile('iPad'), _('iPad')),
(re.compile('(Mac OS X)'), _('OS X')),
(re.compile('NT 5.1'), _('Windows XP')),
(re.compile('NT 6.0'), _('Windows Vista')),
(re.compile('NT 6.1'), _('Windows 7')),
(re.compile('NT 6.2'), _('Windows 8')),
(re.compile('NT 6.3'), _('Windows 8.1')),
(re.compile('Windows'), _('Windows')),
)
def get_device(user_agent):
"""
Transform a User Agent into a human readable text.
"""
infos = []
for regex, name in BROWSERS:
if regex.search(user_agent):
infos.append('%s' % name)
break
for regex, name in DEVICES:
if regex.search(user_agent):
infos.append('%s' % name)
break
return ', '.join(infos)
def get_session_store():
mod = getattr(settings, 'SESSION_ENGINE', 'django.contrib.sessions.backends.db')
engine = importlib.import_module(mod)
store = engine.SessionStore()
return store
<commit_msg>Add get_resolver() util and remove get_device() (now use ua-parser).<commit_after># -*- coding: utf-8 -*-
try:
from django.utils.importlib import import_module
except ImportError:
from importlib import import_module
from django.conf import settings
from django.core.exceptions import ImproperlyConfigured
def get_session_store():
mod = getattr(settings, 'SESSION_ENGINE', 'django.contrib.sessions.backends.db')
engine = import_module(mod)
store = engine.SessionStore()
return store
def get_resolver(request, setting):
module_path = getattr(app_settings, setting)
try:
module, attribute = module_path.rsplit('.', 1)
resolver_module = import_module(module)
resolver = getattr(resolver_module, attribute)
except ImportError:
raise ImproperlyConfigured(
"Please specify a valid %s module. "
"Could not find %s " % (setting, module))
except AttributeError:
raise ImproperlyConfigured(
"Please specify a valid %s "
"function. Could not find %s function in module %s" %
(setting, attribute, module))
return resolver(request)
|
eda91552ae26188afbad74115495e44e07827c4d
|
typ/version.py
|
typ/version.py
|
# Copyright 2014 Google Inc. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
VERSION = '0.9.4'
|
# Copyright 2014 Google Inc. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
VERSION = '0.9.4pre'
|
Add a -vvv mode to log when tests are queued for running.
|
Add a -vvv mode to log when tests are queued for running.
If one is running a bunch of tests in parallel and something
is not working right, it can be useful to see which tests are
currently executing at the same time. There isn't a great way
to do this in typ, because we don't know when tests are actually
picked up for execution by the child processes, but we can at
least log which tests has been queued for running.
This patch adds a -vvv option (verbose=3). This will make
the test log twice as long, since we will now log lines for
when every test starts and stops, but at least that can help
with debugging.
|
Python
|
apache-2.0
|
dpranke/typ
|
# Copyright 2014 Google Inc. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
VERSION = '0.9.4'
Add a -vvv mode to log when tests are queued for running.
If one is running a bunch of tests in parallel and something
is not working right, it can be useful to see which tests are
currently executing at the same time. There isn't a great way
to do this in typ, because we don't know when tests are actually
picked up for execution by the child processes, but we can at
least log which tests has been queued for running.
This patch adds a -vvv option (verbose=3). This will make
the test log twice as long, since we will now log lines for
when every test starts and stops, but at least that can help
with debugging.
|
# Copyright 2014 Google Inc. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
VERSION = '0.9.4pre'
|
<commit_before># Copyright 2014 Google Inc. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
VERSION = '0.9.4'
<commit_msg>Add a -vvv mode to log when tests are queued for running.
If one is running a bunch of tests in parallel and something
is not working right, it can be useful to see which tests are
currently executing at the same time. There isn't a great way
to do this in typ, because we don't know when tests are actually
picked up for execution by the child processes, but we can at
least log which tests has been queued for running.
This patch adds a -vvv option (verbose=3). This will make
the test log twice as long, since we will now log lines for
when every test starts and stops, but at least that can help
with debugging.<commit_after>
|
# Copyright 2014 Google Inc. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
VERSION = '0.9.4pre'
|
# Copyright 2014 Google Inc. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
VERSION = '0.9.4'
Add a -vvv mode to log when tests are queued for running.
If one is running a bunch of tests in parallel and something
is not working right, it can be useful to see which tests are
currently executing at the same time. There isn't a great way
to do this in typ, because we don't know when tests are actually
picked up for execution by the child processes, but we can at
least log which tests has been queued for running.
This patch adds a -vvv option (verbose=3). This will make
the test log twice as long, since we will now log lines for
when every test starts and stops, but at least that can help
with debugging.# Copyright 2014 Google Inc. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
VERSION = '0.9.4pre'
|
<commit_before># Copyright 2014 Google Inc. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
VERSION = '0.9.4'
<commit_msg>Add a -vvv mode to log when tests are queued for running.
If one is running a bunch of tests in parallel and something
is not working right, it can be useful to see which tests are
currently executing at the same time. There isn't a great way
to do this in typ, because we don't know when tests are actually
picked up for execution by the child processes, but we can at
least log which tests has been queued for running.
This patch adds a -vvv option (verbose=3). This will make
the test log twice as long, since we will now log lines for
when every test starts and stops, but at least that can help
with debugging.<commit_after># Copyright 2014 Google Inc. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
VERSION = '0.9.4pre'
|
8b374d041d97307962cdf562c52b2a72345a4efc
|
snowman/urls.py
|
snowman/urls.py
|
"""snowman URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.10/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Import the include() function: from django.conf.urls import url, include
2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls'))
"""
from django.conf.urls import url, include
from django.contrib import admin
from django.views.generic import RedirectView
urlpatterns = [
url(r'^admin/', admin.site.urls),
url(r'^', include('api.router')),
url(r'^.*$', RedirectView.as_view(pattern_name='api-root', permanent=True), name='index')
]
|
"""snowman URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.10/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Import the include() function: from django.conf.urls import url, include
2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls'))
"""
from django.conf.urls import url, include
from django.contrib import admin
from django.views.generic import RedirectView
urlpatterns = [
url(r'^admin/', admin.site.urls),
url(r'^', include('api.router')),
url(r'^api-auth/', include('rest_framework.urls', namespace='rest_framework')),
url(r'^.*$', RedirectView.as_view(pattern_name='api-root', permanent=True), name='index')
]
|
Add simple login form in the API.
|
Add simple login form in the API.
This is usefull for developers to explore the Browlable api directly
on the browser.
|
Python
|
mit
|
johnnywell/snowman
|
"""snowman URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.10/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Import the include() function: from django.conf.urls import url, include
2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls'))
"""
from django.conf.urls import url, include
from django.contrib import admin
from django.views.generic import RedirectView
urlpatterns = [
url(r'^admin/', admin.site.urls),
url(r'^', include('api.router')),
url(r'^.*$', RedirectView.as_view(pattern_name='api-root', permanent=True), name='index')
]
Add simple login form in the API.
This is usefull for developers to explore the Browlable api directly
on the browser.
|
"""snowman URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.10/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Import the include() function: from django.conf.urls import url, include
2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls'))
"""
from django.conf.urls import url, include
from django.contrib import admin
from django.views.generic import RedirectView
urlpatterns = [
url(r'^admin/', admin.site.urls),
url(r'^', include('api.router')),
url(r'^api-auth/', include('rest_framework.urls', namespace='rest_framework')),
url(r'^.*$', RedirectView.as_view(pattern_name='api-root', permanent=True), name='index')
]
|
<commit_before>"""snowman URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.10/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Import the include() function: from django.conf.urls import url, include
2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls'))
"""
from django.conf.urls import url, include
from django.contrib import admin
from django.views.generic import RedirectView
urlpatterns = [
url(r'^admin/', admin.site.urls),
url(r'^', include('api.router')),
url(r'^.*$', RedirectView.as_view(pattern_name='api-root', permanent=True), name='index')
]
<commit_msg>Add simple login form in the API.
This is usefull for developers to explore the Browlable api directly
on the browser.<commit_after>
|
"""snowman URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.10/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Import the include() function: from django.conf.urls import url, include
2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls'))
"""
from django.conf.urls import url, include
from django.contrib import admin
from django.views.generic import RedirectView
urlpatterns = [
url(r'^admin/', admin.site.urls),
url(r'^', include('api.router')),
url(r'^api-auth/', include('rest_framework.urls', namespace='rest_framework')),
url(r'^.*$', RedirectView.as_view(pattern_name='api-root', permanent=True), name='index')
]
|
"""snowman URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.10/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Import the include() function: from django.conf.urls import url, include
2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls'))
"""
from django.conf.urls import url, include
from django.contrib import admin
from django.views.generic import RedirectView
urlpatterns = [
url(r'^admin/', admin.site.urls),
url(r'^', include('api.router')),
url(r'^.*$', RedirectView.as_view(pattern_name='api-root', permanent=True), name='index')
]
Add simple login form in the API.
This is usefull for developers to explore the Browlable api directly
on the browser."""snowman URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.10/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Import the include() function: from django.conf.urls import url, include
2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls'))
"""
from django.conf.urls import url, include
from django.contrib import admin
from django.views.generic import RedirectView
urlpatterns = [
url(r'^admin/', admin.site.urls),
url(r'^', include('api.router')),
url(r'^api-auth/', include('rest_framework.urls', namespace='rest_framework')),
url(r'^.*$', RedirectView.as_view(pattern_name='api-root', permanent=True), name='index')
]
|
<commit_before>"""snowman URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.10/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Import the include() function: from django.conf.urls import url, include
2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls'))
"""
from django.conf.urls import url, include
from django.contrib import admin
from django.views.generic import RedirectView
urlpatterns = [
url(r'^admin/', admin.site.urls),
url(r'^', include('api.router')),
url(r'^.*$', RedirectView.as_view(pattern_name='api-root', permanent=True), name='index')
]
<commit_msg>Add simple login form in the API.
This is usefull for developers to explore the Browlable api directly
on the browser.<commit_after>"""snowman URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.10/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Import the include() function: from django.conf.urls import url, include
2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls'))
"""
from django.conf.urls import url, include
from django.contrib import admin
from django.views.generic import RedirectView
urlpatterns = [
url(r'^admin/', admin.site.urls),
url(r'^', include('api.router')),
url(r'^api-auth/', include('rest_framework.urls', namespace='rest_framework')),
url(r'^.*$', RedirectView.as_view(pattern_name='api-root', permanent=True), name='index')
]
|
0d5946f0c61bcb629d8a1bbacf09bcc5719986fc
|
textRenderer.py
|
textRenderer.py
|
import colorsys
from PIL import Image, ImageFont, ImageDraw
class TextRenderer:
def __init__(self, font="./NotoSansCJK-Bold.otf",
font_color=(0, 120, 0), color_bg=False):
self.image = None
# params
self.color_bg = color_bg
self.font_color = font_color
# new image and font
self.font = ImageFont.truetype(font, 30)
return None
def getFrameCount(self):
return 1
def rainbow_bg(c):
# hue, lightness, saturation to rgb
vals = colorsys.hls_to_rgb(round(c / 360.0, 2), 0.05, 1)
return (int(vals[0] * 255), int(vals[1] * 255), int(vals[2] * 255))
def draw_text(self, text_to_send):
x, y = self.font.getsize(text_to_send)
self.im = Image.new("RGBA", (x, y+10), "black")
# Add padding below, because PIL sucks!
self.draw = ImageDraw.Draw(self.im)
self.draw.text(
(0, 0), text_to_send, font=self.font, fill=self.font_color
)
def render(self, msgText):
self.draw_text(' '.join(msgText))
def getImage(self):
return self.image
def get_queue_token(self, msgToken):
queue_token = {}
# TODO: add possible params
self.render(msgToken["text"])
queue_token["image"] = [self.im]
queue_token["frame_count"] = self.getFrameCount()
queue_token["action"] = "scroll"
queue_token["valid"] = True
return queue_token
|
from PIL import Image, ImageFont, ImageDraw
class TextRenderer:
def __init__(self, font="./NotoSansCJK-Bold.otf",
font_color=(0, 120, 0), color_bg=False):
# params
self.color_bg = color_bg
self.font_color = font_color
# new image and font
self.font = ImageFont.truetype(font, 30)
def draw_text(self, text_to_send):
x, y = self.font.getsize(text_to_send)
# Add padding below, because PIL sucks!
image = Image.new("RGBA", (x, y+10), "black")
ImageDraw.Draw(image).text(
(0, 0), text_to_send, font=self.font, fill=self.font_color
)
return image
def get_queue_token(self, msgToken):
queue_token = {}
# TODO: add possible params
queue_token["image"] = [self.draw_text(' '.join(msgToken["text"]))]
queue_token["frame_count"] = 1
queue_token["action"] = "scroll"
queue_token["valid"] = True
return queue_token
|
Remove all unused code in text renderer.
|
Remove all unused code in text renderer.
|
Python
|
mit
|
marqsm/LED-bot,marqsm/LED-bot,marqsm/LED-bot,marqsm/LED-bot
|
import colorsys
from PIL import Image, ImageFont, ImageDraw
class TextRenderer:
def __init__(self, font="./NotoSansCJK-Bold.otf",
font_color=(0, 120, 0), color_bg=False):
self.image = None
# params
self.color_bg = color_bg
self.font_color = font_color
# new image and font
self.font = ImageFont.truetype(font, 30)
return None
def getFrameCount(self):
return 1
def rainbow_bg(c):
# hue, lightness, saturation to rgb
vals = colorsys.hls_to_rgb(round(c / 360.0, 2), 0.05, 1)
return (int(vals[0] * 255), int(vals[1] * 255), int(vals[2] * 255))
def draw_text(self, text_to_send):
x, y = self.font.getsize(text_to_send)
self.im = Image.new("RGBA", (x, y+10), "black")
# Add padding below, because PIL sucks!
self.draw = ImageDraw.Draw(self.im)
self.draw.text(
(0, 0), text_to_send, font=self.font, fill=self.font_color
)
def render(self, msgText):
self.draw_text(' '.join(msgText))
def getImage(self):
return self.image
def get_queue_token(self, msgToken):
queue_token = {}
# TODO: add possible params
self.render(msgToken["text"])
queue_token["image"] = [self.im]
queue_token["frame_count"] = self.getFrameCount()
queue_token["action"] = "scroll"
queue_token["valid"] = True
return queue_token
Remove all unused code in text renderer.
|
from PIL import Image, ImageFont, ImageDraw
class TextRenderer:
def __init__(self, font="./NotoSansCJK-Bold.otf",
font_color=(0, 120, 0), color_bg=False):
# params
self.color_bg = color_bg
self.font_color = font_color
# new image and font
self.font = ImageFont.truetype(font, 30)
def draw_text(self, text_to_send):
x, y = self.font.getsize(text_to_send)
# Add padding below, because PIL sucks!
image = Image.new("RGBA", (x, y+10), "black")
ImageDraw.Draw(image).text(
(0, 0), text_to_send, font=self.font, fill=self.font_color
)
return image
def get_queue_token(self, msgToken):
queue_token = {}
# TODO: add possible params
queue_token["image"] = [self.draw_text(' '.join(msgToken["text"]))]
queue_token["frame_count"] = 1
queue_token["action"] = "scroll"
queue_token["valid"] = True
return queue_token
|
<commit_before>import colorsys
from PIL import Image, ImageFont, ImageDraw
class TextRenderer:
def __init__(self, font="./NotoSansCJK-Bold.otf",
font_color=(0, 120, 0), color_bg=False):
self.image = None
# params
self.color_bg = color_bg
self.font_color = font_color
# new image and font
self.font = ImageFont.truetype(font, 30)
return None
def getFrameCount(self):
return 1
def rainbow_bg(c):
# hue, lightness, saturation to rgb
vals = colorsys.hls_to_rgb(round(c / 360.0, 2), 0.05, 1)
return (int(vals[0] * 255), int(vals[1] * 255), int(vals[2] * 255))
def draw_text(self, text_to_send):
x, y = self.font.getsize(text_to_send)
self.im = Image.new("RGBA", (x, y+10), "black")
# Add padding below, because PIL sucks!
self.draw = ImageDraw.Draw(self.im)
self.draw.text(
(0, 0), text_to_send, font=self.font, fill=self.font_color
)
def render(self, msgText):
self.draw_text(' '.join(msgText))
def getImage(self):
return self.image
def get_queue_token(self, msgToken):
queue_token = {}
# TODO: add possible params
self.render(msgToken["text"])
queue_token["image"] = [self.im]
queue_token["frame_count"] = self.getFrameCount()
queue_token["action"] = "scroll"
queue_token["valid"] = True
return queue_token
<commit_msg>Remove all unused code in text renderer.<commit_after>
|
from PIL import Image, ImageFont, ImageDraw
class TextRenderer:
def __init__(self, font="./NotoSansCJK-Bold.otf",
font_color=(0, 120, 0), color_bg=False):
# params
self.color_bg = color_bg
self.font_color = font_color
# new image and font
self.font = ImageFont.truetype(font, 30)
def draw_text(self, text_to_send):
x, y = self.font.getsize(text_to_send)
# Add padding below, because PIL sucks!
image = Image.new("RGBA", (x, y+10), "black")
ImageDraw.Draw(image).text(
(0, 0), text_to_send, font=self.font, fill=self.font_color
)
return image
def get_queue_token(self, msgToken):
queue_token = {}
# TODO: add possible params
queue_token["image"] = [self.draw_text(' '.join(msgToken["text"]))]
queue_token["frame_count"] = 1
queue_token["action"] = "scroll"
queue_token["valid"] = True
return queue_token
|
import colorsys
from PIL import Image, ImageFont, ImageDraw
class TextRenderer:
def __init__(self, font="./NotoSansCJK-Bold.otf",
font_color=(0, 120, 0), color_bg=False):
self.image = None
# params
self.color_bg = color_bg
self.font_color = font_color
# new image and font
self.font = ImageFont.truetype(font, 30)
return None
def getFrameCount(self):
return 1
def rainbow_bg(c):
# hue, lightness, saturation to rgb
vals = colorsys.hls_to_rgb(round(c / 360.0, 2), 0.05, 1)
return (int(vals[0] * 255), int(vals[1] * 255), int(vals[2] * 255))
def draw_text(self, text_to_send):
x, y = self.font.getsize(text_to_send)
self.im = Image.new("RGBA", (x, y+10), "black")
# Add padding below, because PIL sucks!
self.draw = ImageDraw.Draw(self.im)
self.draw.text(
(0, 0), text_to_send, font=self.font, fill=self.font_color
)
def render(self, msgText):
self.draw_text(' '.join(msgText))
def getImage(self):
return self.image
def get_queue_token(self, msgToken):
queue_token = {}
# TODO: add possible params
self.render(msgToken["text"])
queue_token["image"] = [self.im]
queue_token["frame_count"] = self.getFrameCount()
queue_token["action"] = "scroll"
queue_token["valid"] = True
return queue_token
Remove all unused code in text renderer.from PIL import Image, ImageFont, ImageDraw
class TextRenderer:
def __init__(self, font="./NotoSansCJK-Bold.otf",
font_color=(0, 120, 0), color_bg=False):
# params
self.color_bg = color_bg
self.font_color = font_color
# new image and font
self.font = ImageFont.truetype(font, 30)
def draw_text(self, text_to_send):
x, y = self.font.getsize(text_to_send)
# Add padding below, because PIL sucks!
image = Image.new("RGBA", (x, y+10), "black")
ImageDraw.Draw(image).text(
(0, 0), text_to_send, font=self.font, fill=self.font_color
)
return image
def get_queue_token(self, msgToken):
queue_token = {}
# TODO: add possible params
queue_token["image"] = [self.draw_text(' '.join(msgToken["text"]))]
queue_token["frame_count"] = 1
queue_token["action"] = "scroll"
queue_token["valid"] = True
return queue_token
|
<commit_before>import colorsys
from PIL import Image, ImageFont, ImageDraw
class TextRenderer:
def __init__(self, font="./NotoSansCJK-Bold.otf",
font_color=(0, 120, 0), color_bg=False):
self.image = None
# params
self.color_bg = color_bg
self.font_color = font_color
# new image and font
self.font = ImageFont.truetype(font, 30)
return None
def getFrameCount(self):
return 1
def rainbow_bg(c):
# hue, lightness, saturation to rgb
vals = colorsys.hls_to_rgb(round(c / 360.0, 2), 0.05, 1)
return (int(vals[0] * 255), int(vals[1] * 255), int(vals[2] * 255))
def draw_text(self, text_to_send):
x, y = self.font.getsize(text_to_send)
self.im = Image.new("RGBA", (x, y+10), "black")
# Add padding below, because PIL sucks!
self.draw = ImageDraw.Draw(self.im)
self.draw.text(
(0, 0), text_to_send, font=self.font, fill=self.font_color
)
def render(self, msgText):
self.draw_text(' '.join(msgText))
def getImage(self):
return self.image
def get_queue_token(self, msgToken):
queue_token = {}
# TODO: add possible params
self.render(msgToken["text"])
queue_token["image"] = [self.im]
queue_token["frame_count"] = self.getFrameCount()
queue_token["action"] = "scroll"
queue_token["valid"] = True
return queue_token
<commit_msg>Remove all unused code in text renderer.<commit_after>from PIL import Image, ImageFont, ImageDraw
class TextRenderer:
def __init__(self, font="./NotoSansCJK-Bold.otf",
font_color=(0, 120, 0), color_bg=False):
# params
self.color_bg = color_bg
self.font_color = font_color
# new image and font
self.font = ImageFont.truetype(font, 30)
def draw_text(self, text_to_send):
x, y = self.font.getsize(text_to_send)
# Add padding below, because PIL sucks!
image = Image.new("RGBA", (x, y+10), "black")
ImageDraw.Draw(image).text(
(0, 0), text_to_send, font=self.font, fill=self.font_color
)
return image
def get_queue_token(self, msgToken):
queue_token = {}
# TODO: add possible params
queue_token["image"] = [self.draw_text(' '.join(msgToken["text"]))]
queue_token["frame_count"] = 1
queue_token["action"] = "scroll"
queue_token["valid"] = True
return queue_token
|
edf38ad11631ad5e793eb9ac95dbc865595d517b
|
glue_vispy_viewers/common/layer_state.py
|
glue_vispy_viewers/common/layer_state.py
|
from __future__ import absolute_import, division, print_function
from glue.external.echo import CallbackProperty, keep_in_sync
from glue.core.state_objects import State
__all__ = ['VispyLayerState']
class VispyLayerState(State):
"""
A base state object for all Vispy layers
"""
layer = CallbackProperty()
visible = CallbackProperty(True)
zorder = CallbackProperty(0)
color = CallbackProperty()
alpha = CallbackProperty()
def __init__(self, **kwargs):
super(VispyLayerState, self).__init__(**kwargs)
self._sync_color = None
self._sync_alpha = None
self.add_callback('layer', self._layer_changed)
self._layer_changed()
def _layer_changed(self):
if self._sync_color is not None:
self._sync_color.stop_syncing()
if self._sync_alpha is not None:
self._sync_alpha.stop_syncing()
if self.layer is not None:
self.color = self.layer.style.color
self.alpha = self.layer.style.alpha
self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color')
self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha')
|
from __future__ import absolute_import, division, print_function
from glue.external.echo import CallbackProperty, keep_in_sync
from glue.core.state_objects import State
from glue.core.message import LayerArtistUpdatedMessage
__all__ = ['VispyLayerState']
class VispyLayerState(State):
"""
A base state object for all Vispy layers
"""
layer = CallbackProperty()
visible = CallbackProperty(True)
zorder = CallbackProperty(0)
color = CallbackProperty()
alpha = CallbackProperty()
def __init__(self, **kwargs):
super(VispyLayerState, self).__init__(**kwargs)
self._sync_color = None
self._sync_alpha = None
self.add_callback('layer', self._layer_changed)
self._layer_changed()
self.add_global_callback(self._notify_layer_update)
def _notify_layer_update(self, **kwargs):
message = LayerArtistUpdatedMessage(self)
if self.layer is not None and self.layer.hub is not None:
self.layer.hub.broadcast(message)
def _layer_changed(self):
if self._sync_color is not None:
self._sync_color.stop_syncing()
if self._sync_alpha is not None:
self._sync_alpha.stop_syncing()
if self.layer is not None:
self.color = self.layer.style.color
self.alpha = self.layer.style.alpha
self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color')
self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha')
|
Make sure layer artist icon updates when changing the color mode or colormaps
|
Make sure layer artist icon updates when changing the color mode or colormaps
|
Python
|
bsd-2-clause
|
glue-viz/glue-vispy-viewers,PennyQ/astro-vispy,astrofrog/glue-3d-viewer,glue-viz/glue-3d-viewer,astrofrog/glue-vispy-viewers
|
from __future__ import absolute_import, division, print_function
from glue.external.echo import CallbackProperty, keep_in_sync
from glue.core.state_objects import State
__all__ = ['VispyLayerState']
class VispyLayerState(State):
"""
A base state object for all Vispy layers
"""
layer = CallbackProperty()
visible = CallbackProperty(True)
zorder = CallbackProperty(0)
color = CallbackProperty()
alpha = CallbackProperty()
def __init__(self, **kwargs):
super(VispyLayerState, self).__init__(**kwargs)
self._sync_color = None
self._sync_alpha = None
self.add_callback('layer', self._layer_changed)
self._layer_changed()
def _layer_changed(self):
if self._sync_color is not None:
self._sync_color.stop_syncing()
if self._sync_alpha is not None:
self._sync_alpha.stop_syncing()
if self.layer is not None:
self.color = self.layer.style.color
self.alpha = self.layer.style.alpha
self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color')
self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha')
Make sure layer artist icon updates when changing the color mode or colormaps
|
from __future__ import absolute_import, division, print_function
from glue.external.echo import CallbackProperty, keep_in_sync
from glue.core.state_objects import State
from glue.core.message import LayerArtistUpdatedMessage
__all__ = ['VispyLayerState']
class VispyLayerState(State):
"""
A base state object for all Vispy layers
"""
layer = CallbackProperty()
visible = CallbackProperty(True)
zorder = CallbackProperty(0)
color = CallbackProperty()
alpha = CallbackProperty()
def __init__(self, **kwargs):
super(VispyLayerState, self).__init__(**kwargs)
self._sync_color = None
self._sync_alpha = None
self.add_callback('layer', self._layer_changed)
self._layer_changed()
self.add_global_callback(self._notify_layer_update)
def _notify_layer_update(self, **kwargs):
message = LayerArtistUpdatedMessage(self)
if self.layer is not None and self.layer.hub is not None:
self.layer.hub.broadcast(message)
def _layer_changed(self):
if self._sync_color is not None:
self._sync_color.stop_syncing()
if self._sync_alpha is not None:
self._sync_alpha.stop_syncing()
if self.layer is not None:
self.color = self.layer.style.color
self.alpha = self.layer.style.alpha
self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color')
self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha')
|
<commit_before>from __future__ import absolute_import, division, print_function
from glue.external.echo import CallbackProperty, keep_in_sync
from glue.core.state_objects import State
__all__ = ['VispyLayerState']
class VispyLayerState(State):
"""
A base state object for all Vispy layers
"""
layer = CallbackProperty()
visible = CallbackProperty(True)
zorder = CallbackProperty(0)
color = CallbackProperty()
alpha = CallbackProperty()
def __init__(self, **kwargs):
super(VispyLayerState, self).__init__(**kwargs)
self._sync_color = None
self._sync_alpha = None
self.add_callback('layer', self._layer_changed)
self._layer_changed()
def _layer_changed(self):
if self._sync_color is not None:
self._sync_color.stop_syncing()
if self._sync_alpha is not None:
self._sync_alpha.stop_syncing()
if self.layer is not None:
self.color = self.layer.style.color
self.alpha = self.layer.style.alpha
self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color')
self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha')
<commit_msg>Make sure layer artist icon updates when changing the color mode or colormaps<commit_after>
|
from __future__ import absolute_import, division, print_function
from glue.external.echo import CallbackProperty, keep_in_sync
from glue.core.state_objects import State
from glue.core.message import LayerArtistUpdatedMessage
__all__ = ['VispyLayerState']
class VispyLayerState(State):
"""
A base state object for all Vispy layers
"""
layer = CallbackProperty()
visible = CallbackProperty(True)
zorder = CallbackProperty(0)
color = CallbackProperty()
alpha = CallbackProperty()
def __init__(self, **kwargs):
super(VispyLayerState, self).__init__(**kwargs)
self._sync_color = None
self._sync_alpha = None
self.add_callback('layer', self._layer_changed)
self._layer_changed()
self.add_global_callback(self._notify_layer_update)
def _notify_layer_update(self, **kwargs):
message = LayerArtistUpdatedMessage(self)
if self.layer is not None and self.layer.hub is not None:
self.layer.hub.broadcast(message)
def _layer_changed(self):
if self._sync_color is not None:
self._sync_color.stop_syncing()
if self._sync_alpha is not None:
self._sync_alpha.stop_syncing()
if self.layer is not None:
self.color = self.layer.style.color
self.alpha = self.layer.style.alpha
self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color')
self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha')
|
from __future__ import absolute_import, division, print_function
from glue.external.echo import CallbackProperty, keep_in_sync
from glue.core.state_objects import State
__all__ = ['VispyLayerState']
class VispyLayerState(State):
"""
A base state object for all Vispy layers
"""
layer = CallbackProperty()
visible = CallbackProperty(True)
zorder = CallbackProperty(0)
color = CallbackProperty()
alpha = CallbackProperty()
def __init__(self, **kwargs):
super(VispyLayerState, self).__init__(**kwargs)
self._sync_color = None
self._sync_alpha = None
self.add_callback('layer', self._layer_changed)
self._layer_changed()
def _layer_changed(self):
if self._sync_color is not None:
self._sync_color.stop_syncing()
if self._sync_alpha is not None:
self._sync_alpha.stop_syncing()
if self.layer is not None:
self.color = self.layer.style.color
self.alpha = self.layer.style.alpha
self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color')
self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha')
Make sure layer artist icon updates when changing the color mode or colormapsfrom __future__ import absolute_import, division, print_function
from glue.external.echo import CallbackProperty, keep_in_sync
from glue.core.state_objects import State
from glue.core.message import LayerArtistUpdatedMessage
__all__ = ['VispyLayerState']
class VispyLayerState(State):
"""
A base state object for all Vispy layers
"""
layer = CallbackProperty()
visible = CallbackProperty(True)
zorder = CallbackProperty(0)
color = CallbackProperty()
alpha = CallbackProperty()
def __init__(self, **kwargs):
super(VispyLayerState, self).__init__(**kwargs)
self._sync_color = None
self._sync_alpha = None
self.add_callback('layer', self._layer_changed)
self._layer_changed()
self.add_global_callback(self._notify_layer_update)
def _notify_layer_update(self, **kwargs):
message = LayerArtistUpdatedMessage(self)
if self.layer is not None and self.layer.hub is not None:
self.layer.hub.broadcast(message)
def _layer_changed(self):
if self._sync_color is not None:
self._sync_color.stop_syncing()
if self._sync_alpha is not None:
self._sync_alpha.stop_syncing()
if self.layer is not None:
self.color = self.layer.style.color
self.alpha = self.layer.style.alpha
self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color')
self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha')
|
<commit_before>from __future__ import absolute_import, division, print_function
from glue.external.echo import CallbackProperty, keep_in_sync
from glue.core.state_objects import State
__all__ = ['VispyLayerState']
class VispyLayerState(State):
"""
A base state object for all Vispy layers
"""
layer = CallbackProperty()
visible = CallbackProperty(True)
zorder = CallbackProperty(0)
color = CallbackProperty()
alpha = CallbackProperty()
def __init__(self, **kwargs):
super(VispyLayerState, self).__init__(**kwargs)
self._sync_color = None
self._sync_alpha = None
self.add_callback('layer', self._layer_changed)
self._layer_changed()
def _layer_changed(self):
if self._sync_color is not None:
self._sync_color.stop_syncing()
if self._sync_alpha is not None:
self._sync_alpha.stop_syncing()
if self.layer is not None:
self.color = self.layer.style.color
self.alpha = self.layer.style.alpha
self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color')
self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha')
<commit_msg>Make sure layer artist icon updates when changing the color mode or colormaps<commit_after>from __future__ import absolute_import, division, print_function
from glue.external.echo import CallbackProperty, keep_in_sync
from glue.core.state_objects import State
from glue.core.message import LayerArtistUpdatedMessage
__all__ = ['VispyLayerState']
class VispyLayerState(State):
"""
A base state object for all Vispy layers
"""
layer = CallbackProperty()
visible = CallbackProperty(True)
zorder = CallbackProperty(0)
color = CallbackProperty()
alpha = CallbackProperty()
def __init__(self, **kwargs):
super(VispyLayerState, self).__init__(**kwargs)
self._sync_color = None
self._sync_alpha = None
self.add_callback('layer', self._layer_changed)
self._layer_changed()
self.add_global_callback(self._notify_layer_update)
def _notify_layer_update(self, **kwargs):
message = LayerArtistUpdatedMessage(self)
if self.layer is not None and self.layer.hub is not None:
self.layer.hub.broadcast(message)
def _layer_changed(self):
if self._sync_color is not None:
self._sync_color.stop_syncing()
if self._sync_alpha is not None:
self._sync_alpha.stop_syncing()
if self.layer is not None:
self.color = self.layer.style.color
self.alpha = self.layer.style.alpha
self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color')
self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha')
|
818d6584164f04001bf0e75f62c526284521ce69
|
demae/dest/s3_dest.py
|
demae/dest/s3_dest.py
|
import pandas as pd
import gzip
import boto3
import re
def default_key_map(key):
return re.sub('_input', '_output', key)
class S3Dest():
def __init__(self, key_map=default_key_map):
self.key_map = key_map
def skip_keys(self, bucket, source_prefix):
s3 = boto3.resource('s3')
objs = s3.Bucket(bucket).objects.filter(Prefix=self.key_map(source_prefix))
return [obj.key for obj in objs]
def put(self, data, obj):
body = self.generate_output_file(data)
dest_key = self.key_map(obj.key)
s3 = boto3.resource('s3')
s3.Object(obj.bucket_name, dest_key).put(Body=body)
def generate_output_file(self, data):
df = pd.DataFrame(data)
tsv = df.to_csv(sep='\t', header=False, index=False)
return gzip.compress(tsv.encode())
|
import pandas as pd
import gzip
import boto3
import re
import io
def default_key_map(key):
return re.sub('_input', '_output', key)
class S3Dest():
def __init__(self, key_map=default_key_map):
self.key_map = key_map
def skip_keys(self, bucket, source_prefix):
s3 = boto3.resource('s3')
objs = s3.Bucket(bucket).objects.filter(Prefix=self.key_map(source_prefix))
return [obj.key for obj in objs]
def put(self, data, obj):
body = self.generate_output_file(data)
dest_key = self.key_map(obj.key)
s3 = boto3.resource('s3')
s3.Object(obj.bucket_name, dest_key).upload_fileobj(io.BytesIO(body))
def generate_output_file(self, data):
df = pd.DataFrame(data)
tsv = df.to_csv(sep='\t', header=False, index=False)
return gzip.compress(tsv.encode())
|
Use managed transfer for uploading
|
Use managed transfer for uploading
|
Python
|
mit
|
uiureo/demae
|
import pandas as pd
import gzip
import boto3
import re
def default_key_map(key):
return re.sub('_input', '_output', key)
class S3Dest():
def __init__(self, key_map=default_key_map):
self.key_map = key_map
def skip_keys(self, bucket, source_prefix):
s3 = boto3.resource('s3')
objs = s3.Bucket(bucket).objects.filter(Prefix=self.key_map(source_prefix))
return [obj.key for obj in objs]
def put(self, data, obj):
body = self.generate_output_file(data)
dest_key = self.key_map(obj.key)
s3 = boto3.resource('s3')
s3.Object(obj.bucket_name, dest_key).put(Body=body)
def generate_output_file(self, data):
df = pd.DataFrame(data)
tsv = df.to_csv(sep='\t', header=False, index=False)
return gzip.compress(tsv.encode())
Use managed transfer for uploading
|
import pandas as pd
import gzip
import boto3
import re
import io
def default_key_map(key):
return re.sub('_input', '_output', key)
class S3Dest():
def __init__(self, key_map=default_key_map):
self.key_map = key_map
def skip_keys(self, bucket, source_prefix):
s3 = boto3.resource('s3')
objs = s3.Bucket(bucket).objects.filter(Prefix=self.key_map(source_prefix))
return [obj.key for obj in objs]
def put(self, data, obj):
body = self.generate_output_file(data)
dest_key = self.key_map(obj.key)
s3 = boto3.resource('s3')
s3.Object(obj.bucket_name, dest_key).upload_fileobj(io.BytesIO(body))
def generate_output_file(self, data):
df = pd.DataFrame(data)
tsv = df.to_csv(sep='\t', header=False, index=False)
return gzip.compress(tsv.encode())
|
<commit_before>import pandas as pd
import gzip
import boto3
import re
def default_key_map(key):
return re.sub('_input', '_output', key)
class S3Dest():
def __init__(self, key_map=default_key_map):
self.key_map = key_map
def skip_keys(self, bucket, source_prefix):
s3 = boto3.resource('s3')
objs = s3.Bucket(bucket).objects.filter(Prefix=self.key_map(source_prefix))
return [obj.key for obj in objs]
def put(self, data, obj):
body = self.generate_output_file(data)
dest_key = self.key_map(obj.key)
s3 = boto3.resource('s3')
s3.Object(obj.bucket_name, dest_key).put(Body=body)
def generate_output_file(self, data):
df = pd.DataFrame(data)
tsv = df.to_csv(sep='\t', header=False, index=False)
return gzip.compress(tsv.encode())
<commit_msg>Use managed transfer for uploading<commit_after>
|
import pandas as pd
import gzip
import boto3
import re
import io
def default_key_map(key):
return re.sub('_input', '_output', key)
class S3Dest():
def __init__(self, key_map=default_key_map):
self.key_map = key_map
def skip_keys(self, bucket, source_prefix):
s3 = boto3.resource('s3')
objs = s3.Bucket(bucket).objects.filter(Prefix=self.key_map(source_prefix))
return [obj.key for obj in objs]
def put(self, data, obj):
body = self.generate_output_file(data)
dest_key = self.key_map(obj.key)
s3 = boto3.resource('s3')
s3.Object(obj.bucket_name, dest_key).upload_fileobj(io.BytesIO(body))
def generate_output_file(self, data):
df = pd.DataFrame(data)
tsv = df.to_csv(sep='\t', header=False, index=False)
return gzip.compress(tsv.encode())
|
import pandas as pd
import gzip
import boto3
import re
def default_key_map(key):
return re.sub('_input', '_output', key)
class S3Dest():
def __init__(self, key_map=default_key_map):
self.key_map = key_map
def skip_keys(self, bucket, source_prefix):
s3 = boto3.resource('s3')
objs = s3.Bucket(bucket).objects.filter(Prefix=self.key_map(source_prefix))
return [obj.key for obj in objs]
def put(self, data, obj):
body = self.generate_output_file(data)
dest_key = self.key_map(obj.key)
s3 = boto3.resource('s3')
s3.Object(obj.bucket_name, dest_key).put(Body=body)
def generate_output_file(self, data):
df = pd.DataFrame(data)
tsv = df.to_csv(sep='\t', header=False, index=False)
return gzip.compress(tsv.encode())
Use managed transfer for uploadingimport pandas as pd
import gzip
import boto3
import re
import io
def default_key_map(key):
return re.sub('_input', '_output', key)
class S3Dest():
def __init__(self, key_map=default_key_map):
self.key_map = key_map
def skip_keys(self, bucket, source_prefix):
s3 = boto3.resource('s3')
objs = s3.Bucket(bucket).objects.filter(Prefix=self.key_map(source_prefix))
return [obj.key for obj in objs]
def put(self, data, obj):
body = self.generate_output_file(data)
dest_key = self.key_map(obj.key)
s3 = boto3.resource('s3')
s3.Object(obj.bucket_name, dest_key).upload_fileobj(io.BytesIO(body))
def generate_output_file(self, data):
df = pd.DataFrame(data)
tsv = df.to_csv(sep='\t', header=False, index=False)
return gzip.compress(tsv.encode())
|
<commit_before>import pandas as pd
import gzip
import boto3
import re
def default_key_map(key):
return re.sub('_input', '_output', key)
class S3Dest():
def __init__(self, key_map=default_key_map):
self.key_map = key_map
def skip_keys(self, bucket, source_prefix):
s3 = boto3.resource('s3')
objs = s3.Bucket(bucket).objects.filter(Prefix=self.key_map(source_prefix))
return [obj.key for obj in objs]
def put(self, data, obj):
body = self.generate_output_file(data)
dest_key = self.key_map(obj.key)
s3 = boto3.resource('s3')
s3.Object(obj.bucket_name, dest_key).put(Body=body)
def generate_output_file(self, data):
df = pd.DataFrame(data)
tsv = df.to_csv(sep='\t', header=False, index=False)
return gzip.compress(tsv.encode())
<commit_msg>Use managed transfer for uploading<commit_after>import pandas as pd
import gzip
import boto3
import re
import io
def default_key_map(key):
return re.sub('_input', '_output', key)
class S3Dest():
def __init__(self, key_map=default_key_map):
self.key_map = key_map
def skip_keys(self, bucket, source_prefix):
s3 = boto3.resource('s3')
objs = s3.Bucket(bucket).objects.filter(Prefix=self.key_map(source_prefix))
return [obj.key for obj in objs]
def put(self, data, obj):
body = self.generate_output_file(data)
dest_key = self.key_map(obj.key)
s3 = boto3.resource('s3')
s3.Object(obj.bucket_name, dest_key).upload_fileobj(io.BytesIO(body))
def generate_output_file(self, data):
df = pd.DataFrame(data)
tsv = df.to_csv(sep='\t', header=False, index=False)
return gzip.compress(tsv.encode())
|
f6e18d142ac965221737205f65d66751ea02f168
|
hack_plot/management/commands/parse_authlog.py
|
hack_plot/management/commands/parse_authlog.py
|
from django.core.management.base import BaseCommand, CommandError
from ...cron import parse_auth_log
class Command(BaseCommand):
def handle(self, *args, **options):
parse_auth_log()
|
from django.conf import settings
from django.core.management.base import BaseCommand, CommandError
try:
import simplejson as json
except ImportError as e:
import json
from rest_framework.renderers import JSONRenderer
from unipath import Path
from ...api.serializers import HackLocationSerializer
from ...cron import parse_auth_log
from ...models import SshHackLocation
class Command(BaseCommand):
def handle(self, *args, **options):
parse_auth_log()
# Write the attempt data to json for fast AJAX loading
serializer = HackLocationSerializer(SshHackLocation.objects.all(), many=True)
data = JSONRenderer().render(serializer.data, 'application/json', {})
json_output_file = Path(settings.STATIC_ROOT).child('hack_location.json')
with open(json_output_file, 'wb') as f:
f.write(data)
|
Write hack locations to json after parsing log file
|
Write hack locations to json after parsing log file
|
Python
|
mit
|
hellsgate1001/graphs,hellsgate1001/graphs,hellsgate1001/graphs
|
from django.core.management.base import BaseCommand, CommandError
from ...cron import parse_auth_log
class Command(BaseCommand):
def handle(self, *args, **options):
parse_auth_log()
Write hack locations to json after parsing log file
|
from django.conf import settings
from django.core.management.base import BaseCommand, CommandError
try:
import simplejson as json
except ImportError as e:
import json
from rest_framework.renderers import JSONRenderer
from unipath import Path
from ...api.serializers import HackLocationSerializer
from ...cron import parse_auth_log
from ...models import SshHackLocation
class Command(BaseCommand):
def handle(self, *args, **options):
parse_auth_log()
# Write the attempt data to json for fast AJAX loading
serializer = HackLocationSerializer(SshHackLocation.objects.all(), many=True)
data = JSONRenderer().render(serializer.data, 'application/json', {})
json_output_file = Path(settings.STATIC_ROOT).child('hack_location.json')
with open(json_output_file, 'wb') as f:
f.write(data)
|
<commit_before>from django.core.management.base import BaseCommand, CommandError
from ...cron import parse_auth_log
class Command(BaseCommand):
def handle(self, *args, **options):
parse_auth_log()
<commit_msg>Write hack locations to json after parsing log file<commit_after>
|
from django.conf import settings
from django.core.management.base import BaseCommand, CommandError
try:
import simplejson as json
except ImportError as e:
import json
from rest_framework.renderers import JSONRenderer
from unipath import Path
from ...api.serializers import HackLocationSerializer
from ...cron import parse_auth_log
from ...models import SshHackLocation
class Command(BaseCommand):
def handle(self, *args, **options):
parse_auth_log()
# Write the attempt data to json for fast AJAX loading
serializer = HackLocationSerializer(SshHackLocation.objects.all(), many=True)
data = JSONRenderer().render(serializer.data, 'application/json', {})
json_output_file = Path(settings.STATIC_ROOT).child('hack_location.json')
with open(json_output_file, 'wb') as f:
f.write(data)
|
from django.core.management.base import BaseCommand, CommandError
from ...cron import parse_auth_log
class Command(BaseCommand):
def handle(self, *args, **options):
parse_auth_log()
Write hack locations to json after parsing log filefrom django.conf import settings
from django.core.management.base import BaseCommand, CommandError
try:
import simplejson as json
except ImportError as e:
import json
from rest_framework.renderers import JSONRenderer
from unipath import Path
from ...api.serializers import HackLocationSerializer
from ...cron import parse_auth_log
from ...models import SshHackLocation
class Command(BaseCommand):
def handle(self, *args, **options):
parse_auth_log()
# Write the attempt data to json for fast AJAX loading
serializer = HackLocationSerializer(SshHackLocation.objects.all(), many=True)
data = JSONRenderer().render(serializer.data, 'application/json', {})
json_output_file = Path(settings.STATIC_ROOT).child('hack_location.json')
with open(json_output_file, 'wb') as f:
f.write(data)
|
<commit_before>from django.core.management.base import BaseCommand, CommandError
from ...cron import parse_auth_log
class Command(BaseCommand):
def handle(self, *args, **options):
parse_auth_log()
<commit_msg>Write hack locations to json after parsing log file<commit_after>from django.conf import settings
from django.core.management.base import BaseCommand, CommandError
try:
import simplejson as json
except ImportError as e:
import json
from rest_framework.renderers import JSONRenderer
from unipath import Path
from ...api.serializers import HackLocationSerializer
from ...cron import parse_auth_log
from ...models import SshHackLocation
class Command(BaseCommand):
def handle(self, *args, **options):
parse_auth_log()
# Write the attempt data to json for fast AJAX loading
serializer = HackLocationSerializer(SshHackLocation.objects.all(), many=True)
data = JSONRenderer().render(serializer.data, 'application/json', {})
json_output_file = Path(settings.STATIC_ROOT).child('hack_location.json')
with open(json_output_file, 'wb') as f:
f.write(data)
|
49069663a3fe3d44be9ab59e59a90d0dfcf49f0c
|
mayatools/qt.py
|
mayatools/qt.py
|
try:
import sip
from uitools.qt import QtCore
import maya.OpenMayaUI as apiUI
# These modules will not exist while building the docs.
except ImportError:
import os
if os.environ.get('SPHINX') != 'True':
raise
def get_maya_window():
"""Get the main Maya window as a QtGui.QMainWindow."""
ptr = apiUI.MQtUtil.mainWindow()
if ptr is not None:
return sip.wrapinstance(long(ptr), QtCore.QObject)
def maya_to_qt(maya_object):
"""Convert a Maya UI path to a Qt object.
:param str maya_object: The path of the Maya UI object to convert.
:returns: QtCore.QObject or None
"""
ptr = (
apiUI.MQtUtil.findControl(maya_object) or
apiUI.MQtUtil.findLayout(maya_object) or
apiUI.MQtUtil.findMenuItem(maya_object)
)
if ptr is not None:
return sip.wrapinstance(long(ptr), QtCore.QObject)
|
try:
from uitools.sip import wrapinstance
from uitools.qt import QtCore
import maya.OpenMayaUI as apiUI
# These modules will not exist while building the docs.
except ImportError:
import os
if os.environ.get('SPHINX') != 'True':
raise
def get_maya_window():
"""Get the main Maya window as a QtGui.QMainWindow."""
ptr = apiUI.MQtUtil.mainWindow()
if ptr is not None:
return wrapinstance(long(ptr), QtCore.QObject)
def maya_to_qt(maya_object):
"""Convert a Maya UI path to a Qt object.
:param str maya_object: The path of the Maya UI object to convert.
:returns: QtCore.QObject or None
"""
ptr = (
apiUI.MQtUtil.findControl(maya_object) or
apiUI.MQtUtil.findLayout(maya_object) or
apiUI.MQtUtil.findMenuItem(maya_object)
)
if ptr is not None:
return wrapinstance(long(ptr), QtCore.QObject)
|
Use uitools.sip instead of straight sip
|
Use uitools.sip instead of straight sip
|
Python
|
bsd-3-clause
|
westernx/mayatools,westernx/mayatools
|
try:
import sip
from uitools.qt import QtCore
import maya.OpenMayaUI as apiUI
# These modules will not exist while building the docs.
except ImportError:
import os
if os.environ.get('SPHINX') != 'True':
raise
def get_maya_window():
"""Get the main Maya window as a QtGui.QMainWindow."""
ptr = apiUI.MQtUtil.mainWindow()
if ptr is not None:
return sip.wrapinstance(long(ptr), QtCore.QObject)
def maya_to_qt(maya_object):
"""Convert a Maya UI path to a Qt object.
:param str maya_object: The path of the Maya UI object to convert.
:returns: QtCore.QObject or None
"""
ptr = (
apiUI.MQtUtil.findControl(maya_object) or
apiUI.MQtUtil.findLayout(maya_object) or
apiUI.MQtUtil.findMenuItem(maya_object)
)
if ptr is not None:
return sip.wrapinstance(long(ptr), QtCore.QObject)Use uitools.sip instead of straight sip
|
try:
from uitools.sip import wrapinstance
from uitools.qt import QtCore
import maya.OpenMayaUI as apiUI
# These modules will not exist while building the docs.
except ImportError:
import os
if os.environ.get('SPHINX') != 'True':
raise
def get_maya_window():
"""Get the main Maya window as a QtGui.QMainWindow."""
ptr = apiUI.MQtUtil.mainWindow()
if ptr is not None:
return wrapinstance(long(ptr), QtCore.QObject)
def maya_to_qt(maya_object):
"""Convert a Maya UI path to a Qt object.
:param str maya_object: The path of the Maya UI object to convert.
:returns: QtCore.QObject or None
"""
ptr = (
apiUI.MQtUtil.findControl(maya_object) or
apiUI.MQtUtil.findLayout(maya_object) or
apiUI.MQtUtil.findMenuItem(maya_object)
)
if ptr is not None:
return wrapinstance(long(ptr), QtCore.QObject)
|
<commit_before>
try:
import sip
from uitools.qt import QtCore
import maya.OpenMayaUI as apiUI
# These modules will not exist while building the docs.
except ImportError:
import os
if os.environ.get('SPHINX') != 'True':
raise
def get_maya_window():
"""Get the main Maya window as a QtGui.QMainWindow."""
ptr = apiUI.MQtUtil.mainWindow()
if ptr is not None:
return sip.wrapinstance(long(ptr), QtCore.QObject)
def maya_to_qt(maya_object):
"""Convert a Maya UI path to a Qt object.
:param str maya_object: The path of the Maya UI object to convert.
:returns: QtCore.QObject or None
"""
ptr = (
apiUI.MQtUtil.findControl(maya_object) or
apiUI.MQtUtil.findLayout(maya_object) or
apiUI.MQtUtil.findMenuItem(maya_object)
)
if ptr is not None:
return sip.wrapinstance(long(ptr), QtCore.QObject)<commit_msg>Use uitools.sip instead of straight sip<commit_after>
|
try:
from uitools.sip import wrapinstance
from uitools.qt import QtCore
import maya.OpenMayaUI as apiUI
# These modules will not exist while building the docs.
except ImportError:
import os
if os.environ.get('SPHINX') != 'True':
raise
def get_maya_window():
"""Get the main Maya window as a QtGui.QMainWindow."""
ptr = apiUI.MQtUtil.mainWindow()
if ptr is not None:
return wrapinstance(long(ptr), QtCore.QObject)
def maya_to_qt(maya_object):
"""Convert a Maya UI path to a Qt object.
:param str maya_object: The path of the Maya UI object to convert.
:returns: QtCore.QObject or None
"""
ptr = (
apiUI.MQtUtil.findControl(maya_object) or
apiUI.MQtUtil.findLayout(maya_object) or
apiUI.MQtUtil.findMenuItem(maya_object)
)
if ptr is not None:
return wrapinstance(long(ptr), QtCore.QObject)
|
try:
import sip
from uitools.qt import QtCore
import maya.OpenMayaUI as apiUI
# These modules will not exist while building the docs.
except ImportError:
import os
if os.environ.get('SPHINX') != 'True':
raise
def get_maya_window():
"""Get the main Maya window as a QtGui.QMainWindow."""
ptr = apiUI.MQtUtil.mainWindow()
if ptr is not None:
return sip.wrapinstance(long(ptr), QtCore.QObject)
def maya_to_qt(maya_object):
"""Convert a Maya UI path to a Qt object.
:param str maya_object: The path of the Maya UI object to convert.
:returns: QtCore.QObject or None
"""
ptr = (
apiUI.MQtUtil.findControl(maya_object) or
apiUI.MQtUtil.findLayout(maya_object) or
apiUI.MQtUtil.findMenuItem(maya_object)
)
if ptr is not None:
return sip.wrapinstance(long(ptr), QtCore.QObject)Use uitools.sip instead of straight sip
try:
from uitools.sip import wrapinstance
from uitools.qt import QtCore
import maya.OpenMayaUI as apiUI
# These modules will not exist while building the docs.
except ImportError:
import os
if os.environ.get('SPHINX') != 'True':
raise
def get_maya_window():
"""Get the main Maya window as a QtGui.QMainWindow."""
ptr = apiUI.MQtUtil.mainWindow()
if ptr is not None:
return wrapinstance(long(ptr), QtCore.QObject)
def maya_to_qt(maya_object):
"""Convert a Maya UI path to a Qt object.
:param str maya_object: The path of the Maya UI object to convert.
:returns: QtCore.QObject or None
"""
ptr = (
apiUI.MQtUtil.findControl(maya_object) or
apiUI.MQtUtil.findLayout(maya_object) or
apiUI.MQtUtil.findMenuItem(maya_object)
)
if ptr is not None:
return wrapinstance(long(ptr), QtCore.QObject)
|
<commit_before>
try:
import sip
from uitools.qt import QtCore
import maya.OpenMayaUI as apiUI
# These modules will not exist while building the docs.
except ImportError:
import os
if os.environ.get('SPHINX') != 'True':
raise
def get_maya_window():
"""Get the main Maya window as a QtGui.QMainWindow."""
ptr = apiUI.MQtUtil.mainWindow()
if ptr is not None:
return sip.wrapinstance(long(ptr), QtCore.QObject)
def maya_to_qt(maya_object):
"""Convert a Maya UI path to a Qt object.
:param str maya_object: The path of the Maya UI object to convert.
:returns: QtCore.QObject or None
"""
ptr = (
apiUI.MQtUtil.findControl(maya_object) or
apiUI.MQtUtil.findLayout(maya_object) or
apiUI.MQtUtil.findMenuItem(maya_object)
)
if ptr is not None:
return sip.wrapinstance(long(ptr), QtCore.QObject)<commit_msg>Use uitools.sip instead of straight sip<commit_after>
try:
from uitools.sip import wrapinstance
from uitools.qt import QtCore
import maya.OpenMayaUI as apiUI
# These modules will not exist while building the docs.
except ImportError:
import os
if os.environ.get('SPHINX') != 'True':
raise
def get_maya_window():
"""Get the main Maya window as a QtGui.QMainWindow."""
ptr = apiUI.MQtUtil.mainWindow()
if ptr is not None:
return wrapinstance(long(ptr), QtCore.QObject)
def maya_to_qt(maya_object):
"""Convert a Maya UI path to a Qt object.
:param str maya_object: The path of the Maya UI object to convert.
:returns: QtCore.QObject or None
"""
ptr = (
apiUI.MQtUtil.findControl(maya_object) or
apiUI.MQtUtil.findLayout(maya_object) or
apiUI.MQtUtil.findMenuItem(maya_object)
)
if ptr is not None:
return wrapinstance(long(ptr), QtCore.QObject)
|
3d4327f6d9d71c6b396b0655de81373210417aba
|
apps/i4p_base/urls.py
|
apps/i4p_base/urls.py
|
#-- encoding: utf-8 --
from django.conf.urls.defaults import patterns, url
from haystack.views import search_view_factory
import views
import ajax
urlpatterns = patterns('',
url(r'^$', views.homepage, name='i4p-index'),
url(r'^homepage/ajax/slider/bestof/$', ajax.slider_bestof, name='i4p-homepage-ajax-slider-bestof'),
url(r'^homepage/ajax/slider/latest/$', ajax.slider_latest, name='i4p-homepage-ajax-slider-latest'),
url(r'^homepage/ajax/slider/commented/$', ajax.slider_most_commented, name='i4p-homepage-ajax-slider-commented'),
url(r'^history/check_version/(?P<pk>[\d]+)$', views.VersionActivityCheckView.as_view(), name='history-check-version'),
url(r'^search/', search_view_factory(view_class=views.SearchView), name='i4p-search'),
url(r'^location/(?P<location_id>\d+)', views.LocationEditView.as_view(), name='i4p-location-edit'),
url(r'^locations/$', views.LocationListView.as_view(), name='i4p-location-list'),
url(r'^locations/missing/(?P<missing_field_name>\w+)$', views.LocationListView.as_view(), name='i4p-location-missing-list'),
)
|
#-- encoding: utf-8 --
from django.conf.urls.defaults import patterns, url
from haystack.views import search_view_factory
import views
import ajax
urlpatterns = patterns('',
#url(r'^$', views.homepage, name='i4p-index'),
url(r'^homepage/ajax/slider/bestof/$', ajax.slider_bestof, name='i4p-homepage-ajax-slider-bestof'),
url(r'^homepage/ajax/slider/latest/$', ajax.slider_latest, name='i4p-homepage-ajax-slider-latest'),
url(r'^homepage/ajax/slider/commented/$', ajax.slider_most_commented, name='i4p-homepage-ajax-slider-commented'),
url(r'^history/check_version/(?P<pk>[\d]+)$', views.VersionActivityCheckView.as_view(), name='history-check-version'),
url(r'^search/', search_view_factory(view_class=views.SearchView), name='i4p-search'),
url(r'^location/(?P<location_id>\d+)', views.LocationEditView.as_view(), name='i4p-location-edit'),
url(r'^locations/$', views.LocationListView.as_view(), name='i4p-location-list'),
url(r'^locations/missing/(?P<missing_field_name>\w+)$', views.LocationListView.as_view(), name='i4p-location-missing-list'),
)
|
Remove explicit link to homepage view in i4p_base
|
Remove explicit link to homepage view in i4p_base
|
Python
|
agpl-3.0
|
ImaginationForPeople/imaginationforpeople,ImaginationForPeople/imaginationforpeople,ImaginationForPeople/imaginationforpeople,ImaginationForPeople/imaginationforpeople,ImaginationForPeople/imaginationforpeople,ImaginationForPeople/imaginationforpeople
|
#-- encoding: utf-8 --
from django.conf.urls.defaults import patterns, url
from haystack.views import search_view_factory
import views
import ajax
urlpatterns = patterns('',
url(r'^$', views.homepage, name='i4p-index'),
url(r'^homepage/ajax/slider/bestof/$', ajax.slider_bestof, name='i4p-homepage-ajax-slider-bestof'),
url(r'^homepage/ajax/slider/latest/$', ajax.slider_latest, name='i4p-homepage-ajax-slider-latest'),
url(r'^homepage/ajax/slider/commented/$', ajax.slider_most_commented, name='i4p-homepage-ajax-slider-commented'),
url(r'^history/check_version/(?P<pk>[\d]+)$', views.VersionActivityCheckView.as_view(), name='history-check-version'),
url(r'^search/', search_view_factory(view_class=views.SearchView), name='i4p-search'),
url(r'^location/(?P<location_id>\d+)', views.LocationEditView.as_view(), name='i4p-location-edit'),
url(r'^locations/$', views.LocationListView.as_view(), name='i4p-location-list'),
url(r'^locations/missing/(?P<missing_field_name>\w+)$', views.LocationListView.as_view(), name='i4p-location-missing-list'),
)
Remove explicit link to homepage view in i4p_base
|
#-- encoding: utf-8 --
from django.conf.urls.defaults import patterns, url
from haystack.views import search_view_factory
import views
import ajax
urlpatterns = patterns('',
#url(r'^$', views.homepage, name='i4p-index'),
url(r'^homepage/ajax/slider/bestof/$', ajax.slider_bestof, name='i4p-homepage-ajax-slider-bestof'),
url(r'^homepage/ajax/slider/latest/$', ajax.slider_latest, name='i4p-homepage-ajax-slider-latest'),
url(r'^homepage/ajax/slider/commented/$', ajax.slider_most_commented, name='i4p-homepage-ajax-slider-commented'),
url(r'^history/check_version/(?P<pk>[\d]+)$', views.VersionActivityCheckView.as_view(), name='history-check-version'),
url(r'^search/', search_view_factory(view_class=views.SearchView), name='i4p-search'),
url(r'^location/(?P<location_id>\d+)', views.LocationEditView.as_view(), name='i4p-location-edit'),
url(r'^locations/$', views.LocationListView.as_view(), name='i4p-location-list'),
url(r'^locations/missing/(?P<missing_field_name>\w+)$', views.LocationListView.as_view(), name='i4p-location-missing-list'),
)
|
<commit_before>#-- encoding: utf-8 --
from django.conf.urls.defaults import patterns, url
from haystack.views import search_view_factory
import views
import ajax
urlpatterns = patterns('',
url(r'^$', views.homepage, name='i4p-index'),
url(r'^homepage/ajax/slider/bestof/$', ajax.slider_bestof, name='i4p-homepage-ajax-slider-bestof'),
url(r'^homepage/ajax/slider/latest/$', ajax.slider_latest, name='i4p-homepage-ajax-slider-latest'),
url(r'^homepage/ajax/slider/commented/$', ajax.slider_most_commented, name='i4p-homepage-ajax-slider-commented'),
url(r'^history/check_version/(?P<pk>[\d]+)$', views.VersionActivityCheckView.as_view(), name='history-check-version'),
url(r'^search/', search_view_factory(view_class=views.SearchView), name='i4p-search'),
url(r'^location/(?P<location_id>\d+)', views.LocationEditView.as_view(), name='i4p-location-edit'),
url(r'^locations/$', views.LocationListView.as_view(), name='i4p-location-list'),
url(r'^locations/missing/(?P<missing_field_name>\w+)$', views.LocationListView.as_view(), name='i4p-location-missing-list'),
)
<commit_msg>Remove explicit link to homepage view in i4p_base<commit_after>
|
#-- encoding: utf-8 --
from django.conf.urls.defaults import patterns, url
from haystack.views import search_view_factory
import views
import ajax
urlpatterns = patterns('',
#url(r'^$', views.homepage, name='i4p-index'),
url(r'^homepage/ajax/slider/bestof/$', ajax.slider_bestof, name='i4p-homepage-ajax-slider-bestof'),
url(r'^homepage/ajax/slider/latest/$', ajax.slider_latest, name='i4p-homepage-ajax-slider-latest'),
url(r'^homepage/ajax/slider/commented/$', ajax.slider_most_commented, name='i4p-homepage-ajax-slider-commented'),
url(r'^history/check_version/(?P<pk>[\d]+)$', views.VersionActivityCheckView.as_view(), name='history-check-version'),
url(r'^search/', search_view_factory(view_class=views.SearchView), name='i4p-search'),
url(r'^location/(?P<location_id>\d+)', views.LocationEditView.as_view(), name='i4p-location-edit'),
url(r'^locations/$', views.LocationListView.as_view(), name='i4p-location-list'),
url(r'^locations/missing/(?P<missing_field_name>\w+)$', views.LocationListView.as_view(), name='i4p-location-missing-list'),
)
|
#-- encoding: utf-8 --
from django.conf.urls.defaults import patterns, url
from haystack.views import search_view_factory
import views
import ajax
urlpatterns = patterns('',
url(r'^$', views.homepage, name='i4p-index'),
url(r'^homepage/ajax/slider/bestof/$', ajax.slider_bestof, name='i4p-homepage-ajax-slider-bestof'),
url(r'^homepage/ajax/slider/latest/$', ajax.slider_latest, name='i4p-homepage-ajax-slider-latest'),
url(r'^homepage/ajax/slider/commented/$', ajax.slider_most_commented, name='i4p-homepage-ajax-slider-commented'),
url(r'^history/check_version/(?P<pk>[\d]+)$', views.VersionActivityCheckView.as_view(), name='history-check-version'),
url(r'^search/', search_view_factory(view_class=views.SearchView), name='i4p-search'),
url(r'^location/(?P<location_id>\d+)', views.LocationEditView.as_view(), name='i4p-location-edit'),
url(r'^locations/$', views.LocationListView.as_view(), name='i4p-location-list'),
url(r'^locations/missing/(?P<missing_field_name>\w+)$', views.LocationListView.as_view(), name='i4p-location-missing-list'),
)
Remove explicit link to homepage view in i4p_base#-- encoding: utf-8 --
from django.conf.urls.defaults import patterns, url
from haystack.views import search_view_factory
import views
import ajax
urlpatterns = patterns('',
#url(r'^$', views.homepage, name='i4p-index'),
url(r'^homepage/ajax/slider/bestof/$', ajax.slider_bestof, name='i4p-homepage-ajax-slider-bestof'),
url(r'^homepage/ajax/slider/latest/$', ajax.slider_latest, name='i4p-homepage-ajax-slider-latest'),
url(r'^homepage/ajax/slider/commented/$', ajax.slider_most_commented, name='i4p-homepage-ajax-slider-commented'),
url(r'^history/check_version/(?P<pk>[\d]+)$', views.VersionActivityCheckView.as_view(), name='history-check-version'),
url(r'^search/', search_view_factory(view_class=views.SearchView), name='i4p-search'),
url(r'^location/(?P<location_id>\d+)', views.LocationEditView.as_view(), name='i4p-location-edit'),
url(r'^locations/$', views.LocationListView.as_view(), name='i4p-location-list'),
url(r'^locations/missing/(?P<missing_field_name>\w+)$', views.LocationListView.as_view(), name='i4p-location-missing-list'),
)
|
<commit_before>#-- encoding: utf-8 --
from django.conf.urls.defaults import patterns, url
from haystack.views import search_view_factory
import views
import ajax
urlpatterns = patterns('',
url(r'^$', views.homepage, name='i4p-index'),
url(r'^homepage/ajax/slider/bestof/$', ajax.slider_bestof, name='i4p-homepage-ajax-slider-bestof'),
url(r'^homepage/ajax/slider/latest/$', ajax.slider_latest, name='i4p-homepage-ajax-slider-latest'),
url(r'^homepage/ajax/slider/commented/$', ajax.slider_most_commented, name='i4p-homepage-ajax-slider-commented'),
url(r'^history/check_version/(?P<pk>[\d]+)$', views.VersionActivityCheckView.as_view(), name='history-check-version'),
url(r'^search/', search_view_factory(view_class=views.SearchView), name='i4p-search'),
url(r'^location/(?P<location_id>\d+)', views.LocationEditView.as_view(), name='i4p-location-edit'),
url(r'^locations/$', views.LocationListView.as_view(), name='i4p-location-list'),
url(r'^locations/missing/(?P<missing_field_name>\w+)$', views.LocationListView.as_view(), name='i4p-location-missing-list'),
)
<commit_msg>Remove explicit link to homepage view in i4p_base<commit_after>#-- encoding: utf-8 --
from django.conf.urls.defaults import patterns, url
from haystack.views import search_view_factory
import views
import ajax
urlpatterns = patterns('',
#url(r'^$', views.homepage, name='i4p-index'),
url(r'^homepage/ajax/slider/bestof/$', ajax.slider_bestof, name='i4p-homepage-ajax-slider-bestof'),
url(r'^homepage/ajax/slider/latest/$', ajax.slider_latest, name='i4p-homepage-ajax-slider-latest'),
url(r'^homepage/ajax/slider/commented/$', ajax.slider_most_commented, name='i4p-homepage-ajax-slider-commented'),
url(r'^history/check_version/(?P<pk>[\d]+)$', views.VersionActivityCheckView.as_view(), name='history-check-version'),
url(r'^search/', search_view_factory(view_class=views.SearchView), name='i4p-search'),
url(r'^location/(?P<location_id>\d+)', views.LocationEditView.as_view(), name='i4p-location-edit'),
url(r'^locations/$', views.LocationListView.as_view(), name='i4p-location-list'),
url(r'^locations/missing/(?P<missing_field_name>\w+)$', views.LocationListView.as_view(), name='i4p-location-missing-list'),
)
|
5446b0cc9335a3fe6c88158c1b864cdc1b0988d5
|
onestop/stopbins.py
|
onestop/stopbins.py
|
"""Stop Bins."""
import util
import errors
import registry
import entities
class StopBin(object):
def __init__(self, prefix):
self.prefix = prefix
self._stops = {}
def stops(self):
return self._stops.values()
def add_stop(self, stop):
key = stop.onestop()
# New stop
if key not in self._stops:
self._stops[key] = stop
else:
self._stops[key].merge(stop)
@classmethod
def from_json(cls, data):
stopbin = cls(prefix=data['prefix'])
for feature in data['features']:
stop = entities.OnestopStop.from_json(feature)
stopbin.add_stop(stop)
return stopbin
def json(self):
return {
'type': 'FeatureCollection',
'properties': {},
'prefix': self.prefix,
'features': [
i.json()
for i in
sorted(self.stops(), key=lambda x:x.onestop())
]
}
|
"""Stop Bins."""
import util
import errors
import registry
import entities
class StopBin(object):
def __init__(self, prefix):
self.prefix = prefix
self._stops = {}
def stops(self):
return self._stops.values()
def add_stop(self, stop):
key = stop.onestop()
# New stop
if key not in self._stops:
self._stops[key] = stop
else:
self._stops[key].merge(stop)
return self._stops[key]
@classmethod
def from_json(cls, data):
stopbin = cls(prefix=data['prefix'])
for feature in data['features']:
stop = entities.OnestopStop.from_json(feature)
stopbin.add_stop(stop)
return stopbin
def json(self):
return {
'type': 'FeatureCollection',
'properties': {},
'prefix': self.prefix,
'features': [
i.json()
for i in
sorted(self.stops(), key=lambda x:x.onestop())
]
}
|
Return added stop in StopBin.add_stop()
|
Return added stop in StopBin.add_stop()
|
Python
|
mit
|
transitland/transitland-python-client,srthurman/transitland-python-client
|
"""Stop Bins."""
import util
import errors
import registry
import entities
class StopBin(object):
def __init__(self, prefix):
self.prefix = prefix
self._stops = {}
def stops(self):
return self._stops.values()
def add_stop(self, stop):
key = stop.onestop()
# New stop
if key not in self._stops:
self._stops[key] = stop
else:
self._stops[key].merge(stop)
@classmethod
def from_json(cls, data):
stopbin = cls(prefix=data['prefix'])
for feature in data['features']:
stop = entities.OnestopStop.from_json(feature)
stopbin.add_stop(stop)
return stopbin
def json(self):
return {
'type': 'FeatureCollection',
'properties': {},
'prefix': self.prefix,
'features': [
i.json()
for i in
sorted(self.stops(), key=lambda x:x.onestop())
]
}
Return added stop in StopBin.add_stop()
|
"""Stop Bins."""
import util
import errors
import registry
import entities
class StopBin(object):
def __init__(self, prefix):
self.prefix = prefix
self._stops = {}
def stops(self):
return self._stops.values()
def add_stop(self, stop):
key = stop.onestop()
# New stop
if key not in self._stops:
self._stops[key] = stop
else:
self._stops[key].merge(stop)
return self._stops[key]
@classmethod
def from_json(cls, data):
stopbin = cls(prefix=data['prefix'])
for feature in data['features']:
stop = entities.OnestopStop.from_json(feature)
stopbin.add_stop(stop)
return stopbin
def json(self):
return {
'type': 'FeatureCollection',
'properties': {},
'prefix': self.prefix,
'features': [
i.json()
for i in
sorted(self.stops(), key=lambda x:x.onestop())
]
}
|
<commit_before>"""Stop Bins."""
import util
import errors
import registry
import entities
class StopBin(object):
def __init__(self, prefix):
self.prefix = prefix
self._stops = {}
def stops(self):
return self._stops.values()
def add_stop(self, stop):
key = stop.onestop()
# New stop
if key not in self._stops:
self._stops[key] = stop
else:
self._stops[key].merge(stop)
@classmethod
def from_json(cls, data):
stopbin = cls(prefix=data['prefix'])
for feature in data['features']:
stop = entities.OnestopStop.from_json(feature)
stopbin.add_stop(stop)
return stopbin
def json(self):
return {
'type': 'FeatureCollection',
'properties': {},
'prefix': self.prefix,
'features': [
i.json()
for i in
sorted(self.stops(), key=lambda x:x.onestop())
]
}
<commit_msg>Return added stop in StopBin.add_stop()<commit_after>
|
"""Stop Bins."""
import util
import errors
import registry
import entities
class StopBin(object):
def __init__(self, prefix):
self.prefix = prefix
self._stops = {}
def stops(self):
return self._stops.values()
def add_stop(self, stop):
key = stop.onestop()
# New stop
if key not in self._stops:
self._stops[key] = stop
else:
self._stops[key].merge(stop)
return self._stops[key]
@classmethod
def from_json(cls, data):
stopbin = cls(prefix=data['prefix'])
for feature in data['features']:
stop = entities.OnestopStop.from_json(feature)
stopbin.add_stop(stop)
return stopbin
def json(self):
return {
'type': 'FeatureCollection',
'properties': {},
'prefix': self.prefix,
'features': [
i.json()
for i in
sorted(self.stops(), key=lambda x:x.onestop())
]
}
|
"""Stop Bins."""
import util
import errors
import registry
import entities
class StopBin(object):
def __init__(self, prefix):
self.prefix = prefix
self._stops = {}
def stops(self):
return self._stops.values()
def add_stop(self, stop):
key = stop.onestop()
# New stop
if key not in self._stops:
self._stops[key] = stop
else:
self._stops[key].merge(stop)
@classmethod
def from_json(cls, data):
stopbin = cls(prefix=data['prefix'])
for feature in data['features']:
stop = entities.OnestopStop.from_json(feature)
stopbin.add_stop(stop)
return stopbin
def json(self):
return {
'type': 'FeatureCollection',
'properties': {},
'prefix': self.prefix,
'features': [
i.json()
for i in
sorted(self.stops(), key=lambda x:x.onestop())
]
}
Return added stop in StopBin.add_stop()"""Stop Bins."""
import util
import errors
import registry
import entities
class StopBin(object):
def __init__(self, prefix):
self.prefix = prefix
self._stops = {}
def stops(self):
return self._stops.values()
def add_stop(self, stop):
key = stop.onestop()
# New stop
if key not in self._stops:
self._stops[key] = stop
else:
self._stops[key].merge(stop)
return self._stops[key]
@classmethod
def from_json(cls, data):
stopbin = cls(prefix=data['prefix'])
for feature in data['features']:
stop = entities.OnestopStop.from_json(feature)
stopbin.add_stop(stop)
return stopbin
def json(self):
return {
'type': 'FeatureCollection',
'properties': {},
'prefix': self.prefix,
'features': [
i.json()
for i in
sorted(self.stops(), key=lambda x:x.onestop())
]
}
|
<commit_before>"""Stop Bins."""
import util
import errors
import registry
import entities
class StopBin(object):
def __init__(self, prefix):
self.prefix = prefix
self._stops = {}
def stops(self):
return self._stops.values()
def add_stop(self, stop):
key = stop.onestop()
# New stop
if key not in self._stops:
self._stops[key] = stop
else:
self._stops[key].merge(stop)
@classmethod
def from_json(cls, data):
stopbin = cls(prefix=data['prefix'])
for feature in data['features']:
stop = entities.OnestopStop.from_json(feature)
stopbin.add_stop(stop)
return stopbin
def json(self):
return {
'type': 'FeatureCollection',
'properties': {},
'prefix': self.prefix,
'features': [
i.json()
for i in
sorted(self.stops(), key=lambda x:x.onestop())
]
}
<commit_msg>Return added stop in StopBin.add_stop()<commit_after>"""Stop Bins."""
import util
import errors
import registry
import entities
class StopBin(object):
def __init__(self, prefix):
self.prefix = prefix
self._stops = {}
def stops(self):
return self._stops.values()
def add_stop(self, stop):
key = stop.onestop()
# New stop
if key not in self._stops:
self._stops[key] = stop
else:
self._stops[key].merge(stop)
return self._stops[key]
@classmethod
def from_json(cls, data):
stopbin = cls(prefix=data['prefix'])
for feature in data['features']:
stop = entities.OnestopStop.from_json(feature)
stopbin.add_stop(stop)
return stopbin
def json(self):
return {
'type': 'FeatureCollection',
'properties': {},
'prefix': self.prefix,
'features': [
i.json()
for i in
sorted(self.stops(), key=lambda x:x.onestop())
]
}
|
73877a82bf9b690827102d1a932a31af94ab78e9
|
partner_event/models/res_partner.py
|
partner_event/models/res_partner.py
|
# -*- encoding: utf-8 -*-
##############################################################################
# For copyright and license notices, see __openerp__.py file in root directory
##############################################################################
from openerp import models, fields, api
class ResPartner(models.Model):
_inherit = 'res.partner'
registrations = fields.One2many(
string="Event registrations",
comodel_name='event.registration', inverse_name="partner_id")
registration_count = fields.Integer(
string='Event registrations number', compute='_count_registration',
store=True)
attended_registration_count = fields.Integer(
string='Event attended registrations number',
compute='_count_registration', store=True)
@api.one
@api.depends('registrations')
def _count_registration(self):
self.registration_count = len(self.registrations)
self.attended_registration_count = len(self.registrations.filtered(
lambda x: x.state == 'done'))
|
# -*- encoding: utf-8 -*-
##############################################################################
# For copyright and license notices, see __openerp__.py file in root directory
##############################################################################
from openerp import models, fields, api
class ResPartner(models.Model):
_inherit = 'res.partner'
registrations = fields.One2many(
string="Event registrations",
comodel_name='event.registration', inverse_name="partner_id")
registration_count = fields.Integer(
string='Event registrations number', compute='_count_registration',
store=True)
attended_registration_count = fields.Integer(
string='Event attended registrations number',
compute='_count_attended_registration', store=True)
@api.one
@api.depends('registrations')
def _count_registration(self):
self.registration_count = len(self.registrations)
@api.one
@api.depends('registrations.state')
def _count_attended_registration(self):
self.attended_registration_count = len(self.registrations.filtered(
lambda x: x.state == 'done'))
|
Revert last commit and tiggers _count_attended_registration method when one registrations.state changes
|
Revert last commit and tiggers _count_attended_registration method when one registrations.state changes
|
Python
|
agpl-3.0
|
open-synergy/event,open-synergy/event,Endika/event,Antiun/event
|
# -*- encoding: utf-8 -*-
##############################################################################
# For copyright and license notices, see __openerp__.py file in root directory
##############################################################################
from openerp import models, fields, api
class ResPartner(models.Model):
_inherit = 'res.partner'
registrations = fields.One2many(
string="Event registrations",
comodel_name='event.registration', inverse_name="partner_id")
registration_count = fields.Integer(
string='Event registrations number', compute='_count_registration',
store=True)
attended_registration_count = fields.Integer(
string='Event attended registrations number',
compute='_count_registration', store=True)
@api.one
@api.depends('registrations')
def _count_registration(self):
self.registration_count = len(self.registrations)
self.attended_registration_count = len(self.registrations.filtered(
lambda x: x.state == 'done'))
Revert last commit and tiggers _count_attended_registration method when one registrations.state changes
|
# -*- encoding: utf-8 -*-
##############################################################################
# For copyright and license notices, see __openerp__.py file in root directory
##############################################################################
from openerp import models, fields, api
class ResPartner(models.Model):
_inherit = 'res.partner'
registrations = fields.One2many(
string="Event registrations",
comodel_name='event.registration', inverse_name="partner_id")
registration_count = fields.Integer(
string='Event registrations number', compute='_count_registration',
store=True)
attended_registration_count = fields.Integer(
string='Event attended registrations number',
compute='_count_attended_registration', store=True)
@api.one
@api.depends('registrations')
def _count_registration(self):
self.registration_count = len(self.registrations)
@api.one
@api.depends('registrations.state')
def _count_attended_registration(self):
self.attended_registration_count = len(self.registrations.filtered(
lambda x: x.state == 'done'))
|
<commit_before># -*- encoding: utf-8 -*-
##############################################################################
# For copyright and license notices, see __openerp__.py file in root directory
##############################################################################
from openerp import models, fields, api
class ResPartner(models.Model):
_inherit = 'res.partner'
registrations = fields.One2many(
string="Event registrations",
comodel_name='event.registration', inverse_name="partner_id")
registration_count = fields.Integer(
string='Event registrations number', compute='_count_registration',
store=True)
attended_registration_count = fields.Integer(
string='Event attended registrations number',
compute='_count_registration', store=True)
@api.one
@api.depends('registrations')
def _count_registration(self):
self.registration_count = len(self.registrations)
self.attended_registration_count = len(self.registrations.filtered(
lambda x: x.state == 'done'))
<commit_msg>Revert last commit and tiggers _count_attended_registration method when one registrations.state changes<commit_after>
|
# -*- encoding: utf-8 -*-
##############################################################################
# For copyright and license notices, see __openerp__.py file in root directory
##############################################################################
from openerp import models, fields, api
class ResPartner(models.Model):
_inherit = 'res.partner'
registrations = fields.One2many(
string="Event registrations",
comodel_name='event.registration', inverse_name="partner_id")
registration_count = fields.Integer(
string='Event registrations number', compute='_count_registration',
store=True)
attended_registration_count = fields.Integer(
string='Event attended registrations number',
compute='_count_attended_registration', store=True)
@api.one
@api.depends('registrations')
def _count_registration(self):
self.registration_count = len(self.registrations)
@api.one
@api.depends('registrations.state')
def _count_attended_registration(self):
self.attended_registration_count = len(self.registrations.filtered(
lambda x: x.state == 'done'))
|
# -*- encoding: utf-8 -*-
##############################################################################
# For copyright and license notices, see __openerp__.py file in root directory
##############################################################################
from openerp import models, fields, api
class ResPartner(models.Model):
_inherit = 'res.partner'
registrations = fields.One2many(
string="Event registrations",
comodel_name='event.registration', inverse_name="partner_id")
registration_count = fields.Integer(
string='Event registrations number', compute='_count_registration',
store=True)
attended_registration_count = fields.Integer(
string='Event attended registrations number',
compute='_count_registration', store=True)
@api.one
@api.depends('registrations')
def _count_registration(self):
self.registration_count = len(self.registrations)
self.attended_registration_count = len(self.registrations.filtered(
lambda x: x.state == 'done'))
Revert last commit and tiggers _count_attended_registration method when one registrations.state changes# -*- encoding: utf-8 -*-
##############################################################################
# For copyright and license notices, see __openerp__.py file in root directory
##############################################################################
from openerp import models, fields, api
class ResPartner(models.Model):
_inherit = 'res.partner'
registrations = fields.One2many(
string="Event registrations",
comodel_name='event.registration', inverse_name="partner_id")
registration_count = fields.Integer(
string='Event registrations number', compute='_count_registration',
store=True)
attended_registration_count = fields.Integer(
string='Event attended registrations number',
compute='_count_attended_registration', store=True)
@api.one
@api.depends('registrations')
def _count_registration(self):
self.registration_count = len(self.registrations)
@api.one
@api.depends('registrations.state')
def _count_attended_registration(self):
self.attended_registration_count = len(self.registrations.filtered(
lambda x: x.state == 'done'))
|
<commit_before># -*- encoding: utf-8 -*-
##############################################################################
# For copyright and license notices, see __openerp__.py file in root directory
##############################################################################
from openerp import models, fields, api
class ResPartner(models.Model):
_inherit = 'res.partner'
registrations = fields.One2many(
string="Event registrations",
comodel_name='event.registration', inverse_name="partner_id")
registration_count = fields.Integer(
string='Event registrations number', compute='_count_registration',
store=True)
attended_registration_count = fields.Integer(
string='Event attended registrations number',
compute='_count_registration', store=True)
@api.one
@api.depends('registrations')
def _count_registration(self):
self.registration_count = len(self.registrations)
self.attended_registration_count = len(self.registrations.filtered(
lambda x: x.state == 'done'))
<commit_msg>Revert last commit and tiggers _count_attended_registration method when one registrations.state changes<commit_after># -*- encoding: utf-8 -*-
##############################################################################
# For copyright and license notices, see __openerp__.py file in root directory
##############################################################################
from openerp import models, fields, api
class ResPartner(models.Model):
_inherit = 'res.partner'
registrations = fields.One2many(
string="Event registrations",
comodel_name='event.registration', inverse_name="partner_id")
registration_count = fields.Integer(
string='Event registrations number', compute='_count_registration',
store=True)
attended_registration_count = fields.Integer(
string='Event attended registrations number',
compute='_count_attended_registration', store=True)
@api.one
@api.depends('registrations')
def _count_registration(self):
self.registration_count = len(self.registrations)
@api.one
@api.depends('registrations.state')
def _count_attended_registration(self):
self.attended_registration_count = len(self.registrations.filtered(
lambda x: x.state == 'done'))
|
e5d3f0f0295cb5943f7926e49da42565a7905c85
|
dummy_celery_worker.py
|
dummy_celery_worker.py
|
import os
from celery import Celery
app = Celery('dummy_tasks', broker=os.environ['TEST_HADES_BROKER_URI'],
backend=os.environ['TEST_HADES_RESULT_BACKEND_URI'])
@app.task
def get_port_auth_attempts(nasipaddress, nasportid):
return ["Success!", "No success! :-(",
"Gotten: {}/{}".format(nasipaddress, nasportid)]
|
import os
from datetime import datetime
from time import sleep
from celery import Celery
app = Celery('dummy_tasks', broker=os.environ['TEST_HADES_BROKER_URI'],
backend=os.environ['TEST_HADES_RESULT_BACKEND_URI'])
@app.task
def get_port_auth_attempts(nasipaddress, nasportid, limit=100):
if nasportid == 'magic_sleep':
# sleep for 10 seconds, which is longer than the default
sleep(10)
return []
if nasipaddress != '141.30.223.206' or nasportid != 'C6':
return []
return [
# (packettype, replymessage, username, auth_date, egress_vlan)
# TODO: What VLAN should there be on Auth-Reject? In any
# case, it will be unused.
("Auth-Reject", "", "00:de:ad:be:ef:00", datetime(2017, 4, 20, 18, 25), None),
("Auth-Access", "Wu5_untagged", "00:de:ad:be:ef:00", datetime(2017, 4, 20, 18, 20), 15),
("Auth-Access", "unknown", "00:de:ad:be:ef:01", datetime(2017, 4, 20, 18, 5), 1001),
("Auth-Access", "traffic", "00:de:ad:be:ef:00", datetime(2017, 4, 20, 18, 0), 1001),
][:limit]
|
Implement correct function signature in dummy API and sleep trigger
|
Implement correct function signature in dummy API and sleep trigger
|
Python
|
apache-2.0
|
lukasjuhrich/pycroft,lukasjuhrich/pycroft,agdsn/pycroft,agdsn/pycroft,agdsn/pycroft,agdsn/pycroft,lukasjuhrich/pycroft,agdsn/pycroft,lukasjuhrich/pycroft
|
import os
from celery import Celery
app = Celery('dummy_tasks', broker=os.environ['TEST_HADES_BROKER_URI'],
backend=os.environ['TEST_HADES_RESULT_BACKEND_URI'])
@app.task
def get_port_auth_attempts(nasipaddress, nasportid):
return ["Success!", "No success! :-(",
"Gotten: {}/{}".format(nasipaddress, nasportid)]
Implement correct function signature in dummy API and sleep trigger
|
import os
from datetime import datetime
from time import sleep
from celery import Celery
app = Celery('dummy_tasks', broker=os.environ['TEST_HADES_BROKER_URI'],
backend=os.environ['TEST_HADES_RESULT_BACKEND_URI'])
@app.task
def get_port_auth_attempts(nasipaddress, nasportid, limit=100):
if nasportid == 'magic_sleep':
# sleep for 10 seconds, which is longer than the default
sleep(10)
return []
if nasipaddress != '141.30.223.206' or nasportid != 'C6':
return []
return [
# (packettype, replymessage, username, auth_date, egress_vlan)
# TODO: What VLAN should there be on Auth-Reject? In any
# case, it will be unused.
("Auth-Reject", "", "00:de:ad:be:ef:00", datetime(2017, 4, 20, 18, 25), None),
("Auth-Access", "Wu5_untagged", "00:de:ad:be:ef:00", datetime(2017, 4, 20, 18, 20), 15),
("Auth-Access", "unknown", "00:de:ad:be:ef:01", datetime(2017, 4, 20, 18, 5), 1001),
("Auth-Access", "traffic", "00:de:ad:be:ef:00", datetime(2017, 4, 20, 18, 0), 1001),
][:limit]
|
<commit_before>import os
from celery import Celery
app = Celery('dummy_tasks', broker=os.environ['TEST_HADES_BROKER_URI'],
backend=os.environ['TEST_HADES_RESULT_BACKEND_URI'])
@app.task
def get_port_auth_attempts(nasipaddress, nasportid):
return ["Success!", "No success! :-(",
"Gotten: {}/{}".format(nasipaddress, nasportid)]
<commit_msg>Implement correct function signature in dummy API and sleep trigger<commit_after>
|
import os
from datetime import datetime
from time import sleep
from celery import Celery
app = Celery('dummy_tasks', broker=os.environ['TEST_HADES_BROKER_URI'],
backend=os.environ['TEST_HADES_RESULT_BACKEND_URI'])
@app.task
def get_port_auth_attempts(nasipaddress, nasportid, limit=100):
if nasportid == 'magic_sleep':
# sleep for 10 seconds, which is longer than the default
sleep(10)
return []
if nasipaddress != '141.30.223.206' or nasportid != 'C6':
return []
return [
# (packettype, replymessage, username, auth_date, egress_vlan)
# TODO: What VLAN should there be on Auth-Reject? In any
# case, it will be unused.
("Auth-Reject", "", "00:de:ad:be:ef:00", datetime(2017, 4, 20, 18, 25), None),
("Auth-Access", "Wu5_untagged", "00:de:ad:be:ef:00", datetime(2017, 4, 20, 18, 20), 15),
("Auth-Access", "unknown", "00:de:ad:be:ef:01", datetime(2017, 4, 20, 18, 5), 1001),
("Auth-Access", "traffic", "00:de:ad:be:ef:00", datetime(2017, 4, 20, 18, 0), 1001),
][:limit]
|
import os
from celery import Celery
app = Celery('dummy_tasks', broker=os.environ['TEST_HADES_BROKER_URI'],
backend=os.environ['TEST_HADES_RESULT_BACKEND_URI'])
@app.task
def get_port_auth_attempts(nasipaddress, nasportid):
return ["Success!", "No success! :-(",
"Gotten: {}/{}".format(nasipaddress, nasportid)]
Implement correct function signature in dummy API and sleep triggerimport os
from datetime import datetime
from time import sleep
from celery import Celery
app = Celery('dummy_tasks', broker=os.environ['TEST_HADES_BROKER_URI'],
backend=os.environ['TEST_HADES_RESULT_BACKEND_URI'])
@app.task
def get_port_auth_attempts(nasipaddress, nasportid, limit=100):
if nasportid == 'magic_sleep':
# sleep for 10 seconds, which is longer than the default
sleep(10)
return []
if nasipaddress != '141.30.223.206' or nasportid != 'C6':
return []
return [
# (packettype, replymessage, username, auth_date, egress_vlan)
# TODO: What VLAN should there be on Auth-Reject? In any
# case, it will be unused.
("Auth-Reject", "", "00:de:ad:be:ef:00", datetime(2017, 4, 20, 18, 25), None),
("Auth-Access", "Wu5_untagged", "00:de:ad:be:ef:00", datetime(2017, 4, 20, 18, 20), 15),
("Auth-Access", "unknown", "00:de:ad:be:ef:01", datetime(2017, 4, 20, 18, 5), 1001),
("Auth-Access", "traffic", "00:de:ad:be:ef:00", datetime(2017, 4, 20, 18, 0), 1001),
][:limit]
|
<commit_before>import os
from celery import Celery
app = Celery('dummy_tasks', broker=os.environ['TEST_HADES_BROKER_URI'],
backend=os.environ['TEST_HADES_RESULT_BACKEND_URI'])
@app.task
def get_port_auth_attempts(nasipaddress, nasportid):
return ["Success!", "No success! :-(",
"Gotten: {}/{}".format(nasipaddress, nasportid)]
<commit_msg>Implement correct function signature in dummy API and sleep trigger<commit_after>import os
from datetime import datetime
from time import sleep
from celery import Celery
app = Celery('dummy_tasks', broker=os.environ['TEST_HADES_BROKER_URI'],
backend=os.environ['TEST_HADES_RESULT_BACKEND_URI'])
@app.task
def get_port_auth_attempts(nasipaddress, nasportid, limit=100):
if nasportid == 'magic_sleep':
# sleep for 10 seconds, which is longer than the default
sleep(10)
return []
if nasipaddress != '141.30.223.206' or nasportid != 'C6':
return []
return [
# (packettype, replymessage, username, auth_date, egress_vlan)
# TODO: What VLAN should there be on Auth-Reject? In any
# case, it will be unused.
("Auth-Reject", "", "00:de:ad:be:ef:00", datetime(2017, 4, 20, 18, 25), None),
("Auth-Access", "Wu5_untagged", "00:de:ad:be:ef:00", datetime(2017, 4, 20, 18, 20), 15),
("Auth-Access", "unknown", "00:de:ad:be:ef:01", datetime(2017, 4, 20, 18, 5), 1001),
("Auth-Access", "traffic", "00:de:ad:be:ef:00", datetime(2017, 4, 20, 18, 0), 1001),
][:limit]
|
ca5851c681452e20a07434b74481860722077bb0
|
server/setup.py
|
server/setup.py
|
import os
from setuptools import setup, find_packages
here = os.path.abspath(os.path.dirname(__file__))
README = open(os.path.join(here, 'README.txt')).read()
version = '0.1'
requires = ['pyramid', 'pyramid_debugtoolbar']
if __name__ == '__main__':
setup(name='pings',
version=version,
description='pings',
long_description=README,
classifiers=[
"Programming Language :: Python",
"Framework :: Pylons",
"Topic :: Internet :: WWW/HTTP",
"Topic :: Internet :: WWW/HTTP :: WSGI :: Application",
],
author='Christian Hudon',
author_email='chrish@pianocktail.org',
url='https://github.com/lisa-lab/pings',
keywords='web pyramid pylons',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=requires,
tests_require=requires,
test_suite="pings",
scripts = ['leaderboards_server', 'storage_server'],
entry_points = """\
[paste.app_factory]
main = pings.web_server:main
""",
paster_plugins=['pyramid'],
)
|
import os
from setuptools import setup, find_packages
here = os.path.abspath(os.path.dirname(__file__))
README = open(os.path.join(here, 'README.txt')).read()
version = '0.1'
# Use requirements.txt for all requirements, at least for now.
requires = []
if __name__ == '__main__':
setup(name='pings',
version=version,
description='pings',
long_description=README,
classifiers=[
"Programming Language :: Python",
"Framework :: Pylons",
"Topic :: Internet :: WWW/HTTP",
"Topic :: Internet :: WWW/HTTP :: WSGI :: Application",
],
author='Christian Hudon',
author_email='chrish@pianocktail.org',
url='https://github.com/lisa-lab/pings',
keywords='web pyramid pylons',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=requires,
tests_require=requires,
test_suite="pings",
scripts = ['leaderboards_server', 'storage_server'],
entry_points = """\
[paste.app_factory]
main = pings.web_server:main
""",
paster_plugins=['pyramid'],
)
|
Use requirements.txt for all requirements, at least for now.
|
Use requirements.txt for all requirements, at least for now.
|
Python
|
bsd-3-clause
|
lisa-lab/pings,lisa-lab/pings,lisa-lab/pings,lisa-lab/pings
|
import os
from setuptools import setup, find_packages
here = os.path.abspath(os.path.dirname(__file__))
README = open(os.path.join(here, 'README.txt')).read()
version = '0.1'
requires = ['pyramid', 'pyramid_debugtoolbar']
if __name__ == '__main__':
setup(name='pings',
version=version,
description='pings',
long_description=README,
classifiers=[
"Programming Language :: Python",
"Framework :: Pylons",
"Topic :: Internet :: WWW/HTTP",
"Topic :: Internet :: WWW/HTTP :: WSGI :: Application",
],
author='Christian Hudon',
author_email='chrish@pianocktail.org',
url='https://github.com/lisa-lab/pings',
keywords='web pyramid pylons',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=requires,
tests_require=requires,
test_suite="pings",
scripts = ['leaderboards_server', 'storage_server'],
entry_points = """\
[paste.app_factory]
main = pings.web_server:main
""",
paster_plugins=['pyramid'],
)
Use requirements.txt for all requirements, at least for now.
|
import os
from setuptools import setup, find_packages
here = os.path.abspath(os.path.dirname(__file__))
README = open(os.path.join(here, 'README.txt')).read()
version = '0.1'
# Use requirements.txt for all requirements, at least for now.
requires = []
if __name__ == '__main__':
setup(name='pings',
version=version,
description='pings',
long_description=README,
classifiers=[
"Programming Language :: Python",
"Framework :: Pylons",
"Topic :: Internet :: WWW/HTTP",
"Topic :: Internet :: WWW/HTTP :: WSGI :: Application",
],
author='Christian Hudon',
author_email='chrish@pianocktail.org',
url='https://github.com/lisa-lab/pings',
keywords='web pyramid pylons',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=requires,
tests_require=requires,
test_suite="pings",
scripts = ['leaderboards_server', 'storage_server'],
entry_points = """\
[paste.app_factory]
main = pings.web_server:main
""",
paster_plugins=['pyramid'],
)
|
<commit_before>import os
from setuptools import setup, find_packages
here = os.path.abspath(os.path.dirname(__file__))
README = open(os.path.join(here, 'README.txt')).read()
version = '0.1'
requires = ['pyramid', 'pyramid_debugtoolbar']
if __name__ == '__main__':
setup(name='pings',
version=version,
description='pings',
long_description=README,
classifiers=[
"Programming Language :: Python",
"Framework :: Pylons",
"Topic :: Internet :: WWW/HTTP",
"Topic :: Internet :: WWW/HTTP :: WSGI :: Application",
],
author='Christian Hudon',
author_email='chrish@pianocktail.org',
url='https://github.com/lisa-lab/pings',
keywords='web pyramid pylons',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=requires,
tests_require=requires,
test_suite="pings",
scripts = ['leaderboards_server', 'storage_server'],
entry_points = """\
[paste.app_factory]
main = pings.web_server:main
""",
paster_plugins=['pyramid'],
)
<commit_msg>Use requirements.txt for all requirements, at least for now.<commit_after>
|
import os
from setuptools import setup, find_packages
here = os.path.abspath(os.path.dirname(__file__))
README = open(os.path.join(here, 'README.txt')).read()
version = '0.1'
# Use requirements.txt for all requirements, at least for now.
requires = []
if __name__ == '__main__':
setup(name='pings',
version=version,
description='pings',
long_description=README,
classifiers=[
"Programming Language :: Python",
"Framework :: Pylons",
"Topic :: Internet :: WWW/HTTP",
"Topic :: Internet :: WWW/HTTP :: WSGI :: Application",
],
author='Christian Hudon',
author_email='chrish@pianocktail.org',
url='https://github.com/lisa-lab/pings',
keywords='web pyramid pylons',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=requires,
tests_require=requires,
test_suite="pings",
scripts = ['leaderboards_server', 'storage_server'],
entry_points = """\
[paste.app_factory]
main = pings.web_server:main
""",
paster_plugins=['pyramid'],
)
|
import os
from setuptools import setup, find_packages
here = os.path.abspath(os.path.dirname(__file__))
README = open(os.path.join(here, 'README.txt')).read()
version = '0.1'
requires = ['pyramid', 'pyramid_debugtoolbar']
if __name__ == '__main__':
setup(name='pings',
version=version,
description='pings',
long_description=README,
classifiers=[
"Programming Language :: Python",
"Framework :: Pylons",
"Topic :: Internet :: WWW/HTTP",
"Topic :: Internet :: WWW/HTTP :: WSGI :: Application",
],
author='Christian Hudon',
author_email='chrish@pianocktail.org',
url='https://github.com/lisa-lab/pings',
keywords='web pyramid pylons',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=requires,
tests_require=requires,
test_suite="pings",
scripts = ['leaderboards_server', 'storage_server'],
entry_points = """\
[paste.app_factory]
main = pings.web_server:main
""",
paster_plugins=['pyramid'],
)
Use requirements.txt for all requirements, at least for now.import os
from setuptools import setup, find_packages
here = os.path.abspath(os.path.dirname(__file__))
README = open(os.path.join(here, 'README.txt')).read()
version = '0.1'
# Use requirements.txt for all requirements, at least for now.
requires = []
if __name__ == '__main__':
setup(name='pings',
version=version,
description='pings',
long_description=README,
classifiers=[
"Programming Language :: Python",
"Framework :: Pylons",
"Topic :: Internet :: WWW/HTTP",
"Topic :: Internet :: WWW/HTTP :: WSGI :: Application",
],
author='Christian Hudon',
author_email='chrish@pianocktail.org',
url='https://github.com/lisa-lab/pings',
keywords='web pyramid pylons',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=requires,
tests_require=requires,
test_suite="pings",
scripts = ['leaderboards_server', 'storage_server'],
entry_points = """\
[paste.app_factory]
main = pings.web_server:main
""",
paster_plugins=['pyramid'],
)
|
<commit_before>import os
from setuptools import setup, find_packages
here = os.path.abspath(os.path.dirname(__file__))
README = open(os.path.join(here, 'README.txt')).read()
version = '0.1'
requires = ['pyramid', 'pyramid_debugtoolbar']
if __name__ == '__main__':
setup(name='pings',
version=version,
description='pings',
long_description=README,
classifiers=[
"Programming Language :: Python",
"Framework :: Pylons",
"Topic :: Internet :: WWW/HTTP",
"Topic :: Internet :: WWW/HTTP :: WSGI :: Application",
],
author='Christian Hudon',
author_email='chrish@pianocktail.org',
url='https://github.com/lisa-lab/pings',
keywords='web pyramid pylons',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=requires,
tests_require=requires,
test_suite="pings",
scripts = ['leaderboards_server', 'storage_server'],
entry_points = """\
[paste.app_factory]
main = pings.web_server:main
""",
paster_plugins=['pyramid'],
)
<commit_msg>Use requirements.txt for all requirements, at least for now.<commit_after>import os
from setuptools import setup, find_packages
here = os.path.abspath(os.path.dirname(__file__))
README = open(os.path.join(here, 'README.txt')).read()
version = '0.1'
# Use requirements.txt for all requirements, at least for now.
requires = []
if __name__ == '__main__':
setup(name='pings',
version=version,
description='pings',
long_description=README,
classifiers=[
"Programming Language :: Python",
"Framework :: Pylons",
"Topic :: Internet :: WWW/HTTP",
"Topic :: Internet :: WWW/HTTP :: WSGI :: Application",
],
author='Christian Hudon',
author_email='chrish@pianocktail.org',
url='https://github.com/lisa-lab/pings',
keywords='web pyramid pylons',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=requires,
tests_require=requires,
test_suite="pings",
scripts = ['leaderboards_server', 'storage_server'],
entry_points = """\
[paste.app_factory]
main = pings.web_server:main
""",
paster_plugins=['pyramid'],
)
|
c6e130682712e8534e773036ba3d87c09b91ff1c
|
knowledge_repo/postprocessors/format_checks.py
|
knowledge_repo/postprocessors/format_checks.py
|
from ..constants import FORMAT_CHECKS
from ..post import HEADER_OPTIONAL_FIELD_TYPES, HEADER_REQUIRED_FIELD_TYPES
from ..postprocessor import KnowledgePostProcessor
class FormatChecks(KnowledgePostProcessor):
_registry_keys = [FORMAT_CHECKS]
def process(self, kp):
headers = kp.headers
for field, typ, input in HEADER_REQUIRED_FIELD_TYPES:
assert field in headers, "Required field `{}` missing from headers.".format(
field)
assert isinstance(headers[field], typ), "Value for field `{}` is of type {}, and needs to be of type {}.".format(
field, type(headers[field]), typ)
for field, typ, input in HEADER_OPTIONAL_FIELD_TYPES:
if field in headers:
assert isinstance(headers[field], typ), "Value for field `{}` is of type {}, and needs to be of type {}.".format(
field, type(headers[field]), typ)
|
from ..constants import FORMAT_CHECKS
from ..post import HEADER_OPTIONAL_FIELD_TYPES, HEADER_REQUIRED_FIELD_TYPES
from ..postprocessor import KnowledgePostProcessor
class FormatChecks(KnowledgePostProcessor):
_registry_keys = [FORMAT_CHECKS]
def process(self, kp):
headers = kp.headers
for field, typ, input in HEADER_REQUIRED_FIELD_TYPES:
assert field in headers, \
"Required field `{field}` missing from headers."
assert isinstance(headers[field], typ), \
f"Value for field `{field}` is of type " + \
f"{type(headers[field])}, and needs to be of type {typ}."
for field, typ, input in HEADER_OPTIONAL_FIELD_TYPES:
if field in headers:
assert isinstance(headers[field], typ), \
f"Value for field `{field}` is of type " + \
f"{type(headers[field])}, and needs to be of type {typ}."
|
Fix lint issues related to long lines
|
Fix lint issues related to long lines
|
Python
|
apache-2.0
|
airbnb/knowledge-repo,airbnb/knowledge-repo,airbnb/knowledge-repo,airbnb/knowledge-repo,airbnb/knowledge-repo
|
from ..constants import FORMAT_CHECKS
from ..post import HEADER_OPTIONAL_FIELD_TYPES, HEADER_REQUIRED_FIELD_TYPES
from ..postprocessor import KnowledgePostProcessor
class FormatChecks(KnowledgePostProcessor):
_registry_keys = [FORMAT_CHECKS]
def process(self, kp):
headers = kp.headers
for field, typ, input in HEADER_REQUIRED_FIELD_TYPES:
assert field in headers, "Required field `{}` missing from headers.".format(
field)
assert isinstance(headers[field], typ), "Value for field `{}` is of type {}, and needs to be of type {}.".format(
field, type(headers[field]), typ)
for field, typ, input in HEADER_OPTIONAL_FIELD_TYPES:
if field in headers:
assert isinstance(headers[field], typ), "Value for field `{}` is of type {}, and needs to be of type {}.".format(
field, type(headers[field]), typ)
Fix lint issues related to long lines
|
from ..constants import FORMAT_CHECKS
from ..post import HEADER_OPTIONAL_FIELD_TYPES, HEADER_REQUIRED_FIELD_TYPES
from ..postprocessor import KnowledgePostProcessor
class FormatChecks(KnowledgePostProcessor):
_registry_keys = [FORMAT_CHECKS]
def process(self, kp):
headers = kp.headers
for field, typ, input in HEADER_REQUIRED_FIELD_TYPES:
assert field in headers, \
"Required field `{field}` missing from headers."
assert isinstance(headers[field], typ), \
f"Value for field `{field}` is of type " + \
f"{type(headers[field])}, and needs to be of type {typ}."
for field, typ, input in HEADER_OPTIONAL_FIELD_TYPES:
if field in headers:
assert isinstance(headers[field], typ), \
f"Value for field `{field}` is of type " + \
f"{type(headers[field])}, and needs to be of type {typ}."
|
<commit_before>from ..constants import FORMAT_CHECKS
from ..post import HEADER_OPTIONAL_FIELD_TYPES, HEADER_REQUIRED_FIELD_TYPES
from ..postprocessor import KnowledgePostProcessor
class FormatChecks(KnowledgePostProcessor):
_registry_keys = [FORMAT_CHECKS]
def process(self, kp):
headers = kp.headers
for field, typ, input in HEADER_REQUIRED_FIELD_TYPES:
assert field in headers, "Required field `{}` missing from headers.".format(
field)
assert isinstance(headers[field], typ), "Value for field `{}` is of type {}, and needs to be of type {}.".format(
field, type(headers[field]), typ)
for field, typ, input in HEADER_OPTIONAL_FIELD_TYPES:
if field in headers:
assert isinstance(headers[field], typ), "Value for field `{}` is of type {}, and needs to be of type {}.".format(
field, type(headers[field]), typ)
<commit_msg>Fix lint issues related to long lines<commit_after>
|
from ..constants import FORMAT_CHECKS
from ..post import HEADER_OPTIONAL_FIELD_TYPES, HEADER_REQUIRED_FIELD_TYPES
from ..postprocessor import KnowledgePostProcessor
class FormatChecks(KnowledgePostProcessor):
_registry_keys = [FORMAT_CHECKS]
def process(self, kp):
headers = kp.headers
for field, typ, input in HEADER_REQUIRED_FIELD_TYPES:
assert field in headers, \
"Required field `{field}` missing from headers."
assert isinstance(headers[field], typ), \
f"Value for field `{field}` is of type " + \
f"{type(headers[field])}, and needs to be of type {typ}."
for field, typ, input in HEADER_OPTIONAL_FIELD_TYPES:
if field in headers:
assert isinstance(headers[field], typ), \
f"Value for field `{field}` is of type " + \
f"{type(headers[field])}, and needs to be of type {typ}."
|
from ..constants import FORMAT_CHECKS
from ..post import HEADER_OPTIONAL_FIELD_TYPES, HEADER_REQUIRED_FIELD_TYPES
from ..postprocessor import KnowledgePostProcessor
class FormatChecks(KnowledgePostProcessor):
_registry_keys = [FORMAT_CHECKS]
def process(self, kp):
headers = kp.headers
for field, typ, input in HEADER_REQUIRED_FIELD_TYPES:
assert field in headers, "Required field `{}` missing from headers.".format(
field)
assert isinstance(headers[field], typ), "Value for field `{}` is of type {}, and needs to be of type {}.".format(
field, type(headers[field]), typ)
for field, typ, input in HEADER_OPTIONAL_FIELD_TYPES:
if field in headers:
assert isinstance(headers[field], typ), "Value for field `{}` is of type {}, and needs to be of type {}.".format(
field, type(headers[field]), typ)
Fix lint issues related to long linesfrom ..constants import FORMAT_CHECKS
from ..post import HEADER_OPTIONAL_FIELD_TYPES, HEADER_REQUIRED_FIELD_TYPES
from ..postprocessor import KnowledgePostProcessor
class FormatChecks(KnowledgePostProcessor):
_registry_keys = [FORMAT_CHECKS]
def process(self, kp):
headers = kp.headers
for field, typ, input in HEADER_REQUIRED_FIELD_TYPES:
assert field in headers, \
"Required field `{field}` missing from headers."
assert isinstance(headers[field], typ), \
f"Value for field `{field}` is of type " + \
f"{type(headers[field])}, and needs to be of type {typ}."
for field, typ, input in HEADER_OPTIONAL_FIELD_TYPES:
if field in headers:
assert isinstance(headers[field], typ), \
f"Value for field `{field}` is of type " + \
f"{type(headers[field])}, and needs to be of type {typ}."
|
<commit_before>from ..constants import FORMAT_CHECKS
from ..post import HEADER_OPTIONAL_FIELD_TYPES, HEADER_REQUIRED_FIELD_TYPES
from ..postprocessor import KnowledgePostProcessor
class FormatChecks(KnowledgePostProcessor):
_registry_keys = [FORMAT_CHECKS]
def process(self, kp):
headers = kp.headers
for field, typ, input in HEADER_REQUIRED_FIELD_TYPES:
assert field in headers, "Required field `{}` missing from headers.".format(
field)
assert isinstance(headers[field], typ), "Value for field `{}` is of type {}, and needs to be of type {}.".format(
field, type(headers[field]), typ)
for field, typ, input in HEADER_OPTIONAL_FIELD_TYPES:
if field in headers:
assert isinstance(headers[field], typ), "Value for field `{}` is of type {}, and needs to be of type {}.".format(
field, type(headers[field]), typ)
<commit_msg>Fix lint issues related to long lines<commit_after>from ..constants import FORMAT_CHECKS
from ..post import HEADER_OPTIONAL_FIELD_TYPES, HEADER_REQUIRED_FIELD_TYPES
from ..postprocessor import KnowledgePostProcessor
class FormatChecks(KnowledgePostProcessor):
_registry_keys = [FORMAT_CHECKS]
def process(self, kp):
headers = kp.headers
for field, typ, input in HEADER_REQUIRED_FIELD_TYPES:
assert field in headers, \
"Required field `{field}` missing from headers."
assert isinstance(headers[field], typ), \
f"Value for field `{field}` is of type " + \
f"{type(headers[field])}, and needs to be of type {typ}."
for field, typ, input in HEADER_OPTIONAL_FIELD_TYPES:
if field in headers:
assert isinstance(headers[field], typ), \
f"Value for field `{field}` is of type " + \
f"{type(headers[field])}, and needs to be of type {typ}."
|
bf264d5683c7fcab69e117f235fbe16298ac90b8
|
wal_e/worker/wabs/wabs_deleter.py
|
wal_e/worker/wabs/wabs_deleter.py
|
from wal_e import retries
from wal_e.worker.base import _Deleter
class Deleter(_Deleter):
def __init__(self, wabs_conn, container):
super(Deleter, self).__init__()
self.wabs_conn = wabs_conn
self.container = container
@retries.retry()
def _delete_batch(self, page):
# Azure Blob Service has no concept of mass-delete, so we must nuke
# each blob one-by-one...
for blob in page:
self.wabs_conn.delete_blob(self.container, blob.name)
|
from wal_e import retries
from wal_e import log_help
from wal_e.worker.base import _Deleter
try:
# New class name in the Azure SDK sometime after v1.0.
#
# See
# https://github.com/Azure/azure-sdk-for-python/blob/master/ChangeLog.txt
from azure.common import AzureMissingResourceHttpError
except ImportError:
# Backwards compatbility for older Azure drivers.
from azure import WindowsAzureMissingResourceError \
as AzureMissingResourceHttpError
logger = log_help.WalELogger(__name__)
class Deleter(_Deleter):
def __init__(self, wabs_conn, container):
super(Deleter, self).__init__()
self.wabs_conn = wabs_conn
self.container = container
@retries.retry()
def _delete_batch(self, page):
# Azure Blob Service has no concept of mass-delete, so we must nuke
# each blob one-by-one...
for blob in page:
try:
self.wabs_conn.delete_blob(self.container, blob.name)
except AzureMissingResourceHttpError:
logger.warning(
msg='failed while deleting resource',
detail='Blob {0} does not exist in container {1}.'.format(
blob.name, self.container))
|
Fix infinite retry while deleting missing resource in WABS
|
Fix infinite retry while deleting missing resource in WABS
|
Python
|
bsd-3-clause
|
wal-e/wal-e
|
from wal_e import retries
from wal_e.worker.base import _Deleter
class Deleter(_Deleter):
def __init__(self, wabs_conn, container):
super(Deleter, self).__init__()
self.wabs_conn = wabs_conn
self.container = container
@retries.retry()
def _delete_batch(self, page):
# Azure Blob Service has no concept of mass-delete, so we must nuke
# each blob one-by-one...
for blob in page:
self.wabs_conn.delete_blob(self.container, blob.name)
Fix infinite retry while deleting missing resource in WABS
|
from wal_e import retries
from wal_e import log_help
from wal_e.worker.base import _Deleter
try:
# New class name in the Azure SDK sometime after v1.0.
#
# See
# https://github.com/Azure/azure-sdk-for-python/blob/master/ChangeLog.txt
from azure.common import AzureMissingResourceHttpError
except ImportError:
# Backwards compatbility for older Azure drivers.
from azure import WindowsAzureMissingResourceError \
as AzureMissingResourceHttpError
logger = log_help.WalELogger(__name__)
class Deleter(_Deleter):
def __init__(self, wabs_conn, container):
super(Deleter, self).__init__()
self.wabs_conn = wabs_conn
self.container = container
@retries.retry()
def _delete_batch(self, page):
# Azure Blob Service has no concept of mass-delete, so we must nuke
# each blob one-by-one...
for blob in page:
try:
self.wabs_conn.delete_blob(self.container, blob.name)
except AzureMissingResourceHttpError:
logger.warning(
msg='failed while deleting resource',
detail='Blob {0} does not exist in container {1}.'.format(
blob.name, self.container))
|
<commit_before>from wal_e import retries
from wal_e.worker.base import _Deleter
class Deleter(_Deleter):
def __init__(self, wabs_conn, container):
super(Deleter, self).__init__()
self.wabs_conn = wabs_conn
self.container = container
@retries.retry()
def _delete_batch(self, page):
# Azure Blob Service has no concept of mass-delete, so we must nuke
# each blob one-by-one...
for blob in page:
self.wabs_conn.delete_blob(self.container, blob.name)
<commit_msg>Fix infinite retry while deleting missing resource in WABS<commit_after>
|
from wal_e import retries
from wal_e import log_help
from wal_e.worker.base import _Deleter
try:
# New class name in the Azure SDK sometime after v1.0.
#
# See
# https://github.com/Azure/azure-sdk-for-python/blob/master/ChangeLog.txt
from azure.common import AzureMissingResourceHttpError
except ImportError:
# Backwards compatbility for older Azure drivers.
from azure import WindowsAzureMissingResourceError \
as AzureMissingResourceHttpError
logger = log_help.WalELogger(__name__)
class Deleter(_Deleter):
def __init__(self, wabs_conn, container):
super(Deleter, self).__init__()
self.wabs_conn = wabs_conn
self.container = container
@retries.retry()
def _delete_batch(self, page):
# Azure Blob Service has no concept of mass-delete, so we must nuke
# each blob one-by-one...
for blob in page:
try:
self.wabs_conn.delete_blob(self.container, blob.name)
except AzureMissingResourceHttpError:
logger.warning(
msg='failed while deleting resource',
detail='Blob {0} does not exist in container {1}.'.format(
blob.name, self.container))
|
from wal_e import retries
from wal_e.worker.base import _Deleter
class Deleter(_Deleter):
def __init__(self, wabs_conn, container):
super(Deleter, self).__init__()
self.wabs_conn = wabs_conn
self.container = container
@retries.retry()
def _delete_batch(self, page):
# Azure Blob Service has no concept of mass-delete, so we must nuke
# each blob one-by-one...
for blob in page:
self.wabs_conn.delete_blob(self.container, blob.name)
Fix infinite retry while deleting missing resource in WABSfrom wal_e import retries
from wal_e import log_help
from wal_e.worker.base import _Deleter
try:
# New class name in the Azure SDK sometime after v1.0.
#
# See
# https://github.com/Azure/azure-sdk-for-python/blob/master/ChangeLog.txt
from azure.common import AzureMissingResourceHttpError
except ImportError:
# Backwards compatbility for older Azure drivers.
from azure import WindowsAzureMissingResourceError \
as AzureMissingResourceHttpError
logger = log_help.WalELogger(__name__)
class Deleter(_Deleter):
def __init__(self, wabs_conn, container):
super(Deleter, self).__init__()
self.wabs_conn = wabs_conn
self.container = container
@retries.retry()
def _delete_batch(self, page):
# Azure Blob Service has no concept of mass-delete, so we must nuke
# each blob one-by-one...
for blob in page:
try:
self.wabs_conn.delete_blob(self.container, blob.name)
except AzureMissingResourceHttpError:
logger.warning(
msg='failed while deleting resource',
detail='Blob {0} does not exist in container {1}.'.format(
blob.name, self.container))
|
<commit_before>from wal_e import retries
from wal_e.worker.base import _Deleter
class Deleter(_Deleter):
def __init__(self, wabs_conn, container):
super(Deleter, self).__init__()
self.wabs_conn = wabs_conn
self.container = container
@retries.retry()
def _delete_batch(self, page):
# Azure Blob Service has no concept of mass-delete, so we must nuke
# each blob one-by-one...
for blob in page:
self.wabs_conn.delete_blob(self.container, blob.name)
<commit_msg>Fix infinite retry while deleting missing resource in WABS<commit_after>from wal_e import retries
from wal_e import log_help
from wal_e.worker.base import _Deleter
try:
# New class name in the Azure SDK sometime after v1.0.
#
# See
# https://github.com/Azure/azure-sdk-for-python/blob/master/ChangeLog.txt
from azure.common import AzureMissingResourceHttpError
except ImportError:
# Backwards compatbility for older Azure drivers.
from azure import WindowsAzureMissingResourceError \
as AzureMissingResourceHttpError
logger = log_help.WalELogger(__name__)
class Deleter(_Deleter):
def __init__(self, wabs_conn, container):
super(Deleter, self).__init__()
self.wabs_conn = wabs_conn
self.container = container
@retries.retry()
def _delete_batch(self, page):
# Azure Blob Service has no concept of mass-delete, so we must nuke
# each blob one-by-one...
for blob in page:
try:
self.wabs_conn.delete_blob(self.container, blob.name)
except AzureMissingResourceHttpError:
logger.warning(
msg='failed while deleting resource',
detail='Blob {0} does not exist in container {1}.'.format(
blob.name, self.container))
|
62845279b46d6f4394e05e666fe459a427bdd358
|
enthought/qt/QtCore.py
|
enthought/qt/QtCore.py
|
import os
qt_api = os.environ.get('QT_API', 'pyqt')
if qt_api == 'pyqt':
from PyQt4.QtCore import *
from PyQt4.QtCore import pyqtSignal as Signal
from PyQt4.Qt import QCoreApplication
from PyQt4.Qt import Qt
else:
from PySide.QtCore import *
|
import os
qt_api = os.environ.get('QT_API', 'pyqt')
if qt_api == 'pyqt':
from PyQt4.QtCore import *
from PyQt4.QtCore import pyqtSignal as Signal
from PyQt4.Qt import QCoreApplication
from PyQt4.Qt import Qt
# Emulate PySide version metadata.
__version__ = QT_VERSION_STR
__version_info__ = tuple(map(int, QT_VERSION_STR.split('.')))
else:
from PySide.QtCore import *
|
Add PySide-style version metadata when PyQt4 is present.
|
Add PySide-style version metadata when PyQt4 is present.
|
Python
|
bsd-3-clause
|
burnpanck/traits,burnpanck/traits
|
import os
qt_api = os.environ.get('QT_API', 'pyqt')
if qt_api == 'pyqt':
from PyQt4.QtCore import *
from PyQt4.QtCore import pyqtSignal as Signal
from PyQt4.Qt import QCoreApplication
from PyQt4.Qt import Qt
else:
from PySide.QtCore import *
Add PySide-style version metadata when PyQt4 is present.
|
import os
qt_api = os.environ.get('QT_API', 'pyqt')
if qt_api == 'pyqt':
from PyQt4.QtCore import *
from PyQt4.QtCore import pyqtSignal as Signal
from PyQt4.Qt import QCoreApplication
from PyQt4.Qt import Qt
# Emulate PySide version metadata.
__version__ = QT_VERSION_STR
__version_info__ = tuple(map(int, QT_VERSION_STR.split('.')))
else:
from PySide.QtCore import *
|
<commit_before>import os
qt_api = os.environ.get('QT_API', 'pyqt')
if qt_api == 'pyqt':
from PyQt4.QtCore import *
from PyQt4.QtCore import pyqtSignal as Signal
from PyQt4.Qt import QCoreApplication
from PyQt4.Qt import Qt
else:
from PySide.QtCore import *
<commit_msg>Add PySide-style version metadata when PyQt4 is present.<commit_after>
|
import os
qt_api = os.environ.get('QT_API', 'pyqt')
if qt_api == 'pyqt':
from PyQt4.QtCore import *
from PyQt4.QtCore import pyqtSignal as Signal
from PyQt4.Qt import QCoreApplication
from PyQt4.Qt import Qt
# Emulate PySide version metadata.
__version__ = QT_VERSION_STR
__version_info__ = tuple(map(int, QT_VERSION_STR.split('.')))
else:
from PySide.QtCore import *
|
import os
qt_api = os.environ.get('QT_API', 'pyqt')
if qt_api == 'pyqt':
from PyQt4.QtCore import *
from PyQt4.QtCore import pyqtSignal as Signal
from PyQt4.Qt import QCoreApplication
from PyQt4.Qt import Qt
else:
from PySide.QtCore import *
Add PySide-style version metadata when PyQt4 is present.import os
qt_api = os.environ.get('QT_API', 'pyqt')
if qt_api == 'pyqt':
from PyQt4.QtCore import *
from PyQt4.QtCore import pyqtSignal as Signal
from PyQt4.Qt import QCoreApplication
from PyQt4.Qt import Qt
# Emulate PySide version metadata.
__version__ = QT_VERSION_STR
__version_info__ = tuple(map(int, QT_VERSION_STR.split('.')))
else:
from PySide.QtCore import *
|
<commit_before>import os
qt_api = os.environ.get('QT_API', 'pyqt')
if qt_api == 'pyqt':
from PyQt4.QtCore import *
from PyQt4.QtCore import pyqtSignal as Signal
from PyQt4.Qt import QCoreApplication
from PyQt4.Qt import Qt
else:
from PySide.QtCore import *
<commit_msg>Add PySide-style version metadata when PyQt4 is present.<commit_after>import os
qt_api = os.environ.get('QT_API', 'pyqt')
if qt_api == 'pyqt':
from PyQt4.QtCore import *
from PyQt4.QtCore import pyqtSignal as Signal
from PyQt4.Qt import QCoreApplication
from PyQt4.Qt import Qt
# Emulate PySide version metadata.
__version__ = QT_VERSION_STR
__version_info__ = tuple(map(int, QT_VERSION_STR.split('.')))
else:
from PySide.QtCore import *
|
cc87cf3967e14274b7819f5424b80bd7e491f0ce
|
alg_kruskal_minimum_spanning_tree.py
|
alg_kruskal_minimum_spanning_tree.py
|
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
def kruskal():
"""Kruskal's algorithm for minimum spanning tree in weighted graph.
Time complexity for graph G(V, E): TBD.
"""
pass
def main():
pass
if __name__ == '__main__':
main()
|
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
def kruskal():
"""Kruskal's algorithm for minimum spanning tree in weighted graph.
Time complexity for graph G(V, E): TBD.
"""
pass
def main():
w_graph_d = {
'a': {'b': 1, 'd': 4, 'e': 3},
'b': {'a': 1, 'd': 4, 'e': 2},
'c': {'e': 4, 'f': 5},
'd': {'a': 4, 'b': 4, 'e': 4},
'e': {'a': 3, 'b': 2, 'c': 4, 'd': 4, 'f': 7},
'f': {'c': 5, 'e': 7}
}
print('w_graph_d:\n{}'.format(w_graph_d))
print('Kruskal\'s minimum spanning tree:')
pass
if __name__ == '__main__':
main()
|
Add weighted graph in main()
|
Add weighted graph in main()
|
Python
|
bsd-2-clause
|
bowen0701/algorithms_data_structures
|
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
def kruskal():
"""Kruskal's algorithm for minimum spanning tree in weighted graph.
Time complexity for graph G(V, E): TBD.
"""
pass
def main():
pass
if __name__ == '__main__':
main()
Add weighted graph in main()
|
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
def kruskal():
"""Kruskal's algorithm for minimum spanning tree in weighted graph.
Time complexity for graph G(V, E): TBD.
"""
pass
def main():
w_graph_d = {
'a': {'b': 1, 'd': 4, 'e': 3},
'b': {'a': 1, 'd': 4, 'e': 2},
'c': {'e': 4, 'f': 5},
'd': {'a': 4, 'b': 4, 'e': 4},
'e': {'a': 3, 'b': 2, 'c': 4, 'd': 4, 'f': 7},
'f': {'c': 5, 'e': 7}
}
print('w_graph_d:\n{}'.format(w_graph_d))
print('Kruskal\'s minimum spanning tree:')
pass
if __name__ == '__main__':
main()
|
<commit_before>from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
def kruskal():
"""Kruskal's algorithm for minimum spanning tree in weighted graph.
Time complexity for graph G(V, E): TBD.
"""
pass
def main():
pass
if __name__ == '__main__':
main()
<commit_msg>Add weighted graph in main()<commit_after>
|
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
def kruskal():
"""Kruskal's algorithm for minimum spanning tree in weighted graph.
Time complexity for graph G(V, E): TBD.
"""
pass
def main():
w_graph_d = {
'a': {'b': 1, 'd': 4, 'e': 3},
'b': {'a': 1, 'd': 4, 'e': 2},
'c': {'e': 4, 'f': 5},
'd': {'a': 4, 'b': 4, 'e': 4},
'e': {'a': 3, 'b': 2, 'c': 4, 'd': 4, 'f': 7},
'f': {'c': 5, 'e': 7}
}
print('w_graph_d:\n{}'.format(w_graph_d))
print('Kruskal\'s minimum spanning tree:')
pass
if __name__ == '__main__':
main()
|
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
def kruskal():
"""Kruskal's algorithm for minimum spanning tree in weighted graph.
Time complexity for graph G(V, E): TBD.
"""
pass
def main():
pass
if __name__ == '__main__':
main()
Add weighted graph in main()from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
def kruskal():
"""Kruskal's algorithm for minimum spanning tree in weighted graph.
Time complexity for graph G(V, E): TBD.
"""
pass
def main():
w_graph_d = {
'a': {'b': 1, 'd': 4, 'e': 3},
'b': {'a': 1, 'd': 4, 'e': 2},
'c': {'e': 4, 'f': 5},
'd': {'a': 4, 'b': 4, 'e': 4},
'e': {'a': 3, 'b': 2, 'c': 4, 'd': 4, 'f': 7},
'f': {'c': 5, 'e': 7}
}
print('w_graph_d:\n{}'.format(w_graph_d))
print('Kruskal\'s minimum spanning tree:')
pass
if __name__ == '__main__':
main()
|
<commit_before>from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
def kruskal():
"""Kruskal's algorithm for minimum spanning tree in weighted graph.
Time complexity for graph G(V, E): TBD.
"""
pass
def main():
pass
if __name__ == '__main__':
main()
<commit_msg>Add weighted graph in main()<commit_after>from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
def kruskal():
"""Kruskal's algorithm for minimum spanning tree in weighted graph.
Time complexity for graph G(V, E): TBD.
"""
pass
def main():
w_graph_d = {
'a': {'b': 1, 'd': 4, 'e': 3},
'b': {'a': 1, 'd': 4, 'e': 2},
'c': {'e': 4, 'f': 5},
'd': {'a': 4, 'b': 4, 'e': 4},
'e': {'a': 3, 'b': 2, 'c': 4, 'd': 4, 'f': 7},
'f': {'c': 5, 'e': 7}
}
print('w_graph_d:\n{}'.format(w_graph_d))
print('Kruskal\'s minimum spanning tree:')
pass
if __name__ == '__main__':
main()
|
4b5f8e14db9cd157d1b3b616726b1c9fb1b3c9b5
|
demos/py_simple/rotate90.py
|
demos/py_simple/rotate90.py
|
#!/usr/bin/env python
import sys
import gfxprim.core as core
import gfxprim.loaders as loaders
import gfxprim.filters as filters
def main():
if len(sys.argv) != 3:
print("USAGE: %s imput_image output_image" % sys.argv[0]);
sys.exit(1)
# Turns on debug messages
core.SetDebugLevel(10);
# Load Image
src = loaders.LoadImage(sys.argv[1], None)
# Rotate by 90 degrees
res = filters.FilterRotate90Alloc(src, None)
# Save Image
res.Save(sys.argv[2])
if __name__ == '__main__':
main()
|
#!/usr/bin/env python
import sys
import gfxprim.core as core
import gfxprim.loaders as loaders
import gfxprim.filters as filters
def main():
if len(sys.argv) != 3:
print("USAGE: %s imput_image output_image" % sys.argv[0]);
sys.exit(1)
# Turns on debug messages
core.SetDebugLevel(10);
# Load Image
src = loaders.LoadImage(sys.argv[1], None)
# Rotate by 90 degrees
res = filters.FilterRotate90_Alloc(src, None)
# Save Image
res.Save(sys.argv[2])
if __name__ == '__main__':
main()
|
Fix python example after the API update.
|
py_simple: Fix python example after the API update.
|
Python
|
lgpl-2.1
|
gfxprim/gfxprim,gfxprim/gfxprim,gfxprim/gfxprim,gfxprim/gfxprim,gfxprim/gfxprim
|
#!/usr/bin/env python
import sys
import gfxprim.core as core
import gfxprim.loaders as loaders
import gfxprim.filters as filters
def main():
if len(sys.argv) != 3:
print("USAGE: %s imput_image output_image" % sys.argv[0]);
sys.exit(1)
# Turns on debug messages
core.SetDebugLevel(10);
# Load Image
src = loaders.LoadImage(sys.argv[1], None)
# Rotate by 90 degrees
res = filters.FilterRotate90Alloc(src, None)
# Save Image
res.Save(sys.argv[2])
if __name__ == '__main__':
main()
py_simple: Fix python example after the API update.
|
#!/usr/bin/env python
import sys
import gfxprim.core as core
import gfxprim.loaders as loaders
import gfxprim.filters as filters
def main():
if len(sys.argv) != 3:
print("USAGE: %s imput_image output_image" % sys.argv[0]);
sys.exit(1)
# Turns on debug messages
core.SetDebugLevel(10);
# Load Image
src = loaders.LoadImage(sys.argv[1], None)
# Rotate by 90 degrees
res = filters.FilterRotate90_Alloc(src, None)
# Save Image
res.Save(sys.argv[2])
if __name__ == '__main__':
main()
|
<commit_before>#!/usr/bin/env python
import sys
import gfxprim.core as core
import gfxprim.loaders as loaders
import gfxprim.filters as filters
def main():
if len(sys.argv) != 3:
print("USAGE: %s imput_image output_image" % sys.argv[0]);
sys.exit(1)
# Turns on debug messages
core.SetDebugLevel(10);
# Load Image
src = loaders.LoadImage(sys.argv[1], None)
# Rotate by 90 degrees
res = filters.FilterRotate90Alloc(src, None)
# Save Image
res.Save(sys.argv[2])
if __name__ == '__main__':
main()
<commit_msg>py_simple: Fix python example after the API update.<commit_after>
|
#!/usr/bin/env python
import sys
import gfxprim.core as core
import gfxprim.loaders as loaders
import gfxprim.filters as filters
def main():
if len(sys.argv) != 3:
print("USAGE: %s imput_image output_image" % sys.argv[0]);
sys.exit(1)
# Turns on debug messages
core.SetDebugLevel(10);
# Load Image
src = loaders.LoadImage(sys.argv[1], None)
# Rotate by 90 degrees
res = filters.FilterRotate90_Alloc(src, None)
# Save Image
res.Save(sys.argv[2])
if __name__ == '__main__':
main()
|
#!/usr/bin/env python
import sys
import gfxprim.core as core
import gfxprim.loaders as loaders
import gfxprim.filters as filters
def main():
if len(sys.argv) != 3:
print("USAGE: %s imput_image output_image" % sys.argv[0]);
sys.exit(1)
# Turns on debug messages
core.SetDebugLevel(10);
# Load Image
src = loaders.LoadImage(sys.argv[1], None)
# Rotate by 90 degrees
res = filters.FilterRotate90Alloc(src, None)
# Save Image
res.Save(sys.argv[2])
if __name__ == '__main__':
main()
py_simple: Fix python example after the API update.#!/usr/bin/env python
import sys
import gfxprim.core as core
import gfxprim.loaders as loaders
import gfxprim.filters as filters
def main():
if len(sys.argv) != 3:
print("USAGE: %s imput_image output_image" % sys.argv[0]);
sys.exit(1)
# Turns on debug messages
core.SetDebugLevel(10);
# Load Image
src = loaders.LoadImage(sys.argv[1], None)
# Rotate by 90 degrees
res = filters.FilterRotate90_Alloc(src, None)
# Save Image
res.Save(sys.argv[2])
if __name__ == '__main__':
main()
|
<commit_before>#!/usr/bin/env python
import sys
import gfxprim.core as core
import gfxprim.loaders as loaders
import gfxprim.filters as filters
def main():
if len(sys.argv) != 3:
print("USAGE: %s imput_image output_image" % sys.argv[0]);
sys.exit(1)
# Turns on debug messages
core.SetDebugLevel(10);
# Load Image
src = loaders.LoadImage(sys.argv[1], None)
# Rotate by 90 degrees
res = filters.FilterRotate90Alloc(src, None)
# Save Image
res.Save(sys.argv[2])
if __name__ == '__main__':
main()
<commit_msg>py_simple: Fix python example after the API update.<commit_after>#!/usr/bin/env python
import sys
import gfxprim.core as core
import gfxprim.loaders as loaders
import gfxprim.filters as filters
def main():
if len(sys.argv) != 3:
print("USAGE: %s imput_image output_image" % sys.argv[0]);
sys.exit(1)
# Turns on debug messages
core.SetDebugLevel(10);
# Load Image
src = loaders.LoadImage(sys.argv[1], None)
# Rotate by 90 degrees
res = filters.FilterRotate90_Alloc(src, None)
# Save Image
res.Save(sys.argv[2])
if __name__ == '__main__':
main()
|
283d7299c732b80d504e971424b18996719fdf80
|
dsub/_dsub_version.py
|
dsub/_dsub_version.py
|
# Copyright 2017 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Single source of truth for dsub's version.
This must remain small and dependency-free so that any dsub module may
import it without creating circular dependencies. Note that this module
is parsed as a text file by setup.py and changes to the format of this
file could break setup.py.
The version should follow formatting requirements specified in PEP-440.
- https://www.python.org/dev/peps/pep-0440
A typical release sequence will be versioned as:
0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...
"""
DSUB_VERSION = '0.3.6'
|
# Copyright 2017 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Single source of truth for dsub's version.
This must remain small and dependency-free so that any dsub module may
import it without creating circular dependencies. Note that this module
is parsed as a text file by setup.py and changes to the format of this
file could break setup.py.
The version should follow formatting requirements specified in PEP-440.
- https://www.python.org/dev/peps/pep-0440
A typical release sequence will be versioned as:
0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...
"""
DSUB_VERSION = '0.3.7.dev0'
|
Update dsub version to 0.3.7.dev0
|
Update dsub version to 0.3.7.dev0
PiperOrigin-RevId: 281987296
|
Python
|
apache-2.0
|
DataBiosphere/dsub,DataBiosphere/dsub
|
# Copyright 2017 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Single source of truth for dsub's version.
This must remain small and dependency-free so that any dsub module may
import it without creating circular dependencies. Note that this module
is parsed as a text file by setup.py and changes to the format of this
file could break setup.py.
The version should follow formatting requirements specified in PEP-440.
- https://www.python.org/dev/peps/pep-0440
A typical release sequence will be versioned as:
0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...
"""
DSUB_VERSION = '0.3.6'
Update dsub version to 0.3.7.dev0
PiperOrigin-RevId: 281987296
|
# Copyright 2017 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Single source of truth for dsub's version.
This must remain small and dependency-free so that any dsub module may
import it without creating circular dependencies. Note that this module
is parsed as a text file by setup.py and changes to the format of this
file could break setup.py.
The version should follow formatting requirements specified in PEP-440.
- https://www.python.org/dev/peps/pep-0440
A typical release sequence will be versioned as:
0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...
"""
DSUB_VERSION = '0.3.7.dev0'
|
<commit_before># Copyright 2017 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Single source of truth for dsub's version.
This must remain small and dependency-free so that any dsub module may
import it without creating circular dependencies. Note that this module
is parsed as a text file by setup.py and changes to the format of this
file could break setup.py.
The version should follow formatting requirements specified in PEP-440.
- https://www.python.org/dev/peps/pep-0440
A typical release sequence will be versioned as:
0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...
"""
DSUB_VERSION = '0.3.6'
<commit_msg>Update dsub version to 0.3.7.dev0
PiperOrigin-RevId: 281987296<commit_after>
|
# Copyright 2017 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Single source of truth for dsub's version.
This must remain small and dependency-free so that any dsub module may
import it without creating circular dependencies. Note that this module
is parsed as a text file by setup.py and changes to the format of this
file could break setup.py.
The version should follow formatting requirements specified in PEP-440.
- https://www.python.org/dev/peps/pep-0440
A typical release sequence will be versioned as:
0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...
"""
DSUB_VERSION = '0.3.7.dev0'
|
# Copyright 2017 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Single source of truth for dsub's version.
This must remain small and dependency-free so that any dsub module may
import it without creating circular dependencies. Note that this module
is parsed as a text file by setup.py and changes to the format of this
file could break setup.py.
The version should follow formatting requirements specified in PEP-440.
- https://www.python.org/dev/peps/pep-0440
A typical release sequence will be versioned as:
0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...
"""
DSUB_VERSION = '0.3.6'
Update dsub version to 0.3.7.dev0
PiperOrigin-RevId: 281987296# Copyright 2017 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Single source of truth for dsub's version.
This must remain small and dependency-free so that any dsub module may
import it without creating circular dependencies. Note that this module
is parsed as a text file by setup.py and changes to the format of this
file could break setup.py.
The version should follow formatting requirements specified in PEP-440.
- https://www.python.org/dev/peps/pep-0440
A typical release sequence will be versioned as:
0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...
"""
DSUB_VERSION = '0.3.7.dev0'
|
<commit_before># Copyright 2017 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Single source of truth for dsub's version.
This must remain small and dependency-free so that any dsub module may
import it without creating circular dependencies. Note that this module
is parsed as a text file by setup.py and changes to the format of this
file could break setup.py.
The version should follow formatting requirements specified in PEP-440.
- https://www.python.org/dev/peps/pep-0440
A typical release sequence will be versioned as:
0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...
"""
DSUB_VERSION = '0.3.6'
<commit_msg>Update dsub version to 0.3.7.dev0
PiperOrigin-RevId: 281987296<commit_after># Copyright 2017 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Single source of truth for dsub's version.
This must remain small and dependency-free so that any dsub module may
import it without creating circular dependencies. Note that this module
is parsed as a text file by setup.py and changes to the format of this
file could break setup.py.
The version should follow formatting requirements specified in PEP-440.
- https://www.python.org/dev/peps/pep-0440
A typical release sequence will be versioned as:
0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...
"""
DSUB_VERSION = '0.3.7.dev0'
|
a5ae06630ef96d1093e4498e0e5437c0a7e65bfa
|
parse.py
|
parse.py
|
from PIL import Image
import sys
import pyocr
import pyocr.builders
tools = pyocr.get_available_tools()
if len(tools) == 0:
print("Error: No OCR tool found")
sys.exit(1)
# should be 'Tesseract (sh)'
tool = tools[0]
orig_image = Image.open('test.png')
# crop to only the section with the number of problems solved
cropped_image = orig_image.crop((47, 40, 97, 60))
# double the size of the image so the OCR has more to go on
resized_image = cropped_image.resize((100, 40), Image.ANTIALIAS)
digits = tool.image_to_string(
resized_image,
builder=pyocr.tesseract.DigitBuilder()
)
print(digits)
|
from PIL import Image
import sys
import pyocr
import pyocr.builders
image_loc = ' '.join(sys.argv[1:])
tools = pyocr.get_available_tools()
if len(tools) == 0:
print("Error: No OCR tool found")
sys.exit(1)
# should be 'Tesseract (sh)'
tool = tools[0]
orig_image = Image.open(image_loc)
# crop to only the section with the number of problems solved
cropped_image = orig_image.crop((47, 40, 97, 60))
# double the size of the image so the OCR has more to go on
resized_image = cropped_image.resize((100, 40), Image.ANTIALIAS)
digits = tool.image_to_string(
resized_image,
builder=pyocr.tesseract.DigitBuilder()
)
print(digits)
|
Allow user to specify image file w argv
|
Allow user to specify image file w argv
|
Python
|
bsd-2-clause
|
iandioch/euler-foiler
|
from PIL import Image
import sys
import pyocr
import pyocr.builders
tools = pyocr.get_available_tools()
if len(tools) == 0:
print("Error: No OCR tool found")
sys.exit(1)
# should be 'Tesseract (sh)'
tool = tools[0]
orig_image = Image.open('test.png')
# crop to only the section with the number of problems solved
cropped_image = orig_image.crop((47, 40, 97, 60))
# double the size of the image so the OCR has more to go on
resized_image = cropped_image.resize((100, 40), Image.ANTIALIAS)
digits = tool.image_to_string(
resized_image,
builder=pyocr.tesseract.DigitBuilder()
)
print(digits)
Allow user to specify image file w argv
|
from PIL import Image
import sys
import pyocr
import pyocr.builders
image_loc = ' '.join(sys.argv[1:])
tools = pyocr.get_available_tools()
if len(tools) == 0:
print("Error: No OCR tool found")
sys.exit(1)
# should be 'Tesseract (sh)'
tool = tools[0]
orig_image = Image.open(image_loc)
# crop to only the section with the number of problems solved
cropped_image = orig_image.crop((47, 40, 97, 60))
# double the size of the image so the OCR has more to go on
resized_image = cropped_image.resize((100, 40), Image.ANTIALIAS)
digits = tool.image_to_string(
resized_image,
builder=pyocr.tesseract.DigitBuilder()
)
print(digits)
|
<commit_before>from PIL import Image
import sys
import pyocr
import pyocr.builders
tools = pyocr.get_available_tools()
if len(tools) == 0:
print("Error: No OCR tool found")
sys.exit(1)
# should be 'Tesseract (sh)'
tool = tools[0]
orig_image = Image.open('test.png')
# crop to only the section with the number of problems solved
cropped_image = orig_image.crop((47, 40, 97, 60))
# double the size of the image so the OCR has more to go on
resized_image = cropped_image.resize((100, 40), Image.ANTIALIAS)
digits = tool.image_to_string(
resized_image,
builder=pyocr.tesseract.DigitBuilder()
)
print(digits)
<commit_msg>Allow user to specify image file w argv<commit_after>
|
from PIL import Image
import sys
import pyocr
import pyocr.builders
image_loc = ' '.join(sys.argv[1:])
tools = pyocr.get_available_tools()
if len(tools) == 0:
print("Error: No OCR tool found")
sys.exit(1)
# should be 'Tesseract (sh)'
tool = tools[0]
orig_image = Image.open(image_loc)
# crop to only the section with the number of problems solved
cropped_image = orig_image.crop((47, 40, 97, 60))
# double the size of the image so the OCR has more to go on
resized_image = cropped_image.resize((100, 40), Image.ANTIALIAS)
digits = tool.image_to_string(
resized_image,
builder=pyocr.tesseract.DigitBuilder()
)
print(digits)
|
from PIL import Image
import sys
import pyocr
import pyocr.builders
tools = pyocr.get_available_tools()
if len(tools) == 0:
print("Error: No OCR tool found")
sys.exit(1)
# should be 'Tesseract (sh)'
tool = tools[0]
orig_image = Image.open('test.png')
# crop to only the section with the number of problems solved
cropped_image = orig_image.crop((47, 40, 97, 60))
# double the size of the image so the OCR has more to go on
resized_image = cropped_image.resize((100, 40), Image.ANTIALIAS)
digits = tool.image_to_string(
resized_image,
builder=pyocr.tesseract.DigitBuilder()
)
print(digits)
Allow user to specify image file w argvfrom PIL import Image
import sys
import pyocr
import pyocr.builders
image_loc = ' '.join(sys.argv[1:])
tools = pyocr.get_available_tools()
if len(tools) == 0:
print("Error: No OCR tool found")
sys.exit(1)
# should be 'Tesseract (sh)'
tool = tools[0]
orig_image = Image.open(image_loc)
# crop to only the section with the number of problems solved
cropped_image = orig_image.crop((47, 40, 97, 60))
# double the size of the image so the OCR has more to go on
resized_image = cropped_image.resize((100, 40), Image.ANTIALIAS)
digits = tool.image_to_string(
resized_image,
builder=pyocr.tesseract.DigitBuilder()
)
print(digits)
|
<commit_before>from PIL import Image
import sys
import pyocr
import pyocr.builders
tools = pyocr.get_available_tools()
if len(tools) == 0:
print("Error: No OCR tool found")
sys.exit(1)
# should be 'Tesseract (sh)'
tool = tools[0]
orig_image = Image.open('test.png')
# crop to only the section with the number of problems solved
cropped_image = orig_image.crop((47, 40, 97, 60))
# double the size of the image so the OCR has more to go on
resized_image = cropped_image.resize((100, 40), Image.ANTIALIAS)
digits = tool.image_to_string(
resized_image,
builder=pyocr.tesseract.DigitBuilder()
)
print(digits)
<commit_msg>Allow user to specify image file w argv<commit_after>from PIL import Image
import sys
import pyocr
import pyocr.builders
image_loc = ' '.join(sys.argv[1:])
tools = pyocr.get_available_tools()
if len(tools) == 0:
print("Error: No OCR tool found")
sys.exit(1)
# should be 'Tesseract (sh)'
tool = tools[0]
orig_image = Image.open(image_loc)
# crop to only the section with the number of problems solved
cropped_image = orig_image.crop((47, 40, 97, 60))
# double the size of the image so the OCR has more to go on
resized_image = cropped_image.resize((100, 40), Image.ANTIALIAS)
digits = tool.image_to_string(
resized_image,
builder=pyocr.tesseract.DigitBuilder()
)
print(digits)
|
b67a5daaa7efc946aebcdfdabbe201057af4aef5
|
globus_sdk/version.py
|
globus_sdk/version.py
|
# single source of truth for package version,
# see https://packaging.python.org/en/latest/single_source_version/
__version__ = "0.6.0"
|
# single source of truth for package version,
# see https://packaging.python.org/en/latest/single_source_version/
__version__ = "0.7.0"
|
Update to v0.7.0 for release
|
Update to v0.7.0 for release
|
Python
|
apache-2.0
|
aaschaer/globus-sdk-python,sirosen/globus-sdk-python,globus/globus-sdk-python,globus/globus-sdk-python,globusonline/globus-sdk-python
|
# single source of truth for package version,
# see https://packaging.python.org/en/latest/single_source_version/
__version__ = "0.6.0"
Update to v0.7.0 for release
|
# single source of truth for package version,
# see https://packaging.python.org/en/latest/single_source_version/
__version__ = "0.7.0"
|
<commit_before># single source of truth for package version,
# see https://packaging.python.org/en/latest/single_source_version/
__version__ = "0.6.0"
<commit_msg>Update to v0.7.0 for release<commit_after>
|
# single source of truth for package version,
# see https://packaging.python.org/en/latest/single_source_version/
__version__ = "0.7.0"
|
# single source of truth for package version,
# see https://packaging.python.org/en/latest/single_source_version/
__version__ = "0.6.0"
Update to v0.7.0 for release# single source of truth for package version,
# see https://packaging.python.org/en/latest/single_source_version/
__version__ = "0.7.0"
|
<commit_before># single source of truth for package version,
# see https://packaging.python.org/en/latest/single_source_version/
__version__ = "0.6.0"
<commit_msg>Update to v0.7.0 for release<commit_after># single source of truth for package version,
# see https://packaging.python.org/en/latest/single_source_version/
__version__ = "0.7.0"
|
a16889f353873e3d08a24440b9aa83177ffd001f
|
engine.py
|
engine.py
|
#!/usr/bin/env python
import json
import sys
import os # For os.path and the like
class DictWrapper(object):
def __init__(self, d):
self.__dict__ = d
def eval_script(self):
return eval(self.script) # With self as context
d = json.load(sys.stdin)
dw = DictWrapper(d)
json.dump(dw.eval_script(), sys.stdout)
|
#!/usr/bin/env python
import json
import sys
import os # For os.path and the like
class DictWrapper(object):
def __init__(self, d):
self.__dict__ = d
def eval_script(self):
return eval(self.script) # With self as context
def __getattr__(self, attr):
return None
if __name__ == '__main__':
input_dict = json.load(sys.stdin)
dw = DictWrapper(input_dict)
json.dump(dw.eval_script(), sys.stdout)
|
Implement __getattr__ to handle KeyErrors
|
Implement __getattr__ to handle KeyErrors
|
Python
|
mit
|
dleehr/py-expr-engine
|
#!/usr/bin/env python
import json
import sys
import os # For os.path and the like
class DictWrapper(object):
def __init__(self, d):
self.__dict__ = d
def eval_script(self):
return eval(self.script) # With self as context
d = json.load(sys.stdin)
dw = DictWrapper(d)
json.dump(dw.eval_script(), sys.stdout)
Implement __getattr__ to handle KeyErrors
|
#!/usr/bin/env python
import json
import sys
import os # For os.path and the like
class DictWrapper(object):
def __init__(self, d):
self.__dict__ = d
def eval_script(self):
return eval(self.script) # With self as context
def __getattr__(self, attr):
return None
if __name__ == '__main__':
input_dict = json.load(sys.stdin)
dw = DictWrapper(input_dict)
json.dump(dw.eval_script(), sys.stdout)
|
<commit_before>#!/usr/bin/env python
import json
import sys
import os # For os.path and the like
class DictWrapper(object):
def __init__(self, d):
self.__dict__ = d
def eval_script(self):
return eval(self.script) # With self as context
d = json.load(sys.stdin)
dw = DictWrapper(d)
json.dump(dw.eval_script(), sys.stdout)
<commit_msg>Implement __getattr__ to handle KeyErrors<commit_after>
|
#!/usr/bin/env python
import json
import sys
import os # For os.path and the like
class DictWrapper(object):
def __init__(self, d):
self.__dict__ = d
def eval_script(self):
return eval(self.script) # With self as context
def __getattr__(self, attr):
return None
if __name__ == '__main__':
input_dict = json.load(sys.stdin)
dw = DictWrapper(input_dict)
json.dump(dw.eval_script(), sys.stdout)
|
#!/usr/bin/env python
import json
import sys
import os # For os.path and the like
class DictWrapper(object):
def __init__(self, d):
self.__dict__ = d
def eval_script(self):
return eval(self.script) # With self as context
d = json.load(sys.stdin)
dw = DictWrapper(d)
json.dump(dw.eval_script(), sys.stdout)
Implement __getattr__ to handle KeyErrors#!/usr/bin/env python
import json
import sys
import os # For os.path and the like
class DictWrapper(object):
def __init__(self, d):
self.__dict__ = d
def eval_script(self):
return eval(self.script) # With self as context
def __getattr__(self, attr):
return None
if __name__ == '__main__':
input_dict = json.load(sys.stdin)
dw = DictWrapper(input_dict)
json.dump(dw.eval_script(), sys.stdout)
|
<commit_before>#!/usr/bin/env python
import json
import sys
import os # For os.path and the like
class DictWrapper(object):
def __init__(self, d):
self.__dict__ = d
def eval_script(self):
return eval(self.script) # With self as context
d = json.load(sys.stdin)
dw = DictWrapper(d)
json.dump(dw.eval_script(), sys.stdout)
<commit_msg>Implement __getattr__ to handle KeyErrors<commit_after>#!/usr/bin/env python
import json
import sys
import os # For os.path and the like
class DictWrapper(object):
def __init__(self, d):
self.__dict__ = d
def eval_script(self):
return eval(self.script) # With self as context
def __getattr__(self, attr):
return None
if __name__ == '__main__':
input_dict = json.load(sys.stdin)
dw = DictWrapper(input_dict)
json.dump(dw.eval_script(), sys.stdout)
|
fc7beded3d286d831df29b8b32614b2eb56ef206
|
enasearch/__main__.py
|
enasearch/__main__.py
|
#!/usr/bin/env python
import click
import ebisearch
from pprint import pprint
@click.group()
def main():
pass
@click.command('get_results', short_help='Get list of results')
def get_results():
"""Return the list of domains in EBI"""
ebisearch.get_results(verbose=True)
@click.command('get_filter_fields', short_help='Get filter fields')
@click.option(
'--result',
help='Id of a result (accessible with get_results)')
def get_filter_fields(result):
"""Get the filter fields of a result to build a query"""
ebisearch.get_filter_fields(verbose=True)
main.add_command(get_results)
main.add_command(get_filter_fields)
if __name__ == "__main__":
main()
|
#!/usr/bin/env python
import click
import ebisearch
from pprint import pprint
@click.group()
def main():
pass
@click.command('get_results', short_help='Get list of results')
def get_results():
"""Return the list of domains in EBI"""
ebisearch.get_results(verbose=True)
@click.command('get_filter_fields', short_help='Get filter fields')
@click.option(
'--result',
help='Id of a result (accessible with get_results)')
def get_filter_fields(result):
"""Get the filter fields of a result to build a query"""
ebisearch.get_filter_fields(verbose=True)
@click.command('get_filter_types', short_help='Get filter types')
def get_filter_types():
"""Get the types of filters usable to build a query"""
ebisearch.get_filter_types(verbose=True)
main.add_command(get_results)
main.add_command(get_filter_fields)
main.add_command(get_filter_types)
if __name__ == "__main__":
main()
|
Add function for get filter types
|
Add function for get filter types
|
Python
|
mit
|
bebatut/enasearch
|
#!/usr/bin/env python
import click
import ebisearch
from pprint import pprint
@click.group()
def main():
pass
@click.command('get_results', short_help='Get list of results')
def get_results():
"""Return the list of domains in EBI"""
ebisearch.get_results(verbose=True)
@click.command('get_filter_fields', short_help='Get filter fields')
@click.option(
'--result',
help='Id of a result (accessible with get_results)')
def get_filter_fields(result):
"""Get the filter fields of a result to build a query"""
ebisearch.get_filter_fields(verbose=True)
main.add_command(get_results)
main.add_command(get_filter_fields)
if __name__ == "__main__":
main()
Add function for get filter types
|
#!/usr/bin/env python
import click
import ebisearch
from pprint import pprint
@click.group()
def main():
pass
@click.command('get_results', short_help='Get list of results')
def get_results():
"""Return the list of domains in EBI"""
ebisearch.get_results(verbose=True)
@click.command('get_filter_fields', short_help='Get filter fields')
@click.option(
'--result',
help='Id of a result (accessible with get_results)')
def get_filter_fields(result):
"""Get the filter fields of a result to build a query"""
ebisearch.get_filter_fields(verbose=True)
@click.command('get_filter_types', short_help='Get filter types')
def get_filter_types():
"""Get the types of filters usable to build a query"""
ebisearch.get_filter_types(verbose=True)
main.add_command(get_results)
main.add_command(get_filter_fields)
main.add_command(get_filter_types)
if __name__ == "__main__":
main()
|
<commit_before>#!/usr/bin/env python
import click
import ebisearch
from pprint import pprint
@click.group()
def main():
pass
@click.command('get_results', short_help='Get list of results')
def get_results():
"""Return the list of domains in EBI"""
ebisearch.get_results(verbose=True)
@click.command('get_filter_fields', short_help='Get filter fields')
@click.option(
'--result',
help='Id of a result (accessible with get_results)')
def get_filter_fields(result):
"""Get the filter fields of a result to build a query"""
ebisearch.get_filter_fields(verbose=True)
main.add_command(get_results)
main.add_command(get_filter_fields)
if __name__ == "__main__":
main()
<commit_msg>Add function for get filter types<commit_after>
|
#!/usr/bin/env python
import click
import ebisearch
from pprint import pprint
@click.group()
def main():
pass
@click.command('get_results', short_help='Get list of results')
def get_results():
"""Return the list of domains in EBI"""
ebisearch.get_results(verbose=True)
@click.command('get_filter_fields', short_help='Get filter fields')
@click.option(
'--result',
help='Id of a result (accessible with get_results)')
def get_filter_fields(result):
"""Get the filter fields of a result to build a query"""
ebisearch.get_filter_fields(verbose=True)
@click.command('get_filter_types', short_help='Get filter types')
def get_filter_types():
"""Get the types of filters usable to build a query"""
ebisearch.get_filter_types(verbose=True)
main.add_command(get_results)
main.add_command(get_filter_fields)
main.add_command(get_filter_types)
if __name__ == "__main__":
main()
|
#!/usr/bin/env python
import click
import ebisearch
from pprint import pprint
@click.group()
def main():
pass
@click.command('get_results', short_help='Get list of results')
def get_results():
"""Return the list of domains in EBI"""
ebisearch.get_results(verbose=True)
@click.command('get_filter_fields', short_help='Get filter fields')
@click.option(
'--result',
help='Id of a result (accessible with get_results)')
def get_filter_fields(result):
"""Get the filter fields of a result to build a query"""
ebisearch.get_filter_fields(verbose=True)
main.add_command(get_results)
main.add_command(get_filter_fields)
if __name__ == "__main__":
main()
Add function for get filter types#!/usr/bin/env python
import click
import ebisearch
from pprint import pprint
@click.group()
def main():
pass
@click.command('get_results', short_help='Get list of results')
def get_results():
"""Return the list of domains in EBI"""
ebisearch.get_results(verbose=True)
@click.command('get_filter_fields', short_help='Get filter fields')
@click.option(
'--result',
help='Id of a result (accessible with get_results)')
def get_filter_fields(result):
"""Get the filter fields of a result to build a query"""
ebisearch.get_filter_fields(verbose=True)
@click.command('get_filter_types', short_help='Get filter types')
def get_filter_types():
"""Get the types of filters usable to build a query"""
ebisearch.get_filter_types(verbose=True)
main.add_command(get_results)
main.add_command(get_filter_fields)
main.add_command(get_filter_types)
if __name__ == "__main__":
main()
|
<commit_before>#!/usr/bin/env python
import click
import ebisearch
from pprint import pprint
@click.group()
def main():
pass
@click.command('get_results', short_help='Get list of results')
def get_results():
"""Return the list of domains in EBI"""
ebisearch.get_results(verbose=True)
@click.command('get_filter_fields', short_help='Get filter fields')
@click.option(
'--result',
help='Id of a result (accessible with get_results)')
def get_filter_fields(result):
"""Get the filter fields of a result to build a query"""
ebisearch.get_filter_fields(verbose=True)
main.add_command(get_results)
main.add_command(get_filter_fields)
if __name__ == "__main__":
main()
<commit_msg>Add function for get filter types<commit_after>#!/usr/bin/env python
import click
import ebisearch
from pprint import pprint
@click.group()
def main():
pass
@click.command('get_results', short_help='Get list of results')
def get_results():
"""Return the list of domains in EBI"""
ebisearch.get_results(verbose=True)
@click.command('get_filter_fields', short_help='Get filter fields')
@click.option(
'--result',
help='Id of a result (accessible with get_results)')
def get_filter_fields(result):
"""Get the filter fields of a result to build a query"""
ebisearch.get_filter_fields(verbose=True)
@click.command('get_filter_types', short_help='Get filter types')
def get_filter_types():
"""Get the types of filters usable to build a query"""
ebisearch.get_filter_types(verbose=True)
main.add_command(get_results)
main.add_command(get_filter_fields)
main.add_command(get_filter_types)
if __name__ == "__main__":
main()
|
9fcfd8e13b5c4684a1cb3890427662ded2d28c24
|
examples/get_dataset.py
|
examples/get_dataset.py
|
#!/usr/bin/env python3
#
# This script is used for downloading the dataset used by the examples.
# Dataset used: UCI / Pima Indians Diabetes (in libsvm format)
import os
import urllib.request
DATASET_URL = 'http://www.csie.ntu.edu.tw/~cjlin/libsvmtools/datasets/binary/diabetes'
TARGET_PATH = os.path.dirname(os.path.realpath(__file__)) + '/dataset.txt'
def main():
urllib.request.urlretrieve(DATASET_URL, TARGET_PATH)
if __name__ == '__main__': main()
|
#!/usr/bin/env python3
#
# This script is used for downloading the dataset used by the examples.
# Dataset used: Statlog / Letter (in libsvm format)
import os
import urllib.request
import random
DATASET_URL = 'http://www.csie.ntu.edu.tw/~cjlin/libsvmtools/datasets/multiclass/letter.scale'
DATASET_SIZE = 1000
TARGET_PATH = os.path.dirname(os.path.realpath(__file__)) + '/dataset.txt'
def main():
rows = list(urllib.request.urlopen(DATASET_URL))
selected = random.sample(rows, DATASET_SIZE)
with open(TARGET_PATH, 'wb') as f:
for row in selected:
f.write(row)
if __name__ == '__main__': main()
|
Change dataset used in example (letter)
|
Change dataset used in example (letter)
XXX: UncertaintySampling(le) weird?
|
Python
|
bsd-2-clause
|
ntucllab/libact,ntucllab/libact,ntucllab/libact
|
#!/usr/bin/env python3
#
# This script is used for downloading the dataset used by the examples.
# Dataset used: UCI / Pima Indians Diabetes (in libsvm format)
import os
import urllib.request
DATASET_URL = 'http://www.csie.ntu.edu.tw/~cjlin/libsvmtools/datasets/binary/diabetes'
TARGET_PATH = os.path.dirname(os.path.realpath(__file__)) + '/dataset.txt'
def main():
urllib.request.urlretrieve(DATASET_URL, TARGET_PATH)
if __name__ == '__main__': main()
Change dataset used in example (letter)
XXX: UncertaintySampling(le) weird?
|
#!/usr/bin/env python3
#
# This script is used for downloading the dataset used by the examples.
# Dataset used: Statlog / Letter (in libsvm format)
import os
import urllib.request
import random
DATASET_URL = 'http://www.csie.ntu.edu.tw/~cjlin/libsvmtools/datasets/multiclass/letter.scale'
DATASET_SIZE = 1000
TARGET_PATH = os.path.dirname(os.path.realpath(__file__)) + '/dataset.txt'
def main():
rows = list(urllib.request.urlopen(DATASET_URL))
selected = random.sample(rows, DATASET_SIZE)
with open(TARGET_PATH, 'wb') as f:
for row in selected:
f.write(row)
if __name__ == '__main__': main()
|
<commit_before>#!/usr/bin/env python3
#
# This script is used for downloading the dataset used by the examples.
# Dataset used: UCI / Pima Indians Diabetes (in libsvm format)
import os
import urllib.request
DATASET_URL = 'http://www.csie.ntu.edu.tw/~cjlin/libsvmtools/datasets/binary/diabetes'
TARGET_PATH = os.path.dirname(os.path.realpath(__file__)) + '/dataset.txt'
def main():
urllib.request.urlretrieve(DATASET_URL, TARGET_PATH)
if __name__ == '__main__': main()
<commit_msg>Change dataset used in example (letter)
XXX: UncertaintySampling(le) weird?<commit_after>
|
#!/usr/bin/env python3
#
# This script is used for downloading the dataset used by the examples.
# Dataset used: Statlog / Letter (in libsvm format)
import os
import urllib.request
import random
DATASET_URL = 'http://www.csie.ntu.edu.tw/~cjlin/libsvmtools/datasets/multiclass/letter.scale'
DATASET_SIZE = 1000
TARGET_PATH = os.path.dirname(os.path.realpath(__file__)) + '/dataset.txt'
def main():
rows = list(urllib.request.urlopen(DATASET_URL))
selected = random.sample(rows, DATASET_SIZE)
with open(TARGET_PATH, 'wb') as f:
for row in selected:
f.write(row)
if __name__ == '__main__': main()
|
#!/usr/bin/env python3
#
# This script is used for downloading the dataset used by the examples.
# Dataset used: UCI / Pima Indians Diabetes (in libsvm format)
import os
import urllib.request
DATASET_URL = 'http://www.csie.ntu.edu.tw/~cjlin/libsvmtools/datasets/binary/diabetes'
TARGET_PATH = os.path.dirname(os.path.realpath(__file__)) + '/dataset.txt'
def main():
urllib.request.urlretrieve(DATASET_URL, TARGET_PATH)
if __name__ == '__main__': main()
Change dataset used in example (letter)
XXX: UncertaintySampling(le) weird?#!/usr/bin/env python3
#
# This script is used for downloading the dataset used by the examples.
# Dataset used: Statlog / Letter (in libsvm format)
import os
import urllib.request
import random
DATASET_URL = 'http://www.csie.ntu.edu.tw/~cjlin/libsvmtools/datasets/multiclass/letter.scale'
DATASET_SIZE = 1000
TARGET_PATH = os.path.dirname(os.path.realpath(__file__)) + '/dataset.txt'
def main():
rows = list(urllib.request.urlopen(DATASET_URL))
selected = random.sample(rows, DATASET_SIZE)
with open(TARGET_PATH, 'wb') as f:
for row in selected:
f.write(row)
if __name__ == '__main__': main()
|
<commit_before>#!/usr/bin/env python3
#
# This script is used for downloading the dataset used by the examples.
# Dataset used: UCI / Pima Indians Diabetes (in libsvm format)
import os
import urllib.request
DATASET_URL = 'http://www.csie.ntu.edu.tw/~cjlin/libsvmtools/datasets/binary/diabetes'
TARGET_PATH = os.path.dirname(os.path.realpath(__file__)) + '/dataset.txt'
def main():
urllib.request.urlretrieve(DATASET_URL, TARGET_PATH)
if __name__ == '__main__': main()
<commit_msg>Change dataset used in example (letter)
XXX: UncertaintySampling(le) weird?<commit_after>#!/usr/bin/env python3
#
# This script is used for downloading the dataset used by the examples.
# Dataset used: Statlog / Letter (in libsvm format)
import os
import urllib.request
import random
DATASET_URL = 'http://www.csie.ntu.edu.tw/~cjlin/libsvmtools/datasets/multiclass/letter.scale'
DATASET_SIZE = 1000
TARGET_PATH = os.path.dirname(os.path.realpath(__file__)) + '/dataset.txt'
def main():
rows = list(urllib.request.urlopen(DATASET_URL))
selected = random.sample(rows, DATASET_SIZE)
with open(TARGET_PATH, 'wb') as f:
for row in selected:
f.write(row)
if __name__ == '__main__': main()
|
32a44354c0a5421c2b8a8ab9d63a26e36ddd6158
|
sponsorship_switzerland/migrations/12.0.1.0.2/pre-migration.py
|
sponsorship_switzerland/migrations/12.0.1.0.2/pre-migration.py
|
from openupgradelib import openupgrade
@openupgrade.migrate(use_env=True)
def migrate(env, version):
if not version:
return
# Associate already created toilets fund to new xml record
covid_fund = env["product.template"].search(
[("default_code", "=", "toilet")]
)
if covid_fund:
openupgrade.add_xmlid(
env.cr,
"sponsorship_switzerland",
"product_template_covid",
"product.template",
covid_fund.id,
)
|
from openupgradelib import openupgrade
@openupgrade.migrate(use_env=True)
def migrate(env, version):
if not version:
return
# Associate already created toilets fund to new xml record
covid_fund = env["product.template"].search(
[("default_code", "=", "coronavirus")]
)
if covid_fund:
openupgrade.add_xmlid(
env.cr,
"sponsorship_switzerland",
"product_template_covid",
"product.template",
covid_fund.id,
)
|
Fix migration of covid product
|
Fix migration of covid product
|
Python
|
agpl-3.0
|
eicher31/compassion-switzerland,eicher31/compassion-switzerland,CompassionCH/compassion-switzerland,eicher31/compassion-switzerland,CompassionCH/compassion-switzerland,CompassionCH/compassion-switzerland
|
from openupgradelib import openupgrade
@openupgrade.migrate(use_env=True)
def migrate(env, version):
if not version:
return
# Associate already created toilets fund to new xml record
covid_fund = env["product.template"].search(
[("default_code", "=", "toilet")]
)
if covid_fund:
openupgrade.add_xmlid(
env.cr,
"sponsorship_switzerland",
"product_template_covid",
"product.template",
covid_fund.id,
)
Fix migration of covid product
|
from openupgradelib import openupgrade
@openupgrade.migrate(use_env=True)
def migrate(env, version):
if not version:
return
# Associate already created toilets fund to new xml record
covid_fund = env["product.template"].search(
[("default_code", "=", "coronavirus")]
)
if covid_fund:
openupgrade.add_xmlid(
env.cr,
"sponsorship_switzerland",
"product_template_covid",
"product.template",
covid_fund.id,
)
|
<commit_before>from openupgradelib import openupgrade
@openupgrade.migrate(use_env=True)
def migrate(env, version):
if not version:
return
# Associate already created toilets fund to new xml record
covid_fund = env["product.template"].search(
[("default_code", "=", "toilet")]
)
if covid_fund:
openupgrade.add_xmlid(
env.cr,
"sponsorship_switzerland",
"product_template_covid",
"product.template",
covid_fund.id,
)
<commit_msg>Fix migration of covid product<commit_after>
|
from openupgradelib import openupgrade
@openupgrade.migrate(use_env=True)
def migrate(env, version):
if not version:
return
# Associate already created toilets fund to new xml record
covid_fund = env["product.template"].search(
[("default_code", "=", "coronavirus")]
)
if covid_fund:
openupgrade.add_xmlid(
env.cr,
"sponsorship_switzerland",
"product_template_covid",
"product.template",
covid_fund.id,
)
|
from openupgradelib import openupgrade
@openupgrade.migrate(use_env=True)
def migrate(env, version):
if not version:
return
# Associate already created toilets fund to new xml record
covid_fund = env["product.template"].search(
[("default_code", "=", "toilet")]
)
if covid_fund:
openupgrade.add_xmlid(
env.cr,
"sponsorship_switzerland",
"product_template_covid",
"product.template",
covid_fund.id,
)
Fix migration of covid productfrom openupgradelib import openupgrade
@openupgrade.migrate(use_env=True)
def migrate(env, version):
if not version:
return
# Associate already created toilets fund to new xml record
covid_fund = env["product.template"].search(
[("default_code", "=", "coronavirus")]
)
if covid_fund:
openupgrade.add_xmlid(
env.cr,
"sponsorship_switzerland",
"product_template_covid",
"product.template",
covid_fund.id,
)
|
<commit_before>from openupgradelib import openupgrade
@openupgrade.migrate(use_env=True)
def migrate(env, version):
if not version:
return
# Associate already created toilets fund to new xml record
covid_fund = env["product.template"].search(
[("default_code", "=", "toilet")]
)
if covid_fund:
openupgrade.add_xmlid(
env.cr,
"sponsorship_switzerland",
"product_template_covid",
"product.template",
covid_fund.id,
)
<commit_msg>Fix migration of covid product<commit_after>from openupgradelib import openupgrade
@openupgrade.migrate(use_env=True)
def migrate(env, version):
if not version:
return
# Associate already created toilets fund to new xml record
covid_fund = env["product.template"].search(
[("default_code", "=", "coronavirus")]
)
if covid_fund:
openupgrade.add_xmlid(
env.cr,
"sponsorship_switzerland",
"product_template_covid",
"product.template",
covid_fund.id,
)
|
8923d10fc831afe7ade5dad4e14167f3525396b6
|
scripts/nipy_4dto3D.py
|
scripts/nipy_4dto3D.py
|
#!/usr/bin/env python
''' Tiny script to write 4D files in any format that we read (nifti,
analyze, MINC, at the moment, as nifti 3D files '''
import os
import sys
import nipy.io.imageformats as nii
if __name__ == '__main__':
try:
fname = sys.argv[1]
except IndexError:
raise OSError('Expecting 4d image filename')
img = nii.load(fname)
imgs = nii.four_to_three(img)
froot, ext = os.path.splitext(fname)
if ext in ('.gz', '.bz2'):
froot, ext = os.path.splitext(froot)
for i, img3d in enumerate(imgs):
fname3d = '%s_%04d.nii' % (froot, i)
nii.save(img3d, fname3d)
|
#!/usr/bin/env python
''' Tiny script to write 4D files in any format that we read (nifti,
analyze, MINC, at the moment, as nifti 3D files '''
import os
import nipy.externals.argparse as argparse
import nipy.io.imageformats as nii
def main():
# create the parser
parser = argparse.ArgumentParser()
# add the arguments
parser.add_argument('filename', type=str,
help='4D image filename')
# parse the command line
args = parser.parse_args()
img = nii.load(args.filename)
imgs = nii.four_to_three(img)
froot, ext = os.path.splitext(args.filename)
if ext in ('.gz', '.bz2'):
froot, ext = os.path.splitext(froot)
for i, img3d in enumerate(imgs):
fname3d = '%s_%04d.nii' % (froot, i)
nii.save(img3d, fname3d)
if __name__ == '__main__':
main()
|
Use argparse for 4D to 3D
|
Use argparse for 4D to 3D
|
Python
|
bsd-3-clause
|
yarikoptic/NiPy-OLD,yarikoptic/NiPy-OLD
|
#!/usr/bin/env python
''' Tiny script to write 4D files in any format that we read (nifti,
analyze, MINC, at the moment, as nifti 3D files '''
import os
import sys
import nipy.io.imageformats as nii
if __name__ == '__main__':
try:
fname = sys.argv[1]
except IndexError:
raise OSError('Expecting 4d image filename')
img = nii.load(fname)
imgs = nii.four_to_three(img)
froot, ext = os.path.splitext(fname)
if ext in ('.gz', '.bz2'):
froot, ext = os.path.splitext(froot)
for i, img3d in enumerate(imgs):
fname3d = '%s_%04d.nii' % (froot, i)
nii.save(img3d, fname3d)
Use argparse for 4D to 3D
|
#!/usr/bin/env python
''' Tiny script to write 4D files in any format that we read (nifti,
analyze, MINC, at the moment, as nifti 3D files '''
import os
import nipy.externals.argparse as argparse
import nipy.io.imageformats as nii
def main():
# create the parser
parser = argparse.ArgumentParser()
# add the arguments
parser.add_argument('filename', type=str,
help='4D image filename')
# parse the command line
args = parser.parse_args()
img = nii.load(args.filename)
imgs = nii.four_to_three(img)
froot, ext = os.path.splitext(args.filename)
if ext in ('.gz', '.bz2'):
froot, ext = os.path.splitext(froot)
for i, img3d in enumerate(imgs):
fname3d = '%s_%04d.nii' % (froot, i)
nii.save(img3d, fname3d)
if __name__ == '__main__':
main()
|
<commit_before>#!/usr/bin/env python
''' Tiny script to write 4D files in any format that we read (nifti,
analyze, MINC, at the moment, as nifti 3D files '''
import os
import sys
import nipy.io.imageformats as nii
if __name__ == '__main__':
try:
fname = sys.argv[1]
except IndexError:
raise OSError('Expecting 4d image filename')
img = nii.load(fname)
imgs = nii.four_to_three(img)
froot, ext = os.path.splitext(fname)
if ext in ('.gz', '.bz2'):
froot, ext = os.path.splitext(froot)
for i, img3d in enumerate(imgs):
fname3d = '%s_%04d.nii' % (froot, i)
nii.save(img3d, fname3d)
<commit_msg>Use argparse for 4D to 3D<commit_after>
|
#!/usr/bin/env python
''' Tiny script to write 4D files in any format that we read (nifti,
analyze, MINC, at the moment, as nifti 3D files '''
import os
import nipy.externals.argparse as argparse
import nipy.io.imageformats as nii
def main():
# create the parser
parser = argparse.ArgumentParser()
# add the arguments
parser.add_argument('filename', type=str,
help='4D image filename')
# parse the command line
args = parser.parse_args()
img = nii.load(args.filename)
imgs = nii.four_to_three(img)
froot, ext = os.path.splitext(args.filename)
if ext in ('.gz', '.bz2'):
froot, ext = os.path.splitext(froot)
for i, img3d in enumerate(imgs):
fname3d = '%s_%04d.nii' % (froot, i)
nii.save(img3d, fname3d)
if __name__ == '__main__':
main()
|
#!/usr/bin/env python
''' Tiny script to write 4D files in any format that we read (nifti,
analyze, MINC, at the moment, as nifti 3D files '''
import os
import sys
import nipy.io.imageformats as nii
if __name__ == '__main__':
try:
fname = sys.argv[1]
except IndexError:
raise OSError('Expecting 4d image filename')
img = nii.load(fname)
imgs = nii.four_to_three(img)
froot, ext = os.path.splitext(fname)
if ext in ('.gz', '.bz2'):
froot, ext = os.path.splitext(froot)
for i, img3d in enumerate(imgs):
fname3d = '%s_%04d.nii' % (froot, i)
nii.save(img3d, fname3d)
Use argparse for 4D to 3D#!/usr/bin/env python
''' Tiny script to write 4D files in any format that we read (nifti,
analyze, MINC, at the moment, as nifti 3D files '''
import os
import nipy.externals.argparse as argparse
import nipy.io.imageformats as nii
def main():
# create the parser
parser = argparse.ArgumentParser()
# add the arguments
parser.add_argument('filename', type=str,
help='4D image filename')
# parse the command line
args = parser.parse_args()
img = nii.load(args.filename)
imgs = nii.four_to_three(img)
froot, ext = os.path.splitext(args.filename)
if ext in ('.gz', '.bz2'):
froot, ext = os.path.splitext(froot)
for i, img3d in enumerate(imgs):
fname3d = '%s_%04d.nii' % (froot, i)
nii.save(img3d, fname3d)
if __name__ == '__main__':
main()
|
<commit_before>#!/usr/bin/env python
''' Tiny script to write 4D files in any format that we read (nifti,
analyze, MINC, at the moment, as nifti 3D files '''
import os
import sys
import nipy.io.imageformats as nii
if __name__ == '__main__':
try:
fname = sys.argv[1]
except IndexError:
raise OSError('Expecting 4d image filename')
img = nii.load(fname)
imgs = nii.four_to_three(img)
froot, ext = os.path.splitext(fname)
if ext in ('.gz', '.bz2'):
froot, ext = os.path.splitext(froot)
for i, img3d in enumerate(imgs):
fname3d = '%s_%04d.nii' % (froot, i)
nii.save(img3d, fname3d)
<commit_msg>Use argparse for 4D to 3D<commit_after>#!/usr/bin/env python
''' Tiny script to write 4D files in any format that we read (nifti,
analyze, MINC, at the moment, as nifti 3D files '''
import os
import nipy.externals.argparse as argparse
import nipy.io.imageformats as nii
def main():
# create the parser
parser = argparse.ArgumentParser()
# add the arguments
parser.add_argument('filename', type=str,
help='4D image filename')
# parse the command line
args = parser.parse_args()
img = nii.load(args.filename)
imgs = nii.four_to_three(img)
froot, ext = os.path.splitext(args.filename)
if ext in ('.gz', '.bz2'):
froot, ext = os.path.splitext(froot)
for i, img3d in enumerate(imgs):
fname3d = '%s_%04d.nii' % (froot, i)
nii.save(img3d, fname3d)
if __name__ == '__main__':
main()
|
ffff9d10862391289e4fba8ac120983ac6368200
|
setup.py
|
setup.py
|
from setuptools import setup
setup(
name='cmsplugin-biography',
version='0.0.1',
packages=['cmsplugin_biography', 'cmsplugin_biography.migrations', ],
install_requires=[
'django-cms',
'djangocms-text-ckeditor==1.0.9',
'easy-thumbnails==1.2',
],
author='Kevin Richardson',
author_email='kevin@magically.us',
description='A Django CMS plugin that manages and displays biographical information',
long_description=open('README.rst').read(),
license='MIT',
url='http://github.com/kfr2/cmsplugin-biography',
include_package_data=True
)
|
from setuptools import setup
setup(
name='cmsplugin-biography',
version='0.0.1',
packages=['cmsplugin_biography', 'cmsplugin_biography.migrations', ],
install_requires=[
'django-cms',
'djangocms-text-ckeditor==1.0.9',
'easy-thumbnails==1.2',
],
author='Kevin Richardson',
author_email='kevin@magically.us',
description='A Django CMS plugin that manages and displays biographical information',
long_description=open('README.rst').read(),
license='MIT',
url='http://github.com/kfr2/cmsplugin-biography',
include_package_data=True,
zip_safe=False
)
|
Mark package as not zip_safe
|
Mark package as not zip_safe
This package needs access to its templates to function. Thus, the
zip_safe flag has been set to False to tell setuptools to not
install the package's egg as a zip file.
See http://pythonhosted.org/distribute/setuptools.html#setting-the-zip-safe-flag
for further information.
|
Python
|
mit
|
kfr2/cmsplugin-biography
|
from setuptools import setup
setup(
name='cmsplugin-biography',
version='0.0.1',
packages=['cmsplugin_biography', 'cmsplugin_biography.migrations', ],
install_requires=[
'django-cms',
'djangocms-text-ckeditor==1.0.9',
'easy-thumbnails==1.2',
],
author='Kevin Richardson',
author_email='kevin@magically.us',
description='A Django CMS plugin that manages and displays biographical information',
long_description=open('README.rst').read(),
license='MIT',
url='http://github.com/kfr2/cmsplugin-biography',
include_package_data=True
)
Mark package as not zip_safe
This package needs access to its templates to function. Thus, the
zip_safe flag has been set to False to tell setuptools to not
install the package's egg as a zip file.
See http://pythonhosted.org/distribute/setuptools.html#setting-the-zip-safe-flag
for further information.
|
from setuptools import setup
setup(
name='cmsplugin-biography',
version='0.0.1',
packages=['cmsplugin_biography', 'cmsplugin_biography.migrations', ],
install_requires=[
'django-cms',
'djangocms-text-ckeditor==1.0.9',
'easy-thumbnails==1.2',
],
author='Kevin Richardson',
author_email='kevin@magically.us',
description='A Django CMS plugin that manages and displays biographical information',
long_description=open('README.rst').read(),
license='MIT',
url='http://github.com/kfr2/cmsplugin-biography',
include_package_data=True,
zip_safe=False
)
|
<commit_before>from setuptools import setup
setup(
name='cmsplugin-biography',
version='0.0.1',
packages=['cmsplugin_biography', 'cmsplugin_biography.migrations', ],
install_requires=[
'django-cms',
'djangocms-text-ckeditor==1.0.9',
'easy-thumbnails==1.2',
],
author='Kevin Richardson',
author_email='kevin@magically.us',
description='A Django CMS plugin that manages and displays biographical information',
long_description=open('README.rst').read(),
license='MIT',
url='http://github.com/kfr2/cmsplugin-biography',
include_package_data=True
)
<commit_msg>Mark package as not zip_safe
This package needs access to its templates to function. Thus, the
zip_safe flag has been set to False to tell setuptools to not
install the package's egg as a zip file.
See http://pythonhosted.org/distribute/setuptools.html#setting-the-zip-safe-flag
for further information.<commit_after>
|
from setuptools import setup
setup(
name='cmsplugin-biography',
version='0.0.1',
packages=['cmsplugin_biography', 'cmsplugin_biography.migrations', ],
install_requires=[
'django-cms',
'djangocms-text-ckeditor==1.0.9',
'easy-thumbnails==1.2',
],
author='Kevin Richardson',
author_email='kevin@magically.us',
description='A Django CMS plugin that manages and displays biographical information',
long_description=open('README.rst').read(),
license='MIT',
url='http://github.com/kfr2/cmsplugin-biography',
include_package_data=True,
zip_safe=False
)
|
from setuptools import setup
setup(
name='cmsplugin-biography',
version='0.0.1',
packages=['cmsplugin_biography', 'cmsplugin_biography.migrations', ],
install_requires=[
'django-cms',
'djangocms-text-ckeditor==1.0.9',
'easy-thumbnails==1.2',
],
author='Kevin Richardson',
author_email='kevin@magically.us',
description='A Django CMS plugin that manages and displays biographical information',
long_description=open('README.rst').read(),
license='MIT',
url='http://github.com/kfr2/cmsplugin-biography',
include_package_data=True
)
Mark package as not zip_safe
This package needs access to its templates to function. Thus, the
zip_safe flag has been set to False to tell setuptools to not
install the package's egg as a zip file.
See http://pythonhosted.org/distribute/setuptools.html#setting-the-zip-safe-flag
for further information.from setuptools import setup
setup(
name='cmsplugin-biography',
version='0.0.1',
packages=['cmsplugin_biography', 'cmsplugin_biography.migrations', ],
install_requires=[
'django-cms',
'djangocms-text-ckeditor==1.0.9',
'easy-thumbnails==1.2',
],
author='Kevin Richardson',
author_email='kevin@magically.us',
description='A Django CMS plugin that manages and displays biographical information',
long_description=open('README.rst').read(),
license='MIT',
url='http://github.com/kfr2/cmsplugin-biography',
include_package_data=True,
zip_safe=False
)
|
<commit_before>from setuptools import setup
setup(
name='cmsplugin-biography',
version='0.0.1',
packages=['cmsplugin_biography', 'cmsplugin_biography.migrations', ],
install_requires=[
'django-cms',
'djangocms-text-ckeditor==1.0.9',
'easy-thumbnails==1.2',
],
author='Kevin Richardson',
author_email='kevin@magically.us',
description='A Django CMS plugin that manages and displays biographical information',
long_description=open('README.rst').read(),
license='MIT',
url='http://github.com/kfr2/cmsplugin-biography',
include_package_data=True
)
<commit_msg>Mark package as not zip_safe
This package needs access to its templates to function. Thus, the
zip_safe flag has been set to False to tell setuptools to not
install the package's egg as a zip file.
See http://pythonhosted.org/distribute/setuptools.html#setting-the-zip-safe-flag
for further information.<commit_after>from setuptools import setup
setup(
name='cmsplugin-biography',
version='0.0.1',
packages=['cmsplugin_biography', 'cmsplugin_biography.migrations', ],
install_requires=[
'django-cms',
'djangocms-text-ckeditor==1.0.9',
'easy-thumbnails==1.2',
],
author='Kevin Richardson',
author_email='kevin@magically.us',
description='A Django CMS plugin that manages and displays biographical information',
long_description=open('README.rst').read(),
license='MIT',
url='http://github.com/kfr2/cmsplugin-biography',
include_package_data=True,
zip_safe=False
)
|
427d3625f26b4a7f3533162e949ed941fa3fe89e
|
setup.py
|
setup.py
|
from setuptools import setup, find_packages
setup(
name='panoptescli',
version='1.1-pre',
url='https://github.com/zooniverse/panoptes-cli',
author='Adam McMaster',
author_email='adam@zooniverse.org',
description=(
'A command-line client for Panoptes, the API behind the Zooniverse'
),
packages=find_packages(),
include_package_data=True,
install_requires=[
'Click>=6.7,<6.8',
'PyYAML>=3.12,<5.2',
'panoptes-client>=1.0,<2.0',
],
entry_points='''
[console_scripts]
panoptes=panoptes_cli.scripts.panoptes:cli
''',
)
|
from setuptools import setup, find_packages
setup(
name='panoptescli',
version='1.1-pre',
url='https://github.com/zooniverse/panoptes-cli',
author='Adam McMaster',
author_email='adam@zooniverse.org',
description=(
'A command-line client for Panoptes, the API behind the Zooniverse'
),
packages=find_packages(),
include_package_data=True,
install_requires=[
'Click>=6.7,<7.1',
'PyYAML>=3.12,<5.2',
'panoptes-client>=1.0,<2.0',
],
entry_points='''
[console_scripts]
panoptes=panoptes_cli.scripts.panoptes:cli
''',
)
|
Update click requirement from <6.8,>=6.7 to >=6.7,<7.1
|
Update click requirement from <6.8,>=6.7 to >=6.7,<7.1
Updates the requirements on [click](https://github.com/pallets/click) to permit the latest version.
- [Release notes](https://github.com/pallets/click/releases)
- [Changelog](https://github.com/pallets/click/blob/master/docs/changelog.rst)
- [Commits](https://github.com/pallets/click/commits/7.0)
Signed-off-by: dependabot[bot] <5bdcd3c0d4d24ae3e71b3b452a024c6324c7e4bb@dependabot.com>
|
Python
|
apache-2.0
|
zooniverse/panoptes-cli
|
from setuptools import setup, find_packages
setup(
name='panoptescli',
version='1.1-pre',
url='https://github.com/zooniverse/panoptes-cli',
author='Adam McMaster',
author_email='adam@zooniverse.org',
description=(
'A command-line client for Panoptes, the API behind the Zooniverse'
),
packages=find_packages(),
include_package_data=True,
install_requires=[
'Click>=6.7,<6.8',
'PyYAML>=3.12,<5.2',
'panoptes-client>=1.0,<2.0',
],
entry_points='''
[console_scripts]
panoptes=panoptes_cli.scripts.panoptes:cli
''',
)
Update click requirement from <6.8,>=6.7 to >=6.7,<7.1
Updates the requirements on [click](https://github.com/pallets/click) to permit the latest version.
- [Release notes](https://github.com/pallets/click/releases)
- [Changelog](https://github.com/pallets/click/blob/master/docs/changelog.rst)
- [Commits](https://github.com/pallets/click/commits/7.0)
Signed-off-by: dependabot[bot] <5bdcd3c0d4d24ae3e71b3b452a024c6324c7e4bb@dependabot.com>
|
from setuptools import setup, find_packages
setup(
name='panoptescli',
version='1.1-pre',
url='https://github.com/zooniverse/panoptes-cli',
author='Adam McMaster',
author_email='adam@zooniverse.org',
description=(
'A command-line client for Panoptes, the API behind the Zooniverse'
),
packages=find_packages(),
include_package_data=True,
install_requires=[
'Click>=6.7,<7.1',
'PyYAML>=3.12,<5.2',
'panoptes-client>=1.0,<2.0',
],
entry_points='''
[console_scripts]
panoptes=panoptes_cli.scripts.panoptes:cli
''',
)
|
<commit_before>from setuptools import setup, find_packages
setup(
name='panoptescli',
version='1.1-pre',
url='https://github.com/zooniverse/panoptes-cli',
author='Adam McMaster',
author_email='adam@zooniverse.org',
description=(
'A command-line client for Panoptes, the API behind the Zooniverse'
),
packages=find_packages(),
include_package_data=True,
install_requires=[
'Click>=6.7,<6.8',
'PyYAML>=3.12,<5.2',
'panoptes-client>=1.0,<2.0',
],
entry_points='''
[console_scripts]
panoptes=panoptes_cli.scripts.panoptes:cli
''',
)
<commit_msg>Update click requirement from <6.8,>=6.7 to >=6.7,<7.1
Updates the requirements on [click](https://github.com/pallets/click) to permit the latest version.
- [Release notes](https://github.com/pallets/click/releases)
- [Changelog](https://github.com/pallets/click/blob/master/docs/changelog.rst)
- [Commits](https://github.com/pallets/click/commits/7.0)
Signed-off-by: dependabot[bot] <5bdcd3c0d4d24ae3e71b3b452a024c6324c7e4bb@dependabot.com><commit_after>
|
from setuptools import setup, find_packages
setup(
name='panoptescli',
version='1.1-pre',
url='https://github.com/zooniverse/panoptes-cli',
author='Adam McMaster',
author_email='adam@zooniverse.org',
description=(
'A command-line client for Panoptes, the API behind the Zooniverse'
),
packages=find_packages(),
include_package_data=True,
install_requires=[
'Click>=6.7,<7.1',
'PyYAML>=3.12,<5.2',
'panoptes-client>=1.0,<2.0',
],
entry_points='''
[console_scripts]
panoptes=panoptes_cli.scripts.panoptes:cli
''',
)
|
from setuptools import setup, find_packages
setup(
name='panoptescli',
version='1.1-pre',
url='https://github.com/zooniverse/panoptes-cli',
author='Adam McMaster',
author_email='adam@zooniverse.org',
description=(
'A command-line client for Panoptes, the API behind the Zooniverse'
),
packages=find_packages(),
include_package_data=True,
install_requires=[
'Click>=6.7,<6.8',
'PyYAML>=3.12,<5.2',
'panoptes-client>=1.0,<2.0',
],
entry_points='''
[console_scripts]
panoptes=panoptes_cli.scripts.panoptes:cli
''',
)
Update click requirement from <6.8,>=6.7 to >=6.7,<7.1
Updates the requirements on [click](https://github.com/pallets/click) to permit the latest version.
- [Release notes](https://github.com/pallets/click/releases)
- [Changelog](https://github.com/pallets/click/blob/master/docs/changelog.rst)
- [Commits](https://github.com/pallets/click/commits/7.0)
Signed-off-by: dependabot[bot] <5bdcd3c0d4d24ae3e71b3b452a024c6324c7e4bb@dependabot.com>from setuptools import setup, find_packages
setup(
name='panoptescli',
version='1.1-pre',
url='https://github.com/zooniverse/panoptes-cli',
author='Adam McMaster',
author_email='adam@zooniverse.org',
description=(
'A command-line client for Panoptes, the API behind the Zooniverse'
),
packages=find_packages(),
include_package_data=True,
install_requires=[
'Click>=6.7,<7.1',
'PyYAML>=3.12,<5.2',
'panoptes-client>=1.0,<2.0',
],
entry_points='''
[console_scripts]
panoptes=panoptes_cli.scripts.panoptes:cli
''',
)
|
<commit_before>from setuptools import setup, find_packages
setup(
name='panoptescli',
version='1.1-pre',
url='https://github.com/zooniverse/panoptes-cli',
author='Adam McMaster',
author_email='adam@zooniverse.org',
description=(
'A command-line client for Panoptes, the API behind the Zooniverse'
),
packages=find_packages(),
include_package_data=True,
install_requires=[
'Click>=6.7,<6.8',
'PyYAML>=3.12,<5.2',
'panoptes-client>=1.0,<2.0',
],
entry_points='''
[console_scripts]
panoptes=panoptes_cli.scripts.panoptes:cli
''',
)
<commit_msg>Update click requirement from <6.8,>=6.7 to >=6.7,<7.1
Updates the requirements on [click](https://github.com/pallets/click) to permit the latest version.
- [Release notes](https://github.com/pallets/click/releases)
- [Changelog](https://github.com/pallets/click/blob/master/docs/changelog.rst)
- [Commits](https://github.com/pallets/click/commits/7.0)
Signed-off-by: dependabot[bot] <5bdcd3c0d4d24ae3e71b3b452a024c6324c7e4bb@dependabot.com><commit_after>from setuptools import setup, find_packages
setup(
name='panoptescli',
version='1.1-pre',
url='https://github.com/zooniverse/panoptes-cli',
author='Adam McMaster',
author_email='adam@zooniverse.org',
description=(
'A command-line client for Panoptes, the API behind the Zooniverse'
),
packages=find_packages(),
include_package_data=True,
install_requires=[
'Click>=6.7,<7.1',
'PyYAML>=3.12,<5.2',
'panoptes-client>=1.0,<2.0',
],
entry_points='''
[console_scripts]
panoptes=panoptes_cli.scripts.panoptes:cli
''',
)
|
ebb3ea0d72835c4acdc38ba241cf8fd4f828c5cd
|
setup.py
|
setup.py
|
from distutils.core import setup, Extension
import sys
ext_modules = [
Extension('classified._platform',
['src/classified._platform.c'],
extra_compile_args=[
'-DPLATFORM_%s' % (sys.platform.upper()),
'-Wunused',
]
)
]
setup(
name = 'classified',
version = '0.0.2',
author = 'Wijnand Modderman',
author_email = 'maze@pyth0n.org',
description = 'Classified data scanner',
license = 'MIT',
keywords = 'classified sensitive pan pci',
packages = [
'classified',
'classified.probe',
],
data_files = [
('/etc/classified', 'etc/classified.conf.sample'),
],
scripts = ['bin/classified'],
ext_modules = ext_modules,
)
|
from distutils.core import setup, Extension
import sys
ext_modules = [
Extension('classified._platform',
['src/classified._platform.c'],
extra_compile_args=[
'-DPLATFORM_%s' % (sys.platform.upper()),
'-Wunused',
]
)
]
setup(
name = 'classified',
version = '0.0.2',
author = 'Wijnand Modderman',
author_email = 'maze@pyth0n.org',
description = 'Classified data scanner',
license = 'MIT',
keywords = 'classified sensitive pan pci',
packages = [
'classified',
'classified.probe',
'classified.probe.pan',
'classified.probe.password',
'classified.probe.ssl',
],
data_files = [
('/etc/classified', 'etc/classified.conf.sample'),
],
scripts = ['bin/classified'],
ext_modules = ext_modules,
)
|
Move probes to their own directory
|
Move probes to their own directory
|
Python
|
mit
|
tehmaze/classified,tehmaze/classified,tehmaze/classified
|
from distutils.core import setup, Extension
import sys
ext_modules = [
Extension('classified._platform',
['src/classified._platform.c'],
extra_compile_args=[
'-DPLATFORM_%s' % (sys.platform.upper()),
'-Wunused',
]
)
]
setup(
name = 'classified',
version = '0.0.2',
author = 'Wijnand Modderman',
author_email = 'maze@pyth0n.org',
description = 'Classified data scanner',
license = 'MIT',
keywords = 'classified sensitive pan pci',
packages = [
'classified',
'classified.probe',
],
data_files = [
('/etc/classified', 'etc/classified.conf.sample'),
],
scripts = ['bin/classified'],
ext_modules = ext_modules,
)
Move probes to their own directory
|
from distutils.core import setup, Extension
import sys
ext_modules = [
Extension('classified._platform',
['src/classified._platform.c'],
extra_compile_args=[
'-DPLATFORM_%s' % (sys.platform.upper()),
'-Wunused',
]
)
]
setup(
name = 'classified',
version = '0.0.2',
author = 'Wijnand Modderman',
author_email = 'maze@pyth0n.org',
description = 'Classified data scanner',
license = 'MIT',
keywords = 'classified sensitive pan pci',
packages = [
'classified',
'classified.probe',
'classified.probe.pan',
'classified.probe.password',
'classified.probe.ssl',
],
data_files = [
('/etc/classified', 'etc/classified.conf.sample'),
],
scripts = ['bin/classified'],
ext_modules = ext_modules,
)
|
<commit_before>from distutils.core import setup, Extension
import sys
ext_modules = [
Extension('classified._platform',
['src/classified._platform.c'],
extra_compile_args=[
'-DPLATFORM_%s' % (sys.platform.upper()),
'-Wunused',
]
)
]
setup(
name = 'classified',
version = '0.0.2',
author = 'Wijnand Modderman',
author_email = 'maze@pyth0n.org',
description = 'Classified data scanner',
license = 'MIT',
keywords = 'classified sensitive pan pci',
packages = [
'classified',
'classified.probe',
],
data_files = [
('/etc/classified', 'etc/classified.conf.sample'),
],
scripts = ['bin/classified'],
ext_modules = ext_modules,
)
<commit_msg>Move probes to their own directory<commit_after>
|
from distutils.core import setup, Extension
import sys
ext_modules = [
Extension('classified._platform',
['src/classified._platform.c'],
extra_compile_args=[
'-DPLATFORM_%s' % (sys.platform.upper()),
'-Wunused',
]
)
]
setup(
name = 'classified',
version = '0.0.2',
author = 'Wijnand Modderman',
author_email = 'maze@pyth0n.org',
description = 'Classified data scanner',
license = 'MIT',
keywords = 'classified sensitive pan pci',
packages = [
'classified',
'classified.probe',
'classified.probe.pan',
'classified.probe.password',
'classified.probe.ssl',
],
data_files = [
('/etc/classified', 'etc/classified.conf.sample'),
],
scripts = ['bin/classified'],
ext_modules = ext_modules,
)
|
from distutils.core import setup, Extension
import sys
ext_modules = [
Extension('classified._platform',
['src/classified._platform.c'],
extra_compile_args=[
'-DPLATFORM_%s' % (sys.platform.upper()),
'-Wunused',
]
)
]
setup(
name = 'classified',
version = '0.0.2',
author = 'Wijnand Modderman',
author_email = 'maze@pyth0n.org',
description = 'Classified data scanner',
license = 'MIT',
keywords = 'classified sensitive pan pci',
packages = [
'classified',
'classified.probe',
],
data_files = [
('/etc/classified', 'etc/classified.conf.sample'),
],
scripts = ['bin/classified'],
ext_modules = ext_modules,
)
Move probes to their own directoryfrom distutils.core import setup, Extension
import sys
ext_modules = [
Extension('classified._platform',
['src/classified._platform.c'],
extra_compile_args=[
'-DPLATFORM_%s' % (sys.platform.upper()),
'-Wunused',
]
)
]
setup(
name = 'classified',
version = '0.0.2',
author = 'Wijnand Modderman',
author_email = 'maze@pyth0n.org',
description = 'Classified data scanner',
license = 'MIT',
keywords = 'classified sensitive pan pci',
packages = [
'classified',
'classified.probe',
'classified.probe.pan',
'classified.probe.password',
'classified.probe.ssl',
],
data_files = [
('/etc/classified', 'etc/classified.conf.sample'),
],
scripts = ['bin/classified'],
ext_modules = ext_modules,
)
|
<commit_before>from distutils.core import setup, Extension
import sys
ext_modules = [
Extension('classified._platform',
['src/classified._platform.c'],
extra_compile_args=[
'-DPLATFORM_%s' % (sys.platform.upper()),
'-Wunused',
]
)
]
setup(
name = 'classified',
version = '0.0.2',
author = 'Wijnand Modderman',
author_email = 'maze@pyth0n.org',
description = 'Classified data scanner',
license = 'MIT',
keywords = 'classified sensitive pan pci',
packages = [
'classified',
'classified.probe',
],
data_files = [
('/etc/classified', 'etc/classified.conf.sample'),
],
scripts = ['bin/classified'],
ext_modules = ext_modules,
)
<commit_msg>Move probes to their own directory<commit_after>from distutils.core import setup, Extension
import sys
ext_modules = [
Extension('classified._platform',
['src/classified._platform.c'],
extra_compile_args=[
'-DPLATFORM_%s' % (sys.platform.upper()),
'-Wunused',
]
)
]
setup(
name = 'classified',
version = '0.0.2',
author = 'Wijnand Modderman',
author_email = 'maze@pyth0n.org',
description = 'Classified data scanner',
license = 'MIT',
keywords = 'classified sensitive pan pci',
packages = [
'classified',
'classified.probe',
'classified.probe.pan',
'classified.probe.password',
'classified.probe.ssl',
],
data_files = [
('/etc/classified', 'etc/classified.conf.sample'),
],
scripts = ['bin/classified'],
ext_modules = ext_modules,
)
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.