commit
stringlengths
40
40
old_file
stringlengths
4
118
new_file
stringlengths
4
118
old_contents
stringlengths
0
2.94k
new_contents
stringlengths
1
4.43k
subject
stringlengths
15
444
message
stringlengths
16
3.45k
lang
stringclasses
1 value
license
stringclasses
13 values
repos
stringlengths
5
43.2k
prompt
stringlengths
17
4.58k
response
stringlengths
1
4.43k
prompt_tagged
stringlengths
58
4.62k
response_tagged
stringlengths
1
4.43k
text
stringlengths
132
7.29k
text_tagged
stringlengths
173
7.33k
b729b8bb5a3b65b6e8c09455b0db21e7df59dd8a
setup.py
setup.py
from setuptools import setup, find_packages with open('README.rst') as readme: description = readme.read() setup( name='mfl', version='2.0.0', description='Master falicity list', long_description=(description), author='Savannah Developers', author_email='developers@savannahinformatics.com', url='', packages=find_packages(), install_requires=[ "django>=1.8", "Fabric>=1.10", "coverage>=3.7", "psycopg2>=2.5", "djangorestframework>=3.1", "django-filter>=0.9", "flake8>=2.3", "django-cors-headers>=1.0", "virtualenv>=12.0", "pip>=6.0", "tox>=1.9", "djangorestframework-xml>=1.0", "djangorestframework-csv>=1.3", "model_mommy>=1.2" ], )
from setuptools import setup, find_packages with open('README.rst') as readme: description = readme.read() setup( name='mfl', version='2.0.0', description='Master falicity list', long_description=(description), author='Savannah Developers', author_email='developers@savannahinformatics.com', url='', packages=find_packages(), install_requires=[ "model_mommy>=1.2", "Fabric>=1.10", "coverage>=3.7", "psycopg2>=2.5", "djangorestframework>=3.1", "django-filter>=0.9", "flake8>=2.3", "django-cors-headers>=1.0", "virtualenv>=12.0", "pip>=6.0", "tox>=1.9", "djangorestframework-xml>=1.0", "djangorestframework-csv>=1.3", "django>=1.8", ], )
Fix model_mommy issues with django==1.8
Fix model_mommy issues with django==1.8
Python
mit
urandu/mfl_api,MasterFacilityList/mfl_api,urandu/mfl_api,urandu/mfl_api,MasterFacilityList/mfl_api,MasterFacilityList/mfl_api,MasterFacilityList/mfl_api,MasterFacilityList/mfl_api,urandu/mfl_api
from setuptools import setup, find_packages with open('README.rst') as readme: description = readme.read() setup( name='mfl', version='2.0.0', description='Master falicity list', long_description=(description), author='Savannah Developers', author_email='developers@savannahinformatics.com', url='', packages=find_packages(), install_requires=[ "django>=1.8", "Fabric>=1.10", "coverage>=3.7", "psycopg2>=2.5", "djangorestframework>=3.1", "django-filter>=0.9", "flake8>=2.3", "django-cors-headers>=1.0", "virtualenv>=12.0", "pip>=6.0", "tox>=1.9", "djangorestframework-xml>=1.0", "djangorestframework-csv>=1.3", "model_mommy>=1.2" ], ) Fix model_mommy issues with django==1.8
from setuptools import setup, find_packages with open('README.rst') as readme: description = readme.read() setup( name='mfl', version='2.0.0', description='Master falicity list', long_description=(description), author='Savannah Developers', author_email='developers@savannahinformatics.com', url='', packages=find_packages(), install_requires=[ "model_mommy>=1.2", "Fabric>=1.10", "coverage>=3.7", "psycopg2>=2.5", "djangorestframework>=3.1", "django-filter>=0.9", "flake8>=2.3", "django-cors-headers>=1.0", "virtualenv>=12.0", "pip>=6.0", "tox>=1.9", "djangorestframework-xml>=1.0", "djangorestframework-csv>=1.3", "django>=1.8", ], )
<commit_before>from setuptools import setup, find_packages with open('README.rst') as readme: description = readme.read() setup( name='mfl', version='2.0.0', description='Master falicity list', long_description=(description), author='Savannah Developers', author_email='developers@savannahinformatics.com', url='', packages=find_packages(), install_requires=[ "django>=1.8", "Fabric>=1.10", "coverage>=3.7", "psycopg2>=2.5", "djangorestframework>=3.1", "django-filter>=0.9", "flake8>=2.3", "django-cors-headers>=1.0", "virtualenv>=12.0", "pip>=6.0", "tox>=1.9", "djangorestframework-xml>=1.0", "djangorestframework-csv>=1.3", "model_mommy>=1.2" ], ) <commit_msg>Fix model_mommy issues with django==1.8<commit_after>
from setuptools import setup, find_packages with open('README.rst') as readme: description = readme.read() setup( name='mfl', version='2.0.0', description='Master falicity list', long_description=(description), author='Savannah Developers', author_email='developers@savannahinformatics.com', url='', packages=find_packages(), install_requires=[ "model_mommy>=1.2", "Fabric>=1.10", "coverage>=3.7", "psycopg2>=2.5", "djangorestframework>=3.1", "django-filter>=0.9", "flake8>=2.3", "django-cors-headers>=1.0", "virtualenv>=12.0", "pip>=6.0", "tox>=1.9", "djangorestframework-xml>=1.0", "djangorestframework-csv>=1.3", "django>=1.8", ], )
from setuptools import setup, find_packages with open('README.rst') as readme: description = readme.read() setup( name='mfl', version='2.0.0', description='Master falicity list', long_description=(description), author='Savannah Developers', author_email='developers@savannahinformatics.com', url='', packages=find_packages(), install_requires=[ "django>=1.8", "Fabric>=1.10", "coverage>=3.7", "psycopg2>=2.5", "djangorestframework>=3.1", "django-filter>=0.9", "flake8>=2.3", "django-cors-headers>=1.0", "virtualenv>=12.0", "pip>=6.0", "tox>=1.9", "djangorestframework-xml>=1.0", "djangorestframework-csv>=1.3", "model_mommy>=1.2" ], ) Fix model_mommy issues with django==1.8from setuptools import setup, find_packages with open('README.rst') as readme: description = readme.read() setup( name='mfl', version='2.0.0', description='Master falicity list', long_description=(description), author='Savannah Developers', author_email='developers@savannahinformatics.com', url='', packages=find_packages(), install_requires=[ "model_mommy>=1.2", "Fabric>=1.10", "coverage>=3.7", "psycopg2>=2.5", "djangorestframework>=3.1", "django-filter>=0.9", "flake8>=2.3", "django-cors-headers>=1.0", "virtualenv>=12.0", "pip>=6.0", "tox>=1.9", "djangorestframework-xml>=1.0", "djangorestframework-csv>=1.3", "django>=1.8", ], )
<commit_before>from setuptools import setup, find_packages with open('README.rst') as readme: description = readme.read() setup( name='mfl', version='2.0.0', description='Master falicity list', long_description=(description), author='Savannah Developers', author_email='developers@savannahinformatics.com', url='', packages=find_packages(), install_requires=[ "django>=1.8", "Fabric>=1.10", "coverage>=3.7", "psycopg2>=2.5", "djangorestframework>=3.1", "django-filter>=0.9", "flake8>=2.3", "django-cors-headers>=1.0", "virtualenv>=12.0", "pip>=6.0", "tox>=1.9", "djangorestframework-xml>=1.0", "djangorestframework-csv>=1.3", "model_mommy>=1.2" ], ) <commit_msg>Fix model_mommy issues with django==1.8<commit_after>from setuptools import setup, find_packages with open('README.rst') as readme: description = readme.read() setup( name='mfl', version='2.0.0', description='Master falicity list', long_description=(description), author='Savannah Developers', author_email='developers@savannahinformatics.com', url='', packages=find_packages(), install_requires=[ "model_mommy>=1.2", "Fabric>=1.10", "coverage>=3.7", "psycopg2>=2.5", "djangorestframework>=3.1", "django-filter>=0.9", "flake8>=2.3", "django-cors-headers>=1.0", "virtualenv>=12.0", "pip>=6.0", "tox>=1.9", "djangorestframework-xml>=1.0", "djangorestframework-csv>=1.3", "django>=1.8", ], )
dfb53cd63c908f13dafcc159ce337af653523748
datasets/forms.py
datasets/forms.py
from django import forms from datasets.models import DatasetRelease, CategoryComment class DatasetReleaseForm(forms.ModelForm): max_number_of_sounds = forms.IntegerField(required=False) class Meta: model = DatasetRelease fields = ['release_tag', 'type'] class PresentNotPresentUnsureForm(forms.Form): vote = forms.ChoiceField( required=True, widget=forms.RadioSelect, choices=( ('1', 'Present and predominant',), ('0.5', 'Present but not predominant',), ('-1', 'Not Present',), ('0', 'Unsure',), ), ) annotation_id = forms.IntegerField( required=True, widget=forms.HiddenInput, ) visited_sound = forms.BooleanField( required=False, initial=False, widget=forms.HiddenInput, ) class CategoryCommentForm(forms.ModelForm): class Meta: model = CategoryComment fields = ['comment', 'category_id', 'dataset'] widgets = { 'comment': forms.Textarea(attrs={ 'cols': 80, 'rows': 3, 'placeholder': 'Add here any general comments you want to make about this category'}), 'category_id': forms.HiddenInput, 'dataset_id': forms.HiddenInput, }
from django import forms from datasets.models import DatasetRelease, CategoryComment class DatasetReleaseForm(forms.ModelForm): max_number_of_sounds = forms.IntegerField(required=False) class Meta: model = DatasetRelease fields = ['release_tag', 'type'] class PresentNotPresentUnsureForm(forms.Form): vote = forms.ChoiceField( required=True, widget=forms.RadioSelect, choices=( ('1', 'Present and predominant',), ('0.5', 'Present but not predominant',), ('-1', 'Not present',), ('0', 'Unsure',), ), ) annotation_id = forms.IntegerField( required=True, widget=forms.HiddenInput, ) visited_sound = forms.BooleanField( required=False, initial=False, widget=forms.HiddenInput, ) class CategoryCommentForm(forms.ModelForm): class Meta: model = CategoryComment fields = ['comment', 'category_id', 'dataset'] widgets = { 'comment': forms.Textarea(attrs={ 'cols': 80, 'rows': 3, 'placeholder': 'Add here any general comments you want to make about this category'}), 'category_id': forms.HiddenInput, 'dataset_id': forms.HiddenInput, }
Remove upper case Not Present
Remove upper case Not Present
Python
agpl-3.0
MTG/freesound-datasets,MTG/freesound-datasets,MTG/freesound-datasets,MTG/freesound-datasets
from django import forms from datasets.models import DatasetRelease, CategoryComment class DatasetReleaseForm(forms.ModelForm): max_number_of_sounds = forms.IntegerField(required=False) class Meta: model = DatasetRelease fields = ['release_tag', 'type'] class PresentNotPresentUnsureForm(forms.Form): vote = forms.ChoiceField( required=True, widget=forms.RadioSelect, choices=( ('1', 'Present and predominant',), ('0.5', 'Present but not predominant',), ('-1', 'Not Present',), ('0', 'Unsure',), ), ) annotation_id = forms.IntegerField( required=True, widget=forms.HiddenInput, ) visited_sound = forms.BooleanField( required=False, initial=False, widget=forms.HiddenInput, ) class CategoryCommentForm(forms.ModelForm): class Meta: model = CategoryComment fields = ['comment', 'category_id', 'dataset'] widgets = { 'comment': forms.Textarea(attrs={ 'cols': 80, 'rows': 3, 'placeholder': 'Add here any general comments you want to make about this category'}), 'category_id': forms.HiddenInput, 'dataset_id': forms.HiddenInput, } Remove upper case Not Present
from django import forms from datasets.models import DatasetRelease, CategoryComment class DatasetReleaseForm(forms.ModelForm): max_number_of_sounds = forms.IntegerField(required=False) class Meta: model = DatasetRelease fields = ['release_tag', 'type'] class PresentNotPresentUnsureForm(forms.Form): vote = forms.ChoiceField( required=True, widget=forms.RadioSelect, choices=( ('1', 'Present and predominant',), ('0.5', 'Present but not predominant',), ('-1', 'Not present',), ('0', 'Unsure',), ), ) annotation_id = forms.IntegerField( required=True, widget=forms.HiddenInput, ) visited_sound = forms.BooleanField( required=False, initial=False, widget=forms.HiddenInput, ) class CategoryCommentForm(forms.ModelForm): class Meta: model = CategoryComment fields = ['comment', 'category_id', 'dataset'] widgets = { 'comment': forms.Textarea(attrs={ 'cols': 80, 'rows': 3, 'placeholder': 'Add here any general comments you want to make about this category'}), 'category_id': forms.HiddenInput, 'dataset_id': forms.HiddenInput, }
<commit_before>from django import forms from datasets.models import DatasetRelease, CategoryComment class DatasetReleaseForm(forms.ModelForm): max_number_of_sounds = forms.IntegerField(required=False) class Meta: model = DatasetRelease fields = ['release_tag', 'type'] class PresentNotPresentUnsureForm(forms.Form): vote = forms.ChoiceField( required=True, widget=forms.RadioSelect, choices=( ('1', 'Present and predominant',), ('0.5', 'Present but not predominant',), ('-1', 'Not Present',), ('0', 'Unsure',), ), ) annotation_id = forms.IntegerField( required=True, widget=forms.HiddenInput, ) visited_sound = forms.BooleanField( required=False, initial=False, widget=forms.HiddenInput, ) class CategoryCommentForm(forms.ModelForm): class Meta: model = CategoryComment fields = ['comment', 'category_id', 'dataset'] widgets = { 'comment': forms.Textarea(attrs={ 'cols': 80, 'rows': 3, 'placeholder': 'Add here any general comments you want to make about this category'}), 'category_id': forms.HiddenInput, 'dataset_id': forms.HiddenInput, } <commit_msg>Remove upper case Not Present<commit_after>
from django import forms from datasets.models import DatasetRelease, CategoryComment class DatasetReleaseForm(forms.ModelForm): max_number_of_sounds = forms.IntegerField(required=False) class Meta: model = DatasetRelease fields = ['release_tag', 'type'] class PresentNotPresentUnsureForm(forms.Form): vote = forms.ChoiceField( required=True, widget=forms.RadioSelect, choices=( ('1', 'Present and predominant',), ('0.5', 'Present but not predominant',), ('-1', 'Not present',), ('0', 'Unsure',), ), ) annotation_id = forms.IntegerField( required=True, widget=forms.HiddenInput, ) visited_sound = forms.BooleanField( required=False, initial=False, widget=forms.HiddenInput, ) class CategoryCommentForm(forms.ModelForm): class Meta: model = CategoryComment fields = ['comment', 'category_id', 'dataset'] widgets = { 'comment': forms.Textarea(attrs={ 'cols': 80, 'rows': 3, 'placeholder': 'Add here any general comments you want to make about this category'}), 'category_id': forms.HiddenInput, 'dataset_id': forms.HiddenInput, }
from django import forms from datasets.models import DatasetRelease, CategoryComment class DatasetReleaseForm(forms.ModelForm): max_number_of_sounds = forms.IntegerField(required=False) class Meta: model = DatasetRelease fields = ['release_tag', 'type'] class PresentNotPresentUnsureForm(forms.Form): vote = forms.ChoiceField( required=True, widget=forms.RadioSelect, choices=( ('1', 'Present and predominant',), ('0.5', 'Present but not predominant',), ('-1', 'Not Present',), ('0', 'Unsure',), ), ) annotation_id = forms.IntegerField( required=True, widget=forms.HiddenInput, ) visited_sound = forms.BooleanField( required=False, initial=False, widget=forms.HiddenInput, ) class CategoryCommentForm(forms.ModelForm): class Meta: model = CategoryComment fields = ['comment', 'category_id', 'dataset'] widgets = { 'comment': forms.Textarea(attrs={ 'cols': 80, 'rows': 3, 'placeholder': 'Add here any general comments you want to make about this category'}), 'category_id': forms.HiddenInput, 'dataset_id': forms.HiddenInput, } Remove upper case Not Presentfrom django import forms from datasets.models import DatasetRelease, CategoryComment class DatasetReleaseForm(forms.ModelForm): max_number_of_sounds = forms.IntegerField(required=False) class Meta: model = DatasetRelease fields = ['release_tag', 'type'] class PresentNotPresentUnsureForm(forms.Form): vote = forms.ChoiceField( required=True, widget=forms.RadioSelect, choices=( ('1', 'Present and predominant',), ('0.5', 'Present but not predominant',), ('-1', 'Not present',), ('0', 'Unsure',), ), ) annotation_id = forms.IntegerField( required=True, widget=forms.HiddenInput, ) visited_sound = forms.BooleanField( required=False, initial=False, widget=forms.HiddenInput, ) class CategoryCommentForm(forms.ModelForm): class Meta: model = CategoryComment fields = ['comment', 'category_id', 'dataset'] widgets = { 'comment': forms.Textarea(attrs={ 'cols': 80, 'rows': 3, 'placeholder': 'Add here any general comments you want to make about this category'}), 'category_id': forms.HiddenInput, 'dataset_id': forms.HiddenInput, }
<commit_before>from django import forms from datasets.models import DatasetRelease, CategoryComment class DatasetReleaseForm(forms.ModelForm): max_number_of_sounds = forms.IntegerField(required=False) class Meta: model = DatasetRelease fields = ['release_tag', 'type'] class PresentNotPresentUnsureForm(forms.Form): vote = forms.ChoiceField( required=True, widget=forms.RadioSelect, choices=( ('1', 'Present and predominant',), ('0.5', 'Present but not predominant',), ('-1', 'Not Present',), ('0', 'Unsure',), ), ) annotation_id = forms.IntegerField( required=True, widget=forms.HiddenInput, ) visited_sound = forms.BooleanField( required=False, initial=False, widget=forms.HiddenInput, ) class CategoryCommentForm(forms.ModelForm): class Meta: model = CategoryComment fields = ['comment', 'category_id', 'dataset'] widgets = { 'comment': forms.Textarea(attrs={ 'cols': 80, 'rows': 3, 'placeholder': 'Add here any general comments you want to make about this category'}), 'category_id': forms.HiddenInput, 'dataset_id': forms.HiddenInput, } <commit_msg>Remove upper case Not Present<commit_after>from django import forms from datasets.models import DatasetRelease, CategoryComment class DatasetReleaseForm(forms.ModelForm): max_number_of_sounds = forms.IntegerField(required=False) class Meta: model = DatasetRelease fields = ['release_tag', 'type'] class PresentNotPresentUnsureForm(forms.Form): vote = forms.ChoiceField( required=True, widget=forms.RadioSelect, choices=( ('1', 'Present and predominant',), ('0.5', 'Present but not predominant',), ('-1', 'Not present',), ('0', 'Unsure',), ), ) annotation_id = forms.IntegerField( required=True, widget=forms.HiddenInput, ) visited_sound = forms.BooleanField( required=False, initial=False, widget=forms.HiddenInput, ) class CategoryCommentForm(forms.ModelForm): class Meta: model = CategoryComment fields = ['comment', 'category_id', 'dataset'] widgets = { 'comment': forms.Textarea(attrs={ 'cols': 80, 'rows': 3, 'placeholder': 'Add here any general comments you want to make about this category'}), 'category_id': forms.HiddenInput, 'dataset_id': forms.HiddenInput, }
bd7cab646b6ead140b62ed19b5378567c76bad9d
setup.py
setup.py
import os from setuptools import setup def read(name): return open(os.path.join(os.path.dirname(__file__), name)).read() setup( name='python-dxf', version='7.5.0', description="Package for accessing a Docker v2 registry", long_description=read('README.rst'), keywords='docker registry', author='David Halls', author_email='dave@davedoesdev.com', url='https://github.com/davedoesdev/dxf', license='MIT', packages=['dxf'], entry_points={'console_scripts': ['dxf=dxf.main:main']}, install_requires=['www-authenticate>=0.9.2', 'requests>=2.18.4', 'jwcrypto>=0.4.2', 'tqdm>=4.19.4'] )
import io import os from setuptools import setup def read(name): file_path = os.path.join(os.path.dirname(__file__), name) return io.open(file_path, encoding='utf8').read() setup( name='python-dxf', version='7.5.0', description="Package for accessing a Docker v2 registry", long_description=read('README.rst'), keywords='docker registry', author='David Halls', author_email='dave@davedoesdev.com', url='https://github.com/davedoesdev/dxf', license='MIT', packages=['dxf'], entry_points={'console_scripts': ['dxf=dxf.main:main']}, install_requires=['www-authenticate>=0.9.2', 'requests>=2.18.4', 'jwcrypto>=0.4.2', 'tqdm>=4.19.4'] )
Fix UnicodeDecodeError error when installing
Fix UnicodeDecodeError error when installing Fix UnicodeDecodeError raised when attempting to install in non-English systems (e.g. Windows 10 Korean).
Python
mit
davedoesdev/dxf,davedoesdev/dxf,davedoesdev/dxf,davedoesdev/dxf
import os from setuptools import setup def read(name): return open(os.path.join(os.path.dirname(__file__), name)).read() setup( name='python-dxf', version='7.5.0', description="Package for accessing a Docker v2 registry", long_description=read('README.rst'), keywords='docker registry', author='David Halls', author_email='dave@davedoesdev.com', url='https://github.com/davedoesdev/dxf', license='MIT', packages=['dxf'], entry_points={'console_scripts': ['dxf=dxf.main:main']}, install_requires=['www-authenticate>=0.9.2', 'requests>=2.18.4', 'jwcrypto>=0.4.2', 'tqdm>=4.19.4'] ) Fix UnicodeDecodeError error when installing Fix UnicodeDecodeError raised when attempting to install in non-English systems (e.g. Windows 10 Korean).
import io import os from setuptools import setup def read(name): file_path = os.path.join(os.path.dirname(__file__), name) return io.open(file_path, encoding='utf8').read() setup( name='python-dxf', version='7.5.0', description="Package for accessing a Docker v2 registry", long_description=read('README.rst'), keywords='docker registry', author='David Halls', author_email='dave@davedoesdev.com', url='https://github.com/davedoesdev/dxf', license='MIT', packages=['dxf'], entry_points={'console_scripts': ['dxf=dxf.main:main']}, install_requires=['www-authenticate>=0.9.2', 'requests>=2.18.4', 'jwcrypto>=0.4.2', 'tqdm>=4.19.4'] )
<commit_before>import os from setuptools import setup def read(name): return open(os.path.join(os.path.dirname(__file__), name)).read() setup( name='python-dxf', version='7.5.0', description="Package for accessing a Docker v2 registry", long_description=read('README.rst'), keywords='docker registry', author='David Halls', author_email='dave@davedoesdev.com', url='https://github.com/davedoesdev/dxf', license='MIT', packages=['dxf'], entry_points={'console_scripts': ['dxf=dxf.main:main']}, install_requires=['www-authenticate>=0.9.2', 'requests>=2.18.4', 'jwcrypto>=0.4.2', 'tqdm>=4.19.4'] ) <commit_msg>Fix UnicodeDecodeError error when installing Fix UnicodeDecodeError raised when attempting to install in non-English systems (e.g. Windows 10 Korean).<commit_after>
import io import os from setuptools import setup def read(name): file_path = os.path.join(os.path.dirname(__file__), name) return io.open(file_path, encoding='utf8').read() setup( name='python-dxf', version='7.5.0', description="Package for accessing a Docker v2 registry", long_description=read('README.rst'), keywords='docker registry', author='David Halls', author_email='dave@davedoesdev.com', url='https://github.com/davedoesdev/dxf', license='MIT', packages=['dxf'], entry_points={'console_scripts': ['dxf=dxf.main:main']}, install_requires=['www-authenticate>=0.9.2', 'requests>=2.18.4', 'jwcrypto>=0.4.2', 'tqdm>=4.19.4'] )
import os from setuptools import setup def read(name): return open(os.path.join(os.path.dirname(__file__), name)).read() setup( name='python-dxf', version='7.5.0', description="Package for accessing a Docker v2 registry", long_description=read('README.rst'), keywords='docker registry', author='David Halls', author_email='dave@davedoesdev.com', url='https://github.com/davedoesdev/dxf', license='MIT', packages=['dxf'], entry_points={'console_scripts': ['dxf=dxf.main:main']}, install_requires=['www-authenticate>=0.9.2', 'requests>=2.18.4', 'jwcrypto>=0.4.2', 'tqdm>=4.19.4'] ) Fix UnicodeDecodeError error when installing Fix UnicodeDecodeError raised when attempting to install in non-English systems (e.g. Windows 10 Korean).import io import os from setuptools import setup def read(name): file_path = os.path.join(os.path.dirname(__file__), name) return io.open(file_path, encoding='utf8').read() setup( name='python-dxf', version='7.5.0', description="Package for accessing a Docker v2 registry", long_description=read('README.rst'), keywords='docker registry', author='David Halls', author_email='dave@davedoesdev.com', url='https://github.com/davedoesdev/dxf', license='MIT', packages=['dxf'], entry_points={'console_scripts': ['dxf=dxf.main:main']}, install_requires=['www-authenticate>=0.9.2', 'requests>=2.18.4', 'jwcrypto>=0.4.2', 'tqdm>=4.19.4'] )
<commit_before>import os from setuptools import setup def read(name): return open(os.path.join(os.path.dirname(__file__), name)).read() setup( name='python-dxf', version='7.5.0', description="Package for accessing a Docker v2 registry", long_description=read('README.rst'), keywords='docker registry', author='David Halls', author_email='dave@davedoesdev.com', url='https://github.com/davedoesdev/dxf', license='MIT', packages=['dxf'], entry_points={'console_scripts': ['dxf=dxf.main:main']}, install_requires=['www-authenticate>=0.9.2', 'requests>=2.18.4', 'jwcrypto>=0.4.2', 'tqdm>=4.19.4'] ) <commit_msg>Fix UnicodeDecodeError error when installing Fix UnicodeDecodeError raised when attempting to install in non-English systems (e.g. Windows 10 Korean).<commit_after>import io import os from setuptools import setup def read(name): file_path = os.path.join(os.path.dirname(__file__), name) return io.open(file_path, encoding='utf8').read() setup( name='python-dxf', version='7.5.0', description="Package for accessing a Docker v2 registry", long_description=read('README.rst'), keywords='docker registry', author='David Halls', author_email='dave@davedoesdev.com', url='https://github.com/davedoesdev/dxf', license='MIT', packages=['dxf'], entry_points={'console_scripts': ['dxf=dxf.main:main']}, install_requires=['www-authenticate>=0.9.2', 'requests>=2.18.4', 'jwcrypto>=0.4.2', 'tqdm>=4.19.4'] )
816e77d6cb5321b57cbccfed56d01cbfbd058aa1
setup.py
setup.py
# Uploading package to PyPi. from setuptools import setup setup(name='schoolopy', version='0.1.0', description='A Python wrapper for Schoology\'s API.', url='https://github.com/ErikBoesen/schoolopy', author='Erik Boesen', author_email='me@erikboesen.com', license='MIT', packages=['schoolopy'], install_requires=['requests', 'requests-oauthlib', 'oauthlib'], zip_safe=False)
# Uploading package to PyPi. from setuptools import setup setup(name='schoolopy', version='0.2.0', description='A Python wrapper for Schoology\'s API.', url='https://github.com/ErikBoesen/schoolopy', author='Erik Boesen', author_email='me@erikboesen.com', license='MIT', packages=['schoolopy'], install_requires=['requests', 'requests-oauthlib', 'oauthlib'], zip_safe=False)
Increment version for PyPI upload
Increment version for PyPI upload
Python
mit
ErikBoesen/schoolopy
# Uploading package to PyPi. from setuptools import setup setup(name='schoolopy', version='0.1.0', description='A Python wrapper for Schoology\'s API.', url='https://github.com/ErikBoesen/schoolopy', author='Erik Boesen', author_email='me@erikboesen.com', license='MIT', packages=['schoolopy'], install_requires=['requests', 'requests-oauthlib', 'oauthlib'], zip_safe=False) Increment version for PyPI upload
# Uploading package to PyPi. from setuptools import setup setup(name='schoolopy', version='0.2.0', description='A Python wrapper for Schoology\'s API.', url='https://github.com/ErikBoesen/schoolopy', author='Erik Boesen', author_email='me@erikboesen.com', license='MIT', packages=['schoolopy'], install_requires=['requests', 'requests-oauthlib', 'oauthlib'], zip_safe=False)
<commit_before># Uploading package to PyPi. from setuptools import setup setup(name='schoolopy', version='0.1.0', description='A Python wrapper for Schoology\'s API.', url='https://github.com/ErikBoesen/schoolopy', author='Erik Boesen', author_email='me@erikboesen.com', license='MIT', packages=['schoolopy'], install_requires=['requests', 'requests-oauthlib', 'oauthlib'], zip_safe=False) <commit_msg>Increment version for PyPI upload<commit_after>
# Uploading package to PyPi. from setuptools import setup setup(name='schoolopy', version='0.2.0', description='A Python wrapper for Schoology\'s API.', url='https://github.com/ErikBoesen/schoolopy', author='Erik Boesen', author_email='me@erikboesen.com', license='MIT', packages=['schoolopy'], install_requires=['requests', 'requests-oauthlib', 'oauthlib'], zip_safe=False)
# Uploading package to PyPi. from setuptools import setup setup(name='schoolopy', version='0.1.0', description='A Python wrapper for Schoology\'s API.', url='https://github.com/ErikBoesen/schoolopy', author='Erik Boesen', author_email='me@erikboesen.com', license='MIT', packages=['schoolopy'], install_requires=['requests', 'requests-oauthlib', 'oauthlib'], zip_safe=False) Increment version for PyPI upload# Uploading package to PyPi. from setuptools import setup setup(name='schoolopy', version='0.2.0', description='A Python wrapper for Schoology\'s API.', url='https://github.com/ErikBoesen/schoolopy', author='Erik Boesen', author_email='me@erikboesen.com', license='MIT', packages=['schoolopy'], install_requires=['requests', 'requests-oauthlib', 'oauthlib'], zip_safe=False)
<commit_before># Uploading package to PyPi. from setuptools import setup setup(name='schoolopy', version='0.1.0', description='A Python wrapper for Schoology\'s API.', url='https://github.com/ErikBoesen/schoolopy', author='Erik Boesen', author_email='me@erikboesen.com', license='MIT', packages=['schoolopy'], install_requires=['requests', 'requests-oauthlib', 'oauthlib'], zip_safe=False) <commit_msg>Increment version for PyPI upload<commit_after># Uploading package to PyPi. from setuptools import setup setup(name='schoolopy', version='0.2.0', description='A Python wrapper for Schoology\'s API.', url='https://github.com/ErikBoesen/schoolopy', author='Erik Boesen', author_email='me@erikboesen.com', license='MIT', packages=['schoolopy'], install_requires=['requests', 'requests-oauthlib', 'oauthlib'], zip_safe=False)
eb0ffe3da8b25a4d673b543faeef169279c63919
python/hxActor/main.py
python/hxActor/main.py
#!/usr/bin/env python import actorcore.ICC class OurActor(actorcore.ICC.ICC): def __init__(self, name, productName=None, configFile=None, modelNames=(), debugLevel=30): """ Setup an Actor instance. See help for actorcore.Actor for details. """ # This sets up the connections to/from the hub, the logger, and the twisted reactor. # actorcore.ICC.ICC.__init__(self, name, productName=productName, configFile=configFile, modelNames=modelNames) # # To work def main(): theActor = OurActor('hx', productName='hxActor') theActor.run() if __name__ == '__main__': main()
#!/usr/bin/env python import actorcore.ICC class OurActor(actorcore.ICC.ICC): def __init__(self, name, productName=None, configFile=None, modelNames=('charis', 'hx'), debugLevel=30): """ Setup an Actor instance. See help for actorcore.Actor for details. """ # This sets up the connections to/from the hub, the logger, and the twisted reactor. # actorcore.ICC.ICC.__init__(self, name, productName=productName, configFile=configFile, modelNames=modelNames) # # To work def main(): theActor = OurActor('hx', productName='hxActor') theActor.run() if __name__ == '__main__': main()
Add model listeners for headers.
Add model listeners for headers.
Python
mit
CraigLoomis/ics_hxActor,CraigLoomis/ics_hxActor
#!/usr/bin/env python import actorcore.ICC class OurActor(actorcore.ICC.ICC): def __init__(self, name, productName=None, configFile=None, modelNames=(), debugLevel=30): """ Setup an Actor instance. See help for actorcore.Actor for details. """ # This sets up the connections to/from the hub, the logger, and the twisted reactor. # actorcore.ICC.ICC.__init__(self, name, productName=productName, configFile=configFile, modelNames=modelNames) # # To work def main(): theActor = OurActor('hx', productName='hxActor') theActor.run() if __name__ == '__main__': main() Add model listeners for headers.
#!/usr/bin/env python import actorcore.ICC class OurActor(actorcore.ICC.ICC): def __init__(self, name, productName=None, configFile=None, modelNames=('charis', 'hx'), debugLevel=30): """ Setup an Actor instance. See help for actorcore.Actor for details. """ # This sets up the connections to/from the hub, the logger, and the twisted reactor. # actorcore.ICC.ICC.__init__(self, name, productName=productName, configFile=configFile, modelNames=modelNames) # # To work def main(): theActor = OurActor('hx', productName='hxActor') theActor.run() if __name__ == '__main__': main()
<commit_before>#!/usr/bin/env python import actorcore.ICC class OurActor(actorcore.ICC.ICC): def __init__(self, name, productName=None, configFile=None, modelNames=(), debugLevel=30): """ Setup an Actor instance. See help for actorcore.Actor for details. """ # This sets up the connections to/from the hub, the logger, and the twisted reactor. # actorcore.ICC.ICC.__init__(self, name, productName=productName, configFile=configFile, modelNames=modelNames) # # To work def main(): theActor = OurActor('hx', productName='hxActor') theActor.run() if __name__ == '__main__': main() <commit_msg>Add model listeners for headers.<commit_after>
#!/usr/bin/env python import actorcore.ICC class OurActor(actorcore.ICC.ICC): def __init__(self, name, productName=None, configFile=None, modelNames=('charis', 'hx'), debugLevel=30): """ Setup an Actor instance. See help for actorcore.Actor for details. """ # This sets up the connections to/from the hub, the logger, and the twisted reactor. # actorcore.ICC.ICC.__init__(self, name, productName=productName, configFile=configFile, modelNames=modelNames) # # To work def main(): theActor = OurActor('hx', productName='hxActor') theActor.run() if __name__ == '__main__': main()
#!/usr/bin/env python import actorcore.ICC class OurActor(actorcore.ICC.ICC): def __init__(self, name, productName=None, configFile=None, modelNames=(), debugLevel=30): """ Setup an Actor instance. See help for actorcore.Actor for details. """ # This sets up the connections to/from the hub, the logger, and the twisted reactor. # actorcore.ICC.ICC.__init__(self, name, productName=productName, configFile=configFile, modelNames=modelNames) # # To work def main(): theActor = OurActor('hx', productName='hxActor') theActor.run() if __name__ == '__main__': main() Add model listeners for headers.#!/usr/bin/env python import actorcore.ICC class OurActor(actorcore.ICC.ICC): def __init__(self, name, productName=None, configFile=None, modelNames=('charis', 'hx'), debugLevel=30): """ Setup an Actor instance. See help for actorcore.Actor for details. """ # This sets up the connections to/from the hub, the logger, and the twisted reactor. # actorcore.ICC.ICC.__init__(self, name, productName=productName, configFile=configFile, modelNames=modelNames) # # To work def main(): theActor = OurActor('hx', productName='hxActor') theActor.run() if __name__ == '__main__': main()
<commit_before>#!/usr/bin/env python import actorcore.ICC class OurActor(actorcore.ICC.ICC): def __init__(self, name, productName=None, configFile=None, modelNames=(), debugLevel=30): """ Setup an Actor instance. See help for actorcore.Actor for details. """ # This sets up the connections to/from the hub, the logger, and the twisted reactor. # actorcore.ICC.ICC.__init__(self, name, productName=productName, configFile=configFile, modelNames=modelNames) # # To work def main(): theActor = OurActor('hx', productName='hxActor') theActor.run() if __name__ == '__main__': main() <commit_msg>Add model listeners for headers.<commit_after>#!/usr/bin/env python import actorcore.ICC class OurActor(actorcore.ICC.ICC): def __init__(self, name, productName=None, configFile=None, modelNames=('charis', 'hx'), debugLevel=30): """ Setup an Actor instance. See help for actorcore.Actor for details. """ # This sets up the connections to/from the hub, the logger, and the twisted reactor. # actorcore.ICC.ICC.__init__(self, name, productName=productName, configFile=configFile, modelNames=modelNames) # # To work def main(): theActor = OurActor('hx', productName='hxActor') theActor.run() if __name__ == '__main__': main()
7f3f2802fb3c00baf440ebdc10920ad0b48118f9
setup.py
setup.py
import codecs from setuptools import find_packages from setuptools import setup import sys install_requires = [ 'cached-property', 'chainer>=2.0.0', 'future', 'gym>=0.9.7', 'numpy>=1.10.4', 'pillow', 'scipy', ] test_requires = [ 'pytest', ] if sys.version_info < (3, 2): install_requires.append('fastcache') if sys.version_info < (3, 4): install_requires.append('statistics') if sys.version_info < (3, 5): install_requires.append('funcsigs') setup(name='chainerrl', version='0.7.0', description='ChainerRL, a deep reinforcement learning library', long_description=codecs.open('README.md', 'r', encoding='utf-8').read(), long_description_content_type='text/markdown', author='Yasuhiro Fujita', author_email='fujita@preferred.jp', license='MIT License', packages=find_packages(), install_requires=install_requires, test_requires=test_requires)
import codecs from setuptools import find_packages from setuptools import setup import sys install_requires = [ 'cached-property', 'chainer>=2.0.0', 'future', 'gym>=0.9.7', 'numpy>=1.10.4', 'pillow', 'scipy', ] test_requires = [ 'pytest', 'attrs<19.2.0', # pytest does not run with attrs==19.2.0 (https://github.com/pytest-dev/pytest/issues/3280) # NOQA ] if sys.version_info < (3, 2): install_requires.append('fastcache') if sys.version_info < (3, 4): install_requires.append('statistics') if sys.version_info < (3, 5): install_requires.append('funcsigs') setup(name='chainerrl', version='0.7.0', description='ChainerRL, a deep reinforcement learning library', long_description=codecs.open('README.md', 'r', encoding='utf-8').read(), long_description_content_type='text/markdown', author='Yasuhiro Fujita', author_email='fujita@preferred.jp', license='MIT License', packages=find_packages(), install_requires=install_requires, test_requires=test_requires)
Use attrs<19.2.0 to avoid pytest error
Use attrs<19.2.0 to avoid pytest error
Python
mit
toslunar/chainerrl,toslunar/chainerrl
import codecs from setuptools import find_packages from setuptools import setup import sys install_requires = [ 'cached-property', 'chainer>=2.0.0', 'future', 'gym>=0.9.7', 'numpy>=1.10.4', 'pillow', 'scipy', ] test_requires = [ 'pytest', ] if sys.version_info < (3, 2): install_requires.append('fastcache') if sys.version_info < (3, 4): install_requires.append('statistics') if sys.version_info < (3, 5): install_requires.append('funcsigs') setup(name='chainerrl', version='0.7.0', description='ChainerRL, a deep reinforcement learning library', long_description=codecs.open('README.md', 'r', encoding='utf-8').read(), long_description_content_type='text/markdown', author='Yasuhiro Fujita', author_email='fujita@preferred.jp', license='MIT License', packages=find_packages(), install_requires=install_requires, test_requires=test_requires) Use attrs<19.2.0 to avoid pytest error
import codecs from setuptools import find_packages from setuptools import setup import sys install_requires = [ 'cached-property', 'chainer>=2.0.0', 'future', 'gym>=0.9.7', 'numpy>=1.10.4', 'pillow', 'scipy', ] test_requires = [ 'pytest', 'attrs<19.2.0', # pytest does not run with attrs==19.2.0 (https://github.com/pytest-dev/pytest/issues/3280) # NOQA ] if sys.version_info < (3, 2): install_requires.append('fastcache') if sys.version_info < (3, 4): install_requires.append('statistics') if sys.version_info < (3, 5): install_requires.append('funcsigs') setup(name='chainerrl', version='0.7.0', description='ChainerRL, a deep reinforcement learning library', long_description=codecs.open('README.md', 'r', encoding='utf-8').read(), long_description_content_type='text/markdown', author='Yasuhiro Fujita', author_email='fujita@preferred.jp', license='MIT License', packages=find_packages(), install_requires=install_requires, test_requires=test_requires)
<commit_before>import codecs from setuptools import find_packages from setuptools import setup import sys install_requires = [ 'cached-property', 'chainer>=2.0.0', 'future', 'gym>=0.9.7', 'numpy>=1.10.4', 'pillow', 'scipy', ] test_requires = [ 'pytest', ] if sys.version_info < (3, 2): install_requires.append('fastcache') if sys.version_info < (3, 4): install_requires.append('statistics') if sys.version_info < (3, 5): install_requires.append('funcsigs') setup(name='chainerrl', version='0.7.0', description='ChainerRL, a deep reinforcement learning library', long_description=codecs.open('README.md', 'r', encoding='utf-8').read(), long_description_content_type='text/markdown', author='Yasuhiro Fujita', author_email='fujita@preferred.jp', license='MIT License', packages=find_packages(), install_requires=install_requires, test_requires=test_requires) <commit_msg>Use attrs<19.2.0 to avoid pytest error<commit_after>
import codecs from setuptools import find_packages from setuptools import setup import sys install_requires = [ 'cached-property', 'chainer>=2.0.0', 'future', 'gym>=0.9.7', 'numpy>=1.10.4', 'pillow', 'scipy', ] test_requires = [ 'pytest', 'attrs<19.2.0', # pytest does not run with attrs==19.2.0 (https://github.com/pytest-dev/pytest/issues/3280) # NOQA ] if sys.version_info < (3, 2): install_requires.append('fastcache') if sys.version_info < (3, 4): install_requires.append('statistics') if sys.version_info < (3, 5): install_requires.append('funcsigs') setup(name='chainerrl', version='0.7.0', description='ChainerRL, a deep reinforcement learning library', long_description=codecs.open('README.md', 'r', encoding='utf-8').read(), long_description_content_type='text/markdown', author='Yasuhiro Fujita', author_email='fujita@preferred.jp', license='MIT License', packages=find_packages(), install_requires=install_requires, test_requires=test_requires)
import codecs from setuptools import find_packages from setuptools import setup import sys install_requires = [ 'cached-property', 'chainer>=2.0.0', 'future', 'gym>=0.9.7', 'numpy>=1.10.4', 'pillow', 'scipy', ] test_requires = [ 'pytest', ] if sys.version_info < (3, 2): install_requires.append('fastcache') if sys.version_info < (3, 4): install_requires.append('statistics') if sys.version_info < (3, 5): install_requires.append('funcsigs') setup(name='chainerrl', version='0.7.0', description='ChainerRL, a deep reinforcement learning library', long_description=codecs.open('README.md', 'r', encoding='utf-8').read(), long_description_content_type='text/markdown', author='Yasuhiro Fujita', author_email='fujita@preferred.jp', license='MIT License', packages=find_packages(), install_requires=install_requires, test_requires=test_requires) Use attrs<19.2.0 to avoid pytest errorimport codecs from setuptools import find_packages from setuptools import setup import sys install_requires = [ 'cached-property', 'chainer>=2.0.0', 'future', 'gym>=0.9.7', 'numpy>=1.10.4', 'pillow', 'scipy', ] test_requires = [ 'pytest', 'attrs<19.2.0', # pytest does not run with attrs==19.2.0 (https://github.com/pytest-dev/pytest/issues/3280) # NOQA ] if sys.version_info < (3, 2): install_requires.append('fastcache') if sys.version_info < (3, 4): install_requires.append('statistics') if sys.version_info < (3, 5): install_requires.append('funcsigs') setup(name='chainerrl', version='0.7.0', description='ChainerRL, a deep reinforcement learning library', long_description=codecs.open('README.md', 'r', encoding='utf-8').read(), long_description_content_type='text/markdown', author='Yasuhiro Fujita', author_email='fujita@preferred.jp', license='MIT License', packages=find_packages(), install_requires=install_requires, test_requires=test_requires)
<commit_before>import codecs from setuptools import find_packages from setuptools import setup import sys install_requires = [ 'cached-property', 'chainer>=2.0.0', 'future', 'gym>=0.9.7', 'numpy>=1.10.4', 'pillow', 'scipy', ] test_requires = [ 'pytest', ] if sys.version_info < (3, 2): install_requires.append('fastcache') if sys.version_info < (3, 4): install_requires.append('statistics') if sys.version_info < (3, 5): install_requires.append('funcsigs') setup(name='chainerrl', version='0.7.0', description='ChainerRL, a deep reinforcement learning library', long_description=codecs.open('README.md', 'r', encoding='utf-8').read(), long_description_content_type='text/markdown', author='Yasuhiro Fujita', author_email='fujita@preferred.jp', license='MIT License', packages=find_packages(), install_requires=install_requires, test_requires=test_requires) <commit_msg>Use attrs<19.2.0 to avoid pytest error<commit_after>import codecs from setuptools import find_packages from setuptools import setup import sys install_requires = [ 'cached-property', 'chainer>=2.0.0', 'future', 'gym>=0.9.7', 'numpy>=1.10.4', 'pillow', 'scipy', ] test_requires = [ 'pytest', 'attrs<19.2.0', # pytest does not run with attrs==19.2.0 (https://github.com/pytest-dev/pytest/issues/3280) # NOQA ] if sys.version_info < (3, 2): install_requires.append('fastcache') if sys.version_info < (3, 4): install_requires.append('statistics') if sys.version_info < (3, 5): install_requires.append('funcsigs') setup(name='chainerrl', version='0.7.0', description='ChainerRL, a deep reinforcement learning library', long_description=codecs.open('README.md', 'r', encoding='utf-8').read(), long_description_content_type='text/markdown', author='Yasuhiro Fujita', author_email='fujita@preferred.jp', license='MIT License', packages=find_packages(), install_requires=install_requires, test_requires=test_requires)
ccd5dce9b4e2392abdf205d2913736a6ce47ae0b
setup.py
setup.py
#!/usr/bin/env python """ DJ-Database-URL =============== This simple Django utility allows you to utilize the `12factor <http://www.12factor.net/backing-services>`_ inspired ``REDIS_URL`` environment variable to configure your Django application. Usage ----- Configure your Redis database in ``settings.py`` from ``REDIS_URL``:: REDIS = {"default": dj_redis_url.config()} Parse an arbitrary Database URL:: REDIS = {"default": dj_redis_url.parse("redis://..."")} Installation ------------ Installation is simple too:: $ pip install dj-redis-url """ from setuptools import setup, find_packages setup( name="dj-redis-url", version="0.1.2", description="Use Redis URLs in your Django Application.", long_description=__doc__, url="https://github.com/dstufft/dj-redis-url", author="Donald Stufft", author_email="donald.stufft@gmail.com", extras_require={ "tests": ["pytest"], }, packages=find_packages(exclude=["tests"]), package_data={"": ["LICENSE"]}, include_package_data=True, zip_safe=False, )
#!/usr/bin/env python """ DJ-Database-URL =============== This simple Django utility allows you to utilize the `12factor <http://www.12factor.net/backing-services>`_ inspired ``REDIS_URL`` environment variable to configure your Django application. Usage ----- Configure your Redis database in ``settings.py`` from ``REDIS_URL``:: REDIS = {"default": dj_redis_url.config()} Parse an arbitrary Database URL:: REDIS = {"default": dj_redis_url.parse("redis://..."")} Installation ------------ Installation is simple too:: $ pip install dj-redis-url """ from setuptools import setup setup( name="dj-redis-url", version="0.1.2", description="Use Redis URLs in your Django Application.", long_description=__doc__, url="https://github.com/dstufft/dj-redis-url", author="Donald Stufft", author_email="donald.stufft@gmail.com", extras_require={ "tests": ["pytest"], }, py_modules=["dj_redis_url"], include_package_data=True, zip_safe=False, )
Switch from packages to modules
Switch from packages to modules
Python
bsd-2-clause
dstufft/dj-redis-url
#!/usr/bin/env python """ DJ-Database-URL =============== This simple Django utility allows you to utilize the `12factor <http://www.12factor.net/backing-services>`_ inspired ``REDIS_URL`` environment variable to configure your Django application. Usage ----- Configure your Redis database in ``settings.py`` from ``REDIS_URL``:: REDIS = {"default": dj_redis_url.config()} Parse an arbitrary Database URL:: REDIS = {"default": dj_redis_url.parse("redis://..."")} Installation ------------ Installation is simple too:: $ pip install dj-redis-url """ from setuptools import setup, find_packages setup( name="dj-redis-url", version="0.1.2", description="Use Redis URLs in your Django Application.", long_description=__doc__, url="https://github.com/dstufft/dj-redis-url", author="Donald Stufft", author_email="donald.stufft@gmail.com", extras_require={ "tests": ["pytest"], }, packages=find_packages(exclude=["tests"]), package_data={"": ["LICENSE"]}, include_package_data=True, zip_safe=False, ) Switch from packages to modules
#!/usr/bin/env python """ DJ-Database-URL =============== This simple Django utility allows you to utilize the `12factor <http://www.12factor.net/backing-services>`_ inspired ``REDIS_URL`` environment variable to configure your Django application. Usage ----- Configure your Redis database in ``settings.py`` from ``REDIS_URL``:: REDIS = {"default": dj_redis_url.config()} Parse an arbitrary Database URL:: REDIS = {"default": dj_redis_url.parse("redis://..."")} Installation ------------ Installation is simple too:: $ pip install dj-redis-url """ from setuptools import setup setup( name="dj-redis-url", version="0.1.2", description="Use Redis URLs in your Django Application.", long_description=__doc__, url="https://github.com/dstufft/dj-redis-url", author="Donald Stufft", author_email="donald.stufft@gmail.com", extras_require={ "tests": ["pytest"], }, py_modules=["dj_redis_url"], include_package_data=True, zip_safe=False, )
<commit_before>#!/usr/bin/env python """ DJ-Database-URL =============== This simple Django utility allows you to utilize the `12factor <http://www.12factor.net/backing-services>`_ inspired ``REDIS_URL`` environment variable to configure your Django application. Usage ----- Configure your Redis database in ``settings.py`` from ``REDIS_URL``:: REDIS = {"default": dj_redis_url.config()} Parse an arbitrary Database URL:: REDIS = {"default": dj_redis_url.parse("redis://..."")} Installation ------------ Installation is simple too:: $ pip install dj-redis-url """ from setuptools import setup, find_packages setup( name="dj-redis-url", version="0.1.2", description="Use Redis URLs in your Django Application.", long_description=__doc__, url="https://github.com/dstufft/dj-redis-url", author="Donald Stufft", author_email="donald.stufft@gmail.com", extras_require={ "tests": ["pytest"], }, packages=find_packages(exclude=["tests"]), package_data={"": ["LICENSE"]}, include_package_data=True, zip_safe=False, ) <commit_msg>Switch from packages to modules<commit_after>
#!/usr/bin/env python """ DJ-Database-URL =============== This simple Django utility allows you to utilize the `12factor <http://www.12factor.net/backing-services>`_ inspired ``REDIS_URL`` environment variable to configure your Django application. Usage ----- Configure your Redis database in ``settings.py`` from ``REDIS_URL``:: REDIS = {"default": dj_redis_url.config()} Parse an arbitrary Database URL:: REDIS = {"default": dj_redis_url.parse("redis://..."")} Installation ------------ Installation is simple too:: $ pip install dj-redis-url """ from setuptools import setup setup( name="dj-redis-url", version="0.1.2", description="Use Redis URLs in your Django Application.", long_description=__doc__, url="https://github.com/dstufft/dj-redis-url", author="Donald Stufft", author_email="donald.stufft@gmail.com", extras_require={ "tests": ["pytest"], }, py_modules=["dj_redis_url"], include_package_data=True, zip_safe=False, )
#!/usr/bin/env python """ DJ-Database-URL =============== This simple Django utility allows you to utilize the `12factor <http://www.12factor.net/backing-services>`_ inspired ``REDIS_URL`` environment variable to configure your Django application. Usage ----- Configure your Redis database in ``settings.py`` from ``REDIS_URL``:: REDIS = {"default": dj_redis_url.config()} Parse an arbitrary Database URL:: REDIS = {"default": dj_redis_url.parse("redis://..."")} Installation ------------ Installation is simple too:: $ pip install dj-redis-url """ from setuptools import setup, find_packages setup( name="dj-redis-url", version="0.1.2", description="Use Redis URLs in your Django Application.", long_description=__doc__, url="https://github.com/dstufft/dj-redis-url", author="Donald Stufft", author_email="donald.stufft@gmail.com", extras_require={ "tests": ["pytest"], }, packages=find_packages(exclude=["tests"]), package_data={"": ["LICENSE"]}, include_package_data=True, zip_safe=False, ) Switch from packages to modules#!/usr/bin/env python """ DJ-Database-URL =============== This simple Django utility allows you to utilize the `12factor <http://www.12factor.net/backing-services>`_ inspired ``REDIS_URL`` environment variable to configure your Django application. Usage ----- Configure your Redis database in ``settings.py`` from ``REDIS_URL``:: REDIS = {"default": dj_redis_url.config()} Parse an arbitrary Database URL:: REDIS = {"default": dj_redis_url.parse("redis://..."")} Installation ------------ Installation is simple too:: $ pip install dj-redis-url """ from setuptools import setup setup( name="dj-redis-url", version="0.1.2", description="Use Redis URLs in your Django Application.", long_description=__doc__, url="https://github.com/dstufft/dj-redis-url", author="Donald Stufft", author_email="donald.stufft@gmail.com", extras_require={ "tests": ["pytest"], }, py_modules=["dj_redis_url"], include_package_data=True, zip_safe=False, )
<commit_before>#!/usr/bin/env python """ DJ-Database-URL =============== This simple Django utility allows you to utilize the `12factor <http://www.12factor.net/backing-services>`_ inspired ``REDIS_URL`` environment variable to configure your Django application. Usage ----- Configure your Redis database in ``settings.py`` from ``REDIS_URL``:: REDIS = {"default": dj_redis_url.config()} Parse an arbitrary Database URL:: REDIS = {"default": dj_redis_url.parse("redis://..."")} Installation ------------ Installation is simple too:: $ pip install dj-redis-url """ from setuptools import setup, find_packages setup( name="dj-redis-url", version="0.1.2", description="Use Redis URLs in your Django Application.", long_description=__doc__, url="https://github.com/dstufft/dj-redis-url", author="Donald Stufft", author_email="donald.stufft@gmail.com", extras_require={ "tests": ["pytest"], }, packages=find_packages(exclude=["tests"]), package_data={"": ["LICENSE"]}, include_package_data=True, zip_safe=False, ) <commit_msg>Switch from packages to modules<commit_after>#!/usr/bin/env python """ DJ-Database-URL =============== This simple Django utility allows you to utilize the `12factor <http://www.12factor.net/backing-services>`_ inspired ``REDIS_URL`` environment variable to configure your Django application. Usage ----- Configure your Redis database in ``settings.py`` from ``REDIS_URL``:: REDIS = {"default": dj_redis_url.config()} Parse an arbitrary Database URL:: REDIS = {"default": dj_redis_url.parse("redis://..."")} Installation ------------ Installation is simple too:: $ pip install dj-redis-url """ from setuptools import setup setup( name="dj-redis-url", version="0.1.2", description="Use Redis URLs in your Django Application.", long_description=__doc__, url="https://github.com/dstufft/dj-redis-url", author="Donald Stufft", author_email="donald.stufft@gmail.com", extras_require={ "tests": ["pytest"], }, py_modules=["dj_redis_url"], include_package_data=True, zip_safe=False, )
df9379a370163ac0d9ca3daf99745c1d1fb7dc19
setup.py
setup.py
from setuptools import setup, find_packages from codecs import open from os import path _HERE = path.abspath(path.dirname(__file__)) def read(*names, **kwds): return open( path.join(_HERE, *names), encoding=kwds.get('encoding', 'utf-8') ).read() def find_version(*file_paths): import re version_file = read(*file_paths) version_match = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]", version_file, re.M) if version_match: return version_match.group(1) raise RuntimeError("Unable to find vresion string.") setup( name='wmtexe', version=find_version("wmtexe/__init__.py"), description='WMT execution server.', long_description=read('README.md'), url='https://github.com/csdms/wmt-exe', author='Eric Hutton', author_email='hutton.eric@gmail.com', license='MIT', packages=find_packages(), entry_points={ 'console_scripts': [ 'wmt-slave=wmtexe.slave:main', ], }, )
from setuptools import setup, find_packages from codecs import open from os import path _HERE = path.abspath(path.dirname(__file__)) def read(*names, **kwds): return open( path.join(_HERE, *names), encoding=kwds.get('encoding', 'utf-8') ).read() def find_version(*file_paths): import re version_file = read(*file_paths) version_match = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]", version_file, re.M) if version_match: return version_match.group(1) raise RuntimeError("Unable to find vresion string.") setup( name='wmtexe', version=find_version("wmtexe/__init__.py"), description='WMT execution server.', long_description=read('README.md'), url='https://github.com/csdms/wmt-exe', author='Eric Hutton', author_email='hutton.eric@gmail.com', license='MIT', packages=find_packages(), entry_points={ 'console_scripts': [ 'wmt-slave=wmtexe.slave:main', ], 'distutils.commands': [ 'configure = wmtexe.configure:Configure', ], }, )
Use the new configure command.
Use the new configure command.
Python
mit
csdms/wmt-exe,csdms/wmt-exe,csdms/wmt-exe,csdms/wmt-exe
from setuptools import setup, find_packages from codecs import open from os import path _HERE = path.abspath(path.dirname(__file__)) def read(*names, **kwds): return open( path.join(_HERE, *names), encoding=kwds.get('encoding', 'utf-8') ).read() def find_version(*file_paths): import re version_file = read(*file_paths) version_match = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]", version_file, re.M) if version_match: return version_match.group(1) raise RuntimeError("Unable to find vresion string.") setup( name='wmtexe', version=find_version("wmtexe/__init__.py"), description='WMT execution server.', long_description=read('README.md'), url='https://github.com/csdms/wmt-exe', author='Eric Hutton', author_email='hutton.eric@gmail.com', license='MIT', packages=find_packages(), entry_points={ 'console_scripts': [ 'wmt-slave=wmtexe.slave:main', ], }, ) Use the new configure command.
from setuptools import setup, find_packages from codecs import open from os import path _HERE = path.abspath(path.dirname(__file__)) def read(*names, **kwds): return open( path.join(_HERE, *names), encoding=kwds.get('encoding', 'utf-8') ).read() def find_version(*file_paths): import re version_file = read(*file_paths) version_match = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]", version_file, re.M) if version_match: return version_match.group(1) raise RuntimeError("Unable to find vresion string.") setup( name='wmtexe', version=find_version("wmtexe/__init__.py"), description='WMT execution server.', long_description=read('README.md'), url='https://github.com/csdms/wmt-exe', author='Eric Hutton', author_email='hutton.eric@gmail.com', license='MIT', packages=find_packages(), entry_points={ 'console_scripts': [ 'wmt-slave=wmtexe.slave:main', ], 'distutils.commands': [ 'configure = wmtexe.configure:Configure', ], }, )
<commit_before>from setuptools import setup, find_packages from codecs import open from os import path _HERE = path.abspath(path.dirname(__file__)) def read(*names, **kwds): return open( path.join(_HERE, *names), encoding=kwds.get('encoding', 'utf-8') ).read() def find_version(*file_paths): import re version_file = read(*file_paths) version_match = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]", version_file, re.M) if version_match: return version_match.group(1) raise RuntimeError("Unable to find vresion string.") setup( name='wmtexe', version=find_version("wmtexe/__init__.py"), description='WMT execution server.', long_description=read('README.md'), url='https://github.com/csdms/wmt-exe', author='Eric Hutton', author_email='hutton.eric@gmail.com', license='MIT', packages=find_packages(), entry_points={ 'console_scripts': [ 'wmt-slave=wmtexe.slave:main', ], }, ) <commit_msg>Use the new configure command.<commit_after>
from setuptools import setup, find_packages from codecs import open from os import path _HERE = path.abspath(path.dirname(__file__)) def read(*names, **kwds): return open( path.join(_HERE, *names), encoding=kwds.get('encoding', 'utf-8') ).read() def find_version(*file_paths): import re version_file = read(*file_paths) version_match = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]", version_file, re.M) if version_match: return version_match.group(1) raise RuntimeError("Unable to find vresion string.") setup( name='wmtexe', version=find_version("wmtexe/__init__.py"), description='WMT execution server.', long_description=read('README.md'), url='https://github.com/csdms/wmt-exe', author='Eric Hutton', author_email='hutton.eric@gmail.com', license='MIT', packages=find_packages(), entry_points={ 'console_scripts': [ 'wmt-slave=wmtexe.slave:main', ], 'distutils.commands': [ 'configure = wmtexe.configure:Configure', ], }, )
from setuptools import setup, find_packages from codecs import open from os import path _HERE = path.abspath(path.dirname(__file__)) def read(*names, **kwds): return open( path.join(_HERE, *names), encoding=kwds.get('encoding', 'utf-8') ).read() def find_version(*file_paths): import re version_file = read(*file_paths) version_match = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]", version_file, re.M) if version_match: return version_match.group(1) raise RuntimeError("Unable to find vresion string.") setup( name='wmtexe', version=find_version("wmtexe/__init__.py"), description='WMT execution server.', long_description=read('README.md'), url='https://github.com/csdms/wmt-exe', author='Eric Hutton', author_email='hutton.eric@gmail.com', license='MIT', packages=find_packages(), entry_points={ 'console_scripts': [ 'wmt-slave=wmtexe.slave:main', ], }, ) Use the new configure command.from setuptools import setup, find_packages from codecs import open from os import path _HERE = path.abspath(path.dirname(__file__)) def read(*names, **kwds): return open( path.join(_HERE, *names), encoding=kwds.get('encoding', 'utf-8') ).read() def find_version(*file_paths): import re version_file = read(*file_paths) version_match = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]", version_file, re.M) if version_match: return version_match.group(1) raise RuntimeError("Unable to find vresion string.") setup( name='wmtexe', version=find_version("wmtexe/__init__.py"), description='WMT execution server.', long_description=read('README.md'), url='https://github.com/csdms/wmt-exe', author='Eric Hutton', author_email='hutton.eric@gmail.com', license='MIT', packages=find_packages(), entry_points={ 'console_scripts': [ 'wmt-slave=wmtexe.slave:main', ], 'distutils.commands': [ 'configure = wmtexe.configure:Configure', ], }, )
<commit_before>from setuptools import setup, find_packages from codecs import open from os import path _HERE = path.abspath(path.dirname(__file__)) def read(*names, **kwds): return open( path.join(_HERE, *names), encoding=kwds.get('encoding', 'utf-8') ).read() def find_version(*file_paths): import re version_file = read(*file_paths) version_match = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]", version_file, re.M) if version_match: return version_match.group(1) raise RuntimeError("Unable to find vresion string.") setup( name='wmtexe', version=find_version("wmtexe/__init__.py"), description='WMT execution server.', long_description=read('README.md'), url='https://github.com/csdms/wmt-exe', author='Eric Hutton', author_email='hutton.eric@gmail.com', license='MIT', packages=find_packages(), entry_points={ 'console_scripts': [ 'wmt-slave=wmtexe.slave:main', ], }, ) <commit_msg>Use the new configure command.<commit_after>from setuptools import setup, find_packages from codecs import open from os import path _HERE = path.abspath(path.dirname(__file__)) def read(*names, **kwds): return open( path.join(_HERE, *names), encoding=kwds.get('encoding', 'utf-8') ).read() def find_version(*file_paths): import re version_file = read(*file_paths) version_match = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]", version_file, re.M) if version_match: return version_match.group(1) raise RuntimeError("Unable to find vresion string.") setup( name='wmtexe', version=find_version("wmtexe/__init__.py"), description='WMT execution server.', long_description=read('README.md'), url='https://github.com/csdms/wmt-exe', author='Eric Hutton', author_email='hutton.eric@gmail.com', license='MIT', packages=find_packages(), entry_points={ 'console_scripts': [ 'wmt-slave=wmtexe.slave:main', ], 'distutils.commands': [ 'configure = wmtexe.configure:Configure', ], }, )
8a8d5bfa1d091f2c701196f58fdf10bfb228d9b5
setup.py
setup.py
""" Flask-Twilio ------------- Make Twilio voice/SMS calls with Flask """ from setuptools import setup exec(open('flask_twilio.py').readline()) setup( name='Flask-Twilio', version=__version__, url='http://example.com/flask-twilio/', license='BSD', author='Leo Singer', author_email='leo.singer@ligo.org', description='Make Twilio voice/SMS calls with Flask', long_description=__doc__, py_modules=['flask_twilio'], zip_safe=False, include_package_data=True, platforms='any', install_requires=[ 'itsdangerous', 'Flask', 'twilio' ], setup_requires=[ 'pytest-runner', ], tests_require=[ 'pytest', ], classifiers=[ 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: Communications :: Telephony' ] )
""" Flask-Twilio ------------- Make Twilio voice/SMS calls with Flask """ from setuptools import setup exec(open('flask_twilio.py').readline()) setup( name='Flask-Twilio', version=__version__, url='http://example.com/flask-twilio/', license='BSD', author='Leo Singer', author_email='leo.singer@ligo.org', description='Make Twilio voice/SMS calls with Flask', long_description=__doc__, py_modules=['flask_twilio'], zip_safe=False, include_package_data=True, platforms='any', install_requires=[ 'itsdangerous', 'Flask', 'six', 'twilio' ], setup_requires=[ 'pytest-runner', ], tests_require=[ 'pytest', ], classifiers=[ 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: Communications :: Telephony' ] )
Add missing dependency on six
Add missing dependency on six
Python
bsd-3-clause
lpsinger/flask-twilio
""" Flask-Twilio ------------- Make Twilio voice/SMS calls with Flask """ from setuptools import setup exec(open('flask_twilio.py').readline()) setup( name='Flask-Twilio', version=__version__, url='http://example.com/flask-twilio/', license='BSD', author='Leo Singer', author_email='leo.singer@ligo.org', description='Make Twilio voice/SMS calls with Flask', long_description=__doc__, py_modules=['flask_twilio'], zip_safe=False, include_package_data=True, platforms='any', install_requires=[ 'itsdangerous', 'Flask', 'twilio' ], setup_requires=[ 'pytest-runner', ], tests_require=[ 'pytest', ], classifiers=[ 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: Communications :: Telephony' ] )Add missing dependency on six
""" Flask-Twilio ------------- Make Twilio voice/SMS calls with Flask """ from setuptools import setup exec(open('flask_twilio.py').readline()) setup( name='Flask-Twilio', version=__version__, url='http://example.com/flask-twilio/', license='BSD', author='Leo Singer', author_email='leo.singer@ligo.org', description='Make Twilio voice/SMS calls with Flask', long_description=__doc__, py_modules=['flask_twilio'], zip_safe=False, include_package_data=True, platforms='any', install_requires=[ 'itsdangerous', 'Flask', 'six', 'twilio' ], setup_requires=[ 'pytest-runner', ], tests_require=[ 'pytest', ], classifiers=[ 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: Communications :: Telephony' ] )
<commit_before>""" Flask-Twilio ------------- Make Twilio voice/SMS calls with Flask """ from setuptools import setup exec(open('flask_twilio.py').readline()) setup( name='Flask-Twilio', version=__version__, url='http://example.com/flask-twilio/', license='BSD', author='Leo Singer', author_email='leo.singer@ligo.org', description='Make Twilio voice/SMS calls with Flask', long_description=__doc__, py_modules=['flask_twilio'], zip_safe=False, include_package_data=True, platforms='any', install_requires=[ 'itsdangerous', 'Flask', 'twilio' ], setup_requires=[ 'pytest-runner', ], tests_require=[ 'pytest', ], classifiers=[ 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: Communications :: Telephony' ] )<commit_msg>Add missing dependency on six<commit_after>
""" Flask-Twilio ------------- Make Twilio voice/SMS calls with Flask """ from setuptools import setup exec(open('flask_twilio.py').readline()) setup( name='Flask-Twilio', version=__version__, url='http://example.com/flask-twilio/', license='BSD', author='Leo Singer', author_email='leo.singer@ligo.org', description='Make Twilio voice/SMS calls with Flask', long_description=__doc__, py_modules=['flask_twilio'], zip_safe=False, include_package_data=True, platforms='any', install_requires=[ 'itsdangerous', 'Flask', 'six', 'twilio' ], setup_requires=[ 'pytest-runner', ], tests_require=[ 'pytest', ], classifiers=[ 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: Communications :: Telephony' ] )
""" Flask-Twilio ------------- Make Twilio voice/SMS calls with Flask """ from setuptools import setup exec(open('flask_twilio.py').readline()) setup( name='Flask-Twilio', version=__version__, url='http://example.com/flask-twilio/', license='BSD', author='Leo Singer', author_email='leo.singer@ligo.org', description='Make Twilio voice/SMS calls with Flask', long_description=__doc__, py_modules=['flask_twilio'], zip_safe=False, include_package_data=True, platforms='any', install_requires=[ 'itsdangerous', 'Flask', 'twilio' ], setup_requires=[ 'pytest-runner', ], tests_require=[ 'pytest', ], classifiers=[ 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: Communications :: Telephony' ] )Add missing dependency on six""" Flask-Twilio ------------- Make Twilio voice/SMS calls with Flask """ from setuptools import setup exec(open('flask_twilio.py').readline()) setup( name='Flask-Twilio', version=__version__, url='http://example.com/flask-twilio/', license='BSD', author='Leo Singer', author_email='leo.singer@ligo.org', description='Make Twilio voice/SMS calls with Flask', long_description=__doc__, py_modules=['flask_twilio'], zip_safe=False, include_package_data=True, platforms='any', install_requires=[ 'itsdangerous', 'Flask', 'six', 'twilio' ], setup_requires=[ 'pytest-runner', ], tests_require=[ 'pytest', ], classifiers=[ 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: Communications :: Telephony' ] )
<commit_before>""" Flask-Twilio ------------- Make Twilio voice/SMS calls with Flask """ from setuptools import setup exec(open('flask_twilio.py').readline()) setup( name='Flask-Twilio', version=__version__, url='http://example.com/flask-twilio/', license='BSD', author='Leo Singer', author_email='leo.singer@ligo.org', description='Make Twilio voice/SMS calls with Flask', long_description=__doc__, py_modules=['flask_twilio'], zip_safe=False, include_package_data=True, platforms='any', install_requires=[ 'itsdangerous', 'Flask', 'twilio' ], setup_requires=[ 'pytest-runner', ], tests_require=[ 'pytest', ], classifiers=[ 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: Communications :: Telephony' ] )<commit_msg>Add missing dependency on six<commit_after>""" Flask-Twilio ------------- Make Twilio voice/SMS calls with Flask """ from setuptools import setup exec(open('flask_twilio.py').readline()) setup( name='Flask-Twilio', version=__version__, url='http://example.com/flask-twilio/', license='BSD', author='Leo Singer', author_email='leo.singer@ligo.org', description='Make Twilio voice/SMS calls with Flask', long_description=__doc__, py_modules=['flask_twilio'], zip_safe=False, include_package_data=True, platforms='any', install_requires=[ 'itsdangerous', 'Flask', 'six', 'twilio' ], setup_requires=[ 'pytest-runner', ], tests_require=[ 'pytest', ], classifiers=[ 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Dynamic Content', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: Communications :: Telephony' ] )
bff48b594f4714e0f2d52793a1a038ce1f436719
setup.py
setup.py
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='edx-oauth2-provider', version='0.5.4', description='Provide OAuth2 access to edX installations', author='edX', url='https://github.com/edx/edx-oauth2-provider', license='AGPL', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: GNU Affero General Public License v3', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ], packages=find_packages(exclude=['tests', '*.tests']), dependency_links=[ 'git+https://github.com/edx/django-oauth2-provider@0.2.7-fork-edx-5#egg=django-oauth2-provider-0.2.7-fork-edx-5', ], install_requires=[ 'django-oauth2-provider==0.2.7-fork-edx-5', 'PyJWT==1.0.1' ] )
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='edx-oauth2-provider', version='0.5.5', description='Provide OAuth2 access to edX installations', author='edX', url='https://github.com/edx/edx-oauth2-provider', license='AGPL', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: GNU Affero General Public License v3', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ], packages=find_packages(exclude=['tests']), dependency_links=[ 'git+https://github.com/edx/django-oauth2-provider@0.2.7-fork-edx-5#egg=django-oauth2-provider-0.2.7-fork-edx-5', ], install_requires=[ 'django-oauth2-provider==0.2.7-fork-edx-5', 'PyJWT==1.0.1' ] )
Include oauth2-provider's tests module in the installed package
Include oauth2-provider's tests module in the installed package
Python
agpl-3.0
edx/edx-oauth2-provider,edx/edx-oauth2-provider
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='edx-oauth2-provider', version='0.5.4', description='Provide OAuth2 access to edX installations', author='edX', url='https://github.com/edx/edx-oauth2-provider', license='AGPL', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: GNU Affero General Public License v3', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ], packages=find_packages(exclude=['tests', '*.tests']), dependency_links=[ 'git+https://github.com/edx/django-oauth2-provider@0.2.7-fork-edx-5#egg=django-oauth2-provider-0.2.7-fork-edx-5', ], install_requires=[ 'django-oauth2-provider==0.2.7-fork-edx-5', 'PyJWT==1.0.1' ] ) Include oauth2-provider's tests module in the installed package
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='edx-oauth2-provider', version='0.5.5', description='Provide OAuth2 access to edX installations', author='edX', url='https://github.com/edx/edx-oauth2-provider', license='AGPL', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: GNU Affero General Public License v3', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ], packages=find_packages(exclude=['tests']), dependency_links=[ 'git+https://github.com/edx/django-oauth2-provider@0.2.7-fork-edx-5#egg=django-oauth2-provider-0.2.7-fork-edx-5', ], install_requires=[ 'django-oauth2-provider==0.2.7-fork-edx-5', 'PyJWT==1.0.1' ] )
<commit_before>#!/usr/bin/env python from setuptools import setup, find_packages setup( name='edx-oauth2-provider', version='0.5.4', description='Provide OAuth2 access to edX installations', author='edX', url='https://github.com/edx/edx-oauth2-provider', license='AGPL', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: GNU Affero General Public License v3', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ], packages=find_packages(exclude=['tests', '*.tests']), dependency_links=[ 'git+https://github.com/edx/django-oauth2-provider@0.2.7-fork-edx-5#egg=django-oauth2-provider-0.2.7-fork-edx-5', ], install_requires=[ 'django-oauth2-provider==0.2.7-fork-edx-5', 'PyJWT==1.0.1' ] ) <commit_msg>Include oauth2-provider's tests module in the installed package<commit_after>
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='edx-oauth2-provider', version='0.5.5', description='Provide OAuth2 access to edX installations', author='edX', url='https://github.com/edx/edx-oauth2-provider', license='AGPL', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: GNU Affero General Public License v3', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ], packages=find_packages(exclude=['tests']), dependency_links=[ 'git+https://github.com/edx/django-oauth2-provider@0.2.7-fork-edx-5#egg=django-oauth2-provider-0.2.7-fork-edx-5', ], install_requires=[ 'django-oauth2-provider==0.2.7-fork-edx-5', 'PyJWT==1.0.1' ] )
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='edx-oauth2-provider', version='0.5.4', description='Provide OAuth2 access to edX installations', author='edX', url='https://github.com/edx/edx-oauth2-provider', license='AGPL', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: GNU Affero General Public License v3', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ], packages=find_packages(exclude=['tests', '*.tests']), dependency_links=[ 'git+https://github.com/edx/django-oauth2-provider@0.2.7-fork-edx-5#egg=django-oauth2-provider-0.2.7-fork-edx-5', ], install_requires=[ 'django-oauth2-provider==0.2.7-fork-edx-5', 'PyJWT==1.0.1' ] ) Include oauth2-provider's tests module in the installed package#!/usr/bin/env python from setuptools import setup, find_packages setup( name='edx-oauth2-provider', version='0.5.5', description='Provide OAuth2 access to edX installations', author='edX', url='https://github.com/edx/edx-oauth2-provider', license='AGPL', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: GNU Affero General Public License v3', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ], packages=find_packages(exclude=['tests']), dependency_links=[ 'git+https://github.com/edx/django-oauth2-provider@0.2.7-fork-edx-5#egg=django-oauth2-provider-0.2.7-fork-edx-5', ], install_requires=[ 'django-oauth2-provider==0.2.7-fork-edx-5', 'PyJWT==1.0.1' ] )
<commit_before>#!/usr/bin/env python from setuptools import setup, find_packages setup( name='edx-oauth2-provider', version='0.5.4', description='Provide OAuth2 access to edX installations', author='edX', url='https://github.com/edx/edx-oauth2-provider', license='AGPL', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: GNU Affero General Public License v3', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ], packages=find_packages(exclude=['tests', '*.tests']), dependency_links=[ 'git+https://github.com/edx/django-oauth2-provider@0.2.7-fork-edx-5#egg=django-oauth2-provider-0.2.7-fork-edx-5', ], install_requires=[ 'django-oauth2-provider==0.2.7-fork-edx-5', 'PyJWT==1.0.1' ] ) <commit_msg>Include oauth2-provider's tests module in the installed package<commit_after>#!/usr/bin/env python from setuptools import setup, find_packages setup( name='edx-oauth2-provider', version='0.5.5', description='Provide OAuth2 access to edX installations', author='edX', url='https://github.com/edx/edx-oauth2-provider', license='AGPL', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Web Environment', 'Intended Audience :: Developers', 'License :: OSI Approved :: GNU Affero General Public License v3', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Framework :: Django', ], packages=find_packages(exclude=['tests']), dependency_links=[ 'git+https://github.com/edx/django-oauth2-provider@0.2.7-fork-edx-5#egg=django-oauth2-provider-0.2.7-fork-edx-5', ], install_requires=[ 'django-oauth2-provider==0.2.7-fork-edx-5', 'PyJWT==1.0.1' ] )
8dffe003fdb42cb9edd4fa0e720d6eceb9719381
setup.py
setup.py
from setuptools import setup, find_packages import thermochem setup( name="thermochem", version=thermochem.__version__, description="Python utilities for thermodynamics and thermochemistry", author="Adel Qalieh", author_email="adelq@sas.upenn.edu", url="https://github.com/adelq/thermochem", license="BSD", packages=find_packages(), include_package_data=True, install_requires=['scipy>=0.6.0', 'numpy>=1.2.1', 'pandas>=0.17.0'], zip_safe=False, keywords='thermo chemistry physics', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Science/Research', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.5', 'Topic :: Scientific/Engineering :: Chemistry', 'Topic :: Scientific/Engineering :: Physics', 'Topic :: Software Development :: Libraries :: Python Modules', ], )
from setuptools import setup, find_packages import thermochem setup( name="thermochem", version=thermochem.__version__, description="Python utilities for thermodynamics and thermochemistry", long_description=open("README.rst").read(), author="Adel Qalieh", author_email="adelq@sas.upenn.edu", url="https://github.com/adelq/thermochem", license="BSD", packages=find_packages(), include_package_data=True, install_requires=['scipy>=0.6.0', 'numpy>=1.2.1', 'pandas>=0.17.0'], zip_safe=False, keywords='thermo chemistry physics', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Science/Research', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.5', 'Topic :: Scientific/Engineering :: Chemistry', 'Topic :: Scientific/Engineering :: Physics', 'Topic :: Software Development :: Libraries :: Python Modules', ], )
Add README as long description
Add README as long description
Python
bsd-3-clause
adelq/thermopy
from setuptools import setup, find_packages import thermochem setup( name="thermochem", version=thermochem.__version__, description="Python utilities for thermodynamics and thermochemistry", author="Adel Qalieh", author_email="adelq@sas.upenn.edu", url="https://github.com/adelq/thermochem", license="BSD", packages=find_packages(), include_package_data=True, install_requires=['scipy>=0.6.0', 'numpy>=1.2.1', 'pandas>=0.17.0'], zip_safe=False, keywords='thermo chemistry physics', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Science/Research', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.5', 'Topic :: Scientific/Engineering :: Chemistry', 'Topic :: Scientific/Engineering :: Physics', 'Topic :: Software Development :: Libraries :: Python Modules', ], ) Add README as long description
from setuptools import setup, find_packages import thermochem setup( name="thermochem", version=thermochem.__version__, description="Python utilities for thermodynamics and thermochemistry", long_description=open("README.rst").read(), author="Adel Qalieh", author_email="adelq@sas.upenn.edu", url="https://github.com/adelq/thermochem", license="BSD", packages=find_packages(), include_package_data=True, install_requires=['scipy>=0.6.0', 'numpy>=1.2.1', 'pandas>=0.17.0'], zip_safe=False, keywords='thermo chemistry physics', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Science/Research', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.5', 'Topic :: Scientific/Engineering :: Chemistry', 'Topic :: Scientific/Engineering :: Physics', 'Topic :: Software Development :: Libraries :: Python Modules', ], )
<commit_before>from setuptools import setup, find_packages import thermochem setup( name="thermochem", version=thermochem.__version__, description="Python utilities for thermodynamics and thermochemistry", author="Adel Qalieh", author_email="adelq@sas.upenn.edu", url="https://github.com/adelq/thermochem", license="BSD", packages=find_packages(), include_package_data=True, install_requires=['scipy>=0.6.0', 'numpy>=1.2.1', 'pandas>=0.17.0'], zip_safe=False, keywords='thermo chemistry physics', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Science/Research', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.5', 'Topic :: Scientific/Engineering :: Chemistry', 'Topic :: Scientific/Engineering :: Physics', 'Topic :: Software Development :: Libraries :: Python Modules', ], ) <commit_msg>Add README as long description<commit_after>
from setuptools import setup, find_packages import thermochem setup( name="thermochem", version=thermochem.__version__, description="Python utilities for thermodynamics and thermochemistry", long_description=open("README.rst").read(), author="Adel Qalieh", author_email="adelq@sas.upenn.edu", url="https://github.com/adelq/thermochem", license="BSD", packages=find_packages(), include_package_data=True, install_requires=['scipy>=0.6.0', 'numpy>=1.2.1', 'pandas>=0.17.0'], zip_safe=False, keywords='thermo chemistry physics', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Science/Research', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.5', 'Topic :: Scientific/Engineering :: Chemistry', 'Topic :: Scientific/Engineering :: Physics', 'Topic :: Software Development :: Libraries :: Python Modules', ], )
from setuptools import setup, find_packages import thermochem setup( name="thermochem", version=thermochem.__version__, description="Python utilities for thermodynamics and thermochemistry", author="Adel Qalieh", author_email="adelq@sas.upenn.edu", url="https://github.com/adelq/thermochem", license="BSD", packages=find_packages(), include_package_data=True, install_requires=['scipy>=0.6.0', 'numpy>=1.2.1', 'pandas>=0.17.0'], zip_safe=False, keywords='thermo chemistry physics', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Science/Research', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.5', 'Topic :: Scientific/Engineering :: Chemistry', 'Topic :: Scientific/Engineering :: Physics', 'Topic :: Software Development :: Libraries :: Python Modules', ], ) Add README as long descriptionfrom setuptools import setup, find_packages import thermochem setup( name="thermochem", version=thermochem.__version__, description="Python utilities for thermodynamics and thermochemistry", long_description=open("README.rst").read(), author="Adel Qalieh", author_email="adelq@sas.upenn.edu", url="https://github.com/adelq/thermochem", license="BSD", packages=find_packages(), include_package_data=True, install_requires=['scipy>=0.6.0', 'numpy>=1.2.1', 'pandas>=0.17.0'], zip_safe=False, keywords='thermo chemistry physics', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Science/Research', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.5', 'Topic :: Scientific/Engineering :: Chemistry', 'Topic :: Scientific/Engineering :: Physics', 'Topic :: Software Development :: Libraries :: Python Modules', ], )
<commit_before>from setuptools import setup, find_packages import thermochem setup( name="thermochem", version=thermochem.__version__, description="Python utilities for thermodynamics and thermochemistry", author="Adel Qalieh", author_email="adelq@sas.upenn.edu", url="https://github.com/adelq/thermochem", license="BSD", packages=find_packages(), include_package_data=True, install_requires=['scipy>=0.6.0', 'numpy>=1.2.1', 'pandas>=0.17.0'], zip_safe=False, keywords='thermo chemistry physics', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Science/Research', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.5', 'Topic :: Scientific/Engineering :: Chemistry', 'Topic :: Scientific/Engineering :: Physics', 'Topic :: Software Development :: Libraries :: Python Modules', ], ) <commit_msg>Add README as long description<commit_after>from setuptools import setup, find_packages import thermochem setup( name="thermochem", version=thermochem.__version__, description="Python utilities for thermodynamics and thermochemistry", long_description=open("README.rst").read(), author="Adel Qalieh", author_email="adelq@sas.upenn.edu", url="https://github.com/adelq/thermochem", license="BSD", packages=find_packages(), include_package_data=True, install_requires=['scipy>=0.6.0', 'numpy>=1.2.1', 'pandas>=0.17.0'], zip_safe=False, keywords='thermo chemistry physics', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Science/Research', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.5', 'Topic :: Scientific/Engineering :: Chemistry', 'Topic :: Scientific/Engineering :: Physics', 'Topic :: Software Development :: Libraries :: Python Modules', ], )
3c33bc1a8cdc33b5519b7a0802b7b71d0193a6c6
setup.py
setup.py
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-slack', url="https://chris-lamb.co.uk/projects/django-slack", version='5.2.0', description="Provides easy-to-use integration between Django projects and " "the Slack group chat and IM tool.", author="Chris Lamb", author_email="chris@chris-lamb.co.uk", license="BSD", packages=find_packages(), include_package_data=True, install_requires=( 'Django>=1.6.8', 'requests', 'six', ), )
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-slack', url="https://chris-lamb.co.uk/projects/django-slack", version='5.2.0', description="Provides easy-to-use integration between Django projects and " "the Slack group chat and IM tool.", author="Chris Lamb", author_email="chris@chris-lamb.co.uk", license="BSD", packages=find_packages(), include_package_data=True, install_requires=( 'Django>=1.8.0', 'requests', 'six', ), )
Update Django requirement to latest LTS
Update Django requirement to latest LTS
Python
bsd-3-clause
lamby/django-slack
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-slack', url="https://chris-lamb.co.uk/projects/django-slack", version='5.2.0', description="Provides easy-to-use integration between Django projects and " "the Slack group chat and IM tool.", author="Chris Lamb", author_email="chris@chris-lamb.co.uk", license="BSD", packages=find_packages(), include_package_data=True, install_requires=( 'Django>=1.6.8', 'requests', 'six', ), ) Update Django requirement to latest LTS
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-slack', url="https://chris-lamb.co.uk/projects/django-slack", version='5.2.0', description="Provides easy-to-use integration between Django projects and " "the Slack group chat and IM tool.", author="Chris Lamb", author_email="chris@chris-lamb.co.uk", license="BSD", packages=find_packages(), include_package_data=True, install_requires=( 'Django>=1.8.0', 'requests', 'six', ), )
<commit_before>#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-slack', url="https://chris-lamb.co.uk/projects/django-slack", version='5.2.0', description="Provides easy-to-use integration between Django projects and " "the Slack group chat and IM tool.", author="Chris Lamb", author_email="chris@chris-lamb.co.uk", license="BSD", packages=find_packages(), include_package_data=True, install_requires=( 'Django>=1.6.8', 'requests', 'six', ), ) <commit_msg>Update Django requirement to latest LTS<commit_after>
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-slack', url="https://chris-lamb.co.uk/projects/django-slack", version='5.2.0', description="Provides easy-to-use integration between Django projects and " "the Slack group chat and IM tool.", author="Chris Lamb", author_email="chris@chris-lamb.co.uk", license="BSD", packages=find_packages(), include_package_data=True, install_requires=( 'Django>=1.8.0', 'requests', 'six', ), )
#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-slack', url="https://chris-lamb.co.uk/projects/django-slack", version='5.2.0', description="Provides easy-to-use integration between Django projects and " "the Slack group chat and IM tool.", author="Chris Lamb", author_email="chris@chris-lamb.co.uk", license="BSD", packages=find_packages(), include_package_data=True, install_requires=( 'Django>=1.6.8', 'requests', 'six', ), ) Update Django requirement to latest LTS#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-slack', url="https://chris-lamb.co.uk/projects/django-slack", version='5.2.0', description="Provides easy-to-use integration between Django projects and " "the Slack group chat and IM tool.", author="Chris Lamb", author_email="chris@chris-lamb.co.uk", license="BSD", packages=find_packages(), include_package_data=True, install_requires=( 'Django>=1.8.0', 'requests', 'six', ), )
<commit_before>#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-slack', url="https://chris-lamb.co.uk/projects/django-slack", version='5.2.0', description="Provides easy-to-use integration between Django projects and " "the Slack group chat and IM tool.", author="Chris Lamb", author_email="chris@chris-lamb.co.uk", license="BSD", packages=find_packages(), include_package_data=True, install_requires=( 'Django>=1.6.8', 'requests', 'six', ), ) <commit_msg>Update Django requirement to latest LTS<commit_after>#!/usr/bin/env python from setuptools import setup, find_packages setup( name='django-slack', url="https://chris-lamb.co.uk/projects/django-slack", version='5.2.0', description="Provides easy-to-use integration between Django projects and " "the Slack group chat and IM tool.", author="Chris Lamb", author_email="chris@chris-lamb.co.uk", license="BSD", packages=find_packages(), include_package_data=True, install_requires=( 'Django>=1.8.0', 'requests', 'six', ), )
f994ab4a70a55706eeff133dd484dfd80a68e108
bin/isy_showevents.py
bin/isy_showevents.py
#!/usr/local/bin/python2.7 -u __author__ = "Peter Shipley" import os import keyring import ConfigParser from ISY.IsyEvent import ISYEvent def main() : config = ConfigParser.ConfigParser() config.read(os.path.expanduser('~/isy.cfg')) server = ISYEvent() # you can subscribe to multiple devices # server.subscribe('10.1.1.25') isy_addr = config.get('isy', 'addr') isy_user = config.get('isy', 'user') server.subscribe( addr=isy_addr, userl=isy_user, userp=keyring.get_password("isy", isy_user) ) server.set_process_func(ISYEvent.print_event, "") try: print('Use Control-C to exit') server.events_loop() #no return # for d in server.event_iter( ignorelist=["_0", "_11"] ): # server.print_event(d, "") except KeyboardInterrupt: print('Exiting') if __name__ == '__main__' : main() exit(0)
#!/usr/local/bin/python2.7 -u __author__ = "Peter Shipley" import os import ConfigParser from ISY.IsyEvent import ISYEvent def main() : config = ConfigParser.ConfigParser() config.read(os.path.expanduser('~/home.cfg')) server = ISYEvent() isy_addr = config.get('isy', 'addr') isy_user = config.get('isy', 'user') isy_pass = config.get('isy', 'pass') server.subscribe( addr=isy_addr, userl=isy_user, userp=isy_pass ) server.set_process_func(ISYEvent.print_event, "") try: print('Use Control-C to exit') server.events_loop() #no return except KeyboardInterrupt: print('Exiting') if __name__ == '__main__' : main() exit(0)
Remove keyring and change to basic cfg file handling
Remove keyring and change to basic cfg file handling
Python
bsd-2-clause
fxstein/ISYlib-python
#!/usr/local/bin/python2.7 -u __author__ = "Peter Shipley" import os import keyring import ConfigParser from ISY.IsyEvent import ISYEvent def main() : config = ConfigParser.ConfigParser() config.read(os.path.expanduser('~/isy.cfg')) server = ISYEvent() # you can subscribe to multiple devices # server.subscribe('10.1.1.25') isy_addr = config.get('isy', 'addr') isy_user = config.get('isy', 'user') server.subscribe( addr=isy_addr, userl=isy_user, userp=keyring.get_password("isy", isy_user) ) server.set_process_func(ISYEvent.print_event, "") try: print('Use Control-C to exit') server.events_loop() #no return # for d in server.event_iter( ignorelist=["_0", "_11"] ): # server.print_event(d, "") except KeyboardInterrupt: print('Exiting') if __name__ == '__main__' : main() exit(0) Remove keyring and change to basic cfg file handling
#!/usr/local/bin/python2.7 -u __author__ = "Peter Shipley" import os import ConfigParser from ISY.IsyEvent import ISYEvent def main() : config = ConfigParser.ConfigParser() config.read(os.path.expanduser('~/home.cfg')) server = ISYEvent() isy_addr = config.get('isy', 'addr') isy_user = config.get('isy', 'user') isy_pass = config.get('isy', 'pass') server.subscribe( addr=isy_addr, userl=isy_user, userp=isy_pass ) server.set_process_func(ISYEvent.print_event, "") try: print('Use Control-C to exit') server.events_loop() #no return except KeyboardInterrupt: print('Exiting') if __name__ == '__main__' : main() exit(0)
<commit_before>#!/usr/local/bin/python2.7 -u __author__ = "Peter Shipley" import os import keyring import ConfigParser from ISY.IsyEvent import ISYEvent def main() : config = ConfigParser.ConfigParser() config.read(os.path.expanduser('~/isy.cfg')) server = ISYEvent() # you can subscribe to multiple devices # server.subscribe('10.1.1.25') isy_addr = config.get('isy', 'addr') isy_user = config.get('isy', 'user') server.subscribe( addr=isy_addr, userl=isy_user, userp=keyring.get_password("isy", isy_user) ) server.set_process_func(ISYEvent.print_event, "") try: print('Use Control-C to exit') server.events_loop() #no return # for d in server.event_iter( ignorelist=["_0", "_11"] ): # server.print_event(d, "") except KeyboardInterrupt: print('Exiting') if __name__ == '__main__' : main() exit(0) <commit_msg>Remove keyring and change to basic cfg file handling<commit_after>
#!/usr/local/bin/python2.7 -u __author__ = "Peter Shipley" import os import ConfigParser from ISY.IsyEvent import ISYEvent def main() : config = ConfigParser.ConfigParser() config.read(os.path.expanduser('~/home.cfg')) server = ISYEvent() isy_addr = config.get('isy', 'addr') isy_user = config.get('isy', 'user') isy_pass = config.get('isy', 'pass') server.subscribe( addr=isy_addr, userl=isy_user, userp=isy_pass ) server.set_process_func(ISYEvent.print_event, "") try: print('Use Control-C to exit') server.events_loop() #no return except KeyboardInterrupt: print('Exiting') if __name__ == '__main__' : main() exit(0)
#!/usr/local/bin/python2.7 -u __author__ = "Peter Shipley" import os import keyring import ConfigParser from ISY.IsyEvent import ISYEvent def main() : config = ConfigParser.ConfigParser() config.read(os.path.expanduser('~/isy.cfg')) server = ISYEvent() # you can subscribe to multiple devices # server.subscribe('10.1.1.25') isy_addr = config.get('isy', 'addr') isy_user = config.get('isy', 'user') server.subscribe( addr=isy_addr, userl=isy_user, userp=keyring.get_password("isy", isy_user) ) server.set_process_func(ISYEvent.print_event, "") try: print('Use Control-C to exit') server.events_loop() #no return # for d in server.event_iter( ignorelist=["_0", "_11"] ): # server.print_event(d, "") except KeyboardInterrupt: print('Exiting') if __name__ == '__main__' : main() exit(0) Remove keyring and change to basic cfg file handling#!/usr/local/bin/python2.7 -u __author__ = "Peter Shipley" import os import ConfigParser from ISY.IsyEvent import ISYEvent def main() : config = ConfigParser.ConfigParser() config.read(os.path.expanduser('~/home.cfg')) server = ISYEvent() isy_addr = config.get('isy', 'addr') isy_user = config.get('isy', 'user') isy_pass = config.get('isy', 'pass') server.subscribe( addr=isy_addr, userl=isy_user, userp=isy_pass ) server.set_process_func(ISYEvent.print_event, "") try: print('Use Control-C to exit') server.events_loop() #no return except KeyboardInterrupt: print('Exiting') if __name__ == '__main__' : main() exit(0)
<commit_before>#!/usr/local/bin/python2.7 -u __author__ = "Peter Shipley" import os import keyring import ConfigParser from ISY.IsyEvent import ISYEvent def main() : config = ConfigParser.ConfigParser() config.read(os.path.expanduser('~/isy.cfg')) server = ISYEvent() # you can subscribe to multiple devices # server.subscribe('10.1.1.25') isy_addr = config.get('isy', 'addr') isy_user = config.get('isy', 'user') server.subscribe( addr=isy_addr, userl=isy_user, userp=keyring.get_password("isy", isy_user) ) server.set_process_func(ISYEvent.print_event, "") try: print('Use Control-C to exit') server.events_loop() #no return # for d in server.event_iter( ignorelist=["_0", "_11"] ): # server.print_event(d, "") except KeyboardInterrupt: print('Exiting') if __name__ == '__main__' : main() exit(0) <commit_msg>Remove keyring and change to basic cfg file handling<commit_after>#!/usr/local/bin/python2.7 -u __author__ = "Peter Shipley" import os import ConfigParser from ISY.IsyEvent import ISYEvent def main() : config = ConfigParser.ConfigParser() config.read(os.path.expanduser('~/home.cfg')) server = ISYEvent() isy_addr = config.get('isy', 'addr') isy_user = config.get('isy', 'user') isy_pass = config.get('isy', 'pass') server.subscribe( addr=isy_addr, userl=isy_user, userp=isy_pass ) server.set_process_func(ISYEvent.print_event, "") try: print('Use Control-C to exit') server.events_loop() #no return except KeyboardInterrupt: print('Exiting') if __name__ == '__main__' : main() exit(0)
ff9f64da2d43591cacfbec1a147fda4b82539c1d
serial_reader.py
serial_reader.py
#!/usr/bin/env python from argparse import ArgumentParser import serial def run(device, baud): with serial.Serial(device, baud, timeout=0.1) as ser: while True: line = ser.readline() if line: print line if __name__ == '__main__': parser = ArgumentParser() parser.add_argument('device', help='serial device, typically /dev/tty.usbserial-*') parser.add_argument('--baud', dest='baud', type=int, default=74880) args = parser.parse_args() run(args.device, args.baud)
#!/usr/bin/env python from argparse import ArgumentParser import sys import serial def run(device, baud): with serial.Serial(device, baud, timeout=0.1) as ser: while True: line = ser.readline() if line: sys.stdout.write(line) if __name__ == '__main__': parser = ArgumentParser() parser.add_argument('device', help='serial device, typically /dev/tty.usbserial-*') parser.add_argument('--baud', dest='baud', type=int, default=74880) args = parser.parse_args() run(args.device, args.baud)
Use sys.stdout.write to get rid of extra newlines
Use sys.stdout.write to get rid of extra newlines
Python
unlicense
recursify/serial-debug-tool
#!/usr/bin/env python from argparse import ArgumentParser import serial def run(device, baud): with serial.Serial(device, baud, timeout=0.1) as ser: while True: line = ser.readline() if line: print line if __name__ == '__main__': parser = ArgumentParser() parser.add_argument('device', help='serial device, typically /dev/tty.usbserial-*') parser.add_argument('--baud', dest='baud', type=int, default=74880) args = parser.parse_args() run(args.device, args.baud) Use sys.stdout.write to get rid of extra newlines
#!/usr/bin/env python from argparse import ArgumentParser import sys import serial def run(device, baud): with serial.Serial(device, baud, timeout=0.1) as ser: while True: line = ser.readline() if line: sys.stdout.write(line) if __name__ == '__main__': parser = ArgumentParser() parser.add_argument('device', help='serial device, typically /dev/tty.usbserial-*') parser.add_argument('--baud', dest='baud', type=int, default=74880) args = parser.parse_args() run(args.device, args.baud)
<commit_before>#!/usr/bin/env python from argparse import ArgumentParser import serial def run(device, baud): with serial.Serial(device, baud, timeout=0.1) as ser: while True: line = ser.readline() if line: print line if __name__ == '__main__': parser = ArgumentParser() parser.add_argument('device', help='serial device, typically /dev/tty.usbserial-*') parser.add_argument('--baud', dest='baud', type=int, default=74880) args = parser.parse_args() run(args.device, args.baud) <commit_msg>Use sys.stdout.write to get rid of extra newlines<commit_after>
#!/usr/bin/env python from argparse import ArgumentParser import sys import serial def run(device, baud): with serial.Serial(device, baud, timeout=0.1) as ser: while True: line = ser.readline() if line: sys.stdout.write(line) if __name__ == '__main__': parser = ArgumentParser() parser.add_argument('device', help='serial device, typically /dev/tty.usbserial-*') parser.add_argument('--baud', dest='baud', type=int, default=74880) args = parser.parse_args() run(args.device, args.baud)
#!/usr/bin/env python from argparse import ArgumentParser import serial def run(device, baud): with serial.Serial(device, baud, timeout=0.1) as ser: while True: line = ser.readline() if line: print line if __name__ == '__main__': parser = ArgumentParser() parser.add_argument('device', help='serial device, typically /dev/tty.usbserial-*') parser.add_argument('--baud', dest='baud', type=int, default=74880) args = parser.parse_args() run(args.device, args.baud) Use sys.stdout.write to get rid of extra newlines#!/usr/bin/env python from argparse import ArgumentParser import sys import serial def run(device, baud): with serial.Serial(device, baud, timeout=0.1) as ser: while True: line = ser.readline() if line: sys.stdout.write(line) if __name__ == '__main__': parser = ArgumentParser() parser.add_argument('device', help='serial device, typically /dev/tty.usbserial-*') parser.add_argument('--baud', dest='baud', type=int, default=74880) args = parser.parse_args() run(args.device, args.baud)
<commit_before>#!/usr/bin/env python from argparse import ArgumentParser import serial def run(device, baud): with serial.Serial(device, baud, timeout=0.1) as ser: while True: line = ser.readline() if line: print line if __name__ == '__main__': parser = ArgumentParser() parser.add_argument('device', help='serial device, typically /dev/tty.usbserial-*') parser.add_argument('--baud', dest='baud', type=int, default=74880) args = parser.parse_args() run(args.device, args.baud) <commit_msg>Use sys.stdout.write to get rid of extra newlines<commit_after>#!/usr/bin/env python from argparse import ArgumentParser import sys import serial def run(device, baud): with serial.Serial(device, baud, timeout=0.1) as ser: while True: line = ser.readline() if line: sys.stdout.write(line) if __name__ == '__main__': parser = ArgumentParser() parser.add_argument('device', help='serial device, typically /dev/tty.usbserial-*') parser.add_argument('--baud', dest='baud', type=int, default=74880) args = parser.parse_args() run(args.device, args.baud)
8ffd26f4fddb0c367e61a46af6427eab6c244ea8
south/signals.py
south/signals.py
""" South-specific signals """ from django.dispatch import Signal # Sent at the start of the migration of an app pre_migrate = Signal(providing_args=["app"]) # Sent after each successful migration of an app post_migrate = Signal(providing_args=["app"]) # Sent after each run of a particular migration in a direction ran_migration = Signal(providing_args=["app","migration","method"])
""" South-specific signals """ from django.dispatch import Signal from django.conf import settings # Sent at the start of the migration of an app pre_migrate = Signal(providing_args=["app"]) # Sent after each successful migration of an app post_migrate = Signal(providing_args=["app"]) # Sent after each run of a particular migration in a direction ran_migration = Signal(providing_args=["app","migration","method"]) # Compatibility code for django.contrib.auth if 'django.contrib.auth' in settings.INSTALLED_APPS: def create_permissions_compat(app, **kwargs): from django.db.models import get_app from django.contrib.auth.management import create_permissions create_permissions(get_app(app), (), 0) post_migrate.connect(create_permissions_compat)
Add a compatibility hook to deal with creating django.contrib.auth permissions on migrated models.
Add a compatibility hook to deal with creating django.contrib.auth permissions on migrated models.
Python
apache-2.0
theatlantic/django-south,theatlantic/django-south
""" South-specific signals """ from django.dispatch import Signal # Sent at the start of the migration of an app pre_migrate = Signal(providing_args=["app"]) # Sent after each successful migration of an app post_migrate = Signal(providing_args=["app"]) # Sent after each run of a particular migration in a direction ran_migration = Signal(providing_args=["app","migration","method"]) Add a compatibility hook to deal with creating django.contrib.auth permissions on migrated models.
""" South-specific signals """ from django.dispatch import Signal from django.conf import settings # Sent at the start of the migration of an app pre_migrate = Signal(providing_args=["app"]) # Sent after each successful migration of an app post_migrate = Signal(providing_args=["app"]) # Sent after each run of a particular migration in a direction ran_migration = Signal(providing_args=["app","migration","method"]) # Compatibility code for django.contrib.auth if 'django.contrib.auth' in settings.INSTALLED_APPS: def create_permissions_compat(app, **kwargs): from django.db.models import get_app from django.contrib.auth.management import create_permissions create_permissions(get_app(app), (), 0) post_migrate.connect(create_permissions_compat)
<commit_before>""" South-specific signals """ from django.dispatch import Signal # Sent at the start of the migration of an app pre_migrate = Signal(providing_args=["app"]) # Sent after each successful migration of an app post_migrate = Signal(providing_args=["app"]) # Sent after each run of a particular migration in a direction ran_migration = Signal(providing_args=["app","migration","method"]) <commit_msg>Add a compatibility hook to deal with creating django.contrib.auth permissions on migrated models.<commit_after>
""" South-specific signals """ from django.dispatch import Signal from django.conf import settings # Sent at the start of the migration of an app pre_migrate = Signal(providing_args=["app"]) # Sent after each successful migration of an app post_migrate = Signal(providing_args=["app"]) # Sent after each run of a particular migration in a direction ran_migration = Signal(providing_args=["app","migration","method"]) # Compatibility code for django.contrib.auth if 'django.contrib.auth' in settings.INSTALLED_APPS: def create_permissions_compat(app, **kwargs): from django.db.models import get_app from django.contrib.auth.management import create_permissions create_permissions(get_app(app), (), 0) post_migrate.connect(create_permissions_compat)
""" South-specific signals """ from django.dispatch import Signal # Sent at the start of the migration of an app pre_migrate = Signal(providing_args=["app"]) # Sent after each successful migration of an app post_migrate = Signal(providing_args=["app"]) # Sent after each run of a particular migration in a direction ran_migration = Signal(providing_args=["app","migration","method"]) Add a compatibility hook to deal with creating django.contrib.auth permissions on migrated models.""" South-specific signals """ from django.dispatch import Signal from django.conf import settings # Sent at the start of the migration of an app pre_migrate = Signal(providing_args=["app"]) # Sent after each successful migration of an app post_migrate = Signal(providing_args=["app"]) # Sent after each run of a particular migration in a direction ran_migration = Signal(providing_args=["app","migration","method"]) # Compatibility code for django.contrib.auth if 'django.contrib.auth' in settings.INSTALLED_APPS: def create_permissions_compat(app, **kwargs): from django.db.models import get_app from django.contrib.auth.management import create_permissions create_permissions(get_app(app), (), 0) post_migrate.connect(create_permissions_compat)
<commit_before>""" South-specific signals """ from django.dispatch import Signal # Sent at the start of the migration of an app pre_migrate = Signal(providing_args=["app"]) # Sent after each successful migration of an app post_migrate = Signal(providing_args=["app"]) # Sent after each run of a particular migration in a direction ran_migration = Signal(providing_args=["app","migration","method"]) <commit_msg>Add a compatibility hook to deal with creating django.contrib.auth permissions on migrated models.<commit_after>""" South-specific signals """ from django.dispatch import Signal from django.conf import settings # Sent at the start of the migration of an app pre_migrate = Signal(providing_args=["app"]) # Sent after each successful migration of an app post_migrate = Signal(providing_args=["app"]) # Sent after each run of a particular migration in a direction ran_migration = Signal(providing_args=["app","migration","method"]) # Compatibility code for django.contrib.auth if 'django.contrib.auth' in settings.INSTALLED_APPS: def create_permissions_compat(app, **kwargs): from django.db.models import get_app from django.contrib.auth.management import create_permissions create_permissions(get_app(app), (), 0) post_migrate.connect(create_permissions_compat)
46a568690a9a284ddc350519a15e092e1211d073
reviewboard/site/urlresolvers.py
reviewboard/site/urlresolvers.py
from __future__ import unicode_literals from django.core.urlresolvers import NoReverseMatch, reverse def local_site_reverse(viewname, request=None, local_site_name=None, args=None, kwargs=None, *func_args, **func_kwargs): """Reverses a URL name, returning a working URL. This works much like Django's reverse(), but handles returning a localsite version of a URL when invoked with a request within a localsite. """ if request or local_site_name: if request and not local_site_name: local_site_name = getattr(request, '_local_site_name', None) if local_site_name: if args: new_args = [local_site_name] + args new_kwargs = kwargs else: new_args = args new_kwargs = { 'local_site_name': local_site_name, } if kwargs: new_kwargs.update(kwargs) try: return reverse(viewname, args=new_args, kwargs=new_kwargs, *func_args, **func_kwargs) except NoReverseMatch: # We'll try it again without those arguments. pass return reverse(viewname, args=args, kwargs=kwargs, *func_args, **func_kwargs)
from __future__ import unicode_literals from django.core.urlresolvers import NoReverseMatch, reverse def local_site_reverse(viewname, request=None, local_site_name=None, local_site=None, args=None, kwargs=None, *func_args, **func_kwargs): """Reverses a URL name, returning a working URL. This works much like Django's reverse(), but handles returning a localsite version of a URL when invoked with a request within a localsite. """ assert not (local_site_name and local_site) if request or local_site_name or local_site: if local_site: local_site_name = local_site.name elif request and not local_site_name: local_site_name = getattr(request, '_local_site_name', None) if local_site_name: if args: new_args = [local_site_name] + args new_kwargs = kwargs else: new_args = args new_kwargs = { 'local_site_name': local_site_name, } if kwargs: new_kwargs.update(kwargs) try: return reverse(viewname, args=new_args, kwargs=new_kwargs, *func_args, **func_kwargs) except NoReverseMatch: # We'll try it again without those arguments. pass return reverse(viewname, args=args, kwargs=kwargs, *func_args, **func_kwargs)
Allow local_site_reverse to take an actual LocalSite.
Allow local_site_reverse to take an actual LocalSite. local_site_reverse was able to take a LocalSite's name, or a request object, but if you actually had a LocalSite (or None), you'd have to write your own conditional to extract the name and pass it. Now, local_site_reverse can take a LocalSite. This saves a database query, if one is already available, and simplifies calling code. Reviewed at https://reviews.reviewboard.org/r/6302/
Python
mit
custode/reviewboard,custode/reviewboard,bkochendorfer/reviewboard,custode/reviewboard,brennie/reviewboard,reviewboard/reviewboard,KnowNo/reviewboard,sgallagher/reviewboard,custode/reviewboard,sgallagher/reviewboard,brennie/reviewboard,davidt/reviewboard,chipx86/reviewboard,KnowNo/reviewboard,chipx86/reviewboard,KnowNo/reviewboard,beol/reviewboard,bkochendorfer/reviewboard,davidt/reviewboard,bkochendorfer/reviewboard,beol/reviewboard,chipx86/reviewboard,davidt/reviewboard,bkochendorfer/reviewboard,reviewboard/reviewboard,chipx86/reviewboard,davidt/reviewboard,brennie/reviewboard,reviewboard/reviewboard,brennie/reviewboard,beol/reviewboard,KnowNo/reviewboard,sgallagher/reviewboard,reviewboard/reviewboard,beol/reviewboard,sgallagher/reviewboard
from __future__ import unicode_literals from django.core.urlresolvers import NoReverseMatch, reverse def local_site_reverse(viewname, request=None, local_site_name=None, args=None, kwargs=None, *func_args, **func_kwargs): """Reverses a URL name, returning a working URL. This works much like Django's reverse(), but handles returning a localsite version of a URL when invoked with a request within a localsite. """ if request or local_site_name: if request and not local_site_name: local_site_name = getattr(request, '_local_site_name', None) if local_site_name: if args: new_args = [local_site_name] + args new_kwargs = kwargs else: new_args = args new_kwargs = { 'local_site_name': local_site_name, } if kwargs: new_kwargs.update(kwargs) try: return reverse(viewname, args=new_args, kwargs=new_kwargs, *func_args, **func_kwargs) except NoReverseMatch: # We'll try it again without those arguments. pass return reverse(viewname, args=args, kwargs=kwargs, *func_args, **func_kwargs) Allow local_site_reverse to take an actual LocalSite. local_site_reverse was able to take a LocalSite's name, or a request object, but if you actually had a LocalSite (or None), you'd have to write your own conditional to extract the name and pass it. Now, local_site_reverse can take a LocalSite. This saves a database query, if one is already available, and simplifies calling code. Reviewed at https://reviews.reviewboard.org/r/6302/
from __future__ import unicode_literals from django.core.urlresolvers import NoReverseMatch, reverse def local_site_reverse(viewname, request=None, local_site_name=None, local_site=None, args=None, kwargs=None, *func_args, **func_kwargs): """Reverses a URL name, returning a working URL. This works much like Django's reverse(), but handles returning a localsite version of a URL when invoked with a request within a localsite. """ assert not (local_site_name and local_site) if request or local_site_name or local_site: if local_site: local_site_name = local_site.name elif request and not local_site_name: local_site_name = getattr(request, '_local_site_name', None) if local_site_name: if args: new_args = [local_site_name] + args new_kwargs = kwargs else: new_args = args new_kwargs = { 'local_site_name': local_site_name, } if kwargs: new_kwargs.update(kwargs) try: return reverse(viewname, args=new_args, kwargs=new_kwargs, *func_args, **func_kwargs) except NoReverseMatch: # We'll try it again without those arguments. pass return reverse(viewname, args=args, kwargs=kwargs, *func_args, **func_kwargs)
<commit_before>from __future__ import unicode_literals from django.core.urlresolvers import NoReverseMatch, reverse def local_site_reverse(viewname, request=None, local_site_name=None, args=None, kwargs=None, *func_args, **func_kwargs): """Reverses a URL name, returning a working URL. This works much like Django's reverse(), but handles returning a localsite version of a URL when invoked with a request within a localsite. """ if request or local_site_name: if request and not local_site_name: local_site_name = getattr(request, '_local_site_name', None) if local_site_name: if args: new_args = [local_site_name] + args new_kwargs = kwargs else: new_args = args new_kwargs = { 'local_site_name': local_site_name, } if kwargs: new_kwargs.update(kwargs) try: return reverse(viewname, args=new_args, kwargs=new_kwargs, *func_args, **func_kwargs) except NoReverseMatch: # We'll try it again without those arguments. pass return reverse(viewname, args=args, kwargs=kwargs, *func_args, **func_kwargs) <commit_msg>Allow local_site_reverse to take an actual LocalSite. local_site_reverse was able to take a LocalSite's name, or a request object, but if you actually had a LocalSite (or None), you'd have to write your own conditional to extract the name and pass it. Now, local_site_reverse can take a LocalSite. This saves a database query, if one is already available, and simplifies calling code. Reviewed at https://reviews.reviewboard.org/r/6302/<commit_after>
from __future__ import unicode_literals from django.core.urlresolvers import NoReverseMatch, reverse def local_site_reverse(viewname, request=None, local_site_name=None, local_site=None, args=None, kwargs=None, *func_args, **func_kwargs): """Reverses a URL name, returning a working URL. This works much like Django's reverse(), but handles returning a localsite version of a URL when invoked with a request within a localsite. """ assert not (local_site_name and local_site) if request or local_site_name or local_site: if local_site: local_site_name = local_site.name elif request and not local_site_name: local_site_name = getattr(request, '_local_site_name', None) if local_site_name: if args: new_args = [local_site_name] + args new_kwargs = kwargs else: new_args = args new_kwargs = { 'local_site_name': local_site_name, } if kwargs: new_kwargs.update(kwargs) try: return reverse(viewname, args=new_args, kwargs=new_kwargs, *func_args, **func_kwargs) except NoReverseMatch: # We'll try it again without those arguments. pass return reverse(viewname, args=args, kwargs=kwargs, *func_args, **func_kwargs)
from __future__ import unicode_literals from django.core.urlresolvers import NoReverseMatch, reverse def local_site_reverse(viewname, request=None, local_site_name=None, args=None, kwargs=None, *func_args, **func_kwargs): """Reverses a URL name, returning a working URL. This works much like Django's reverse(), but handles returning a localsite version of a URL when invoked with a request within a localsite. """ if request or local_site_name: if request and not local_site_name: local_site_name = getattr(request, '_local_site_name', None) if local_site_name: if args: new_args = [local_site_name] + args new_kwargs = kwargs else: new_args = args new_kwargs = { 'local_site_name': local_site_name, } if kwargs: new_kwargs.update(kwargs) try: return reverse(viewname, args=new_args, kwargs=new_kwargs, *func_args, **func_kwargs) except NoReverseMatch: # We'll try it again without those arguments. pass return reverse(viewname, args=args, kwargs=kwargs, *func_args, **func_kwargs) Allow local_site_reverse to take an actual LocalSite. local_site_reverse was able to take a LocalSite's name, or a request object, but if you actually had a LocalSite (or None), you'd have to write your own conditional to extract the name and pass it. Now, local_site_reverse can take a LocalSite. This saves a database query, if one is already available, and simplifies calling code. Reviewed at https://reviews.reviewboard.org/r/6302/from __future__ import unicode_literals from django.core.urlresolvers import NoReverseMatch, reverse def local_site_reverse(viewname, request=None, local_site_name=None, local_site=None, args=None, kwargs=None, *func_args, **func_kwargs): """Reverses a URL name, returning a working URL. This works much like Django's reverse(), but handles returning a localsite version of a URL when invoked with a request within a localsite. """ assert not (local_site_name and local_site) if request or local_site_name or local_site: if local_site: local_site_name = local_site.name elif request and not local_site_name: local_site_name = getattr(request, '_local_site_name', None) if local_site_name: if args: new_args = [local_site_name] + args new_kwargs = kwargs else: new_args = args new_kwargs = { 'local_site_name': local_site_name, } if kwargs: new_kwargs.update(kwargs) try: return reverse(viewname, args=new_args, kwargs=new_kwargs, *func_args, **func_kwargs) except NoReverseMatch: # We'll try it again without those arguments. pass return reverse(viewname, args=args, kwargs=kwargs, *func_args, **func_kwargs)
<commit_before>from __future__ import unicode_literals from django.core.urlresolvers import NoReverseMatch, reverse def local_site_reverse(viewname, request=None, local_site_name=None, args=None, kwargs=None, *func_args, **func_kwargs): """Reverses a URL name, returning a working URL. This works much like Django's reverse(), but handles returning a localsite version of a URL when invoked with a request within a localsite. """ if request or local_site_name: if request and not local_site_name: local_site_name = getattr(request, '_local_site_name', None) if local_site_name: if args: new_args = [local_site_name] + args new_kwargs = kwargs else: new_args = args new_kwargs = { 'local_site_name': local_site_name, } if kwargs: new_kwargs.update(kwargs) try: return reverse(viewname, args=new_args, kwargs=new_kwargs, *func_args, **func_kwargs) except NoReverseMatch: # We'll try it again without those arguments. pass return reverse(viewname, args=args, kwargs=kwargs, *func_args, **func_kwargs) <commit_msg>Allow local_site_reverse to take an actual LocalSite. local_site_reverse was able to take a LocalSite's name, or a request object, but if you actually had a LocalSite (or None), you'd have to write your own conditional to extract the name and pass it. Now, local_site_reverse can take a LocalSite. This saves a database query, if one is already available, and simplifies calling code. Reviewed at https://reviews.reviewboard.org/r/6302/<commit_after>from __future__ import unicode_literals from django.core.urlresolvers import NoReverseMatch, reverse def local_site_reverse(viewname, request=None, local_site_name=None, local_site=None, args=None, kwargs=None, *func_args, **func_kwargs): """Reverses a URL name, returning a working URL. This works much like Django's reverse(), but handles returning a localsite version of a URL when invoked with a request within a localsite. """ assert not (local_site_name and local_site) if request or local_site_name or local_site: if local_site: local_site_name = local_site.name elif request and not local_site_name: local_site_name = getattr(request, '_local_site_name', None) if local_site_name: if args: new_args = [local_site_name] + args new_kwargs = kwargs else: new_args = args new_kwargs = { 'local_site_name': local_site_name, } if kwargs: new_kwargs.update(kwargs) try: return reverse(viewname, args=new_args, kwargs=new_kwargs, *func_args, **func_kwargs) except NoReverseMatch: # We'll try it again without those arguments. pass return reverse(viewname, args=args, kwargs=kwargs, *func_args, **func_kwargs)
c488252462a2cb860111a4826f01883e7c16b3aa
numpy/distutils/setup.py
numpy/distutils/setup.py
#!/usr/bin/env python def configuration(parent_package='',top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('distutils',parent_package,top_path) config.add_subpackage('command') config.add_subpackage('fcompiler') config.add_data_dir('tests') config.add_data_files('site.cfg') config.make_config_py() return config if __name__ == '__main__': from numpy.distutils.core import setup setup(configuration=configuration)
#!/usr/bin/env python def configuration(parent_package='',top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('distutils',parent_package,top_path) config.add_subpackage('command') config.add_subpackage('fcompiler') config.add_data_dir('tests') config.add_data_files('site.cfg') config.add_data_files('mingw/gfortran_vs2003_hack.c') config.make_config_py() return config if __name__ == '__main__': from numpy.distutils.core import setup setup(configuration=configuration)
Make the gfortran/vs2003 hack source file known to distutils.
Make the gfortran/vs2003 hack source file known to distutils. git-svn-id: 77a43f9646713b91fea7788fad5dfbf67e151ece@6650 94b884b6-d6fd-0310-90d3-974f1d3f35e1
Python
bsd-3-clause
chadnetzer/numpy-gaurdro,Ademan/NumPy-GSoC,Ademan/NumPy-GSoC,teoliphant/numpy-refactor,illume/numpy3k,Ademan/NumPy-GSoC,teoliphant/numpy-refactor,chadnetzer/numpy-gaurdro,Ademan/NumPy-GSoC,illume/numpy3k,jasonmccampbell/numpy-refactor-sprint,efiring/numpy-work,efiring/numpy-work,chadnetzer/numpy-gaurdro,efiring/numpy-work,teoliphant/numpy-refactor,teoliphant/numpy-refactor,efiring/numpy-work,teoliphant/numpy-refactor,jasonmccampbell/numpy-refactor-sprint,illume/numpy3k,jasonmccampbell/numpy-refactor-sprint,illume/numpy3k,jasonmccampbell/numpy-refactor-sprint,chadnetzer/numpy-gaurdro
#!/usr/bin/env python def configuration(parent_package='',top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('distutils',parent_package,top_path) config.add_subpackage('command') config.add_subpackage('fcompiler') config.add_data_dir('tests') config.add_data_files('site.cfg') config.make_config_py() return config if __name__ == '__main__': from numpy.distutils.core import setup setup(configuration=configuration) Make the gfortran/vs2003 hack source file known to distutils. git-svn-id: 77a43f9646713b91fea7788fad5dfbf67e151ece@6650 94b884b6-d6fd-0310-90d3-974f1d3f35e1
#!/usr/bin/env python def configuration(parent_package='',top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('distutils',parent_package,top_path) config.add_subpackage('command') config.add_subpackage('fcompiler') config.add_data_dir('tests') config.add_data_files('site.cfg') config.add_data_files('mingw/gfortran_vs2003_hack.c') config.make_config_py() return config if __name__ == '__main__': from numpy.distutils.core import setup setup(configuration=configuration)
<commit_before>#!/usr/bin/env python def configuration(parent_package='',top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('distutils',parent_package,top_path) config.add_subpackage('command') config.add_subpackage('fcompiler') config.add_data_dir('tests') config.add_data_files('site.cfg') config.make_config_py() return config if __name__ == '__main__': from numpy.distutils.core import setup setup(configuration=configuration) <commit_msg>Make the gfortran/vs2003 hack source file known to distutils. git-svn-id: 77a43f9646713b91fea7788fad5dfbf67e151ece@6650 94b884b6-d6fd-0310-90d3-974f1d3f35e1<commit_after>
#!/usr/bin/env python def configuration(parent_package='',top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('distutils',parent_package,top_path) config.add_subpackage('command') config.add_subpackage('fcompiler') config.add_data_dir('tests') config.add_data_files('site.cfg') config.add_data_files('mingw/gfortran_vs2003_hack.c') config.make_config_py() return config if __name__ == '__main__': from numpy.distutils.core import setup setup(configuration=configuration)
#!/usr/bin/env python def configuration(parent_package='',top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('distutils',parent_package,top_path) config.add_subpackage('command') config.add_subpackage('fcompiler') config.add_data_dir('tests') config.add_data_files('site.cfg') config.make_config_py() return config if __name__ == '__main__': from numpy.distutils.core import setup setup(configuration=configuration) Make the gfortran/vs2003 hack source file known to distutils. git-svn-id: 77a43f9646713b91fea7788fad5dfbf67e151ece@6650 94b884b6-d6fd-0310-90d3-974f1d3f35e1#!/usr/bin/env python def configuration(parent_package='',top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('distutils',parent_package,top_path) config.add_subpackage('command') config.add_subpackage('fcompiler') config.add_data_dir('tests') config.add_data_files('site.cfg') config.add_data_files('mingw/gfortran_vs2003_hack.c') config.make_config_py() return config if __name__ == '__main__': from numpy.distutils.core import setup setup(configuration=configuration)
<commit_before>#!/usr/bin/env python def configuration(parent_package='',top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('distutils',parent_package,top_path) config.add_subpackage('command') config.add_subpackage('fcompiler') config.add_data_dir('tests') config.add_data_files('site.cfg') config.make_config_py() return config if __name__ == '__main__': from numpy.distutils.core import setup setup(configuration=configuration) <commit_msg>Make the gfortran/vs2003 hack source file known to distutils. git-svn-id: 77a43f9646713b91fea7788fad5dfbf67e151ece@6650 94b884b6-d6fd-0310-90d3-974f1d3f35e1<commit_after>#!/usr/bin/env python def configuration(parent_package='',top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('distutils',parent_package,top_path) config.add_subpackage('command') config.add_subpackage('fcompiler') config.add_data_dir('tests') config.add_data_files('site.cfg') config.add_data_files('mingw/gfortran_vs2003_hack.c') config.make_config_py() return config if __name__ == '__main__': from numpy.distutils.core import setup setup(configuration=configuration)
56215f65ea1513adfd574cf6fc3d118121119290
stestr/commands/__init__.py
stestr/commands/__init__.py
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from stestr.commands.failing import failing as failing_command from stestr.commands.init import init as init_command from stestr.commands.last import last as last_command from stestr.commands.list import list_command from stestr.commands.load import load as load_command from stestr.commands.run import run_command from stestr.commands.slowest import slowest as slowest_command __all__ = [failing_command, init_command, last_command, list_command, load_command, run_command, slowest_command]
Allow to import commands APIs from a single module
Allow to import commands APIs from a single module The APIs available to run various stestr commands are distributed in multiple modules. For ease of use make them all accessible from the commands module.
Python
apache-2.0
masayukig/stestr,mtreinish/stestr,mtreinish/stestr,masayukig/stestr
Allow to import commands APIs from a single module The APIs available to run various stestr commands are distributed in multiple modules. For ease of use make them all accessible from the commands module.
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from stestr.commands.failing import failing as failing_command from stestr.commands.init import init as init_command from stestr.commands.last import last as last_command from stestr.commands.list import list_command from stestr.commands.load import load as load_command from stestr.commands.run import run_command from stestr.commands.slowest import slowest as slowest_command __all__ = [failing_command, init_command, last_command, list_command, load_command, run_command, slowest_command]
<commit_before><commit_msg>Allow to import commands APIs from a single module The APIs available to run various stestr commands are distributed in multiple modules. For ease of use make them all accessible from the commands module.<commit_after>
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from stestr.commands.failing import failing as failing_command from stestr.commands.init import init as init_command from stestr.commands.last import last as last_command from stestr.commands.list import list_command from stestr.commands.load import load as load_command from stestr.commands.run import run_command from stestr.commands.slowest import slowest as slowest_command __all__ = [failing_command, init_command, last_command, list_command, load_command, run_command, slowest_command]
Allow to import commands APIs from a single module The APIs available to run various stestr commands are distributed in multiple modules. For ease of use make them all accessible from the commands module.# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from stestr.commands.failing import failing as failing_command from stestr.commands.init import init as init_command from stestr.commands.last import last as last_command from stestr.commands.list import list_command from stestr.commands.load import load as load_command from stestr.commands.run import run_command from stestr.commands.slowest import slowest as slowest_command __all__ = [failing_command, init_command, last_command, list_command, load_command, run_command, slowest_command]
<commit_before><commit_msg>Allow to import commands APIs from a single module The APIs available to run various stestr commands are distributed in multiple modules. For ease of use make them all accessible from the commands module.<commit_after># Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from stestr.commands.failing import failing as failing_command from stestr.commands.init import init as init_command from stestr.commands.last import last as last_command from stestr.commands.list import list_command from stestr.commands.load import load as load_command from stestr.commands.run import run_command from stestr.commands.slowest import slowest as slowest_command __all__ = [failing_command, init_command, last_command, list_command, load_command, run_command, slowest_command]
a7f7d8ff9f8279ec2c1f3981b1507001f1f94394
test/completion/docstring.py
test/completion/docstring.py
""" Test docstrings in functions and classes, which are used to infer types """ def f(a, b): """ asdfasdf :param a: blablabla :type a: str """ #? str() a #? b def g(a, b): """ asdfasdf Arguments: a (str): blablabla """ #? str() a #? b
""" Test docstrings in functions and classes, which are used to infer types """ def f(a, b): """ asdfasdf :param a: blablabla :type a: str """ #? str() a #? b def g(a, b): """ asdfasdf Arguments: a (str): blablabla """ #? str() a #? b def e(a, b): """ asdfasdf @type a: str @param a: blablabla """ #? str() a #? b
Add tests for epydoc formated dosctring
Add tests for epydoc formated dosctring
Python
mit
mfussenegger/jedi,flurischt/jedi,dwillmer/jedi,tjwei/jedi,WoLpH/jedi,jonashaag/jedi,mfussenegger/jedi,flurischt/jedi,jonashaag/jedi,tjwei/jedi,dwillmer/jedi,WoLpH/jedi
""" Test docstrings in functions and classes, which are used to infer types """ def f(a, b): """ asdfasdf :param a: blablabla :type a: str """ #? str() a #? b def g(a, b): """ asdfasdf Arguments: a (str): blablabla """ #? str() a #? b Add tests for epydoc formated dosctring
""" Test docstrings in functions and classes, which are used to infer types """ def f(a, b): """ asdfasdf :param a: blablabla :type a: str """ #? str() a #? b def g(a, b): """ asdfasdf Arguments: a (str): blablabla """ #? str() a #? b def e(a, b): """ asdfasdf @type a: str @param a: blablabla """ #? str() a #? b
<commit_before>""" Test docstrings in functions and classes, which are used to infer types """ def f(a, b): """ asdfasdf :param a: blablabla :type a: str """ #? str() a #? b def g(a, b): """ asdfasdf Arguments: a (str): blablabla """ #? str() a #? b <commit_msg>Add tests for epydoc formated dosctring<commit_after>
""" Test docstrings in functions and classes, which are used to infer types """ def f(a, b): """ asdfasdf :param a: blablabla :type a: str """ #? str() a #? b def g(a, b): """ asdfasdf Arguments: a (str): blablabla """ #? str() a #? b def e(a, b): """ asdfasdf @type a: str @param a: blablabla """ #? str() a #? b
""" Test docstrings in functions and classes, which are used to infer types """ def f(a, b): """ asdfasdf :param a: blablabla :type a: str """ #? str() a #? b def g(a, b): """ asdfasdf Arguments: a (str): blablabla """ #? str() a #? b Add tests for epydoc formated dosctring""" Test docstrings in functions and classes, which are used to infer types """ def f(a, b): """ asdfasdf :param a: blablabla :type a: str """ #? str() a #? b def g(a, b): """ asdfasdf Arguments: a (str): blablabla """ #? str() a #? b def e(a, b): """ asdfasdf @type a: str @param a: blablabla """ #? str() a #? b
<commit_before>""" Test docstrings in functions and classes, which are used to infer types """ def f(a, b): """ asdfasdf :param a: blablabla :type a: str """ #? str() a #? b def g(a, b): """ asdfasdf Arguments: a (str): blablabla """ #? str() a #? b <commit_msg>Add tests for epydoc formated dosctring<commit_after>""" Test docstrings in functions and classes, which are used to infer types """ def f(a, b): """ asdfasdf :param a: blablabla :type a: str """ #? str() a #? b def g(a, b): """ asdfasdf Arguments: a (str): blablabla """ #? str() a #? b def e(a, b): """ asdfasdf @type a: str @param a: blablabla """ #? str() a #? b
92d5991a37c3f269e9a7e59ab5edd90b45699930
test/style_test.py
test/style_test.py
import glob import os import pep8 class TestCodeFormat: def test_pep8(self): def match(*p): s = list(p) + ['*.py'] return glob.glob(os.path.join(*s)) pep8style = pep8.StyleGuide(config_file='setup.cfg') result = pep8style.check_files( match('examples') + match('test') + match('pagoda')) assert result.total_errors == 0
import glob import os import pep8 class TestCodeFormat: def test_pep8(self): def match(*p): s = list(p) + ['*.py'] return glob.glob(os.path.join(*s)) pep8style = pep8.StyleGuide(config_file='setup.cfg') result = pep8style.check_files( match('examples') + match('scripts') + match('test') + match('pagoda')) assert result.total_errors == 0
Include scripts in style check.
Include scripts in style check.
Python
mit
EmbodiedCognition/pagoda,EmbodiedCognition/pagoda
import glob import os import pep8 class TestCodeFormat: def test_pep8(self): def match(*p): s = list(p) + ['*.py'] return glob.glob(os.path.join(*s)) pep8style = pep8.StyleGuide(config_file='setup.cfg') result = pep8style.check_files( match('examples') + match('test') + match('pagoda')) assert result.total_errors == 0 Include scripts in style check.
import glob import os import pep8 class TestCodeFormat: def test_pep8(self): def match(*p): s = list(p) + ['*.py'] return glob.glob(os.path.join(*s)) pep8style = pep8.StyleGuide(config_file='setup.cfg') result = pep8style.check_files( match('examples') + match('scripts') + match('test') + match('pagoda')) assert result.total_errors == 0
<commit_before>import glob import os import pep8 class TestCodeFormat: def test_pep8(self): def match(*p): s = list(p) + ['*.py'] return glob.glob(os.path.join(*s)) pep8style = pep8.StyleGuide(config_file='setup.cfg') result = pep8style.check_files( match('examples') + match('test') + match('pagoda')) assert result.total_errors == 0 <commit_msg>Include scripts in style check.<commit_after>
import glob import os import pep8 class TestCodeFormat: def test_pep8(self): def match(*p): s = list(p) + ['*.py'] return glob.glob(os.path.join(*s)) pep8style = pep8.StyleGuide(config_file='setup.cfg') result = pep8style.check_files( match('examples') + match('scripts') + match('test') + match('pagoda')) assert result.total_errors == 0
import glob import os import pep8 class TestCodeFormat: def test_pep8(self): def match(*p): s = list(p) + ['*.py'] return glob.glob(os.path.join(*s)) pep8style = pep8.StyleGuide(config_file='setup.cfg') result = pep8style.check_files( match('examples') + match('test') + match('pagoda')) assert result.total_errors == 0 Include scripts in style check.import glob import os import pep8 class TestCodeFormat: def test_pep8(self): def match(*p): s = list(p) + ['*.py'] return glob.glob(os.path.join(*s)) pep8style = pep8.StyleGuide(config_file='setup.cfg') result = pep8style.check_files( match('examples') + match('scripts') + match('test') + match('pagoda')) assert result.total_errors == 0
<commit_before>import glob import os import pep8 class TestCodeFormat: def test_pep8(self): def match(*p): s = list(p) + ['*.py'] return glob.glob(os.path.join(*s)) pep8style = pep8.StyleGuide(config_file='setup.cfg') result = pep8style.check_files( match('examples') + match('test') + match('pagoda')) assert result.total_errors == 0 <commit_msg>Include scripts in style check.<commit_after>import glob import os import pep8 class TestCodeFormat: def test_pep8(self): def match(*p): s = list(p) + ['*.py'] return glob.glob(os.path.join(*s)) pep8style = pep8.StyleGuide(config_file='setup.cfg') result = pep8style.check_files( match('examples') + match('scripts') + match('test') + match('pagoda')) assert result.total_errors == 0
cfd710e0035c885fab926690d7ea450f3f9c3845
setup.py
setup.py
from distutils.core import setup import os, glob, string, shutil # Packages packages = ['neuroimaging', 'neuroimaging.statistics', 'neuroimaging.image', 'neuroimaging.reference', 'neuroimaging.data', 'neuroimaging.image.formats', 'neuroimaging.image.formats.analyze', 'neuroimaging.fmri', 'neuroimaging.fmri.fmristat', 'neuroimaging.visualization', 'neuroimaging.visualization.cmap'] def main(): setup (name = 'neuroimaging', version = '0.01a', description = 'This is a neuroimaging python package', author = 'Various, one of whom is Jonathan Taylor', author_email = 'jonathan.taylor@stanford.edu', ext_package = 'neuroimaging', packages=packages, package_dir = {'neuroimaging': 'lib'}, url = 'http://neuroimaging.scipy.org', long_description = ''' ''') if __name__ == "__main__": main()
import os, glob, string, shutil from distutils.core import setup # Packages packages = ['neuroimaging', 'neuroimaging.statistics', 'neuroimaging.image', 'neuroimaging.reference', 'neuroimaging.data', 'neuroimaging.image.formats', 'neuroimaging.image.formats.analyze', 'neuroimaging.fmri', 'neuroimaging.fmri.fmristat', 'neuroimaging.visualization', 'neuroimaging.visualization.cmap'] def main(): setup (name = 'neuroimaging', version = '0.01a', description = 'This is a neuroimaging python package', author = 'Various, one of whom is Jonathan Taylor', author_email = 'jonathan.taylor@stanford.edu', ext_package = 'neuroimaging', packages=packages, package_dir = {'neuroimaging': 'lib'}, url = 'http://neuroimaging.scipy.org', long_description = ''' ''') if __name__ == "__main__": main()
Test edit - to check svn email hook
Test edit - to check svn email hook
Python
bsd-3-clause
matthew-brett/draft-statsmodels,matthew-brett/draft-statsmodels
from distutils.core import setup import os, glob, string, shutil # Packages packages = ['neuroimaging', 'neuroimaging.statistics', 'neuroimaging.image', 'neuroimaging.reference', 'neuroimaging.data', 'neuroimaging.image.formats', 'neuroimaging.image.formats.analyze', 'neuroimaging.fmri', 'neuroimaging.fmri.fmristat', 'neuroimaging.visualization', 'neuroimaging.visualization.cmap'] def main(): setup (name = 'neuroimaging', version = '0.01a', description = 'This is a neuroimaging python package', author = 'Various, one of whom is Jonathan Taylor', author_email = 'jonathan.taylor@stanford.edu', ext_package = 'neuroimaging', packages=packages, package_dir = {'neuroimaging': 'lib'}, url = 'http://neuroimaging.scipy.org', long_description = ''' ''') if __name__ == "__main__": main() Test edit - to check svn email hook
import os, glob, string, shutil from distutils.core import setup # Packages packages = ['neuroimaging', 'neuroimaging.statistics', 'neuroimaging.image', 'neuroimaging.reference', 'neuroimaging.data', 'neuroimaging.image.formats', 'neuroimaging.image.formats.analyze', 'neuroimaging.fmri', 'neuroimaging.fmri.fmristat', 'neuroimaging.visualization', 'neuroimaging.visualization.cmap'] def main(): setup (name = 'neuroimaging', version = '0.01a', description = 'This is a neuroimaging python package', author = 'Various, one of whom is Jonathan Taylor', author_email = 'jonathan.taylor@stanford.edu', ext_package = 'neuroimaging', packages=packages, package_dir = {'neuroimaging': 'lib'}, url = 'http://neuroimaging.scipy.org', long_description = ''' ''') if __name__ == "__main__": main()
<commit_before>from distutils.core import setup import os, glob, string, shutil # Packages packages = ['neuroimaging', 'neuroimaging.statistics', 'neuroimaging.image', 'neuroimaging.reference', 'neuroimaging.data', 'neuroimaging.image.formats', 'neuroimaging.image.formats.analyze', 'neuroimaging.fmri', 'neuroimaging.fmri.fmristat', 'neuroimaging.visualization', 'neuroimaging.visualization.cmap'] def main(): setup (name = 'neuroimaging', version = '0.01a', description = 'This is a neuroimaging python package', author = 'Various, one of whom is Jonathan Taylor', author_email = 'jonathan.taylor@stanford.edu', ext_package = 'neuroimaging', packages=packages, package_dir = {'neuroimaging': 'lib'}, url = 'http://neuroimaging.scipy.org', long_description = ''' ''') if __name__ == "__main__": main() <commit_msg>Test edit - to check svn email hook<commit_after>
import os, glob, string, shutil from distutils.core import setup # Packages packages = ['neuroimaging', 'neuroimaging.statistics', 'neuroimaging.image', 'neuroimaging.reference', 'neuroimaging.data', 'neuroimaging.image.formats', 'neuroimaging.image.formats.analyze', 'neuroimaging.fmri', 'neuroimaging.fmri.fmristat', 'neuroimaging.visualization', 'neuroimaging.visualization.cmap'] def main(): setup (name = 'neuroimaging', version = '0.01a', description = 'This is a neuroimaging python package', author = 'Various, one of whom is Jonathan Taylor', author_email = 'jonathan.taylor@stanford.edu', ext_package = 'neuroimaging', packages=packages, package_dir = {'neuroimaging': 'lib'}, url = 'http://neuroimaging.scipy.org', long_description = ''' ''') if __name__ == "__main__": main()
from distutils.core import setup import os, glob, string, shutil # Packages packages = ['neuroimaging', 'neuroimaging.statistics', 'neuroimaging.image', 'neuroimaging.reference', 'neuroimaging.data', 'neuroimaging.image.formats', 'neuroimaging.image.formats.analyze', 'neuroimaging.fmri', 'neuroimaging.fmri.fmristat', 'neuroimaging.visualization', 'neuroimaging.visualization.cmap'] def main(): setup (name = 'neuroimaging', version = '0.01a', description = 'This is a neuroimaging python package', author = 'Various, one of whom is Jonathan Taylor', author_email = 'jonathan.taylor@stanford.edu', ext_package = 'neuroimaging', packages=packages, package_dir = {'neuroimaging': 'lib'}, url = 'http://neuroimaging.scipy.org', long_description = ''' ''') if __name__ == "__main__": main() Test edit - to check svn email hookimport os, glob, string, shutil from distutils.core import setup # Packages packages = ['neuroimaging', 'neuroimaging.statistics', 'neuroimaging.image', 'neuroimaging.reference', 'neuroimaging.data', 'neuroimaging.image.formats', 'neuroimaging.image.formats.analyze', 'neuroimaging.fmri', 'neuroimaging.fmri.fmristat', 'neuroimaging.visualization', 'neuroimaging.visualization.cmap'] def main(): setup (name = 'neuroimaging', version = '0.01a', description = 'This is a neuroimaging python package', author = 'Various, one of whom is Jonathan Taylor', author_email = 'jonathan.taylor@stanford.edu', ext_package = 'neuroimaging', packages=packages, package_dir = {'neuroimaging': 'lib'}, url = 'http://neuroimaging.scipy.org', long_description = ''' ''') if __name__ == "__main__": main()
<commit_before>from distutils.core import setup import os, glob, string, shutil # Packages packages = ['neuroimaging', 'neuroimaging.statistics', 'neuroimaging.image', 'neuroimaging.reference', 'neuroimaging.data', 'neuroimaging.image.formats', 'neuroimaging.image.formats.analyze', 'neuroimaging.fmri', 'neuroimaging.fmri.fmristat', 'neuroimaging.visualization', 'neuroimaging.visualization.cmap'] def main(): setup (name = 'neuroimaging', version = '0.01a', description = 'This is a neuroimaging python package', author = 'Various, one of whom is Jonathan Taylor', author_email = 'jonathan.taylor@stanford.edu', ext_package = 'neuroimaging', packages=packages, package_dir = {'neuroimaging': 'lib'}, url = 'http://neuroimaging.scipy.org', long_description = ''' ''') if __name__ == "__main__": main() <commit_msg>Test edit - to check svn email hook<commit_after>import os, glob, string, shutil from distutils.core import setup # Packages packages = ['neuroimaging', 'neuroimaging.statistics', 'neuroimaging.image', 'neuroimaging.reference', 'neuroimaging.data', 'neuroimaging.image.formats', 'neuroimaging.image.formats.analyze', 'neuroimaging.fmri', 'neuroimaging.fmri.fmristat', 'neuroimaging.visualization', 'neuroimaging.visualization.cmap'] def main(): setup (name = 'neuroimaging', version = '0.01a', description = 'This is a neuroimaging python package', author = 'Various, one of whom is Jonathan Taylor', author_email = 'jonathan.taylor@stanford.edu', ext_package = 'neuroimaging', packages=packages, package_dir = {'neuroimaging': 'lib'}, url = 'http://neuroimaging.scipy.org', long_description = ''' ''') if __name__ == "__main__": main()
a848ab0489b0419592dec6f2b5b728863bf0bebf
setup.py
setup.py
from distutils.core import setup setup( name = 'missingno', packages = ['missingno'], # this must be the same as the name above install_requires=['numpy', 'matplotlib'], py_modules=['missingno'], version = '0.1.1', description = 'Missing data visualization module for Python.', author = 'Aleksey Bilogur', author_email = 'aleksey.bilogur@gmail.com', url = 'https://github.com/ResidentMario/missingno', download_url = 'https://github.com/ResidentMario/missingno/tarball/0.1.1', keywords = ['data', 'data visualization', 'data analysis', 'missing data', 'data science', 'pandas', 'python', 'jupyter'], classifiers = [], )
from setuptools import setup setup( name = 'missingno', packages = ['missingno'], # this must be the same as the name above install_requires=['numpy', 'matplotlib'], py_modules=['missingno'], version = '0.1.1', description = 'Missing data visualization module for Python.', author = 'Aleksey Bilogur', author_email = 'aleksey.bilogur@gmail.com', url = 'https://github.com/ResidentMario/missingno', download_url = 'https://github.com/ResidentMario/missingno/tarball/0.1.1', keywords = ['data', 'data visualization', 'data analysis', 'missing data', 'data science', 'pandas', 'python', 'jupyter'], classifiers = [], )
Switch to 'more correct' distributor.
Switch to 'more correct' distributor.
Python
mit
ResidentMario/missingno
from distutils.core import setup setup( name = 'missingno', packages = ['missingno'], # this must be the same as the name above install_requires=['numpy', 'matplotlib'], py_modules=['missingno'], version = '0.1.1', description = 'Missing data visualization module for Python.', author = 'Aleksey Bilogur', author_email = 'aleksey.bilogur@gmail.com', url = 'https://github.com/ResidentMario/missingno', download_url = 'https://github.com/ResidentMario/missingno/tarball/0.1.1', keywords = ['data', 'data visualization', 'data analysis', 'missing data', 'data science', 'pandas', 'python', 'jupyter'], classifiers = [], )Switch to 'more correct' distributor.
from setuptools import setup setup( name = 'missingno', packages = ['missingno'], # this must be the same as the name above install_requires=['numpy', 'matplotlib'], py_modules=['missingno'], version = '0.1.1', description = 'Missing data visualization module for Python.', author = 'Aleksey Bilogur', author_email = 'aleksey.bilogur@gmail.com', url = 'https://github.com/ResidentMario/missingno', download_url = 'https://github.com/ResidentMario/missingno/tarball/0.1.1', keywords = ['data', 'data visualization', 'data analysis', 'missing data', 'data science', 'pandas', 'python', 'jupyter'], classifiers = [], )
<commit_before>from distutils.core import setup setup( name = 'missingno', packages = ['missingno'], # this must be the same as the name above install_requires=['numpy', 'matplotlib'], py_modules=['missingno'], version = '0.1.1', description = 'Missing data visualization module for Python.', author = 'Aleksey Bilogur', author_email = 'aleksey.bilogur@gmail.com', url = 'https://github.com/ResidentMario/missingno', download_url = 'https://github.com/ResidentMario/missingno/tarball/0.1.1', keywords = ['data', 'data visualization', 'data analysis', 'missing data', 'data science', 'pandas', 'python', 'jupyter'], classifiers = [], )<commit_msg>Switch to 'more correct' distributor.<commit_after>
from setuptools import setup setup( name = 'missingno', packages = ['missingno'], # this must be the same as the name above install_requires=['numpy', 'matplotlib'], py_modules=['missingno'], version = '0.1.1', description = 'Missing data visualization module for Python.', author = 'Aleksey Bilogur', author_email = 'aleksey.bilogur@gmail.com', url = 'https://github.com/ResidentMario/missingno', download_url = 'https://github.com/ResidentMario/missingno/tarball/0.1.1', keywords = ['data', 'data visualization', 'data analysis', 'missing data', 'data science', 'pandas', 'python', 'jupyter'], classifiers = [], )
from distutils.core import setup setup( name = 'missingno', packages = ['missingno'], # this must be the same as the name above install_requires=['numpy', 'matplotlib'], py_modules=['missingno'], version = '0.1.1', description = 'Missing data visualization module for Python.', author = 'Aleksey Bilogur', author_email = 'aleksey.bilogur@gmail.com', url = 'https://github.com/ResidentMario/missingno', download_url = 'https://github.com/ResidentMario/missingno/tarball/0.1.1', keywords = ['data', 'data visualization', 'data analysis', 'missing data', 'data science', 'pandas', 'python', 'jupyter'], classifiers = [], )Switch to 'more correct' distributor.from setuptools import setup setup( name = 'missingno', packages = ['missingno'], # this must be the same as the name above install_requires=['numpy', 'matplotlib'], py_modules=['missingno'], version = '0.1.1', description = 'Missing data visualization module for Python.', author = 'Aleksey Bilogur', author_email = 'aleksey.bilogur@gmail.com', url = 'https://github.com/ResidentMario/missingno', download_url = 'https://github.com/ResidentMario/missingno/tarball/0.1.1', keywords = ['data', 'data visualization', 'data analysis', 'missing data', 'data science', 'pandas', 'python', 'jupyter'], classifiers = [], )
<commit_before>from distutils.core import setup setup( name = 'missingno', packages = ['missingno'], # this must be the same as the name above install_requires=['numpy', 'matplotlib'], py_modules=['missingno'], version = '0.1.1', description = 'Missing data visualization module for Python.', author = 'Aleksey Bilogur', author_email = 'aleksey.bilogur@gmail.com', url = 'https://github.com/ResidentMario/missingno', download_url = 'https://github.com/ResidentMario/missingno/tarball/0.1.1', keywords = ['data', 'data visualization', 'data analysis', 'missing data', 'data science', 'pandas', 'python', 'jupyter'], classifiers = [], )<commit_msg>Switch to 'more correct' distributor.<commit_after>from setuptools import setup setup( name = 'missingno', packages = ['missingno'], # this must be the same as the name above install_requires=['numpy', 'matplotlib'], py_modules=['missingno'], version = '0.1.1', description = 'Missing data visualization module for Python.', author = 'Aleksey Bilogur', author_email = 'aleksey.bilogur@gmail.com', url = 'https://github.com/ResidentMario/missingno', download_url = 'https://github.com/ResidentMario/missingno/tarball/0.1.1', keywords = ['data', 'data visualization', 'data analysis', 'missing data', 'data science', 'pandas', 'python', 'jupyter'], classifiers = [], )
a74ad2850521eea430c0144b433707e0e9d5bb88
setup.py
setup.py
from setuptools import setup, find_packages exec(open('keras_vggface/version.py').read()) setup(name='keras_vggface', version=__version__, description='VGGFace implementation with Keras framework', url='https://github.com/rcmalli/keras-vggface', author='Refik Can MALLI', author_email = "mallir@itu.edu.tr", license='MIT', keywords = ['keras', 'vggface', 'deeplearning'], packages=find_packages(exclude=["temp", "test", "data", "visualize"]), zip_safe=False, install_requires=['numpy>=1.9.1', 'scipy>=0.14', 'h5py', 'pillow', 'tensorflow', 'keras', 'six>=1.9.0', 'pyyaml'])
from setuptools import setup, find_packages exec(open('keras_vggface/version.py').read()) setup( name='keras_vggface', version=__version__, description='VGGFace implementation with Keras framework', url='https://github.com/rcmalli/keras-vggface', author='Refik Can MALLI', author_email="mallir@itu.edu.tr", license='MIT', keywords=['keras', 'vggface', 'deeplearning'], packages=find_packages(exclude=["temp", "test", "data", "visualize"]), zip_safe=False, install_requires=[ 'numpy>=1.9.1', 'scipy>=0.14', 'h5py', 'pillow', 'keras', 'six>=1.9.0', 'pyyaml' ], extras_require={ "tf": ["tensorflow"], "tf_gpu": ["tensorflow-gpu"], })
Make tensorflow be optional to allow using tensorflow-gpu
Make tensorflow be optional to allow using tensorflow-gpu
Python
mit
rcmalli/keras-vggface
from setuptools import setup, find_packages exec(open('keras_vggface/version.py').read()) setup(name='keras_vggface', version=__version__, description='VGGFace implementation with Keras framework', url='https://github.com/rcmalli/keras-vggface', author='Refik Can MALLI', author_email = "mallir@itu.edu.tr", license='MIT', keywords = ['keras', 'vggface', 'deeplearning'], packages=find_packages(exclude=["temp", "test", "data", "visualize"]), zip_safe=False, install_requires=['numpy>=1.9.1', 'scipy>=0.14', 'h5py', 'pillow', 'tensorflow', 'keras', 'six>=1.9.0', 'pyyaml']) Make tensorflow be optional to allow using tensorflow-gpu
from setuptools import setup, find_packages exec(open('keras_vggface/version.py').read()) setup( name='keras_vggface', version=__version__, description='VGGFace implementation with Keras framework', url='https://github.com/rcmalli/keras-vggface', author='Refik Can MALLI', author_email="mallir@itu.edu.tr", license='MIT', keywords=['keras', 'vggface', 'deeplearning'], packages=find_packages(exclude=["temp", "test", "data", "visualize"]), zip_safe=False, install_requires=[ 'numpy>=1.9.1', 'scipy>=0.14', 'h5py', 'pillow', 'keras', 'six>=1.9.0', 'pyyaml' ], extras_require={ "tf": ["tensorflow"], "tf_gpu": ["tensorflow-gpu"], })
<commit_before>from setuptools import setup, find_packages exec(open('keras_vggface/version.py').read()) setup(name='keras_vggface', version=__version__, description='VGGFace implementation with Keras framework', url='https://github.com/rcmalli/keras-vggface', author='Refik Can MALLI', author_email = "mallir@itu.edu.tr", license='MIT', keywords = ['keras', 'vggface', 'deeplearning'], packages=find_packages(exclude=["temp", "test", "data", "visualize"]), zip_safe=False, install_requires=['numpy>=1.9.1', 'scipy>=0.14', 'h5py', 'pillow', 'tensorflow', 'keras', 'six>=1.9.0', 'pyyaml']) <commit_msg>Make tensorflow be optional to allow using tensorflow-gpu<commit_after>
from setuptools import setup, find_packages exec(open('keras_vggface/version.py').read()) setup( name='keras_vggface', version=__version__, description='VGGFace implementation with Keras framework', url='https://github.com/rcmalli/keras-vggface', author='Refik Can MALLI', author_email="mallir@itu.edu.tr", license='MIT', keywords=['keras', 'vggface', 'deeplearning'], packages=find_packages(exclude=["temp", "test", "data", "visualize"]), zip_safe=False, install_requires=[ 'numpy>=1.9.1', 'scipy>=0.14', 'h5py', 'pillow', 'keras', 'six>=1.9.0', 'pyyaml' ], extras_require={ "tf": ["tensorflow"], "tf_gpu": ["tensorflow-gpu"], })
from setuptools import setup, find_packages exec(open('keras_vggface/version.py').read()) setup(name='keras_vggface', version=__version__, description='VGGFace implementation with Keras framework', url='https://github.com/rcmalli/keras-vggface', author='Refik Can MALLI', author_email = "mallir@itu.edu.tr", license='MIT', keywords = ['keras', 'vggface', 'deeplearning'], packages=find_packages(exclude=["temp", "test", "data", "visualize"]), zip_safe=False, install_requires=['numpy>=1.9.1', 'scipy>=0.14', 'h5py', 'pillow', 'tensorflow', 'keras', 'six>=1.9.0', 'pyyaml']) Make tensorflow be optional to allow using tensorflow-gpufrom setuptools import setup, find_packages exec(open('keras_vggface/version.py').read()) setup( name='keras_vggface', version=__version__, description='VGGFace implementation with Keras framework', url='https://github.com/rcmalli/keras-vggface', author='Refik Can MALLI', author_email="mallir@itu.edu.tr", license='MIT', keywords=['keras', 'vggface', 'deeplearning'], packages=find_packages(exclude=["temp", "test", "data", "visualize"]), zip_safe=False, install_requires=[ 'numpy>=1.9.1', 'scipy>=0.14', 'h5py', 'pillow', 'keras', 'six>=1.9.0', 'pyyaml' ], extras_require={ "tf": ["tensorflow"], "tf_gpu": ["tensorflow-gpu"], })
<commit_before>from setuptools import setup, find_packages exec(open('keras_vggface/version.py').read()) setup(name='keras_vggface', version=__version__, description='VGGFace implementation with Keras framework', url='https://github.com/rcmalli/keras-vggface', author='Refik Can MALLI', author_email = "mallir@itu.edu.tr", license='MIT', keywords = ['keras', 'vggface', 'deeplearning'], packages=find_packages(exclude=["temp", "test", "data", "visualize"]), zip_safe=False, install_requires=['numpy>=1.9.1', 'scipy>=0.14', 'h5py', 'pillow', 'tensorflow', 'keras', 'six>=1.9.0', 'pyyaml']) <commit_msg>Make tensorflow be optional to allow using tensorflow-gpu<commit_after>from setuptools import setup, find_packages exec(open('keras_vggface/version.py').read()) setup( name='keras_vggface', version=__version__, description='VGGFace implementation with Keras framework', url='https://github.com/rcmalli/keras-vggface', author='Refik Can MALLI', author_email="mallir@itu.edu.tr", license='MIT', keywords=['keras', 'vggface', 'deeplearning'], packages=find_packages(exclude=["temp", "test", "data", "visualize"]), zip_safe=False, install_requires=[ 'numpy>=1.9.1', 'scipy>=0.14', 'h5py', 'pillow', 'keras', 'six>=1.9.0', 'pyyaml' ], extras_require={ "tf": ["tensorflow"], "tf_gpu": ["tensorflow-gpu"], })
ce966cd73ff3dc37fd534fc7ae41e47b05782a4e
setup.py
setup.py
from setuptools import setup, find_packages def read(fname): with open(fname) as fp: content = fp.read() return content setup( name='munerator', version="0.10.11", description='Manager of OpenArena battles', long_description=read("README.rst"), author='Johan Bloemberg', author_email='mail@ijohan.nl', url='https://github.com/aequitas/munerator', install_requires=[ 'six', 'docopt', 'pyzmq', 'webcolors', 'python-Levenshtein', 'eve==0.3', 'flask', 'tornado', 'mongoengine', 'eve-mongoengine', 'python-dateutil', 'rcfile' ], packages=find_packages(), license=read("LICENSE"), include_package_data=True, classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Natural Language :: English', "Programming Language :: Python :: 2", 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: Implementation :: CPython', ], entry_points={ 'console_scripts': [ "munerator = munerator:main", ] }, )
from setuptools import setup, find_packages def read(fname): with open(fname) as fp: content = fp.read() return content setup( name='munerator', version="0.10.12", description='Manager of OpenArena battles', long_description=read("README.rst"), author='Johan Bloemberg', author_email='mail@ijohan.nl', url='https://github.com/aequitas/munerator', install_requires=[ 'six', 'docopt', 'pyzmq', 'webcolors', 'python-Levenshtein', 'python-dateutil', 'rcfile' ], extras_require={ 'restapi': [ 'mongoengine', 'eve-mongoengine', 'eve==0.3', 'flask', ], 'store': [ 'mongoengine', 'eve-mongoengine', 'eve==0.3', ] }, packages=find_packages(), license=read("LICENSE"), include_package_data=True, classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Natural Language :: English', "Programming Language :: Python :: 2", 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: Implementation :: CPython', ], entry_points={ 'console_scripts': [ "munerator = munerator:main", ] }, )
Split extra requires for store and restapi
Split extra requires for store and restapi
Python
mit
aequitas/munerator,aequitas/munerator,aequitas/munerator,aequitas/munerator
from setuptools import setup, find_packages def read(fname): with open(fname) as fp: content = fp.read() return content setup( name='munerator', version="0.10.11", description='Manager of OpenArena battles', long_description=read("README.rst"), author='Johan Bloemberg', author_email='mail@ijohan.nl', url='https://github.com/aequitas/munerator', install_requires=[ 'six', 'docopt', 'pyzmq', 'webcolors', 'python-Levenshtein', 'eve==0.3', 'flask', 'tornado', 'mongoengine', 'eve-mongoengine', 'python-dateutil', 'rcfile' ], packages=find_packages(), license=read("LICENSE"), include_package_data=True, classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Natural Language :: English', "Programming Language :: Python :: 2", 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: Implementation :: CPython', ], entry_points={ 'console_scripts': [ "munerator = munerator:main", ] }, ) Split extra requires for store and restapi
from setuptools import setup, find_packages def read(fname): with open(fname) as fp: content = fp.read() return content setup( name='munerator', version="0.10.12", description='Manager of OpenArena battles', long_description=read("README.rst"), author='Johan Bloemberg', author_email='mail@ijohan.nl', url='https://github.com/aequitas/munerator', install_requires=[ 'six', 'docopt', 'pyzmq', 'webcolors', 'python-Levenshtein', 'python-dateutil', 'rcfile' ], extras_require={ 'restapi': [ 'mongoengine', 'eve-mongoengine', 'eve==0.3', 'flask', ], 'store': [ 'mongoengine', 'eve-mongoengine', 'eve==0.3', ] }, packages=find_packages(), license=read("LICENSE"), include_package_data=True, classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Natural Language :: English', "Programming Language :: Python :: 2", 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: Implementation :: CPython', ], entry_points={ 'console_scripts': [ "munerator = munerator:main", ] }, )
<commit_before>from setuptools import setup, find_packages def read(fname): with open(fname) as fp: content = fp.read() return content setup( name='munerator', version="0.10.11", description='Manager of OpenArena battles', long_description=read("README.rst"), author='Johan Bloemberg', author_email='mail@ijohan.nl', url='https://github.com/aequitas/munerator', install_requires=[ 'six', 'docopt', 'pyzmq', 'webcolors', 'python-Levenshtein', 'eve==0.3', 'flask', 'tornado', 'mongoengine', 'eve-mongoengine', 'python-dateutil', 'rcfile' ], packages=find_packages(), license=read("LICENSE"), include_package_data=True, classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Natural Language :: English', "Programming Language :: Python :: 2", 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: Implementation :: CPython', ], entry_points={ 'console_scripts': [ "munerator = munerator:main", ] }, ) <commit_msg>Split extra requires for store and restapi<commit_after>
from setuptools import setup, find_packages def read(fname): with open(fname) as fp: content = fp.read() return content setup( name='munerator', version="0.10.12", description='Manager of OpenArena battles', long_description=read("README.rst"), author='Johan Bloemberg', author_email='mail@ijohan.nl', url='https://github.com/aequitas/munerator', install_requires=[ 'six', 'docopt', 'pyzmq', 'webcolors', 'python-Levenshtein', 'python-dateutil', 'rcfile' ], extras_require={ 'restapi': [ 'mongoengine', 'eve-mongoengine', 'eve==0.3', 'flask', ], 'store': [ 'mongoengine', 'eve-mongoengine', 'eve==0.3', ] }, packages=find_packages(), license=read("LICENSE"), include_package_data=True, classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Natural Language :: English', "Programming Language :: Python :: 2", 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: Implementation :: CPython', ], entry_points={ 'console_scripts': [ "munerator = munerator:main", ] }, )
from setuptools import setup, find_packages def read(fname): with open(fname) as fp: content = fp.read() return content setup( name='munerator', version="0.10.11", description='Manager of OpenArena battles', long_description=read("README.rst"), author='Johan Bloemberg', author_email='mail@ijohan.nl', url='https://github.com/aequitas/munerator', install_requires=[ 'six', 'docopt', 'pyzmq', 'webcolors', 'python-Levenshtein', 'eve==0.3', 'flask', 'tornado', 'mongoengine', 'eve-mongoengine', 'python-dateutil', 'rcfile' ], packages=find_packages(), license=read("LICENSE"), include_package_data=True, classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Natural Language :: English', "Programming Language :: Python :: 2", 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: Implementation :: CPython', ], entry_points={ 'console_scripts': [ "munerator = munerator:main", ] }, ) Split extra requires for store and restapifrom setuptools import setup, find_packages def read(fname): with open(fname) as fp: content = fp.read() return content setup( name='munerator', version="0.10.12", description='Manager of OpenArena battles', long_description=read("README.rst"), author='Johan Bloemberg', author_email='mail@ijohan.nl', url='https://github.com/aequitas/munerator', install_requires=[ 'six', 'docopt', 'pyzmq', 'webcolors', 'python-Levenshtein', 'python-dateutil', 'rcfile' ], extras_require={ 'restapi': [ 'mongoengine', 'eve-mongoengine', 'eve==0.3', 'flask', ], 'store': [ 'mongoengine', 'eve-mongoengine', 'eve==0.3', ] }, packages=find_packages(), license=read("LICENSE"), include_package_data=True, classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Natural Language :: English', "Programming Language :: Python :: 2", 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: Implementation :: CPython', ], entry_points={ 'console_scripts': [ "munerator = munerator:main", ] }, )
<commit_before>from setuptools import setup, find_packages def read(fname): with open(fname) as fp: content = fp.read() return content setup( name='munerator', version="0.10.11", description='Manager of OpenArena battles', long_description=read("README.rst"), author='Johan Bloemberg', author_email='mail@ijohan.nl', url='https://github.com/aequitas/munerator', install_requires=[ 'six', 'docopt', 'pyzmq', 'webcolors', 'python-Levenshtein', 'eve==0.3', 'flask', 'tornado', 'mongoengine', 'eve-mongoengine', 'python-dateutil', 'rcfile' ], packages=find_packages(), license=read("LICENSE"), include_package_data=True, classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Natural Language :: English', "Programming Language :: Python :: 2", 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: Implementation :: CPython', ], entry_points={ 'console_scripts': [ "munerator = munerator:main", ] }, ) <commit_msg>Split extra requires for store and restapi<commit_after>from setuptools import setup, find_packages def read(fname): with open(fname) as fp: content = fp.read() return content setup( name='munerator', version="0.10.12", description='Manager of OpenArena battles', long_description=read("README.rst"), author='Johan Bloemberg', author_email='mail@ijohan.nl', url='https://github.com/aequitas/munerator', install_requires=[ 'six', 'docopt', 'pyzmq', 'webcolors', 'python-Levenshtein', 'python-dateutil', 'rcfile' ], extras_require={ 'restapi': [ 'mongoengine', 'eve-mongoengine', 'eve==0.3', 'flask', ], 'store': [ 'mongoengine', 'eve-mongoengine', 'eve==0.3', ] }, packages=find_packages(), license=read("LICENSE"), include_package_data=True, classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Natural Language :: English', "Programming Language :: Python :: 2", 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Programming Language :: Python :: Implementation :: CPython', ], entry_points={ 'console_scripts': [ "munerator = munerator:main", ] }, )
4ffdbaf30c46ff886c7eb2e9e9686679a5045e49
setup.py
setup.py
from setuptools import setup PACKAGE = 'TracBacklog' VERSION = '0.1' setup(name=PACKAGE, version=VERSION, packages=['backlog'], package_data={ 'backlog': [ 'htdocs/css/*.css', 'htdocs/img/*.png', 'htdocs/js/*.js', 'htdocs/js/dojo/*.js', 'htdocs/js/dojo/dnd/*.js', 'htdocs/js/dojo/date/*.js', 'templates/*.html', 'scripts/*' ]}, entry_points=''' [trac.plugins] backlog = backlog.web_ui ''')
from setuptools import setup PACKAGE = 'TracBacklog' VERSION = '0.1' setup(name=PACKAGE, version=VERSION, packages=['backlog'], package_data={ 'backlog': [ 'htdocs/css/*.css', 'htdocs/img/*.png', 'htdocs/js/*.js', 'htdocs/js/dojo/*.js', 'htdocs/js/dojo/dnd/*.js', 'htdocs/js/dojo/date/*.js', 'templates/*.html', 'scripts/*' ]}, entry_points={ 'trac.plugins': ['backlog = backlog.web_ui'] } )
Use a different method of specifying the entry point.
Use a different method of specifying the entry point.
Python
bsd-3-clause
jszakmeister/trac-backlog,jszakmeister/trac-backlog
from setuptools import setup PACKAGE = 'TracBacklog' VERSION = '0.1' setup(name=PACKAGE, version=VERSION, packages=['backlog'], package_data={ 'backlog': [ 'htdocs/css/*.css', 'htdocs/img/*.png', 'htdocs/js/*.js', 'htdocs/js/dojo/*.js', 'htdocs/js/dojo/dnd/*.js', 'htdocs/js/dojo/date/*.js', 'templates/*.html', 'scripts/*' ]}, entry_points=''' [trac.plugins] backlog = backlog.web_ui ''') Use a different method of specifying the entry point.
from setuptools import setup PACKAGE = 'TracBacklog' VERSION = '0.1' setup(name=PACKAGE, version=VERSION, packages=['backlog'], package_data={ 'backlog': [ 'htdocs/css/*.css', 'htdocs/img/*.png', 'htdocs/js/*.js', 'htdocs/js/dojo/*.js', 'htdocs/js/dojo/dnd/*.js', 'htdocs/js/dojo/date/*.js', 'templates/*.html', 'scripts/*' ]}, entry_points={ 'trac.plugins': ['backlog = backlog.web_ui'] } )
<commit_before>from setuptools import setup PACKAGE = 'TracBacklog' VERSION = '0.1' setup(name=PACKAGE, version=VERSION, packages=['backlog'], package_data={ 'backlog': [ 'htdocs/css/*.css', 'htdocs/img/*.png', 'htdocs/js/*.js', 'htdocs/js/dojo/*.js', 'htdocs/js/dojo/dnd/*.js', 'htdocs/js/dojo/date/*.js', 'templates/*.html', 'scripts/*' ]}, entry_points=''' [trac.plugins] backlog = backlog.web_ui ''') <commit_msg>Use a different method of specifying the entry point.<commit_after>
from setuptools import setup PACKAGE = 'TracBacklog' VERSION = '0.1' setup(name=PACKAGE, version=VERSION, packages=['backlog'], package_data={ 'backlog': [ 'htdocs/css/*.css', 'htdocs/img/*.png', 'htdocs/js/*.js', 'htdocs/js/dojo/*.js', 'htdocs/js/dojo/dnd/*.js', 'htdocs/js/dojo/date/*.js', 'templates/*.html', 'scripts/*' ]}, entry_points={ 'trac.plugins': ['backlog = backlog.web_ui'] } )
from setuptools import setup PACKAGE = 'TracBacklog' VERSION = '0.1' setup(name=PACKAGE, version=VERSION, packages=['backlog'], package_data={ 'backlog': [ 'htdocs/css/*.css', 'htdocs/img/*.png', 'htdocs/js/*.js', 'htdocs/js/dojo/*.js', 'htdocs/js/dojo/dnd/*.js', 'htdocs/js/dojo/date/*.js', 'templates/*.html', 'scripts/*' ]}, entry_points=''' [trac.plugins] backlog = backlog.web_ui ''') Use a different method of specifying the entry point.from setuptools import setup PACKAGE = 'TracBacklog' VERSION = '0.1' setup(name=PACKAGE, version=VERSION, packages=['backlog'], package_data={ 'backlog': [ 'htdocs/css/*.css', 'htdocs/img/*.png', 'htdocs/js/*.js', 'htdocs/js/dojo/*.js', 'htdocs/js/dojo/dnd/*.js', 'htdocs/js/dojo/date/*.js', 'templates/*.html', 'scripts/*' ]}, entry_points={ 'trac.plugins': ['backlog = backlog.web_ui'] } )
<commit_before>from setuptools import setup PACKAGE = 'TracBacklog' VERSION = '0.1' setup(name=PACKAGE, version=VERSION, packages=['backlog'], package_data={ 'backlog': [ 'htdocs/css/*.css', 'htdocs/img/*.png', 'htdocs/js/*.js', 'htdocs/js/dojo/*.js', 'htdocs/js/dojo/dnd/*.js', 'htdocs/js/dojo/date/*.js', 'templates/*.html', 'scripts/*' ]}, entry_points=''' [trac.plugins] backlog = backlog.web_ui ''') <commit_msg>Use a different method of specifying the entry point.<commit_after>from setuptools import setup PACKAGE = 'TracBacklog' VERSION = '0.1' setup(name=PACKAGE, version=VERSION, packages=['backlog'], package_data={ 'backlog': [ 'htdocs/css/*.css', 'htdocs/img/*.png', 'htdocs/js/*.js', 'htdocs/js/dojo/*.js', 'htdocs/js/dojo/dnd/*.js', 'htdocs/js/dojo/date/*.js', 'templates/*.html', 'scripts/*' ]}, entry_points={ 'trac.plugins': ['backlog = backlog.web_ui'] } )
d1fb5fbef04fde818abd1eb6ac38805058addafb
setup.py
setup.py
import re from setuptools import setup, find_packages versionPattern = re.compile(r"""^__version__ = ['"](.*?)['"]$""", re.M) with open("fusion_index/_version.py", "rt") as f: version = versionPattern.search(f.read()).group(1) setup( name='fusion-index', version=version, description='Lookup/search index service for Fusion', url='https://bitbucket.org/fusionapp/fusion-index', install_requires=[ 'Twisted[tls] >= 15.0.0', 'txspinneret >= 0.1.2', 'Axiom >= 0.7.4', 'eliot >= 0.8.0', 'testtools', ], license='MIT', packages=find_packages() + ['axiom.plugins'], include_package_data=True)
import re from setuptools import setup, find_packages versionPattern = re.compile(r"""^__version__ = ['"](.*?)['"]$""", re.M) with open("fusion_index/_version.py", "rt") as f: version = versionPattern.search(f.read()).group(1) setup( name='fusion-index', version=version, description='Lookup/search index service for Fusion', url='https://bitbucket.org/fusionapp/fusion-index', install_requires=[ 'Twisted[tls] >= 15.0.0', 'txspinneret >= 0.1.2', 'Axiom >= 0.7.4', 'eliot >= 0.8.0', 'testtools', 'characteristic', ], license='MIT', packages=find_packages() + ['axiom.plugins'], include_package_data=True)
Add Characteristic to depends since we are using it directly now.
Add Characteristic to depends since we are using it directly now. --HG-- branch : 1-lookup-index
Python
mit
fusionapp/fusion-index
import re from setuptools import setup, find_packages versionPattern = re.compile(r"""^__version__ = ['"](.*?)['"]$""", re.M) with open("fusion_index/_version.py", "rt") as f: version = versionPattern.search(f.read()).group(1) setup( name='fusion-index', version=version, description='Lookup/search index service for Fusion', url='https://bitbucket.org/fusionapp/fusion-index', install_requires=[ 'Twisted[tls] >= 15.0.0', 'txspinneret >= 0.1.2', 'Axiom >= 0.7.4', 'eliot >= 0.8.0', 'testtools', ], license='MIT', packages=find_packages() + ['axiom.plugins'], include_package_data=True) Add Characteristic to depends since we are using it directly now. --HG-- branch : 1-lookup-index
import re from setuptools import setup, find_packages versionPattern = re.compile(r"""^__version__ = ['"](.*?)['"]$""", re.M) with open("fusion_index/_version.py", "rt") as f: version = versionPattern.search(f.read()).group(1) setup( name='fusion-index', version=version, description='Lookup/search index service for Fusion', url='https://bitbucket.org/fusionapp/fusion-index', install_requires=[ 'Twisted[tls] >= 15.0.0', 'txspinneret >= 0.1.2', 'Axiom >= 0.7.4', 'eliot >= 0.8.0', 'testtools', 'characteristic', ], license='MIT', packages=find_packages() + ['axiom.plugins'], include_package_data=True)
<commit_before>import re from setuptools import setup, find_packages versionPattern = re.compile(r"""^__version__ = ['"](.*?)['"]$""", re.M) with open("fusion_index/_version.py", "rt") as f: version = versionPattern.search(f.read()).group(1) setup( name='fusion-index', version=version, description='Lookup/search index service for Fusion', url='https://bitbucket.org/fusionapp/fusion-index', install_requires=[ 'Twisted[tls] >= 15.0.0', 'txspinneret >= 0.1.2', 'Axiom >= 0.7.4', 'eliot >= 0.8.0', 'testtools', ], license='MIT', packages=find_packages() + ['axiom.plugins'], include_package_data=True) <commit_msg>Add Characteristic to depends since we are using it directly now. --HG-- branch : 1-lookup-index<commit_after>
import re from setuptools import setup, find_packages versionPattern = re.compile(r"""^__version__ = ['"](.*?)['"]$""", re.M) with open("fusion_index/_version.py", "rt") as f: version = versionPattern.search(f.read()).group(1) setup( name='fusion-index', version=version, description='Lookup/search index service for Fusion', url='https://bitbucket.org/fusionapp/fusion-index', install_requires=[ 'Twisted[tls] >= 15.0.0', 'txspinneret >= 0.1.2', 'Axiom >= 0.7.4', 'eliot >= 0.8.0', 'testtools', 'characteristic', ], license='MIT', packages=find_packages() + ['axiom.plugins'], include_package_data=True)
import re from setuptools import setup, find_packages versionPattern = re.compile(r"""^__version__ = ['"](.*?)['"]$""", re.M) with open("fusion_index/_version.py", "rt") as f: version = versionPattern.search(f.read()).group(1) setup( name='fusion-index', version=version, description='Lookup/search index service for Fusion', url='https://bitbucket.org/fusionapp/fusion-index', install_requires=[ 'Twisted[tls] >= 15.0.0', 'txspinneret >= 0.1.2', 'Axiom >= 0.7.4', 'eliot >= 0.8.0', 'testtools', ], license='MIT', packages=find_packages() + ['axiom.plugins'], include_package_data=True) Add Characteristic to depends since we are using it directly now. --HG-- branch : 1-lookup-indeximport re from setuptools import setup, find_packages versionPattern = re.compile(r"""^__version__ = ['"](.*?)['"]$""", re.M) with open("fusion_index/_version.py", "rt") as f: version = versionPattern.search(f.read()).group(1) setup( name='fusion-index', version=version, description='Lookup/search index service for Fusion', url='https://bitbucket.org/fusionapp/fusion-index', install_requires=[ 'Twisted[tls] >= 15.0.0', 'txspinneret >= 0.1.2', 'Axiom >= 0.7.4', 'eliot >= 0.8.0', 'testtools', 'characteristic', ], license='MIT', packages=find_packages() + ['axiom.plugins'], include_package_data=True)
<commit_before>import re from setuptools import setup, find_packages versionPattern = re.compile(r"""^__version__ = ['"](.*?)['"]$""", re.M) with open("fusion_index/_version.py", "rt") as f: version = versionPattern.search(f.read()).group(1) setup( name='fusion-index', version=version, description='Lookup/search index service for Fusion', url='https://bitbucket.org/fusionapp/fusion-index', install_requires=[ 'Twisted[tls] >= 15.0.0', 'txspinneret >= 0.1.2', 'Axiom >= 0.7.4', 'eliot >= 0.8.0', 'testtools', ], license='MIT', packages=find_packages() + ['axiom.plugins'], include_package_data=True) <commit_msg>Add Characteristic to depends since we are using it directly now. --HG-- branch : 1-lookup-index<commit_after>import re from setuptools import setup, find_packages versionPattern = re.compile(r"""^__version__ = ['"](.*?)['"]$""", re.M) with open("fusion_index/_version.py", "rt") as f: version = versionPattern.search(f.read()).group(1) setup( name='fusion-index', version=version, description='Lookup/search index service for Fusion', url='https://bitbucket.org/fusionapp/fusion-index', install_requires=[ 'Twisted[tls] >= 15.0.0', 'txspinneret >= 0.1.2', 'Axiom >= 0.7.4', 'eliot >= 0.8.0', 'testtools', 'characteristic', ], license='MIT', packages=find_packages() + ['axiom.plugins'], include_package_data=True)
11d4763b093d0f1006051e892277d33ca273916c
setup.py
setup.py
from setuptools import setup, find_packages import re with open("README.rst", mode='r') as readme_file: text=readme_file.read() #below version code pulled from requests module with open('__init__.py', 'r') as fd: version_number = re.search(r'^__version__\s*=\s*[\'"]([^\'"]*)[\'"]', fd.read(), re.MULTILINE).group(1) if not version_number: raise RuntimeError('Cannot find version information') setup( name='pyautoupdate', version=version_number, packages=find_packages(), description='Interface to allow python programs to automatically update', long_description=text, url='https://github.com/rlee287/pyautoupdate', install_requires=['requests>=2.6'], package_data={ 'testing':['*.rst']}, license="LGPL 2.1" )
from setuptools import setup, find_packages import re with open("README.rst", mode='r') as readme_file: text=readme_file.read() #below version code pulled from requests module with open('__init__.py', 'r') as fd: version_number = re.search(r'^__version__\s*=\s*[\'"]([^\'"]*)[\'"]', fd.read(), re.MULTILINE).group(1) if not version_number: raise RuntimeError('Cannot find version information') setup( name='pyautoupdate', version=version_number, packages=find_packages(), description='Interface to allow python programs to automatically update', long_description=text, url='https://github.com/rlee287/pyautoupdate', install_requires=['requests'], extras_require={ 'testing': ['pytest','coverage'] }, package_data={ 'testing':['*.rst']}, license="LGPL 2.1" )
Fix continuation whitespace and add extras_require
Fix continuation whitespace and add extras_require
Python
lgpl-2.1
rlee287/pyautoupdate,rlee287/pyautoupdate
from setuptools import setup, find_packages import re with open("README.rst", mode='r') as readme_file: text=readme_file.read() #below version code pulled from requests module with open('__init__.py', 'r') as fd: version_number = re.search(r'^__version__\s*=\s*[\'"]([^\'"]*)[\'"]', fd.read(), re.MULTILINE).group(1) if not version_number: raise RuntimeError('Cannot find version information') setup( name='pyautoupdate', version=version_number, packages=find_packages(), description='Interface to allow python programs to automatically update', long_description=text, url='https://github.com/rlee287/pyautoupdate', install_requires=['requests>=2.6'], package_data={ 'testing':['*.rst']}, license="LGPL 2.1" ) Fix continuation whitespace and add extras_require
from setuptools import setup, find_packages import re with open("README.rst", mode='r') as readme_file: text=readme_file.read() #below version code pulled from requests module with open('__init__.py', 'r') as fd: version_number = re.search(r'^__version__\s*=\s*[\'"]([^\'"]*)[\'"]', fd.read(), re.MULTILINE).group(1) if not version_number: raise RuntimeError('Cannot find version information') setup( name='pyautoupdate', version=version_number, packages=find_packages(), description='Interface to allow python programs to automatically update', long_description=text, url='https://github.com/rlee287/pyautoupdate', install_requires=['requests'], extras_require={ 'testing': ['pytest','coverage'] }, package_data={ 'testing':['*.rst']}, license="LGPL 2.1" )
<commit_before>from setuptools import setup, find_packages import re with open("README.rst", mode='r') as readme_file: text=readme_file.read() #below version code pulled from requests module with open('__init__.py', 'r') as fd: version_number = re.search(r'^__version__\s*=\s*[\'"]([^\'"]*)[\'"]', fd.read(), re.MULTILINE).group(1) if not version_number: raise RuntimeError('Cannot find version information') setup( name='pyautoupdate', version=version_number, packages=find_packages(), description='Interface to allow python programs to automatically update', long_description=text, url='https://github.com/rlee287/pyautoupdate', install_requires=['requests>=2.6'], package_data={ 'testing':['*.rst']}, license="LGPL 2.1" ) <commit_msg>Fix continuation whitespace and add extras_require<commit_after>
from setuptools import setup, find_packages import re with open("README.rst", mode='r') as readme_file: text=readme_file.read() #below version code pulled from requests module with open('__init__.py', 'r') as fd: version_number = re.search(r'^__version__\s*=\s*[\'"]([^\'"]*)[\'"]', fd.read(), re.MULTILINE).group(1) if not version_number: raise RuntimeError('Cannot find version information') setup( name='pyautoupdate', version=version_number, packages=find_packages(), description='Interface to allow python programs to automatically update', long_description=text, url='https://github.com/rlee287/pyautoupdate', install_requires=['requests'], extras_require={ 'testing': ['pytest','coverage'] }, package_data={ 'testing':['*.rst']}, license="LGPL 2.1" )
from setuptools import setup, find_packages import re with open("README.rst", mode='r') as readme_file: text=readme_file.read() #below version code pulled from requests module with open('__init__.py', 'r') as fd: version_number = re.search(r'^__version__\s*=\s*[\'"]([^\'"]*)[\'"]', fd.read(), re.MULTILINE).group(1) if not version_number: raise RuntimeError('Cannot find version information') setup( name='pyautoupdate', version=version_number, packages=find_packages(), description='Interface to allow python programs to automatically update', long_description=text, url='https://github.com/rlee287/pyautoupdate', install_requires=['requests>=2.6'], package_data={ 'testing':['*.rst']}, license="LGPL 2.1" ) Fix continuation whitespace and add extras_requirefrom setuptools import setup, find_packages import re with open("README.rst", mode='r') as readme_file: text=readme_file.read() #below version code pulled from requests module with open('__init__.py', 'r') as fd: version_number = re.search(r'^__version__\s*=\s*[\'"]([^\'"]*)[\'"]', fd.read(), re.MULTILINE).group(1) if not version_number: raise RuntimeError('Cannot find version information') setup( name='pyautoupdate', version=version_number, packages=find_packages(), description='Interface to allow python programs to automatically update', long_description=text, url='https://github.com/rlee287/pyautoupdate', install_requires=['requests'], extras_require={ 'testing': ['pytest','coverage'] }, package_data={ 'testing':['*.rst']}, license="LGPL 2.1" )
<commit_before>from setuptools import setup, find_packages import re with open("README.rst", mode='r') as readme_file: text=readme_file.read() #below version code pulled from requests module with open('__init__.py', 'r') as fd: version_number = re.search(r'^__version__\s*=\s*[\'"]([^\'"]*)[\'"]', fd.read(), re.MULTILINE).group(1) if not version_number: raise RuntimeError('Cannot find version information') setup( name='pyautoupdate', version=version_number, packages=find_packages(), description='Interface to allow python programs to automatically update', long_description=text, url='https://github.com/rlee287/pyautoupdate', install_requires=['requests>=2.6'], package_data={ 'testing':['*.rst']}, license="LGPL 2.1" ) <commit_msg>Fix continuation whitespace and add extras_require<commit_after>from setuptools import setup, find_packages import re with open("README.rst", mode='r') as readme_file: text=readme_file.read() #below version code pulled from requests module with open('__init__.py', 'r') as fd: version_number = re.search(r'^__version__\s*=\s*[\'"]([^\'"]*)[\'"]', fd.read(), re.MULTILINE).group(1) if not version_number: raise RuntimeError('Cannot find version information') setup( name='pyautoupdate', version=version_number, packages=find_packages(), description='Interface to allow python programs to automatically update', long_description=text, url='https://github.com/rlee287/pyautoupdate', install_requires=['requests'], extras_require={ 'testing': ['pytest','coverage'] }, package_data={ 'testing':['*.rst']}, license="LGPL 2.1" )
5958f80d456a43654c5013d38569554940e754f4
tests/dojo_test.py
tests/dojo_test.py
import unittest from src.dojo import Dojo class TestCreateRoom (unittest.TestCase): def setUp(self): self.dojo = Dojo() self.test_office = self.dojo.create_room("office", "test") self.test_living_space = self.dojo.create_room("living_space", "test living space") def test_create_room_successfully(self): initial_room_count = len(self.dojo.all_rooms) blue_office = self.dojo.create_room("office", "Blue") self.assertTrue(blue_office) new_room_count = len(self.dojo.all_rooms) self.assertEqual(new_room_count - initial_room_count, 1) def test_create_rooms_successfully(self): my_class_instance = Dojo() initial_room_count = len(my_class_instance.all_rooms) offices = my_class_instance.create_room("office", "Blue", "Black", "Brown") self.assertTrue(offices) new_room_count = len(my_class_instance.all_rooms) self.assertEqual(new_room_count - initial_room_count, 3)
import unittest from src.dojo import Dojo class TestCreateRoom (unittest.TestCase): def setUp(self): self.dojo = Dojo() self.test_office = self.dojo.create_room("office", "test") self.test_living_space = self.dojo.create_room("living_space", "test living space") def test_create_room_successfully(self): initial_room_count = len(self.dojo.all_rooms) blue_office = self.dojo.create_room("office", "Blue") self.assertTrue(blue_office) new_room_count = len(self.dojo.all_rooms) self.assertEqual(new_room_count - initial_room_count, 1) def test_create_rooms_successfully(self): initial_room_count = len(self.dojo.all_rooms) offices = self.dojo.create_room("office", "Blue", "Black", "Brown") self.assertTrue(offices) new_room_count = len(self.dojo.all_rooms) self.assertEqual(new_room_count - initial_room_count, 3)
Change test for multiple rooms to use test set-up
Change test for multiple rooms to use test set-up
Python
mit
EdwinKato/Space-Allocator,EdwinKato/Space-Allocator
import unittest from src.dojo import Dojo class TestCreateRoom (unittest.TestCase): def setUp(self): self.dojo = Dojo() self.test_office = self.dojo.create_room("office", "test") self.test_living_space = self.dojo.create_room("living_space", "test living space") def test_create_room_successfully(self): initial_room_count = len(self.dojo.all_rooms) blue_office = self.dojo.create_room("office", "Blue") self.assertTrue(blue_office) new_room_count = len(self.dojo.all_rooms) self.assertEqual(new_room_count - initial_room_count, 1) def test_create_rooms_successfully(self): my_class_instance = Dojo() initial_room_count = len(my_class_instance.all_rooms) offices = my_class_instance.create_room("office", "Blue", "Black", "Brown") self.assertTrue(offices) new_room_count = len(my_class_instance.all_rooms) self.assertEqual(new_room_count - initial_room_count, 3)Change test for multiple rooms to use test set-up
import unittest from src.dojo import Dojo class TestCreateRoom (unittest.TestCase): def setUp(self): self.dojo = Dojo() self.test_office = self.dojo.create_room("office", "test") self.test_living_space = self.dojo.create_room("living_space", "test living space") def test_create_room_successfully(self): initial_room_count = len(self.dojo.all_rooms) blue_office = self.dojo.create_room("office", "Blue") self.assertTrue(blue_office) new_room_count = len(self.dojo.all_rooms) self.assertEqual(new_room_count - initial_room_count, 1) def test_create_rooms_successfully(self): initial_room_count = len(self.dojo.all_rooms) offices = self.dojo.create_room("office", "Blue", "Black", "Brown") self.assertTrue(offices) new_room_count = len(self.dojo.all_rooms) self.assertEqual(new_room_count - initial_room_count, 3)
<commit_before>import unittest from src.dojo import Dojo class TestCreateRoom (unittest.TestCase): def setUp(self): self.dojo = Dojo() self.test_office = self.dojo.create_room("office", "test") self.test_living_space = self.dojo.create_room("living_space", "test living space") def test_create_room_successfully(self): initial_room_count = len(self.dojo.all_rooms) blue_office = self.dojo.create_room("office", "Blue") self.assertTrue(blue_office) new_room_count = len(self.dojo.all_rooms) self.assertEqual(new_room_count - initial_room_count, 1) def test_create_rooms_successfully(self): my_class_instance = Dojo() initial_room_count = len(my_class_instance.all_rooms) offices = my_class_instance.create_room("office", "Blue", "Black", "Brown") self.assertTrue(offices) new_room_count = len(my_class_instance.all_rooms) self.assertEqual(new_room_count - initial_room_count, 3)<commit_msg>Change test for multiple rooms to use test set-up<commit_after>
import unittest from src.dojo import Dojo class TestCreateRoom (unittest.TestCase): def setUp(self): self.dojo = Dojo() self.test_office = self.dojo.create_room("office", "test") self.test_living_space = self.dojo.create_room("living_space", "test living space") def test_create_room_successfully(self): initial_room_count = len(self.dojo.all_rooms) blue_office = self.dojo.create_room("office", "Blue") self.assertTrue(blue_office) new_room_count = len(self.dojo.all_rooms) self.assertEqual(new_room_count - initial_room_count, 1) def test_create_rooms_successfully(self): initial_room_count = len(self.dojo.all_rooms) offices = self.dojo.create_room("office", "Blue", "Black", "Brown") self.assertTrue(offices) new_room_count = len(self.dojo.all_rooms) self.assertEqual(new_room_count - initial_room_count, 3)
import unittest from src.dojo import Dojo class TestCreateRoom (unittest.TestCase): def setUp(self): self.dojo = Dojo() self.test_office = self.dojo.create_room("office", "test") self.test_living_space = self.dojo.create_room("living_space", "test living space") def test_create_room_successfully(self): initial_room_count = len(self.dojo.all_rooms) blue_office = self.dojo.create_room("office", "Blue") self.assertTrue(blue_office) new_room_count = len(self.dojo.all_rooms) self.assertEqual(new_room_count - initial_room_count, 1) def test_create_rooms_successfully(self): my_class_instance = Dojo() initial_room_count = len(my_class_instance.all_rooms) offices = my_class_instance.create_room("office", "Blue", "Black", "Brown") self.assertTrue(offices) new_room_count = len(my_class_instance.all_rooms) self.assertEqual(new_room_count - initial_room_count, 3)Change test for multiple rooms to use test set-upimport unittest from src.dojo import Dojo class TestCreateRoom (unittest.TestCase): def setUp(self): self.dojo = Dojo() self.test_office = self.dojo.create_room("office", "test") self.test_living_space = self.dojo.create_room("living_space", "test living space") def test_create_room_successfully(self): initial_room_count = len(self.dojo.all_rooms) blue_office = self.dojo.create_room("office", "Blue") self.assertTrue(blue_office) new_room_count = len(self.dojo.all_rooms) self.assertEqual(new_room_count - initial_room_count, 1) def test_create_rooms_successfully(self): initial_room_count = len(self.dojo.all_rooms) offices = self.dojo.create_room("office", "Blue", "Black", "Brown") self.assertTrue(offices) new_room_count = len(self.dojo.all_rooms) self.assertEqual(new_room_count - initial_room_count, 3)
<commit_before>import unittest from src.dojo import Dojo class TestCreateRoom (unittest.TestCase): def setUp(self): self.dojo = Dojo() self.test_office = self.dojo.create_room("office", "test") self.test_living_space = self.dojo.create_room("living_space", "test living space") def test_create_room_successfully(self): initial_room_count = len(self.dojo.all_rooms) blue_office = self.dojo.create_room("office", "Blue") self.assertTrue(blue_office) new_room_count = len(self.dojo.all_rooms) self.assertEqual(new_room_count - initial_room_count, 1) def test_create_rooms_successfully(self): my_class_instance = Dojo() initial_room_count = len(my_class_instance.all_rooms) offices = my_class_instance.create_room("office", "Blue", "Black", "Brown") self.assertTrue(offices) new_room_count = len(my_class_instance.all_rooms) self.assertEqual(new_room_count - initial_room_count, 3)<commit_msg>Change test for multiple rooms to use test set-up<commit_after>import unittest from src.dojo import Dojo class TestCreateRoom (unittest.TestCase): def setUp(self): self.dojo = Dojo() self.test_office = self.dojo.create_room("office", "test") self.test_living_space = self.dojo.create_room("living_space", "test living space") def test_create_room_successfully(self): initial_room_count = len(self.dojo.all_rooms) blue_office = self.dojo.create_room("office", "Blue") self.assertTrue(blue_office) new_room_count = len(self.dojo.all_rooms) self.assertEqual(new_room_count - initial_room_count, 1) def test_create_rooms_successfully(self): initial_room_count = len(self.dojo.all_rooms) offices = self.dojo.create_room("office", "Blue", "Black", "Brown") self.assertTrue(offices) new_room_count = len(self.dojo.all_rooms) self.assertEqual(new_room_count - initial_room_count, 3)
2f75cf84d37869e49dd3820d00a2fa69ab57a219
setup.py
setup.py
#!/usr/bin/env python # encoding=utf-8 from __future__ import print_function import os import sys from setuptools import setup here = os.path.abspath(os.path.dirname(__file__)) README = open(os.path.join(here, 'README.md')).read() needs_pytest = set(['pytest', 'test', 'ptr']).intersection(sys.argv) pytest_runner = ['pytest-runner'] if needs_pytest else [] setup(name='mwclient', version='0.9.3', # Use bumpversion to update description='MediaWiki API client', long_description=README, long_description_content_type='text/markdown', classifiers=[ 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', ], keywords='mediawiki wikipedia', author='Bryan Tong Minh', author_email='bryan.tongminh@gmail.com', url='https://github.com/btongminh/mwclient', license='MIT', packages=['mwclient'], install_requires=['requests_oauthlib', 'six'], setup_requires=pytest_runner, tests_require=['pytest', 'pytest-cov', 'flake8', 'responses>=0.3.0', 'responses!=0.6.0', 'mock'], zip_safe=True )
#!/usr/bin/env python # encoding=utf-8 from __future__ import print_function import os import sys from setuptools import setup here = os.path.abspath(os.path.dirname(__file__)) README = open(os.path.join(here, 'README.md')).read() needs_pytest = set(['pytest', 'test', 'ptr']).intersection(sys.argv) pytest_runner = ['pytest-runner'] if needs_pytest else [] setup(name='mwclient', version='0.9.3', # Use bumpversion to update description='MediaWiki API client', long_description=README, long_description_content_type='text/markdown', classifiers=[ 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', ], keywords='mediawiki wikipedia', author='Bryan Tong Minh', author_email='bryan.tongminh@gmail.com', url='https://github.com/btongminh/mwclient', license='MIT', packages=['mwclient'], install_requires=['requests-oauthlib', 'six'], setup_requires=pytest_runner, tests_require=['pytest', 'pytest-cov', 'flake8', 'responses>=0.3.0', 'responses!=0.6.0', 'mock'], zip_safe=True )
Update dependency to work on conda
Update dependency to work on conda requests_oauthlib dependency name will only install from pip, requests-oauthlib works on both conda and pip
Python
mit
mwclient/mwclient
#!/usr/bin/env python # encoding=utf-8 from __future__ import print_function import os import sys from setuptools import setup here = os.path.abspath(os.path.dirname(__file__)) README = open(os.path.join(here, 'README.md')).read() needs_pytest = set(['pytest', 'test', 'ptr']).intersection(sys.argv) pytest_runner = ['pytest-runner'] if needs_pytest else [] setup(name='mwclient', version='0.9.3', # Use bumpversion to update description='MediaWiki API client', long_description=README, long_description_content_type='text/markdown', classifiers=[ 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', ], keywords='mediawiki wikipedia', author='Bryan Tong Minh', author_email='bryan.tongminh@gmail.com', url='https://github.com/btongminh/mwclient', license='MIT', packages=['mwclient'], install_requires=['requests_oauthlib', 'six'], setup_requires=pytest_runner, tests_require=['pytest', 'pytest-cov', 'flake8', 'responses>=0.3.0', 'responses!=0.6.0', 'mock'], zip_safe=True ) Update dependency to work on conda requests_oauthlib dependency name will only install from pip, requests-oauthlib works on both conda and pip
#!/usr/bin/env python # encoding=utf-8 from __future__ import print_function import os import sys from setuptools import setup here = os.path.abspath(os.path.dirname(__file__)) README = open(os.path.join(here, 'README.md')).read() needs_pytest = set(['pytest', 'test', 'ptr']).intersection(sys.argv) pytest_runner = ['pytest-runner'] if needs_pytest else [] setup(name='mwclient', version='0.9.3', # Use bumpversion to update description='MediaWiki API client', long_description=README, long_description_content_type='text/markdown', classifiers=[ 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', ], keywords='mediawiki wikipedia', author='Bryan Tong Minh', author_email='bryan.tongminh@gmail.com', url='https://github.com/btongminh/mwclient', license='MIT', packages=['mwclient'], install_requires=['requests-oauthlib', 'six'], setup_requires=pytest_runner, tests_require=['pytest', 'pytest-cov', 'flake8', 'responses>=0.3.0', 'responses!=0.6.0', 'mock'], zip_safe=True )
<commit_before>#!/usr/bin/env python # encoding=utf-8 from __future__ import print_function import os import sys from setuptools import setup here = os.path.abspath(os.path.dirname(__file__)) README = open(os.path.join(here, 'README.md')).read() needs_pytest = set(['pytest', 'test', 'ptr']).intersection(sys.argv) pytest_runner = ['pytest-runner'] if needs_pytest else [] setup(name='mwclient', version='0.9.3', # Use bumpversion to update description='MediaWiki API client', long_description=README, long_description_content_type='text/markdown', classifiers=[ 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', ], keywords='mediawiki wikipedia', author='Bryan Tong Minh', author_email='bryan.tongminh@gmail.com', url='https://github.com/btongminh/mwclient', license='MIT', packages=['mwclient'], install_requires=['requests_oauthlib', 'six'], setup_requires=pytest_runner, tests_require=['pytest', 'pytest-cov', 'flake8', 'responses>=0.3.0', 'responses!=0.6.0', 'mock'], zip_safe=True ) <commit_msg>Update dependency to work on conda requests_oauthlib dependency name will only install from pip, requests-oauthlib works on both conda and pip<commit_after>
#!/usr/bin/env python # encoding=utf-8 from __future__ import print_function import os import sys from setuptools import setup here = os.path.abspath(os.path.dirname(__file__)) README = open(os.path.join(here, 'README.md')).read() needs_pytest = set(['pytest', 'test', 'ptr']).intersection(sys.argv) pytest_runner = ['pytest-runner'] if needs_pytest else [] setup(name='mwclient', version='0.9.3', # Use bumpversion to update description='MediaWiki API client', long_description=README, long_description_content_type='text/markdown', classifiers=[ 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', ], keywords='mediawiki wikipedia', author='Bryan Tong Minh', author_email='bryan.tongminh@gmail.com', url='https://github.com/btongminh/mwclient', license='MIT', packages=['mwclient'], install_requires=['requests-oauthlib', 'six'], setup_requires=pytest_runner, tests_require=['pytest', 'pytest-cov', 'flake8', 'responses>=0.3.0', 'responses!=0.6.0', 'mock'], zip_safe=True )
#!/usr/bin/env python # encoding=utf-8 from __future__ import print_function import os import sys from setuptools import setup here = os.path.abspath(os.path.dirname(__file__)) README = open(os.path.join(here, 'README.md')).read() needs_pytest = set(['pytest', 'test', 'ptr']).intersection(sys.argv) pytest_runner = ['pytest-runner'] if needs_pytest else [] setup(name='mwclient', version='0.9.3', # Use bumpversion to update description='MediaWiki API client', long_description=README, long_description_content_type='text/markdown', classifiers=[ 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', ], keywords='mediawiki wikipedia', author='Bryan Tong Minh', author_email='bryan.tongminh@gmail.com', url='https://github.com/btongminh/mwclient', license='MIT', packages=['mwclient'], install_requires=['requests_oauthlib', 'six'], setup_requires=pytest_runner, tests_require=['pytest', 'pytest-cov', 'flake8', 'responses>=0.3.0', 'responses!=0.6.0', 'mock'], zip_safe=True ) Update dependency to work on conda requests_oauthlib dependency name will only install from pip, requests-oauthlib works on both conda and pip#!/usr/bin/env python # encoding=utf-8 from __future__ import print_function import os import sys from setuptools import setup here = os.path.abspath(os.path.dirname(__file__)) README = open(os.path.join(here, 'README.md')).read() needs_pytest = set(['pytest', 'test', 'ptr']).intersection(sys.argv) pytest_runner = ['pytest-runner'] if needs_pytest else [] setup(name='mwclient', version='0.9.3', # Use bumpversion to update description='MediaWiki API client', long_description=README, long_description_content_type='text/markdown', classifiers=[ 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', ], keywords='mediawiki wikipedia', author='Bryan Tong Minh', author_email='bryan.tongminh@gmail.com', url='https://github.com/btongminh/mwclient', license='MIT', packages=['mwclient'], install_requires=['requests-oauthlib', 'six'], setup_requires=pytest_runner, tests_require=['pytest', 'pytest-cov', 'flake8', 'responses>=0.3.0', 'responses!=0.6.0', 'mock'], zip_safe=True )
<commit_before>#!/usr/bin/env python # encoding=utf-8 from __future__ import print_function import os import sys from setuptools import setup here = os.path.abspath(os.path.dirname(__file__)) README = open(os.path.join(here, 'README.md')).read() needs_pytest = set(['pytest', 'test', 'ptr']).intersection(sys.argv) pytest_runner = ['pytest-runner'] if needs_pytest else [] setup(name='mwclient', version='0.9.3', # Use bumpversion to update description='MediaWiki API client', long_description=README, long_description_content_type='text/markdown', classifiers=[ 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', ], keywords='mediawiki wikipedia', author='Bryan Tong Minh', author_email='bryan.tongminh@gmail.com', url='https://github.com/btongminh/mwclient', license='MIT', packages=['mwclient'], install_requires=['requests_oauthlib', 'six'], setup_requires=pytest_runner, tests_require=['pytest', 'pytest-cov', 'flake8', 'responses>=0.3.0', 'responses!=0.6.0', 'mock'], zip_safe=True ) <commit_msg>Update dependency to work on conda requests_oauthlib dependency name will only install from pip, requests-oauthlib works on both conda and pip<commit_after>#!/usr/bin/env python # encoding=utf-8 from __future__ import print_function import os import sys from setuptools import setup here = os.path.abspath(os.path.dirname(__file__)) README = open(os.path.join(here, 'README.md')).read() needs_pytest = set(['pytest', 'test', 'ptr']).intersection(sys.argv) pytest_runner = ['pytest-runner'] if needs_pytest else [] setup(name='mwclient', version='0.9.3', # Use bumpversion to update description='MediaWiki API client', long_description=README, long_description_content_type='text/markdown', classifiers=[ 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.5', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', ], keywords='mediawiki wikipedia', author='Bryan Tong Minh', author_email='bryan.tongminh@gmail.com', url='https://github.com/btongminh/mwclient', license='MIT', packages=['mwclient'], install_requires=['requests-oauthlib', 'six'], setup_requires=pytest_runner, tests_require=['pytest', 'pytest-cov', 'flake8', 'responses>=0.3.0', 'responses!=0.6.0', 'mock'], zip_safe=True )
2ee7bba787a29f715805b038fb98eda767534f10
setup.py
setup.py
#!/usr/bin/env python #-*- coding:utf-8 -*- from setuptools import setup setup( name='pybreaker', version='0.6.0', description='Python implementation of the Circuit Breaker pattern', long_description=open('README.rst', 'r').read(), keywords=['design', 'pattern', 'circuit', 'breaker', 'integration'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries', ], platforms=[ 'Any', ], license='BSD', author='Daniel Fernandes Martins', author_email='daniel.tritone@gmail.com', url='http://github.com/danielfm/pybreaker', package_dir={'': 'src'}, py_modules=['pybreaker'], install_requires=[ 'six', ], include_package_data=True, zip_safe=False, test_suite='tests', tests_require=['mock', 'fakeredis==0.16.0', 'redis==2.10.6', 'tornado'], )
#!/usr/bin/env python #-*- coding:utf-8 -*- from setuptools import setup setup( name='pybreaker', version='0.6.0', description='Python implementation of the Circuit Breaker pattern', long_description=open('README.rst', 'r').read(), keywords=['design', 'pattern', 'circuit', 'breaker', 'integration'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries', ], platforms=[ 'Any', ], license='BSD', author='Daniel Fernandes Martins', author_email='daniel.tritone@gmail.com', url='http://github.com/danielfm/pybreaker', package_dir={'': 'src'}, py_modules=['pybreaker'], install_requires=[ 'six', ], include_package_data=True, zip_safe=False, test_suite='tests', tests_require=['mock', 'fakeredis==0.16.0', 'redis==2.10.6', 'tornado'], )
Update classifiers to drop support for Python 2.7
Update classifiers to drop support for Python 2.7
Python
bsd-3-clause
danielfm/pybreaker
#!/usr/bin/env python #-*- coding:utf-8 -*- from setuptools import setup setup( name='pybreaker', version='0.6.0', description='Python implementation of the Circuit Breaker pattern', long_description=open('README.rst', 'r').read(), keywords=['design', 'pattern', 'circuit', 'breaker', 'integration'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries', ], platforms=[ 'Any', ], license='BSD', author='Daniel Fernandes Martins', author_email='daniel.tritone@gmail.com', url='http://github.com/danielfm/pybreaker', package_dir={'': 'src'}, py_modules=['pybreaker'], install_requires=[ 'six', ], include_package_data=True, zip_safe=False, test_suite='tests', tests_require=['mock', 'fakeredis==0.16.0', 'redis==2.10.6', 'tornado'], ) Update classifiers to drop support for Python 2.7
#!/usr/bin/env python #-*- coding:utf-8 -*- from setuptools import setup setup( name='pybreaker', version='0.6.0', description='Python implementation of the Circuit Breaker pattern', long_description=open('README.rst', 'r').read(), keywords=['design', 'pattern', 'circuit', 'breaker', 'integration'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries', ], platforms=[ 'Any', ], license='BSD', author='Daniel Fernandes Martins', author_email='daniel.tritone@gmail.com', url='http://github.com/danielfm/pybreaker', package_dir={'': 'src'}, py_modules=['pybreaker'], install_requires=[ 'six', ], include_package_data=True, zip_safe=False, test_suite='tests', tests_require=['mock', 'fakeredis==0.16.0', 'redis==2.10.6', 'tornado'], )
<commit_before>#!/usr/bin/env python #-*- coding:utf-8 -*- from setuptools import setup setup( name='pybreaker', version='0.6.0', description='Python implementation of the Circuit Breaker pattern', long_description=open('README.rst', 'r').read(), keywords=['design', 'pattern', 'circuit', 'breaker', 'integration'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries', ], platforms=[ 'Any', ], license='BSD', author='Daniel Fernandes Martins', author_email='daniel.tritone@gmail.com', url='http://github.com/danielfm/pybreaker', package_dir={'': 'src'}, py_modules=['pybreaker'], install_requires=[ 'six', ], include_package_data=True, zip_safe=False, test_suite='tests', tests_require=['mock', 'fakeredis==0.16.0', 'redis==2.10.6', 'tornado'], ) <commit_msg>Update classifiers to drop support for Python 2.7<commit_after>
#!/usr/bin/env python #-*- coding:utf-8 -*- from setuptools import setup setup( name='pybreaker', version='0.6.0', description='Python implementation of the Circuit Breaker pattern', long_description=open('README.rst', 'r').read(), keywords=['design', 'pattern', 'circuit', 'breaker', 'integration'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries', ], platforms=[ 'Any', ], license='BSD', author='Daniel Fernandes Martins', author_email='daniel.tritone@gmail.com', url='http://github.com/danielfm/pybreaker', package_dir={'': 'src'}, py_modules=['pybreaker'], install_requires=[ 'six', ], include_package_data=True, zip_safe=False, test_suite='tests', tests_require=['mock', 'fakeredis==0.16.0', 'redis==2.10.6', 'tornado'], )
#!/usr/bin/env python #-*- coding:utf-8 -*- from setuptools import setup setup( name='pybreaker', version='0.6.0', description='Python implementation of the Circuit Breaker pattern', long_description=open('README.rst', 'r').read(), keywords=['design', 'pattern', 'circuit', 'breaker', 'integration'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries', ], platforms=[ 'Any', ], license='BSD', author='Daniel Fernandes Martins', author_email='daniel.tritone@gmail.com', url='http://github.com/danielfm/pybreaker', package_dir={'': 'src'}, py_modules=['pybreaker'], install_requires=[ 'six', ], include_package_data=True, zip_safe=False, test_suite='tests', tests_require=['mock', 'fakeredis==0.16.0', 'redis==2.10.6', 'tornado'], ) Update classifiers to drop support for Python 2.7#!/usr/bin/env python #-*- coding:utf-8 -*- from setuptools import setup setup( name='pybreaker', version='0.6.0', description='Python implementation of the Circuit Breaker pattern', long_description=open('README.rst', 'r').read(), keywords=['design', 'pattern', 'circuit', 'breaker', 'integration'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries', ], platforms=[ 'Any', ], license='BSD', author='Daniel Fernandes Martins', author_email='daniel.tritone@gmail.com', url='http://github.com/danielfm/pybreaker', package_dir={'': 'src'}, py_modules=['pybreaker'], install_requires=[ 'six', ], include_package_data=True, zip_safe=False, test_suite='tests', tests_require=['mock', 'fakeredis==0.16.0', 'redis==2.10.6', 'tornado'], )
<commit_before>#!/usr/bin/env python #-*- coding:utf-8 -*- from setuptools import setup setup( name='pybreaker', version='0.6.0', description='Python implementation of the Circuit Breaker pattern', long_description=open('README.rst', 'r').read(), keywords=['design', 'pattern', 'circuit', 'breaker', 'integration'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries', ], platforms=[ 'Any', ], license='BSD', author='Daniel Fernandes Martins', author_email='daniel.tritone@gmail.com', url='http://github.com/danielfm/pybreaker', package_dir={'': 'src'}, py_modules=['pybreaker'], install_requires=[ 'six', ], include_package_data=True, zip_safe=False, test_suite='tests', tests_require=['mock', 'fakeredis==0.16.0', 'redis==2.10.6', 'tornado'], ) <commit_msg>Update classifiers to drop support for Python 2.7<commit_after>#!/usr/bin/env python #-*- coding:utf-8 -*- from setuptools import setup setup( name='pybreaker', version='0.6.0', description='Python implementation of the Circuit Breaker pattern', long_description=open('README.rst', 'r').read(), keywords=['design', 'pattern', 'circuit', 'breaker', 'integration'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries', ], platforms=[ 'Any', ], license='BSD', author='Daniel Fernandes Martins', author_email='daniel.tritone@gmail.com', url='http://github.com/danielfm/pybreaker', package_dir={'': 'src'}, py_modules=['pybreaker'], install_requires=[ 'six', ], include_package_data=True, zip_safe=False, test_suite='tests', tests_require=['mock', 'fakeredis==0.16.0', 'redis==2.10.6', 'tornado'], )
97f5bbf5c99a40b8204c1e818e3bc8d897f822f0
setup.py
setup.py
import sys from setuptools import setup from betfairlightweight.__init__ import __version__ INSTALL_REQUIRES = [ 'requests', ] TEST_REQUIRES = [ 'mock' ] if sys.version_info < (3,4): INSTALL_REQUIRES.extend([ 'enum34', ]) setup( name='betfairlightweight', version=__version__, packages=['betfairlightweight', 'betfairlightweight.endpoints', 'betfairlightweight.resources', 'betfairlightweight.streaming'], package_dir={'betfairlightweight': 'betfairlightweight'}, install_requires=INSTALL_REQUIRES, requires=['requests'], url='https://github.com/liampauling/betfairlightweight', license='MIT', author='liampauling', author_email='', description='Lightweight python wrapper for Betfair API-NG', classifiers=[ 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], test_suite='tests' )
import sys import re from setuptools import setup INSTALL_REQUIRES = [ 'requests', ] TEST_REQUIRES = [ 'mock' ] if sys.version_info < (3,4): INSTALL_REQUIRES.extend([ 'enum34', ]) with open('betfairlightweight/__init__.py', 'r') as f: version = re.search(r'^__version__\s*=\s*[\'"]([^\'"]*)[\'"]', f.read(), re.MULTILINE).group(1) setup( name='betfairlightweight', version=version, packages=['betfairlightweight', 'betfairlightweight.endpoints', 'betfairlightweight.resources', 'betfairlightweight.streaming'], package_dir={'betfairlightweight': 'betfairlightweight'}, install_requires=INSTALL_REQUIRES, requires=['requests'], url='https://github.com/liampauling/betfairlightweight', license='MIT', author='liampauling', author_email='', description='Lightweight python wrapper for Betfair API-NG', classifiers=[ 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], test_suite='tests' )
Read __version__ with regex to avoid ImportError
Read __version__ with regex to avoid ImportError
Python
mit
liampauling/betfairlightweight,liampauling/betfair
import sys from setuptools import setup from betfairlightweight.__init__ import __version__ INSTALL_REQUIRES = [ 'requests', ] TEST_REQUIRES = [ 'mock' ] if sys.version_info < (3,4): INSTALL_REQUIRES.extend([ 'enum34', ]) setup( name='betfairlightweight', version=__version__, packages=['betfairlightweight', 'betfairlightweight.endpoints', 'betfairlightweight.resources', 'betfairlightweight.streaming'], package_dir={'betfairlightweight': 'betfairlightweight'}, install_requires=INSTALL_REQUIRES, requires=['requests'], url='https://github.com/liampauling/betfairlightweight', license='MIT', author='liampauling', author_email='', description='Lightweight python wrapper for Betfair API-NG', classifiers=[ 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], test_suite='tests' ) Read __version__ with regex to avoid ImportError
import sys import re from setuptools import setup INSTALL_REQUIRES = [ 'requests', ] TEST_REQUIRES = [ 'mock' ] if sys.version_info < (3,4): INSTALL_REQUIRES.extend([ 'enum34', ]) with open('betfairlightweight/__init__.py', 'r') as f: version = re.search(r'^__version__\s*=\s*[\'"]([^\'"]*)[\'"]', f.read(), re.MULTILINE).group(1) setup( name='betfairlightweight', version=version, packages=['betfairlightweight', 'betfairlightweight.endpoints', 'betfairlightweight.resources', 'betfairlightweight.streaming'], package_dir={'betfairlightweight': 'betfairlightweight'}, install_requires=INSTALL_REQUIRES, requires=['requests'], url='https://github.com/liampauling/betfairlightweight', license='MIT', author='liampauling', author_email='', description='Lightweight python wrapper for Betfair API-NG', classifiers=[ 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], test_suite='tests' )
<commit_before>import sys from setuptools import setup from betfairlightweight.__init__ import __version__ INSTALL_REQUIRES = [ 'requests', ] TEST_REQUIRES = [ 'mock' ] if sys.version_info < (3,4): INSTALL_REQUIRES.extend([ 'enum34', ]) setup( name='betfairlightweight', version=__version__, packages=['betfairlightweight', 'betfairlightweight.endpoints', 'betfairlightweight.resources', 'betfairlightweight.streaming'], package_dir={'betfairlightweight': 'betfairlightweight'}, install_requires=INSTALL_REQUIRES, requires=['requests'], url='https://github.com/liampauling/betfairlightweight', license='MIT', author='liampauling', author_email='', description='Lightweight python wrapper for Betfair API-NG', classifiers=[ 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], test_suite='tests' ) <commit_msg>Read __version__ with regex to avoid ImportError<commit_after>
import sys import re from setuptools import setup INSTALL_REQUIRES = [ 'requests', ] TEST_REQUIRES = [ 'mock' ] if sys.version_info < (3,4): INSTALL_REQUIRES.extend([ 'enum34', ]) with open('betfairlightweight/__init__.py', 'r') as f: version = re.search(r'^__version__\s*=\s*[\'"]([^\'"]*)[\'"]', f.read(), re.MULTILINE).group(1) setup( name='betfairlightweight', version=version, packages=['betfairlightweight', 'betfairlightweight.endpoints', 'betfairlightweight.resources', 'betfairlightweight.streaming'], package_dir={'betfairlightweight': 'betfairlightweight'}, install_requires=INSTALL_REQUIRES, requires=['requests'], url='https://github.com/liampauling/betfairlightweight', license='MIT', author='liampauling', author_email='', description='Lightweight python wrapper for Betfair API-NG', classifiers=[ 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], test_suite='tests' )
import sys from setuptools import setup from betfairlightweight.__init__ import __version__ INSTALL_REQUIRES = [ 'requests', ] TEST_REQUIRES = [ 'mock' ] if sys.version_info < (3,4): INSTALL_REQUIRES.extend([ 'enum34', ]) setup( name='betfairlightweight', version=__version__, packages=['betfairlightweight', 'betfairlightweight.endpoints', 'betfairlightweight.resources', 'betfairlightweight.streaming'], package_dir={'betfairlightweight': 'betfairlightweight'}, install_requires=INSTALL_REQUIRES, requires=['requests'], url='https://github.com/liampauling/betfairlightweight', license='MIT', author='liampauling', author_email='', description='Lightweight python wrapper for Betfair API-NG', classifiers=[ 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], test_suite='tests' ) Read __version__ with regex to avoid ImportErrorimport sys import re from setuptools import setup INSTALL_REQUIRES = [ 'requests', ] TEST_REQUIRES = [ 'mock' ] if sys.version_info < (3,4): INSTALL_REQUIRES.extend([ 'enum34', ]) with open('betfairlightweight/__init__.py', 'r') as f: version = re.search(r'^__version__\s*=\s*[\'"]([^\'"]*)[\'"]', f.read(), re.MULTILINE).group(1) setup( name='betfairlightweight', version=version, packages=['betfairlightweight', 'betfairlightweight.endpoints', 'betfairlightweight.resources', 'betfairlightweight.streaming'], package_dir={'betfairlightweight': 'betfairlightweight'}, install_requires=INSTALL_REQUIRES, requires=['requests'], url='https://github.com/liampauling/betfairlightweight', license='MIT', author='liampauling', author_email='', description='Lightweight python wrapper for Betfair API-NG', classifiers=[ 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], test_suite='tests' )
<commit_before>import sys from setuptools import setup from betfairlightweight.__init__ import __version__ INSTALL_REQUIRES = [ 'requests', ] TEST_REQUIRES = [ 'mock' ] if sys.version_info < (3,4): INSTALL_REQUIRES.extend([ 'enum34', ]) setup( name='betfairlightweight', version=__version__, packages=['betfairlightweight', 'betfairlightweight.endpoints', 'betfairlightweight.resources', 'betfairlightweight.streaming'], package_dir={'betfairlightweight': 'betfairlightweight'}, install_requires=INSTALL_REQUIRES, requires=['requests'], url='https://github.com/liampauling/betfairlightweight', license='MIT', author='liampauling', author_email='', description='Lightweight python wrapper for Betfair API-NG', classifiers=[ 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], test_suite='tests' ) <commit_msg>Read __version__ with regex to avoid ImportError<commit_after>import sys import re from setuptools import setup INSTALL_REQUIRES = [ 'requests', ] TEST_REQUIRES = [ 'mock' ] if sys.version_info < (3,4): INSTALL_REQUIRES.extend([ 'enum34', ]) with open('betfairlightweight/__init__.py', 'r') as f: version = re.search(r'^__version__\s*=\s*[\'"]([^\'"]*)[\'"]', f.read(), re.MULTILINE).group(1) setup( name='betfairlightweight', version=version, packages=['betfairlightweight', 'betfairlightweight.endpoints', 'betfairlightweight.resources', 'betfairlightweight.streaming'], package_dir={'betfairlightweight': 'betfairlightweight'}, install_requires=INSTALL_REQUIRES, requires=['requests'], url='https://github.com/liampauling/betfairlightweight', license='MIT', author='liampauling', author_email='', description='Lightweight python wrapper for Betfair API-NG', classifiers=[ 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.4', 'Programming Language :: Python :: 3.5', ], test_suite='tests' )
67a12d44699e4bb8e3b6895ab10c9bb2477ed7fc
tests/fd_io.py
tests/fd_io.py
from filedes.test.base import BaseFDTestCase from filedes import FD import os import errno class TestFDIO(BaseFDTestCase): def testReadWrite(self): r, w = os.pipe() self.assertEquals(FD(w).write("OK"), 2) self.assertEquals(FD(r).read(2), "OK") FD(r).close() FD(w).close() def testWriteAfterClose(self): r, w = os.pipe() fw = FD(w) fw.close() with self.assertRaises(OSError): fw.write("oops") FD(r).close() def testNonblocking(self): r, w = os.pipe() fr = FD(r) try: fr.set_nonblocking() with self.assertRaises(OSError) as ar: fr.read(1) self.assertEquals(ar.exception.errno, errno.EAGAIN) finally: fr.close() FD(w).close()
from filedes.test.base import BaseFDTestCase from filedes import FD import os import errno class TestFDIO(BaseFDTestCase): def testReadWrite(self): r, w = os.pipe() self.assertEquals(FD(w).write("OK"), 2) self.assertEquals(FD(r).read(2), "OK") FD(r).close() FD(w).close() def testWriteAfterClose(self): r, w = os.pipe() fw = FD(w) fw.close() try: with self.assertRaises(OSError) as ar: fw.write("oops") self.assertEquals(ar.exception.errno, errno.EBADF) finally: FD(r).close() def testNonblocking(self): r, w = os.pipe() fr = FD(r) try: fr.set_nonblocking() with self.assertRaises(OSError) as ar: fr.read(1) self.assertEquals(ar.exception.errno, errno.EAGAIN) finally: fr.close() FD(w).close()
Check for EBADF in write-to-closed fd test case
Check for EBADF in write-to-closed fd test case
Python
isc
fmoo/python-filedes,fmoo/python-filedes
from filedes.test.base import BaseFDTestCase from filedes import FD import os import errno class TestFDIO(BaseFDTestCase): def testReadWrite(self): r, w = os.pipe() self.assertEquals(FD(w).write("OK"), 2) self.assertEquals(FD(r).read(2), "OK") FD(r).close() FD(w).close() def testWriteAfterClose(self): r, w = os.pipe() fw = FD(w) fw.close() with self.assertRaises(OSError): fw.write("oops") FD(r).close() def testNonblocking(self): r, w = os.pipe() fr = FD(r) try: fr.set_nonblocking() with self.assertRaises(OSError) as ar: fr.read(1) self.assertEquals(ar.exception.errno, errno.EAGAIN) finally: fr.close() FD(w).close() Check for EBADF in write-to-closed fd test case
from filedes.test.base import BaseFDTestCase from filedes import FD import os import errno class TestFDIO(BaseFDTestCase): def testReadWrite(self): r, w = os.pipe() self.assertEquals(FD(w).write("OK"), 2) self.assertEquals(FD(r).read(2), "OK") FD(r).close() FD(w).close() def testWriteAfterClose(self): r, w = os.pipe() fw = FD(w) fw.close() try: with self.assertRaises(OSError) as ar: fw.write("oops") self.assertEquals(ar.exception.errno, errno.EBADF) finally: FD(r).close() def testNonblocking(self): r, w = os.pipe() fr = FD(r) try: fr.set_nonblocking() with self.assertRaises(OSError) as ar: fr.read(1) self.assertEquals(ar.exception.errno, errno.EAGAIN) finally: fr.close() FD(w).close()
<commit_before>from filedes.test.base import BaseFDTestCase from filedes import FD import os import errno class TestFDIO(BaseFDTestCase): def testReadWrite(self): r, w = os.pipe() self.assertEquals(FD(w).write("OK"), 2) self.assertEquals(FD(r).read(2), "OK") FD(r).close() FD(w).close() def testWriteAfterClose(self): r, w = os.pipe() fw = FD(w) fw.close() with self.assertRaises(OSError): fw.write("oops") FD(r).close() def testNonblocking(self): r, w = os.pipe() fr = FD(r) try: fr.set_nonblocking() with self.assertRaises(OSError) as ar: fr.read(1) self.assertEquals(ar.exception.errno, errno.EAGAIN) finally: fr.close() FD(w).close() <commit_msg>Check for EBADF in write-to-closed fd test case<commit_after>
from filedes.test.base import BaseFDTestCase from filedes import FD import os import errno class TestFDIO(BaseFDTestCase): def testReadWrite(self): r, w = os.pipe() self.assertEquals(FD(w).write("OK"), 2) self.assertEquals(FD(r).read(2), "OK") FD(r).close() FD(w).close() def testWriteAfterClose(self): r, w = os.pipe() fw = FD(w) fw.close() try: with self.assertRaises(OSError) as ar: fw.write("oops") self.assertEquals(ar.exception.errno, errno.EBADF) finally: FD(r).close() def testNonblocking(self): r, w = os.pipe() fr = FD(r) try: fr.set_nonblocking() with self.assertRaises(OSError) as ar: fr.read(1) self.assertEquals(ar.exception.errno, errno.EAGAIN) finally: fr.close() FD(w).close()
from filedes.test.base import BaseFDTestCase from filedes import FD import os import errno class TestFDIO(BaseFDTestCase): def testReadWrite(self): r, w = os.pipe() self.assertEquals(FD(w).write("OK"), 2) self.assertEquals(FD(r).read(2), "OK") FD(r).close() FD(w).close() def testWriteAfterClose(self): r, w = os.pipe() fw = FD(w) fw.close() with self.assertRaises(OSError): fw.write("oops") FD(r).close() def testNonblocking(self): r, w = os.pipe() fr = FD(r) try: fr.set_nonblocking() with self.assertRaises(OSError) as ar: fr.read(1) self.assertEquals(ar.exception.errno, errno.EAGAIN) finally: fr.close() FD(w).close() Check for EBADF in write-to-closed fd test casefrom filedes.test.base import BaseFDTestCase from filedes import FD import os import errno class TestFDIO(BaseFDTestCase): def testReadWrite(self): r, w = os.pipe() self.assertEquals(FD(w).write("OK"), 2) self.assertEquals(FD(r).read(2), "OK") FD(r).close() FD(w).close() def testWriteAfterClose(self): r, w = os.pipe() fw = FD(w) fw.close() try: with self.assertRaises(OSError) as ar: fw.write("oops") self.assertEquals(ar.exception.errno, errno.EBADF) finally: FD(r).close() def testNonblocking(self): r, w = os.pipe() fr = FD(r) try: fr.set_nonblocking() with self.assertRaises(OSError) as ar: fr.read(1) self.assertEquals(ar.exception.errno, errno.EAGAIN) finally: fr.close() FD(w).close()
<commit_before>from filedes.test.base import BaseFDTestCase from filedes import FD import os import errno class TestFDIO(BaseFDTestCase): def testReadWrite(self): r, w = os.pipe() self.assertEquals(FD(w).write("OK"), 2) self.assertEquals(FD(r).read(2), "OK") FD(r).close() FD(w).close() def testWriteAfterClose(self): r, w = os.pipe() fw = FD(w) fw.close() with self.assertRaises(OSError): fw.write("oops") FD(r).close() def testNonblocking(self): r, w = os.pipe() fr = FD(r) try: fr.set_nonblocking() with self.assertRaises(OSError) as ar: fr.read(1) self.assertEquals(ar.exception.errno, errno.EAGAIN) finally: fr.close() FD(w).close() <commit_msg>Check for EBADF in write-to-closed fd test case<commit_after>from filedes.test.base import BaseFDTestCase from filedes import FD import os import errno class TestFDIO(BaseFDTestCase): def testReadWrite(self): r, w = os.pipe() self.assertEquals(FD(w).write("OK"), 2) self.assertEquals(FD(r).read(2), "OK") FD(r).close() FD(w).close() def testWriteAfterClose(self): r, w = os.pipe() fw = FD(w) fw.close() try: with self.assertRaises(OSError) as ar: fw.write("oops") self.assertEquals(ar.exception.errno, errno.EBADF) finally: FD(r).close() def testNonblocking(self): r, w = os.pipe() fr = FD(r) try: fr.set_nonblocking() with self.assertRaises(OSError) as ar: fr.read(1) self.assertEquals(ar.exception.errno, errno.EAGAIN) finally: fr.close() FD(w).close()
912b8a90472fc39f7c5d3b8e1e44b57aa88c0b02
setup.py
setup.py
#!/usr/bin/env python from distutils.core import setup setup(name='Numspell', version='0.9', description='A Python module for spelling numbers', author='Alexei Sholik', author_email='alcosholik@gmail.com', url='https://github.com/alco/numspell', license="MIT", packages=['numspell'], data_files=[('/usr/bin', ['spellnum'])], )
#!/usr/bin/env python from distutils.core import setup setup(name='Numspell', version='0.9', description='A Python module for spelling numbers', author='Alexei Sholik', author_email='alcosholik@gmail.com', url='https://github.com/alco/numspell', license="MIT", packages=['numspell'], scripts=['spellnum'], )
Add more idiomatic (and also portable) way to install `spellnum` script
Add more idiomatic (and also portable) way to install `spellnum` script
Python
mit
alco/numspell,alco/numspell
#!/usr/bin/env python from distutils.core import setup setup(name='Numspell', version='0.9', description='A Python module for spelling numbers', author='Alexei Sholik', author_email='alcosholik@gmail.com', url='https://github.com/alco/numspell', license="MIT", packages=['numspell'], data_files=[('/usr/bin', ['spellnum'])], ) Add more idiomatic (and also portable) way to install `spellnum` script
#!/usr/bin/env python from distutils.core import setup setup(name='Numspell', version='0.9', description='A Python module for spelling numbers', author='Alexei Sholik', author_email='alcosholik@gmail.com', url='https://github.com/alco/numspell', license="MIT", packages=['numspell'], scripts=['spellnum'], )
<commit_before>#!/usr/bin/env python from distutils.core import setup setup(name='Numspell', version='0.9', description='A Python module for spelling numbers', author='Alexei Sholik', author_email='alcosholik@gmail.com', url='https://github.com/alco/numspell', license="MIT", packages=['numspell'], data_files=[('/usr/bin', ['spellnum'])], ) <commit_msg>Add more idiomatic (and also portable) way to install `spellnum` script<commit_after>
#!/usr/bin/env python from distutils.core import setup setup(name='Numspell', version='0.9', description='A Python module for spelling numbers', author='Alexei Sholik', author_email='alcosholik@gmail.com', url='https://github.com/alco/numspell', license="MIT", packages=['numspell'], scripts=['spellnum'], )
#!/usr/bin/env python from distutils.core import setup setup(name='Numspell', version='0.9', description='A Python module for spelling numbers', author='Alexei Sholik', author_email='alcosholik@gmail.com', url='https://github.com/alco/numspell', license="MIT", packages=['numspell'], data_files=[('/usr/bin', ['spellnum'])], ) Add more idiomatic (and also portable) way to install `spellnum` script#!/usr/bin/env python from distutils.core import setup setup(name='Numspell', version='0.9', description='A Python module for spelling numbers', author='Alexei Sholik', author_email='alcosholik@gmail.com', url='https://github.com/alco/numspell', license="MIT", packages=['numspell'], scripts=['spellnum'], )
<commit_before>#!/usr/bin/env python from distutils.core import setup setup(name='Numspell', version='0.9', description='A Python module for spelling numbers', author='Alexei Sholik', author_email='alcosholik@gmail.com', url='https://github.com/alco/numspell', license="MIT", packages=['numspell'], data_files=[('/usr/bin', ['spellnum'])], ) <commit_msg>Add more idiomatic (and also portable) way to install `spellnum` script<commit_after>#!/usr/bin/env python from distutils.core import setup setup(name='Numspell', version='0.9', description='A Python module for spelling numbers', author='Alexei Sholik', author_email='alcosholik@gmail.com', url='https://github.com/alco/numspell', license="MIT", packages=['numspell'], scripts=['spellnum'], )
7bbdd8b41e6cb2afb64a7b860e51553748a160bc
setup.py
setup.py
import setuptools setuptools.setup( name="Mongothon", version="0.7.16", author="Tom Leach", author_email="tom@gc.io", description="A MongoDB object-document mapping layer for Python", license="BSD", keywords="mongo mongodb database pymongo odm validation", url="http://github.com/gamechanger/mongothon", packages=["mongothon"], long_description="Mongothon is a MongoDB object-document mapping " + "API for Python, loosely based on the awesome " + "mongoose.js library.", install_requires=['pymongo>=2.5.0', 'inflection==0.2.0', 'schemer==0.2.4'], tests_require=['mock', 'nose'] )
import setuptools setuptools.setup( name="Mongothon", version="0.7.17", author="Tom Leach", author_email="tom@gc.io", description="A MongoDB object-document mapping layer for Python", license="BSD", keywords="mongo mongodb database pymongo odm validation", url="http://github.com/gamechanger/mongothon", packages=["mongothon"], long_description="Mongothon is a MongoDB object-document mapping " + "API for Python, loosely based on the awesome " + "mongoose.js library.", install_requires=['pymongo>=2.5.0', 'inflection==0.2.0', 'schemer>=0.2.0, <0.3.0'], tests_require=['mock', 'nose'] )
Use new version of Schemer and make the version float
Use new version of Schemer and make the version float
Python
mit
gamechanger/mongothon
import setuptools setuptools.setup( name="Mongothon", version="0.7.16", author="Tom Leach", author_email="tom@gc.io", description="A MongoDB object-document mapping layer for Python", license="BSD", keywords="mongo mongodb database pymongo odm validation", url="http://github.com/gamechanger/mongothon", packages=["mongothon"], long_description="Mongothon is a MongoDB object-document mapping " + "API for Python, loosely based on the awesome " + "mongoose.js library.", install_requires=['pymongo>=2.5.0', 'inflection==0.2.0', 'schemer==0.2.4'], tests_require=['mock', 'nose'] ) Use new version of Schemer and make the version float
import setuptools setuptools.setup( name="Mongothon", version="0.7.17", author="Tom Leach", author_email="tom@gc.io", description="A MongoDB object-document mapping layer for Python", license="BSD", keywords="mongo mongodb database pymongo odm validation", url="http://github.com/gamechanger/mongothon", packages=["mongothon"], long_description="Mongothon is a MongoDB object-document mapping " + "API for Python, loosely based on the awesome " + "mongoose.js library.", install_requires=['pymongo>=2.5.0', 'inflection==0.2.0', 'schemer>=0.2.0, <0.3.0'], tests_require=['mock', 'nose'] )
<commit_before>import setuptools setuptools.setup( name="Mongothon", version="0.7.16", author="Tom Leach", author_email="tom@gc.io", description="A MongoDB object-document mapping layer for Python", license="BSD", keywords="mongo mongodb database pymongo odm validation", url="http://github.com/gamechanger/mongothon", packages=["mongothon"], long_description="Mongothon is a MongoDB object-document mapping " + "API for Python, loosely based on the awesome " + "mongoose.js library.", install_requires=['pymongo>=2.5.0', 'inflection==0.2.0', 'schemer==0.2.4'], tests_require=['mock', 'nose'] ) <commit_msg>Use new version of Schemer and make the version float<commit_after>
import setuptools setuptools.setup( name="Mongothon", version="0.7.17", author="Tom Leach", author_email="tom@gc.io", description="A MongoDB object-document mapping layer for Python", license="BSD", keywords="mongo mongodb database pymongo odm validation", url="http://github.com/gamechanger/mongothon", packages=["mongothon"], long_description="Mongothon is a MongoDB object-document mapping " + "API for Python, loosely based on the awesome " + "mongoose.js library.", install_requires=['pymongo>=2.5.0', 'inflection==0.2.0', 'schemer>=0.2.0, <0.3.0'], tests_require=['mock', 'nose'] )
import setuptools setuptools.setup( name="Mongothon", version="0.7.16", author="Tom Leach", author_email="tom@gc.io", description="A MongoDB object-document mapping layer for Python", license="BSD", keywords="mongo mongodb database pymongo odm validation", url="http://github.com/gamechanger/mongothon", packages=["mongothon"], long_description="Mongothon is a MongoDB object-document mapping " + "API for Python, loosely based on the awesome " + "mongoose.js library.", install_requires=['pymongo>=2.5.0', 'inflection==0.2.0', 'schemer==0.2.4'], tests_require=['mock', 'nose'] ) Use new version of Schemer and make the version floatimport setuptools setuptools.setup( name="Mongothon", version="0.7.17", author="Tom Leach", author_email="tom@gc.io", description="A MongoDB object-document mapping layer for Python", license="BSD", keywords="mongo mongodb database pymongo odm validation", url="http://github.com/gamechanger/mongothon", packages=["mongothon"], long_description="Mongothon is a MongoDB object-document mapping " + "API for Python, loosely based on the awesome " + "mongoose.js library.", install_requires=['pymongo>=2.5.0', 'inflection==0.2.0', 'schemer>=0.2.0, <0.3.0'], tests_require=['mock', 'nose'] )
<commit_before>import setuptools setuptools.setup( name="Mongothon", version="0.7.16", author="Tom Leach", author_email="tom@gc.io", description="A MongoDB object-document mapping layer for Python", license="BSD", keywords="mongo mongodb database pymongo odm validation", url="http://github.com/gamechanger/mongothon", packages=["mongothon"], long_description="Mongothon is a MongoDB object-document mapping " + "API for Python, loosely based on the awesome " + "mongoose.js library.", install_requires=['pymongo>=2.5.0', 'inflection==0.2.0', 'schemer==0.2.4'], tests_require=['mock', 'nose'] ) <commit_msg>Use new version of Schemer and make the version float<commit_after>import setuptools setuptools.setup( name="Mongothon", version="0.7.17", author="Tom Leach", author_email="tom@gc.io", description="A MongoDB object-document mapping layer for Python", license="BSD", keywords="mongo mongodb database pymongo odm validation", url="http://github.com/gamechanger/mongothon", packages=["mongothon"], long_description="Mongothon is a MongoDB object-document mapping " + "API for Python, loosely based on the awesome " + "mongoose.js library.", install_requires=['pymongo>=2.5.0', 'inflection==0.2.0', 'schemer>=0.2.0, <0.3.0'], tests_require=['mock', 'nose'] )
9dcd5a0f963a716e669d59429b3a8242cd65ffab
setup.py
setup.py
#!/usr/bin/env python from setuptools import setup, find_packages packages = find_packages() desc = open("README.md").read(), setup(name='cooperhewitt-swatchbook', version='0.3', description='Cooper Hewitt\'s Python tools for wrangling colours', long_description=desc, author='Cooper Hewitt Smithsonian Design Museum', url='https://github.com/cooperhewitt/py-cooperhewitt-swatchbook', requires=['colorsys', 'webcolors', 'colormath (<2.0)'], packages=packages, scripts=[], download_url='https://github.com/cooperhewitt/py-cooperhewitt-swatchbook/releases/tag/v0.3', license='BSD')
#!/usr/bin/env python from setuptools import setup, find_packages packages = find_packages() desc = open("README.md").read(), setup(name='cooperhewitt-swatchbook', version='0.3', description='Cooper Hewitt\'s Python tools for wrangling colours', long_description=desc, author='Cooper Hewitt Smithsonian Design Museum', url='https://github.com/cooperhewitt/py-cooperhewitt-swatchbook', requires=['colorsys', 'webcolors', 'colormath (<2.0)'], install_requires=['webcolors', 'colormath<2.0'], packages=packages, scripts=[], download_url='https://github.com/cooperhewitt/py-cooperhewitt-swatchbook/releases/tag/v0.3', license='BSD')
Add pip requirements support with install_requires
Add pip requirements support with install_requires
Python
bsd-3-clause
cooperhewitt/py-cooperhewitt-swatchbook
#!/usr/bin/env python from setuptools import setup, find_packages packages = find_packages() desc = open("README.md").read(), setup(name='cooperhewitt-swatchbook', version='0.3', description='Cooper Hewitt\'s Python tools for wrangling colours', long_description=desc, author='Cooper Hewitt Smithsonian Design Museum', url='https://github.com/cooperhewitt/py-cooperhewitt-swatchbook', requires=['colorsys', 'webcolors', 'colormath (<2.0)'], packages=packages, scripts=[], download_url='https://github.com/cooperhewitt/py-cooperhewitt-swatchbook/releases/tag/v0.3', license='BSD') Add pip requirements support with install_requires
#!/usr/bin/env python from setuptools import setup, find_packages packages = find_packages() desc = open("README.md").read(), setup(name='cooperhewitt-swatchbook', version='0.3', description='Cooper Hewitt\'s Python tools for wrangling colours', long_description=desc, author='Cooper Hewitt Smithsonian Design Museum', url='https://github.com/cooperhewitt/py-cooperhewitt-swatchbook', requires=['colorsys', 'webcolors', 'colormath (<2.0)'], install_requires=['webcolors', 'colormath<2.0'], packages=packages, scripts=[], download_url='https://github.com/cooperhewitt/py-cooperhewitt-swatchbook/releases/tag/v0.3', license='BSD')
<commit_before>#!/usr/bin/env python from setuptools import setup, find_packages packages = find_packages() desc = open("README.md").read(), setup(name='cooperhewitt-swatchbook', version='0.3', description='Cooper Hewitt\'s Python tools for wrangling colours', long_description=desc, author='Cooper Hewitt Smithsonian Design Museum', url='https://github.com/cooperhewitt/py-cooperhewitt-swatchbook', requires=['colorsys', 'webcolors', 'colormath (<2.0)'], packages=packages, scripts=[], download_url='https://github.com/cooperhewitt/py-cooperhewitt-swatchbook/releases/tag/v0.3', license='BSD') <commit_msg>Add pip requirements support with install_requires<commit_after>
#!/usr/bin/env python from setuptools import setup, find_packages packages = find_packages() desc = open("README.md").read(), setup(name='cooperhewitt-swatchbook', version='0.3', description='Cooper Hewitt\'s Python tools for wrangling colours', long_description=desc, author='Cooper Hewitt Smithsonian Design Museum', url='https://github.com/cooperhewitt/py-cooperhewitt-swatchbook', requires=['colorsys', 'webcolors', 'colormath (<2.0)'], install_requires=['webcolors', 'colormath<2.0'], packages=packages, scripts=[], download_url='https://github.com/cooperhewitt/py-cooperhewitt-swatchbook/releases/tag/v0.3', license='BSD')
#!/usr/bin/env python from setuptools import setup, find_packages packages = find_packages() desc = open("README.md").read(), setup(name='cooperhewitt-swatchbook', version='0.3', description='Cooper Hewitt\'s Python tools for wrangling colours', long_description=desc, author='Cooper Hewitt Smithsonian Design Museum', url='https://github.com/cooperhewitt/py-cooperhewitt-swatchbook', requires=['colorsys', 'webcolors', 'colormath (<2.0)'], packages=packages, scripts=[], download_url='https://github.com/cooperhewitt/py-cooperhewitt-swatchbook/releases/tag/v0.3', license='BSD') Add pip requirements support with install_requires#!/usr/bin/env python from setuptools import setup, find_packages packages = find_packages() desc = open("README.md").read(), setup(name='cooperhewitt-swatchbook', version='0.3', description='Cooper Hewitt\'s Python tools for wrangling colours', long_description=desc, author='Cooper Hewitt Smithsonian Design Museum', url='https://github.com/cooperhewitt/py-cooperhewitt-swatchbook', requires=['colorsys', 'webcolors', 'colormath (<2.0)'], install_requires=['webcolors', 'colormath<2.0'], packages=packages, scripts=[], download_url='https://github.com/cooperhewitt/py-cooperhewitt-swatchbook/releases/tag/v0.3', license='BSD')
<commit_before>#!/usr/bin/env python from setuptools import setup, find_packages packages = find_packages() desc = open("README.md").read(), setup(name='cooperhewitt-swatchbook', version='0.3', description='Cooper Hewitt\'s Python tools for wrangling colours', long_description=desc, author='Cooper Hewitt Smithsonian Design Museum', url='https://github.com/cooperhewitt/py-cooperhewitt-swatchbook', requires=['colorsys', 'webcolors', 'colormath (<2.0)'], packages=packages, scripts=[], download_url='https://github.com/cooperhewitt/py-cooperhewitt-swatchbook/releases/tag/v0.3', license='BSD') <commit_msg>Add pip requirements support with install_requires<commit_after>#!/usr/bin/env python from setuptools import setup, find_packages packages = find_packages() desc = open("README.md").read(), setup(name='cooperhewitt-swatchbook', version='0.3', description='Cooper Hewitt\'s Python tools for wrangling colours', long_description=desc, author='Cooper Hewitt Smithsonian Design Museum', url='https://github.com/cooperhewitt/py-cooperhewitt-swatchbook', requires=['colorsys', 'webcolors', 'colormath (<2.0)'], install_requires=['webcolors', 'colormath<2.0'], packages=packages, scripts=[], download_url='https://github.com/cooperhewitt/py-cooperhewitt-swatchbook/releases/tag/v0.3', license='BSD')
3a34bed3e0fdcfdc1d13b50f0b16b367d1133609
setup.py
setup.py
#!/usr/bin/env python # -*- coding:utf-8 -*- """ Setup script for csft """ import sys from setuptools import find_packages, setup import csft requires = [ 'argparse >= 1.2.1', 'pandas >= 0.20.1', 'pathlib >= 1.0.1', ] if sys.version <= '3.5': requires.append('scandir >= 1.5') setup( name=csft.__name__, description=csft.__doc__, version=csft.__version__, author=csft.__author__, license=csft.__license__, author_email=csft.__email__, url=csft.__url__, packages=find_packages(), entry_points={ 'console_scripts': ( 'csft = csft.__main__:main', ), }, install_requires=requires, )
#!/usr/bin/env python # -*- coding:utf-8 -*- """ Setup script for csft """ import sys from setuptools import find_packages, setup import csft requires = [ 'argparse >= 1.2.1', 'pandas >= 0.20.1', ] if sys.version < '3.4': requires.append('pathlib >= 1.0.1') if sys.version < '3.5': requires.append('scandir >= 1.5') setup( name=csft.__name__, description=csft.__doc__, version=csft.__version__, author=csft.__author__, license=csft.__license__, author_email=csft.__email__, url=csft.__url__, packages=find_packages(), entry_points={ 'console_scripts': ( 'csft = csft.__main__:main', ), }, install_requires=requires, )
Use the standard pathlib after Python 3.4
Use the standard pathlib after Python 3.4
Python
mit
yanqd0/csft
#!/usr/bin/env python # -*- coding:utf-8 -*- """ Setup script for csft """ import sys from setuptools import find_packages, setup import csft requires = [ 'argparse >= 1.2.1', 'pandas >= 0.20.1', 'pathlib >= 1.0.1', ] if sys.version <= '3.5': requires.append('scandir >= 1.5') setup( name=csft.__name__, description=csft.__doc__, version=csft.__version__, author=csft.__author__, license=csft.__license__, author_email=csft.__email__, url=csft.__url__, packages=find_packages(), entry_points={ 'console_scripts': ( 'csft = csft.__main__:main', ), }, install_requires=requires, ) Use the standard pathlib after Python 3.4
#!/usr/bin/env python # -*- coding:utf-8 -*- """ Setup script for csft """ import sys from setuptools import find_packages, setup import csft requires = [ 'argparse >= 1.2.1', 'pandas >= 0.20.1', ] if sys.version < '3.4': requires.append('pathlib >= 1.0.1') if sys.version < '3.5': requires.append('scandir >= 1.5') setup( name=csft.__name__, description=csft.__doc__, version=csft.__version__, author=csft.__author__, license=csft.__license__, author_email=csft.__email__, url=csft.__url__, packages=find_packages(), entry_points={ 'console_scripts': ( 'csft = csft.__main__:main', ), }, install_requires=requires, )
<commit_before>#!/usr/bin/env python # -*- coding:utf-8 -*- """ Setup script for csft """ import sys from setuptools import find_packages, setup import csft requires = [ 'argparse >= 1.2.1', 'pandas >= 0.20.1', 'pathlib >= 1.0.1', ] if sys.version <= '3.5': requires.append('scandir >= 1.5') setup( name=csft.__name__, description=csft.__doc__, version=csft.__version__, author=csft.__author__, license=csft.__license__, author_email=csft.__email__, url=csft.__url__, packages=find_packages(), entry_points={ 'console_scripts': ( 'csft = csft.__main__:main', ), }, install_requires=requires, ) <commit_msg>Use the standard pathlib after Python 3.4<commit_after>
#!/usr/bin/env python # -*- coding:utf-8 -*- """ Setup script for csft """ import sys from setuptools import find_packages, setup import csft requires = [ 'argparse >= 1.2.1', 'pandas >= 0.20.1', ] if sys.version < '3.4': requires.append('pathlib >= 1.0.1') if sys.version < '3.5': requires.append('scandir >= 1.5') setup( name=csft.__name__, description=csft.__doc__, version=csft.__version__, author=csft.__author__, license=csft.__license__, author_email=csft.__email__, url=csft.__url__, packages=find_packages(), entry_points={ 'console_scripts': ( 'csft = csft.__main__:main', ), }, install_requires=requires, )
#!/usr/bin/env python # -*- coding:utf-8 -*- """ Setup script for csft """ import sys from setuptools import find_packages, setup import csft requires = [ 'argparse >= 1.2.1', 'pandas >= 0.20.1', 'pathlib >= 1.0.1', ] if sys.version <= '3.5': requires.append('scandir >= 1.5') setup( name=csft.__name__, description=csft.__doc__, version=csft.__version__, author=csft.__author__, license=csft.__license__, author_email=csft.__email__, url=csft.__url__, packages=find_packages(), entry_points={ 'console_scripts': ( 'csft = csft.__main__:main', ), }, install_requires=requires, ) Use the standard pathlib after Python 3.4#!/usr/bin/env python # -*- coding:utf-8 -*- """ Setup script for csft """ import sys from setuptools import find_packages, setup import csft requires = [ 'argparse >= 1.2.1', 'pandas >= 0.20.1', ] if sys.version < '3.4': requires.append('pathlib >= 1.0.1') if sys.version < '3.5': requires.append('scandir >= 1.5') setup( name=csft.__name__, description=csft.__doc__, version=csft.__version__, author=csft.__author__, license=csft.__license__, author_email=csft.__email__, url=csft.__url__, packages=find_packages(), entry_points={ 'console_scripts': ( 'csft = csft.__main__:main', ), }, install_requires=requires, )
<commit_before>#!/usr/bin/env python # -*- coding:utf-8 -*- """ Setup script for csft """ import sys from setuptools import find_packages, setup import csft requires = [ 'argparse >= 1.2.1', 'pandas >= 0.20.1', 'pathlib >= 1.0.1', ] if sys.version <= '3.5': requires.append('scandir >= 1.5') setup( name=csft.__name__, description=csft.__doc__, version=csft.__version__, author=csft.__author__, license=csft.__license__, author_email=csft.__email__, url=csft.__url__, packages=find_packages(), entry_points={ 'console_scripts': ( 'csft = csft.__main__:main', ), }, install_requires=requires, ) <commit_msg>Use the standard pathlib after Python 3.4<commit_after>#!/usr/bin/env python # -*- coding:utf-8 -*- """ Setup script for csft """ import sys from setuptools import find_packages, setup import csft requires = [ 'argparse >= 1.2.1', 'pandas >= 0.20.1', ] if sys.version < '3.4': requires.append('pathlib >= 1.0.1') if sys.version < '3.5': requires.append('scandir >= 1.5') setup( name=csft.__name__, description=csft.__doc__, version=csft.__version__, author=csft.__author__, license=csft.__license__, author_email=csft.__email__, url=csft.__url__, packages=find_packages(), entry_points={ 'console_scripts': ( 'csft = csft.__main__:main', ), }, install_requires=requires, )
a3064f6feeabc47b01fb01e75771646400c80949
setup.py
setup.py
from setuptools import setup setup(name='uvp_adb', version='1.0', description='adb functions for uvp phone', url='https://www.linkedin.com/in/sajjad-ziyaei-amiri-a1a16b9b', author='Sajjad Ziyaei amiri', author_email='szaforward@gmail.com', license='MIT', packages=['uvp_adb'], zip_safe=False)
from setuptools import setup setup(name='uvp_adb', version='1.0', description='adb functions for uvp phone', url='https://github.com/inforfang/uvp_adb', download_url = 'https://github.com/inforfang/uvp_adb/tarball/1.0', author='inforfang', author_email='szaforward2@gmail.com', license='MIT', packages=['uvp_adb'], zip_safe=False)
Correct urldownload and email address
Correct urldownload and email address
Python
mit
inforfang/uvp_adb
from setuptools import setup setup(name='uvp_adb', version='1.0', description='adb functions for uvp phone', url='https://www.linkedin.com/in/sajjad-ziyaei-amiri-a1a16b9b', author='Sajjad Ziyaei amiri', author_email='szaforward@gmail.com', license='MIT', packages=['uvp_adb'], zip_safe=False)Correct urldownload and email address
from setuptools import setup setup(name='uvp_adb', version='1.0', description='adb functions for uvp phone', url='https://github.com/inforfang/uvp_adb', download_url = 'https://github.com/inforfang/uvp_adb/tarball/1.0', author='inforfang', author_email='szaforward2@gmail.com', license='MIT', packages=['uvp_adb'], zip_safe=False)
<commit_before>from setuptools import setup setup(name='uvp_adb', version='1.0', description='adb functions for uvp phone', url='https://www.linkedin.com/in/sajjad-ziyaei-amiri-a1a16b9b', author='Sajjad Ziyaei amiri', author_email='szaforward@gmail.com', license='MIT', packages=['uvp_adb'], zip_safe=False)<commit_msg>Correct urldownload and email address<commit_after>
from setuptools import setup setup(name='uvp_adb', version='1.0', description='adb functions for uvp phone', url='https://github.com/inforfang/uvp_adb', download_url = 'https://github.com/inforfang/uvp_adb/tarball/1.0', author='inforfang', author_email='szaforward2@gmail.com', license='MIT', packages=['uvp_adb'], zip_safe=False)
from setuptools import setup setup(name='uvp_adb', version='1.0', description='adb functions for uvp phone', url='https://www.linkedin.com/in/sajjad-ziyaei-amiri-a1a16b9b', author='Sajjad Ziyaei amiri', author_email='szaforward@gmail.com', license='MIT', packages=['uvp_adb'], zip_safe=False)Correct urldownload and email addressfrom setuptools import setup setup(name='uvp_adb', version='1.0', description='adb functions for uvp phone', url='https://github.com/inforfang/uvp_adb', download_url = 'https://github.com/inforfang/uvp_adb/tarball/1.0', author='inforfang', author_email='szaforward2@gmail.com', license='MIT', packages=['uvp_adb'], zip_safe=False)
<commit_before>from setuptools import setup setup(name='uvp_adb', version='1.0', description='adb functions for uvp phone', url='https://www.linkedin.com/in/sajjad-ziyaei-amiri-a1a16b9b', author='Sajjad Ziyaei amiri', author_email='szaforward@gmail.com', license='MIT', packages=['uvp_adb'], zip_safe=False)<commit_msg>Correct urldownload and email address<commit_after>from setuptools import setup setup(name='uvp_adb', version='1.0', description='adb functions for uvp phone', url='https://github.com/inforfang/uvp_adb', download_url = 'https://github.com/inforfang/uvp_adb/tarball/1.0', author='inforfang', author_email='szaforward2@gmail.com', license='MIT', packages=['uvp_adb'], zip_safe=False)
5004b500284ec1d0f709b3069867757770601c44
mla_game/apps/transcript/management/commands/recalculate_phrase_game_mappings.py
mla_game/apps/transcript/management/commands/recalculate_phrase_game_mappings.py
from django.core.management.base import BaseCommand from django.conf import settings from ...models import TranscriptPhrase from ...tasks import assign_current_game phrase_positive_limit = settings.TRANSCRIPT_PHRASE_POSITIVE_CONFIDENCE_LIMIT phrase_negative_limit = settings.TRANSCRIPT_PHRASE_NEGATIVE_CONFIDENCE_LIMIT correction_lower_limit = settings.TRANSCRIPT_PHRASE_CORRECTION_LOWER_LIMIT correction_upper_limit = settings.TRANSCRIPT_PHRASE_CORRECTION_UPPER_LIMIT class Command(BaseCommand): help = '''Recalculate game eligiblity for all phrases''' def handle(self, *args, **options): for phrase in TranscriptPhrase.objects.all(): assign_current_game(phrase)
from django.core.management.base import BaseCommand from django.conf import settings from ...models import TranscriptPhraseVote from ...tasks import assign_current_game phrase_positive_limit = settings.TRANSCRIPT_PHRASE_POSITIVE_CONFIDENCE_LIMIT phrase_negative_limit = settings.TRANSCRIPT_PHRASE_NEGATIVE_CONFIDENCE_LIMIT correction_lower_limit = settings.TRANSCRIPT_PHRASE_CORRECTION_LOWER_LIMIT correction_upper_limit = settings.TRANSCRIPT_PHRASE_CORRECTION_UPPER_LIMIT class Command(BaseCommand): help = '''Recalculate game eligiblity for all phrases''' def handle(self, *args, **options): to_assign = set() for vote in TranscriptPhraseVote.objects.all(): to_assign.add(vote.transcript_phrase) for phrase in to_assign: assign_current_game(phrase)
Update only phrases with user input
Update only phrases with user input
Python
mit
WGBH/FixIt,WGBH/FixIt,WGBH/FixIt
from django.core.management.base import BaseCommand from django.conf import settings from ...models import TranscriptPhrase from ...tasks import assign_current_game phrase_positive_limit = settings.TRANSCRIPT_PHRASE_POSITIVE_CONFIDENCE_LIMIT phrase_negative_limit = settings.TRANSCRIPT_PHRASE_NEGATIVE_CONFIDENCE_LIMIT correction_lower_limit = settings.TRANSCRIPT_PHRASE_CORRECTION_LOWER_LIMIT correction_upper_limit = settings.TRANSCRIPT_PHRASE_CORRECTION_UPPER_LIMIT class Command(BaseCommand): help = '''Recalculate game eligiblity for all phrases''' def handle(self, *args, **options): for phrase in TranscriptPhrase.objects.all(): assign_current_game(phrase) Update only phrases with user input
from django.core.management.base import BaseCommand from django.conf import settings from ...models import TranscriptPhraseVote from ...tasks import assign_current_game phrase_positive_limit = settings.TRANSCRIPT_PHRASE_POSITIVE_CONFIDENCE_LIMIT phrase_negative_limit = settings.TRANSCRIPT_PHRASE_NEGATIVE_CONFIDENCE_LIMIT correction_lower_limit = settings.TRANSCRIPT_PHRASE_CORRECTION_LOWER_LIMIT correction_upper_limit = settings.TRANSCRIPT_PHRASE_CORRECTION_UPPER_LIMIT class Command(BaseCommand): help = '''Recalculate game eligiblity for all phrases''' def handle(self, *args, **options): to_assign = set() for vote in TranscriptPhraseVote.objects.all(): to_assign.add(vote.transcript_phrase) for phrase in to_assign: assign_current_game(phrase)
<commit_before>from django.core.management.base import BaseCommand from django.conf import settings from ...models import TranscriptPhrase from ...tasks import assign_current_game phrase_positive_limit = settings.TRANSCRIPT_PHRASE_POSITIVE_CONFIDENCE_LIMIT phrase_negative_limit = settings.TRANSCRIPT_PHRASE_NEGATIVE_CONFIDENCE_LIMIT correction_lower_limit = settings.TRANSCRIPT_PHRASE_CORRECTION_LOWER_LIMIT correction_upper_limit = settings.TRANSCRIPT_PHRASE_CORRECTION_UPPER_LIMIT class Command(BaseCommand): help = '''Recalculate game eligiblity for all phrases''' def handle(self, *args, **options): for phrase in TranscriptPhrase.objects.all(): assign_current_game(phrase) <commit_msg>Update only phrases with user input<commit_after>
from django.core.management.base import BaseCommand from django.conf import settings from ...models import TranscriptPhraseVote from ...tasks import assign_current_game phrase_positive_limit = settings.TRANSCRIPT_PHRASE_POSITIVE_CONFIDENCE_LIMIT phrase_negative_limit = settings.TRANSCRIPT_PHRASE_NEGATIVE_CONFIDENCE_LIMIT correction_lower_limit = settings.TRANSCRIPT_PHRASE_CORRECTION_LOWER_LIMIT correction_upper_limit = settings.TRANSCRIPT_PHRASE_CORRECTION_UPPER_LIMIT class Command(BaseCommand): help = '''Recalculate game eligiblity for all phrases''' def handle(self, *args, **options): to_assign = set() for vote in TranscriptPhraseVote.objects.all(): to_assign.add(vote.transcript_phrase) for phrase in to_assign: assign_current_game(phrase)
from django.core.management.base import BaseCommand from django.conf import settings from ...models import TranscriptPhrase from ...tasks import assign_current_game phrase_positive_limit = settings.TRANSCRIPT_PHRASE_POSITIVE_CONFIDENCE_LIMIT phrase_negative_limit = settings.TRANSCRIPT_PHRASE_NEGATIVE_CONFIDENCE_LIMIT correction_lower_limit = settings.TRANSCRIPT_PHRASE_CORRECTION_LOWER_LIMIT correction_upper_limit = settings.TRANSCRIPT_PHRASE_CORRECTION_UPPER_LIMIT class Command(BaseCommand): help = '''Recalculate game eligiblity for all phrases''' def handle(self, *args, **options): for phrase in TranscriptPhrase.objects.all(): assign_current_game(phrase) Update only phrases with user inputfrom django.core.management.base import BaseCommand from django.conf import settings from ...models import TranscriptPhraseVote from ...tasks import assign_current_game phrase_positive_limit = settings.TRANSCRIPT_PHRASE_POSITIVE_CONFIDENCE_LIMIT phrase_negative_limit = settings.TRANSCRIPT_PHRASE_NEGATIVE_CONFIDENCE_LIMIT correction_lower_limit = settings.TRANSCRIPT_PHRASE_CORRECTION_LOWER_LIMIT correction_upper_limit = settings.TRANSCRIPT_PHRASE_CORRECTION_UPPER_LIMIT class Command(BaseCommand): help = '''Recalculate game eligiblity for all phrases''' def handle(self, *args, **options): to_assign = set() for vote in TranscriptPhraseVote.objects.all(): to_assign.add(vote.transcript_phrase) for phrase in to_assign: assign_current_game(phrase)
<commit_before>from django.core.management.base import BaseCommand from django.conf import settings from ...models import TranscriptPhrase from ...tasks import assign_current_game phrase_positive_limit = settings.TRANSCRIPT_PHRASE_POSITIVE_CONFIDENCE_LIMIT phrase_negative_limit = settings.TRANSCRIPT_PHRASE_NEGATIVE_CONFIDENCE_LIMIT correction_lower_limit = settings.TRANSCRIPT_PHRASE_CORRECTION_LOWER_LIMIT correction_upper_limit = settings.TRANSCRIPT_PHRASE_CORRECTION_UPPER_LIMIT class Command(BaseCommand): help = '''Recalculate game eligiblity for all phrases''' def handle(self, *args, **options): for phrase in TranscriptPhrase.objects.all(): assign_current_game(phrase) <commit_msg>Update only phrases with user input<commit_after>from django.core.management.base import BaseCommand from django.conf import settings from ...models import TranscriptPhraseVote from ...tasks import assign_current_game phrase_positive_limit = settings.TRANSCRIPT_PHRASE_POSITIVE_CONFIDENCE_LIMIT phrase_negative_limit = settings.TRANSCRIPT_PHRASE_NEGATIVE_CONFIDENCE_LIMIT correction_lower_limit = settings.TRANSCRIPT_PHRASE_CORRECTION_LOWER_LIMIT correction_upper_limit = settings.TRANSCRIPT_PHRASE_CORRECTION_UPPER_LIMIT class Command(BaseCommand): help = '''Recalculate game eligiblity for all phrases''' def handle(self, *args, **options): to_assign = set() for vote in TranscriptPhraseVote.objects.all(): to_assign.add(vote.transcript_phrase) for phrase in to_assign: assign_current_game(phrase)
05b5a4390c7109bb8c0778c547883c41272769fb
examples/oauth.py
examples/oauth.py
import tweepy # == OAuth Authentication == # # This mode of authentication is the new preferred way # of authenticating with Twitter. # The consumer keys can be found on your application's Details # page located at https://dev.twitter.com/apps (under "OAuth settings") consumer_key="" consumer_secret="" # The access tokens can be found on your applications's Details # page located at https://dev.twitter.com/apps (located # under "Your access token") access_token="" access_token_secret="" auth = tweepy.OAuthHandler(consumer_key, consumer_secret) auth.set_access_token(access_token, access_token_secret) api = tweepy.API(auth) # If the authentication was successful, you should # see the name of the account print out print api.me().name # If the application settings are set for "Read and Write" then # this line should tweet out the message to your account's # timeline. The "Read and Write" setting is on https://dev.twitter.com/apps api.update_status('Updating using OAuth authentication via Tweepy!')
import tweepy # == OAuth Authentication == # # This mode of authentication is the new preferred way # of authenticating with Twitter. # The consumer keys can be found on your application's Details # page located at https://dev.twitter.com/apps (under "OAuth settings") consumer_key="" consumer_secret="" # The access tokens can be found on your applications's Details # page located at https://dev.twitter.com/apps (located # under "Your access token") access_token="" access_token_secret="" auth = tweepy.OAuthHandler(consumer_key, consumer_secret) auth.secure = True auth.set_access_token(access_token, access_token_secret) api = tweepy.API(auth) # If the authentication was successful, you should # see the name of the account print out print api.me().name # If the application settings are set for "Read and Write" then # this line should tweet out the message to your account's # timeline. The "Read and Write" setting is on https://dev.twitter.com/apps api.update_status('Updating using OAuth authentication via Tweepy!')
Change to make sure tweepy use SSL
Change to make sure tweepy use SSL I had some problems using Tweepy and I realized it was because it didn't always use SSL. So I propose to add a line to make sure the connexion we use is secure.
Python
mit
vishnugonela/tweepy,iamjakob/tweepy,srimanthd/tweepy,sa8/tweepy,awangga/tweepy,zhenv5/tweepy,kcompher/tweepy,damchilly/tweepy,edsu/tweepy,cogniteev/tweepy,yared-bezum/tweepy,truekonrads/tweepy,tweepy/tweepy,Choko256/tweepy,elijah513/tweepy,tsablic/tweepy,hackebrot/tweepy,rudraksh125/tweepy,atomicjets/tweepy,IsaacHaze/tweepy,aleczadikian/tweepy,bconnelly/tweepy,LikeABird/tweepy,kylemanna/tweepy,conversocial/tweepy,vikasgorur/tweepy,arunxarun/tweepy,kskk02/tweepy,robbiewoods05/tweepy,sidewire/tweepy,aganzha/tweepy,obskyr/tweepy,raymondethan/tweepy,thelostscientist/tweepy,tuxos/tweepy,vivek8943/tweepy,techieshark/tweepy,svven/tweepy,wjt/tweepy,takeshineshiro/tweepy
import tweepy # == OAuth Authentication == # # This mode of authentication is the new preferred way # of authenticating with Twitter. # The consumer keys can be found on your application's Details # page located at https://dev.twitter.com/apps (under "OAuth settings") consumer_key="" consumer_secret="" # The access tokens can be found on your applications's Details # page located at https://dev.twitter.com/apps (located # under "Your access token") access_token="" access_token_secret="" auth = tweepy.OAuthHandler(consumer_key, consumer_secret) auth.set_access_token(access_token, access_token_secret) api = tweepy.API(auth) # If the authentication was successful, you should # see the name of the account print out print api.me().name # If the application settings are set for "Read and Write" then # this line should tweet out the message to your account's # timeline. The "Read and Write" setting is on https://dev.twitter.com/apps api.update_status('Updating using OAuth authentication via Tweepy!') Change to make sure tweepy use SSL I had some problems using Tweepy and I realized it was because it didn't always use SSL. So I propose to add a line to make sure the connexion we use is secure.
import tweepy # == OAuth Authentication == # # This mode of authentication is the new preferred way # of authenticating with Twitter. # The consumer keys can be found on your application's Details # page located at https://dev.twitter.com/apps (under "OAuth settings") consumer_key="" consumer_secret="" # The access tokens can be found on your applications's Details # page located at https://dev.twitter.com/apps (located # under "Your access token") access_token="" access_token_secret="" auth = tweepy.OAuthHandler(consumer_key, consumer_secret) auth.secure = True auth.set_access_token(access_token, access_token_secret) api = tweepy.API(auth) # If the authentication was successful, you should # see the name of the account print out print api.me().name # If the application settings are set for "Read and Write" then # this line should tweet out the message to your account's # timeline. The "Read and Write" setting is on https://dev.twitter.com/apps api.update_status('Updating using OAuth authentication via Tweepy!')
<commit_before>import tweepy # == OAuth Authentication == # # This mode of authentication is the new preferred way # of authenticating with Twitter. # The consumer keys can be found on your application's Details # page located at https://dev.twitter.com/apps (under "OAuth settings") consumer_key="" consumer_secret="" # The access tokens can be found on your applications's Details # page located at https://dev.twitter.com/apps (located # under "Your access token") access_token="" access_token_secret="" auth = tweepy.OAuthHandler(consumer_key, consumer_secret) auth.set_access_token(access_token, access_token_secret) api = tweepy.API(auth) # If the authentication was successful, you should # see the name of the account print out print api.me().name # If the application settings are set for "Read and Write" then # this line should tweet out the message to your account's # timeline. The "Read and Write" setting is on https://dev.twitter.com/apps api.update_status('Updating using OAuth authentication via Tweepy!') <commit_msg>Change to make sure tweepy use SSL I had some problems using Tweepy and I realized it was because it didn't always use SSL. So I propose to add a line to make sure the connexion we use is secure.<commit_after>
import tweepy # == OAuth Authentication == # # This mode of authentication is the new preferred way # of authenticating with Twitter. # The consumer keys can be found on your application's Details # page located at https://dev.twitter.com/apps (under "OAuth settings") consumer_key="" consumer_secret="" # The access tokens can be found on your applications's Details # page located at https://dev.twitter.com/apps (located # under "Your access token") access_token="" access_token_secret="" auth = tweepy.OAuthHandler(consumer_key, consumer_secret) auth.secure = True auth.set_access_token(access_token, access_token_secret) api = tweepy.API(auth) # If the authentication was successful, you should # see the name of the account print out print api.me().name # If the application settings are set for "Read and Write" then # this line should tweet out the message to your account's # timeline. The "Read and Write" setting is on https://dev.twitter.com/apps api.update_status('Updating using OAuth authentication via Tweepy!')
import tweepy # == OAuth Authentication == # # This mode of authentication is the new preferred way # of authenticating with Twitter. # The consumer keys can be found on your application's Details # page located at https://dev.twitter.com/apps (under "OAuth settings") consumer_key="" consumer_secret="" # The access tokens can be found on your applications's Details # page located at https://dev.twitter.com/apps (located # under "Your access token") access_token="" access_token_secret="" auth = tweepy.OAuthHandler(consumer_key, consumer_secret) auth.set_access_token(access_token, access_token_secret) api = tweepy.API(auth) # If the authentication was successful, you should # see the name of the account print out print api.me().name # If the application settings are set for "Read and Write" then # this line should tweet out the message to your account's # timeline. The "Read and Write" setting is on https://dev.twitter.com/apps api.update_status('Updating using OAuth authentication via Tweepy!') Change to make sure tweepy use SSL I had some problems using Tweepy and I realized it was because it didn't always use SSL. So I propose to add a line to make sure the connexion we use is secure.import tweepy # == OAuth Authentication == # # This mode of authentication is the new preferred way # of authenticating with Twitter. # The consumer keys can be found on your application's Details # page located at https://dev.twitter.com/apps (under "OAuth settings") consumer_key="" consumer_secret="" # The access tokens can be found on your applications's Details # page located at https://dev.twitter.com/apps (located # under "Your access token") access_token="" access_token_secret="" auth = tweepy.OAuthHandler(consumer_key, consumer_secret) auth.secure = True auth.set_access_token(access_token, access_token_secret) api = tweepy.API(auth) # If the authentication was successful, you should # see the name of the account print out print api.me().name # If the application settings are set for "Read and Write" then # this line should tweet out the message to your account's # timeline. The "Read and Write" setting is on https://dev.twitter.com/apps api.update_status('Updating using OAuth authentication via Tweepy!')
<commit_before>import tweepy # == OAuth Authentication == # # This mode of authentication is the new preferred way # of authenticating with Twitter. # The consumer keys can be found on your application's Details # page located at https://dev.twitter.com/apps (under "OAuth settings") consumer_key="" consumer_secret="" # The access tokens can be found on your applications's Details # page located at https://dev.twitter.com/apps (located # under "Your access token") access_token="" access_token_secret="" auth = tweepy.OAuthHandler(consumer_key, consumer_secret) auth.set_access_token(access_token, access_token_secret) api = tweepy.API(auth) # If the authentication was successful, you should # see the name of the account print out print api.me().name # If the application settings are set for "Read and Write" then # this line should tweet out the message to your account's # timeline. The "Read and Write" setting is on https://dev.twitter.com/apps api.update_status('Updating using OAuth authentication via Tweepy!') <commit_msg>Change to make sure tweepy use SSL I had some problems using Tweepy and I realized it was because it didn't always use SSL. So I propose to add a line to make sure the connexion we use is secure.<commit_after>import tweepy # == OAuth Authentication == # # This mode of authentication is the new preferred way # of authenticating with Twitter. # The consumer keys can be found on your application's Details # page located at https://dev.twitter.com/apps (under "OAuth settings") consumer_key="" consumer_secret="" # The access tokens can be found on your applications's Details # page located at https://dev.twitter.com/apps (located # under "Your access token") access_token="" access_token_secret="" auth = tweepy.OAuthHandler(consumer_key, consumer_secret) auth.secure = True auth.set_access_token(access_token, access_token_secret) api = tweepy.API(auth) # If the authentication was successful, you should # see the name of the account print out print api.me().name # If the application settings are set for "Read and Write" then # this line should tweet out the message to your account's # timeline. The "Read and Write" setting is on https://dev.twitter.com/apps api.update_status('Updating using OAuth authentication via Tweepy!')
1d8f7d4a57b145fa3f8cce12a55b02eb0a754581
crypto/envelope.py
crypto/envelope.py
""" Sealed envelope example from Princeton course 'Intro to Crypto and Cryptocurrencies' https://www.youtube.com/watch?v=fOMVZXLjKYo """ import hashlib def commit(key, msg): m = hashlib.sha256() m.update(key) m.update(msg) return { "hash": m.hexdigest(), "key": key } def verify(com, key, msg): v = hashlib.sha256() v.update(key) v.update(msg) try: assert v.hexdigest() == com return "Message has been verified" except: raise sealed_msg = 'there' msg = commit('hey', sealed_msg) public_msg = msg.get('hash') print(verify(public_msg, msg.get('key'), sealed_msg))
""" Sealed envelope example from Princeton course 'Intro to Crypto and Cryptocurrencies' https://www.youtube.com/watch?v=fOMVZXLjKYo """ import hashlib def commit(key, msg): m = hashlib.sha256() m.update(key) m.update(msg) return { "hash": m.hexdigest(), "key": key } def verify(com, key, msg): v = hashlib.sha256() v.update(key) v.update(msg) try: assert v.hexdigest() == com return "Message has been verified" except: raise sealed_msg = 'there' key = hashlib.sha256().hexdigest() msg = commit(key, sealed_msg) public_msg = msg.get('hash') print(verify(public_msg, msg.get('key'), sealed_msg))
Make the key 'more secure'
Make the key 'more secure'
Python
mit
b-ritter/python-notes,b-ritter/python-notes
""" Sealed envelope example from Princeton course 'Intro to Crypto and Cryptocurrencies' https://www.youtube.com/watch?v=fOMVZXLjKYo """ import hashlib def commit(key, msg): m = hashlib.sha256() m.update(key) m.update(msg) return { "hash": m.hexdigest(), "key": key } def verify(com, key, msg): v = hashlib.sha256() v.update(key) v.update(msg) try: assert v.hexdigest() == com return "Message has been verified" except: raise sealed_msg = 'there' msg = commit('hey', sealed_msg) public_msg = msg.get('hash') print(verify(public_msg, msg.get('key'), sealed_msg))Make the key 'more secure'
""" Sealed envelope example from Princeton course 'Intro to Crypto and Cryptocurrencies' https://www.youtube.com/watch?v=fOMVZXLjKYo """ import hashlib def commit(key, msg): m = hashlib.sha256() m.update(key) m.update(msg) return { "hash": m.hexdigest(), "key": key } def verify(com, key, msg): v = hashlib.sha256() v.update(key) v.update(msg) try: assert v.hexdigest() == com return "Message has been verified" except: raise sealed_msg = 'there' key = hashlib.sha256().hexdigest() msg = commit(key, sealed_msg) public_msg = msg.get('hash') print(verify(public_msg, msg.get('key'), sealed_msg))
<commit_before>""" Sealed envelope example from Princeton course 'Intro to Crypto and Cryptocurrencies' https://www.youtube.com/watch?v=fOMVZXLjKYo """ import hashlib def commit(key, msg): m = hashlib.sha256() m.update(key) m.update(msg) return { "hash": m.hexdigest(), "key": key } def verify(com, key, msg): v = hashlib.sha256() v.update(key) v.update(msg) try: assert v.hexdigest() == com return "Message has been verified" except: raise sealed_msg = 'there' msg = commit('hey', sealed_msg) public_msg = msg.get('hash') print(verify(public_msg, msg.get('key'), sealed_msg))<commit_msg>Make the key 'more secure'<commit_after>
""" Sealed envelope example from Princeton course 'Intro to Crypto and Cryptocurrencies' https://www.youtube.com/watch?v=fOMVZXLjKYo """ import hashlib def commit(key, msg): m = hashlib.sha256() m.update(key) m.update(msg) return { "hash": m.hexdigest(), "key": key } def verify(com, key, msg): v = hashlib.sha256() v.update(key) v.update(msg) try: assert v.hexdigest() == com return "Message has been verified" except: raise sealed_msg = 'there' key = hashlib.sha256().hexdigest() msg = commit(key, sealed_msg) public_msg = msg.get('hash') print(verify(public_msg, msg.get('key'), sealed_msg))
""" Sealed envelope example from Princeton course 'Intro to Crypto and Cryptocurrencies' https://www.youtube.com/watch?v=fOMVZXLjKYo """ import hashlib def commit(key, msg): m = hashlib.sha256() m.update(key) m.update(msg) return { "hash": m.hexdigest(), "key": key } def verify(com, key, msg): v = hashlib.sha256() v.update(key) v.update(msg) try: assert v.hexdigest() == com return "Message has been verified" except: raise sealed_msg = 'there' msg = commit('hey', sealed_msg) public_msg = msg.get('hash') print(verify(public_msg, msg.get('key'), sealed_msg))Make the key 'more secure'""" Sealed envelope example from Princeton course 'Intro to Crypto and Cryptocurrencies' https://www.youtube.com/watch?v=fOMVZXLjKYo """ import hashlib def commit(key, msg): m = hashlib.sha256() m.update(key) m.update(msg) return { "hash": m.hexdigest(), "key": key } def verify(com, key, msg): v = hashlib.sha256() v.update(key) v.update(msg) try: assert v.hexdigest() == com return "Message has been verified" except: raise sealed_msg = 'there' key = hashlib.sha256().hexdigest() msg = commit(key, sealed_msg) public_msg = msg.get('hash') print(verify(public_msg, msg.get('key'), sealed_msg))
<commit_before>""" Sealed envelope example from Princeton course 'Intro to Crypto and Cryptocurrencies' https://www.youtube.com/watch?v=fOMVZXLjKYo """ import hashlib def commit(key, msg): m = hashlib.sha256() m.update(key) m.update(msg) return { "hash": m.hexdigest(), "key": key } def verify(com, key, msg): v = hashlib.sha256() v.update(key) v.update(msg) try: assert v.hexdigest() == com return "Message has been verified" except: raise sealed_msg = 'there' msg = commit('hey', sealed_msg) public_msg = msg.get('hash') print(verify(public_msg, msg.get('key'), sealed_msg))<commit_msg>Make the key 'more secure'<commit_after>""" Sealed envelope example from Princeton course 'Intro to Crypto and Cryptocurrencies' https://www.youtube.com/watch?v=fOMVZXLjKYo """ import hashlib def commit(key, msg): m = hashlib.sha256() m.update(key) m.update(msg) return { "hash": m.hexdigest(), "key": key } def verify(com, key, msg): v = hashlib.sha256() v.update(key) v.update(msg) try: assert v.hexdigest() == com return "Message has been verified" except: raise sealed_msg = 'there' key = hashlib.sha256().hexdigest() msg = commit(key, sealed_msg) public_msg = msg.get('hash') print(verify(public_msg, msg.get('key'), sealed_msg))
2997b84a5ead65f6c17128baef3e3039957d97f8
dduplicated/cli.py
dduplicated/cli.py
# The client of DDuplicated tool. from os import path as opath, getcwd from sys import argv from dduplicated import commands def get_paths(params): paths = [] for param in params: path = opath.join(getcwd(), param) if opath.exists(path) and opath.isdir(path) and not opath.islink(path): paths.append(path) return paths def main(): params = argv processed_files = [] # Remove the command name del params[0] if len(params) == 0 or "help" in params: commands.help() elif "detect" in params: processed_files = commands.detect(get_paths(params)) elif "delete" in params: processed_files = commands.delete(commands.detect(get_paths(params))) elif "link" in params: processed_files = commands.link(commands.detect(get_paths(params))) else: commands.help() if len(processed_files) > 0: print(processed_files) else: print("No duplicates found") print("Great! Bye!") exit(0)
# The client of DDuplicated tool. from os import path as opath, getcwd from pprint import pprint from sys import argv from dduplicated import commands def get_paths(params): paths = [] for param in params: path = opath.join(getcwd(), param) if opath.exists(path) and opath.isdir(path) and not opath.islink(path): paths.append(path) return paths def main(): params = argv processed_files = [] # Remove the command name del params[0] if len(params) == 0 or "help" in params: commands.help() elif "detect" in params: processed_files = commands.detect(get_paths(params)) elif "delete" in params: processed_files = commands.delete(commands.detect(get_paths(params))) elif "link" in params: processed_files = commands.link(commands.detect(get_paths(params))) else: commands.help() if len(processed_files) > 0: pprint(processed_files) else: print("No duplicates found") print("Great! Bye!") exit(0)
Update in output to terminal.
Update in output to terminal. Signed-off-by: messiasthi <8562fc1efba9a3c99753c749fdfb1b6932b70fbf@gmail.com>
Python
mit
messiasthi/dduplicated-cli
# The client of DDuplicated tool. from os import path as opath, getcwd from sys import argv from dduplicated import commands def get_paths(params): paths = [] for param in params: path = opath.join(getcwd(), param) if opath.exists(path) and opath.isdir(path) and not opath.islink(path): paths.append(path) return paths def main(): params = argv processed_files = [] # Remove the command name del params[0] if len(params) == 0 or "help" in params: commands.help() elif "detect" in params: processed_files = commands.detect(get_paths(params)) elif "delete" in params: processed_files = commands.delete(commands.detect(get_paths(params))) elif "link" in params: processed_files = commands.link(commands.detect(get_paths(params))) else: commands.help() if len(processed_files) > 0: print(processed_files) else: print("No duplicates found") print("Great! Bye!") exit(0) Update in output to terminal. Signed-off-by: messiasthi <8562fc1efba9a3c99753c749fdfb1b6932b70fbf@gmail.com>
# The client of DDuplicated tool. from os import path as opath, getcwd from pprint import pprint from sys import argv from dduplicated import commands def get_paths(params): paths = [] for param in params: path = opath.join(getcwd(), param) if opath.exists(path) and opath.isdir(path) and not opath.islink(path): paths.append(path) return paths def main(): params = argv processed_files = [] # Remove the command name del params[0] if len(params) == 0 or "help" in params: commands.help() elif "detect" in params: processed_files = commands.detect(get_paths(params)) elif "delete" in params: processed_files = commands.delete(commands.detect(get_paths(params))) elif "link" in params: processed_files = commands.link(commands.detect(get_paths(params))) else: commands.help() if len(processed_files) > 0: pprint(processed_files) else: print("No duplicates found") print("Great! Bye!") exit(0)
<commit_before># The client of DDuplicated tool. from os import path as opath, getcwd from sys import argv from dduplicated import commands def get_paths(params): paths = [] for param in params: path = opath.join(getcwd(), param) if opath.exists(path) and opath.isdir(path) and not opath.islink(path): paths.append(path) return paths def main(): params = argv processed_files = [] # Remove the command name del params[0] if len(params) == 0 or "help" in params: commands.help() elif "detect" in params: processed_files = commands.detect(get_paths(params)) elif "delete" in params: processed_files = commands.delete(commands.detect(get_paths(params))) elif "link" in params: processed_files = commands.link(commands.detect(get_paths(params))) else: commands.help() if len(processed_files) > 0: print(processed_files) else: print("No duplicates found") print("Great! Bye!") exit(0) <commit_msg>Update in output to terminal. Signed-off-by: messiasthi <8562fc1efba9a3c99753c749fdfb1b6932b70fbf@gmail.com><commit_after>
# The client of DDuplicated tool. from os import path as opath, getcwd from pprint import pprint from sys import argv from dduplicated import commands def get_paths(params): paths = [] for param in params: path = opath.join(getcwd(), param) if opath.exists(path) and opath.isdir(path) and not opath.islink(path): paths.append(path) return paths def main(): params = argv processed_files = [] # Remove the command name del params[0] if len(params) == 0 or "help" in params: commands.help() elif "detect" in params: processed_files = commands.detect(get_paths(params)) elif "delete" in params: processed_files = commands.delete(commands.detect(get_paths(params))) elif "link" in params: processed_files = commands.link(commands.detect(get_paths(params))) else: commands.help() if len(processed_files) > 0: pprint(processed_files) else: print("No duplicates found") print("Great! Bye!") exit(0)
# The client of DDuplicated tool. from os import path as opath, getcwd from sys import argv from dduplicated import commands def get_paths(params): paths = [] for param in params: path = opath.join(getcwd(), param) if opath.exists(path) and opath.isdir(path) and not opath.islink(path): paths.append(path) return paths def main(): params = argv processed_files = [] # Remove the command name del params[0] if len(params) == 0 or "help" in params: commands.help() elif "detect" in params: processed_files = commands.detect(get_paths(params)) elif "delete" in params: processed_files = commands.delete(commands.detect(get_paths(params))) elif "link" in params: processed_files = commands.link(commands.detect(get_paths(params))) else: commands.help() if len(processed_files) > 0: print(processed_files) else: print("No duplicates found") print("Great! Bye!") exit(0) Update in output to terminal. Signed-off-by: messiasthi <8562fc1efba9a3c99753c749fdfb1b6932b70fbf@gmail.com># The client of DDuplicated tool. from os import path as opath, getcwd from pprint import pprint from sys import argv from dduplicated import commands def get_paths(params): paths = [] for param in params: path = opath.join(getcwd(), param) if opath.exists(path) and opath.isdir(path) and not opath.islink(path): paths.append(path) return paths def main(): params = argv processed_files = [] # Remove the command name del params[0] if len(params) == 0 or "help" in params: commands.help() elif "detect" in params: processed_files = commands.detect(get_paths(params)) elif "delete" in params: processed_files = commands.delete(commands.detect(get_paths(params))) elif "link" in params: processed_files = commands.link(commands.detect(get_paths(params))) else: commands.help() if len(processed_files) > 0: pprint(processed_files) else: print("No duplicates found") print("Great! Bye!") exit(0)
<commit_before># The client of DDuplicated tool. from os import path as opath, getcwd from sys import argv from dduplicated import commands def get_paths(params): paths = [] for param in params: path = opath.join(getcwd(), param) if opath.exists(path) and opath.isdir(path) and not opath.islink(path): paths.append(path) return paths def main(): params = argv processed_files = [] # Remove the command name del params[0] if len(params) == 0 or "help" in params: commands.help() elif "detect" in params: processed_files = commands.detect(get_paths(params)) elif "delete" in params: processed_files = commands.delete(commands.detect(get_paths(params))) elif "link" in params: processed_files = commands.link(commands.detect(get_paths(params))) else: commands.help() if len(processed_files) > 0: print(processed_files) else: print("No duplicates found") print("Great! Bye!") exit(0) <commit_msg>Update in output to terminal. Signed-off-by: messiasthi <8562fc1efba9a3c99753c749fdfb1b6932b70fbf@gmail.com><commit_after># The client of DDuplicated tool. from os import path as opath, getcwd from pprint import pprint from sys import argv from dduplicated import commands def get_paths(params): paths = [] for param in params: path = opath.join(getcwd(), param) if opath.exists(path) and opath.isdir(path) and not opath.islink(path): paths.append(path) return paths def main(): params = argv processed_files = [] # Remove the command name del params[0] if len(params) == 0 or "help" in params: commands.help() elif "detect" in params: processed_files = commands.detect(get_paths(params)) elif "delete" in params: processed_files = commands.delete(commands.detect(get_paths(params))) elif "link" in params: processed_files = commands.link(commands.detect(get_paths(params))) else: commands.help() if len(processed_files) > 0: pprint(processed_files) else: print("No duplicates found") print("Great! Bye!") exit(0)
b1efc997c510fbdcaeb8d3ba9b4202ac810bb9ff
util/html_clean.py
util/html_clean.py
import bleach DESCR_ALLOWED_TAGS = bleach.ALLOWED_TAGS + ['h1', 'h2', 'h3', 'h4', 'h5', 'h6'] USER_DESCR_ALLOWED_TAGS = bleach.ALLOWED_TAGS + ['h2', 'h3', 'h4', 'h5', 'h6'] def clean_for_user_description(html): """ Removes dangerous tags, including h1. """ return bleach.clean(html, tags=USER_DESCR_ALLOWED_TAGS, strip=True) def clean_for_description(html): """ Removes dangerous tags. """ return bleach.clean(html, tags=DESCR_ALLOWED_TAGS, strip=True) def clean_all(html): """ Removes *all* html tags. """ return bleach.clean(html, tags=[], styles=[], attributes=[], strip=True)
import bleach DESCR_ALLOWED_TAGS = bleach.ALLOWED_TAGS + ['h1', 'h2', 'h3', 'h4', 'h5', 'h6', 'br', 'p'] USER_DESCR_ALLOWED_TAGS = bleach.ALLOWED_TAGS + ['h2', 'h3', 'h4', 'h5', 'h6', 'br', 'p'] def clean_for_user_description(html): """ Removes dangerous tags, including h1. """ return bleach.clean(html, tags=USER_DESCR_ALLOWED_TAGS, strip=True) def clean_for_description(html): """ Removes dangerous tags. """ return bleach.clean(html, tags=DESCR_ALLOWED_TAGS, strip=True) def clean_all(html): """ Removes *all* html tags. """ return bleach.clean(html, tags=[], styles=[], attributes=[], strip=True)
Allow <br /> and <p> in descriptions
Allow <br /> and <p> in descriptions
Python
bsd-3-clause
fsr/course-management,fsr/course-management
import bleach DESCR_ALLOWED_TAGS = bleach.ALLOWED_TAGS + ['h1', 'h2', 'h3', 'h4', 'h5', 'h6'] USER_DESCR_ALLOWED_TAGS = bleach.ALLOWED_TAGS + ['h2', 'h3', 'h4', 'h5', 'h6'] def clean_for_user_description(html): """ Removes dangerous tags, including h1. """ return bleach.clean(html, tags=USER_DESCR_ALLOWED_TAGS, strip=True) def clean_for_description(html): """ Removes dangerous tags. """ return bleach.clean(html, tags=DESCR_ALLOWED_TAGS, strip=True) def clean_all(html): """ Removes *all* html tags. """ return bleach.clean(html, tags=[], styles=[], attributes=[], strip=True) Allow <br /> and <p> in descriptions
import bleach DESCR_ALLOWED_TAGS = bleach.ALLOWED_TAGS + ['h1', 'h2', 'h3', 'h4', 'h5', 'h6', 'br', 'p'] USER_DESCR_ALLOWED_TAGS = bleach.ALLOWED_TAGS + ['h2', 'h3', 'h4', 'h5', 'h6', 'br', 'p'] def clean_for_user_description(html): """ Removes dangerous tags, including h1. """ return bleach.clean(html, tags=USER_DESCR_ALLOWED_TAGS, strip=True) def clean_for_description(html): """ Removes dangerous tags. """ return bleach.clean(html, tags=DESCR_ALLOWED_TAGS, strip=True) def clean_all(html): """ Removes *all* html tags. """ return bleach.clean(html, tags=[], styles=[], attributes=[], strip=True)
<commit_before>import bleach DESCR_ALLOWED_TAGS = bleach.ALLOWED_TAGS + ['h1', 'h2', 'h3', 'h4', 'h5', 'h6'] USER_DESCR_ALLOWED_TAGS = bleach.ALLOWED_TAGS + ['h2', 'h3', 'h4', 'h5', 'h6'] def clean_for_user_description(html): """ Removes dangerous tags, including h1. """ return bleach.clean(html, tags=USER_DESCR_ALLOWED_TAGS, strip=True) def clean_for_description(html): """ Removes dangerous tags. """ return bleach.clean(html, tags=DESCR_ALLOWED_TAGS, strip=True) def clean_all(html): """ Removes *all* html tags. """ return bleach.clean(html, tags=[], styles=[], attributes=[], strip=True) <commit_msg>Allow <br /> and <p> in descriptions<commit_after>
import bleach DESCR_ALLOWED_TAGS = bleach.ALLOWED_TAGS + ['h1', 'h2', 'h3', 'h4', 'h5', 'h6', 'br', 'p'] USER_DESCR_ALLOWED_TAGS = bleach.ALLOWED_TAGS + ['h2', 'h3', 'h4', 'h5', 'h6', 'br', 'p'] def clean_for_user_description(html): """ Removes dangerous tags, including h1. """ return bleach.clean(html, tags=USER_DESCR_ALLOWED_TAGS, strip=True) def clean_for_description(html): """ Removes dangerous tags. """ return bleach.clean(html, tags=DESCR_ALLOWED_TAGS, strip=True) def clean_all(html): """ Removes *all* html tags. """ return bleach.clean(html, tags=[], styles=[], attributes=[], strip=True)
import bleach DESCR_ALLOWED_TAGS = bleach.ALLOWED_TAGS + ['h1', 'h2', 'h3', 'h4', 'h5', 'h6'] USER_DESCR_ALLOWED_TAGS = bleach.ALLOWED_TAGS + ['h2', 'h3', 'h4', 'h5', 'h6'] def clean_for_user_description(html): """ Removes dangerous tags, including h1. """ return bleach.clean(html, tags=USER_DESCR_ALLOWED_TAGS, strip=True) def clean_for_description(html): """ Removes dangerous tags. """ return bleach.clean(html, tags=DESCR_ALLOWED_TAGS, strip=True) def clean_all(html): """ Removes *all* html tags. """ return bleach.clean(html, tags=[], styles=[], attributes=[], strip=True) Allow <br /> and <p> in descriptionsimport bleach DESCR_ALLOWED_TAGS = bleach.ALLOWED_TAGS + ['h1', 'h2', 'h3', 'h4', 'h5', 'h6', 'br', 'p'] USER_DESCR_ALLOWED_TAGS = bleach.ALLOWED_TAGS + ['h2', 'h3', 'h4', 'h5', 'h6', 'br', 'p'] def clean_for_user_description(html): """ Removes dangerous tags, including h1. """ return bleach.clean(html, tags=USER_DESCR_ALLOWED_TAGS, strip=True) def clean_for_description(html): """ Removes dangerous tags. """ return bleach.clean(html, tags=DESCR_ALLOWED_TAGS, strip=True) def clean_all(html): """ Removes *all* html tags. """ return bleach.clean(html, tags=[], styles=[], attributes=[], strip=True)
<commit_before>import bleach DESCR_ALLOWED_TAGS = bleach.ALLOWED_TAGS + ['h1', 'h2', 'h3', 'h4', 'h5', 'h6'] USER_DESCR_ALLOWED_TAGS = bleach.ALLOWED_TAGS + ['h2', 'h3', 'h4', 'h5', 'h6'] def clean_for_user_description(html): """ Removes dangerous tags, including h1. """ return bleach.clean(html, tags=USER_DESCR_ALLOWED_TAGS, strip=True) def clean_for_description(html): """ Removes dangerous tags. """ return bleach.clean(html, tags=DESCR_ALLOWED_TAGS, strip=True) def clean_all(html): """ Removes *all* html tags. """ return bleach.clean(html, tags=[], styles=[], attributes=[], strip=True) <commit_msg>Allow <br /> and <p> in descriptions<commit_after>import bleach DESCR_ALLOWED_TAGS = bleach.ALLOWED_TAGS + ['h1', 'h2', 'h3', 'h4', 'h5', 'h6', 'br', 'p'] USER_DESCR_ALLOWED_TAGS = bleach.ALLOWED_TAGS + ['h2', 'h3', 'h4', 'h5', 'h6', 'br', 'p'] def clean_for_user_description(html): """ Removes dangerous tags, including h1. """ return bleach.clean(html, tags=USER_DESCR_ALLOWED_TAGS, strip=True) def clean_for_description(html): """ Removes dangerous tags. """ return bleach.clean(html, tags=DESCR_ALLOWED_TAGS, strip=True) def clean_all(html): """ Removes *all* html tags. """ return bleach.clean(html, tags=[], styles=[], attributes=[], strip=True)
4673b980252f5fd4c490652d9af63eac02506614
dimod/reference/composites/structure.py
dimod/reference/composites/structure.py
from dimod.core.sampler import Sampler from dimod.core.composite import Composite from dimod.core.structured import Structured from dimod.decorators import bqm_structured class StructureComposite(Sampler, Composite, Structured): """Creates a structured composed sampler from an unstructured sampler. todo """ def __init__(self, sampler, nodelist, edgelist): Sampler.__init__(self) Composite.__init__(self, sampler) Structured.__init__(self, nodelist, edgelist) @bqm_structured def sample(self, bqm, **sample_kwargs): return self.child.sample(bqm, **sample_kwargs)
from dimod.core.sampler import Sampler from dimod.core.composite import Composite from dimod.core.structured import Structured from dimod.decorators import bqm_structured class StructureComposite(Sampler, Composite, Structured): """Creates a structured composed sampler from an unstructured sampler. """ # we will override these in the __init__, but because they are abstract properties we need to # signal that we are overriding them edgelist = None nodelist = None children = None def __init__(self, sampler, nodelist, edgelist): self.children = [sampler] self.nodelist = nodelist self.edgelist = edgelist @property def parameters(self): return self.child.parameters @property def properties(self): return self.child.properties @bqm_structured def sample(self, bqm, **sample_kwargs): return self.child.sample(bqm, **sample_kwargs)
Update Structure composite to use the new abc
Update Structure composite to use the new abc
Python
apache-2.0
dwavesystems/dimod,dwavesystems/dimod
from dimod.core.sampler import Sampler from dimod.core.composite import Composite from dimod.core.structured import Structured from dimod.decorators import bqm_structured class StructureComposite(Sampler, Composite, Structured): """Creates a structured composed sampler from an unstructured sampler. todo """ def __init__(self, sampler, nodelist, edgelist): Sampler.__init__(self) Composite.__init__(self, sampler) Structured.__init__(self, nodelist, edgelist) @bqm_structured def sample(self, bqm, **sample_kwargs): return self.child.sample(bqm, **sample_kwargs) Update Structure composite to use the new abc
from dimod.core.sampler import Sampler from dimod.core.composite import Composite from dimod.core.structured import Structured from dimod.decorators import bqm_structured class StructureComposite(Sampler, Composite, Structured): """Creates a structured composed sampler from an unstructured sampler. """ # we will override these in the __init__, but because they are abstract properties we need to # signal that we are overriding them edgelist = None nodelist = None children = None def __init__(self, sampler, nodelist, edgelist): self.children = [sampler] self.nodelist = nodelist self.edgelist = edgelist @property def parameters(self): return self.child.parameters @property def properties(self): return self.child.properties @bqm_structured def sample(self, bqm, **sample_kwargs): return self.child.sample(bqm, **sample_kwargs)
<commit_before>from dimod.core.sampler import Sampler from dimod.core.composite import Composite from dimod.core.structured import Structured from dimod.decorators import bqm_structured class StructureComposite(Sampler, Composite, Structured): """Creates a structured composed sampler from an unstructured sampler. todo """ def __init__(self, sampler, nodelist, edgelist): Sampler.__init__(self) Composite.__init__(self, sampler) Structured.__init__(self, nodelist, edgelist) @bqm_structured def sample(self, bqm, **sample_kwargs): return self.child.sample(bqm, **sample_kwargs) <commit_msg>Update Structure composite to use the new abc<commit_after>
from dimod.core.sampler import Sampler from dimod.core.composite import Composite from dimod.core.structured import Structured from dimod.decorators import bqm_structured class StructureComposite(Sampler, Composite, Structured): """Creates a structured composed sampler from an unstructured sampler. """ # we will override these in the __init__, but because they are abstract properties we need to # signal that we are overriding them edgelist = None nodelist = None children = None def __init__(self, sampler, nodelist, edgelist): self.children = [sampler] self.nodelist = nodelist self.edgelist = edgelist @property def parameters(self): return self.child.parameters @property def properties(self): return self.child.properties @bqm_structured def sample(self, bqm, **sample_kwargs): return self.child.sample(bqm, **sample_kwargs)
from dimod.core.sampler import Sampler from dimod.core.composite import Composite from dimod.core.structured import Structured from dimod.decorators import bqm_structured class StructureComposite(Sampler, Composite, Structured): """Creates a structured composed sampler from an unstructured sampler. todo """ def __init__(self, sampler, nodelist, edgelist): Sampler.__init__(self) Composite.__init__(self, sampler) Structured.__init__(self, nodelist, edgelist) @bqm_structured def sample(self, bqm, **sample_kwargs): return self.child.sample(bqm, **sample_kwargs) Update Structure composite to use the new abcfrom dimod.core.sampler import Sampler from dimod.core.composite import Composite from dimod.core.structured import Structured from dimod.decorators import bqm_structured class StructureComposite(Sampler, Composite, Structured): """Creates a structured composed sampler from an unstructured sampler. """ # we will override these in the __init__, but because they are abstract properties we need to # signal that we are overriding them edgelist = None nodelist = None children = None def __init__(self, sampler, nodelist, edgelist): self.children = [sampler] self.nodelist = nodelist self.edgelist = edgelist @property def parameters(self): return self.child.parameters @property def properties(self): return self.child.properties @bqm_structured def sample(self, bqm, **sample_kwargs): return self.child.sample(bqm, **sample_kwargs)
<commit_before>from dimod.core.sampler import Sampler from dimod.core.composite import Composite from dimod.core.structured import Structured from dimod.decorators import bqm_structured class StructureComposite(Sampler, Composite, Structured): """Creates a structured composed sampler from an unstructured sampler. todo """ def __init__(self, sampler, nodelist, edgelist): Sampler.__init__(self) Composite.__init__(self, sampler) Structured.__init__(self, nodelist, edgelist) @bqm_structured def sample(self, bqm, **sample_kwargs): return self.child.sample(bqm, **sample_kwargs) <commit_msg>Update Structure composite to use the new abc<commit_after>from dimod.core.sampler import Sampler from dimod.core.composite import Composite from dimod.core.structured import Structured from dimod.decorators import bqm_structured class StructureComposite(Sampler, Composite, Structured): """Creates a structured composed sampler from an unstructured sampler. """ # we will override these in the __init__, but because they are abstract properties we need to # signal that we are overriding them edgelist = None nodelist = None children = None def __init__(self, sampler, nodelist, edgelist): self.children = [sampler] self.nodelist = nodelist self.edgelist = edgelist @property def parameters(self): return self.child.parameters @property def properties(self): return self.child.properties @bqm_structured def sample(self, bqm, **sample_kwargs): return self.child.sample(bqm, **sample_kwargs)
6d38920c1867921235c002b6ad411fd08378dd1f
fluent_contents/tests/test_models.py
fluent_contents/tests/test_models.py
from django.contrib.contenttypes.models import ContentType from fluent_contents.models import ContentItem from fluent_contents.tests.utils import AppTestCase class ModelTests(AppTestCase): """ Testing the data model. """ def test_stale_model_str(self): """ No matter what, the ContentItem.__str__() should work. This would break the admin delete screen otherwise. """ c = ContentType() a = ContentItem(polymorphic_ctype=c) self.assertEqual(str(a), "'(type deleted) 0' in 'None None'")
from django.contrib.contenttypes.models import ContentType from fluent_contents.models import ContentItem from fluent_contents.tests.utils import AppTestCase class ModelTests(AppTestCase): """ Testing the data model. """ def test_stale_model_str(self): """ No matter what, the ContentItem.__str__() should work. This would break the admin delete screen otherwise. """ c = ContentType() c.save() a = ContentItem(polymorphic_ctype=c) self.assertEqual(str(a), "'(type deleted) 0' in 'None None'")
Fix Django 1.8+ tests for stale content type
Fix Django 1.8+ tests for stale content type
Python
apache-2.0
django-fluent/django-fluent-contents,edoburu/django-fluent-contents,edoburu/django-fluent-contents,django-fluent/django-fluent-contents,django-fluent/django-fluent-contents,edoburu/django-fluent-contents
from django.contrib.contenttypes.models import ContentType from fluent_contents.models import ContentItem from fluent_contents.tests.utils import AppTestCase class ModelTests(AppTestCase): """ Testing the data model. """ def test_stale_model_str(self): """ No matter what, the ContentItem.__str__() should work. This would break the admin delete screen otherwise. """ c = ContentType() a = ContentItem(polymorphic_ctype=c) self.assertEqual(str(a), "'(type deleted) 0' in 'None None'") Fix Django 1.8+ tests for stale content type
from django.contrib.contenttypes.models import ContentType from fluent_contents.models import ContentItem from fluent_contents.tests.utils import AppTestCase class ModelTests(AppTestCase): """ Testing the data model. """ def test_stale_model_str(self): """ No matter what, the ContentItem.__str__() should work. This would break the admin delete screen otherwise. """ c = ContentType() c.save() a = ContentItem(polymorphic_ctype=c) self.assertEqual(str(a), "'(type deleted) 0' in 'None None'")
<commit_before>from django.contrib.contenttypes.models import ContentType from fluent_contents.models import ContentItem from fluent_contents.tests.utils import AppTestCase class ModelTests(AppTestCase): """ Testing the data model. """ def test_stale_model_str(self): """ No matter what, the ContentItem.__str__() should work. This would break the admin delete screen otherwise. """ c = ContentType() a = ContentItem(polymorphic_ctype=c) self.assertEqual(str(a), "'(type deleted) 0' in 'None None'") <commit_msg>Fix Django 1.8+ tests for stale content type<commit_after>
from django.contrib.contenttypes.models import ContentType from fluent_contents.models import ContentItem from fluent_contents.tests.utils import AppTestCase class ModelTests(AppTestCase): """ Testing the data model. """ def test_stale_model_str(self): """ No matter what, the ContentItem.__str__() should work. This would break the admin delete screen otherwise. """ c = ContentType() c.save() a = ContentItem(polymorphic_ctype=c) self.assertEqual(str(a), "'(type deleted) 0' in 'None None'")
from django.contrib.contenttypes.models import ContentType from fluent_contents.models import ContentItem from fluent_contents.tests.utils import AppTestCase class ModelTests(AppTestCase): """ Testing the data model. """ def test_stale_model_str(self): """ No matter what, the ContentItem.__str__() should work. This would break the admin delete screen otherwise. """ c = ContentType() a = ContentItem(polymorphic_ctype=c) self.assertEqual(str(a), "'(type deleted) 0' in 'None None'") Fix Django 1.8+ tests for stale content typefrom django.contrib.contenttypes.models import ContentType from fluent_contents.models import ContentItem from fluent_contents.tests.utils import AppTestCase class ModelTests(AppTestCase): """ Testing the data model. """ def test_stale_model_str(self): """ No matter what, the ContentItem.__str__() should work. This would break the admin delete screen otherwise. """ c = ContentType() c.save() a = ContentItem(polymorphic_ctype=c) self.assertEqual(str(a), "'(type deleted) 0' in 'None None'")
<commit_before>from django.contrib.contenttypes.models import ContentType from fluent_contents.models import ContentItem from fluent_contents.tests.utils import AppTestCase class ModelTests(AppTestCase): """ Testing the data model. """ def test_stale_model_str(self): """ No matter what, the ContentItem.__str__() should work. This would break the admin delete screen otherwise. """ c = ContentType() a = ContentItem(polymorphic_ctype=c) self.assertEqual(str(a), "'(type deleted) 0' in 'None None'") <commit_msg>Fix Django 1.8+ tests for stale content type<commit_after>from django.contrib.contenttypes.models import ContentType from fluent_contents.models import ContentItem from fluent_contents.tests.utils import AppTestCase class ModelTests(AppTestCase): """ Testing the data model. """ def test_stale_model_str(self): """ No matter what, the ContentItem.__str__() should work. This would break the admin delete screen otherwise. """ c = ContentType() c.save() a = ContentItem(polymorphic_ctype=c) self.assertEqual(str(a), "'(type deleted) 0' in 'None None'")
a16a822ede8ad987b2234f29a31f9fe79c27cbd5
dbaas/workflow/steps/util/clone/clone_database.py
dbaas/workflow/steps/util/clone/clone_database.py
# -*- coding: utf-8 -*- import logging from util import full_stack from util import call_script from django.conf import settings from drivers import factory_for from system.models import Configuration from notification.util import get_clone_args from ...base import BaseStep from ....exceptions.error_codes import DBAAS_0017 LOG = logging.getLogger(__name__) class CloneDatabase(BaseStep): def __unicode__(self): return "Replicating database data..." def do(self, workflow_dict): try: if 'databaseinfra' not in workflow_dict \ or 'clone' not in workflow_dict : return False args = get_clone_args(workflow_dict['clone'], workflow_dict['database']) script_name = factory_for(workflow_dict['clone'].databaseinfra).clone() python_bin= Configuration.get_by_name('python_venv_bin') return_code, output = call_script(script_name, working_dir=settings.SCRIPTS_PATH , args=args, split_lines=False, python_bin=python_bin) LOG.info("Script Output: {}".format(output)) LOG.info("Return code: {}".format(return_code)) if return_code != 0: return False return True except Exception: traceback = full_stack() workflow_dict['exceptions']['error_codes'].append(DBAAS_0017) workflow_dict['exceptions']['traceback'].append(traceback) return False def undo(self, workflow_dict): LOG.info("Nothing to do here...") return True
# -*- coding: utf-8 -*- import logging from util import full_stack from util import call_script from django.conf import settings from drivers import factory_for from system.models import Configuration from notification.util import get_clone_args from ...base import BaseStep from ....exceptions.error_codes import DBAAS_0017 LOG = logging.getLogger(__name__) class CloneDatabase(BaseStep): def __unicode__(self): return "Replicating database data..." def do(self, workflow_dict): try: if 'databaseinfra' not in workflow_dict \ or 'clone' not in workflow_dict : return False args = get_clone_args(workflow_dict['clone'], workflow_dict['database']) script_name = factory_for(workflow_dict['clone'].databaseinfra).clone() python_bin= Configuration.get_by_name('python_venv_bin') return_code, output = call_script(script_name, working_dir=settings.SCRIPTS_PATH , args=args, split_lines=False, python_bin=python_bin) LOG.info("Script Output: {}".format(output)) LOG.info("Return code: {}".format(return_code)) if return_code != 0: workflow_dict['exceptions']['traceback'].append(output) return False return True except Exception: traceback = full_stack() workflow_dict['exceptions']['error_codes'].append(DBAAS_0017) workflow_dict['exceptions']['traceback'].append(traceback) return False def undo(self, workflow_dict): LOG.info("Nothing to do here...") return True
Add call_script output to workflow_dict traceback
Add call_script output to workflow_dict traceback
Python
bsd-3-clause
globocom/database-as-a-service,globocom/database-as-a-service,globocom/database-as-a-service,globocom/database-as-a-service
# -*- coding: utf-8 -*- import logging from util import full_stack from util import call_script from django.conf import settings from drivers import factory_for from system.models import Configuration from notification.util import get_clone_args from ...base import BaseStep from ....exceptions.error_codes import DBAAS_0017 LOG = logging.getLogger(__name__) class CloneDatabase(BaseStep): def __unicode__(self): return "Replicating database data..." def do(self, workflow_dict): try: if 'databaseinfra' not in workflow_dict \ or 'clone' not in workflow_dict : return False args = get_clone_args(workflow_dict['clone'], workflow_dict['database']) script_name = factory_for(workflow_dict['clone'].databaseinfra).clone() python_bin= Configuration.get_by_name('python_venv_bin') return_code, output = call_script(script_name, working_dir=settings.SCRIPTS_PATH , args=args, split_lines=False, python_bin=python_bin) LOG.info("Script Output: {}".format(output)) LOG.info("Return code: {}".format(return_code)) if return_code != 0: return False return True except Exception: traceback = full_stack() workflow_dict['exceptions']['error_codes'].append(DBAAS_0017) workflow_dict['exceptions']['traceback'].append(traceback) return False def undo(self, workflow_dict): LOG.info("Nothing to do here...") return True Add call_script output to workflow_dict traceback
# -*- coding: utf-8 -*- import logging from util import full_stack from util import call_script from django.conf import settings from drivers import factory_for from system.models import Configuration from notification.util import get_clone_args from ...base import BaseStep from ....exceptions.error_codes import DBAAS_0017 LOG = logging.getLogger(__name__) class CloneDatabase(BaseStep): def __unicode__(self): return "Replicating database data..." def do(self, workflow_dict): try: if 'databaseinfra' not in workflow_dict \ or 'clone' not in workflow_dict : return False args = get_clone_args(workflow_dict['clone'], workflow_dict['database']) script_name = factory_for(workflow_dict['clone'].databaseinfra).clone() python_bin= Configuration.get_by_name('python_venv_bin') return_code, output = call_script(script_name, working_dir=settings.SCRIPTS_PATH , args=args, split_lines=False, python_bin=python_bin) LOG.info("Script Output: {}".format(output)) LOG.info("Return code: {}".format(return_code)) if return_code != 0: workflow_dict['exceptions']['traceback'].append(output) return False return True except Exception: traceback = full_stack() workflow_dict['exceptions']['error_codes'].append(DBAAS_0017) workflow_dict['exceptions']['traceback'].append(traceback) return False def undo(self, workflow_dict): LOG.info("Nothing to do here...") return True
<commit_before># -*- coding: utf-8 -*- import logging from util import full_stack from util import call_script from django.conf import settings from drivers import factory_for from system.models import Configuration from notification.util import get_clone_args from ...base import BaseStep from ....exceptions.error_codes import DBAAS_0017 LOG = logging.getLogger(__name__) class CloneDatabase(BaseStep): def __unicode__(self): return "Replicating database data..." def do(self, workflow_dict): try: if 'databaseinfra' not in workflow_dict \ or 'clone' not in workflow_dict : return False args = get_clone_args(workflow_dict['clone'], workflow_dict['database']) script_name = factory_for(workflow_dict['clone'].databaseinfra).clone() python_bin= Configuration.get_by_name('python_venv_bin') return_code, output = call_script(script_name, working_dir=settings.SCRIPTS_PATH , args=args, split_lines=False, python_bin=python_bin) LOG.info("Script Output: {}".format(output)) LOG.info("Return code: {}".format(return_code)) if return_code != 0: return False return True except Exception: traceback = full_stack() workflow_dict['exceptions']['error_codes'].append(DBAAS_0017) workflow_dict['exceptions']['traceback'].append(traceback) return False def undo(self, workflow_dict): LOG.info("Nothing to do here...") return True <commit_msg>Add call_script output to workflow_dict traceback<commit_after>
# -*- coding: utf-8 -*- import logging from util import full_stack from util import call_script from django.conf import settings from drivers import factory_for from system.models import Configuration from notification.util import get_clone_args from ...base import BaseStep from ....exceptions.error_codes import DBAAS_0017 LOG = logging.getLogger(__name__) class CloneDatabase(BaseStep): def __unicode__(self): return "Replicating database data..." def do(self, workflow_dict): try: if 'databaseinfra' not in workflow_dict \ or 'clone' not in workflow_dict : return False args = get_clone_args(workflow_dict['clone'], workflow_dict['database']) script_name = factory_for(workflow_dict['clone'].databaseinfra).clone() python_bin= Configuration.get_by_name('python_venv_bin') return_code, output = call_script(script_name, working_dir=settings.SCRIPTS_PATH , args=args, split_lines=False, python_bin=python_bin) LOG.info("Script Output: {}".format(output)) LOG.info("Return code: {}".format(return_code)) if return_code != 0: workflow_dict['exceptions']['traceback'].append(output) return False return True except Exception: traceback = full_stack() workflow_dict['exceptions']['error_codes'].append(DBAAS_0017) workflow_dict['exceptions']['traceback'].append(traceback) return False def undo(self, workflow_dict): LOG.info("Nothing to do here...") return True
# -*- coding: utf-8 -*- import logging from util import full_stack from util import call_script from django.conf import settings from drivers import factory_for from system.models import Configuration from notification.util import get_clone_args from ...base import BaseStep from ....exceptions.error_codes import DBAAS_0017 LOG = logging.getLogger(__name__) class CloneDatabase(BaseStep): def __unicode__(self): return "Replicating database data..." def do(self, workflow_dict): try: if 'databaseinfra' not in workflow_dict \ or 'clone' not in workflow_dict : return False args = get_clone_args(workflow_dict['clone'], workflow_dict['database']) script_name = factory_for(workflow_dict['clone'].databaseinfra).clone() python_bin= Configuration.get_by_name('python_venv_bin') return_code, output = call_script(script_name, working_dir=settings.SCRIPTS_PATH , args=args, split_lines=False, python_bin=python_bin) LOG.info("Script Output: {}".format(output)) LOG.info("Return code: {}".format(return_code)) if return_code != 0: return False return True except Exception: traceback = full_stack() workflow_dict['exceptions']['error_codes'].append(DBAAS_0017) workflow_dict['exceptions']['traceback'].append(traceback) return False def undo(self, workflow_dict): LOG.info("Nothing to do here...") return True Add call_script output to workflow_dict traceback# -*- coding: utf-8 -*- import logging from util import full_stack from util import call_script from django.conf import settings from drivers import factory_for from system.models import Configuration from notification.util import get_clone_args from ...base import BaseStep from ....exceptions.error_codes import DBAAS_0017 LOG = logging.getLogger(__name__) class CloneDatabase(BaseStep): def __unicode__(self): return "Replicating database data..." def do(self, workflow_dict): try: if 'databaseinfra' not in workflow_dict \ or 'clone' not in workflow_dict : return False args = get_clone_args(workflow_dict['clone'], workflow_dict['database']) script_name = factory_for(workflow_dict['clone'].databaseinfra).clone() python_bin= Configuration.get_by_name('python_venv_bin') return_code, output = call_script(script_name, working_dir=settings.SCRIPTS_PATH , args=args, split_lines=False, python_bin=python_bin) LOG.info("Script Output: {}".format(output)) LOG.info("Return code: {}".format(return_code)) if return_code != 0: workflow_dict['exceptions']['traceback'].append(output) return False return True except Exception: traceback = full_stack() workflow_dict['exceptions']['error_codes'].append(DBAAS_0017) workflow_dict['exceptions']['traceback'].append(traceback) return False def undo(self, workflow_dict): LOG.info("Nothing to do here...") return True
<commit_before># -*- coding: utf-8 -*- import logging from util import full_stack from util import call_script from django.conf import settings from drivers import factory_for from system.models import Configuration from notification.util import get_clone_args from ...base import BaseStep from ....exceptions.error_codes import DBAAS_0017 LOG = logging.getLogger(__name__) class CloneDatabase(BaseStep): def __unicode__(self): return "Replicating database data..." def do(self, workflow_dict): try: if 'databaseinfra' not in workflow_dict \ or 'clone' not in workflow_dict : return False args = get_clone_args(workflow_dict['clone'], workflow_dict['database']) script_name = factory_for(workflow_dict['clone'].databaseinfra).clone() python_bin= Configuration.get_by_name('python_venv_bin') return_code, output = call_script(script_name, working_dir=settings.SCRIPTS_PATH , args=args, split_lines=False, python_bin=python_bin) LOG.info("Script Output: {}".format(output)) LOG.info("Return code: {}".format(return_code)) if return_code != 0: return False return True except Exception: traceback = full_stack() workflow_dict['exceptions']['error_codes'].append(DBAAS_0017) workflow_dict['exceptions']['traceback'].append(traceback) return False def undo(self, workflow_dict): LOG.info("Nothing to do here...") return True <commit_msg>Add call_script output to workflow_dict traceback<commit_after># -*- coding: utf-8 -*- import logging from util import full_stack from util import call_script from django.conf import settings from drivers import factory_for from system.models import Configuration from notification.util import get_clone_args from ...base import BaseStep from ....exceptions.error_codes import DBAAS_0017 LOG = logging.getLogger(__name__) class CloneDatabase(BaseStep): def __unicode__(self): return "Replicating database data..." def do(self, workflow_dict): try: if 'databaseinfra' not in workflow_dict \ or 'clone' not in workflow_dict : return False args = get_clone_args(workflow_dict['clone'], workflow_dict['database']) script_name = factory_for(workflow_dict['clone'].databaseinfra).clone() python_bin= Configuration.get_by_name('python_venv_bin') return_code, output = call_script(script_name, working_dir=settings.SCRIPTS_PATH , args=args, split_lines=False, python_bin=python_bin) LOG.info("Script Output: {}".format(output)) LOG.info("Return code: {}".format(return_code)) if return_code != 0: workflow_dict['exceptions']['traceback'].append(output) return False return True except Exception: traceback = full_stack() workflow_dict['exceptions']['error_codes'].append(DBAAS_0017) workflow_dict['exceptions']['traceback'].append(traceback) return False def undo(self, workflow_dict): LOG.info("Nothing to do here...") return True
870c89649032480587bdb03ae31f4eecf21eebf7
tldr/parser.py
tldr/parser.py
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import import click def parse_page(page): """Parse the command man page.""" with open(page) as f: lines = f.readlines() for line in lines: if line.startswith('#'): continue elif line.startswith('>'): click.secho(line.replace('>', ' '), fg='blue', nl=False) elif line.startswith('-'): click.secho(line, fg='green', nl=False) elif line.startswith('`'): click.secho(' ' + line.replace('`', ''), fg='cyan', nl=False) else: click.secho(line, nl=False)
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import import click from tldr.config import get_config def parse_page(page): """Parse the command man page.""" colors = get_config()['colors'] with open(page) as f: lines = f.readlines() for line in lines: if line.startswith('#'): continue elif line.startswith('>'): click.secho(line.replace('>', ' '), fg=colors['description'], nl=False) elif line.startswith('-'): click.secho(line, fg=colors['usage'], nl=False) elif line.startswith('`'): click.secho(' ' + line.replace('`', ''), fg=colors['command'], nl=False) else: click.secho(line, nl=False)
Use color configurations from config file
Use color configurations from config file
Python
mit
lord63/tldr.py
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import import click def parse_page(page): """Parse the command man page.""" with open(page) as f: lines = f.readlines() for line in lines: if line.startswith('#'): continue elif line.startswith('>'): click.secho(line.replace('>', ' '), fg='blue', nl=False) elif line.startswith('-'): click.secho(line, fg='green', nl=False) elif line.startswith('`'): click.secho(' ' + line.replace('`', ''), fg='cyan', nl=False) else: click.secho(line, nl=False) Use color configurations from config file
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import import click from tldr.config import get_config def parse_page(page): """Parse the command man page.""" colors = get_config()['colors'] with open(page) as f: lines = f.readlines() for line in lines: if line.startswith('#'): continue elif line.startswith('>'): click.secho(line.replace('>', ' '), fg=colors['description'], nl=False) elif line.startswith('-'): click.secho(line, fg=colors['usage'], nl=False) elif line.startswith('`'): click.secho(' ' + line.replace('`', ''), fg=colors['command'], nl=False) else: click.secho(line, nl=False)
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import import click def parse_page(page): """Parse the command man page.""" with open(page) as f: lines = f.readlines() for line in lines: if line.startswith('#'): continue elif line.startswith('>'): click.secho(line.replace('>', ' '), fg='blue', nl=False) elif line.startswith('-'): click.secho(line, fg='green', nl=False) elif line.startswith('`'): click.secho(' ' + line.replace('`', ''), fg='cyan', nl=False) else: click.secho(line, nl=False) <commit_msg>Use color configurations from config file<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import import click from tldr.config import get_config def parse_page(page): """Parse the command man page.""" colors = get_config()['colors'] with open(page) as f: lines = f.readlines() for line in lines: if line.startswith('#'): continue elif line.startswith('>'): click.secho(line.replace('>', ' '), fg=colors['description'], nl=False) elif line.startswith('-'): click.secho(line, fg=colors['usage'], nl=False) elif line.startswith('`'): click.secho(' ' + line.replace('`', ''), fg=colors['command'], nl=False) else: click.secho(line, nl=False)
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import import click def parse_page(page): """Parse the command man page.""" with open(page) as f: lines = f.readlines() for line in lines: if line.startswith('#'): continue elif line.startswith('>'): click.secho(line.replace('>', ' '), fg='blue', nl=False) elif line.startswith('-'): click.secho(line, fg='green', nl=False) elif line.startswith('`'): click.secho(' ' + line.replace('`', ''), fg='cyan', nl=False) else: click.secho(line, nl=False) Use color configurations from config file#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import import click from tldr.config import get_config def parse_page(page): """Parse the command man page.""" colors = get_config()['colors'] with open(page) as f: lines = f.readlines() for line in lines: if line.startswith('#'): continue elif line.startswith('>'): click.secho(line.replace('>', ' '), fg=colors['description'], nl=False) elif line.startswith('-'): click.secho(line, fg=colors['usage'], nl=False) elif line.startswith('`'): click.secho(' ' + line.replace('`', ''), fg=colors['command'], nl=False) else: click.secho(line, nl=False)
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import import click def parse_page(page): """Parse the command man page.""" with open(page) as f: lines = f.readlines() for line in lines: if line.startswith('#'): continue elif line.startswith('>'): click.secho(line.replace('>', ' '), fg='blue', nl=False) elif line.startswith('-'): click.secho(line, fg='green', nl=False) elif line.startswith('`'): click.secho(' ' + line.replace('`', ''), fg='cyan', nl=False) else: click.secho(line, nl=False) <commit_msg>Use color configurations from config file<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import import click from tldr.config import get_config def parse_page(page): """Parse the command man page.""" colors = get_config()['colors'] with open(page) as f: lines = f.readlines() for line in lines: if line.startswith('#'): continue elif line.startswith('>'): click.secho(line.replace('>', ' '), fg=colors['description'], nl=False) elif line.startswith('-'): click.secho(line, fg=colors['usage'], nl=False) elif line.startswith('`'): click.secho(' ' + line.replace('`', ''), fg=colors['command'], nl=False) else: click.secho(line, nl=False)
9710589f34e50bdf2fa19b1fcd827b07d4248ab6
dodo_commands/extra/dodo_standard_commands/commit-config.py
dodo_commands/extra/dodo_standard_commands/commit-config.py
from argparse import ArgumentParser from dodo_commands.framework import Dodo from dodo_commands.framework.config import Paths import os def _args(): parser = ArgumentParser() parser.add_argument('-m', dest='message') args = Dodo.parse_args(parser) return args if Dodo.is_main(__name__, safe=True): args = _args() if not os.path.exists(os.path.join(Paths().res_dir(), '.git')): Dodo.run(['git', 'init'], cwd=Paths().res_dir()) Dodo.run(['git', 'add', '-A'], cwd=Paths().res_dir()) Dodo.run( ['git', 'commit', '-m', args.message or 'Update configuration'], cwd=Paths().res_dir())
from argparse import ArgumentParser from dodo_commands.framework import Dodo from dodo_commands.framework.config import Paths import os def _args(): parser = ArgumentParser() parser.add_argument( '--message', '-m', dest='message', help='The commit message') args = Dodo.parse_args(parser) return args if Dodo.is_main(__name__, safe=True): args = _args() if not os.path.exists(os.path.join(Paths().res_dir(), '.git')): Dodo.run(['git', 'init'], cwd=Paths().res_dir()) Dodo.run(['git', 'add', '-A'], cwd=Paths().res_dir()) Dodo.run(['git', 'commit', '-m', args.message or 'Update configuration'], cwd=Paths().res_dir())
Allow both -m and --message
Allow both -m and --message
Python
mit
mnieber/dodo_commands
from argparse import ArgumentParser from dodo_commands.framework import Dodo from dodo_commands.framework.config import Paths import os def _args(): parser = ArgumentParser() parser.add_argument('-m', dest='message') args = Dodo.parse_args(parser) return args if Dodo.is_main(__name__, safe=True): args = _args() if not os.path.exists(os.path.join(Paths().res_dir(), '.git')): Dodo.run(['git', 'init'], cwd=Paths().res_dir()) Dodo.run(['git', 'add', '-A'], cwd=Paths().res_dir()) Dodo.run( ['git', 'commit', '-m', args.message or 'Update configuration'], cwd=Paths().res_dir()) Allow both -m and --message
from argparse import ArgumentParser from dodo_commands.framework import Dodo from dodo_commands.framework.config import Paths import os def _args(): parser = ArgumentParser() parser.add_argument( '--message', '-m', dest='message', help='The commit message') args = Dodo.parse_args(parser) return args if Dodo.is_main(__name__, safe=True): args = _args() if not os.path.exists(os.path.join(Paths().res_dir(), '.git')): Dodo.run(['git', 'init'], cwd=Paths().res_dir()) Dodo.run(['git', 'add', '-A'], cwd=Paths().res_dir()) Dodo.run(['git', 'commit', '-m', args.message or 'Update configuration'], cwd=Paths().res_dir())
<commit_before>from argparse import ArgumentParser from dodo_commands.framework import Dodo from dodo_commands.framework.config import Paths import os def _args(): parser = ArgumentParser() parser.add_argument('-m', dest='message') args = Dodo.parse_args(parser) return args if Dodo.is_main(__name__, safe=True): args = _args() if not os.path.exists(os.path.join(Paths().res_dir(), '.git')): Dodo.run(['git', 'init'], cwd=Paths().res_dir()) Dodo.run(['git', 'add', '-A'], cwd=Paths().res_dir()) Dodo.run( ['git', 'commit', '-m', args.message or 'Update configuration'], cwd=Paths().res_dir()) <commit_msg>Allow both -m and --message<commit_after>
from argparse import ArgumentParser from dodo_commands.framework import Dodo from dodo_commands.framework.config import Paths import os def _args(): parser = ArgumentParser() parser.add_argument( '--message', '-m', dest='message', help='The commit message') args = Dodo.parse_args(parser) return args if Dodo.is_main(__name__, safe=True): args = _args() if not os.path.exists(os.path.join(Paths().res_dir(), '.git')): Dodo.run(['git', 'init'], cwd=Paths().res_dir()) Dodo.run(['git', 'add', '-A'], cwd=Paths().res_dir()) Dodo.run(['git', 'commit', '-m', args.message or 'Update configuration'], cwd=Paths().res_dir())
from argparse import ArgumentParser from dodo_commands.framework import Dodo from dodo_commands.framework.config import Paths import os def _args(): parser = ArgumentParser() parser.add_argument('-m', dest='message') args = Dodo.parse_args(parser) return args if Dodo.is_main(__name__, safe=True): args = _args() if not os.path.exists(os.path.join(Paths().res_dir(), '.git')): Dodo.run(['git', 'init'], cwd=Paths().res_dir()) Dodo.run(['git', 'add', '-A'], cwd=Paths().res_dir()) Dodo.run( ['git', 'commit', '-m', args.message or 'Update configuration'], cwd=Paths().res_dir()) Allow both -m and --messagefrom argparse import ArgumentParser from dodo_commands.framework import Dodo from dodo_commands.framework.config import Paths import os def _args(): parser = ArgumentParser() parser.add_argument( '--message', '-m', dest='message', help='The commit message') args = Dodo.parse_args(parser) return args if Dodo.is_main(__name__, safe=True): args = _args() if not os.path.exists(os.path.join(Paths().res_dir(), '.git')): Dodo.run(['git', 'init'], cwd=Paths().res_dir()) Dodo.run(['git', 'add', '-A'], cwd=Paths().res_dir()) Dodo.run(['git', 'commit', '-m', args.message or 'Update configuration'], cwd=Paths().res_dir())
<commit_before>from argparse import ArgumentParser from dodo_commands.framework import Dodo from dodo_commands.framework.config import Paths import os def _args(): parser = ArgumentParser() parser.add_argument('-m', dest='message') args = Dodo.parse_args(parser) return args if Dodo.is_main(__name__, safe=True): args = _args() if not os.path.exists(os.path.join(Paths().res_dir(), '.git')): Dodo.run(['git', 'init'], cwd=Paths().res_dir()) Dodo.run(['git', 'add', '-A'], cwd=Paths().res_dir()) Dodo.run( ['git', 'commit', '-m', args.message or 'Update configuration'], cwd=Paths().res_dir()) <commit_msg>Allow both -m and --message<commit_after>from argparse import ArgumentParser from dodo_commands.framework import Dodo from dodo_commands.framework.config import Paths import os def _args(): parser = ArgumentParser() parser.add_argument( '--message', '-m', dest='message', help='The commit message') args = Dodo.parse_args(parser) return args if Dodo.is_main(__name__, safe=True): args = _args() if not os.path.exists(os.path.join(Paths().res_dir(), '.git')): Dodo.run(['git', 'init'], cwd=Paths().res_dir()) Dodo.run(['git', 'add', '-A'], cwd=Paths().res_dir()) Dodo.run(['git', 'commit', '-m', args.message or 'Update configuration'], cwd=Paths().res_dir())
2ea8ad7608f73fea2657e8e67a6c2bfef01b08d8
uchicagohvz/users/backend.py
uchicagohvz/users/backend.py
from django.contrib.auth import get_user_model import ldap User = get_user_model() class UChicagoLDAPBackend(object): LDAP_SERVER = "ldaps://ldap.uchicago.edu:636" def authenticate(self, cnetid=None, password=None): if cnetid and password: cnetid = ldap.filter.escape_filter_chars(cnetid) try: conn.simple_bind_s("uid=%s,ou=people,dc=uchicago,dc=edu" % cnetid, password) except ldap.INVALID_CREDENTIALS: return None try: user = User.objects.get(username=cnetid) except User.DoesNotExist: query = "(&(uid=%s)(objectclass=inetOrgPerson))" % (cnetid) result = conn.search_ext_s("dc=uchicago,dc=edu", ldap.SCOPE_SUBTREE, query) if result: user_data = result[0] user = User.objects.create_user(username=cnetid, email=user_data['email'], first_name=user_data['givenName'], last_name=user_data['sn']) return user return None def get_user(self, user_id): try: return User.objects.get(pk=user_id) except User.DoesNotExist: return None
from django.contrib.auth import get_user_model import ldap User = get_user_model() class UChicagoLDAPBackend(object): LDAP_SERVER = "ldaps://ldap.uchicago.edu:636" def authenticate(self, cnetid=None, password=None): if cnetid and password: cnetid = ldap.filter.escape_filter_chars(cnetid) try: conn.simple_bind_s("uid=%s,ou=people,dc=uchicago,dc=edu" % cnetid, password) except ldap.INVALID_CREDENTIALS: return None try: user = User.objects.get(username=cnetid) except User.DoesNotExist: query = "(&(uid=%s)(objectclass=inetOrgPerson))" % (cnetid) result = conn.search_ext_s("dc=uchicago,dc=edu", ldap.SCOPE_SUBTREE, query) if result: user_data = result[0] user = User.objects.create_user(username=cnetid, email=user_data['mail'], first_name=user_data['givenName'], last_name=user_data['sn']) return user return None def get_user(self, user_id): try: return User.objects.get(pk=user_id) except User.DoesNotExist: return None
Change email to mail in LDAP
Change email to mail in LDAP
Python
mit
kz26/uchicago-hvz,kz26/uchicago-hvz,kz26/uchicago-hvz
from django.contrib.auth import get_user_model import ldap User = get_user_model() class UChicagoLDAPBackend(object): LDAP_SERVER = "ldaps://ldap.uchicago.edu:636" def authenticate(self, cnetid=None, password=None): if cnetid and password: cnetid = ldap.filter.escape_filter_chars(cnetid) try: conn.simple_bind_s("uid=%s,ou=people,dc=uchicago,dc=edu" % cnetid, password) except ldap.INVALID_CREDENTIALS: return None try: user = User.objects.get(username=cnetid) except User.DoesNotExist: query = "(&(uid=%s)(objectclass=inetOrgPerson))" % (cnetid) result = conn.search_ext_s("dc=uchicago,dc=edu", ldap.SCOPE_SUBTREE, query) if result: user_data = result[0] user = User.objects.create_user(username=cnetid, email=user_data['email'], first_name=user_data['givenName'], last_name=user_data['sn']) return user return None def get_user(self, user_id): try: return User.objects.get(pk=user_id) except User.DoesNotExist: return NoneChange email to mail in LDAP
from django.contrib.auth import get_user_model import ldap User = get_user_model() class UChicagoLDAPBackend(object): LDAP_SERVER = "ldaps://ldap.uchicago.edu:636" def authenticate(self, cnetid=None, password=None): if cnetid and password: cnetid = ldap.filter.escape_filter_chars(cnetid) try: conn.simple_bind_s("uid=%s,ou=people,dc=uchicago,dc=edu" % cnetid, password) except ldap.INVALID_CREDENTIALS: return None try: user = User.objects.get(username=cnetid) except User.DoesNotExist: query = "(&(uid=%s)(objectclass=inetOrgPerson))" % (cnetid) result = conn.search_ext_s("dc=uchicago,dc=edu", ldap.SCOPE_SUBTREE, query) if result: user_data = result[0] user = User.objects.create_user(username=cnetid, email=user_data['mail'], first_name=user_data['givenName'], last_name=user_data['sn']) return user return None def get_user(self, user_id): try: return User.objects.get(pk=user_id) except User.DoesNotExist: return None
<commit_before>from django.contrib.auth import get_user_model import ldap User = get_user_model() class UChicagoLDAPBackend(object): LDAP_SERVER = "ldaps://ldap.uchicago.edu:636" def authenticate(self, cnetid=None, password=None): if cnetid and password: cnetid = ldap.filter.escape_filter_chars(cnetid) try: conn.simple_bind_s("uid=%s,ou=people,dc=uchicago,dc=edu" % cnetid, password) except ldap.INVALID_CREDENTIALS: return None try: user = User.objects.get(username=cnetid) except User.DoesNotExist: query = "(&(uid=%s)(objectclass=inetOrgPerson))" % (cnetid) result = conn.search_ext_s("dc=uchicago,dc=edu", ldap.SCOPE_SUBTREE, query) if result: user_data = result[0] user = User.objects.create_user(username=cnetid, email=user_data['email'], first_name=user_data['givenName'], last_name=user_data['sn']) return user return None def get_user(self, user_id): try: return User.objects.get(pk=user_id) except User.DoesNotExist: return None<commit_msg>Change email to mail in LDAP<commit_after>
from django.contrib.auth import get_user_model import ldap User = get_user_model() class UChicagoLDAPBackend(object): LDAP_SERVER = "ldaps://ldap.uchicago.edu:636" def authenticate(self, cnetid=None, password=None): if cnetid and password: cnetid = ldap.filter.escape_filter_chars(cnetid) try: conn.simple_bind_s("uid=%s,ou=people,dc=uchicago,dc=edu" % cnetid, password) except ldap.INVALID_CREDENTIALS: return None try: user = User.objects.get(username=cnetid) except User.DoesNotExist: query = "(&(uid=%s)(objectclass=inetOrgPerson))" % (cnetid) result = conn.search_ext_s("dc=uchicago,dc=edu", ldap.SCOPE_SUBTREE, query) if result: user_data = result[0] user = User.objects.create_user(username=cnetid, email=user_data['mail'], first_name=user_data['givenName'], last_name=user_data['sn']) return user return None def get_user(self, user_id): try: return User.objects.get(pk=user_id) except User.DoesNotExist: return None
from django.contrib.auth import get_user_model import ldap User = get_user_model() class UChicagoLDAPBackend(object): LDAP_SERVER = "ldaps://ldap.uchicago.edu:636" def authenticate(self, cnetid=None, password=None): if cnetid and password: cnetid = ldap.filter.escape_filter_chars(cnetid) try: conn.simple_bind_s("uid=%s,ou=people,dc=uchicago,dc=edu" % cnetid, password) except ldap.INVALID_CREDENTIALS: return None try: user = User.objects.get(username=cnetid) except User.DoesNotExist: query = "(&(uid=%s)(objectclass=inetOrgPerson))" % (cnetid) result = conn.search_ext_s("dc=uchicago,dc=edu", ldap.SCOPE_SUBTREE, query) if result: user_data = result[0] user = User.objects.create_user(username=cnetid, email=user_data['email'], first_name=user_data['givenName'], last_name=user_data['sn']) return user return None def get_user(self, user_id): try: return User.objects.get(pk=user_id) except User.DoesNotExist: return NoneChange email to mail in LDAPfrom django.contrib.auth import get_user_model import ldap User = get_user_model() class UChicagoLDAPBackend(object): LDAP_SERVER = "ldaps://ldap.uchicago.edu:636" def authenticate(self, cnetid=None, password=None): if cnetid and password: cnetid = ldap.filter.escape_filter_chars(cnetid) try: conn.simple_bind_s("uid=%s,ou=people,dc=uchicago,dc=edu" % cnetid, password) except ldap.INVALID_CREDENTIALS: return None try: user = User.objects.get(username=cnetid) except User.DoesNotExist: query = "(&(uid=%s)(objectclass=inetOrgPerson))" % (cnetid) result = conn.search_ext_s("dc=uchicago,dc=edu", ldap.SCOPE_SUBTREE, query) if result: user_data = result[0] user = User.objects.create_user(username=cnetid, email=user_data['mail'], first_name=user_data['givenName'], last_name=user_data['sn']) return user return None def get_user(self, user_id): try: return User.objects.get(pk=user_id) except User.DoesNotExist: return None
<commit_before>from django.contrib.auth import get_user_model import ldap User = get_user_model() class UChicagoLDAPBackend(object): LDAP_SERVER = "ldaps://ldap.uchicago.edu:636" def authenticate(self, cnetid=None, password=None): if cnetid and password: cnetid = ldap.filter.escape_filter_chars(cnetid) try: conn.simple_bind_s("uid=%s,ou=people,dc=uchicago,dc=edu" % cnetid, password) except ldap.INVALID_CREDENTIALS: return None try: user = User.objects.get(username=cnetid) except User.DoesNotExist: query = "(&(uid=%s)(objectclass=inetOrgPerson))" % (cnetid) result = conn.search_ext_s("dc=uchicago,dc=edu", ldap.SCOPE_SUBTREE, query) if result: user_data = result[0] user = User.objects.create_user(username=cnetid, email=user_data['email'], first_name=user_data['givenName'], last_name=user_data['sn']) return user return None def get_user(self, user_id): try: return User.objects.get(pk=user_id) except User.DoesNotExist: return None<commit_msg>Change email to mail in LDAP<commit_after>from django.contrib.auth import get_user_model import ldap User = get_user_model() class UChicagoLDAPBackend(object): LDAP_SERVER = "ldaps://ldap.uchicago.edu:636" def authenticate(self, cnetid=None, password=None): if cnetid and password: cnetid = ldap.filter.escape_filter_chars(cnetid) try: conn.simple_bind_s("uid=%s,ou=people,dc=uchicago,dc=edu" % cnetid, password) except ldap.INVALID_CREDENTIALS: return None try: user = User.objects.get(username=cnetid) except User.DoesNotExist: query = "(&(uid=%s)(objectclass=inetOrgPerson))" % (cnetid) result = conn.search_ext_s("dc=uchicago,dc=edu", ldap.SCOPE_SUBTREE, query) if result: user_data = result[0] user = User.objects.create_user(username=cnetid, email=user_data['mail'], first_name=user_data['givenName'], last_name=user_data['sn']) return user return None def get_user(self, user_id): try: return User.objects.get(pk=user_id) except User.DoesNotExist: return None
52ffc2b264cbacaee56017cd4a67df4511d60392
celery/managers.py
celery/managers.py
from django.db import models from celery.registry import tasks from datetime import datetime, timedelta __all__ = ["TaskManager", "PeriodicTaskManager"] class TaskManager(models.Manager): def get_task(self, task_id): task, created = self.get_or_create(task_id=task_id) return task def is_done(self, task_id): return self.get_task(task_id).is_done def get_all_expired(self): return self.filter(date_done__lt=datetime.now() - timedelta(days=5)) def delete_expired(self): self.get_all_expired().delete() def mark_as_done(self, task_id): task, created = self.get_or_create(task_id=task_id, defaults={ "is_done": True}) if not created: task.is_done = True task.save() class PeriodicTaskManager(models.Manager): def get_waiting_tasks(self): periodic_tasks = tasks.get_all_periodic() waiting = [] for task_name, task in periodic_tasks.items(): task_meta, created = self.get_or_create(name=task_name) # task_run.every must be a timedelta object. run_at = task_meta.last_run_at + task.run_every if datetime.now() > run_at: waiting.append(task_meta) return waiting
from django.db import models from celery.registry import tasks from datetime import datetime, timedelta __all__ = ["TaskManager", "PeriodicTaskManager"] class TaskManager(models.Manager): def get_task(self, task_id): task, created = self.get_or_create(task_id=task_id) return task def is_done(self, task_id): return self.get_task(task_id).is_done def get_all_expired(self): return self.filter(date_done__lt=datetime.now() - timedelta(days=5), is_done=True) def delete_expired(self): self.get_all_expired().delete() def mark_as_done(self, task_id): task, created = self.get_or_create(task_id=task_id, defaults={ "is_done": True}) if not created: task.is_done = True task.save() class PeriodicTaskManager(models.Manager): def get_waiting_tasks(self): periodic_tasks = tasks.get_all_periodic() waiting = [] for task_name, task in periodic_tasks.items(): task_meta, created = self.get_or_create(name=task_name) # task_run.every must be a timedelta object. run_at = task_meta.last_run_at + task.run_every if datetime.now() > run_at: waiting.append(task_meta) return waiting
Add is_done=True to get_all_expired filter.
Add is_done=True to get_all_expired filter.
Python
bsd-3-clause
WoLpH/celery,cbrepo/celery,cbrepo/celery,ask/celery,frac/celery,WoLpH/celery,mitsuhiko/celery,ask/celery,mitsuhiko/celery,frac/celery
from django.db import models from celery.registry import tasks from datetime import datetime, timedelta __all__ = ["TaskManager", "PeriodicTaskManager"] class TaskManager(models.Manager): def get_task(self, task_id): task, created = self.get_or_create(task_id=task_id) return task def is_done(self, task_id): return self.get_task(task_id).is_done def get_all_expired(self): return self.filter(date_done__lt=datetime.now() - timedelta(days=5)) def delete_expired(self): self.get_all_expired().delete() def mark_as_done(self, task_id): task, created = self.get_or_create(task_id=task_id, defaults={ "is_done": True}) if not created: task.is_done = True task.save() class PeriodicTaskManager(models.Manager): def get_waiting_tasks(self): periodic_tasks = tasks.get_all_periodic() waiting = [] for task_name, task in periodic_tasks.items(): task_meta, created = self.get_or_create(name=task_name) # task_run.every must be a timedelta object. run_at = task_meta.last_run_at + task.run_every if datetime.now() > run_at: waiting.append(task_meta) return waiting Add is_done=True to get_all_expired filter.
from django.db import models from celery.registry import tasks from datetime import datetime, timedelta __all__ = ["TaskManager", "PeriodicTaskManager"] class TaskManager(models.Manager): def get_task(self, task_id): task, created = self.get_or_create(task_id=task_id) return task def is_done(self, task_id): return self.get_task(task_id).is_done def get_all_expired(self): return self.filter(date_done__lt=datetime.now() - timedelta(days=5), is_done=True) def delete_expired(self): self.get_all_expired().delete() def mark_as_done(self, task_id): task, created = self.get_or_create(task_id=task_id, defaults={ "is_done": True}) if not created: task.is_done = True task.save() class PeriodicTaskManager(models.Manager): def get_waiting_tasks(self): periodic_tasks = tasks.get_all_periodic() waiting = [] for task_name, task in periodic_tasks.items(): task_meta, created = self.get_or_create(name=task_name) # task_run.every must be a timedelta object. run_at = task_meta.last_run_at + task.run_every if datetime.now() > run_at: waiting.append(task_meta) return waiting
<commit_before>from django.db import models from celery.registry import tasks from datetime import datetime, timedelta __all__ = ["TaskManager", "PeriodicTaskManager"] class TaskManager(models.Manager): def get_task(self, task_id): task, created = self.get_or_create(task_id=task_id) return task def is_done(self, task_id): return self.get_task(task_id).is_done def get_all_expired(self): return self.filter(date_done__lt=datetime.now() - timedelta(days=5)) def delete_expired(self): self.get_all_expired().delete() def mark_as_done(self, task_id): task, created = self.get_or_create(task_id=task_id, defaults={ "is_done": True}) if not created: task.is_done = True task.save() class PeriodicTaskManager(models.Manager): def get_waiting_tasks(self): periodic_tasks = tasks.get_all_periodic() waiting = [] for task_name, task in periodic_tasks.items(): task_meta, created = self.get_or_create(name=task_name) # task_run.every must be a timedelta object. run_at = task_meta.last_run_at + task.run_every if datetime.now() > run_at: waiting.append(task_meta) return waiting <commit_msg>Add is_done=True to get_all_expired filter.<commit_after>
from django.db import models from celery.registry import tasks from datetime import datetime, timedelta __all__ = ["TaskManager", "PeriodicTaskManager"] class TaskManager(models.Manager): def get_task(self, task_id): task, created = self.get_or_create(task_id=task_id) return task def is_done(self, task_id): return self.get_task(task_id).is_done def get_all_expired(self): return self.filter(date_done__lt=datetime.now() - timedelta(days=5), is_done=True) def delete_expired(self): self.get_all_expired().delete() def mark_as_done(self, task_id): task, created = self.get_or_create(task_id=task_id, defaults={ "is_done": True}) if not created: task.is_done = True task.save() class PeriodicTaskManager(models.Manager): def get_waiting_tasks(self): periodic_tasks = tasks.get_all_periodic() waiting = [] for task_name, task in periodic_tasks.items(): task_meta, created = self.get_or_create(name=task_name) # task_run.every must be a timedelta object. run_at = task_meta.last_run_at + task.run_every if datetime.now() > run_at: waiting.append(task_meta) return waiting
from django.db import models from celery.registry import tasks from datetime import datetime, timedelta __all__ = ["TaskManager", "PeriodicTaskManager"] class TaskManager(models.Manager): def get_task(self, task_id): task, created = self.get_or_create(task_id=task_id) return task def is_done(self, task_id): return self.get_task(task_id).is_done def get_all_expired(self): return self.filter(date_done__lt=datetime.now() - timedelta(days=5)) def delete_expired(self): self.get_all_expired().delete() def mark_as_done(self, task_id): task, created = self.get_or_create(task_id=task_id, defaults={ "is_done": True}) if not created: task.is_done = True task.save() class PeriodicTaskManager(models.Manager): def get_waiting_tasks(self): periodic_tasks = tasks.get_all_periodic() waiting = [] for task_name, task in periodic_tasks.items(): task_meta, created = self.get_or_create(name=task_name) # task_run.every must be a timedelta object. run_at = task_meta.last_run_at + task.run_every if datetime.now() > run_at: waiting.append(task_meta) return waiting Add is_done=True to get_all_expired filter.from django.db import models from celery.registry import tasks from datetime import datetime, timedelta __all__ = ["TaskManager", "PeriodicTaskManager"] class TaskManager(models.Manager): def get_task(self, task_id): task, created = self.get_or_create(task_id=task_id) return task def is_done(self, task_id): return self.get_task(task_id).is_done def get_all_expired(self): return self.filter(date_done__lt=datetime.now() - timedelta(days=5), is_done=True) def delete_expired(self): self.get_all_expired().delete() def mark_as_done(self, task_id): task, created = self.get_or_create(task_id=task_id, defaults={ "is_done": True}) if not created: task.is_done = True task.save() class PeriodicTaskManager(models.Manager): def get_waiting_tasks(self): periodic_tasks = tasks.get_all_periodic() waiting = [] for task_name, task in periodic_tasks.items(): task_meta, created = self.get_or_create(name=task_name) # task_run.every must be a timedelta object. run_at = task_meta.last_run_at + task.run_every if datetime.now() > run_at: waiting.append(task_meta) return waiting
<commit_before>from django.db import models from celery.registry import tasks from datetime import datetime, timedelta __all__ = ["TaskManager", "PeriodicTaskManager"] class TaskManager(models.Manager): def get_task(self, task_id): task, created = self.get_or_create(task_id=task_id) return task def is_done(self, task_id): return self.get_task(task_id).is_done def get_all_expired(self): return self.filter(date_done__lt=datetime.now() - timedelta(days=5)) def delete_expired(self): self.get_all_expired().delete() def mark_as_done(self, task_id): task, created = self.get_or_create(task_id=task_id, defaults={ "is_done": True}) if not created: task.is_done = True task.save() class PeriodicTaskManager(models.Manager): def get_waiting_tasks(self): periodic_tasks = tasks.get_all_periodic() waiting = [] for task_name, task in periodic_tasks.items(): task_meta, created = self.get_or_create(name=task_name) # task_run.every must be a timedelta object. run_at = task_meta.last_run_at + task.run_every if datetime.now() > run_at: waiting.append(task_meta) return waiting <commit_msg>Add is_done=True to get_all_expired filter.<commit_after>from django.db import models from celery.registry import tasks from datetime import datetime, timedelta __all__ = ["TaskManager", "PeriodicTaskManager"] class TaskManager(models.Manager): def get_task(self, task_id): task, created = self.get_or_create(task_id=task_id) return task def is_done(self, task_id): return self.get_task(task_id).is_done def get_all_expired(self): return self.filter(date_done__lt=datetime.now() - timedelta(days=5), is_done=True) def delete_expired(self): self.get_all_expired().delete() def mark_as_done(self, task_id): task, created = self.get_or_create(task_id=task_id, defaults={ "is_done": True}) if not created: task.is_done = True task.save() class PeriodicTaskManager(models.Manager): def get_waiting_tasks(self): periodic_tasks = tasks.get_all_periodic() waiting = [] for task_name, task in periodic_tasks.items(): task_meta, created = self.get_or_create(name=task_name) # task_run.every must be a timedelta object. run_at = task_meta.last_run_at + task.run_every if datetime.now() > run_at: waiting.append(task_meta) return waiting
697fbacc04bc41dea056377e85ba6b29949d8feb
wars/device.py
wars/device.py
import pygame from pygame.locals import * from wars.block import Block class Device(object): # Static params height = 60 # Object params blocks = [] title = None pos_y = None def __init__(self, title, pos): self.title = title self.pos_y = pos self.blocks = [] def move_blocks(self, amount): for block in self.blocks: block.move(amount) if block.right_edge() <= 0: self.blocks.remove(block) def add_block(self, block): block.pos_y = self.pos_y * self.height self.blocks.append(block) def reset(self): del self.blocks[:] def draw(self, surface): # Do all the drawing of yourself ans such # TODO: Implement this.. # Draw the blocks into yourself for block in self.blocks: block.draw(surface)
import pygame from pygame.locals import * from wars.block import Block class Device(object): # Static params height = 60 # Object params blocks = [] title = None pos_y = None def __init__(self, title, pos): self.title = title self.pos_y = pos self.blocks = [] def move_blocks(self, amount): for block in self.blocks: block.move(amount) if block.right_edge() <= 0: self.blocks.remove(block) del block def add_block(self, block): block.pos_y = self.pos_y * self.height self.blocks.append(block) def reset(self): del self.blocks[:] def draw(self, surface): # Do all the drawing of yourself and such. # As there is currently only one ST, this is omitted # Draw the blocks into yourself for block in self.blocks: block.draw(surface)
Remove objects when they are not used
Remove objects when they are not used
Python
mit
cmol/wifi-wars
import pygame from pygame.locals import * from wars.block import Block class Device(object): # Static params height = 60 # Object params blocks = [] title = None pos_y = None def __init__(self, title, pos): self.title = title self.pos_y = pos self.blocks = [] def move_blocks(self, amount): for block in self.blocks: block.move(amount) if block.right_edge() <= 0: self.blocks.remove(block) def add_block(self, block): block.pos_y = self.pos_y * self.height self.blocks.append(block) def reset(self): del self.blocks[:] def draw(self, surface): # Do all the drawing of yourself ans such # TODO: Implement this.. # Draw the blocks into yourself for block in self.blocks: block.draw(surface) Remove objects when they are not used
import pygame from pygame.locals import * from wars.block import Block class Device(object): # Static params height = 60 # Object params blocks = [] title = None pos_y = None def __init__(self, title, pos): self.title = title self.pos_y = pos self.blocks = [] def move_blocks(self, amount): for block in self.blocks: block.move(amount) if block.right_edge() <= 0: self.blocks.remove(block) del block def add_block(self, block): block.pos_y = self.pos_y * self.height self.blocks.append(block) def reset(self): del self.blocks[:] def draw(self, surface): # Do all the drawing of yourself and such. # As there is currently only one ST, this is omitted # Draw the blocks into yourself for block in self.blocks: block.draw(surface)
<commit_before>import pygame from pygame.locals import * from wars.block import Block class Device(object): # Static params height = 60 # Object params blocks = [] title = None pos_y = None def __init__(self, title, pos): self.title = title self.pos_y = pos self.blocks = [] def move_blocks(self, amount): for block in self.blocks: block.move(amount) if block.right_edge() <= 0: self.blocks.remove(block) def add_block(self, block): block.pos_y = self.pos_y * self.height self.blocks.append(block) def reset(self): del self.blocks[:] def draw(self, surface): # Do all the drawing of yourself ans such # TODO: Implement this.. # Draw the blocks into yourself for block in self.blocks: block.draw(surface) <commit_msg>Remove objects when they are not used<commit_after>
import pygame from pygame.locals import * from wars.block import Block class Device(object): # Static params height = 60 # Object params blocks = [] title = None pos_y = None def __init__(self, title, pos): self.title = title self.pos_y = pos self.blocks = [] def move_blocks(self, amount): for block in self.blocks: block.move(amount) if block.right_edge() <= 0: self.blocks.remove(block) del block def add_block(self, block): block.pos_y = self.pos_y * self.height self.blocks.append(block) def reset(self): del self.blocks[:] def draw(self, surface): # Do all the drawing of yourself and such. # As there is currently only one ST, this is omitted # Draw the blocks into yourself for block in self.blocks: block.draw(surface)
import pygame from pygame.locals import * from wars.block import Block class Device(object): # Static params height = 60 # Object params blocks = [] title = None pos_y = None def __init__(self, title, pos): self.title = title self.pos_y = pos self.blocks = [] def move_blocks(self, amount): for block in self.blocks: block.move(amount) if block.right_edge() <= 0: self.blocks.remove(block) def add_block(self, block): block.pos_y = self.pos_y * self.height self.blocks.append(block) def reset(self): del self.blocks[:] def draw(self, surface): # Do all the drawing of yourself ans such # TODO: Implement this.. # Draw the blocks into yourself for block in self.blocks: block.draw(surface) Remove objects when they are not usedimport pygame from pygame.locals import * from wars.block import Block class Device(object): # Static params height = 60 # Object params blocks = [] title = None pos_y = None def __init__(self, title, pos): self.title = title self.pos_y = pos self.blocks = [] def move_blocks(self, amount): for block in self.blocks: block.move(amount) if block.right_edge() <= 0: self.blocks.remove(block) del block def add_block(self, block): block.pos_y = self.pos_y * self.height self.blocks.append(block) def reset(self): del self.blocks[:] def draw(self, surface): # Do all the drawing of yourself and such. # As there is currently only one ST, this is omitted # Draw the blocks into yourself for block in self.blocks: block.draw(surface)
<commit_before>import pygame from pygame.locals import * from wars.block import Block class Device(object): # Static params height = 60 # Object params blocks = [] title = None pos_y = None def __init__(self, title, pos): self.title = title self.pos_y = pos self.blocks = [] def move_blocks(self, amount): for block in self.blocks: block.move(amount) if block.right_edge() <= 0: self.blocks.remove(block) def add_block(self, block): block.pos_y = self.pos_y * self.height self.blocks.append(block) def reset(self): del self.blocks[:] def draw(self, surface): # Do all the drawing of yourself ans such # TODO: Implement this.. # Draw the blocks into yourself for block in self.blocks: block.draw(surface) <commit_msg>Remove objects when they are not used<commit_after>import pygame from pygame.locals import * from wars.block import Block class Device(object): # Static params height = 60 # Object params blocks = [] title = None pos_y = None def __init__(self, title, pos): self.title = title self.pos_y = pos self.blocks = [] def move_blocks(self, amount): for block in self.blocks: block.move(amount) if block.right_edge() <= 0: self.blocks.remove(block) del block def add_block(self, block): block.pos_y = self.pos_y * self.height self.blocks.append(block) def reset(self): del self.blocks[:] def draw(self, surface): # Do all the drawing of yourself and such. # As there is currently only one ST, this is omitted # Draw the blocks into yourself for block in self.blocks: block.draw(surface)
5fc1735082c001d81d9cded4c694c12b7afac8a0
scripts/scrape-ecdc-list.py
scripts/scrape-ecdc-list.py
#!/usr/bin/env python import requests import lxml.html import pandas as pd import sys URL = "http://ecdc.europa.eu/en/healthtopics/zika_virus_infection/zika-outbreak/Pages/Zika-countries-with-transmission.aspx" columns = [ "country", "current_zika_transmission", ] def scrape(): html = requests.get(URL).content dom = lxml.html.fromstring(html) table = dom.cssselect(".ms-rteTable-1")[0] rows = table.cssselect("tr")[1:] data = [ [ td.text_content().strip() for td in tr.cssselect("td, th") ] for tr in rows ] df = pd.DataFrame(data, columns=columns)[columns] return df if __name__ == "__main__": df = scrape() df.to_csv(sys.stdout, index=False, encoding="utf-8")
#!/usr/bin/env python import requests import lxml.html import pandas as pd import sys URL = "http://ecdc.europa.eu/en/healthtopics/zika_virus_infection/zika-outbreak/Pages/Zika-countries-with-transmission.aspx" columns = [ "country", "current_zika_transmission", ] def scrape(): html = requests.get(URL).content dom = lxml.html.fromstring(html) table = dom.cssselect(".ms-rteTable-1")[-1] rows = table.cssselect("tr")[1:] data = [ [ td.text_content().strip() for td in tr.cssselect("td, th") ] for tr in rows ] df = pd.DataFrame(data, columns=columns)[columns] return df if __name__ == "__main__": df = scrape() df.to_csv(sys.stdout, index=False, encoding="utf-8")
Update ECDC scraper for tweaked page structure
Update ECDC scraper for tweaked page structure
Python
mit
BuzzFeedNews/zika-data
#!/usr/bin/env python import requests import lxml.html import pandas as pd import sys URL = "http://ecdc.europa.eu/en/healthtopics/zika_virus_infection/zika-outbreak/Pages/Zika-countries-with-transmission.aspx" columns = [ "country", "current_zika_transmission", ] def scrape(): html = requests.get(URL).content dom = lxml.html.fromstring(html) table = dom.cssselect(".ms-rteTable-1")[0] rows = table.cssselect("tr")[1:] data = [ [ td.text_content().strip() for td in tr.cssselect("td, th") ] for tr in rows ] df = pd.DataFrame(data, columns=columns)[columns] return df if __name__ == "__main__": df = scrape() df.to_csv(sys.stdout, index=False, encoding="utf-8") Update ECDC scraper for tweaked page structure
#!/usr/bin/env python import requests import lxml.html import pandas as pd import sys URL = "http://ecdc.europa.eu/en/healthtopics/zika_virus_infection/zika-outbreak/Pages/Zika-countries-with-transmission.aspx" columns = [ "country", "current_zika_transmission", ] def scrape(): html = requests.get(URL).content dom = lxml.html.fromstring(html) table = dom.cssselect(".ms-rteTable-1")[-1] rows = table.cssselect("tr")[1:] data = [ [ td.text_content().strip() for td in tr.cssselect("td, th") ] for tr in rows ] df = pd.DataFrame(data, columns=columns)[columns] return df if __name__ == "__main__": df = scrape() df.to_csv(sys.stdout, index=False, encoding="utf-8")
<commit_before>#!/usr/bin/env python import requests import lxml.html import pandas as pd import sys URL = "http://ecdc.europa.eu/en/healthtopics/zika_virus_infection/zika-outbreak/Pages/Zika-countries-with-transmission.aspx" columns = [ "country", "current_zika_transmission", ] def scrape(): html = requests.get(URL).content dom = lxml.html.fromstring(html) table = dom.cssselect(".ms-rteTable-1")[0] rows = table.cssselect("tr")[1:] data = [ [ td.text_content().strip() for td in tr.cssselect("td, th") ] for tr in rows ] df = pd.DataFrame(data, columns=columns)[columns] return df if __name__ == "__main__": df = scrape() df.to_csv(sys.stdout, index=False, encoding="utf-8") <commit_msg>Update ECDC scraper for tweaked page structure<commit_after>
#!/usr/bin/env python import requests import lxml.html import pandas as pd import sys URL = "http://ecdc.europa.eu/en/healthtopics/zika_virus_infection/zika-outbreak/Pages/Zika-countries-with-transmission.aspx" columns = [ "country", "current_zika_transmission", ] def scrape(): html = requests.get(URL).content dom = lxml.html.fromstring(html) table = dom.cssselect(".ms-rteTable-1")[-1] rows = table.cssselect("tr")[1:] data = [ [ td.text_content().strip() for td in tr.cssselect("td, th") ] for tr in rows ] df = pd.DataFrame(data, columns=columns)[columns] return df if __name__ == "__main__": df = scrape() df.to_csv(sys.stdout, index=False, encoding="utf-8")
#!/usr/bin/env python import requests import lxml.html import pandas as pd import sys URL = "http://ecdc.europa.eu/en/healthtopics/zika_virus_infection/zika-outbreak/Pages/Zika-countries-with-transmission.aspx" columns = [ "country", "current_zika_transmission", ] def scrape(): html = requests.get(URL).content dom = lxml.html.fromstring(html) table = dom.cssselect(".ms-rteTable-1")[0] rows = table.cssselect("tr")[1:] data = [ [ td.text_content().strip() for td in tr.cssselect("td, th") ] for tr in rows ] df = pd.DataFrame(data, columns=columns)[columns] return df if __name__ == "__main__": df = scrape() df.to_csv(sys.stdout, index=False, encoding="utf-8") Update ECDC scraper for tweaked page structure#!/usr/bin/env python import requests import lxml.html import pandas as pd import sys URL = "http://ecdc.europa.eu/en/healthtopics/zika_virus_infection/zika-outbreak/Pages/Zika-countries-with-transmission.aspx" columns = [ "country", "current_zika_transmission", ] def scrape(): html = requests.get(URL).content dom = lxml.html.fromstring(html) table = dom.cssselect(".ms-rteTable-1")[-1] rows = table.cssselect("tr")[1:] data = [ [ td.text_content().strip() for td in tr.cssselect("td, th") ] for tr in rows ] df = pd.DataFrame(data, columns=columns)[columns] return df if __name__ == "__main__": df = scrape() df.to_csv(sys.stdout, index=False, encoding="utf-8")
<commit_before>#!/usr/bin/env python import requests import lxml.html import pandas as pd import sys URL = "http://ecdc.europa.eu/en/healthtopics/zika_virus_infection/zika-outbreak/Pages/Zika-countries-with-transmission.aspx" columns = [ "country", "current_zika_transmission", ] def scrape(): html = requests.get(URL).content dom = lxml.html.fromstring(html) table = dom.cssselect(".ms-rteTable-1")[0] rows = table.cssselect("tr")[1:] data = [ [ td.text_content().strip() for td in tr.cssselect("td, th") ] for tr in rows ] df = pd.DataFrame(data, columns=columns)[columns] return df if __name__ == "__main__": df = scrape() df.to_csv(sys.stdout, index=False, encoding="utf-8") <commit_msg>Update ECDC scraper for tweaked page structure<commit_after>#!/usr/bin/env python import requests import lxml.html import pandas as pd import sys URL = "http://ecdc.europa.eu/en/healthtopics/zika_virus_infection/zika-outbreak/Pages/Zika-countries-with-transmission.aspx" columns = [ "country", "current_zika_transmission", ] def scrape(): html = requests.get(URL).content dom = lxml.html.fromstring(html) table = dom.cssselect(".ms-rteTable-1")[-1] rows = table.cssselect("tr")[1:] data = [ [ td.text_content().strip() for td in tr.cssselect("td, th") ] for tr in rows ] df = pd.DataFrame(data, columns=columns)[columns] return df if __name__ == "__main__": df = scrape() df.to_csv(sys.stdout, index=False, encoding="utf-8")
5d901fddb0c863f811b0de40c063bb00b50e5394
ion/processes/bootstrap/plugins/bootstrap_core.py
ion/processes/bootstrap/plugins/bootstrap_core.py
#!/usr/bin/env python __author__ = 'Michael Meisinger' from ion.core.bootstrap_process import BootstrapPlugin, AbortBootstrap from pyon.public import IonObject, RT from pyon.util.containers import get_safe from interface.objects import ActorIdentity, Org class BootstrapCore(BootstrapPlugin): """ Bootstrap plugin for core system resources. No service dependency """ def on_initial_bootstrap(self, process, config, **kwargs): # Detect if system has been started before by the presence of the ION system actor system_actor, _ = process.container.resource_registry.find_resources( restype=RT.ActorIdentity, id_only=True) if system_actor: raise AbortBootstrap("System already initialized. Start with bootmode=restart or force_clean!") # Possibly start the event persister here # Create ION actor actor_name = get_safe(config, "system.system_actor", "ionsystem") sys_actor = ActorIdentity(name=actor_name, description="ION System Agent") process.container.resource_registry.create(sys_actor) # Store all resource types # Store all event types def on_restart(self, process, config, **kwargs): pass
#!/usr/bin/env python __author__ = 'Michael Meisinger' from ion.core.bootstrap_process import BootstrapPlugin, AbortBootstrap from pyon.public import IonObject, RT from pyon.util.containers import get_safe from interface.objects import ActorIdentity, Org class BootstrapCore(BootstrapPlugin): """ Bootstrap plugin for core system resources. No service dependency """ def on_initial_bootstrap(self, process, config, **kwargs): # Detect if system has been started before by the presence of the ION system actor system_actor, _ = process.container.resource_registry.find_resources( restype=RT.ActorIdentity, id_only=True) if system_actor: raise AbortBootstrap("System already initialized. Start with bootmode=restart or force_clean!") # Possibly start the event persister here # Create ION actor actor_name = get_safe(config, "system.system_actor", "ionsystem") sys_actor = ActorIdentity(name=actor_name, description="ION System Agent") process.container.resource_registry.create(sys_actor) webauth_actor_name = get_safe(config, "system.web_authentication_actor", "web_authentication") web_auth_actor = ActorIdentity(name=webauth_actor_name, description="Web Authentication Actor") process.container.resource_registry.create(web_auth_actor) # Store all resource types # Store all event types def on_restart(self, process, config, **kwargs): pass
Add web authentication actor on bootstrap
Add web authentication actor on bootstrap
Python
bsd-2-clause
ooici/coi-services,ooici/coi-services,ooici/coi-services,ooici/coi-services,ooici/coi-services
#!/usr/bin/env python __author__ = 'Michael Meisinger' from ion.core.bootstrap_process import BootstrapPlugin, AbortBootstrap from pyon.public import IonObject, RT from pyon.util.containers import get_safe from interface.objects import ActorIdentity, Org class BootstrapCore(BootstrapPlugin): """ Bootstrap plugin for core system resources. No service dependency """ def on_initial_bootstrap(self, process, config, **kwargs): # Detect if system has been started before by the presence of the ION system actor system_actor, _ = process.container.resource_registry.find_resources( restype=RT.ActorIdentity, id_only=True) if system_actor: raise AbortBootstrap("System already initialized. Start with bootmode=restart or force_clean!") # Possibly start the event persister here # Create ION actor actor_name = get_safe(config, "system.system_actor", "ionsystem") sys_actor = ActorIdentity(name=actor_name, description="ION System Agent") process.container.resource_registry.create(sys_actor) # Store all resource types # Store all event types def on_restart(self, process, config, **kwargs): pass Add web authentication actor on bootstrap
#!/usr/bin/env python __author__ = 'Michael Meisinger' from ion.core.bootstrap_process import BootstrapPlugin, AbortBootstrap from pyon.public import IonObject, RT from pyon.util.containers import get_safe from interface.objects import ActorIdentity, Org class BootstrapCore(BootstrapPlugin): """ Bootstrap plugin for core system resources. No service dependency """ def on_initial_bootstrap(self, process, config, **kwargs): # Detect if system has been started before by the presence of the ION system actor system_actor, _ = process.container.resource_registry.find_resources( restype=RT.ActorIdentity, id_only=True) if system_actor: raise AbortBootstrap("System already initialized. Start with bootmode=restart or force_clean!") # Possibly start the event persister here # Create ION actor actor_name = get_safe(config, "system.system_actor", "ionsystem") sys_actor = ActorIdentity(name=actor_name, description="ION System Agent") process.container.resource_registry.create(sys_actor) webauth_actor_name = get_safe(config, "system.web_authentication_actor", "web_authentication") web_auth_actor = ActorIdentity(name=webauth_actor_name, description="Web Authentication Actor") process.container.resource_registry.create(web_auth_actor) # Store all resource types # Store all event types def on_restart(self, process, config, **kwargs): pass
<commit_before>#!/usr/bin/env python __author__ = 'Michael Meisinger' from ion.core.bootstrap_process import BootstrapPlugin, AbortBootstrap from pyon.public import IonObject, RT from pyon.util.containers import get_safe from interface.objects import ActorIdentity, Org class BootstrapCore(BootstrapPlugin): """ Bootstrap plugin for core system resources. No service dependency """ def on_initial_bootstrap(self, process, config, **kwargs): # Detect if system has been started before by the presence of the ION system actor system_actor, _ = process.container.resource_registry.find_resources( restype=RT.ActorIdentity, id_only=True) if system_actor: raise AbortBootstrap("System already initialized. Start with bootmode=restart or force_clean!") # Possibly start the event persister here # Create ION actor actor_name = get_safe(config, "system.system_actor", "ionsystem") sys_actor = ActorIdentity(name=actor_name, description="ION System Agent") process.container.resource_registry.create(sys_actor) # Store all resource types # Store all event types def on_restart(self, process, config, **kwargs): pass <commit_msg>Add web authentication actor on bootstrap<commit_after>
#!/usr/bin/env python __author__ = 'Michael Meisinger' from ion.core.bootstrap_process import BootstrapPlugin, AbortBootstrap from pyon.public import IonObject, RT from pyon.util.containers import get_safe from interface.objects import ActorIdentity, Org class BootstrapCore(BootstrapPlugin): """ Bootstrap plugin for core system resources. No service dependency """ def on_initial_bootstrap(self, process, config, **kwargs): # Detect if system has been started before by the presence of the ION system actor system_actor, _ = process.container.resource_registry.find_resources( restype=RT.ActorIdentity, id_only=True) if system_actor: raise AbortBootstrap("System already initialized. Start with bootmode=restart or force_clean!") # Possibly start the event persister here # Create ION actor actor_name = get_safe(config, "system.system_actor", "ionsystem") sys_actor = ActorIdentity(name=actor_name, description="ION System Agent") process.container.resource_registry.create(sys_actor) webauth_actor_name = get_safe(config, "system.web_authentication_actor", "web_authentication") web_auth_actor = ActorIdentity(name=webauth_actor_name, description="Web Authentication Actor") process.container.resource_registry.create(web_auth_actor) # Store all resource types # Store all event types def on_restart(self, process, config, **kwargs): pass
#!/usr/bin/env python __author__ = 'Michael Meisinger' from ion.core.bootstrap_process import BootstrapPlugin, AbortBootstrap from pyon.public import IonObject, RT from pyon.util.containers import get_safe from interface.objects import ActorIdentity, Org class BootstrapCore(BootstrapPlugin): """ Bootstrap plugin for core system resources. No service dependency """ def on_initial_bootstrap(self, process, config, **kwargs): # Detect if system has been started before by the presence of the ION system actor system_actor, _ = process.container.resource_registry.find_resources( restype=RT.ActorIdentity, id_only=True) if system_actor: raise AbortBootstrap("System already initialized. Start with bootmode=restart or force_clean!") # Possibly start the event persister here # Create ION actor actor_name = get_safe(config, "system.system_actor", "ionsystem") sys_actor = ActorIdentity(name=actor_name, description="ION System Agent") process.container.resource_registry.create(sys_actor) # Store all resource types # Store all event types def on_restart(self, process, config, **kwargs): pass Add web authentication actor on bootstrap#!/usr/bin/env python __author__ = 'Michael Meisinger' from ion.core.bootstrap_process import BootstrapPlugin, AbortBootstrap from pyon.public import IonObject, RT from pyon.util.containers import get_safe from interface.objects import ActorIdentity, Org class BootstrapCore(BootstrapPlugin): """ Bootstrap plugin for core system resources. No service dependency """ def on_initial_bootstrap(self, process, config, **kwargs): # Detect if system has been started before by the presence of the ION system actor system_actor, _ = process.container.resource_registry.find_resources( restype=RT.ActorIdentity, id_only=True) if system_actor: raise AbortBootstrap("System already initialized. Start with bootmode=restart or force_clean!") # Possibly start the event persister here # Create ION actor actor_name = get_safe(config, "system.system_actor", "ionsystem") sys_actor = ActorIdentity(name=actor_name, description="ION System Agent") process.container.resource_registry.create(sys_actor) webauth_actor_name = get_safe(config, "system.web_authentication_actor", "web_authentication") web_auth_actor = ActorIdentity(name=webauth_actor_name, description="Web Authentication Actor") process.container.resource_registry.create(web_auth_actor) # Store all resource types # Store all event types def on_restart(self, process, config, **kwargs): pass
<commit_before>#!/usr/bin/env python __author__ = 'Michael Meisinger' from ion.core.bootstrap_process import BootstrapPlugin, AbortBootstrap from pyon.public import IonObject, RT from pyon.util.containers import get_safe from interface.objects import ActorIdentity, Org class BootstrapCore(BootstrapPlugin): """ Bootstrap plugin for core system resources. No service dependency """ def on_initial_bootstrap(self, process, config, **kwargs): # Detect if system has been started before by the presence of the ION system actor system_actor, _ = process.container.resource_registry.find_resources( restype=RT.ActorIdentity, id_only=True) if system_actor: raise AbortBootstrap("System already initialized. Start with bootmode=restart or force_clean!") # Possibly start the event persister here # Create ION actor actor_name = get_safe(config, "system.system_actor", "ionsystem") sys_actor = ActorIdentity(name=actor_name, description="ION System Agent") process.container.resource_registry.create(sys_actor) # Store all resource types # Store all event types def on_restart(self, process, config, **kwargs): pass <commit_msg>Add web authentication actor on bootstrap<commit_after>#!/usr/bin/env python __author__ = 'Michael Meisinger' from ion.core.bootstrap_process import BootstrapPlugin, AbortBootstrap from pyon.public import IonObject, RT from pyon.util.containers import get_safe from interface.objects import ActorIdentity, Org class BootstrapCore(BootstrapPlugin): """ Bootstrap plugin for core system resources. No service dependency """ def on_initial_bootstrap(self, process, config, **kwargs): # Detect if system has been started before by the presence of the ION system actor system_actor, _ = process.container.resource_registry.find_resources( restype=RT.ActorIdentity, id_only=True) if system_actor: raise AbortBootstrap("System already initialized. Start with bootmode=restart or force_clean!") # Possibly start the event persister here # Create ION actor actor_name = get_safe(config, "system.system_actor", "ionsystem") sys_actor = ActorIdentity(name=actor_name, description="ION System Agent") process.container.resource_registry.create(sys_actor) webauth_actor_name = get_safe(config, "system.web_authentication_actor", "web_authentication") web_auth_actor = ActorIdentity(name=webauth_actor_name, description="Web Authentication Actor") process.container.resource_registry.create(web_auth_actor) # Store all resource types # Store all event types def on_restart(self, process, config, **kwargs): pass
17457815b8eccabf137fdd91053796389d3c8de5
EventParsers/ParserHelper.py
EventParsers/ParserHelper.py
# The note off event contains no data, except for the least significant bits # represening the polyphonic ID, so that all notes with that particular # polyphonic ID can be turned off. def parse_noteOff(bmsfile, read, strict, commandID): return commandID & 0b111 # Several commands use three bytes of data and one byte for # something else. Because Python's struct module does not have # a way to parse three bytes at once, we need to do it as follows. def parse_1Byte_1Tripplet(bmsfile, read, strict, commandID): byte = read.byte() tripplet = read.tripplet() return (byte, tripplet) # Variable-length delay def parse_VL_delay(bmsfile, read, strict, commandID): start = bmsfile.tell() value = read.byte() while (value >> 7) == 1: value = read.byte() dataLen = bmsfile.tell() - start bmsfile.seek(start) data = read.byteArray(dataLen) return (data, )
# The note off event contains no data, except for the least significant bits # represening the polyphonic ID, so that all notes with that particular # polyphonic ID can be turned off. def parse_noteOff(bmsfile, read, strict, commandID): return (commandID & 0b111,) # Several commands use three bytes of data and one byte for # something else. Because Python's struct module does not have # a way to parse three bytes at once, we need to do it as follows. def parse_1Byte_1Tripplet(bmsfile, read, strict, commandID): byte = read.byte() tripplet = read.tripplet() return (byte, tripplet) # Variable-length delay def parse_VL_delay(bmsfile, read, strict, commandID): start = bmsfile.tell() value = read.byte() while (value >> 7) == 1: value = read.byte() dataLen = bmsfile.tell() - start bmsfile.seek(start) data = read.byteArray(dataLen) return (data, )
Make return value of function a tuple
Make return value of function a tuple This should make the return value consistent with what struct.unpack would return.
Python
bsd-3-clause
RenolY2/py-playBMS
# The note off event contains no data, except for the least significant bits # represening the polyphonic ID, so that all notes with that particular # polyphonic ID can be turned off. def parse_noteOff(bmsfile, read, strict, commandID): return commandID & 0b111 # Several commands use three bytes of data and one byte for # something else. Because Python's struct module does not have # a way to parse three bytes at once, we need to do it as follows. def parse_1Byte_1Tripplet(bmsfile, read, strict, commandID): byte = read.byte() tripplet = read.tripplet() return (byte, tripplet) # Variable-length delay def parse_VL_delay(bmsfile, read, strict, commandID): start = bmsfile.tell() value = read.byte() while (value >> 7) == 1: value = read.byte() dataLen = bmsfile.tell() - start bmsfile.seek(start) data = read.byteArray(dataLen) return (data, ) Make return value of function a tuple This should make the return value consistent with what struct.unpack would return.
# The note off event contains no data, except for the least significant bits # represening the polyphonic ID, so that all notes with that particular # polyphonic ID can be turned off. def parse_noteOff(bmsfile, read, strict, commandID): return (commandID & 0b111,) # Several commands use three bytes of data and one byte for # something else. Because Python's struct module does not have # a way to parse three bytes at once, we need to do it as follows. def parse_1Byte_1Tripplet(bmsfile, read, strict, commandID): byte = read.byte() tripplet = read.tripplet() return (byte, tripplet) # Variable-length delay def parse_VL_delay(bmsfile, read, strict, commandID): start = bmsfile.tell() value = read.byte() while (value >> 7) == 1: value = read.byte() dataLen = bmsfile.tell() - start bmsfile.seek(start) data = read.byteArray(dataLen) return (data, )
<commit_before> # The note off event contains no data, except for the least significant bits # represening the polyphonic ID, so that all notes with that particular # polyphonic ID can be turned off. def parse_noteOff(bmsfile, read, strict, commandID): return commandID & 0b111 # Several commands use three bytes of data and one byte for # something else. Because Python's struct module does not have # a way to parse three bytes at once, we need to do it as follows. def parse_1Byte_1Tripplet(bmsfile, read, strict, commandID): byte = read.byte() tripplet = read.tripplet() return (byte, tripplet) # Variable-length delay def parse_VL_delay(bmsfile, read, strict, commandID): start = bmsfile.tell() value = read.byte() while (value >> 7) == 1: value = read.byte() dataLen = bmsfile.tell() - start bmsfile.seek(start) data = read.byteArray(dataLen) return (data, ) <commit_msg>Make return value of function a tuple This should make the return value consistent with what struct.unpack would return.<commit_after>
# The note off event contains no data, except for the least significant bits # represening the polyphonic ID, so that all notes with that particular # polyphonic ID can be turned off. def parse_noteOff(bmsfile, read, strict, commandID): return (commandID & 0b111,) # Several commands use three bytes of data and one byte for # something else. Because Python's struct module does not have # a way to parse three bytes at once, we need to do it as follows. def parse_1Byte_1Tripplet(bmsfile, read, strict, commandID): byte = read.byte() tripplet = read.tripplet() return (byte, tripplet) # Variable-length delay def parse_VL_delay(bmsfile, read, strict, commandID): start = bmsfile.tell() value = read.byte() while (value >> 7) == 1: value = read.byte() dataLen = bmsfile.tell() - start bmsfile.seek(start) data = read.byteArray(dataLen) return (data, )
# The note off event contains no data, except for the least significant bits # represening the polyphonic ID, so that all notes with that particular # polyphonic ID can be turned off. def parse_noteOff(bmsfile, read, strict, commandID): return commandID & 0b111 # Several commands use three bytes of data and one byte for # something else. Because Python's struct module does not have # a way to parse three bytes at once, we need to do it as follows. def parse_1Byte_1Tripplet(bmsfile, read, strict, commandID): byte = read.byte() tripplet = read.tripplet() return (byte, tripplet) # Variable-length delay def parse_VL_delay(bmsfile, read, strict, commandID): start = bmsfile.tell() value = read.byte() while (value >> 7) == 1: value = read.byte() dataLen = bmsfile.tell() - start bmsfile.seek(start) data = read.byteArray(dataLen) return (data, ) Make return value of function a tuple This should make the return value consistent with what struct.unpack would return. # The note off event contains no data, except for the least significant bits # represening the polyphonic ID, so that all notes with that particular # polyphonic ID can be turned off. def parse_noteOff(bmsfile, read, strict, commandID): return (commandID & 0b111,) # Several commands use three bytes of data and one byte for # something else. Because Python's struct module does not have # a way to parse three bytes at once, we need to do it as follows. def parse_1Byte_1Tripplet(bmsfile, read, strict, commandID): byte = read.byte() tripplet = read.tripplet() return (byte, tripplet) # Variable-length delay def parse_VL_delay(bmsfile, read, strict, commandID): start = bmsfile.tell() value = read.byte() while (value >> 7) == 1: value = read.byte() dataLen = bmsfile.tell() - start bmsfile.seek(start) data = read.byteArray(dataLen) return (data, )
<commit_before> # The note off event contains no data, except for the least significant bits # represening the polyphonic ID, so that all notes with that particular # polyphonic ID can be turned off. def parse_noteOff(bmsfile, read, strict, commandID): return commandID & 0b111 # Several commands use three bytes of data and one byte for # something else. Because Python's struct module does not have # a way to parse three bytes at once, we need to do it as follows. def parse_1Byte_1Tripplet(bmsfile, read, strict, commandID): byte = read.byte() tripplet = read.tripplet() return (byte, tripplet) # Variable-length delay def parse_VL_delay(bmsfile, read, strict, commandID): start = bmsfile.tell() value = read.byte() while (value >> 7) == 1: value = read.byte() dataLen = bmsfile.tell() - start bmsfile.seek(start) data = read.byteArray(dataLen) return (data, ) <commit_msg>Make return value of function a tuple This should make the return value consistent with what struct.unpack would return.<commit_after> # The note off event contains no data, except for the least significant bits # represening the polyphonic ID, so that all notes with that particular # polyphonic ID can be turned off. def parse_noteOff(bmsfile, read, strict, commandID): return (commandID & 0b111,) # Several commands use three bytes of data and one byte for # something else. Because Python's struct module does not have # a way to parse three bytes at once, we need to do it as follows. def parse_1Byte_1Tripplet(bmsfile, read, strict, commandID): byte = read.byte() tripplet = read.tripplet() return (byte, tripplet) # Variable-length delay def parse_VL_delay(bmsfile, read, strict, commandID): start = bmsfile.tell() value = read.byte() while (value >> 7) == 1: value = read.byte() dataLen = bmsfile.tell() - start bmsfile.seek(start) data = read.byteArray(dataLen) return (data, )
35974efcbae0c8a1b3009d7a2f38c73a52ff5790
powerdns/admin.py
powerdns/admin.py
# -*- coding: utf-8 -*- from django.contrib import admin from powerdns.models import Domain, Record, Supermaster class RecordAdmin(admin.ModelAdmin): list_display = ('name', 'type', 'content', 'ttl', 'prio', 'change_date',) list_filter = ['type', 'ttl',] search_fields = ('name','content',) class DomainAdmin(admin.ModelAdmin): list_display = ('name', 'type', 'last_check', 'account',) list_filter = ['type', 'last_check', 'account',] search_fields = ('name',) class SupermasterAdmin(admin.ModelAdmin): list_display = ('ip', 'nameserver', 'account',) list_filter = ['ip', 'account',] search_fields = ('ip', 'nameserver',) admin.site.register(Domain,DomainAdmin) admin.site.register(Record,RecordAdmin) admin.site.register(Supermaster,SupermasterAdmin)
# -*- coding: utf-8 -*- from django.contrib import admin from powerdns.models import Domain, Record, Supermaster class RecordAdmin(admin.ModelAdmin): list_display = ('name', 'type', 'content', 'ttl', 'prio', 'change_date',) list_filter = ['type', 'ttl',] search_fields = ('name','content',) readonly_fields = ('change_date',) class DomainAdmin(admin.ModelAdmin): list_display = ('name', 'type', 'last_check', 'account',) list_filter = ['type', 'last_check', 'account',] search_fields = ('name',) readonly_fields = ('notified_serial',) class SupermasterAdmin(admin.ModelAdmin): list_display = ('ip', 'nameserver', 'account',) list_filter = ['ip', 'account',] search_fields = ('ip', 'nameserver',) admin.site.register(Domain,DomainAdmin) admin.site.register(Record,RecordAdmin) admin.site.register(Supermaster,SupermasterAdmin)
Mark some fields as readonly in the Admin panel as they are set by the server (Requires Django 1.2 or greater)
Mark some fields as readonly in the Admin panel as they are set by the server (Requires Django 1.2 or greater)
Python
bsd-2-clause
dominikkowalski/django-powerdns-dnssec,zefciu/django-powerdns-dnssec,zefciu/django-powerdns-dnssec,dominikkowalski/django-powerdns-dnssec,dominikkowalski/django-powerdns-dnssec,allegro/django-powerdns-dnssec,allegro/django-powerdns-dnssec,dominikkowalski/django-powerdns-dnssec,allegro/django-powerdns-dnssec,allegro/django-powerdns-dnssec,dominikkowalski/django-powerdns-dnssec,zefciu/django-powerdns-dnssec
# -*- coding: utf-8 -*- from django.contrib import admin from powerdns.models import Domain, Record, Supermaster class RecordAdmin(admin.ModelAdmin): list_display = ('name', 'type', 'content', 'ttl', 'prio', 'change_date',) list_filter = ['type', 'ttl',] search_fields = ('name','content',) class DomainAdmin(admin.ModelAdmin): list_display = ('name', 'type', 'last_check', 'account',) list_filter = ['type', 'last_check', 'account',] search_fields = ('name',) class SupermasterAdmin(admin.ModelAdmin): list_display = ('ip', 'nameserver', 'account',) list_filter = ['ip', 'account',] search_fields = ('ip', 'nameserver',) admin.site.register(Domain,DomainAdmin) admin.site.register(Record,RecordAdmin) admin.site.register(Supermaster,SupermasterAdmin) Mark some fields as readonly in the Admin panel as they are set by the server (Requires Django 1.2 or greater)
# -*- coding: utf-8 -*- from django.contrib import admin from powerdns.models import Domain, Record, Supermaster class RecordAdmin(admin.ModelAdmin): list_display = ('name', 'type', 'content', 'ttl', 'prio', 'change_date',) list_filter = ['type', 'ttl',] search_fields = ('name','content',) readonly_fields = ('change_date',) class DomainAdmin(admin.ModelAdmin): list_display = ('name', 'type', 'last_check', 'account',) list_filter = ['type', 'last_check', 'account',] search_fields = ('name',) readonly_fields = ('notified_serial',) class SupermasterAdmin(admin.ModelAdmin): list_display = ('ip', 'nameserver', 'account',) list_filter = ['ip', 'account',] search_fields = ('ip', 'nameserver',) admin.site.register(Domain,DomainAdmin) admin.site.register(Record,RecordAdmin) admin.site.register(Supermaster,SupermasterAdmin)
<commit_before># -*- coding: utf-8 -*- from django.contrib import admin from powerdns.models import Domain, Record, Supermaster class RecordAdmin(admin.ModelAdmin): list_display = ('name', 'type', 'content', 'ttl', 'prio', 'change_date',) list_filter = ['type', 'ttl',] search_fields = ('name','content',) class DomainAdmin(admin.ModelAdmin): list_display = ('name', 'type', 'last_check', 'account',) list_filter = ['type', 'last_check', 'account',] search_fields = ('name',) class SupermasterAdmin(admin.ModelAdmin): list_display = ('ip', 'nameserver', 'account',) list_filter = ['ip', 'account',] search_fields = ('ip', 'nameserver',) admin.site.register(Domain,DomainAdmin) admin.site.register(Record,RecordAdmin) admin.site.register(Supermaster,SupermasterAdmin) <commit_msg>Mark some fields as readonly in the Admin panel as they are set by the server (Requires Django 1.2 or greater)<commit_after>
# -*- coding: utf-8 -*- from django.contrib import admin from powerdns.models import Domain, Record, Supermaster class RecordAdmin(admin.ModelAdmin): list_display = ('name', 'type', 'content', 'ttl', 'prio', 'change_date',) list_filter = ['type', 'ttl',] search_fields = ('name','content',) readonly_fields = ('change_date',) class DomainAdmin(admin.ModelAdmin): list_display = ('name', 'type', 'last_check', 'account',) list_filter = ['type', 'last_check', 'account',] search_fields = ('name',) readonly_fields = ('notified_serial',) class SupermasterAdmin(admin.ModelAdmin): list_display = ('ip', 'nameserver', 'account',) list_filter = ['ip', 'account',] search_fields = ('ip', 'nameserver',) admin.site.register(Domain,DomainAdmin) admin.site.register(Record,RecordAdmin) admin.site.register(Supermaster,SupermasterAdmin)
# -*- coding: utf-8 -*- from django.contrib import admin from powerdns.models import Domain, Record, Supermaster class RecordAdmin(admin.ModelAdmin): list_display = ('name', 'type', 'content', 'ttl', 'prio', 'change_date',) list_filter = ['type', 'ttl',] search_fields = ('name','content',) class DomainAdmin(admin.ModelAdmin): list_display = ('name', 'type', 'last_check', 'account',) list_filter = ['type', 'last_check', 'account',] search_fields = ('name',) class SupermasterAdmin(admin.ModelAdmin): list_display = ('ip', 'nameserver', 'account',) list_filter = ['ip', 'account',] search_fields = ('ip', 'nameserver',) admin.site.register(Domain,DomainAdmin) admin.site.register(Record,RecordAdmin) admin.site.register(Supermaster,SupermasterAdmin) Mark some fields as readonly in the Admin panel as they are set by the server (Requires Django 1.2 or greater)# -*- coding: utf-8 -*- from django.contrib import admin from powerdns.models import Domain, Record, Supermaster class RecordAdmin(admin.ModelAdmin): list_display = ('name', 'type', 'content', 'ttl', 'prio', 'change_date',) list_filter = ['type', 'ttl',] search_fields = ('name','content',) readonly_fields = ('change_date',) class DomainAdmin(admin.ModelAdmin): list_display = ('name', 'type', 'last_check', 'account',) list_filter = ['type', 'last_check', 'account',] search_fields = ('name',) readonly_fields = ('notified_serial',) class SupermasterAdmin(admin.ModelAdmin): list_display = ('ip', 'nameserver', 'account',) list_filter = ['ip', 'account',] search_fields = ('ip', 'nameserver',) admin.site.register(Domain,DomainAdmin) admin.site.register(Record,RecordAdmin) admin.site.register(Supermaster,SupermasterAdmin)
<commit_before># -*- coding: utf-8 -*- from django.contrib import admin from powerdns.models import Domain, Record, Supermaster class RecordAdmin(admin.ModelAdmin): list_display = ('name', 'type', 'content', 'ttl', 'prio', 'change_date',) list_filter = ['type', 'ttl',] search_fields = ('name','content',) class DomainAdmin(admin.ModelAdmin): list_display = ('name', 'type', 'last_check', 'account',) list_filter = ['type', 'last_check', 'account',] search_fields = ('name',) class SupermasterAdmin(admin.ModelAdmin): list_display = ('ip', 'nameserver', 'account',) list_filter = ['ip', 'account',] search_fields = ('ip', 'nameserver',) admin.site.register(Domain,DomainAdmin) admin.site.register(Record,RecordAdmin) admin.site.register(Supermaster,SupermasterAdmin) <commit_msg>Mark some fields as readonly in the Admin panel as they are set by the server (Requires Django 1.2 or greater)<commit_after># -*- coding: utf-8 -*- from django.contrib import admin from powerdns.models import Domain, Record, Supermaster class RecordAdmin(admin.ModelAdmin): list_display = ('name', 'type', 'content', 'ttl', 'prio', 'change_date',) list_filter = ['type', 'ttl',] search_fields = ('name','content',) readonly_fields = ('change_date',) class DomainAdmin(admin.ModelAdmin): list_display = ('name', 'type', 'last_check', 'account',) list_filter = ['type', 'last_check', 'account',] search_fields = ('name',) readonly_fields = ('notified_serial',) class SupermasterAdmin(admin.ModelAdmin): list_display = ('ip', 'nameserver', 'account',) list_filter = ['ip', 'account',] search_fields = ('ip', 'nameserver',) admin.site.register(Domain,DomainAdmin) admin.site.register(Record,RecordAdmin) admin.site.register(Supermaster,SupermasterAdmin)
59854731162c25c75304798b62f6541bb9646eee
flask_boost/project/config/__init__.py
flask_boost/project/config/__init__.py
# coding: UTF-8 import os def load_config(): """Load config.""" mode = os.environ.get('MODE') try: if mode == 'PRODUCTION': from .production import ProductionConfig return ProductionConfig elif mode == 'TESTING': from .testing import TestingConfig return TestingConfig else: from .development import DevelopmentConfig return DevelopmentConfig except ImportError, e: from .default import Config return Config
# coding: UTF-8 import os def load_config(): """Load config.""" mode = os.environ.get('MODE') try: if mode == 'PRODUCTION': from .production import ProductionConfig return ProductionConfig elif mode == 'TESTING': from .testing import TestingConfig return TestingConfig else: from .development import DevelopmentConfig return DevelopmentConfig except ImportError: from .default import Config return Config
Update except statements to support Python3.
Update except statements to support Python3.
Python
mit
hustlzp/Flask-Boost,hustlzp/Flask-Boost,hustlzp/Flask-Boost,1045347128/Flask-Boost,hustlzp/Flask-Boost,1045347128/Flask-Boost,1045347128/Flask-Boost,1045347128/Flask-Boost
# coding: UTF-8 import os def load_config(): """Load config.""" mode = os.environ.get('MODE') try: if mode == 'PRODUCTION': from .production import ProductionConfig return ProductionConfig elif mode == 'TESTING': from .testing import TestingConfig return TestingConfig else: from .development import DevelopmentConfig return DevelopmentConfig except ImportError, e: from .default import Config return ConfigUpdate except statements to support Python3.
# coding: UTF-8 import os def load_config(): """Load config.""" mode = os.environ.get('MODE') try: if mode == 'PRODUCTION': from .production import ProductionConfig return ProductionConfig elif mode == 'TESTING': from .testing import TestingConfig return TestingConfig else: from .development import DevelopmentConfig return DevelopmentConfig except ImportError: from .default import Config return Config
<commit_before># coding: UTF-8 import os def load_config(): """Load config.""" mode = os.environ.get('MODE') try: if mode == 'PRODUCTION': from .production import ProductionConfig return ProductionConfig elif mode == 'TESTING': from .testing import TestingConfig return TestingConfig else: from .development import DevelopmentConfig return DevelopmentConfig except ImportError, e: from .default import Config return Config<commit_msg>Update except statements to support Python3.<commit_after>
# coding: UTF-8 import os def load_config(): """Load config.""" mode = os.environ.get('MODE') try: if mode == 'PRODUCTION': from .production import ProductionConfig return ProductionConfig elif mode == 'TESTING': from .testing import TestingConfig return TestingConfig else: from .development import DevelopmentConfig return DevelopmentConfig except ImportError: from .default import Config return Config
# coding: UTF-8 import os def load_config(): """Load config.""" mode = os.environ.get('MODE') try: if mode == 'PRODUCTION': from .production import ProductionConfig return ProductionConfig elif mode == 'TESTING': from .testing import TestingConfig return TestingConfig else: from .development import DevelopmentConfig return DevelopmentConfig except ImportError, e: from .default import Config return ConfigUpdate except statements to support Python3.# coding: UTF-8 import os def load_config(): """Load config.""" mode = os.environ.get('MODE') try: if mode == 'PRODUCTION': from .production import ProductionConfig return ProductionConfig elif mode == 'TESTING': from .testing import TestingConfig return TestingConfig else: from .development import DevelopmentConfig return DevelopmentConfig except ImportError: from .default import Config return Config
<commit_before># coding: UTF-8 import os def load_config(): """Load config.""" mode = os.environ.get('MODE') try: if mode == 'PRODUCTION': from .production import ProductionConfig return ProductionConfig elif mode == 'TESTING': from .testing import TestingConfig return TestingConfig else: from .development import DevelopmentConfig return DevelopmentConfig except ImportError, e: from .default import Config return Config<commit_msg>Update except statements to support Python3.<commit_after># coding: UTF-8 import os def load_config(): """Load config.""" mode = os.environ.get('MODE') try: if mode == 'PRODUCTION': from .production import ProductionConfig return ProductionConfig elif mode == 'TESTING': from .testing import TestingConfig return TestingConfig else: from .development import DevelopmentConfig return DevelopmentConfig except ImportError: from .default import Config return Config
cda678a3ea9b01a684c6e2d3ed636834f41c96f8
wkhtmltopdf/test_settings.py
wkhtmltopdf/test_settings.py
import os DEBUG = True DIRNAME = os.path.abspath(os.path.dirname(__file__)) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': ':memory:', } } # MEDIA_URL = '/media/' # MEDIA_ROOT = '/tmp/media' # STATIC_URL = '/static/' # STATIC_ROOT = '/tmp/static' MEDIA_ROOT = os.path.join(DIRNAME, 'media') MEDIA_URL = '/media/' STATIC_ROOT = os.path.join(DIRNAME, 'static') STATIC_URL = '/static/' INSTALLED_APPS = ( 'wkhtmltopdf.tests', 'wkhtmltopdf', ) TEMPLATE_DIRS = [ os.path.join(DIRNAME, 'testproject', 'tests', 'templates'), ] WKHTMLTOPDF_DEBUG = DEBUG
import os DEBUG = True DIRNAME = os.path.abspath(os.path.dirname(__file__)) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': ':memory:', } } MEDIA_ROOT = os.path.join(DIRNAME, 'media') MEDIA_URL = '/media/' STATIC_ROOT = os.path.join(DIRNAME, 'static') STATIC_URL = '/static/' INSTALLED_APPS = ( 'wkhtmltopdf.tests', 'wkhtmltopdf', ) TEMPLATE_DIRS = [ os.path.join(DIRNAME, 'testproject', 'tests', 'templates'), ] WKHTMLTOPDF_DEBUG = DEBUG
Remove unwanted settings vars from test project
Remove unwanted settings vars from test project
Python
bsd-2-clause
powderflask/django-wkhtmltopdf,denisbalyko/django-wkhtmltopdf,halfnibble/django-wkhtmltopdf,halfnibble/django-wkhtmltopdf,denisbalyko/django-wkhtmltopdf,unrealsolver/django-wkhtmltopdf,tclancy/django-wkhtmltopdf,incuna/django-wkhtmltopdf,powderflask/django-wkhtmltopdf,unrealsolver/django-wkhtmltopdf,fankcoder/django-wkhtmltopdf,incuna/django-wkhtmltopdf,fankcoder/django-wkhtmltopdf,tclancy/django-wkhtmltopdf
import os DEBUG = True DIRNAME = os.path.abspath(os.path.dirname(__file__)) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': ':memory:', } } # MEDIA_URL = '/media/' # MEDIA_ROOT = '/tmp/media' # STATIC_URL = '/static/' # STATIC_ROOT = '/tmp/static' MEDIA_ROOT = os.path.join(DIRNAME, 'media') MEDIA_URL = '/media/' STATIC_ROOT = os.path.join(DIRNAME, 'static') STATIC_URL = '/static/' INSTALLED_APPS = ( 'wkhtmltopdf.tests', 'wkhtmltopdf', ) TEMPLATE_DIRS = [ os.path.join(DIRNAME, 'testproject', 'tests', 'templates'), ] WKHTMLTOPDF_DEBUG = DEBUG Remove unwanted settings vars from test project
import os DEBUG = True DIRNAME = os.path.abspath(os.path.dirname(__file__)) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': ':memory:', } } MEDIA_ROOT = os.path.join(DIRNAME, 'media') MEDIA_URL = '/media/' STATIC_ROOT = os.path.join(DIRNAME, 'static') STATIC_URL = '/static/' INSTALLED_APPS = ( 'wkhtmltopdf.tests', 'wkhtmltopdf', ) TEMPLATE_DIRS = [ os.path.join(DIRNAME, 'testproject', 'tests', 'templates'), ] WKHTMLTOPDF_DEBUG = DEBUG
<commit_before>import os DEBUG = True DIRNAME = os.path.abspath(os.path.dirname(__file__)) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': ':memory:', } } # MEDIA_URL = '/media/' # MEDIA_ROOT = '/tmp/media' # STATIC_URL = '/static/' # STATIC_ROOT = '/tmp/static' MEDIA_ROOT = os.path.join(DIRNAME, 'media') MEDIA_URL = '/media/' STATIC_ROOT = os.path.join(DIRNAME, 'static') STATIC_URL = '/static/' INSTALLED_APPS = ( 'wkhtmltopdf.tests', 'wkhtmltopdf', ) TEMPLATE_DIRS = [ os.path.join(DIRNAME, 'testproject', 'tests', 'templates'), ] WKHTMLTOPDF_DEBUG = DEBUG <commit_msg>Remove unwanted settings vars from test project<commit_after>
import os DEBUG = True DIRNAME = os.path.abspath(os.path.dirname(__file__)) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': ':memory:', } } MEDIA_ROOT = os.path.join(DIRNAME, 'media') MEDIA_URL = '/media/' STATIC_ROOT = os.path.join(DIRNAME, 'static') STATIC_URL = '/static/' INSTALLED_APPS = ( 'wkhtmltopdf.tests', 'wkhtmltopdf', ) TEMPLATE_DIRS = [ os.path.join(DIRNAME, 'testproject', 'tests', 'templates'), ] WKHTMLTOPDF_DEBUG = DEBUG
import os DEBUG = True DIRNAME = os.path.abspath(os.path.dirname(__file__)) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': ':memory:', } } # MEDIA_URL = '/media/' # MEDIA_ROOT = '/tmp/media' # STATIC_URL = '/static/' # STATIC_ROOT = '/tmp/static' MEDIA_ROOT = os.path.join(DIRNAME, 'media') MEDIA_URL = '/media/' STATIC_ROOT = os.path.join(DIRNAME, 'static') STATIC_URL = '/static/' INSTALLED_APPS = ( 'wkhtmltopdf.tests', 'wkhtmltopdf', ) TEMPLATE_DIRS = [ os.path.join(DIRNAME, 'testproject', 'tests', 'templates'), ] WKHTMLTOPDF_DEBUG = DEBUG Remove unwanted settings vars from test projectimport os DEBUG = True DIRNAME = os.path.abspath(os.path.dirname(__file__)) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': ':memory:', } } MEDIA_ROOT = os.path.join(DIRNAME, 'media') MEDIA_URL = '/media/' STATIC_ROOT = os.path.join(DIRNAME, 'static') STATIC_URL = '/static/' INSTALLED_APPS = ( 'wkhtmltopdf.tests', 'wkhtmltopdf', ) TEMPLATE_DIRS = [ os.path.join(DIRNAME, 'testproject', 'tests', 'templates'), ] WKHTMLTOPDF_DEBUG = DEBUG
<commit_before>import os DEBUG = True DIRNAME = os.path.abspath(os.path.dirname(__file__)) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': ':memory:', } } # MEDIA_URL = '/media/' # MEDIA_ROOT = '/tmp/media' # STATIC_URL = '/static/' # STATIC_ROOT = '/tmp/static' MEDIA_ROOT = os.path.join(DIRNAME, 'media') MEDIA_URL = '/media/' STATIC_ROOT = os.path.join(DIRNAME, 'static') STATIC_URL = '/static/' INSTALLED_APPS = ( 'wkhtmltopdf.tests', 'wkhtmltopdf', ) TEMPLATE_DIRS = [ os.path.join(DIRNAME, 'testproject', 'tests', 'templates'), ] WKHTMLTOPDF_DEBUG = DEBUG <commit_msg>Remove unwanted settings vars from test project<commit_after>import os DEBUG = True DIRNAME = os.path.abspath(os.path.dirname(__file__)) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': ':memory:', } } MEDIA_ROOT = os.path.join(DIRNAME, 'media') MEDIA_URL = '/media/' STATIC_ROOT = os.path.join(DIRNAME, 'static') STATIC_URL = '/static/' INSTALLED_APPS = ( 'wkhtmltopdf.tests', 'wkhtmltopdf', ) TEMPLATE_DIRS = [ os.path.join(DIRNAME, 'testproject', 'tests', 'templates'), ] WKHTMLTOPDF_DEBUG = DEBUG
bb3ffdea2a76a86c1911426ce030c29abbd1074f
tempo/django/forms.py
tempo/django/forms.py
#!/usr/bin/env python # coding=utf-8 from decimal import Decimal from django.utils.translation import ugettext_lazy as _ from django.forms.fields import Field, ValidationError from tempo.django.widgets import ScheduleSetWidget from tempo.schedule import Schedule from tempo.scheduleset import ScheduleSet class ScheduleSetField(Field): widget = ScheduleSetWidget def to_python(self, value): if value is None: return None repeats = value['repeats'] if repeats == 'monthly': return ScheduleSet(include=[ Schedule(days=[int(value['repeatOn'])]) ]) elif repeats == 'weekly': schedules = [] for repeat_on in value['repeatOn']: if Decimal(repeat_on['from']) > Decimal(repeat_on['to']): raise ValidationError(_('"From" is greater than "to".'), code='invalid') schedule = Schedule( weekdays=[int(repeat_on['weekday'])], days=[], seconds_of_the_day=list(range( int(Decimal(repeat_on['from']) * 60 * 60), int(Decimal(repeat_on['to']) * 60 * 60) + 1) ), seconds=[], minutes=[], hours=[] ) schedules.append(schedule) return ScheduleSet(include=schedules) else: raise ValueError
#!/usr/bin/env python # coding=utf-8 from decimal import Decimal from django.utils.translation import ugettext_lazy as _ from django.forms.fields import Field, ValidationError from tempo.django.widgets import ScheduleSetWidget from tempo.schedule import Schedule from tempo.scheduleset import ScheduleSet class ScheduleSetField(Field): widget = ScheduleSetWidget def to_python(self, value): if value is None: return None repeats = value['repeats'] if repeats == 'monthly': return ScheduleSet(include=[ Schedule(days=[int(value['repeatOn'])]) ]) elif repeats == 'weekly': schedules = [] for repeat_on in value['repeatOn']: if Decimal(repeat_on['from']) >= Decimal(repeat_on['to']): raise ValidationError(_('"From" is greater than "to" ' 'or equal to it.'), code='invalid') schedule = Schedule( weekdays=[int(repeat_on['weekday'])], days=[], seconds_of_the_day=list(range( int(Decimal(repeat_on['from']) * 60 * 60), int(Decimal(repeat_on['to']) * 60 * 60)) ), seconds=[], minutes=[], hours=[] ) schedules.append(schedule) return ScheduleSet(include=schedules) else: raise ValueError
Make "to" not inclusive in widget
Make "to" not inclusive in widget To avoid exceeding possible values.
Python
bsd-3-clause
AndrewPashkin/python-tempo,AndrewPashkin/python-tempo
#!/usr/bin/env python # coding=utf-8 from decimal import Decimal from django.utils.translation import ugettext_lazy as _ from django.forms.fields import Field, ValidationError from tempo.django.widgets import ScheduleSetWidget from tempo.schedule import Schedule from tempo.scheduleset import ScheduleSet class ScheduleSetField(Field): widget = ScheduleSetWidget def to_python(self, value): if value is None: return None repeats = value['repeats'] if repeats == 'monthly': return ScheduleSet(include=[ Schedule(days=[int(value['repeatOn'])]) ]) elif repeats == 'weekly': schedules = [] for repeat_on in value['repeatOn']: if Decimal(repeat_on['from']) > Decimal(repeat_on['to']): raise ValidationError(_('"From" is greater than "to".'), code='invalid') schedule = Schedule( weekdays=[int(repeat_on['weekday'])], days=[], seconds_of_the_day=list(range( int(Decimal(repeat_on['from']) * 60 * 60), int(Decimal(repeat_on['to']) * 60 * 60) + 1) ), seconds=[], minutes=[], hours=[] ) schedules.append(schedule) return ScheduleSet(include=schedules) else: raise ValueError Make "to" not inclusive in widget To avoid exceeding possible values.
#!/usr/bin/env python # coding=utf-8 from decimal import Decimal from django.utils.translation import ugettext_lazy as _ from django.forms.fields import Field, ValidationError from tempo.django.widgets import ScheduleSetWidget from tempo.schedule import Schedule from tempo.scheduleset import ScheduleSet class ScheduleSetField(Field): widget = ScheduleSetWidget def to_python(self, value): if value is None: return None repeats = value['repeats'] if repeats == 'monthly': return ScheduleSet(include=[ Schedule(days=[int(value['repeatOn'])]) ]) elif repeats == 'weekly': schedules = [] for repeat_on in value['repeatOn']: if Decimal(repeat_on['from']) >= Decimal(repeat_on['to']): raise ValidationError(_('"From" is greater than "to" ' 'or equal to it.'), code='invalid') schedule = Schedule( weekdays=[int(repeat_on['weekday'])], days=[], seconds_of_the_day=list(range( int(Decimal(repeat_on['from']) * 60 * 60), int(Decimal(repeat_on['to']) * 60 * 60)) ), seconds=[], minutes=[], hours=[] ) schedules.append(schedule) return ScheduleSet(include=schedules) else: raise ValueError
<commit_before>#!/usr/bin/env python # coding=utf-8 from decimal import Decimal from django.utils.translation import ugettext_lazy as _ from django.forms.fields import Field, ValidationError from tempo.django.widgets import ScheduleSetWidget from tempo.schedule import Schedule from tempo.scheduleset import ScheduleSet class ScheduleSetField(Field): widget = ScheduleSetWidget def to_python(self, value): if value is None: return None repeats = value['repeats'] if repeats == 'monthly': return ScheduleSet(include=[ Schedule(days=[int(value['repeatOn'])]) ]) elif repeats == 'weekly': schedules = [] for repeat_on in value['repeatOn']: if Decimal(repeat_on['from']) > Decimal(repeat_on['to']): raise ValidationError(_('"From" is greater than "to".'), code='invalid') schedule = Schedule( weekdays=[int(repeat_on['weekday'])], days=[], seconds_of_the_day=list(range( int(Decimal(repeat_on['from']) * 60 * 60), int(Decimal(repeat_on['to']) * 60 * 60) + 1) ), seconds=[], minutes=[], hours=[] ) schedules.append(schedule) return ScheduleSet(include=schedules) else: raise ValueError <commit_msg>Make "to" not inclusive in widget To avoid exceeding possible values.<commit_after>
#!/usr/bin/env python # coding=utf-8 from decimal import Decimal from django.utils.translation import ugettext_lazy as _ from django.forms.fields import Field, ValidationError from tempo.django.widgets import ScheduleSetWidget from tempo.schedule import Schedule from tempo.scheduleset import ScheduleSet class ScheduleSetField(Field): widget = ScheduleSetWidget def to_python(self, value): if value is None: return None repeats = value['repeats'] if repeats == 'monthly': return ScheduleSet(include=[ Schedule(days=[int(value['repeatOn'])]) ]) elif repeats == 'weekly': schedules = [] for repeat_on in value['repeatOn']: if Decimal(repeat_on['from']) >= Decimal(repeat_on['to']): raise ValidationError(_('"From" is greater than "to" ' 'or equal to it.'), code='invalid') schedule = Schedule( weekdays=[int(repeat_on['weekday'])], days=[], seconds_of_the_day=list(range( int(Decimal(repeat_on['from']) * 60 * 60), int(Decimal(repeat_on['to']) * 60 * 60)) ), seconds=[], minutes=[], hours=[] ) schedules.append(schedule) return ScheduleSet(include=schedules) else: raise ValueError
#!/usr/bin/env python # coding=utf-8 from decimal import Decimal from django.utils.translation import ugettext_lazy as _ from django.forms.fields import Field, ValidationError from tempo.django.widgets import ScheduleSetWidget from tempo.schedule import Schedule from tempo.scheduleset import ScheduleSet class ScheduleSetField(Field): widget = ScheduleSetWidget def to_python(self, value): if value is None: return None repeats = value['repeats'] if repeats == 'monthly': return ScheduleSet(include=[ Schedule(days=[int(value['repeatOn'])]) ]) elif repeats == 'weekly': schedules = [] for repeat_on in value['repeatOn']: if Decimal(repeat_on['from']) > Decimal(repeat_on['to']): raise ValidationError(_('"From" is greater than "to".'), code='invalid') schedule = Schedule( weekdays=[int(repeat_on['weekday'])], days=[], seconds_of_the_day=list(range( int(Decimal(repeat_on['from']) * 60 * 60), int(Decimal(repeat_on['to']) * 60 * 60) + 1) ), seconds=[], minutes=[], hours=[] ) schedules.append(schedule) return ScheduleSet(include=schedules) else: raise ValueError Make "to" not inclusive in widget To avoid exceeding possible values.#!/usr/bin/env python # coding=utf-8 from decimal import Decimal from django.utils.translation import ugettext_lazy as _ from django.forms.fields import Field, ValidationError from tempo.django.widgets import ScheduleSetWidget from tempo.schedule import Schedule from tempo.scheduleset import ScheduleSet class ScheduleSetField(Field): widget = ScheduleSetWidget def to_python(self, value): if value is None: return None repeats = value['repeats'] if repeats == 'monthly': return ScheduleSet(include=[ Schedule(days=[int(value['repeatOn'])]) ]) elif repeats == 'weekly': schedules = [] for repeat_on in value['repeatOn']: if Decimal(repeat_on['from']) >= Decimal(repeat_on['to']): raise ValidationError(_('"From" is greater than "to" ' 'or equal to it.'), code='invalid') schedule = Schedule( weekdays=[int(repeat_on['weekday'])], days=[], seconds_of_the_day=list(range( int(Decimal(repeat_on['from']) * 60 * 60), int(Decimal(repeat_on['to']) * 60 * 60)) ), seconds=[], minutes=[], hours=[] ) schedules.append(schedule) return ScheduleSet(include=schedules) else: raise ValueError
<commit_before>#!/usr/bin/env python # coding=utf-8 from decimal import Decimal from django.utils.translation import ugettext_lazy as _ from django.forms.fields import Field, ValidationError from tempo.django.widgets import ScheduleSetWidget from tempo.schedule import Schedule from tempo.scheduleset import ScheduleSet class ScheduleSetField(Field): widget = ScheduleSetWidget def to_python(self, value): if value is None: return None repeats = value['repeats'] if repeats == 'monthly': return ScheduleSet(include=[ Schedule(days=[int(value['repeatOn'])]) ]) elif repeats == 'weekly': schedules = [] for repeat_on in value['repeatOn']: if Decimal(repeat_on['from']) > Decimal(repeat_on['to']): raise ValidationError(_('"From" is greater than "to".'), code='invalid') schedule = Schedule( weekdays=[int(repeat_on['weekday'])], days=[], seconds_of_the_day=list(range( int(Decimal(repeat_on['from']) * 60 * 60), int(Decimal(repeat_on['to']) * 60 * 60) + 1) ), seconds=[], minutes=[], hours=[] ) schedules.append(schedule) return ScheduleSet(include=schedules) else: raise ValueError <commit_msg>Make "to" not inclusive in widget To avoid exceeding possible values.<commit_after>#!/usr/bin/env python # coding=utf-8 from decimal import Decimal from django.utils.translation import ugettext_lazy as _ from django.forms.fields import Field, ValidationError from tempo.django.widgets import ScheduleSetWidget from tempo.schedule import Schedule from tempo.scheduleset import ScheduleSet class ScheduleSetField(Field): widget = ScheduleSetWidget def to_python(self, value): if value is None: return None repeats = value['repeats'] if repeats == 'monthly': return ScheduleSet(include=[ Schedule(days=[int(value['repeatOn'])]) ]) elif repeats == 'weekly': schedules = [] for repeat_on in value['repeatOn']: if Decimal(repeat_on['from']) >= Decimal(repeat_on['to']): raise ValidationError(_('"From" is greater than "to" ' 'or equal to it.'), code='invalid') schedule = Schedule( weekdays=[int(repeat_on['weekday'])], days=[], seconds_of_the_day=list(range( int(Decimal(repeat_on['from']) * 60 * 60), int(Decimal(repeat_on['to']) * 60 * 60)) ), seconds=[], minutes=[], hours=[] ) schedules.append(schedule) return ScheduleSet(include=schedules) else: raise ValueError
f58b24402d8d41778a45d4ce6e6080651c79d529
scripts/common.py
scripts/common.py
#!/usr/bin/env python # # file: common.py # # author: Copyright (C) 2017-2018 Kamil Szczygiel http://www.distortec.com http://www.freddiechopin.info # # This Source Code Form is subject to the terms of the Mozilla Public License, v. 2.0. If a copy of the MPL was not # distributed with this file, You can obtain one at http://mozilla.org/MPL/2.0/. # import re def sanitize(string, pattern = '[^0-9A-Za-z-]', replacement = '_'): """Sanitize and return provided string. * `string` is the string that will be sanitized * `pattern` is the pattern which will be replaced, default - `'[^0-9A-Za-z-]'` * `replacement` is the replacement string, default - `'_'` """ regex = re.compile(pattern) return regex.sub(replacement, string)
#!/usr/bin/env python # # file: common.py # # author: Copyright (C) 2017-2018 Kamil Szczygiel http://www.distortec.com http://www.freddiechopin.info # # This Source Code Form is subject to the terms of the Mozilla Public License, v. 2.0. If a copy of the MPL was not # distributed with this file, You can obtain one at http://mozilla.org/MPL/2.0/. # import re def sanitize(string, pattern = '[^0-9A-Za-z-]', replacement = '_'): """Sanitize and return provided string. * `string` is the string that will be sanitized * `pattern` is the pattern which will be replaced, default - `'[^0-9A-Za-z-]'` * `replacement` is the replacement string, default - `'_'` """ regex = re.compile(pattern) return regex.sub(replacement, str(string))
Convert sanitize() argument to string
Convert sanitize() argument to string This fixes the bug when the name of the node is an integer, in which case it's not possible to do regex.sub() on it.
Python
mpl-2.0
CezaryGapinski/distortos,CezaryGapinski/distortos,jasmin-j/distortos,jasmin-j/distortos,jasmin-j/distortos,DISTORTEC/distortos,CezaryGapinski/distortos,DISTORTEC/distortos,CezaryGapinski/distortos,CezaryGapinski/distortos,DISTORTEC/distortos,jasmin-j/distortos,jasmin-j/distortos,DISTORTEC/distortos
#!/usr/bin/env python # # file: common.py # # author: Copyright (C) 2017-2018 Kamil Szczygiel http://www.distortec.com http://www.freddiechopin.info # # This Source Code Form is subject to the terms of the Mozilla Public License, v. 2.0. If a copy of the MPL was not # distributed with this file, You can obtain one at http://mozilla.org/MPL/2.0/. # import re def sanitize(string, pattern = '[^0-9A-Za-z-]', replacement = '_'): """Sanitize and return provided string. * `string` is the string that will be sanitized * `pattern` is the pattern which will be replaced, default - `'[^0-9A-Za-z-]'` * `replacement` is the replacement string, default - `'_'` """ regex = re.compile(pattern) return regex.sub(replacement, string) Convert sanitize() argument to string This fixes the bug when the name of the node is an integer, in which case it's not possible to do regex.sub() on it.
#!/usr/bin/env python # # file: common.py # # author: Copyright (C) 2017-2018 Kamil Szczygiel http://www.distortec.com http://www.freddiechopin.info # # This Source Code Form is subject to the terms of the Mozilla Public License, v. 2.0. If a copy of the MPL was not # distributed with this file, You can obtain one at http://mozilla.org/MPL/2.0/. # import re def sanitize(string, pattern = '[^0-9A-Za-z-]', replacement = '_'): """Sanitize and return provided string. * `string` is the string that will be sanitized * `pattern` is the pattern which will be replaced, default - `'[^0-9A-Za-z-]'` * `replacement` is the replacement string, default - `'_'` """ regex = re.compile(pattern) return regex.sub(replacement, str(string))
<commit_before>#!/usr/bin/env python # # file: common.py # # author: Copyright (C) 2017-2018 Kamil Szczygiel http://www.distortec.com http://www.freddiechopin.info # # This Source Code Form is subject to the terms of the Mozilla Public License, v. 2.0. If a copy of the MPL was not # distributed with this file, You can obtain one at http://mozilla.org/MPL/2.0/. # import re def sanitize(string, pattern = '[^0-9A-Za-z-]', replacement = '_'): """Sanitize and return provided string. * `string` is the string that will be sanitized * `pattern` is the pattern which will be replaced, default - `'[^0-9A-Za-z-]'` * `replacement` is the replacement string, default - `'_'` """ regex = re.compile(pattern) return regex.sub(replacement, string) <commit_msg>Convert sanitize() argument to string This fixes the bug when the name of the node is an integer, in which case it's not possible to do regex.sub() on it.<commit_after>
#!/usr/bin/env python # # file: common.py # # author: Copyright (C) 2017-2018 Kamil Szczygiel http://www.distortec.com http://www.freddiechopin.info # # This Source Code Form is subject to the terms of the Mozilla Public License, v. 2.0. If a copy of the MPL was not # distributed with this file, You can obtain one at http://mozilla.org/MPL/2.0/. # import re def sanitize(string, pattern = '[^0-9A-Za-z-]', replacement = '_'): """Sanitize and return provided string. * `string` is the string that will be sanitized * `pattern` is the pattern which will be replaced, default - `'[^0-9A-Za-z-]'` * `replacement` is the replacement string, default - `'_'` """ regex = re.compile(pattern) return regex.sub(replacement, str(string))
#!/usr/bin/env python # # file: common.py # # author: Copyright (C) 2017-2018 Kamil Szczygiel http://www.distortec.com http://www.freddiechopin.info # # This Source Code Form is subject to the terms of the Mozilla Public License, v. 2.0. If a copy of the MPL was not # distributed with this file, You can obtain one at http://mozilla.org/MPL/2.0/. # import re def sanitize(string, pattern = '[^0-9A-Za-z-]', replacement = '_'): """Sanitize and return provided string. * `string` is the string that will be sanitized * `pattern` is the pattern which will be replaced, default - `'[^0-9A-Za-z-]'` * `replacement` is the replacement string, default - `'_'` """ regex = re.compile(pattern) return regex.sub(replacement, string) Convert sanitize() argument to string This fixes the bug when the name of the node is an integer, in which case it's not possible to do regex.sub() on it.#!/usr/bin/env python # # file: common.py # # author: Copyright (C) 2017-2018 Kamil Szczygiel http://www.distortec.com http://www.freddiechopin.info # # This Source Code Form is subject to the terms of the Mozilla Public License, v. 2.0. If a copy of the MPL was not # distributed with this file, You can obtain one at http://mozilla.org/MPL/2.0/. # import re def sanitize(string, pattern = '[^0-9A-Za-z-]', replacement = '_'): """Sanitize and return provided string. * `string` is the string that will be sanitized * `pattern` is the pattern which will be replaced, default - `'[^0-9A-Za-z-]'` * `replacement` is the replacement string, default - `'_'` """ regex = re.compile(pattern) return regex.sub(replacement, str(string))
<commit_before>#!/usr/bin/env python # # file: common.py # # author: Copyright (C) 2017-2018 Kamil Szczygiel http://www.distortec.com http://www.freddiechopin.info # # This Source Code Form is subject to the terms of the Mozilla Public License, v. 2.0. If a copy of the MPL was not # distributed with this file, You can obtain one at http://mozilla.org/MPL/2.0/. # import re def sanitize(string, pattern = '[^0-9A-Za-z-]', replacement = '_'): """Sanitize and return provided string. * `string` is the string that will be sanitized * `pattern` is the pattern which will be replaced, default - `'[^0-9A-Za-z-]'` * `replacement` is the replacement string, default - `'_'` """ regex = re.compile(pattern) return regex.sub(replacement, string) <commit_msg>Convert sanitize() argument to string This fixes the bug when the name of the node is an integer, in which case it's not possible to do regex.sub() on it.<commit_after>#!/usr/bin/env python # # file: common.py # # author: Copyright (C) 2017-2018 Kamil Szczygiel http://www.distortec.com http://www.freddiechopin.info # # This Source Code Form is subject to the terms of the Mozilla Public License, v. 2.0. If a copy of the MPL was not # distributed with this file, You can obtain one at http://mozilla.org/MPL/2.0/. # import re def sanitize(string, pattern = '[^0-9A-Za-z-]', replacement = '_'): """Sanitize and return provided string. * `string` is the string that will be sanitized * `pattern` is the pattern which will be replaced, default - `'[^0-9A-Za-z-]'` * `replacement` is the replacement string, default - `'_'` """ regex = re.compile(pattern) return regex.sub(replacement, str(string))
d66355e4758b37be39d17d681ede1dbbd6b9b311
setmagic/admin.py
setmagic/admin.py
from django import forms from django.contrib import admin from setmagic import settings from setmagic.models import Setting _denied = lambda *args: False class SetMagicAdmin(admin.ModelAdmin): list_display = 'label', 'current_value', list_editable = 'current_value', list_display_links = None has_add_permission = _denied has_delete_permission = _denied # Make all fields read-only at the change form def get_readonly_fields(self, *args, **kwargs): return self.opts.get_all_field_names() def changelist_view(self, *args, **kwargs): settings._sync() return super(SetMagicAdmin, self).changelist_view(*args, **kwargs) def get_queryset(self, request): return Setting.objects.filter(name__in=settings.defs) def get_changelist_form(self, *args, **kwargs): class Form(forms.ModelForm): class Meta: fields = self.list_editable def __init__(self, *args, **kwargs): super(Form, self).__init__(*args, **kwargs) # Do nothing for empty forms if not self.instance.pk: return # Set a custom field custom_field = settings.defs[self.instance.name].get('field') if custom_field: self.fields['current_value'] = custom_field return Form admin.site.register(Setting, SetMagicAdmin)
from django import forms from django.contrib import admin from django.utils.importlib import import_module from setmagic import settings from setmagic.models import Setting _denied = lambda *args: False class SetMagicAdmin(admin.ModelAdmin): list_display = 'label', 'current_value', list_editable = 'current_value', list_display_links = None has_add_permission = _denied has_delete_permission = _denied # Make all fields read-only at the change form def get_readonly_fields(self, *args, **kwargs): return self.opts.get_all_field_names() def changelist_view(self, *args, **kwargs): settings._sync() return super(SetMagicAdmin, self).changelist_view(*args, **kwargs) def get_queryset(self, request): return Setting.objects.filter(name__in=settings.defs) def get_changelist_form(self, *args, **kwargs): class Form(forms.ModelForm): class Meta: fields = self.list_editable def __init__(self, *args, **kwargs): super(Form, self).__init__(*args, **kwargs) # Do nothing for empty forms if not self.instance.pk: return # Set a custom field custom_field = settings.defs[self.instance.name].get('field') if custom_field: if isinstance(custom_field, str): module, name = custom_field.rsplit('.', 1) custom_field = getattr(import_module(module), name)() self.fields['current_value'] = custom_field return Form admin.site.register(Setting, SetMagicAdmin)
Use importlib to load custom fields by str
Use importlib to load custom fields by str
Python
mit
7ws/django-setmagic
from django import forms from django.contrib import admin from setmagic import settings from setmagic.models import Setting _denied = lambda *args: False class SetMagicAdmin(admin.ModelAdmin): list_display = 'label', 'current_value', list_editable = 'current_value', list_display_links = None has_add_permission = _denied has_delete_permission = _denied # Make all fields read-only at the change form def get_readonly_fields(self, *args, **kwargs): return self.opts.get_all_field_names() def changelist_view(self, *args, **kwargs): settings._sync() return super(SetMagicAdmin, self).changelist_view(*args, **kwargs) def get_queryset(self, request): return Setting.objects.filter(name__in=settings.defs) def get_changelist_form(self, *args, **kwargs): class Form(forms.ModelForm): class Meta: fields = self.list_editable def __init__(self, *args, **kwargs): super(Form, self).__init__(*args, **kwargs) # Do nothing for empty forms if not self.instance.pk: return # Set a custom field custom_field = settings.defs[self.instance.name].get('field') if custom_field: self.fields['current_value'] = custom_field return Form admin.site.register(Setting, SetMagicAdmin) Use importlib to load custom fields by str
from django import forms from django.contrib import admin from django.utils.importlib import import_module from setmagic import settings from setmagic.models import Setting _denied = lambda *args: False class SetMagicAdmin(admin.ModelAdmin): list_display = 'label', 'current_value', list_editable = 'current_value', list_display_links = None has_add_permission = _denied has_delete_permission = _denied # Make all fields read-only at the change form def get_readonly_fields(self, *args, **kwargs): return self.opts.get_all_field_names() def changelist_view(self, *args, **kwargs): settings._sync() return super(SetMagicAdmin, self).changelist_view(*args, **kwargs) def get_queryset(self, request): return Setting.objects.filter(name__in=settings.defs) def get_changelist_form(self, *args, **kwargs): class Form(forms.ModelForm): class Meta: fields = self.list_editable def __init__(self, *args, **kwargs): super(Form, self).__init__(*args, **kwargs) # Do nothing for empty forms if not self.instance.pk: return # Set a custom field custom_field = settings.defs[self.instance.name].get('field') if custom_field: if isinstance(custom_field, str): module, name = custom_field.rsplit('.', 1) custom_field = getattr(import_module(module), name)() self.fields['current_value'] = custom_field return Form admin.site.register(Setting, SetMagicAdmin)
<commit_before>from django import forms from django.contrib import admin from setmagic import settings from setmagic.models import Setting _denied = lambda *args: False class SetMagicAdmin(admin.ModelAdmin): list_display = 'label', 'current_value', list_editable = 'current_value', list_display_links = None has_add_permission = _denied has_delete_permission = _denied # Make all fields read-only at the change form def get_readonly_fields(self, *args, **kwargs): return self.opts.get_all_field_names() def changelist_view(self, *args, **kwargs): settings._sync() return super(SetMagicAdmin, self).changelist_view(*args, **kwargs) def get_queryset(self, request): return Setting.objects.filter(name__in=settings.defs) def get_changelist_form(self, *args, **kwargs): class Form(forms.ModelForm): class Meta: fields = self.list_editable def __init__(self, *args, **kwargs): super(Form, self).__init__(*args, **kwargs) # Do nothing for empty forms if not self.instance.pk: return # Set a custom field custom_field = settings.defs[self.instance.name].get('field') if custom_field: self.fields['current_value'] = custom_field return Form admin.site.register(Setting, SetMagicAdmin) <commit_msg>Use importlib to load custom fields by str<commit_after>
from django import forms from django.contrib import admin from django.utils.importlib import import_module from setmagic import settings from setmagic.models import Setting _denied = lambda *args: False class SetMagicAdmin(admin.ModelAdmin): list_display = 'label', 'current_value', list_editable = 'current_value', list_display_links = None has_add_permission = _denied has_delete_permission = _denied # Make all fields read-only at the change form def get_readonly_fields(self, *args, **kwargs): return self.opts.get_all_field_names() def changelist_view(self, *args, **kwargs): settings._sync() return super(SetMagicAdmin, self).changelist_view(*args, **kwargs) def get_queryset(self, request): return Setting.objects.filter(name__in=settings.defs) def get_changelist_form(self, *args, **kwargs): class Form(forms.ModelForm): class Meta: fields = self.list_editable def __init__(self, *args, **kwargs): super(Form, self).__init__(*args, **kwargs) # Do nothing for empty forms if not self.instance.pk: return # Set a custom field custom_field = settings.defs[self.instance.name].get('field') if custom_field: if isinstance(custom_field, str): module, name = custom_field.rsplit('.', 1) custom_field = getattr(import_module(module), name)() self.fields['current_value'] = custom_field return Form admin.site.register(Setting, SetMagicAdmin)
from django import forms from django.contrib import admin from setmagic import settings from setmagic.models import Setting _denied = lambda *args: False class SetMagicAdmin(admin.ModelAdmin): list_display = 'label', 'current_value', list_editable = 'current_value', list_display_links = None has_add_permission = _denied has_delete_permission = _denied # Make all fields read-only at the change form def get_readonly_fields(self, *args, **kwargs): return self.opts.get_all_field_names() def changelist_view(self, *args, **kwargs): settings._sync() return super(SetMagicAdmin, self).changelist_view(*args, **kwargs) def get_queryset(self, request): return Setting.objects.filter(name__in=settings.defs) def get_changelist_form(self, *args, **kwargs): class Form(forms.ModelForm): class Meta: fields = self.list_editable def __init__(self, *args, **kwargs): super(Form, self).__init__(*args, **kwargs) # Do nothing for empty forms if not self.instance.pk: return # Set a custom field custom_field = settings.defs[self.instance.name].get('field') if custom_field: self.fields['current_value'] = custom_field return Form admin.site.register(Setting, SetMagicAdmin) Use importlib to load custom fields by strfrom django import forms from django.contrib import admin from django.utils.importlib import import_module from setmagic import settings from setmagic.models import Setting _denied = lambda *args: False class SetMagicAdmin(admin.ModelAdmin): list_display = 'label', 'current_value', list_editable = 'current_value', list_display_links = None has_add_permission = _denied has_delete_permission = _denied # Make all fields read-only at the change form def get_readonly_fields(self, *args, **kwargs): return self.opts.get_all_field_names() def changelist_view(self, *args, **kwargs): settings._sync() return super(SetMagicAdmin, self).changelist_view(*args, **kwargs) def get_queryset(self, request): return Setting.objects.filter(name__in=settings.defs) def get_changelist_form(self, *args, **kwargs): class Form(forms.ModelForm): class Meta: fields = self.list_editable def __init__(self, *args, **kwargs): super(Form, self).__init__(*args, **kwargs) # Do nothing for empty forms if not self.instance.pk: return # Set a custom field custom_field = settings.defs[self.instance.name].get('field') if custom_field: if isinstance(custom_field, str): module, name = custom_field.rsplit('.', 1) custom_field = getattr(import_module(module), name)() self.fields['current_value'] = custom_field return Form admin.site.register(Setting, SetMagicAdmin)
<commit_before>from django import forms from django.contrib import admin from setmagic import settings from setmagic.models import Setting _denied = lambda *args: False class SetMagicAdmin(admin.ModelAdmin): list_display = 'label', 'current_value', list_editable = 'current_value', list_display_links = None has_add_permission = _denied has_delete_permission = _denied # Make all fields read-only at the change form def get_readonly_fields(self, *args, **kwargs): return self.opts.get_all_field_names() def changelist_view(self, *args, **kwargs): settings._sync() return super(SetMagicAdmin, self).changelist_view(*args, **kwargs) def get_queryset(self, request): return Setting.objects.filter(name__in=settings.defs) def get_changelist_form(self, *args, **kwargs): class Form(forms.ModelForm): class Meta: fields = self.list_editable def __init__(self, *args, **kwargs): super(Form, self).__init__(*args, **kwargs) # Do nothing for empty forms if not self.instance.pk: return # Set a custom field custom_field = settings.defs[self.instance.name].get('field') if custom_field: self.fields['current_value'] = custom_field return Form admin.site.register(Setting, SetMagicAdmin) <commit_msg>Use importlib to load custom fields by str<commit_after>from django import forms from django.contrib import admin from django.utils.importlib import import_module from setmagic import settings from setmagic.models import Setting _denied = lambda *args: False class SetMagicAdmin(admin.ModelAdmin): list_display = 'label', 'current_value', list_editable = 'current_value', list_display_links = None has_add_permission = _denied has_delete_permission = _denied # Make all fields read-only at the change form def get_readonly_fields(self, *args, **kwargs): return self.opts.get_all_field_names() def changelist_view(self, *args, **kwargs): settings._sync() return super(SetMagicAdmin, self).changelist_view(*args, **kwargs) def get_queryset(self, request): return Setting.objects.filter(name__in=settings.defs) def get_changelist_form(self, *args, **kwargs): class Form(forms.ModelForm): class Meta: fields = self.list_editable def __init__(self, *args, **kwargs): super(Form, self).__init__(*args, **kwargs) # Do nothing for empty forms if not self.instance.pk: return # Set a custom field custom_field = settings.defs[self.instance.name].get('field') if custom_field: if isinstance(custom_field, str): module, name = custom_field.rsplit('.', 1) custom_field = getattr(import_module(module), name)() self.fields['current_value'] = custom_field return Form admin.site.register(Setting, SetMagicAdmin)
65137b42bd5ebee37cba0fd462e8f7484a3f9aaa
pelops/etl/hog.py
pelops/etl/hog.py
from skimage.feature import hog from skimage import colos from pelops.etl.feature_producer import FeatureProducer class HOGFeatureProducer(FeatureProducer): def __init__(self, chip_producer): super().__init__(chip_producer) def produce_features(self, chip): """Takes a chip object and returns a feature vector of size self.feat_size. """ img = PIL_Image.open(chip.filepath) img = img.resize((256, 256), PIL_Image.BICUBIC) img_x, img_y = img.size img = color.rgb2gray(np.array(img)) features = hog( img, orientations=8, pixels_per_cell=(int(img_x / 16), int(img_y / 16)), cells_per_block=(16, 16), # Normalize over the whole image ) return features def set_variables(self): """Child classes should use this to set self.feat_size, and any other needed variables. """ self.feat_size = 2048 # Set this in your inherited class
from skimage.feature import hog from skimage import colos from pelops.etl.feature_producer import FeatureProducer class HOGFeatureProducer(FeatureProducer): def __init__(self, chip_producer, image_size=(256,256), cells=(16, 16), orientations=8): super().__init__(chip_producer) self.image_size = image_size self.cells = cells self.orientations = orientations def produce_features(self, chip): """Takes a chip object and returns a feature vector of size self.feat_size. """ img = PIL_Image.open(chip.filepath) img = img.resize(self.image_size, PIL_Image.BICUBIC) img_x, img_y = img.size img = color.rgb2gray(np.array(img)) features = hog( img, orientations=self.orientations, pixels_per_cell=(img_x / cells[0], img_y / cells[1]), cells_per_block=cells, # Normalize over the whole image ) return features def set_variables(self): self.feat_size = self.cells[0] * self.cells[1] * orientations
Replace magic numbers in `HOGFeatureProducer`
Replace magic numbers in `HOGFeatureProducer`
Python
apache-2.0
d-grossman/pelops,d-grossman/pelops,dave-lab41/pelops,Lab41/pelops,dave-lab41/pelops,Lab41/pelops
from skimage.feature import hog from skimage import colos from pelops.etl.feature_producer import FeatureProducer class HOGFeatureProducer(FeatureProducer): def __init__(self, chip_producer): super().__init__(chip_producer) def produce_features(self, chip): """Takes a chip object and returns a feature vector of size self.feat_size. """ img = PIL_Image.open(chip.filepath) img = img.resize((256, 256), PIL_Image.BICUBIC) img_x, img_y = img.size img = color.rgb2gray(np.array(img)) features = hog( img, orientations=8, pixels_per_cell=(int(img_x / 16), int(img_y / 16)), cells_per_block=(16, 16), # Normalize over the whole image ) return features def set_variables(self): """Child classes should use this to set self.feat_size, and any other needed variables. """ self.feat_size = 2048 # Set this in your inherited class Replace magic numbers in `HOGFeatureProducer`
from skimage.feature import hog from skimage import colos from pelops.etl.feature_producer import FeatureProducer class HOGFeatureProducer(FeatureProducer): def __init__(self, chip_producer, image_size=(256,256), cells=(16, 16), orientations=8): super().__init__(chip_producer) self.image_size = image_size self.cells = cells self.orientations = orientations def produce_features(self, chip): """Takes a chip object and returns a feature vector of size self.feat_size. """ img = PIL_Image.open(chip.filepath) img = img.resize(self.image_size, PIL_Image.BICUBIC) img_x, img_y = img.size img = color.rgb2gray(np.array(img)) features = hog( img, orientations=self.orientations, pixels_per_cell=(img_x / cells[0], img_y / cells[1]), cells_per_block=cells, # Normalize over the whole image ) return features def set_variables(self): self.feat_size = self.cells[0] * self.cells[1] * orientations
<commit_before>from skimage.feature import hog from skimage import colos from pelops.etl.feature_producer import FeatureProducer class HOGFeatureProducer(FeatureProducer): def __init__(self, chip_producer): super().__init__(chip_producer) def produce_features(self, chip): """Takes a chip object and returns a feature vector of size self.feat_size. """ img = PIL_Image.open(chip.filepath) img = img.resize((256, 256), PIL_Image.BICUBIC) img_x, img_y = img.size img = color.rgb2gray(np.array(img)) features = hog( img, orientations=8, pixels_per_cell=(int(img_x / 16), int(img_y / 16)), cells_per_block=(16, 16), # Normalize over the whole image ) return features def set_variables(self): """Child classes should use this to set self.feat_size, and any other needed variables. """ self.feat_size = 2048 # Set this in your inherited class <commit_msg>Replace magic numbers in `HOGFeatureProducer`<commit_after>
from skimage.feature import hog from skimage import colos from pelops.etl.feature_producer import FeatureProducer class HOGFeatureProducer(FeatureProducer): def __init__(self, chip_producer, image_size=(256,256), cells=(16, 16), orientations=8): super().__init__(chip_producer) self.image_size = image_size self.cells = cells self.orientations = orientations def produce_features(self, chip): """Takes a chip object and returns a feature vector of size self.feat_size. """ img = PIL_Image.open(chip.filepath) img = img.resize(self.image_size, PIL_Image.BICUBIC) img_x, img_y = img.size img = color.rgb2gray(np.array(img)) features = hog( img, orientations=self.orientations, pixels_per_cell=(img_x / cells[0], img_y / cells[1]), cells_per_block=cells, # Normalize over the whole image ) return features def set_variables(self): self.feat_size = self.cells[0] * self.cells[1] * orientations
from skimage.feature import hog from skimage import colos from pelops.etl.feature_producer import FeatureProducer class HOGFeatureProducer(FeatureProducer): def __init__(self, chip_producer): super().__init__(chip_producer) def produce_features(self, chip): """Takes a chip object and returns a feature vector of size self.feat_size. """ img = PIL_Image.open(chip.filepath) img = img.resize((256, 256), PIL_Image.BICUBIC) img_x, img_y = img.size img = color.rgb2gray(np.array(img)) features = hog( img, orientations=8, pixels_per_cell=(int(img_x / 16), int(img_y / 16)), cells_per_block=(16, 16), # Normalize over the whole image ) return features def set_variables(self): """Child classes should use this to set self.feat_size, and any other needed variables. """ self.feat_size = 2048 # Set this in your inherited class Replace magic numbers in `HOGFeatureProducer`from skimage.feature import hog from skimage import colos from pelops.etl.feature_producer import FeatureProducer class HOGFeatureProducer(FeatureProducer): def __init__(self, chip_producer, image_size=(256,256), cells=(16, 16), orientations=8): super().__init__(chip_producer) self.image_size = image_size self.cells = cells self.orientations = orientations def produce_features(self, chip): """Takes a chip object and returns a feature vector of size self.feat_size. """ img = PIL_Image.open(chip.filepath) img = img.resize(self.image_size, PIL_Image.BICUBIC) img_x, img_y = img.size img = color.rgb2gray(np.array(img)) features = hog( img, orientations=self.orientations, pixels_per_cell=(img_x / cells[0], img_y / cells[1]), cells_per_block=cells, # Normalize over the whole image ) return features def set_variables(self): self.feat_size = self.cells[0] * self.cells[1] * orientations
<commit_before>from skimage.feature import hog from skimage import colos from pelops.etl.feature_producer import FeatureProducer class HOGFeatureProducer(FeatureProducer): def __init__(self, chip_producer): super().__init__(chip_producer) def produce_features(self, chip): """Takes a chip object and returns a feature vector of size self.feat_size. """ img = PIL_Image.open(chip.filepath) img = img.resize((256, 256), PIL_Image.BICUBIC) img_x, img_y = img.size img = color.rgb2gray(np.array(img)) features = hog( img, orientations=8, pixels_per_cell=(int(img_x / 16), int(img_y / 16)), cells_per_block=(16, 16), # Normalize over the whole image ) return features def set_variables(self): """Child classes should use this to set self.feat_size, and any other needed variables. """ self.feat_size = 2048 # Set this in your inherited class <commit_msg>Replace magic numbers in `HOGFeatureProducer`<commit_after>from skimage.feature import hog from skimage import colos from pelops.etl.feature_producer import FeatureProducer class HOGFeatureProducer(FeatureProducer): def __init__(self, chip_producer, image_size=(256,256), cells=(16, 16), orientations=8): super().__init__(chip_producer) self.image_size = image_size self.cells = cells self.orientations = orientations def produce_features(self, chip): """Takes a chip object and returns a feature vector of size self.feat_size. """ img = PIL_Image.open(chip.filepath) img = img.resize(self.image_size, PIL_Image.BICUBIC) img_x, img_y = img.size img = color.rgb2gray(np.array(img)) features = hog( img, orientations=self.orientations, pixels_per_cell=(img_x / cells[0], img_y / cells[1]), cells_per_block=cells, # Normalize over the whole image ) return features def set_variables(self): self.feat_size = self.cells[0] * self.cells[1] * orientations
71c9b12056de1e1fdcc1effd2fda4c4dd284afab
froide/problem/utils.py
froide/problem/utils.py
from django.core.mail import mail_managers from django.conf import settings from django.template.loader import render_to_string from django.utils.translation import ugettext_lazy as _ def inform_managers(report): mail_managers( _('New problem: {label} [#{reqid}]').format( label=report.get_kind_display(), reqid=report.message.request_id ), '{}\n{}'.format( report.description, report.get_absolute_domain_url() ) ) def inform_user_problem_resolved(report): if report.auto_submitted or not report.user: return False foirequest = report.message.request subject = _('Problem resolved on your request') body = render_to_string("problem/email_problem_resolved.txt", { "user": report.user, "title": foirequest.title, "report": report, "url": report.user.get_autologin_url( report.message.get_absolute_short_url() ), "site_name": settings.SITE_NAME }) report.user.send_mail(subject, body) return True
from django.core.mail import mail_managers from django.conf import settings from django.urls import reverse from django.template.loader import render_to_string from django.utils.translation import ugettext_lazy as _ def inform_managers(report): admin_url = settings.SITE_URL + reverse( 'admin:problem_problemreport_change', args=(report.id,)) mail_managers( _('New problem: {label} [#{reqid}]').format( label=report.get_kind_display(), reqid=report.message.request_id ), '{}\n\n---\n\n{}\n'.format( report.description, report.get_absolute_domain_url(), admin_url ) ) def inform_user_problem_resolved(report): if report.auto_submitted or not report.user: return False foirequest = report.message.request subject = _('Problem resolved on your request') body = render_to_string("problem/email_problem_resolved.txt", { "user": report.user, "title": foirequest.title, "report": report, "url": report.user.get_autologin_url( report.message.get_absolute_short_url() ), "site_name": settings.SITE_NAME }) report.user.send_mail(subject, body) return True
Add link to report admin page to report info mail
Add link to report admin page to report info mail
Python
mit
stefanw/froide,stefanw/froide,stefanw/froide,fin/froide,stefanw/froide,fin/froide,stefanw/froide,fin/froide,fin/froide
from django.core.mail import mail_managers from django.conf import settings from django.template.loader import render_to_string from django.utils.translation import ugettext_lazy as _ def inform_managers(report): mail_managers( _('New problem: {label} [#{reqid}]').format( label=report.get_kind_display(), reqid=report.message.request_id ), '{}\n{}'.format( report.description, report.get_absolute_domain_url() ) ) def inform_user_problem_resolved(report): if report.auto_submitted or not report.user: return False foirequest = report.message.request subject = _('Problem resolved on your request') body = render_to_string("problem/email_problem_resolved.txt", { "user": report.user, "title": foirequest.title, "report": report, "url": report.user.get_autologin_url( report.message.get_absolute_short_url() ), "site_name": settings.SITE_NAME }) report.user.send_mail(subject, body) return True Add link to report admin page to report info mail
from django.core.mail import mail_managers from django.conf import settings from django.urls import reverse from django.template.loader import render_to_string from django.utils.translation import ugettext_lazy as _ def inform_managers(report): admin_url = settings.SITE_URL + reverse( 'admin:problem_problemreport_change', args=(report.id,)) mail_managers( _('New problem: {label} [#{reqid}]').format( label=report.get_kind_display(), reqid=report.message.request_id ), '{}\n\n---\n\n{}\n'.format( report.description, report.get_absolute_domain_url(), admin_url ) ) def inform_user_problem_resolved(report): if report.auto_submitted or not report.user: return False foirequest = report.message.request subject = _('Problem resolved on your request') body = render_to_string("problem/email_problem_resolved.txt", { "user": report.user, "title": foirequest.title, "report": report, "url": report.user.get_autologin_url( report.message.get_absolute_short_url() ), "site_name": settings.SITE_NAME }) report.user.send_mail(subject, body) return True
<commit_before>from django.core.mail import mail_managers from django.conf import settings from django.template.loader import render_to_string from django.utils.translation import ugettext_lazy as _ def inform_managers(report): mail_managers( _('New problem: {label} [#{reqid}]').format( label=report.get_kind_display(), reqid=report.message.request_id ), '{}\n{}'.format( report.description, report.get_absolute_domain_url() ) ) def inform_user_problem_resolved(report): if report.auto_submitted or not report.user: return False foirequest = report.message.request subject = _('Problem resolved on your request') body = render_to_string("problem/email_problem_resolved.txt", { "user": report.user, "title": foirequest.title, "report": report, "url": report.user.get_autologin_url( report.message.get_absolute_short_url() ), "site_name": settings.SITE_NAME }) report.user.send_mail(subject, body) return True <commit_msg>Add link to report admin page to report info mail<commit_after>
from django.core.mail import mail_managers from django.conf import settings from django.urls import reverse from django.template.loader import render_to_string from django.utils.translation import ugettext_lazy as _ def inform_managers(report): admin_url = settings.SITE_URL + reverse( 'admin:problem_problemreport_change', args=(report.id,)) mail_managers( _('New problem: {label} [#{reqid}]').format( label=report.get_kind_display(), reqid=report.message.request_id ), '{}\n\n---\n\n{}\n'.format( report.description, report.get_absolute_domain_url(), admin_url ) ) def inform_user_problem_resolved(report): if report.auto_submitted or not report.user: return False foirequest = report.message.request subject = _('Problem resolved on your request') body = render_to_string("problem/email_problem_resolved.txt", { "user": report.user, "title": foirequest.title, "report": report, "url": report.user.get_autologin_url( report.message.get_absolute_short_url() ), "site_name": settings.SITE_NAME }) report.user.send_mail(subject, body) return True
from django.core.mail import mail_managers from django.conf import settings from django.template.loader import render_to_string from django.utils.translation import ugettext_lazy as _ def inform_managers(report): mail_managers( _('New problem: {label} [#{reqid}]').format( label=report.get_kind_display(), reqid=report.message.request_id ), '{}\n{}'.format( report.description, report.get_absolute_domain_url() ) ) def inform_user_problem_resolved(report): if report.auto_submitted or not report.user: return False foirequest = report.message.request subject = _('Problem resolved on your request') body = render_to_string("problem/email_problem_resolved.txt", { "user": report.user, "title": foirequest.title, "report": report, "url": report.user.get_autologin_url( report.message.get_absolute_short_url() ), "site_name": settings.SITE_NAME }) report.user.send_mail(subject, body) return True Add link to report admin page to report info mailfrom django.core.mail import mail_managers from django.conf import settings from django.urls import reverse from django.template.loader import render_to_string from django.utils.translation import ugettext_lazy as _ def inform_managers(report): admin_url = settings.SITE_URL + reverse( 'admin:problem_problemreport_change', args=(report.id,)) mail_managers( _('New problem: {label} [#{reqid}]').format( label=report.get_kind_display(), reqid=report.message.request_id ), '{}\n\n---\n\n{}\n'.format( report.description, report.get_absolute_domain_url(), admin_url ) ) def inform_user_problem_resolved(report): if report.auto_submitted or not report.user: return False foirequest = report.message.request subject = _('Problem resolved on your request') body = render_to_string("problem/email_problem_resolved.txt", { "user": report.user, "title": foirequest.title, "report": report, "url": report.user.get_autologin_url( report.message.get_absolute_short_url() ), "site_name": settings.SITE_NAME }) report.user.send_mail(subject, body) return True
<commit_before>from django.core.mail import mail_managers from django.conf import settings from django.template.loader import render_to_string from django.utils.translation import ugettext_lazy as _ def inform_managers(report): mail_managers( _('New problem: {label} [#{reqid}]').format( label=report.get_kind_display(), reqid=report.message.request_id ), '{}\n{}'.format( report.description, report.get_absolute_domain_url() ) ) def inform_user_problem_resolved(report): if report.auto_submitted or not report.user: return False foirequest = report.message.request subject = _('Problem resolved on your request') body = render_to_string("problem/email_problem_resolved.txt", { "user": report.user, "title": foirequest.title, "report": report, "url": report.user.get_autologin_url( report.message.get_absolute_short_url() ), "site_name": settings.SITE_NAME }) report.user.send_mail(subject, body) return True <commit_msg>Add link to report admin page to report info mail<commit_after>from django.core.mail import mail_managers from django.conf import settings from django.urls import reverse from django.template.loader import render_to_string from django.utils.translation import ugettext_lazy as _ def inform_managers(report): admin_url = settings.SITE_URL + reverse( 'admin:problem_problemreport_change', args=(report.id,)) mail_managers( _('New problem: {label} [#{reqid}]').format( label=report.get_kind_display(), reqid=report.message.request_id ), '{}\n\n---\n\n{}\n'.format( report.description, report.get_absolute_domain_url(), admin_url ) ) def inform_user_problem_resolved(report): if report.auto_submitted or not report.user: return False foirequest = report.message.request subject = _('Problem resolved on your request') body = render_to_string("problem/email_problem_resolved.txt", { "user": report.user, "title": foirequest.title, "report": report, "url": report.user.get_autologin_url( report.message.get_absolute_short_url() ), "site_name": settings.SITE_NAME }) report.user.send_mail(subject, body) return True
53e2ef4079bc191b0b88395cf23e7bac576033dc
mne/simulation/simulation_metrics.py
mne/simulation/simulation_metrics.py
import numpy as np def _check_stc(stc1, stc2): # XXX What should we check? that the data is having the same size? if stc1.data.shape != stc2.data.shape: raise ValueError('Data in stcs must have the same size') if stc1.times != stc2.times: raise ValueError('Times of two stcs must match.') def source_estimate_quantification(stc1, stc2, metric='rms'): """Helper function to calculate matrix similarities. Parameters ---------- stc1 : SourceEstimate First source estimate for comparison stc2 : SourceEstimate First source estimate for comparison metric : str Metric to calculate. 'rms', 'avg_corrcoef', Returns ------- """ # TODO Add checks for source space _check_stc(stc1, stc2) score = _calc_metric(stc1.data, stc2.data, metric) def _calc_metric(data1, data2, metric): """Calculate metric of choice""" # Calculate root mean square difference between two matrices if metric == 'rms': return np.mean((stc1.data - stc2.data) ** 2) # Calculate correlation coefficient between matrix elements elif metric == 'avg_corrcoef': return np.mean(np.corrcoeff(stc1.data, stc2.data))
# Authors: Yousra Bekhti # Mark Wronkiewicz <wronk.mark@gmail.com> # # License: BSD (3-clause) import numpy as np def _check_stc(stc1, stc2): # XXX What should we check? that the data is having the same size? if stc1.data.shape != stc2.data.shape: raise ValueError('Data in stcs must have the same size') if stc1.times != stc2.times: raise ValueError('Times of two stcs must match.') def source_estimate_quantification(stc1, stc2, metric='rms'): """Helper function to calculate matrix similarities. Parameters ---------- stc1 : SourceEstimate First source estimate for comparison stc2 : SourceEstimate First source estimate for comparison metric : str Metric to calculate. 'rms', 'avg_corrcoef', Returns ------- """ # TODO Add checks for source space _check_stc(stc1, stc2) score = _calc_metric(stc1.data, stc2.data, metric) def _calc_metric(data1, data2, metric): """Helper to calculate metric of choice. Parameters ---------- data1 : ndarray, shape(n_sources, ntimes) Second data matrix data2 : ndarray, shape(n_sources, ntimes) Second data matrix metric : str Metric to calculate. 'rms', 'corr', Returns ------- score : float Calculated metric """ # Calculate root mean square difference between two matrices if metric == 'rms': return np.mean((stc1.data - stc2.data) ** 2) # Calculate correlation coefficient between matrix elements elif metric == 'corr': return np.correlate(stc1.data.flatten(), stc2.data.flatten())
Add documentation, tests, fix corr
Add documentation, tests, fix corr
Python
bsd-3-clause
lorenzo-desantis/mne-python,wronk/mne-python,Teekuningas/mne-python,olafhauk/mne-python,mne-tools/mne-python,jaeilepp/mne-python,ARudiuk/mne-python,jmontoyam/mne-python,rkmaddox/mne-python,pravsripad/mne-python,larsoner/mne-python,wmvanvliet/mne-python,larsoner/mne-python,kingjr/mne-python,teonlamont/mne-python,kingjr/mne-python,adykstra/mne-python,yousrabk/mne-python,wmvanvliet/mne-python,rkmaddox/mne-python,bloyl/mne-python,pravsripad/mne-python,adykstra/mne-python,nicproulx/mne-python,Teekuningas/mne-python,jniediek/mne-python,kambysese/mne-python,Teekuningas/mne-python,ARudiuk/mne-python,kambysese/mne-python,drammock/mne-python,jniediek/mne-python,Eric89GXL/mne-python,wronk/mne-python,cmoutard/mne-python,mne-tools/mne-python,drammock/mne-python,pravsripad/mne-python,jaeilepp/mne-python,cjayb/mne-python,teonlamont/mne-python,cjayb/mne-python,mne-tools/mne-python,nicproulx/mne-python,alexandrebarachant/mne-python,drammock/mne-python,olafhauk/mne-python,olafhauk/mne-python,lorenzo-desantis/mne-python,jmontoyam/mne-python,leggitta/mne-python,kingjr/mne-python,bloyl/mne-python,cmoutard/mne-python,leggitta/mne-python,larsoner/mne-python,yousrabk/mne-python,alexandrebarachant/mne-python,Eric89GXL/mne-python,wmvanvliet/mne-python
import numpy as np def _check_stc(stc1, stc2): # XXX What should we check? that the data is having the same size? if stc1.data.shape != stc2.data.shape: raise ValueError('Data in stcs must have the same size') if stc1.times != stc2.times: raise ValueError('Times of two stcs must match.') def source_estimate_quantification(stc1, stc2, metric='rms'): """Helper function to calculate matrix similarities. Parameters ---------- stc1 : SourceEstimate First source estimate for comparison stc2 : SourceEstimate First source estimate for comparison metric : str Metric to calculate. 'rms', 'avg_corrcoef', Returns ------- """ # TODO Add checks for source space _check_stc(stc1, stc2) score = _calc_metric(stc1.data, stc2.data, metric) def _calc_metric(data1, data2, metric): """Calculate metric of choice""" # Calculate root mean square difference between two matrices if metric == 'rms': return np.mean((stc1.data - stc2.data) ** 2) # Calculate correlation coefficient between matrix elements elif metric == 'avg_corrcoef': return np.mean(np.corrcoeff(stc1.data, stc2.data)) Add documentation, tests, fix corr
# Authors: Yousra Bekhti # Mark Wronkiewicz <wronk.mark@gmail.com> # # License: BSD (3-clause) import numpy as np def _check_stc(stc1, stc2): # XXX What should we check? that the data is having the same size? if stc1.data.shape != stc2.data.shape: raise ValueError('Data in stcs must have the same size') if stc1.times != stc2.times: raise ValueError('Times of two stcs must match.') def source_estimate_quantification(stc1, stc2, metric='rms'): """Helper function to calculate matrix similarities. Parameters ---------- stc1 : SourceEstimate First source estimate for comparison stc2 : SourceEstimate First source estimate for comparison metric : str Metric to calculate. 'rms', 'avg_corrcoef', Returns ------- """ # TODO Add checks for source space _check_stc(stc1, stc2) score = _calc_metric(stc1.data, stc2.data, metric) def _calc_metric(data1, data2, metric): """Helper to calculate metric of choice. Parameters ---------- data1 : ndarray, shape(n_sources, ntimes) Second data matrix data2 : ndarray, shape(n_sources, ntimes) Second data matrix metric : str Metric to calculate. 'rms', 'corr', Returns ------- score : float Calculated metric """ # Calculate root mean square difference between two matrices if metric == 'rms': return np.mean((stc1.data - stc2.data) ** 2) # Calculate correlation coefficient between matrix elements elif metric == 'corr': return np.correlate(stc1.data.flatten(), stc2.data.flatten())
<commit_before> import numpy as np def _check_stc(stc1, stc2): # XXX What should we check? that the data is having the same size? if stc1.data.shape != stc2.data.shape: raise ValueError('Data in stcs must have the same size') if stc1.times != stc2.times: raise ValueError('Times of two stcs must match.') def source_estimate_quantification(stc1, stc2, metric='rms'): """Helper function to calculate matrix similarities. Parameters ---------- stc1 : SourceEstimate First source estimate for comparison stc2 : SourceEstimate First source estimate for comparison metric : str Metric to calculate. 'rms', 'avg_corrcoef', Returns ------- """ # TODO Add checks for source space _check_stc(stc1, stc2) score = _calc_metric(stc1.data, stc2.data, metric) def _calc_metric(data1, data2, metric): """Calculate metric of choice""" # Calculate root mean square difference between two matrices if metric == 'rms': return np.mean((stc1.data - stc2.data) ** 2) # Calculate correlation coefficient between matrix elements elif metric == 'avg_corrcoef': return np.mean(np.corrcoeff(stc1.data, stc2.data)) <commit_msg>Add documentation, tests, fix corr<commit_after>
# Authors: Yousra Bekhti # Mark Wronkiewicz <wronk.mark@gmail.com> # # License: BSD (3-clause) import numpy as np def _check_stc(stc1, stc2): # XXX What should we check? that the data is having the same size? if stc1.data.shape != stc2.data.shape: raise ValueError('Data in stcs must have the same size') if stc1.times != stc2.times: raise ValueError('Times of two stcs must match.') def source_estimate_quantification(stc1, stc2, metric='rms'): """Helper function to calculate matrix similarities. Parameters ---------- stc1 : SourceEstimate First source estimate for comparison stc2 : SourceEstimate First source estimate for comparison metric : str Metric to calculate. 'rms', 'avg_corrcoef', Returns ------- """ # TODO Add checks for source space _check_stc(stc1, stc2) score = _calc_metric(stc1.data, stc2.data, metric) def _calc_metric(data1, data2, metric): """Helper to calculate metric of choice. Parameters ---------- data1 : ndarray, shape(n_sources, ntimes) Second data matrix data2 : ndarray, shape(n_sources, ntimes) Second data matrix metric : str Metric to calculate. 'rms', 'corr', Returns ------- score : float Calculated metric """ # Calculate root mean square difference between two matrices if metric == 'rms': return np.mean((stc1.data - stc2.data) ** 2) # Calculate correlation coefficient between matrix elements elif metric == 'corr': return np.correlate(stc1.data.flatten(), stc2.data.flatten())
import numpy as np def _check_stc(stc1, stc2): # XXX What should we check? that the data is having the same size? if stc1.data.shape != stc2.data.shape: raise ValueError('Data in stcs must have the same size') if stc1.times != stc2.times: raise ValueError('Times of two stcs must match.') def source_estimate_quantification(stc1, stc2, metric='rms'): """Helper function to calculate matrix similarities. Parameters ---------- stc1 : SourceEstimate First source estimate for comparison stc2 : SourceEstimate First source estimate for comparison metric : str Metric to calculate. 'rms', 'avg_corrcoef', Returns ------- """ # TODO Add checks for source space _check_stc(stc1, stc2) score = _calc_metric(stc1.data, stc2.data, metric) def _calc_metric(data1, data2, metric): """Calculate metric of choice""" # Calculate root mean square difference between two matrices if metric == 'rms': return np.mean((stc1.data - stc2.data) ** 2) # Calculate correlation coefficient between matrix elements elif metric == 'avg_corrcoef': return np.mean(np.corrcoeff(stc1.data, stc2.data)) Add documentation, tests, fix corr# Authors: Yousra Bekhti # Mark Wronkiewicz <wronk.mark@gmail.com> # # License: BSD (3-clause) import numpy as np def _check_stc(stc1, stc2): # XXX What should we check? that the data is having the same size? if stc1.data.shape != stc2.data.shape: raise ValueError('Data in stcs must have the same size') if stc1.times != stc2.times: raise ValueError('Times of two stcs must match.') def source_estimate_quantification(stc1, stc2, metric='rms'): """Helper function to calculate matrix similarities. Parameters ---------- stc1 : SourceEstimate First source estimate for comparison stc2 : SourceEstimate First source estimate for comparison metric : str Metric to calculate. 'rms', 'avg_corrcoef', Returns ------- """ # TODO Add checks for source space _check_stc(stc1, stc2) score = _calc_metric(stc1.data, stc2.data, metric) def _calc_metric(data1, data2, metric): """Helper to calculate metric of choice. Parameters ---------- data1 : ndarray, shape(n_sources, ntimes) Second data matrix data2 : ndarray, shape(n_sources, ntimes) Second data matrix metric : str Metric to calculate. 'rms', 'corr', Returns ------- score : float Calculated metric """ # Calculate root mean square difference between two matrices if metric == 'rms': return np.mean((stc1.data - stc2.data) ** 2) # Calculate correlation coefficient between matrix elements elif metric == 'corr': return np.correlate(stc1.data.flatten(), stc2.data.flatten())
<commit_before> import numpy as np def _check_stc(stc1, stc2): # XXX What should we check? that the data is having the same size? if stc1.data.shape != stc2.data.shape: raise ValueError('Data in stcs must have the same size') if stc1.times != stc2.times: raise ValueError('Times of two stcs must match.') def source_estimate_quantification(stc1, stc2, metric='rms'): """Helper function to calculate matrix similarities. Parameters ---------- stc1 : SourceEstimate First source estimate for comparison stc2 : SourceEstimate First source estimate for comparison metric : str Metric to calculate. 'rms', 'avg_corrcoef', Returns ------- """ # TODO Add checks for source space _check_stc(stc1, stc2) score = _calc_metric(stc1.data, stc2.data, metric) def _calc_metric(data1, data2, metric): """Calculate metric of choice""" # Calculate root mean square difference between two matrices if metric == 'rms': return np.mean((stc1.data - stc2.data) ** 2) # Calculate correlation coefficient between matrix elements elif metric == 'avg_corrcoef': return np.mean(np.corrcoeff(stc1.data, stc2.data)) <commit_msg>Add documentation, tests, fix corr<commit_after># Authors: Yousra Bekhti # Mark Wronkiewicz <wronk.mark@gmail.com> # # License: BSD (3-clause) import numpy as np def _check_stc(stc1, stc2): # XXX What should we check? that the data is having the same size? if stc1.data.shape != stc2.data.shape: raise ValueError('Data in stcs must have the same size') if stc1.times != stc2.times: raise ValueError('Times of two stcs must match.') def source_estimate_quantification(stc1, stc2, metric='rms'): """Helper function to calculate matrix similarities. Parameters ---------- stc1 : SourceEstimate First source estimate for comparison stc2 : SourceEstimate First source estimate for comparison metric : str Metric to calculate. 'rms', 'avg_corrcoef', Returns ------- """ # TODO Add checks for source space _check_stc(stc1, stc2) score = _calc_metric(stc1.data, stc2.data, metric) def _calc_metric(data1, data2, metric): """Helper to calculate metric of choice. Parameters ---------- data1 : ndarray, shape(n_sources, ntimes) Second data matrix data2 : ndarray, shape(n_sources, ntimes) Second data matrix metric : str Metric to calculate. 'rms', 'corr', Returns ------- score : float Calculated metric """ # Calculate root mean square difference between two matrices if metric == 'rms': return np.mean((stc1.data - stc2.data) ** 2) # Calculate correlation coefficient between matrix elements elif metric == 'corr': return np.correlate(stc1.data.flatten(), stc2.data.flatten())
8f3b6912ac7dd9fa63143a7e482d880658e69ac4
pyeda/__init__.py
pyeda/__init__.py
""" Python EDA Package common.py boolfunc.py -- Boolean functions expr.py -- Boolean logic expressions vexpr.py -- Boolean vector logic expressions table.py -- Boolean tables """ __copyright__ = "Copyright (c) 2012, Chris Drake" __version__ = "0.3.1"
""" Python EDA Package common.py boolfunc.py -- Boolean functions constant.py -- Boolean constant functions expr.py -- Boolean logic expressions vexpr.py -- Boolean vector logic expressions table.py -- Boolean tables """ __copyright__ = "Copyright (c) 2012, Chris Drake" __version__ = "0.3.1"
Add constant.py to package docstring
Add constant.py to package docstring
Python
bsd-2-clause
sschnug/pyeda,pombredanne/pyeda,pombredanne/pyeda,GtTmy/pyeda,cjdrake/pyeda,pombredanne/pyeda,GtTmy/pyeda,cjdrake/pyeda,sschnug/pyeda,karissa/pyeda,sschnug/pyeda,karissa/pyeda,cjdrake/pyeda,GtTmy/pyeda,karissa/pyeda
""" Python EDA Package common.py boolfunc.py -- Boolean functions expr.py -- Boolean logic expressions vexpr.py -- Boolean vector logic expressions table.py -- Boolean tables """ __copyright__ = "Copyright (c) 2012, Chris Drake" __version__ = "0.3.1" Add constant.py to package docstring
""" Python EDA Package common.py boolfunc.py -- Boolean functions constant.py -- Boolean constant functions expr.py -- Boolean logic expressions vexpr.py -- Boolean vector logic expressions table.py -- Boolean tables """ __copyright__ = "Copyright (c) 2012, Chris Drake" __version__ = "0.3.1"
<commit_before>""" Python EDA Package common.py boolfunc.py -- Boolean functions expr.py -- Boolean logic expressions vexpr.py -- Boolean vector logic expressions table.py -- Boolean tables """ __copyright__ = "Copyright (c) 2012, Chris Drake" __version__ = "0.3.1" <commit_msg>Add constant.py to package docstring<commit_after>
""" Python EDA Package common.py boolfunc.py -- Boolean functions constant.py -- Boolean constant functions expr.py -- Boolean logic expressions vexpr.py -- Boolean vector logic expressions table.py -- Boolean tables """ __copyright__ = "Copyright (c) 2012, Chris Drake" __version__ = "0.3.1"
""" Python EDA Package common.py boolfunc.py -- Boolean functions expr.py -- Boolean logic expressions vexpr.py -- Boolean vector logic expressions table.py -- Boolean tables """ __copyright__ = "Copyright (c) 2012, Chris Drake" __version__ = "0.3.1" Add constant.py to package docstring""" Python EDA Package common.py boolfunc.py -- Boolean functions constant.py -- Boolean constant functions expr.py -- Boolean logic expressions vexpr.py -- Boolean vector logic expressions table.py -- Boolean tables """ __copyright__ = "Copyright (c) 2012, Chris Drake" __version__ = "0.3.1"
<commit_before>""" Python EDA Package common.py boolfunc.py -- Boolean functions expr.py -- Boolean logic expressions vexpr.py -- Boolean vector logic expressions table.py -- Boolean tables """ __copyright__ = "Copyright (c) 2012, Chris Drake" __version__ = "0.3.1" <commit_msg>Add constant.py to package docstring<commit_after>""" Python EDA Package common.py boolfunc.py -- Boolean functions constant.py -- Boolean constant functions expr.py -- Boolean logic expressions vexpr.py -- Boolean vector logic expressions table.py -- Boolean tables """ __copyright__ = "Copyright (c) 2012, Chris Drake" __version__ = "0.3.1"
ac1dcd9f7acb8e8867996c70aa0a35eb45fcdc53
generator/php_di_gen.py
generator/php_di_gen.py
#! /usr/bin/python """ This script accepts a string with the following syntax: Get the list of classes that a PHP class depends on. Generate PHP code that defines the fields with corresponding names ( same name as the class name but with the first letter converted to lower case ). defines the constructor. Print out the code to the console. """ import sys from generator.generator import Generator from pyperclip.pyperclip import copy def main(): """ Parse arguments from command line """ argv = sys.argv length = len(argv) if length != 2: print_help() exit() dependent_list_string = sys.argv[1] statement = Generator.generate_statements(dependent_list_string) copy(statement) print statement def print_help(): """ Prints the help string for this script """ print "php-di-gen.py <dependent class list string>" main()
#! /usr/bin/python """ This script accepts a string with the following syntax: Get the list of classes that a PHP class depends on. Generate PHP code that defines the fields with corresponding names ( same name as the class name but with the first letter converted to lower case ). defines the constructor. Print out the code to the console. """ import sys from generator import Generator from pyperclip.pyperclip import copy def main(): """ Parse arguments from command line """ argv = sys.argv length = len(argv) if length != 2: print_help() exit() dependent_list_string = sys.argv[1] statement = Generator.generate_statements(dependent_list_string) copy(statement) print statement def print_help(): """ Prints the help string for this script """ print "Run this script/executable with the following parameter <dependent class list string>." main()
Fix import and update help
Fix import and update help Use relative module name The script may be invoked from a single executable without py extension.
Python
apache-2.0
HappyRay/php-di-generator
#! /usr/bin/python """ This script accepts a string with the following syntax: Get the list of classes that a PHP class depends on. Generate PHP code that defines the fields with corresponding names ( same name as the class name but with the first letter converted to lower case ). defines the constructor. Print out the code to the console. """ import sys from generator.generator import Generator from pyperclip.pyperclip import copy def main(): """ Parse arguments from command line """ argv = sys.argv length = len(argv) if length != 2: print_help() exit() dependent_list_string = sys.argv[1] statement = Generator.generate_statements(dependent_list_string) copy(statement) print statement def print_help(): """ Prints the help string for this script """ print "php-di-gen.py <dependent class list string>" main() Fix import and update help Use relative module name The script may be invoked from a single executable without py extension.
#! /usr/bin/python """ This script accepts a string with the following syntax: Get the list of classes that a PHP class depends on. Generate PHP code that defines the fields with corresponding names ( same name as the class name but with the first letter converted to lower case ). defines the constructor. Print out the code to the console. """ import sys from generator import Generator from pyperclip.pyperclip import copy def main(): """ Parse arguments from command line """ argv = sys.argv length = len(argv) if length != 2: print_help() exit() dependent_list_string = sys.argv[1] statement = Generator.generate_statements(dependent_list_string) copy(statement) print statement def print_help(): """ Prints the help string for this script """ print "Run this script/executable with the following parameter <dependent class list string>." main()
<commit_before>#! /usr/bin/python """ This script accepts a string with the following syntax: Get the list of classes that a PHP class depends on. Generate PHP code that defines the fields with corresponding names ( same name as the class name but with the first letter converted to lower case ). defines the constructor. Print out the code to the console. """ import sys from generator.generator import Generator from pyperclip.pyperclip import copy def main(): """ Parse arguments from command line """ argv = sys.argv length = len(argv) if length != 2: print_help() exit() dependent_list_string = sys.argv[1] statement = Generator.generate_statements(dependent_list_string) copy(statement) print statement def print_help(): """ Prints the help string for this script """ print "php-di-gen.py <dependent class list string>" main() <commit_msg>Fix import and update help Use relative module name The script may be invoked from a single executable without py extension.<commit_after>
#! /usr/bin/python """ This script accepts a string with the following syntax: Get the list of classes that a PHP class depends on. Generate PHP code that defines the fields with corresponding names ( same name as the class name but with the first letter converted to lower case ). defines the constructor. Print out the code to the console. """ import sys from generator import Generator from pyperclip.pyperclip import copy def main(): """ Parse arguments from command line """ argv = sys.argv length = len(argv) if length != 2: print_help() exit() dependent_list_string = sys.argv[1] statement = Generator.generate_statements(dependent_list_string) copy(statement) print statement def print_help(): """ Prints the help string for this script """ print "Run this script/executable with the following parameter <dependent class list string>." main()
#! /usr/bin/python """ This script accepts a string with the following syntax: Get the list of classes that a PHP class depends on. Generate PHP code that defines the fields with corresponding names ( same name as the class name but with the first letter converted to lower case ). defines the constructor. Print out the code to the console. """ import sys from generator.generator import Generator from pyperclip.pyperclip import copy def main(): """ Parse arguments from command line """ argv = sys.argv length = len(argv) if length != 2: print_help() exit() dependent_list_string = sys.argv[1] statement = Generator.generate_statements(dependent_list_string) copy(statement) print statement def print_help(): """ Prints the help string for this script """ print "php-di-gen.py <dependent class list string>" main() Fix import and update help Use relative module name The script may be invoked from a single executable without py extension.#! /usr/bin/python """ This script accepts a string with the following syntax: Get the list of classes that a PHP class depends on. Generate PHP code that defines the fields with corresponding names ( same name as the class name but with the first letter converted to lower case ). defines the constructor. Print out the code to the console. """ import sys from generator import Generator from pyperclip.pyperclip import copy def main(): """ Parse arguments from command line """ argv = sys.argv length = len(argv) if length != 2: print_help() exit() dependent_list_string = sys.argv[1] statement = Generator.generate_statements(dependent_list_string) copy(statement) print statement def print_help(): """ Prints the help string for this script """ print "Run this script/executable with the following parameter <dependent class list string>." main()
<commit_before>#! /usr/bin/python """ This script accepts a string with the following syntax: Get the list of classes that a PHP class depends on. Generate PHP code that defines the fields with corresponding names ( same name as the class name but with the first letter converted to lower case ). defines the constructor. Print out the code to the console. """ import sys from generator.generator import Generator from pyperclip.pyperclip import copy def main(): """ Parse arguments from command line """ argv = sys.argv length = len(argv) if length != 2: print_help() exit() dependent_list_string = sys.argv[1] statement = Generator.generate_statements(dependent_list_string) copy(statement) print statement def print_help(): """ Prints the help string for this script """ print "php-di-gen.py <dependent class list string>" main() <commit_msg>Fix import and update help Use relative module name The script may be invoked from a single executable without py extension.<commit_after>#! /usr/bin/python """ This script accepts a string with the following syntax: Get the list of classes that a PHP class depends on. Generate PHP code that defines the fields with corresponding names ( same name as the class name but with the first letter converted to lower case ). defines the constructor. Print out the code to the console. """ import sys from generator import Generator from pyperclip.pyperclip import copy def main(): """ Parse arguments from command line """ argv = sys.argv length = len(argv) if length != 2: print_help() exit() dependent_list_string = sys.argv[1] statement = Generator.generate_statements(dependent_list_string) copy(statement) print statement def print_help(): """ Prints the help string for this script """ print "Run this script/executable with the following parameter <dependent class list string>." main()
f029a204289abb91d57e117d45772d24372a1a43
requests-RT/rt.py
requests-RT/rt.py
#!/usr/bin/python import requests class RT: def __init__(self, apikey): self.apikey = apikey @staticmethod def make_request(url, params=None): req = requests.get(url, params=params) return req.content def search(self, query, page_limit=30, page=1): url = 'http://api.rottentomatoes.com/api/public/v1.0/movies.json' params = { 'q': query, 'page_limit': page_limit, 'page': page, 'apikey': self.apikey} return make_request(url, params) def box_office(self, limit=10, country='us'): url = 'http://api.rottentomatoes.com/api/public/v1.0/lists/movies/box_office.json' params = { 'limit': limit, 'country': country, 'apikey': self.apikey} return make_request(url, params=params)
#!/usr/bin/python __author__ = 'Mahmoud Hossam' __version__ = '0.1' import requests class RT: def __init__(self, apikey): self.apikey = apikey @staticmethod def make_request(url, params=None): req = requests.get(url, params=params) return req.content def search(self, query, page_limit=30, page=1): url = 'http://api.rottentomatoes.com/api/public/v1.0/movies.json' params = { 'q': query, 'page_limit': page_limit, 'page': page, 'apikey': self.apikey} return make_request(url, params) def box_office(self, limit=10, country='us'): url = 'http://api.rottentomatoes.com/api/public/v1.0/lists/movies/box_office.json' params = { 'limit': limit, 'country': country, 'apikey': self.apikey} return make_request(url, params=params)
Add author and version number
Add author and version number
Python
bsd-2-clause
mahmoudhossam/requests-RT
#!/usr/bin/python import requests class RT: def __init__(self, apikey): self.apikey = apikey @staticmethod def make_request(url, params=None): req = requests.get(url, params=params) return req.content def search(self, query, page_limit=30, page=1): url = 'http://api.rottentomatoes.com/api/public/v1.0/movies.json' params = { 'q': query, 'page_limit': page_limit, 'page': page, 'apikey': self.apikey} return make_request(url, params) def box_office(self, limit=10, country='us'): url = 'http://api.rottentomatoes.com/api/public/v1.0/lists/movies/box_office.json' params = { 'limit': limit, 'country': country, 'apikey': self.apikey} return make_request(url, params=params)Add author and version number
#!/usr/bin/python __author__ = 'Mahmoud Hossam' __version__ = '0.1' import requests class RT: def __init__(self, apikey): self.apikey = apikey @staticmethod def make_request(url, params=None): req = requests.get(url, params=params) return req.content def search(self, query, page_limit=30, page=1): url = 'http://api.rottentomatoes.com/api/public/v1.0/movies.json' params = { 'q': query, 'page_limit': page_limit, 'page': page, 'apikey': self.apikey} return make_request(url, params) def box_office(self, limit=10, country='us'): url = 'http://api.rottentomatoes.com/api/public/v1.0/lists/movies/box_office.json' params = { 'limit': limit, 'country': country, 'apikey': self.apikey} return make_request(url, params=params)
<commit_before>#!/usr/bin/python import requests class RT: def __init__(self, apikey): self.apikey = apikey @staticmethod def make_request(url, params=None): req = requests.get(url, params=params) return req.content def search(self, query, page_limit=30, page=1): url = 'http://api.rottentomatoes.com/api/public/v1.0/movies.json' params = { 'q': query, 'page_limit': page_limit, 'page': page, 'apikey': self.apikey} return make_request(url, params) def box_office(self, limit=10, country='us'): url = 'http://api.rottentomatoes.com/api/public/v1.0/lists/movies/box_office.json' params = { 'limit': limit, 'country': country, 'apikey': self.apikey} return make_request(url, params=params)<commit_msg>Add author and version number<commit_after>
#!/usr/bin/python __author__ = 'Mahmoud Hossam' __version__ = '0.1' import requests class RT: def __init__(self, apikey): self.apikey = apikey @staticmethod def make_request(url, params=None): req = requests.get(url, params=params) return req.content def search(self, query, page_limit=30, page=1): url = 'http://api.rottentomatoes.com/api/public/v1.0/movies.json' params = { 'q': query, 'page_limit': page_limit, 'page': page, 'apikey': self.apikey} return make_request(url, params) def box_office(self, limit=10, country='us'): url = 'http://api.rottentomatoes.com/api/public/v1.0/lists/movies/box_office.json' params = { 'limit': limit, 'country': country, 'apikey': self.apikey} return make_request(url, params=params)
#!/usr/bin/python import requests class RT: def __init__(self, apikey): self.apikey = apikey @staticmethod def make_request(url, params=None): req = requests.get(url, params=params) return req.content def search(self, query, page_limit=30, page=1): url = 'http://api.rottentomatoes.com/api/public/v1.0/movies.json' params = { 'q': query, 'page_limit': page_limit, 'page': page, 'apikey': self.apikey} return make_request(url, params) def box_office(self, limit=10, country='us'): url = 'http://api.rottentomatoes.com/api/public/v1.0/lists/movies/box_office.json' params = { 'limit': limit, 'country': country, 'apikey': self.apikey} return make_request(url, params=params)Add author and version number#!/usr/bin/python __author__ = 'Mahmoud Hossam' __version__ = '0.1' import requests class RT: def __init__(self, apikey): self.apikey = apikey @staticmethod def make_request(url, params=None): req = requests.get(url, params=params) return req.content def search(self, query, page_limit=30, page=1): url = 'http://api.rottentomatoes.com/api/public/v1.0/movies.json' params = { 'q': query, 'page_limit': page_limit, 'page': page, 'apikey': self.apikey} return make_request(url, params) def box_office(self, limit=10, country='us'): url = 'http://api.rottentomatoes.com/api/public/v1.0/lists/movies/box_office.json' params = { 'limit': limit, 'country': country, 'apikey': self.apikey} return make_request(url, params=params)
<commit_before>#!/usr/bin/python import requests class RT: def __init__(self, apikey): self.apikey = apikey @staticmethod def make_request(url, params=None): req = requests.get(url, params=params) return req.content def search(self, query, page_limit=30, page=1): url = 'http://api.rottentomatoes.com/api/public/v1.0/movies.json' params = { 'q': query, 'page_limit': page_limit, 'page': page, 'apikey': self.apikey} return make_request(url, params) def box_office(self, limit=10, country='us'): url = 'http://api.rottentomatoes.com/api/public/v1.0/lists/movies/box_office.json' params = { 'limit': limit, 'country': country, 'apikey': self.apikey} return make_request(url, params=params)<commit_msg>Add author and version number<commit_after>#!/usr/bin/python __author__ = 'Mahmoud Hossam' __version__ = '0.1' import requests class RT: def __init__(self, apikey): self.apikey = apikey @staticmethod def make_request(url, params=None): req = requests.get(url, params=params) return req.content def search(self, query, page_limit=30, page=1): url = 'http://api.rottentomatoes.com/api/public/v1.0/movies.json' params = { 'q': query, 'page_limit': page_limit, 'page': page, 'apikey': self.apikey} return make_request(url, params) def box_office(self, limit=10, country='us'): url = 'http://api.rottentomatoes.com/api/public/v1.0/lists/movies/box_office.json' params = { 'limit': limit, 'country': country, 'apikey': self.apikey} return make_request(url, params=params)
6194aa67ce54ce5c2fb0de58bfb2d75607d7b882
bin/combine_results.py
bin/combine_results.py
#!/usr/bin/env python """ Simple script to combine JUnit test results into a single XML file. Useful for Jenkins. TODO: Pretty indentation """ import os from xml.etree import cElementTree as ET def find_all(name, path): result = [] for root, dirs, files in os.walk(path): if name in files: yield os.path.join(root, name) def main(path, output): testsuite = ET.Element("testsuite", name="all", package="all", tests="0") for fname in find_all("results.xml", path): tree = ET.parse(fname) for element in tree.iter("testcase"): testsuite.append(element) result = ET.Element("testsuites", name="results") result.append(testsuite) ET.ElementTree(result).write(output, encoding="UTF-8") if __name__ == "__main__": main(".", "combined_results.xml")
#!/usr/bin/env python """ Simple script to combine JUnit test results into a single XML file. Useful for Jenkins. TODO: Pretty indentation """ import os from xml.etree import cElementTree as ET def find_all(name, path): result = [] for root, dirs, files in os.walk(path): if name in files: yield os.path.join(root, name) def main(path, output): testsuite = ET.Element("testsuite", name="all", package="all", tests="0") for fname in find_all("results.xml", path): tree = ET.parse(fname) for element in tree.getiterator("testcase"): testsuite.append(element) result = ET.Element("testsuites", name="results") result.append(testsuite) ET.ElementTree(result).write(output, encoding="UTF-8") if __name__ == "__main__": main(".", "combined_results.xml")
Enable regression run to work on python 2.6
Cleanup: Enable regression run to work on python 2.6
Python
bsd-3-clause
mkreider/cocotb2,stuarthodgson/cocotb,mkreider/cocotb2,mkreider/cocotb2,stuarthodgson/cocotb,stuarthodgson/cocotb
#!/usr/bin/env python """ Simple script to combine JUnit test results into a single XML file. Useful for Jenkins. TODO: Pretty indentation """ import os from xml.etree import cElementTree as ET def find_all(name, path): result = [] for root, dirs, files in os.walk(path): if name in files: yield os.path.join(root, name) def main(path, output): testsuite = ET.Element("testsuite", name="all", package="all", tests="0") for fname in find_all("results.xml", path): tree = ET.parse(fname) for element in tree.iter("testcase"): testsuite.append(element) result = ET.Element("testsuites", name="results") result.append(testsuite) ET.ElementTree(result).write(output, encoding="UTF-8") if __name__ == "__main__": main(".", "combined_results.xml") Cleanup: Enable regression run to work on python 2.6
#!/usr/bin/env python """ Simple script to combine JUnit test results into a single XML file. Useful for Jenkins. TODO: Pretty indentation """ import os from xml.etree import cElementTree as ET def find_all(name, path): result = [] for root, dirs, files in os.walk(path): if name in files: yield os.path.join(root, name) def main(path, output): testsuite = ET.Element("testsuite", name="all", package="all", tests="0") for fname in find_all("results.xml", path): tree = ET.parse(fname) for element in tree.getiterator("testcase"): testsuite.append(element) result = ET.Element("testsuites", name="results") result.append(testsuite) ET.ElementTree(result).write(output, encoding="UTF-8") if __name__ == "__main__": main(".", "combined_results.xml")
<commit_before>#!/usr/bin/env python """ Simple script to combine JUnit test results into a single XML file. Useful for Jenkins. TODO: Pretty indentation """ import os from xml.etree import cElementTree as ET def find_all(name, path): result = [] for root, dirs, files in os.walk(path): if name in files: yield os.path.join(root, name) def main(path, output): testsuite = ET.Element("testsuite", name="all", package="all", tests="0") for fname in find_all("results.xml", path): tree = ET.parse(fname) for element in tree.iter("testcase"): testsuite.append(element) result = ET.Element("testsuites", name="results") result.append(testsuite) ET.ElementTree(result).write(output, encoding="UTF-8") if __name__ == "__main__": main(".", "combined_results.xml") <commit_msg>Cleanup: Enable regression run to work on python 2.6<commit_after>
#!/usr/bin/env python """ Simple script to combine JUnit test results into a single XML file. Useful for Jenkins. TODO: Pretty indentation """ import os from xml.etree import cElementTree as ET def find_all(name, path): result = [] for root, dirs, files in os.walk(path): if name in files: yield os.path.join(root, name) def main(path, output): testsuite = ET.Element("testsuite", name="all", package="all", tests="0") for fname in find_all("results.xml", path): tree = ET.parse(fname) for element in tree.getiterator("testcase"): testsuite.append(element) result = ET.Element("testsuites", name="results") result.append(testsuite) ET.ElementTree(result).write(output, encoding="UTF-8") if __name__ == "__main__": main(".", "combined_results.xml")
#!/usr/bin/env python """ Simple script to combine JUnit test results into a single XML file. Useful for Jenkins. TODO: Pretty indentation """ import os from xml.etree import cElementTree as ET def find_all(name, path): result = [] for root, dirs, files in os.walk(path): if name in files: yield os.path.join(root, name) def main(path, output): testsuite = ET.Element("testsuite", name="all", package="all", tests="0") for fname in find_all("results.xml", path): tree = ET.parse(fname) for element in tree.iter("testcase"): testsuite.append(element) result = ET.Element("testsuites", name="results") result.append(testsuite) ET.ElementTree(result).write(output, encoding="UTF-8") if __name__ == "__main__": main(".", "combined_results.xml") Cleanup: Enable regression run to work on python 2.6#!/usr/bin/env python """ Simple script to combine JUnit test results into a single XML file. Useful for Jenkins. TODO: Pretty indentation """ import os from xml.etree import cElementTree as ET def find_all(name, path): result = [] for root, dirs, files in os.walk(path): if name in files: yield os.path.join(root, name) def main(path, output): testsuite = ET.Element("testsuite", name="all", package="all", tests="0") for fname in find_all("results.xml", path): tree = ET.parse(fname) for element in tree.getiterator("testcase"): testsuite.append(element) result = ET.Element("testsuites", name="results") result.append(testsuite) ET.ElementTree(result).write(output, encoding="UTF-8") if __name__ == "__main__": main(".", "combined_results.xml")
<commit_before>#!/usr/bin/env python """ Simple script to combine JUnit test results into a single XML file. Useful for Jenkins. TODO: Pretty indentation """ import os from xml.etree import cElementTree as ET def find_all(name, path): result = [] for root, dirs, files in os.walk(path): if name in files: yield os.path.join(root, name) def main(path, output): testsuite = ET.Element("testsuite", name="all", package="all", tests="0") for fname in find_all("results.xml", path): tree = ET.parse(fname) for element in tree.iter("testcase"): testsuite.append(element) result = ET.Element("testsuites", name="results") result.append(testsuite) ET.ElementTree(result).write(output, encoding="UTF-8") if __name__ == "__main__": main(".", "combined_results.xml") <commit_msg>Cleanup: Enable regression run to work on python 2.6<commit_after>#!/usr/bin/env python """ Simple script to combine JUnit test results into a single XML file. Useful for Jenkins. TODO: Pretty indentation """ import os from xml.etree import cElementTree as ET def find_all(name, path): result = [] for root, dirs, files in os.walk(path): if name in files: yield os.path.join(root, name) def main(path, output): testsuite = ET.Element("testsuite", name="all", package="all", tests="0") for fname in find_all("results.xml", path): tree = ET.parse(fname) for element in tree.getiterator("testcase"): testsuite.append(element) result = ET.Element("testsuites", name="results") result.append(testsuite) ET.ElementTree(result).write(output, encoding="UTF-8") if __name__ == "__main__": main(".", "combined_results.xml")
e9ec19e68ccefaee9a975a6adc26cb6e5f1f7f33
pymodels/middlelayer/devices/__init__.py
pymodels/middlelayer/devices/__init__.py
from .dcct import DCCT from .li_llrf import LiLLRF from .rf import RF from .sofb import SOFB from .kicker import Kicker from .septum import Septum from .screen import Screen from .bpm import BPM from .ict import ICT from .ict import TranspEff from .egun import Bias from .egun import Filament from .egun import HVPS
from .dcct import DCCT from .li_llrf import LiLLRF from .rf import RF from .sofb import SOFB from .kicker import Kicker from .septum import Septum from .screen import Screen from .bpm import BPM from .ict import ICT from .ict import TranspEff from .egun import Bias from .egun import Filament from .egun import HVPS from .timing import Timing
Add timing in devices init
DEV.ENH: Add timing in devices init
Python
mit
lnls-fac/sirius
from .dcct import DCCT from .li_llrf import LiLLRF from .rf import RF from .sofb import SOFB from .kicker import Kicker from .septum import Septum from .screen import Screen from .bpm import BPM from .ict import ICT from .ict import TranspEff from .egun import Bias from .egun import Filament from .egun import HVPS DEV.ENH: Add timing in devices init
from .dcct import DCCT from .li_llrf import LiLLRF from .rf import RF from .sofb import SOFB from .kicker import Kicker from .septum import Septum from .screen import Screen from .bpm import BPM from .ict import ICT from .ict import TranspEff from .egun import Bias from .egun import Filament from .egun import HVPS from .timing import Timing
<commit_before>from .dcct import DCCT from .li_llrf import LiLLRF from .rf import RF from .sofb import SOFB from .kicker import Kicker from .septum import Septum from .screen import Screen from .bpm import BPM from .ict import ICT from .ict import TranspEff from .egun import Bias from .egun import Filament from .egun import HVPS <commit_msg>DEV.ENH: Add timing in devices init<commit_after>
from .dcct import DCCT from .li_llrf import LiLLRF from .rf import RF from .sofb import SOFB from .kicker import Kicker from .septum import Septum from .screen import Screen from .bpm import BPM from .ict import ICT from .ict import TranspEff from .egun import Bias from .egun import Filament from .egun import HVPS from .timing import Timing
from .dcct import DCCT from .li_llrf import LiLLRF from .rf import RF from .sofb import SOFB from .kicker import Kicker from .septum import Septum from .screen import Screen from .bpm import BPM from .ict import ICT from .ict import TranspEff from .egun import Bias from .egun import Filament from .egun import HVPS DEV.ENH: Add timing in devices initfrom .dcct import DCCT from .li_llrf import LiLLRF from .rf import RF from .sofb import SOFB from .kicker import Kicker from .septum import Septum from .screen import Screen from .bpm import BPM from .ict import ICT from .ict import TranspEff from .egun import Bias from .egun import Filament from .egun import HVPS from .timing import Timing
<commit_before>from .dcct import DCCT from .li_llrf import LiLLRF from .rf import RF from .sofb import SOFB from .kicker import Kicker from .septum import Septum from .screen import Screen from .bpm import BPM from .ict import ICT from .ict import TranspEff from .egun import Bias from .egun import Filament from .egun import HVPS <commit_msg>DEV.ENH: Add timing in devices init<commit_after>from .dcct import DCCT from .li_llrf import LiLLRF from .rf import RF from .sofb import SOFB from .kicker import Kicker from .septum import Septum from .screen import Screen from .bpm import BPM from .ict import ICT from .ict import TranspEff from .egun import Bias from .egun import Filament from .egun import HVPS from .timing import Timing
a9354124f4905f4befe9ff2ca8274406fbbb0dad
readux/annotations/migrations/0003_annotation_group_and_permissions.py
readux/annotations/migrations/0003_annotation_group_and_permissions.py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('auth', '0006_require_contenttypes_0002'), ('annotations', '0002_add_volume_uri'), ] operations = [ migrations.CreateModel( name='AnnotationGroup', fields=[ ('group_ptr', models.OneToOneField(parent_link=True, auto_created=True, primary_key=True, serialize=False, to='auth.Group')), ('notes', models.TextField(blank=True)), ('created', models.DateTimeField(auto_now_add=True)), ('updated', models.DateTimeField(auto_now=True)), ], bases=('auth.group',), ), migrations.AlterModelOptions( name='annotation', options={'permissions': (('view_annotation', 'View annotation'),)}, ), migrations.AlterModelOptions( name='annotation', options={'permissions': (('view_annotation', 'View annotation'), ('admin_annotation', 'Manage annotation'))}, ), ]
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('auth', '0006_require_contenttypes_0002'), ('annotations', '0002_add_volume_uri'), ] operations = [ migrations.CreateModel( name='AnnotationGroup', fields=[ ('group_ptr', models.OneToOneField(parent_link=True, auto_created=True, primary_key=True, serialize=False, to='auth.Group')), ('notes', models.TextField(blank=True)), ('created', models.DateTimeField(auto_now_add=True)), ('updated', models.DateTimeField(auto_now=True)), ], bases=('auth.group',), ), migrations.AlterModelOptions( name='annotation', options={'permissions': (('view_annotation', 'View annotation'), ('admin_annotation', 'Manage annotation'))}, ), ]
Fix migration that adds custom annotation permissions
Fix migration that adds custom annotation permissions
Python
apache-2.0
emory-libraries/readux,emory-libraries/readux,emory-libraries/readux
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('auth', '0006_require_contenttypes_0002'), ('annotations', '0002_add_volume_uri'), ] operations = [ migrations.CreateModel( name='AnnotationGroup', fields=[ ('group_ptr', models.OneToOneField(parent_link=True, auto_created=True, primary_key=True, serialize=False, to='auth.Group')), ('notes', models.TextField(blank=True)), ('created', models.DateTimeField(auto_now_add=True)), ('updated', models.DateTimeField(auto_now=True)), ], bases=('auth.group',), ), migrations.AlterModelOptions( name='annotation', options={'permissions': (('view_annotation', 'View annotation'),)}, ), migrations.AlterModelOptions( name='annotation', options={'permissions': (('view_annotation', 'View annotation'), ('admin_annotation', 'Manage annotation'))}, ), ] Fix migration that adds custom annotation permissions
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('auth', '0006_require_contenttypes_0002'), ('annotations', '0002_add_volume_uri'), ] operations = [ migrations.CreateModel( name='AnnotationGroup', fields=[ ('group_ptr', models.OneToOneField(parent_link=True, auto_created=True, primary_key=True, serialize=False, to='auth.Group')), ('notes', models.TextField(blank=True)), ('created', models.DateTimeField(auto_now_add=True)), ('updated', models.DateTimeField(auto_now=True)), ], bases=('auth.group',), ), migrations.AlterModelOptions( name='annotation', options={'permissions': (('view_annotation', 'View annotation'), ('admin_annotation', 'Manage annotation'))}, ), ]
<commit_before># -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('auth', '0006_require_contenttypes_0002'), ('annotations', '0002_add_volume_uri'), ] operations = [ migrations.CreateModel( name='AnnotationGroup', fields=[ ('group_ptr', models.OneToOneField(parent_link=True, auto_created=True, primary_key=True, serialize=False, to='auth.Group')), ('notes', models.TextField(blank=True)), ('created', models.DateTimeField(auto_now_add=True)), ('updated', models.DateTimeField(auto_now=True)), ], bases=('auth.group',), ), migrations.AlterModelOptions( name='annotation', options={'permissions': (('view_annotation', 'View annotation'),)}, ), migrations.AlterModelOptions( name='annotation', options={'permissions': (('view_annotation', 'View annotation'), ('admin_annotation', 'Manage annotation'))}, ), ] <commit_msg>Fix migration that adds custom annotation permissions<commit_after>
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('auth', '0006_require_contenttypes_0002'), ('annotations', '0002_add_volume_uri'), ] operations = [ migrations.CreateModel( name='AnnotationGroup', fields=[ ('group_ptr', models.OneToOneField(parent_link=True, auto_created=True, primary_key=True, serialize=False, to='auth.Group')), ('notes', models.TextField(blank=True)), ('created', models.DateTimeField(auto_now_add=True)), ('updated', models.DateTimeField(auto_now=True)), ], bases=('auth.group',), ), migrations.AlterModelOptions( name='annotation', options={'permissions': (('view_annotation', 'View annotation'), ('admin_annotation', 'Manage annotation'))}, ), ]
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('auth', '0006_require_contenttypes_0002'), ('annotations', '0002_add_volume_uri'), ] operations = [ migrations.CreateModel( name='AnnotationGroup', fields=[ ('group_ptr', models.OneToOneField(parent_link=True, auto_created=True, primary_key=True, serialize=False, to='auth.Group')), ('notes', models.TextField(blank=True)), ('created', models.DateTimeField(auto_now_add=True)), ('updated', models.DateTimeField(auto_now=True)), ], bases=('auth.group',), ), migrations.AlterModelOptions( name='annotation', options={'permissions': (('view_annotation', 'View annotation'),)}, ), migrations.AlterModelOptions( name='annotation', options={'permissions': (('view_annotation', 'View annotation'), ('admin_annotation', 'Manage annotation'))}, ), ] Fix migration that adds custom annotation permissions# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('auth', '0006_require_contenttypes_0002'), ('annotations', '0002_add_volume_uri'), ] operations = [ migrations.CreateModel( name='AnnotationGroup', fields=[ ('group_ptr', models.OneToOneField(parent_link=True, auto_created=True, primary_key=True, serialize=False, to='auth.Group')), ('notes', models.TextField(blank=True)), ('created', models.DateTimeField(auto_now_add=True)), ('updated', models.DateTimeField(auto_now=True)), ], bases=('auth.group',), ), migrations.AlterModelOptions( name='annotation', options={'permissions': (('view_annotation', 'View annotation'), ('admin_annotation', 'Manage annotation'))}, ), ]
<commit_before># -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('auth', '0006_require_contenttypes_0002'), ('annotations', '0002_add_volume_uri'), ] operations = [ migrations.CreateModel( name='AnnotationGroup', fields=[ ('group_ptr', models.OneToOneField(parent_link=True, auto_created=True, primary_key=True, serialize=False, to='auth.Group')), ('notes', models.TextField(blank=True)), ('created', models.DateTimeField(auto_now_add=True)), ('updated', models.DateTimeField(auto_now=True)), ], bases=('auth.group',), ), migrations.AlterModelOptions( name='annotation', options={'permissions': (('view_annotation', 'View annotation'),)}, ), migrations.AlterModelOptions( name='annotation', options={'permissions': (('view_annotation', 'View annotation'), ('admin_annotation', 'Manage annotation'))}, ), ] <commit_msg>Fix migration that adds custom annotation permissions<commit_after># -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('auth', '0006_require_contenttypes_0002'), ('annotations', '0002_add_volume_uri'), ] operations = [ migrations.CreateModel( name='AnnotationGroup', fields=[ ('group_ptr', models.OneToOneField(parent_link=True, auto_created=True, primary_key=True, serialize=False, to='auth.Group')), ('notes', models.TextField(blank=True)), ('created', models.DateTimeField(auto_now_add=True)), ('updated', models.DateTimeField(auto_now=True)), ], bases=('auth.group',), ), migrations.AlterModelOptions( name='annotation', options={'permissions': (('view_annotation', 'View annotation'), ('admin_annotation', 'Manage annotation'))}, ), ]
5531f188c7bf3030cb9fc3b46d92a1db60817b7c
confirmation/views.py
confirmation/views.py
# -*- coding: utf-8 -*- # Copyright: (c) 2008, Jarek Zgoda <jarek.zgoda@gmail.com> __revision__ = '$Id: views.py 21 2008-12-05 09:21:03Z jarek.zgoda $' from django.shortcuts import render_to_response from django.template import RequestContext from django.conf import settings from confirmation.models import Confirmation def confirm(request, confirmation_key): confirmation_key = confirmation_key.lower() obj = Confirmation.objects.confirm(confirmation_key) confirmed = True if not obj: # confirmation failed confirmed = False try: # try to get the object we was supposed to confirm obj = Confirmation.objects.get(confirmation_key=confirmation_key) except Confirmation.DoesNotExist: pass ctx = { 'object': obj, 'confirmed': confirmed, 'days': getattr(settings, 'EMAIL_CONFIRMATION_DAYS', 10), } templates = [ 'confirmation/confirm.html', ] if obj: # if we have an object, we can use specific template templates.insert(0, 'confirmation/confirm_%s.html' % obj._meta.module_name) return render_to_response(templates, ctx, context_instance=RequestContext(request))
# -*- coding: utf-8 -*- # Copyright: (c) 2008, Jarek Zgoda <jarek.zgoda@gmail.com> __revision__ = '$Id: views.py 21 2008-12-05 09:21:03Z jarek.zgoda $' from django.shortcuts import render_to_response from django.template import RequestContext from django.conf import settings from confirmation.models import Confirmation def confirm(request, confirmation_key): confirmation_key = confirmation_key.lower() obj = Confirmation.objects.confirm(confirmation_key) confirmed = True if not obj: # confirmation failed confirmed = False try: # try to get the object we was supposed to confirm obj = Confirmation.objects.get(confirmation_key=confirmation_key) except Confirmation.DoesNotExist: pass ctx = { 'object': obj, 'confirmed': confirmed, 'days': getattr(settings, 'EMAIL_CONFIRMATION_DAYS', 10), 'key': confirmation_key, } templates = [ 'confirmation/confirm.html', ] if obj: # if we have an object, we can use specific template templates.insert(0, 'confirmation/confirm_%s.html' % obj._meta.module_name) return render_to_response(templates, ctx, context_instance=RequestContext(request))
Include confirmation key in context object.
Include confirmation key in context object. This way our templates can reference the confirmation key later. (imported from commit 4d57e1309386f2236829b6fdf4e4ad43c5b125c8)
Python
apache-2.0
schatt/zulip,deer-hope/zulip,PaulPetring/zulip,adnanh/zulip,jeffcao/zulip,hafeez3000/zulip,fw1121/zulip,KJin99/zulip,ufosky-server/zulip,mahim97/zulip,dhcrzf/zulip,dawran6/zulip,firstblade/zulip,akuseru/zulip,eastlhu/zulip,showell/zulip,dawran6/zulip,vikas-parashar/zulip,esander91/zulip,eastlhu/zulip,so0k/zulip,johnny9/zulip,amallia/zulip,sharmaeklavya2/zulip,mansilladev/zulip,krtkmj/zulip,dawran6/zulip,proliming/zulip,rht/zulip,jackrzhang/zulip,noroot/zulip,eastlhu/zulip,deer-hope/zulip,firstblade/zulip,wweiradio/zulip,jackrzhang/zulip,punchagan/zulip,peiwei/zulip,showell/zulip,Cheppers/zulip,guiquanz/zulip,seapasulli/zulip,hackerkid/zulip,jrowan/zulip,ipernet/zulip,dotcool/zulip,esander91/zulip,kou/zulip,Galexrt/zulip,nicholasbs/zulip,PhilSk/zulip,zwily/zulip,mdavid/zulip,amanharitsh123/zulip,Drooids/zulip,kokoar/zulip,karamcnair/zulip,stamhe/zulip,easyfmxu/zulip,PaulPetring/zulip,sharmaeklavya2/zulip,ahmadassaf/zulip,saitodisse/zulip,levixie/zulip,arpith/zulip,Drooids/zulip,zwily/zulip,gigawhitlocks/zulip,Drooids/zulip,dxq-git/zulip,swinghu/zulip,jonesgithub/zulip,thomasboyt/zulip,itnihao/zulip,dwrpayne/zulip,ashwinirudrappa/zulip,armooo/zulip,jimmy54/zulip,jessedhillon/zulip,LeeRisk/zulip,samatdav/zulip,zorojean/zulip,proliming/zulip,jerryge/zulip,fw1121/zulip,Suninus/zulip,ryanbackman/zulip,andersk/zulip,luyifan/zulip,pradiptad/zulip,joshisa/zulip,swinghu/zulip,armooo/zulip,grave-w-grave/zulip,Suninus/zulip,udxxabp/zulip,firstblade/zulip,akuseru/zulip,shubhamdhama/zulip,atomic-labs/zulip,bluesea/zulip,AZtheAsian/zulip,rht/zulip,hayderimran7/zulip,akuseru/zulip,peiwei/zulip,codeKonami/zulip,jerryge/zulip,fw1121/zulip,kou/zulip,shaunstanislaus/zulip,johnnygaddarr/zulip,kokoar/zulip,Frouk/zulip,samatdav/zulip,hackerkid/zulip,Frouk/zulip,saitodisse/zulip,hackerkid/zulip,lfranchi/zulip,ApsOps/zulip,sonali0901/zulip,voidException/zulip,atomic-labs/zulip,zachallaun/zulip,wdaher/zulip,noroot/zulip,zulip/zulip,xuxiao/zulip,gigawhitlocks/zulip,hackerkid/zulip,Vallher/zulip,susansls/zulip,reyha/zulip,paxapy/zulip,ryansnowboarder/zulip,xuxiao/zulip,ahmadassaf/zulip,dattatreya303/zulip,rht/zulip,shubhamdhama/zulip,jainayush975/zulip,noroot/zulip,TigorC/zulip,vabs22/zulip,luyifan/zulip,synicalsyntax/zulip,lfranchi/zulip,dwrpayne/zulip,kaiyuanheshang/zulip,dnmfarrell/zulip,kokoar/zulip,moria/zulip,jainayush975/zulip,thomasboyt/zulip,j831/zulip,wangdeshui/zulip,timabbott/zulip,levixie/zulip,vakila/zulip,akuseru/zulip,m1ssou/zulip,luyifan/zulip,bssrdf/zulip,johnnygaddarr/zulip,punchagan/zulip,vabs22/zulip,sharmaeklavya2/zulip,amallia/zulip,schatt/zulip,qq1012803704/zulip,wweiradio/zulip,showell/zulip,deer-hope/zulip,zachallaun/zulip,christi3k/zulip,Qgap/zulip,christi3k/zulip,zwily/zulip,schatt/zulip,MariaFaBella85/zulip,peguin40/zulip,ryanbackman/zulip,babbage/zulip,natanovia/zulip,wdaher/zulip,jeffcao/zulip,noroot/zulip,he15his/zulip,TigorC/zulip,babbage/zulip,tbutter/zulip,wavelets/zulip,developerfm/zulip,hj3938/zulip,sup95/zulip,wdaher/zulip,samatdav/zulip,mohsenSy/zulip,jerryge/zulip,huangkebo/zulip,adnanh/zulip,mansilladev/zulip,shaunstanislaus/zulip,Frouk/zulip,Juanvulcano/zulip,hj3938/zulip,Drooids/zulip,levixie/zulip,qq1012803704/zulip,ryanbackman/zulip,TigorC/zulip,Gabriel0402/zulip,johnnygaddarr/zulip,stamhe/zulip,m1ssou/zulip,mahim97/zulip,mdavid/zulip,blaze225/zulip,glovebx/zulip,bluesea/zulip,jphilipsen05/zulip,nicholasbs/zulip,brockwhittaker/zulip,showell/zulip,sonali0901/zulip,sonali0901/zulip,bastianh/zulip,johnnygaddarr/zulip,udxxabp/zulip,Cheppers/zulip,amanharitsh123/zulip,deer-hope/zulip,alliejones/zulip,JPJPJPOPOP/zulip,lfranchi/zulip,Suninus/zulip,ryansnowboarder/zulip,babbage/zulip,KingxBanana/zulip,bitemyapp/zulip,shrikrishnaholla/zulip,sup95/zulip,so0k/zulip,dnmfarrell/zulip,Suninus/zulip,ikasumiwt/zulip,bowlofstew/zulip,tommyip/zulip,ApsOps/zulip,suxinde2009/zulip,suxinde2009/zulip,vaidap/zulip,akuseru/zulip,seapasulli/zulip,jeffcao/zulip,themass/zulip,Batterfii/zulip,zulip/zulip,ericzhou2008/zulip,vakila/zulip,Vallher/zulip,arpith/zulip,EasonYi/zulip,ryansnowboarder/zulip,willingc/zulip,eeshangarg/zulip,deer-hope/zulip,showell/zulip,itnihao/zulip,hayderimran7/zulip,shrikrishnaholla/zulip,technicalpickles/zulip,PhilSk/zulip,Qgap/zulip,kokoar/zulip,Qgap/zulip,sup95/zulip,tiansiyuan/zulip,aps-sids/zulip,karamcnair/zulip,Batterfii/zulip,bastianh/zulip,pradiptad/zulip,codeKonami/zulip,punchagan/zulip,dxq-git/zulip,ashwinirudrappa/zulip,samatdav/zulip,hackerkid/zulip,levixie/zulip,alliejones/zulip,stamhe/zulip,bluesea/zulip,DazWorrall/zulip,zachallaun/zulip,AZtheAsian/zulip,aliceriot/zulip,yuvipanda/zulip,Juanvulcano/zulip,shrikrishnaholla/zulip,arpith/zulip,ApsOps/zulip,zofuthan/zulip,zulip/zulip,zulip/zulip,seapasulli/zulip,arpitpanwar/zulip,ericzhou2008/zulip,thomasboyt/zulip,amyliu345/zulip,hayderimran7/zulip,willingc/zulip,hafeez3000/zulip,ashwinirudrappa/zulip,aps-sids/zulip,MayB/zulip,yocome/zulip,mahim97/zulip,ikasumiwt/zulip,tbutter/zulip,samatdav/zulip,blaze225/zulip,kou/zulip,cosmicAsymmetry/zulip,ipernet/zulip,ericzhou2008/zulip,MariaFaBella85/zulip,KJin99/zulip,joshisa/zulip,ufosky-server/zulip,seapasulli/zulip,andersk/zulip,niftynei/zulip,JanzTam/zulip,Suninus/zulip,suxinde2009/zulip,eeshangarg/zulip,vabs22/zulip,rishig/zulip,tommyip/zulip,dwrpayne/zulip,seapasulli/zulip,tommyip/zulip,bowlofstew/zulip,souravbadami/zulip,synicalsyntax/zulip,saitodisse/zulip,ikasumiwt/zulip,aakash-cr7/zulip,verma-varsha/zulip,timabbott/zulip,eeshangarg/zulip,seapasulli/zulip,pradiptad/zulip,timabbott/zulip,DazWorrall/zulip,vaidap/zulip,zacps/zulip,deer-hope/zulip,Cheppers/zulip,Galexrt/zulip,eastlhu/zulip,timabbott/zulip,glovebx/zulip,adnanh/zulip,vikas-parashar/zulip,ryansnowboarder/zulip,lfranchi/zulip,willingc/zulip,AZtheAsian/zulip,swinghu/zulip,fw1121/zulip,KingxBanana/zulip,wangdeshui/zulip,bitemyapp/zulip,verma-varsha/zulip,shrikrishnaholla/zulip,TigorC/zulip,Qgap/zulip,peguin40/zulip,nicholasbs/zulip,moria/zulip,firstblade/zulip,krtkmj/zulip,Juanvulcano/zulip,zachallaun/zulip,yuvipanda/zulip,natanovia/zulip,Juanvulcano/zulip,adnanh/zulip,suxinde2009/zulip,punchagan/zulip,j831/zulip,huangkebo/zulip,jeffcao/zulip,zhaoweigg/zulip,shrikrishnaholla/zulip,jeffcao/zulip,adnanh/zulip,eastlhu/zulip,firstblade/zulip,tommyip/zulip,huangkebo/zulip,bastianh/zulip,susansls/zulip,kaiyuanheshang/zulip,aps-sids/zulip,brainwane/zulip,zorojean/zulip,suxinde2009/zulip,niftynei/zulip,LeeRisk/zulip,MariaFaBella85/zulip,jerryge/zulip,developerfm/zulip,karamcnair/zulip,themass/zulip,MayB/zulip,hafeez3000/zulip,jackrzhang/zulip,so0k/zulip,gkotian/zulip,hengqujushi/zulip,itnihao/zulip,johnnygaddarr/zulip,LAndreas/zulip,rishig/zulip,amallia/zulip,avastu/zulip,yocome/zulip,proliming/zulip,Galexrt/zulip,stamhe/zulip,firstblade/zulip,dxq-git/zulip,bluesea/zulip,dxq-git/zulip,praveenaki/zulip,bluesea/zulip,xuanhan863/zulip,Frouk/zulip,ipernet/zulip,voidException/zulip,aps-sids/zulip,natanovia/zulip,karamcnair/zulip,peguin40/zulip,wavelets/zulip,bastianh/zulip,joshisa/zulip,jrowan/zulip,eeshangarg/zulip,huangkebo/zulip,Qgap/zulip,ufosky-server/zulip,DazWorrall/zulip,dawran6/zulip,proliming/zulip,avastu/zulip,PaulPetring/zulip,umkay/zulip,hafeez3000/zulip,zulip/zulip,susansls/zulip,jphilipsen05/zulip,vaidap/zulip,Frouk/zulip,proliming/zulip,souravbadami/zulip,amyliu345/zulip,Cheppers/zulip,verma-varsha/zulip,arpith/zulip,EasonYi/zulip,ashwinirudrappa/zulip,zachallaun/zulip,dwrpayne/zulip,zacps/zulip,johnny9/zulip,jimmy54/zulip,wdaher/zulip,cosmicAsymmetry/zulip,fw1121/zulip,ryansnowboarder/zulip,jonesgithub/zulip,mdavid/zulip,yuvipanda/zulip,johnny9/zulip,Cheppers/zulip,zorojean/zulip,tbutter/zulip,sonali0901/zulip,Juanvulcano/zulip,hustlzp/zulip,reyha/zulip,m1ssou/zulip,wavelets/zulip,wangdeshui/zulip,eeshangarg/zulip,guiquanz/zulip,schatt/zulip,saitodisse/zulip,JanzTam/zulip,verma-varsha/zulip,Jianchun1/zulip,esander91/zulip,johnny9/zulip,hengqujushi/zulip,yocome/zulip,vaidap/zulip,KingxBanana/zulip,natanovia/zulip,ikasumiwt/zulip,PaulPetring/zulip,AZtheAsian/zulip,aakash-cr7/zulip,j831/zulip,yuvipanda/zulip,armooo/zulip,hafeez3000/zulip,rishig/zulip,mohsenSy/zulip,mahim97/zulip,christi3k/zulip,Diptanshu8/zulip,littledogboy/zulip,he15his/zulip,dxq-git/zulip,guiquanz/zulip,peguin40/zulip,noroot/zulip,SmartPeople/zulip,technicalpickles/zulip,KJin99/zulip,umkay/zulip,wweiradio/zulip,hengqujushi/zulip,LAndreas/zulip,zulip/zulip,tiansiyuan/zulip,babbage/zulip,themass/zulip,Drooids/zulip,LeeRisk/zulip,jainayush975/zulip,TigorC/zulip,ikasumiwt/zulip,LAndreas/zulip,Cheppers/zulip,Gabriel0402/zulip,cosmicAsymmetry/zulip,joyhchen/zulip,mahim97/zulip,paxapy/zulip,kou/zulip,SmartPeople/zulip,zofuthan/zulip,AZtheAsian/zulip,wweiradio/zulip,dnmfarrell/zulip,vaidap/zulip,rishig/zulip,krtkmj/zulip,technicalpickles/zulip,brainwane/zulip,schatt/zulip,tdr130/zulip,andersk/zulip,amyliu345/zulip,karamcnair/zulip,jonesgithub/zulip,Vallher/zulip,ryanbackman/zulip,arpitpanwar/zulip,PaulPetring/zulip,synicalsyntax/zulip,akuseru/zulip,Suninus/zulip,EasonYi/zulip,umkay/zulip,joyhchen/zulip,alliejones/zulip,pradiptad/zulip,so0k/zulip,timabbott/zulip,arpith/zulip,hustlzp/zulip,blaze225/zulip,Vallher/zulip,johnny9/zulip,umkay/zulip,armooo/zulip,hayderimran7/zulip,aakash-cr7/zulip,littledogboy/zulip,punchagan/zulip,avastu/zulip,codeKonami/zulip,verma-varsha/zulip,PhilSk/zulip,dotcool/zulip,dattatreya303/zulip,tommyip/zulip,luyifan/zulip,rht/zulip,zhaoweigg/zulip,zacps/zulip,rht/zulip,joshisa/zulip,he15his/zulip,voidException/zulip,willingc/zulip,EasonYi/zulip,voidException/zulip,zwily/zulip,lfranchi/zulip,sonali0901/zulip,esander91/zulip,joshisa/zulip,souravbadami/zulip,esander91/zulip,RobotCaleb/zulip,praveenaki/zulip,Gabriel0402/zulip,amyliu345/zulip,zachallaun/zulip,swinghu/zulip,wangdeshui/zulip,synicalsyntax/zulip,brainwane/zulip,peiwei/zulip,arpitpanwar/zulip,punchagan/zulip,souravbadami/zulip,zorojean/zulip,AZtheAsian/zulip,zofuthan/zulip,xuanhan863/zulip,KingxBanana/zulip,saitodisse/zulip,Vallher/zulip,sup95/zulip,armooo/zulip,blaze225/zulip,showell/zulip,kaiyuanheshang/zulip,jessedhillon/zulip,ryanbackman/zulip,huangkebo/zulip,dhcrzf/zulip,arpitpanwar/zulip,ipernet/zulip,joyhchen/zulip,guiquanz/zulip,atomic-labs/zulip,amyliu345/zulip,KingxBanana/zulip,reyha/zulip,krtkmj/zulip,KJin99/zulip,johnny9/zulip,Cheppers/zulip,MayB/zulip,easyfmxu/zulip,esander91/zulip,themass/zulip,Juanvulcano/zulip,dotcool/zulip,so0k/zulip,amallia/zulip,souravbadami/zulip,voidException/zulip,DazWorrall/zulip,saitodisse/zulip,so0k/zulip,ApsOps/zulip,zhaoweigg/zulip,rishig/zulip,kou/zulip,tbutter/zulip,schatt/zulip,brockwhittaker/zulip,joshisa/zulip,hj3938/zulip,developerfm/zulip,aakash-cr7/zulip,ufosky-server/zulip,dxq-git/zulip,zacps/zulip,Qgap/zulip,grave-w-grave/zulip,fw1121/zulip,PhilSk/zulip,j831/zulip,christi3k/zulip,gigawhitlocks/zulip,hayderimran7/zulip,udxxabp/zulip,synicalsyntax/zulip,JPJPJPOPOP/zulip,paxapy/zulip,hustlzp/zulip,xuanhan863/zulip,ipernet/zulip,paxapy/zulip,glovebx/zulip,cosmicAsymmetry/zulip,grave-w-grave/zulip,amallia/zulip,mansilladev/zulip,mohsenSy/zulip,RobotCaleb/zulip,aakash-cr7/zulip,shaunstanislaus/zulip,Galexrt/zulip,jonesgithub/zulip,rht/zulip,easyfmxu/zulip,wangdeshui/zulip,eeshangarg/zulip,praveenaki/zulip,rishig/zulip,dnmfarrell/zulip,noroot/zulip,ApsOps/zulip,dwrpayne/zulip,zofuthan/zulip,xuxiao/zulip,zofuthan/zulip,yocome/zulip,guiquanz/zulip,mansilladev/zulip,easyfmxu/zulip,vikas-parashar/zulip,seapasulli/zulip,johnny9/zulip,ApsOps/zulip,ipernet/zulip,glovebx/zulip,thomasboyt/zulip,LeeRisk/zulip,sup95/zulip,reyha/zulip,eastlhu/zulip,he15his/zulip,amallia/zulip,zorojean/zulip,wweiradio/zulip,MariaFaBella85/zulip,mdavid/zulip,ufosky-server/zulip,zwily/zulip,JanzTam/zulip,jackrzhang/zulip,shaunstanislaus/zulip,JPJPJPOPOP/zulip,jonesgithub/zulip,tdr130/zulip,paxapy/zulip,jainayush975/zulip,kou/zulip,tdr130/zulip,hengqujushi/zulip,hj3938/zulip,dnmfarrell/zulip,shubhamdhama/zulip,ashwinirudrappa/zulip,synicalsyntax/zulip,bitemyapp/zulip,Gabriel0402/zulip,jimmy54/zulip,lfranchi/zulip,bitemyapp/zulip,Frouk/zulip,MayB/zulip,bastianh/zulip,technicalpickles/zulip,dotcool/zulip,voidException/zulip,ufosky-server/zulip,technicalpickles/zulip,Batterfii/zulip,willingc/zulip,reyha/zulip,ahmadassaf/zulip,gkotian/zulip,LAndreas/zulip,jessedhillon/zulip,EasonYi/zulip,ipernet/zulip,jrowan/zulip,vakila/zulip,gkotian/zulip,isht3/zulip,moria/zulip,dnmfarrell/zulip,gigawhitlocks/zulip,developerfm/zulip,bitemyapp/zulip,peiwei/zulip,ryanbackman/zulip,gigawhitlocks/zulip,niftynei/zulip,praveenaki/zulip,wweiradio/zulip,tbutter/zulip,nicholasbs/zulip,bssrdf/zulip,qq1012803704/zulip,ericzhou2008/zulip,wavelets/zulip,brainwane/zulip,ahmadassaf/zulip,RobotCaleb/zulip,huangkebo/zulip,isht3/zulip,calvinleenyc/zulip,stamhe/zulip,dhcrzf/zulip,udxxabp/zulip,sharmaeklavya2/zulip,calvinleenyc/zulip,thomasboyt/zulip,aliceriot/zulip,vikas-parashar/zulip,verma-varsha/zulip,hackerkid/zulip,andersk/zulip,itnihao/zulip,krtkmj/zulip,natanovia/zulip,qq1012803704/zulip,dotcool/zulip,jerryge/zulip,pradiptad/zulip,levixie/zulip,RobotCaleb/zulip,easyfmxu/zulip,Vallher/zulip,he15his/zulip,so0k/zulip,avastu/zulip,bowlofstew/zulip,EasonYi/zulip,Batterfii/zulip,brockwhittaker/zulip,dnmfarrell/zulip,ashwinirudrappa/zulip,stamhe/zulip,peiwei/zulip,shubhamdhama/zulip,xuanhan863/zulip,brainwane/zulip,ericzhou2008/zulip,vikas-parashar/zulip,jphilipsen05/zulip,amallia/zulip,aliceriot/zulip,wangdeshui/zulip,DazWorrall/zulip,bluesea/zulip,karamcnair/zulip,zachallaun/zulip,andersk/zulip,ikasumiwt/zulip,thomasboyt/zulip,grave-w-grave/zulip,DazWorrall/zulip,littledogboy/zulip,he15his/zulip,peiwei/zulip,aakash-cr7/zulip,udxxabp/zulip,hustlzp/zulip,praveenaki/zulip,umkay/zulip,MayB/zulip,jrowan/zulip,niftynei/zulip,mansilladev/zulip,jimmy54/zulip,andersk/zulip,reyha/zulip,dwrpayne/zulip,m1ssou/zulip,kaiyuanheshang/zulip,krtkmj/zulip,mansilladev/zulip,Gabriel0402/zulip,gigawhitlocks/zulip,SmartPeople/zulip,LeeRisk/zulip,joshisa/zulip,themass/zulip,aps-sids/zulip,aliceriot/zulip,jeffcao/zulip,alliejones/zulip,hj3938/zulip,showell/zulip,Diptanshu8/zulip,bssrdf/zulip,stamhe/zulip,codeKonami/zulip,esander91/zulip,littledogboy/zulip,schatt/zulip,gkotian/zulip,KJin99/zulip,zwily/zulip,jackrzhang/zulip,armooo/zulip,mohsenSy/zulip,ApsOps/zulip,luyifan/zulip,vakila/zulip,JPJPJPOPOP/zulip,RobotCaleb/zulip,zhaoweigg/zulip,tiansiyuan/zulip,mdavid/zulip,calvinleenyc/zulip,joyhchen/zulip,developerfm/zulip,shaunstanislaus/zulip,hafeez3000/zulip,hayderimran7/zulip,isht3/zulip,bssrdf/zulip,jackrzhang/zulip,umkay/zulip,xuanhan863/zulip,themass/zulip,kokoar/zulip,cosmicAsymmetry/zulip,vakila/zulip,zacps/zulip,noroot/zulip,SmartPeople/zulip,MariaFaBella85/zulip,KingxBanana/zulip,dotcool/zulip,LAndreas/zulip,johnnygaddarr/zulip,hustlzp/zulip,proliming/zulip,jessedhillon/zulip,Diptanshu8/zulip,hayderimran7/zulip,developerfm/zulip,arpith/zulip,ahmadassaf/zulip,Vallher/zulip,RobotCaleb/zulip,bowlofstew/zulip,shubhamdhama/zulip,jonesgithub/zulip,Jianchun1/zulip,bastianh/zulip,tdr130/zulip,aps-sids/zulip,qq1012803704/zulip,praveenaki/zulip,xuanhan863/zulip,Frouk/zulip,Drooids/zulip,xuxiao/zulip,willingc/zulip,MariaFaBella85/zulip,moria/zulip,krtkmj/zulip,moria/zulip,dattatreya303/zulip,Drooids/zulip,PhilSk/zulip,jimmy54/zulip,shaunstanislaus/zulip,Gabriel0402/zulip,hustlzp/zulip,shaunstanislaus/zulip,amyliu345/zulip,Diptanshu8/zulip,pradiptad/zulip,dhcrzf/zulip,jainayush975/zulip,timabbott/zulip,dhcrzf/zulip,babbage/zulip,PaulPetring/zulip,ahmadassaf/zulip,zwily/zulip,brainwane/zulip,samatdav/zulip,he15his/zulip,shrikrishnaholla/zulip,qq1012803704/zulip,qq1012803704/zulip,tdr130/zulip,kaiyuanheshang/zulip,peguin40/zulip,MariaFaBella85/zulip,Gabriel0402/zulip,PaulPetring/zulip,hj3938/zulip,zorojean/zulip,kaiyuanheshang/zulip,dhcrzf/zulip,amanharitsh123/zulip,dotcool/zulip,adnanh/zulip,aliceriot/zulip,technicalpickles/zulip,sonali0901/zulip,brockwhittaker/zulip,JanzTam/zulip,KJin99/zulip,hengqujushi/zulip,udxxabp/zulip,codeKonami/zulip,timabbott/zulip,SmartPeople/zulip,susansls/zulip,vakila/zulip,xuxiao/zulip,blaze225/zulip,hengqujushi/zulip,natanovia/zulip,jrowan/zulip,peiwei/zulip,aliceriot/zulip,nicholasbs/zulip,rht/zulip,amanharitsh123/zulip,vabs22/zulip,adnanh/zulip,m1ssou/zulip,Batterfii/zulip,yocome/zulip,tommyip/zulip,mdavid/zulip,yuvipanda/zulip,LAndreas/zulip,atomic-labs/zulip,ryansnowboarder/zulip,Jianchun1/zulip,bowlofstew/zulip,shubhamdhama/zulip,isht3/zulip,Diptanshu8/zulip,jphilipsen05/zulip,luyifan/zulip,yocome/zulip,vabs22/zulip,tiansiyuan/zulip,dxq-git/zulip,vakila/zulip,atomic-labs/zulip,hafeez3000/zulip,Diptanshu8/zulip,eeshangarg/zulip,MayB/zulip,mdavid/zulip,tiansiyuan/zulip,SmartPeople/zulip,ikasumiwt/zulip,udxxabp/zulip,brockwhittaker/zulip,blaze225/zulip,bitemyapp/zulip,suxinde2009/zulip,tiansiyuan/zulip,wweiradio/zulip,j831/zulip,mansilladev/zulip,susansls/zulip,ashwinirudrappa/zulip,luyifan/zulip,PhilSk/zulip,voidException/zulip,jonesgithub/zulip,aliceriot/zulip,calvinleenyc/zulip,avastu/zulip,zulip/zulip,synicalsyntax/zulip,bowlofstew/zulip,m1ssou/zulip,Jianchun1/zulip,codeKonami/zulip,dwrpayne/zulip,lfranchi/zulip,jphilipsen05/zulip,suxinde2009/zulip,babbage/zulip,vikas-parashar/zulip,kaiyuanheshang/zulip,JanzTam/zulip,jackrzhang/zulip,mohsenSy/zulip,niftynei/zulip,firstblade/zulip,yocome/zulip,jessedhillon/zulip,grave-w-grave/zulip,LeeRisk/zulip,jerryge/zulip,Batterfii/zulip,punchagan/zulip,avastu/zulip,sharmaeklavya2/zulip,bowlofstew/zulip,ericzhou2008/zulip,amanharitsh123/zulip,hustlzp/zulip,jessedhillon/zulip,JanzTam/zulip,saitodisse/zulip,bssrdf/zulip,shrikrishnaholla/zulip,sup95/zulip,gigawhitlocks/zulip,amanharitsh123/zulip,kokoar/zulip,tommyip/zulip,wdaher/zulip,brainwane/zulip,swinghu/zulip,xuanhan863/zulip,huangkebo/zulip,ufosky-server/zulip,brockwhittaker/zulip,atomic-labs/zulip,jimmy54/zulip,Suninus/zulip,zofuthan/zulip,jainayush975/zulip,littledogboy/zulip,wavelets/zulip,wavelets/zulip,yuvipanda/zulip,moria/zulip,alliejones/zulip,Batterfii/zulip,johnnygaddarr/zulip,TigorC/zulip,christi3k/zulip,glovebx/zulip,tiansiyuan/zulip,willingc/zulip,umkay/zulip,swinghu/zulip,jerryge/zulip,Galexrt/zulip,joyhchen/zulip,thomasboyt/zulip,Jianchun1/zulip,natanovia/zulip,DazWorrall/zulip,zorojean/zulip,nicholasbs/zulip,cosmicAsymmetry/zulip,gkotian/zulip,arpitpanwar/zulip,bssrdf/zulip,zofuthan/zulip,dawran6/zulip,itnihao/zulip,tdr130/zulip,mahim97/zulip,jeffcao/zulip,isht3/zulip,JPJPJPOPOP/zulip,bastianh/zulip,alliejones/zulip,m1ssou/zulip,vaidap/zulip,andersk/zulip,arpitpanwar/zulip,zhaoweigg/zulip,dattatreya303/zulip,littledogboy/zulip,calvinleenyc/zulip,xuxiao/zulip,akuseru/zulip,tdr130/zulip,isht3/zulip,dattatreya303/zulip,pradiptad/zulip,alliejones/zulip,themass/zulip,susansls/zulip,codeKonami/zulip,littledogboy/zulip,technicalpickles/zulip,dawran6/zulip,hengqujushi/zulip,developerfm/zulip,dhcrzf/zulip,hackerkid/zulip,kokoar/zulip,babbage/zulip,kou/zulip,itnihao/zulip,hj3938/zulip,vabs22/zulip,atomic-labs/zulip,fw1121/zulip,praveenaki/zulip,xuxiao/zulip,JPJPJPOPOP/zulip,rishig/zulip,moria/zulip,jphilipsen05/zulip,calvinleenyc/zulip,EasonYi/zulip,JanzTam/zulip,easyfmxu/zulip,zhaoweigg/zulip,grave-w-grave/zulip,swinghu/zulip,aps-sids/zulip,Jianchun1/zulip,joyhchen/zulip,Qgap/zulip,levixie/zulip,christi3k/zulip,wdaher/zulip,levixie/zulip,sharmaeklavya2/zulip,proliming/zulip,zacps/zulip,jessedhillon/zulip,tbutter/zulip,mohsenSy/zulip,ahmadassaf/zulip,bssrdf/zulip,arpitpanwar/zulip,wavelets/zulip,paxapy/zulip,bitemyapp/zulip,souravbadami/zulip,guiquanz/zulip,jimmy54/zulip,nicholasbs/zulip,bluesea/zulip,j831/zulip,jrowan/zulip,KJin99/zulip,Galexrt/zulip,gkotian/zulip,shubhamdhama/zulip,LAndreas/zulip,glovebx/zulip,easyfmxu/zulip,peguin40/zulip,dattatreya303/zulip,ericzhou2008/zulip,glovebx/zulip,deer-hope/zulip,wdaher/zulip,RobotCaleb/zulip,avastu/zulip,eastlhu/zulip,armooo/zulip,guiquanz/zulip,wangdeshui/zulip,MayB/zulip,gkotian/zulip,yuvipanda/zulip,zhaoweigg/zulip,LeeRisk/zulip,karamcnair/zulip,itnihao/zulip,tbutter/zulip,Galexrt/zulip,ryansnowboarder/zulip,niftynei/zulip
# -*- coding: utf-8 -*- # Copyright: (c) 2008, Jarek Zgoda <jarek.zgoda@gmail.com> __revision__ = '$Id: views.py 21 2008-12-05 09:21:03Z jarek.zgoda $' from django.shortcuts import render_to_response from django.template import RequestContext from django.conf import settings from confirmation.models import Confirmation def confirm(request, confirmation_key): confirmation_key = confirmation_key.lower() obj = Confirmation.objects.confirm(confirmation_key) confirmed = True if not obj: # confirmation failed confirmed = False try: # try to get the object we was supposed to confirm obj = Confirmation.objects.get(confirmation_key=confirmation_key) except Confirmation.DoesNotExist: pass ctx = { 'object': obj, 'confirmed': confirmed, 'days': getattr(settings, 'EMAIL_CONFIRMATION_DAYS', 10), } templates = [ 'confirmation/confirm.html', ] if obj: # if we have an object, we can use specific template templates.insert(0, 'confirmation/confirm_%s.html' % obj._meta.module_name) return render_to_response(templates, ctx, context_instance=RequestContext(request)) Include confirmation key in context object. This way our templates can reference the confirmation key later. (imported from commit 4d57e1309386f2236829b6fdf4e4ad43c5b125c8)
# -*- coding: utf-8 -*- # Copyright: (c) 2008, Jarek Zgoda <jarek.zgoda@gmail.com> __revision__ = '$Id: views.py 21 2008-12-05 09:21:03Z jarek.zgoda $' from django.shortcuts import render_to_response from django.template import RequestContext from django.conf import settings from confirmation.models import Confirmation def confirm(request, confirmation_key): confirmation_key = confirmation_key.lower() obj = Confirmation.objects.confirm(confirmation_key) confirmed = True if not obj: # confirmation failed confirmed = False try: # try to get the object we was supposed to confirm obj = Confirmation.objects.get(confirmation_key=confirmation_key) except Confirmation.DoesNotExist: pass ctx = { 'object': obj, 'confirmed': confirmed, 'days': getattr(settings, 'EMAIL_CONFIRMATION_DAYS', 10), 'key': confirmation_key, } templates = [ 'confirmation/confirm.html', ] if obj: # if we have an object, we can use specific template templates.insert(0, 'confirmation/confirm_%s.html' % obj._meta.module_name) return render_to_response(templates, ctx, context_instance=RequestContext(request))
<commit_before># -*- coding: utf-8 -*- # Copyright: (c) 2008, Jarek Zgoda <jarek.zgoda@gmail.com> __revision__ = '$Id: views.py 21 2008-12-05 09:21:03Z jarek.zgoda $' from django.shortcuts import render_to_response from django.template import RequestContext from django.conf import settings from confirmation.models import Confirmation def confirm(request, confirmation_key): confirmation_key = confirmation_key.lower() obj = Confirmation.objects.confirm(confirmation_key) confirmed = True if not obj: # confirmation failed confirmed = False try: # try to get the object we was supposed to confirm obj = Confirmation.objects.get(confirmation_key=confirmation_key) except Confirmation.DoesNotExist: pass ctx = { 'object': obj, 'confirmed': confirmed, 'days': getattr(settings, 'EMAIL_CONFIRMATION_DAYS', 10), } templates = [ 'confirmation/confirm.html', ] if obj: # if we have an object, we can use specific template templates.insert(0, 'confirmation/confirm_%s.html' % obj._meta.module_name) return render_to_response(templates, ctx, context_instance=RequestContext(request)) <commit_msg>Include confirmation key in context object. This way our templates can reference the confirmation key later. (imported from commit 4d57e1309386f2236829b6fdf4e4ad43c5b125c8)<commit_after>
# -*- coding: utf-8 -*- # Copyright: (c) 2008, Jarek Zgoda <jarek.zgoda@gmail.com> __revision__ = '$Id: views.py 21 2008-12-05 09:21:03Z jarek.zgoda $' from django.shortcuts import render_to_response from django.template import RequestContext from django.conf import settings from confirmation.models import Confirmation def confirm(request, confirmation_key): confirmation_key = confirmation_key.lower() obj = Confirmation.objects.confirm(confirmation_key) confirmed = True if not obj: # confirmation failed confirmed = False try: # try to get the object we was supposed to confirm obj = Confirmation.objects.get(confirmation_key=confirmation_key) except Confirmation.DoesNotExist: pass ctx = { 'object': obj, 'confirmed': confirmed, 'days': getattr(settings, 'EMAIL_CONFIRMATION_DAYS', 10), 'key': confirmation_key, } templates = [ 'confirmation/confirm.html', ] if obj: # if we have an object, we can use specific template templates.insert(0, 'confirmation/confirm_%s.html' % obj._meta.module_name) return render_to_response(templates, ctx, context_instance=RequestContext(request))
# -*- coding: utf-8 -*- # Copyright: (c) 2008, Jarek Zgoda <jarek.zgoda@gmail.com> __revision__ = '$Id: views.py 21 2008-12-05 09:21:03Z jarek.zgoda $' from django.shortcuts import render_to_response from django.template import RequestContext from django.conf import settings from confirmation.models import Confirmation def confirm(request, confirmation_key): confirmation_key = confirmation_key.lower() obj = Confirmation.objects.confirm(confirmation_key) confirmed = True if not obj: # confirmation failed confirmed = False try: # try to get the object we was supposed to confirm obj = Confirmation.objects.get(confirmation_key=confirmation_key) except Confirmation.DoesNotExist: pass ctx = { 'object': obj, 'confirmed': confirmed, 'days': getattr(settings, 'EMAIL_CONFIRMATION_DAYS', 10), } templates = [ 'confirmation/confirm.html', ] if obj: # if we have an object, we can use specific template templates.insert(0, 'confirmation/confirm_%s.html' % obj._meta.module_name) return render_to_response(templates, ctx, context_instance=RequestContext(request)) Include confirmation key in context object. This way our templates can reference the confirmation key later. (imported from commit 4d57e1309386f2236829b6fdf4e4ad43c5b125c8)# -*- coding: utf-8 -*- # Copyright: (c) 2008, Jarek Zgoda <jarek.zgoda@gmail.com> __revision__ = '$Id: views.py 21 2008-12-05 09:21:03Z jarek.zgoda $' from django.shortcuts import render_to_response from django.template import RequestContext from django.conf import settings from confirmation.models import Confirmation def confirm(request, confirmation_key): confirmation_key = confirmation_key.lower() obj = Confirmation.objects.confirm(confirmation_key) confirmed = True if not obj: # confirmation failed confirmed = False try: # try to get the object we was supposed to confirm obj = Confirmation.objects.get(confirmation_key=confirmation_key) except Confirmation.DoesNotExist: pass ctx = { 'object': obj, 'confirmed': confirmed, 'days': getattr(settings, 'EMAIL_CONFIRMATION_DAYS', 10), 'key': confirmation_key, } templates = [ 'confirmation/confirm.html', ] if obj: # if we have an object, we can use specific template templates.insert(0, 'confirmation/confirm_%s.html' % obj._meta.module_name) return render_to_response(templates, ctx, context_instance=RequestContext(request))
<commit_before># -*- coding: utf-8 -*- # Copyright: (c) 2008, Jarek Zgoda <jarek.zgoda@gmail.com> __revision__ = '$Id: views.py 21 2008-12-05 09:21:03Z jarek.zgoda $' from django.shortcuts import render_to_response from django.template import RequestContext from django.conf import settings from confirmation.models import Confirmation def confirm(request, confirmation_key): confirmation_key = confirmation_key.lower() obj = Confirmation.objects.confirm(confirmation_key) confirmed = True if not obj: # confirmation failed confirmed = False try: # try to get the object we was supposed to confirm obj = Confirmation.objects.get(confirmation_key=confirmation_key) except Confirmation.DoesNotExist: pass ctx = { 'object': obj, 'confirmed': confirmed, 'days': getattr(settings, 'EMAIL_CONFIRMATION_DAYS', 10), } templates = [ 'confirmation/confirm.html', ] if obj: # if we have an object, we can use specific template templates.insert(0, 'confirmation/confirm_%s.html' % obj._meta.module_name) return render_to_response(templates, ctx, context_instance=RequestContext(request)) <commit_msg>Include confirmation key in context object. This way our templates can reference the confirmation key later. (imported from commit 4d57e1309386f2236829b6fdf4e4ad43c5b125c8)<commit_after># -*- coding: utf-8 -*- # Copyright: (c) 2008, Jarek Zgoda <jarek.zgoda@gmail.com> __revision__ = '$Id: views.py 21 2008-12-05 09:21:03Z jarek.zgoda $' from django.shortcuts import render_to_response from django.template import RequestContext from django.conf import settings from confirmation.models import Confirmation def confirm(request, confirmation_key): confirmation_key = confirmation_key.lower() obj = Confirmation.objects.confirm(confirmation_key) confirmed = True if not obj: # confirmation failed confirmed = False try: # try to get the object we was supposed to confirm obj = Confirmation.objects.get(confirmation_key=confirmation_key) except Confirmation.DoesNotExist: pass ctx = { 'object': obj, 'confirmed': confirmed, 'days': getattr(settings, 'EMAIL_CONFIRMATION_DAYS', 10), 'key': confirmation_key, } templates = [ 'confirmation/confirm.html', ] if obj: # if we have an object, we can use specific template templates.insert(0, 'confirmation/confirm_%s.html' % obj._meta.module_name) return render_to_response(templates, ctx, context_instance=RequestContext(request))
6ea9492ae32ad744da4803dcab3cf57334dd69e5
script_helpers.py
script_helpers.py
"""A set of functions to standardize some options for python scripts.""" def setup_parser_help(parser, additional_docs=None): """ Set formatting for parser to raw and add docstring to help output Parameters ---------- parser : `ArgumentParser` The parser to be modified. additional_docs: str Any documentation to be added to the documentation produced by `argparse` """ from argparse import RawDescriptionHelpFormatter parser.formatter_class = RawDescriptionHelpFormatter if additional_docs is not None: parser.epilog = additional_docs def add_verbose(parser): """ Add a verbose option (--verbose or -v) to parser. Parameters: ----------- parser : `ArgumentParser` """ verbose_help = "provide more information during processing" parser.add_argument("-v", "--verbose", help=verbose_help, action="store_true")
"""A set of functions to standardize some options for python scripts.""" def setup_parser_help(parser, additional_docs=None): """ Set formatting for parser to raw and add docstring to help output Parameters ---------- parser : `ArgumentParser` The parser to be modified. additional_docs: str Any documentation to be added to the documentation produced by `argparse` """ from argparse import RawDescriptionHelpFormatter parser.formatter_class = RawDescriptionHelpFormatter if additional_docs is not None: parser.epilog = additional_docs def add_verbose(parser): """ Add a verbose option (--verbose or -v) to parser. Parameters: ----------- parser : `ArgumentParser` """ verbose_help = "provide more information during processing" parser.add_argument("-v", "--verbose", help=verbose_help, action="store_true") def add_directories(parser): """ Add a positional argument that is one or more directories. Parameters ---------- parser : `ArgumentParser` """ parser.add_argument("dir", metavar='dir', nargs='+', help="Directory to process")
Change name of and add documentation to directory argument
Change name of and add documentation to directory argument
Python
bsd-3-clause
mwcraig/msumastro
"""A set of functions to standardize some options for python scripts.""" def setup_parser_help(parser, additional_docs=None): """ Set formatting for parser to raw and add docstring to help output Parameters ---------- parser : `ArgumentParser` The parser to be modified. additional_docs: str Any documentation to be added to the documentation produced by `argparse` """ from argparse import RawDescriptionHelpFormatter parser.formatter_class = RawDescriptionHelpFormatter if additional_docs is not None: parser.epilog = additional_docs def add_verbose(parser): """ Add a verbose option (--verbose or -v) to parser. Parameters: ----------- parser : `ArgumentParser` """ verbose_help = "provide more information during processing" parser.add_argument("-v", "--verbose", help=verbose_help, action="store_true") Change name of and add documentation to directory argument
"""A set of functions to standardize some options for python scripts.""" def setup_parser_help(parser, additional_docs=None): """ Set formatting for parser to raw and add docstring to help output Parameters ---------- parser : `ArgumentParser` The parser to be modified. additional_docs: str Any documentation to be added to the documentation produced by `argparse` """ from argparse import RawDescriptionHelpFormatter parser.formatter_class = RawDescriptionHelpFormatter if additional_docs is not None: parser.epilog = additional_docs def add_verbose(parser): """ Add a verbose option (--verbose or -v) to parser. Parameters: ----------- parser : `ArgumentParser` """ verbose_help = "provide more information during processing" parser.add_argument("-v", "--verbose", help=verbose_help, action="store_true") def add_directories(parser): """ Add a positional argument that is one or more directories. Parameters ---------- parser : `ArgumentParser` """ parser.add_argument("dir", metavar='dir', nargs='+', help="Directory to process")
<commit_before>"""A set of functions to standardize some options for python scripts.""" def setup_parser_help(parser, additional_docs=None): """ Set formatting for parser to raw and add docstring to help output Parameters ---------- parser : `ArgumentParser` The parser to be modified. additional_docs: str Any documentation to be added to the documentation produced by `argparse` """ from argparse import RawDescriptionHelpFormatter parser.formatter_class = RawDescriptionHelpFormatter if additional_docs is not None: parser.epilog = additional_docs def add_verbose(parser): """ Add a verbose option (--verbose or -v) to parser. Parameters: ----------- parser : `ArgumentParser` """ verbose_help = "provide more information during processing" parser.add_argument("-v", "--verbose", help=verbose_help, action="store_true") <commit_msg>Change name of and add documentation to directory argument<commit_after>
"""A set of functions to standardize some options for python scripts.""" def setup_parser_help(parser, additional_docs=None): """ Set formatting for parser to raw and add docstring to help output Parameters ---------- parser : `ArgumentParser` The parser to be modified. additional_docs: str Any documentation to be added to the documentation produced by `argparse` """ from argparse import RawDescriptionHelpFormatter parser.formatter_class = RawDescriptionHelpFormatter if additional_docs is not None: parser.epilog = additional_docs def add_verbose(parser): """ Add a verbose option (--verbose or -v) to parser. Parameters: ----------- parser : `ArgumentParser` """ verbose_help = "provide more information during processing" parser.add_argument("-v", "--verbose", help=verbose_help, action="store_true") def add_directories(parser): """ Add a positional argument that is one or more directories. Parameters ---------- parser : `ArgumentParser` """ parser.add_argument("dir", metavar='dir', nargs='+', help="Directory to process")
"""A set of functions to standardize some options for python scripts.""" def setup_parser_help(parser, additional_docs=None): """ Set formatting for parser to raw and add docstring to help output Parameters ---------- parser : `ArgumentParser` The parser to be modified. additional_docs: str Any documentation to be added to the documentation produced by `argparse` """ from argparse import RawDescriptionHelpFormatter parser.formatter_class = RawDescriptionHelpFormatter if additional_docs is not None: parser.epilog = additional_docs def add_verbose(parser): """ Add a verbose option (--verbose or -v) to parser. Parameters: ----------- parser : `ArgumentParser` """ verbose_help = "provide more information during processing" parser.add_argument("-v", "--verbose", help=verbose_help, action="store_true") Change name of and add documentation to directory argument"""A set of functions to standardize some options for python scripts.""" def setup_parser_help(parser, additional_docs=None): """ Set formatting for parser to raw and add docstring to help output Parameters ---------- parser : `ArgumentParser` The parser to be modified. additional_docs: str Any documentation to be added to the documentation produced by `argparse` """ from argparse import RawDescriptionHelpFormatter parser.formatter_class = RawDescriptionHelpFormatter if additional_docs is not None: parser.epilog = additional_docs def add_verbose(parser): """ Add a verbose option (--verbose or -v) to parser. Parameters: ----------- parser : `ArgumentParser` """ verbose_help = "provide more information during processing" parser.add_argument("-v", "--verbose", help=verbose_help, action="store_true") def add_directories(parser): """ Add a positional argument that is one or more directories. Parameters ---------- parser : `ArgumentParser` """ parser.add_argument("dir", metavar='dir', nargs='+', help="Directory to process")
<commit_before>"""A set of functions to standardize some options for python scripts.""" def setup_parser_help(parser, additional_docs=None): """ Set formatting for parser to raw and add docstring to help output Parameters ---------- parser : `ArgumentParser` The parser to be modified. additional_docs: str Any documentation to be added to the documentation produced by `argparse` """ from argparse import RawDescriptionHelpFormatter parser.formatter_class = RawDescriptionHelpFormatter if additional_docs is not None: parser.epilog = additional_docs def add_verbose(parser): """ Add a verbose option (--verbose or -v) to parser. Parameters: ----------- parser : `ArgumentParser` """ verbose_help = "provide more information during processing" parser.add_argument("-v", "--verbose", help=verbose_help, action="store_true") <commit_msg>Change name of and add documentation to directory argument<commit_after>"""A set of functions to standardize some options for python scripts.""" def setup_parser_help(parser, additional_docs=None): """ Set formatting for parser to raw and add docstring to help output Parameters ---------- parser : `ArgumentParser` The parser to be modified. additional_docs: str Any documentation to be added to the documentation produced by `argparse` """ from argparse import RawDescriptionHelpFormatter parser.formatter_class = RawDescriptionHelpFormatter if additional_docs is not None: parser.epilog = additional_docs def add_verbose(parser): """ Add a verbose option (--verbose or -v) to parser. Parameters: ----------- parser : `ArgumentParser` """ verbose_help = "provide more information during processing" parser.add_argument("-v", "--verbose", help=verbose_help, action="store_true") def add_directories(parser): """ Add a positional argument that is one or more directories. Parameters ---------- parser : `ArgumentParser` """ parser.add_argument("dir", metavar='dir', nargs='+', help="Directory to process")
07538222d07b0a565cadadd40df9bedeb12a4f60
talk_timer.py
talk_timer.py
#!/usr/bin/env python2.7 # -*- coding: utf-8 -*- import argparse from time import sleep import pyttsx parser = argparse.ArgumentParser(description='Presentation timer with spoken warnings.') parser.add_argument('minutes', type=int, nargs=1, help='Length of talk in minutes') args = parser.parse_args() minutes = args.minutes[0] print "Timing a {0}-minute talk.".format(minutes) # Sleep for X-minute talk duration sleep(minutes * 60) # Text to speech engine = pyttsx.init() engine.say("Time is up") engine.runAndWait() print("Time is up")
#!/usr/bin/env python2.7 # -*- coding: utf-8 -*- import argparse from time import sleep import pyttsx parser = argparse.ArgumentParser(description='Presentation timer with spoken warnings.') parser.add_argument('minutes', type=int, help='Length of talk in minutes') args = parser.parse_args() print "Timing a {0}-minute talk.".format(args.minutes) # Sleep for X-minute talk duration sleep(args.minutes * 60) # Text to speech engine = pyttsx.init() engine.say("Time is up") engine.runAndWait() print("Time is up")
Remove nargs param so that minutes is an int, not list.
Remove nargs param so that minutes is an int, not list.
Python
bsd-3-clause
audreyr/useful
#!/usr/bin/env python2.7 # -*- coding: utf-8 -*- import argparse from time import sleep import pyttsx parser = argparse.ArgumentParser(description='Presentation timer with spoken warnings.') parser.add_argument('minutes', type=int, nargs=1, help='Length of talk in minutes') args = parser.parse_args() minutes = args.minutes[0] print "Timing a {0}-minute talk.".format(minutes) # Sleep for X-minute talk duration sleep(minutes * 60) # Text to speech engine = pyttsx.init() engine.say("Time is up") engine.runAndWait() print("Time is up") Remove nargs param so that minutes is an int, not list.
#!/usr/bin/env python2.7 # -*- coding: utf-8 -*- import argparse from time import sleep import pyttsx parser = argparse.ArgumentParser(description='Presentation timer with spoken warnings.') parser.add_argument('minutes', type=int, help='Length of talk in minutes') args = parser.parse_args() print "Timing a {0}-minute talk.".format(args.minutes) # Sleep for X-minute talk duration sleep(args.minutes * 60) # Text to speech engine = pyttsx.init() engine.say("Time is up") engine.runAndWait() print("Time is up")
<commit_before>#!/usr/bin/env python2.7 # -*- coding: utf-8 -*- import argparse from time import sleep import pyttsx parser = argparse.ArgumentParser(description='Presentation timer with spoken warnings.') parser.add_argument('minutes', type=int, nargs=1, help='Length of talk in minutes') args = parser.parse_args() minutes = args.minutes[0] print "Timing a {0}-minute talk.".format(minutes) # Sleep for X-minute talk duration sleep(minutes * 60) # Text to speech engine = pyttsx.init() engine.say("Time is up") engine.runAndWait() print("Time is up") <commit_msg>Remove nargs param so that minutes is an int, not list.<commit_after>
#!/usr/bin/env python2.7 # -*- coding: utf-8 -*- import argparse from time import sleep import pyttsx parser = argparse.ArgumentParser(description='Presentation timer with spoken warnings.') parser.add_argument('minutes', type=int, help='Length of talk in minutes') args = parser.parse_args() print "Timing a {0}-minute talk.".format(args.minutes) # Sleep for X-minute talk duration sleep(args.minutes * 60) # Text to speech engine = pyttsx.init() engine.say("Time is up") engine.runAndWait() print("Time is up")
#!/usr/bin/env python2.7 # -*- coding: utf-8 -*- import argparse from time import sleep import pyttsx parser = argparse.ArgumentParser(description='Presentation timer with spoken warnings.') parser.add_argument('minutes', type=int, nargs=1, help='Length of talk in minutes') args = parser.parse_args() minutes = args.minutes[0] print "Timing a {0}-minute talk.".format(minutes) # Sleep for X-minute talk duration sleep(minutes * 60) # Text to speech engine = pyttsx.init() engine.say("Time is up") engine.runAndWait() print("Time is up") Remove nargs param so that minutes is an int, not list.#!/usr/bin/env python2.7 # -*- coding: utf-8 -*- import argparse from time import sleep import pyttsx parser = argparse.ArgumentParser(description='Presentation timer with spoken warnings.') parser.add_argument('minutes', type=int, help='Length of talk in minutes') args = parser.parse_args() print "Timing a {0}-minute talk.".format(args.minutes) # Sleep for X-minute talk duration sleep(args.minutes * 60) # Text to speech engine = pyttsx.init() engine.say("Time is up") engine.runAndWait() print("Time is up")
<commit_before>#!/usr/bin/env python2.7 # -*- coding: utf-8 -*- import argparse from time import sleep import pyttsx parser = argparse.ArgumentParser(description='Presentation timer with spoken warnings.') parser.add_argument('minutes', type=int, nargs=1, help='Length of talk in minutes') args = parser.parse_args() minutes = args.minutes[0] print "Timing a {0}-minute talk.".format(minutes) # Sleep for X-minute talk duration sleep(minutes * 60) # Text to speech engine = pyttsx.init() engine.say("Time is up") engine.runAndWait() print("Time is up") <commit_msg>Remove nargs param so that minutes is an int, not list.<commit_after>#!/usr/bin/env python2.7 # -*- coding: utf-8 -*- import argparse from time import sleep import pyttsx parser = argparse.ArgumentParser(description='Presentation timer with spoken warnings.') parser.add_argument('minutes', type=int, help='Length of talk in minutes') args = parser.parse_args() print "Timing a {0}-minute talk.".format(args.minutes) # Sleep for X-minute talk duration sleep(args.minutes * 60) # Text to speech engine = pyttsx.init() engine.say("Time is up") engine.runAndWait() print("Time is up")
b06eb92ec878a06a2fa1ce9b7eb4d253d5481daa
tests/activity/test_activity_deposit_assets.py
tests/activity/test_activity_deposit_assets.py
import unittest from activity.activity_DepositAssets import activity_DepositAssets import settings_mock from ddt import ddt, data, unpack @ddt class TestDepositAssets(unittest.TestCase): def setUp(self): self.depositassets = activity_DepositAssets(settings_mock, None, None, None, None) @unpack @data({'input': '.tif', 'expected': ['.tif']}, {'input': '.jpg, .tiff, .png', 'expected':['.jpg', '.tiff', '.png']}) def test_get_no_download_extensions(self, input, expected): result = self.depositassets.get_no_download_extensions(input) self.assertListEqual(result, expected) @unpack @data( (None, None), ('image.jpg', 'image/jpeg'), ('/folder/file.pdf.zip', 'application/x-zip-compressed'), ('/folder/weird_file.wdl', 'binary/octet-stream'), ('a_file', 'binary/octet-stream') ) def test_content_type_from_file_name(self, input, expected): result = self.depositassets.content_type_from_file_name(input) self.assertEqual(result, expected) if __name__ == '__main__': unittest.main()
import unittest from activity.activity_DepositAssets import activity_DepositAssets import settings_mock from ddt import ddt, data, unpack @ddt class TestDepositAssets(unittest.TestCase): def setUp(self): self.depositassets = activity_DepositAssets(settings_mock, None, None, None, None) @unpack @data({'input': '.tif', 'expected': ['.tif']}, {'input': '.jpg, .tiff, .png', 'expected':['.jpg', '.tiff', '.png']}) def test_get_no_download_extensions(self, input, expected): result = self.depositassets.get_no_download_extensions(input) self.assertListEqual(result, expected) @unpack @data( (None, None), ('image.jpg', 'image/jpeg'), ('/folder/file.test.pdf', 'application/pdf'), ('/folder/weird_file.wdl', 'binary/octet-stream'), ('a_file', 'binary/octet-stream') ) def test_content_type_from_file_name(self, input, expected): result = self.depositassets.content_type_from_file_name(input) self.assertEqual(result, expected) if __name__ == '__main__': unittest.main()
Change test data for mimetype to content_type.
Change test data for mimetype to content_type.
Python
mit
gnott/elife-bot,gnott/elife-bot,jhroot/elife-bot,gnott/elife-bot,jhroot/elife-bot,jhroot/elife-bot
import unittest from activity.activity_DepositAssets import activity_DepositAssets import settings_mock from ddt import ddt, data, unpack @ddt class TestDepositAssets(unittest.TestCase): def setUp(self): self.depositassets = activity_DepositAssets(settings_mock, None, None, None, None) @unpack @data({'input': '.tif', 'expected': ['.tif']}, {'input': '.jpg, .tiff, .png', 'expected':['.jpg', '.tiff', '.png']}) def test_get_no_download_extensions(self, input, expected): result = self.depositassets.get_no_download_extensions(input) self.assertListEqual(result, expected) @unpack @data( (None, None), ('image.jpg', 'image/jpeg'), ('/folder/file.pdf.zip', 'application/x-zip-compressed'), ('/folder/weird_file.wdl', 'binary/octet-stream'), ('a_file', 'binary/octet-stream') ) def test_content_type_from_file_name(self, input, expected): result = self.depositassets.content_type_from_file_name(input) self.assertEqual(result, expected) if __name__ == '__main__': unittest.main() Change test data for mimetype to content_type.
import unittest from activity.activity_DepositAssets import activity_DepositAssets import settings_mock from ddt import ddt, data, unpack @ddt class TestDepositAssets(unittest.TestCase): def setUp(self): self.depositassets = activity_DepositAssets(settings_mock, None, None, None, None) @unpack @data({'input': '.tif', 'expected': ['.tif']}, {'input': '.jpg, .tiff, .png', 'expected':['.jpg', '.tiff', '.png']}) def test_get_no_download_extensions(self, input, expected): result = self.depositassets.get_no_download_extensions(input) self.assertListEqual(result, expected) @unpack @data( (None, None), ('image.jpg', 'image/jpeg'), ('/folder/file.test.pdf', 'application/pdf'), ('/folder/weird_file.wdl', 'binary/octet-stream'), ('a_file', 'binary/octet-stream') ) def test_content_type_from_file_name(self, input, expected): result = self.depositassets.content_type_from_file_name(input) self.assertEqual(result, expected) if __name__ == '__main__': unittest.main()
<commit_before>import unittest from activity.activity_DepositAssets import activity_DepositAssets import settings_mock from ddt import ddt, data, unpack @ddt class TestDepositAssets(unittest.TestCase): def setUp(self): self.depositassets = activity_DepositAssets(settings_mock, None, None, None, None) @unpack @data({'input': '.tif', 'expected': ['.tif']}, {'input': '.jpg, .tiff, .png', 'expected':['.jpg', '.tiff', '.png']}) def test_get_no_download_extensions(self, input, expected): result = self.depositassets.get_no_download_extensions(input) self.assertListEqual(result, expected) @unpack @data( (None, None), ('image.jpg', 'image/jpeg'), ('/folder/file.pdf.zip', 'application/x-zip-compressed'), ('/folder/weird_file.wdl', 'binary/octet-stream'), ('a_file', 'binary/octet-stream') ) def test_content_type_from_file_name(self, input, expected): result = self.depositassets.content_type_from_file_name(input) self.assertEqual(result, expected) if __name__ == '__main__': unittest.main() <commit_msg>Change test data for mimetype to content_type.<commit_after>
import unittest from activity.activity_DepositAssets import activity_DepositAssets import settings_mock from ddt import ddt, data, unpack @ddt class TestDepositAssets(unittest.TestCase): def setUp(self): self.depositassets = activity_DepositAssets(settings_mock, None, None, None, None) @unpack @data({'input': '.tif', 'expected': ['.tif']}, {'input': '.jpg, .tiff, .png', 'expected':['.jpg', '.tiff', '.png']}) def test_get_no_download_extensions(self, input, expected): result = self.depositassets.get_no_download_extensions(input) self.assertListEqual(result, expected) @unpack @data( (None, None), ('image.jpg', 'image/jpeg'), ('/folder/file.test.pdf', 'application/pdf'), ('/folder/weird_file.wdl', 'binary/octet-stream'), ('a_file', 'binary/octet-stream') ) def test_content_type_from_file_name(self, input, expected): result = self.depositassets.content_type_from_file_name(input) self.assertEqual(result, expected) if __name__ == '__main__': unittest.main()
import unittest from activity.activity_DepositAssets import activity_DepositAssets import settings_mock from ddt import ddt, data, unpack @ddt class TestDepositAssets(unittest.TestCase): def setUp(self): self.depositassets = activity_DepositAssets(settings_mock, None, None, None, None) @unpack @data({'input': '.tif', 'expected': ['.tif']}, {'input': '.jpg, .tiff, .png', 'expected':['.jpg', '.tiff', '.png']}) def test_get_no_download_extensions(self, input, expected): result = self.depositassets.get_no_download_extensions(input) self.assertListEqual(result, expected) @unpack @data( (None, None), ('image.jpg', 'image/jpeg'), ('/folder/file.pdf.zip', 'application/x-zip-compressed'), ('/folder/weird_file.wdl', 'binary/octet-stream'), ('a_file', 'binary/octet-stream') ) def test_content_type_from_file_name(self, input, expected): result = self.depositassets.content_type_from_file_name(input) self.assertEqual(result, expected) if __name__ == '__main__': unittest.main() Change test data for mimetype to content_type.import unittest from activity.activity_DepositAssets import activity_DepositAssets import settings_mock from ddt import ddt, data, unpack @ddt class TestDepositAssets(unittest.TestCase): def setUp(self): self.depositassets = activity_DepositAssets(settings_mock, None, None, None, None) @unpack @data({'input': '.tif', 'expected': ['.tif']}, {'input': '.jpg, .tiff, .png', 'expected':['.jpg', '.tiff', '.png']}) def test_get_no_download_extensions(self, input, expected): result = self.depositassets.get_no_download_extensions(input) self.assertListEqual(result, expected) @unpack @data( (None, None), ('image.jpg', 'image/jpeg'), ('/folder/file.test.pdf', 'application/pdf'), ('/folder/weird_file.wdl', 'binary/octet-stream'), ('a_file', 'binary/octet-stream') ) def test_content_type_from_file_name(self, input, expected): result = self.depositassets.content_type_from_file_name(input) self.assertEqual(result, expected) if __name__ == '__main__': unittest.main()
<commit_before>import unittest from activity.activity_DepositAssets import activity_DepositAssets import settings_mock from ddt import ddt, data, unpack @ddt class TestDepositAssets(unittest.TestCase): def setUp(self): self.depositassets = activity_DepositAssets(settings_mock, None, None, None, None) @unpack @data({'input': '.tif', 'expected': ['.tif']}, {'input': '.jpg, .tiff, .png', 'expected':['.jpg', '.tiff', '.png']}) def test_get_no_download_extensions(self, input, expected): result = self.depositassets.get_no_download_extensions(input) self.assertListEqual(result, expected) @unpack @data( (None, None), ('image.jpg', 'image/jpeg'), ('/folder/file.pdf.zip', 'application/x-zip-compressed'), ('/folder/weird_file.wdl', 'binary/octet-stream'), ('a_file', 'binary/octet-stream') ) def test_content_type_from_file_name(self, input, expected): result = self.depositassets.content_type_from_file_name(input) self.assertEqual(result, expected) if __name__ == '__main__': unittest.main() <commit_msg>Change test data for mimetype to content_type.<commit_after>import unittest from activity.activity_DepositAssets import activity_DepositAssets import settings_mock from ddt import ddt, data, unpack @ddt class TestDepositAssets(unittest.TestCase): def setUp(self): self.depositassets = activity_DepositAssets(settings_mock, None, None, None, None) @unpack @data({'input': '.tif', 'expected': ['.tif']}, {'input': '.jpg, .tiff, .png', 'expected':['.jpg', '.tiff', '.png']}) def test_get_no_download_extensions(self, input, expected): result = self.depositassets.get_no_download_extensions(input) self.assertListEqual(result, expected) @unpack @data( (None, None), ('image.jpg', 'image/jpeg'), ('/folder/file.test.pdf', 'application/pdf'), ('/folder/weird_file.wdl', 'binary/octet-stream'), ('a_file', 'binary/octet-stream') ) def test_content_type_from_file_name(self, input, expected): result = self.depositassets.content_type_from_file_name(input) self.assertEqual(result, expected) if __name__ == '__main__': unittest.main()
a09274fbc9277de2cbd3336fca4922094b0db8d1
crmapp/urls.py
crmapp/urls.py
from django.conf.urls import patterns, include, url from django.contrib import admin urlpatterns = patterns('', # Examples: # url(r'^$', 'crmapp.views.home', name='home'), # url(r'^blog/', include('blog.urls')), url(r'^admin/', include(admin.site.urls)), )
from django.conf.urls import patterns, include, url from marketing.views import HomePage urlpatterns = patterns('', # Marketing pages url(r'^$', HomePage.as_view(), name="home"), # Subscriber related URLs # Admin URL # Login/Logout URLs # Account related URLs # Contact related URLS # Communication related URLs )
Create the Home Page > Create the URL Conf
Create the Home Page > Create the URL Conf
Python
mit
tabdon/crmeasyapp,tabdon/crmeasyapp,deenaariff/Django
from django.conf.urls import patterns, include, url from django.contrib import admin urlpatterns = patterns('', # Examples: # url(r'^$', 'crmapp.views.home', name='home'), # url(r'^blog/', include('blog.urls')), url(r'^admin/', include(admin.site.urls)), ) Create the Home Page > Create the URL Conf
from django.conf.urls import patterns, include, url from marketing.views import HomePage urlpatterns = patterns('', # Marketing pages url(r'^$', HomePage.as_view(), name="home"), # Subscriber related URLs # Admin URL # Login/Logout URLs # Account related URLs # Contact related URLS # Communication related URLs )
<commit_before>from django.conf.urls import patterns, include, url from django.contrib import admin urlpatterns = patterns('', # Examples: # url(r'^$', 'crmapp.views.home', name='home'), # url(r'^blog/', include('blog.urls')), url(r'^admin/', include(admin.site.urls)), ) <commit_msg>Create the Home Page > Create the URL Conf<commit_after>
from django.conf.urls import patterns, include, url from marketing.views import HomePage urlpatterns = patterns('', # Marketing pages url(r'^$', HomePage.as_view(), name="home"), # Subscriber related URLs # Admin URL # Login/Logout URLs # Account related URLs # Contact related URLS # Communication related URLs )
from django.conf.urls import patterns, include, url from django.contrib import admin urlpatterns = patterns('', # Examples: # url(r'^$', 'crmapp.views.home', name='home'), # url(r'^blog/', include('blog.urls')), url(r'^admin/', include(admin.site.urls)), ) Create the Home Page > Create the URL Conffrom django.conf.urls import patterns, include, url from marketing.views import HomePage urlpatterns = patterns('', # Marketing pages url(r'^$', HomePage.as_view(), name="home"), # Subscriber related URLs # Admin URL # Login/Logout URLs # Account related URLs # Contact related URLS # Communication related URLs )
<commit_before>from django.conf.urls import patterns, include, url from django.contrib import admin urlpatterns = patterns('', # Examples: # url(r'^$', 'crmapp.views.home', name='home'), # url(r'^blog/', include('blog.urls')), url(r'^admin/', include(admin.site.urls)), ) <commit_msg>Create the Home Page > Create the URL Conf<commit_after>from django.conf.urls import patterns, include, url from marketing.views import HomePage urlpatterns = patterns('', # Marketing pages url(r'^$', HomePage.as_view(), name="home"), # Subscriber related URLs # Admin URL # Login/Logout URLs # Account related URLs # Contact related URLS # Communication related URLs )
f155c200dc6f8b7b9461f399a3d3642d23e64942
snipsskills/templates/intent_template.py
snipsskills/templates/intent_template.py
# -*-: coding utf-8 -*- """ Auto-generated intent class. """ # ***************************************************************************** # ***************************************************************************** # ***************************************************************************** # # WARNING: THIS IS AN AUTO-GENERATED FILE # DO NOT ATTEMPT TO EDIT IT, AS CHANGES WILL BE OVERWRITTEN. # # ***************************************************************************** # ***************************************************************************** # ***************************************************************************** # pylint: disable=line-too-long from snipsskillscore.intent_parser import IntentParser class {{to_camelcase_capitalized(intent.name)}}Intent: intentName = "{{ intent.name }}" {% if intent.slots is defined and intent.slots|length > 0 -%} def __init__(self{% for slot in intent.slots -%}, {{slot.name}}=None{% endfor %}): {% for slot in intent.slots -%} self.{{slot.name}} = {{slot.name}} {% endfor %} {% endif -%} @staticmethod def parse(payload): intentName = IntentParser.get_intent_name(payload) if intentName != {{to_camelcase_capitalized(intent.name)}}Intent.intentName: return None return {{to_camelcase_capitalized(intent.name)}}Intent( {% for slot in intent.slots -%} IntentParser.get_slot_value(payload, "{{ slot.name}}"){{"," if not loop.last}} {% endfor -%} )
# -*-: coding utf-8 -*- """ Auto-generated intent class. """ # ***************************************************************************** # ***************************************************************************** # ***************************************************************************** # # WARNING: THIS IS AN AUTO-GENERATED FILE # DO NOT ATTEMPT TO EDIT IT, AS CHANGES WILL BE OVERWRITTEN. # # ***************************************************************************** # ***************************************************************************** # ***************************************************************************** # pylint: disable=line-too-long from snipsskillscore.intent_parser import IntentParser class {{to_camelcase_capitalized(intent.name)}}Intent: intentName = "{{ intent.name }}" {% if intent.slots is defined and intent.slots|length > 0 -%} def __init__(self, sessionId, siteId, customData{% for slot in intent.slots -%}, {{slot.name}}=None{% endfor %}): self.sessionId = sessionId self.siteId = siteId self.customData = customData {% for slot in intent.slots -%} self.{{slot.name}} = {{slot.name}} {% endfor %} {% endif -%} @staticmethod def parse(payload): intentName = IntentParser.get_intent_name(payload) if intentName != {{to_camelcase_capitalized(intent.name)}}Intent.intentName: return None return {{to_camelcase_capitalized(intent.name)}}Intent( IntentParser.get_session_id(payload), IntentParser.get_site_id(payload), IntentParser.get_custom_data(payload), {% for slot in intent.slots -%} IntentParser.get_slot_value(payload, "{{ slot.name}}"){{"," if not loop.last}} {% endfor -%} )
Add intent sessionId siteId & customData
Add intent sessionId siteId & customData
Python
mit
snipsco/snipsskills,snipsco/snipsskills,snipsco/snipsskills,snipsco/snipsskills
# -*-: coding utf-8 -*- """ Auto-generated intent class. """ # ***************************************************************************** # ***************************************************************************** # ***************************************************************************** # # WARNING: THIS IS AN AUTO-GENERATED FILE # DO NOT ATTEMPT TO EDIT IT, AS CHANGES WILL BE OVERWRITTEN. # # ***************************************************************************** # ***************************************************************************** # ***************************************************************************** # pylint: disable=line-too-long from snipsskillscore.intent_parser import IntentParser class {{to_camelcase_capitalized(intent.name)}}Intent: intentName = "{{ intent.name }}" {% if intent.slots is defined and intent.slots|length > 0 -%} def __init__(self{% for slot in intent.slots -%}, {{slot.name}}=None{% endfor %}): {% for slot in intent.slots -%} self.{{slot.name}} = {{slot.name}} {% endfor %} {% endif -%} @staticmethod def parse(payload): intentName = IntentParser.get_intent_name(payload) if intentName != {{to_camelcase_capitalized(intent.name)}}Intent.intentName: return None return {{to_camelcase_capitalized(intent.name)}}Intent( {% for slot in intent.slots -%} IntentParser.get_slot_value(payload, "{{ slot.name}}"){{"," if not loop.last}} {% endfor -%} ) Add intent sessionId siteId & customData
# -*-: coding utf-8 -*- """ Auto-generated intent class. """ # ***************************************************************************** # ***************************************************************************** # ***************************************************************************** # # WARNING: THIS IS AN AUTO-GENERATED FILE # DO NOT ATTEMPT TO EDIT IT, AS CHANGES WILL BE OVERWRITTEN. # # ***************************************************************************** # ***************************************************************************** # ***************************************************************************** # pylint: disable=line-too-long from snipsskillscore.intent_parser import IntentParser class {{to_camelcase_capitalized(intent.name)}}Intent: intentName = "{{ intent.name }}" {% if intent.slots is defined and intent.slots|length > 0 -%} def __init__(self, sessionId, siteId, customData{% for slot in intent.slots -%}, {{slot.name}}=None{% endfor %}): self.sessionId = sessionId self.siteId = siteId self.customData = customData {% for slot in intent.slots -%} self.{{slot.name}} = {{slot.name}} {% endfor %} {% endif -%} @staticmethod def parse(payload): intentName = IntentParser.get_intent_name(payload) if intentName != {{to_camelcase_capitalized(intent.name)}}Intent.intentName: return None return {{to_camelcase_capitalized(intent.name)}}Intent( IntentParser.get_session_id(payload), IntentParser.get_site_id(payload), IntentParser.get_custom_data(payload), {% for slot in intent.slots -%} IntentParser.get_slot_value(payload, "{{ slot.name}}"){{"," if not loop.last}} {% endfor -%} )
<commit_before># -*-: coding utf-8 -*- """ Auto-generated intent class. """ # ***************************************************************************** # ***************************************************************************** # ***************************************************************************** # # WARNING: THIS IS AN AUTO-GENERATED FILE # DO NOT ATTEMPT TO EDIT IT, AS CHANGES WILL BE OVERWRITTEN. # # ***************************************************************************** # ***************************************************************************** # ***************************************************************************** # pylint: disable=line-too-long from snipsskillscore.intent_parser import IntentParser class {{to_camelcase_capitalized(intent.name)}}Intent: intentName = "{{ intent.name }}" {% if intent.slots is defined and intent.slots|length > 0 -%} def __init__(self{% for slot in intent.slots -%}, {{slot.name}}=None{% endfor %}): {% for slot in intent.slots -%} self.{{slot.name}} = {{slot.name}} {% endfor %} {% endif -%} @staticmethod def parse(payload): intentName = IntentParser.get_intent_name(payload) if intentName != {{to_camelcase_capitalized(intent.name)}}Intent.intentName: return None return {{to_camelcase_capitalized(intent.name)}}Intent( {% for slot in intent.slots -%} IntentParser.get_slot_value(payload, "{{ slot.name}}"){{"," if not loop.last}} {% endfor -%} ) <commit_msg>Add intent sessionId siteId & customData<commit_after>
# -*-: coding utf-8 -*- """ Auto-generated intent class. """ # ***************************************************************************** # ***************************************************************************** # ***************************************************************************** # # WARNING: THIS IS AN AUTO-GENERATED FILE # DO NOT ATTEMPT TO EDIT IT, AS CHANGES WILL BE OVERWRITTEN. # # ***************************************************************************** # ***************************************************************************** # ***************************************************************************** # pylint: disable=line-too-long from snipsskillscore.intent_parser import IntentParser class {{to_camelcase_capitalized(intent.name)}}Intent: intentName = "{{ intent.name }}" {% if intent.slots is defined and intent.slots|length > 0 -%} def __init__(self, sessionId, siteId, customData{% for slot in intent.slots -%}, {{slot.name}}=None{% endfor %}): self.sessionId = sessionId self.siteId = siteId self.customData = customData {% for slot in intent.slots -%} self.{{slot.name}} = {{slot.name}} {% endfor %} {% endif -%} @staticmethod def parse(payload): intentName = IntentParser.get_intent_name(payload) if intentName != {{to_camelcase_capitalized(intent.name)}}Intent.intentName: return None return {{to_camelcase_capitalized(intent.name)}}Intent( IntentParser.get_session_id(payload), IntentParser.get_site_id(payload), IntentParser.get_custom_data(payload), {% for slot in intent.slots -%} IntentParser.get_slot_value(payload, "{{ slot.name}}"){{"," if not loop.last}} {% endfor -%} )
# -*-: coding utf-8 -*- """ Auto-generated intent class. """ # ***************************************************************************** # ***************************************************************************** # ***************************************************************************** # # WARNING: THIS IS AN AUTO-GENERATED FILE # DO NOT ATTEMPT TO EDIT IT, AS CHANGES WILL BE OVERWRITTEN. # # ***************************************************************************** # ***************************************************************************** # ***************************************************************************** # pylint: disable=line-too-long from snipsskillscore.intent_parser import IntentParser class {{to_camelcase_capitalized(intent.name)}}Intent: intentName = "{{ intent.name }}" {% if intent.slots is defined and intent.slots|length > 0 -%} def __init__(self{% for slot in intent.slots -%}, {{slot.name}}=None{% endfor %}): {% for slot in intent.slots -%} self.{{slot.name}} = {{slot.name}} {% endfor %} {% endif -%} @staticmethod def parse(payload): intentName = IntentParser.get_intent_name(payload) if intentName != {{to_camelcase_capitalized(intent.name)}}Intent.intentName: return None return {{to_camelcase_capitalized(intent.name)}}Intent( {% for slot in intent.slots -%} IntentParser.get_slot_value(payload, "{{ slot.name}}"){{"," if not loop.last}} {% endfor -%} ) Add intent sessionId siteId & customData# -*-: coding utf-8 -*- """ Auto-generated intent class. """ # ***************************************************************************** # ***************************************************************************** # ***************************************************************************** # # WARNING: THIS IS AN AUTO-GENERATED FILE # DO NOT ATTEMPT TO EDIT IT, AS CHANGES WILL BE OVERWRITTEN. # # ***************************************************************************** # ***************************************************************************** # ***************************************************************************** # pylint: disable=line-too-long from snipsskillscore.intent_parser import IntentParser class {{to_camelcase_capitalized(intent.name)}}Intent: intentName = "{{ intent.name }}" {% if intent.slots is defined and intent.slots|length > 0 -%} def __init__(self, sessionId, siteId, customData{% for slot in intent.slots -%}, {{slot.name}}=None{% endfor %}): self.sessionId = sessionId self.siteId = siteId self.customData = customData {% for slot in intent.slots -%} self.{{slot.name}} = {{slot.name}} {% endfor %} {% endif -%} @staticmethod def parse(payload): intentName = IntentParser.get_intent_name(payload) if intentName != {{to_camelcase_capitalized(intent.name)}}Intent.intentName: return None return {{to_camelcase_capitalized(intent.name)}}Intent( IntentParser.get_session_id(payload), IntentParser.get_site_id(payload), IntentParser.get_custom_data(payload), {% for slot in intent.slots -%} IntentParser.get_slot_value(payload, "{{ slot.name}}"){{"," if not loop.last}} {% endfor -%} )
<commit_before># -*-: coding utf-8 -*- """ Auto-generated intent class. """ # ***************************************************************************** # ***************************************************************************** # ***************************************************************************** # # WARNING: THIS IS AN AUTO-GENERATED FILE # DO NOT ATTEMPT TO EDIT IT, AS CHANGES WILL BE OVERWRITTEN. # # ***************************************************************************** # ***************************************************************************** # ***************************************************************************** # pylint: disable=line-too-long from snipsskillscore.intent_parser import IntentParser class {{to_camelcase_capitalized(intent.name)}}Intent: intentName = "{{ intent.name }}" {% if intent.slots is defined and intent.slots|length > 0 -%} def __init__(self{% for slot in intent.slots -%}, {{slot.name}}=None{% endfor %}): {% for slot in intent.slots -%} self.{{slot.name}} = {{slot.name}} {% endfor %} {% endif -%} @staticmethod def parse(payload): intentName = IntentParser.get_intent_name(payload) if intentName != {{to_camelcase_capitalized(intent.name)}}Intent.intentName: return None return {{to_camelcase_capitalized(intent.name)}}Intent( {% for slot in intent.slots -%} IntentParser.get_slot_value(payload, "{{ slot.name}}"){{"," if not loop.last}} {% endfor -%} ) <commit_msg>Add intent sessionId siteId & customData<commit_after># -*-: coding utf-8 -*- """ Auto-generated intent class. """ # ***************************************************************************** # ***************************************************************************** # ***************************************************************************** # # WARNING: THIS IS AN AUTO-GENERATED FILE # DO NOT ATTEMPT TO EDIT IT, AS CHANGES WILL BE OVERWRITTEN. # # ***************************************************************************** # ***************************************************************************** # ***************************************************************************** # pylint: disable=line-too-long from snipsskillscore.intent_parser import IntentParser class {{to_camelcase_capitalized(intent.name)}}Intent: intentName = "{{ intent.name }}" {% if intent.slots is defined and intent.slots|length > 0 -%} def __init__(self, sessionId, siteId, customData{% for slot in intent.slots -%}, {{slot.name}}=None{% endfor %}): self.sessionId = sessionId self.siteId = siteId self.customData = customData {% for slot in intent.slots -%} self.{{slot.name}} = {{slot.name}} {% endfor %} {% endif -%} @staticmethod def parse(payload): intentName = IntentParser.get_intent_name(payload) if intentName != {{to_camelcase_capitalized(intent.name)}}Intent.intentName: return None return {{to_camelcase_capitalized(intent.name)}}Intent( IntentParser.get_session_id(payload), IntentParser.get_site_id(payload), IntentParser.get_custom_data(payload), {% for slot in intent.slots -%} IntentParser.get_slot_value(payload, "{{ slot.name}}"){{"," if not loop.last}} {% endfor -%} )
a51d280ca7c7f487e6743e4f377f70641a8b4edd
turbustat/statistics/statistics_list.py
turbustat/statistics/statistics_list.py
# Licensed under an MIT open source license - see LICENSE ''' Returns a list of all available distance metrics ''' statistics_list = ["Wavelet", "MVC", "PSpec", "Bispectrum", "DeltaVariance", "Genus", "VCS", "VCA", "Tsallis", "PCA", "SCF", "Cramer", "Skewness", "Kurtosis", "VCS_Small_Scale", "VCS_Break", "VCS_Large_Scale", "PDF_Hellinger", "PDF_KS", # "PDF_AD", "Dendrogram_Hist", "Dendrogram_Num"] twoD_statistics_list = \ ["Wavelet", "MVC", "PSpec", "Bispectrum", "DeltaVariance", "Genus", "Tsallis", "Skewness", "Kurtosis", "PDF_Hellinger", "PDF_KS", # "PDF_AD", "Dendrogram_Hist", "Dendrogram_Num"]
# Licensed under an MIT open source license - see LICENSE ''' Returns a list of all available distance metrics ''' statistics_list = ["Wavelet", "MVC", "PSpec", "Bispectrum", "DeltaVariance", "Genus", "VCS", "VCA", "Tsallis", "PCA", "SCF", "Cramer", "Skewness", "Kurtosis", "VCS_Small_Scale", "VCS_Break", "VCS_Large_Scale", "PDF_Hellinger", "PDF_KS", "PDF_Lognormal", # "PDF_AD", "Dendrogram_Hist", "Dendrogram_Num"] twoD_statistics_list = \ ["Wavelet", "MVC", "PSpec", "Bispectrum", "DeltaVariance", "Genus", "Tsallis", "Skewness", "Kurtosis", "PDF_Hellinger", "PDF_KS", "PDF_Lognormal", # "PDF_AD", "Dendrogram_Hist", "Dendrogram_Num"] threeD_statistics_list = \ ["VCS", "VCA", "PCA", "SCF", "Cramer", "VCS_Small_Scale", "VCS_Large_Scale", "VCS_Break", "PDF_Hellinger", "PDF_KS", "PDF_Lognormal", "Dendrogram_Hist", "Dendrogram_Num"]
Update the stats lists and add a 3D only version
Update the stats lists and add a 3D only version
Python
mit
Astroua/TurbuStat,e-koch/TurbuStat
# Licensed under an MIT open source license - see LICENSE ''' Returns a list of all available distance metrics ''' statistics_list = ["Wavelet", "MVC", "PSpec", "Bispectrum", "DeltaVariance", "Genus", "VCS", "VCA", "Tsallis", "PCA", "SCF", "Cramer", "Skewness", "Kurtosis", "VCS_Small_Scale", "VCS_Break", "VCS_Large_Scale", "PDF_Hellinger", "PDF_KS", # "PDF_AD", "Dendrogram_Hist", "Dendrogram_Num"] twoD_statistics_list = \ ["Wavelet", "MVC", "PSpec", "Bispectrum", "DeltaVariance", "Genus", "Tsallis", "Skewness", "Kurtosis", "PDF_Hellinger", "PDF_KS", # "PDF_AD", "Dendrogram_Hist", "Dendrogram_Num"] Update the stats lists and add a 3D only version
# Licensed under an MIT open source license - see LICENSE ''' Returns a list of all available distance metrics ''' statistics_list = ["Wavelet", "MVC", "PSpec", "Bispectrum", "DeltaVariance", "Genus", "VCS", "VCA", "Tsallis", "PCA", "SCF", "Cramer", "Skewness", "Kurtosis", "VCS_Small_Scale", "VCS_Break", "VCS_Large_Scale", "PDF_Hellinger", "PDF_KS", "PDF_Lognormal", # "PDF_AD", "Dendrogram_Hist", "Dendrogram_Num"] twoD_statistics_list = \ ["Wavelet", "MVC", "PSpec", "Bispectrum", "DeltaVariance", "Genus", "Tsallis", "Skewness", "Kurtosis", "PDF_Hellinger", "PDF_KS", "PDF_Lognormal", # "PDF_AD", "Dendrogram_Hist", "Dendrogram_Num"] threeD_statistics_list = \ ["VCS", "VCA", "PCA", "SCF", "Cramer", "VCS_Small_Scale", "VCS_Large_Scale", "VCS_Break", "PDF_Hellinger", "PDF_KS", "PDF_Lognormal", "Dendrogram_Hist", "Dendrogram_Num"]
<commit_before># Licensed under an MIT open source license - see LICENSE ''' Returns a list of all available distance metrics ''' statistics_list = ["Wavelet", "MVC", "PSpec", "Bispectrum", "DeltaVariance", "Genus", "VCS", "VCA", "Tsallis", "PCA", "SCF", "Cramer", "Skewness", "Kurtosis", "VCS_Small_Scale", "VCS_Break", "VCS_Large_Scale", "PDF_Hellinger", "PDF_KS", # "PDF_AD", "Dendrogram_Hist", "Dendrogram_Num"] twoD_statistics_list = \ ["Wavelet", "MVC", "PSpec", "Bispectrum", "DeltaVariance", "Genus", "Tsallis", "Skewness", "Kurtosis", "PDF_Hellinger", "PDF_KS", # "PDF_AD", "Dendrogram_Hist", "Dendrogram_Num"] <commit_msg>Update the stats lists and add a 3D only version<commit_after>
# Licensed under an MIT open source license - see LICENSE ''' Returns a list of all available distance metrics ''' statistics_list = ["Wavelet", "MVC", "PSpec", "Bispectrum", "DeltaVariance", "Genus", "VCS", "VCA", "Tsallis", "PCA", "SCF", "Cramer", "Skewness", "Kurtosis", "VCS_Small_Scale", "VCS_Break", "VCS_Large_Scale", "PDF_Hellinger", "PDF_KS", "PDF_Lognormal", # "PDF_AD", "Dendrogram_Hist", "Dendrogram_Num"] twoD_statistics_list = \ ["Wavelet", "MVC", "PSpec", "Bispectrum", "DeltaVariance", "Genus", "Tsallis", "Skewness", "Kurtosis", "PDF_Hellinger", "PDF_KS", "PDF_Lognormal", # "PDF_AD", "Dendrogram_Hist", "Dendrogram_Num"] threeD_statistics_list = \ ["VCS", "VCA", "PCA", "SCF", "Cramer", "VCS_Small_Scale", "VCS_Large_Scale", "VCS_Break", "PDF_Hellinger", "PDF_KS", "PDF_Lognormal", "Dendrogram_Hist", "Dendrogram_Num"]
# Licensed under an MIT open source license - see LICENSE ''' Returns a list of all available distance metrics ''' statistics_list = ["Wavelet", "MVC", "PSpec", "Bispectrum", "DeltaVariance", "Genus", "VCS", "VCA", "Tsallis", "PCA", "SCF", "Cramer", "Skewness", "Kurtosis", "VCS_Small_Scale", "VCS_Break", "VCS_Large_Scale", "PDF_Hellinger", "PDF_KS", # "PDF_AD", "Dendrogram_Hist", "Dendrogram_Num"] twoD_statistics_list = \ ["Wavelet", "MVC", "PSpec", "Bispectrum", "DeltaVariance", "Genus", "Tsallis", "Skewness", "Kurtosis", "PDF_Hellinger", "PDF_KS", # "PDF_AD", "Dendrogram_Hist", "Dendrogram_Num"] Update the stats lists and add a 3D only version# Licensed under an MIT open source license - see LICENSE ''' Returns a list of all available distance metrics ''' statistics_list = ["Wavelet", "MVC", "PSpec", "Bispectrum", "DeltaVariance", "Genus", "VCS", "VCA", "Tsallis", "PCA", "SCF", "Cramer", "Skewness", "Kurtosis", "VCS_Small_Scale", "VCS_Break", "VCS_Large_Scale", "PDF_Hellinger", "PDF_KS", "PDF_Lognormal", # "PDF_AD", "Dendrogram_Hist", "Dendrogram_Num"] twoD_statistics_list = \ ["Wavelet", "MVC", "PSpec", "Bispectrum", "DeltaVariance", "Genus", "Tsallis", "Skewness", "Kurtosis", "PDF_Hellinger", "PDF_KS", "PDF_Lognormal", # "PDF_AD", "Dendrogram_Hist", "Dendrogram_Num"] threeD_statistics_list = \ ["VCS", "VCA", "PCA", "SCF", "Cramer", "VCS_Small_Scale", "VCS_Large_Scale", "VCS_Break", "PDF_Hellinger", "PDF_KS", "PDF_Lognormal", "Dendrogram_Hist", "Dendrogram_Num"]
<commit_before># Licensed under an MIT open source license - see LICENSE ''' Returns a list of all available distance metrics ''' statistics_list = ["Wavelet", "MVC", "PSpec", "Bispectrum", "DeltaVariance", "Genus", "VCS", "VCA", "Tsallis", "PCA", "SCF", "Cramer", "Skewness", "Kurtosis", "VCS_Small_Scale", "VCS_Break", "VCS_Large_Scale", "PDF_Hellinger", "PDF_KS", # "PDF_AD", "Dendrogram_Hist", "Dendrogram_Num"] twoD_statistics_list = \ ["Wavelet", "MVC", "PSpec", "Bispectrum", "DeltaVariance", "Genus", "Tsallis", "Skewness", "Kurtosis", "PDF_Hellinger", "PDF_KS", # "PDF_AD", "Dendrogram_Hist", "Dendrogram_Num"] <commit_msg>Update the stats lists and add a 3D only version<commit_after># Licensed under an MIT open source license - see LICENSE ''' Returns a list of all available distance metrics ''' statistics_list = ["Wavelet", "MVC", "PSpec", "Bispectrum", "DeltaVariance", "Genus", "VCS", "VCA", "Tsallis", "PCA", "SCF", "Cramer", "Skewness", "Kurtosis", "VCS_Small_Scale", "VCS_Break", "VCS_Large_Scale", "PDF_Hellinger", "PDF_KS", "PDF_Lognormal", # "PDF_AD", "Dendrogram_Hist", "Dendrogram_Num"] twoD_statistics_list = \ ["Wavelet", "MVC", "PSpec", "Bispectrum", "DeltaVariance", "Genus", "Tsallis", "Skewness", "Kurtosis", "PDF_Hellinger", "PDF_KS", "PDF_Lognormal", # "PDF_AD", "Dendrogram_Hist", "Dendrogram_Num"] threeD_statistics_list = \ ["VCS", "VCA", "PCA", "SCF", "Cramer", "VCS_Small_Scale", "VCS_Large_Scale", "VCS_Break", "PDF_Hellinger", "PDF_KS", "PDF_Lognormal", "Dendrogram_Hist", "Dendrogram_Num"]
50d69badbaeb41736269c99f7b301f09c9b24ecb
testing/test_models.py
testing/test_models.py
from k2catalogue import models def test_safe_float_good(): assert models.safe_float("2") == 2.0 def test_safe_float_bad(): assert models.safe_float('this is not convertable to a float') is None
from k2catalogue import models def test_safe_float_good(): assert models.safe_float("2") == 2.0 def test_safe_float_bad(): assert models.safe_float('this is not convertable to a float') is None def test_proposal_printing(): proposal = models.Proposal(proposal_id='abc') assert repr(proposal) == '<Proposal: abc>'
Add test for proposal printing
Add test for proposal printing
Python
mit
mindriot101/k2catalogue
from k2catalogue import models def test_safe_float_good(): assert models.safe_float("2") == 2.0 def test_safe_float_bad(): assert models.safe_float('this is not convertable to a float') is None Add test for proposal printing
from k2catalogue import models def test_safe_float_good(): assert models.safe_float("2") == 2.0 def test_safe_float_bad(): assert models.safe_float('this is not convertable to a float') is None def test_proposal_printing(): proposal = models.Proposal(proposal_id='abc') assert repr(proposal) == '<Proposal: abc>'
<commit_before>from k2catalogue import models def test_safe_float_good(): assert models.safe_float("2") == 2.0 def test_safe_float_bad(): assert models.safe_float('this is not convertable to a float') is None <commit_msg>Add test for proposal printing<commit_after>
from k2catalogue import models def test_safe_float_good(): assert models.safe_float("2") == 2.0 def test_safe_float_bad(): assert models.safe_float('this is not convertable to a float') is None def test_proposal_printing(): proposal = models.Proposal(proposal_id='abc') assert repr(proposal) == '<Proposal: abc>'
from k2catalogue import models def test_safe_float_good(): assert models.safe_float("2") == 2.0 def test_safe_float_bad(): assert models.safe_float('this is not convertable to a float') is None Add test for proposal printingfrom k2catalogue import models def test_safe_float_good(): assert models.safe_float("2") == 2.0 def test_safe_float_bad(): assert models.safe_float('this is not convertable to a float') is None def test_proposal_printing(): proposal = models.Proposal(proposal_id='abc') assert repr(proposal) == '<Proposal: abc>'
<commit_before>from k2catalogue import models def test_safe_float_good(): assert models.safe_float("2") == 2.0 def test_safe_float_bad(): assert models.safe_float('this is not convertable to a float') is None <commit_msg>Add test for proposal printing<commit_after>from k2catalogue import models def test_safe_float_good(): assert models.safe_float("2") == 2.0 def test_safe_float_bad(): assert models.safe_float('this is not convertable to a float') is None def test_proposal_printing(): proposal = models.Proposal(proposal_id='abc') assert repr(proposal) == '<Proposal: abc>'
855cea2b603453ce7ed907fc153962596de31f00
src/json.py
src/json.py
msg = 'This is not yet implemented' def validate(sketch): """ - sketch: a `dict` """ raise NotImplementedError(msg) def readsketch_iter(iterable): """ - iterable: as return by ijson.parser Returns a `dict` with a sketch information """ raise NotImplementedError(msg) def readjamschema(schema): """ Read a JAM definition schema """ raise NotImplementedError(msg)
MSG = 'This is not yet implemented' def validate(sketch): """ - sketch: a `dict` """ raise NotImplementedError(MSG) def readsketch_iter(iterable): """ - iterable: as return by ijson.parser Returns a `dict` with a sketch information """ raise NotImplementedError(MSG) def readjamschema(schema): """ Read a JAM definition schema """ raise NotImplementedError(MSG)
Rename msg -> MSG (since a constant).
Rename msg -> MSG (since a constant).
Python
mit
lgautier/mashing-pumpkins,lgautier/mashing-pumpkins,lgautier/mashing-pumpkins
msg = 'This is not yet implemented' def validate(sketch): """ - sketch: a `dict` """ raise NotImplementedError(msg) def readsketch_iter(iterable): """ - iterable: as return by ijson.parser Returns a `dict` with a sketch information """ raise NotImplementedError(msg) def readjamschema(schema): """ Read a JAM definition schema """ raise NotImplementedError(msg) Rename msg -> MSG (since a constant).
MSG = 'This is not yet implemented' def validate(sketch): """ - sketch: a `dict` """ raise NotImplementedError(MSG) def readsketch_iter(iterable): """ - iterable: as return by ijson.parser Returns a `dict` with a sketch information """ raise NotImplementedError(MSG) def readjamschema(schema): """ Read a JAM definition schema """ raise NotImplementedError(MSG)
<commit_before>msg = 'This is not yet implemented' def validate(sketch): """ - sketch: a `dict` """ raise NotImplementedError(msg) def readsketch_iter(iterable): """ - iterable: as return by ijson.parser Returns a `dict` with a sketch information """ raise NotImplementedError(msg) def readjamschema(schema): """ Read a JAM definition schema """ raise NotImplementedError(msg) <commit_msg>Rename msg -> MSG (since a constant).<commit_after>
MSG = 'This is not yet implemented' def validate(sketch): """ - sketch: a `dict` """ raise NotImplementedError(MSG) def readsketch_iter(iterable): """ - iterable: as return by ijson.parser Returns a `dict` with a sketch information """ raise NotImplementedError(MSG) def readjamschema(schema): """ Read a JAM definition schema """ raise NotImplementedError(MSG)
msg = 'This is not yet implemented' def validate(sketch): """ - sketch: a `dict` """ raise NotImplementedError(msg) def readsketch_iter(iterable): """ - iterable: as return by ijson.parser Returns a `dict` with a sketch information """ raise NotImplementedError(msg) def readjamschema(schema): """ Read a JAM definition schema """ raise NotImplementedError(msg) Rename msg -> MSG (since a constant).MSG = 'This is not yet implemented' def validate(sketch): """ - sketch: a `dict` """ raise NotImplementedError(MSG) def readsketch_iter(iterable): """ - iterable: as return by ijson.parser Returns a `dict` with a sketch information """ raise NotImplementedError(MSG) def readjamschema(schema): """ Read a JAM definition schema """ raise NotImplementedError(MSG)
<commit_before>msg = 'This is not yet implemented' def validate(sketch): """ - sketch: a `dict` """ raise NotImplementedError(msg) def readsketch_iter(iterable): """ - iterable: as return by ijson.parser Returns a `dict` with a sketch information """ raise NotImplementedError(msg) def readjamschema(schema): """ Read a JAM definition schema """ raise NotImplementedError(msg) <commit_msg>Rename msg -> MSG (since a constant).<commit_after>MSG = 'This is not yet implemented' def validate(sketch): """ - sketch: a `dict` """ raise NotImplementedError(MSG) def readsketch_iter(iterable): """ - iterable: as return by ijson.parser Returns a `dict` with a sketch information """ raise NotImplementedError(MSG) def readjamschema(schema): """ Read a JAM definition schema """ raise NotImplementedError(MSG)
7cacf4f78f03e95eec51aac211779539954aee38
pymanopt/manifolds/__init__.py
pymanopt/manifolds/__init__.py
from .grassmann import Grassmann from .sphere import Sphere from .stiefel import Stiefel from .psd import PSDFixedRank, PSDFixedRankComplex, Elliptope, PositiveDefinite from .oblique import Oblique from .euclidean import Euclidean, Symmetric from .product import Product __all__ = ["Grassmann", "Sphere", "Stiefel", "PSDFixedRank", "PSDFixedRankComplex", "Elliptope", "PositiveDefinite", "Oblique", "Euclidean", "Product", "Symmetric"]
from .grassmann import Grassmann from .sphere import Sphere from .stiefel import Stiefel from .psd import PSDFixedRank, PSDFixedRankComplex, Elliptope, PositiveDefinite from .oblique import Oblique from .euclidean import Euclidean, Symmetric from .product import Product from .fixed_rank import FixedRankEmbedded __all__ = ["Grassmann", "Sphere", "Stiefel", "PSDFixedRank", "PSDFixedRankComplex", "Elliptope", "PositiveDefinite", "Oblique", "Euclidean", "Product", "Symmetric", "FixedRankEmbedded"]
Add FixedRankEmbedded to manifolds init.
Add FixedRankEmbedded to manifolds init. Signed-off-by: Sebastian Weichwald <46f1a0bd5592a2f9244ca321b129902a06b53e03@sweichwald.de>
Python
bsd-3-clause
j-towns/pymanopt,nkoep/pymanopt,tingelst/pymanopt,nkoep/pymanopt,pymanopt/pymanopt,nkoep/pymanopt,pymanopt/pymanopt
from .grassmann import Grassmann from .sphere import Sphere from .stiefel import Stiefel from .psd import PSDFixedRank, PSDFixedRankComplex, Elliptope, PositiveDefinite from .oblique import Oblique from .euclidean import Euclidean, Symmetric from .product import Product __all__ = ["Grassmann", "Sphere", "Stiefel", "PSDFixedRank", "PSDFixedRankComplex", "Elliptope", "PositiveDefinite", "Oblique", "Euclidean", "Product", "Symmetric"] Add FixedRankEmbedded to manifolds init. Signed-off-by: Sebastian Weichwald <46f1a0bd5592a2f9244ca321b129902a06b53e03@sweichwald.de>
from .grassmann import Grassmann from .sphere import Sphere from .stiefel import Stiefel from .psd import PSDFixedRank, PSDFixedRankComplex, Elliptope, PositiveDefinite from .oblique import Oblique from .euclidean import Euclidean, Symmetric from .product import Product from .fixed_rank import FixedRankEmbedded __all__ = ["Grassmann", "Sphere", "Stiefel", "PSDFixedRank", "PSDFixedRankComplex", "Elliptope", "PositiveDefinite", "Oblique", "Euclidean", "Product", "Symmetric", "FixedRankEmbedded"]
<commit_before>from .grassmann import Grassmann from .sphere import Sphere from .stiefel import Stiefel from .psd import PSDFixedRank, PSDFixedRankComplex, Elliptope, PositiveDefinite from .oblique import Oblique from .euclidean import Euclidean, Symmetric from .product import Product __all__ = ["Grassmann", "Sphere", "Stiefel", "PSDFixedRank", "PSDFixedRankComplex", "Elliptope", "PositiveDefinite", "Oblique", "Euclidean", "Product", "Symmetric"] <commit_msg>Add FixedRankEmbedded to manifolds init. Signed-off-by: Sebastian Weichwald <46f1a0bd5592a2f9244ca321b129902a06b53e03@sweichwald.de><commit_after>
from .grassmann import Grassmann from .sphere import Sphere from .stiefel import Stiefel from .psd import PSDFixedRank, PSDFixedRankComplex, Elliptope, PositiveDefinite from .oblique import Oblique from .euclidean import Euclidean, Symmetric from .product import Product from .fixed_rank import FixedRankEmbedded __all__ = ["Grassmann", "Sphere", "Stiefel", "PSDFixedRank", "PSDFixedRankComplex", "Elliptope", "PositiveDefinite", "Oblique", "Euclidean", "Product", "Symmetric", "FixedRankEmbedded"]
from .grassmann import Grassmann from .sphere import Sphere from .stiefel import Stiefel from .psd import PSDFixedRank, PSDFixedRankComplex, Elliptope, PositiveDefinite from .oblique import Oblique from .euclidean import Euclidean, Symmetric from .product import Product __all__ = ["Grassmann", "Sphere", "Stiefel", "PSDFixedRank", "PSDFixedRankComplex", "Elliptope", "PositiveDefinite", "Oblique", "Euclidean", "Product", "Symmetric"] Add FixedRankEmbedded to manifolds init. Signed-off-by: Sebastian Weichwald <46f1a0bd5592a2f9244ca321b129902a06b53e03@sweichwald.de>from .grassmann import Grassmann from .sphere import Sphere from .stiefel import Stiefel from .psd import PSDFixedRank, PSDFixedRankComplex, Elliptope, PositiveDefinite from .oblique import Oblique from .euclidean import Euclidean, Symmetric from .product import Product from .fixed_rank import FixedRankEmbedded __all__ = ["Grassmann", "Sphere", "Stiefel", "PSDFixedRank", "PSDFixedRankComplex", "Elliptope", "PositiveDefinite", "Oblique", "Euclidean", "Product", "Symmetric", "FixedRankEmbedded"]
<commit_before>from .grassmann import Grassmann from .sphere import Sphere from .stiefel import Stiefel from .psd import PSDFixedRank, PSDFixedRankComplex, Elliptope, PositiveDefinite from .oblique import Oblique from .euclidean import Euclidean, Symmetric from .product import Product __all__ = ["Grassmann", "Sphere", "Stiefel", "PSDFixedRank", "PSDFixedRankComplex", "Elliptope", "PositiveDefinite", "Oblique", "Euclidean", "Product", "Symmetric"] <commit_msg>Add FixedRankEmbedded to manifolds init. Signed-off-by: Sebastian Weichwald <46f1a0bd5592a2f9244ca321b129902a06b53e03@sweichwald.de><commit_after>from .grassmann import Grassmann from .sphere import Sphere from .stiefel import Stiefel from .psd import PSDFixedRank, PSDFixedRankComplex, Elliptope, PositiveDefinite from .oblique import Oblique from .euclidean import Euclidean, Symmetric from .product import Product from .fixed_rank import FixedRankEmbedded __all__ = ["Grassmann", "Sphere", "Stiefel", "PSDFixedRank", "PSDFixedRankComplex", "Elliptope", "PositiveDefinite", "Oblique", "Euclidean", "Product", "Symmetric", "FixedRankEmbedded"]
c2b9bcb5c931f89dcbbf473cefe8b238e01e5429
vpr/tests/migrate/shell_migrate.py
vpr/tests/migrate/shell_migrate.py
from django.db import connection from vpr_content import models def removeDuplicatedTitleInMaterial(): cur = connection.cursor() qr0 = 'select id from vpr_content_material' qr1 = 'select text from vpr_content_material where id=%d' qr2 = 'update vpr_content_material set text=\'%s\' where id=%d' pt0 = '<div class="title">' pt1 = '</div>' cur.execute(qr0) mids = cur.fetchall() for mid in mids: try: mid = mid[0] cur.execute(qr1 % mid) text = cur.fetchone()[0] p0 = text.find(pt0) p1 = text.find(pt1, p0) text = text[:p0] + text[p1+len(pt1)+1:] material = models.Material.objects.get(pk=mid) material.text = text material.save() print mid except: raise print 'Updating failed at ' + str(mid)
from django.db import connection from vpr_content import models def removeDuplicatedTitleInMaterial(): cur = connection.cursor() qr0 = 'select id from vpr_content_material' qr1 = 'select text from vpr_content_material where id=%d' qr2 = 'update vpr_content_material set text=\'%s\' where id=%d' pt0 = '<div class="title">' pt1 = '</div>' cur.execute(qr0) mids = cur.fetchall() for mid in mids: try: mid = mid[0] cur.execute(qr1 % mid) text = cur.fetchone()[0] p0 = text.find(pt0) p1 = text.find(pt1, p0) text = text[:p0] + text[p1+len(pt1)+1:] material = models.Material.objects.get(pk=mid) material.text = text material.save() print mid except: raise print 'Updating failed at ' + str(mid) def determineLanguage(title, text): VI_ELEMENTS_0 = (u'\u1ea1', u'\u1ebf', u'\u01b0') VI_ELEMENTS_1 = ('&#226;', '&#432;', '&#224;') for item in VI_ELEMENTS_0: if item in title: return 'vi' for item in VI_ELEMENTS_1: if item in text: return 'vi' return 'en' def correctMaterialLanguage(material, dry=False): if type(material) == str: material = models.Material.objects.get(material_id=material) new_lang = determineLanguage(material.title, material.text) print str(material.language) + ' > ' + new_lang if not dry: material.language = new_lang material.save() def correctAllLanguages(dry=True): targets = models.Material.objects.exclude(language='vi') for material in targets: correctMaterialLanguage(material, dry)
Add script for correcting language
Add script for correcting language
Python
agpl-3.0
voer-platform/vp.repo,voer-platform/vp.repo,voer-platform/vp.repo,voer-platform/vp.repo
from django.db import connection from vpr_content import models def removeDuplicatedTitleInMaterial(): cur = connection.cursor() qr0 = 'select id from vpr_content_material' qr1 = 'select text from vpr_content_material where id=%d' qr2 = 'update vpr_content_material set text=\'%s\' where id=%d' pt0 = '<div class="title">' pt1 = '</div>' cur.execute(qr0) mids = cur.fetchall() for mid in mids: try: mid = mid[0] cur.execute(qr1 % mid) text = cur.fetchone()[0] p0 = text.find(pt0) p1 = text.find(pt1, p0) text = text[:p0] + text[p1+len(pt1)+1:] material = models.Material.objects.get(pk=mid) material.text = text material.save() print mid except: raise print 'Updating failed at ' + str(mid) Add script for correcting language
from django.db import connection from vpr_content import models def removeDuplicatedTitleInMaterial(): cur = connection.cursor() qr0 = 'select id from vpr_content_material' qr1 = 'select text from vpr_content_material where id=%d' qr2 = 'update vpr_content_material set text=\'%s\' where id=%d' pt0 = '<div class="title">' pt1 = '</div>' cur.execute(qr0) mids = cur.fetchall() for mid in mids: try: mid = mid[0] cur.execute(qr1 % mid) text = cur.fetchone()[0] p0 = text.find(pt0) p1 = text.find(pt1, p0) text = text[:p0] + text[p1+len(pt1)+1:] material = models.Material.objects.get(pk=mid) material.text = text material.save() print mid except: raise print 'Updating failed at ' + str(mid) def determineLanguage(title, text): VI_ELEMENTS_0 = (u'\u1ea1', u'\u1ebf', u'\u01b0') VI_ELEMENTS_1 = ('&#226;', '&#432;', '&#224;') for item in VI_ELEMENTS_0: if item in title: return 'vi' for item in VI_ELEMENTS_1: if item in text: return 'vi' return 'en' def correctMaterialLanguage(material, dry=False): if type(material) == str: material = models.Material.objects.get(material_id=material) new_lang = determineLanguage(material.title, material.text) print str(material.language) + ' > ' + new_lang if not dry: material.language = new_lang material.save() def correctAllLanguages(dry=True): targets = models.Material.objects.exclude(language='vi') for material in targets: correctMaterialLanguage(material, dry)
<commit_before>from django.db import connection from vpr_content import models def removeDuplicatedTitleInMaterial(): cur = connection.cursor() qr0 = 'select id from vpr_content_material' qr1 = 'select text from vpr_content_material where id=%d' qr2 = 'update vpr_content_material set text=\'%s\' where id=%d' pt0 = '<div class="title">' pt1 = '</div>' cur.execute(qr0) mids = cur.fetchall() for mid in mids: try: mid = mid[0] cur.execute(qr1 % mid) text = cur.fetchone()[0] p0 = text.find(pt0) p1 = text.find(pt1, p0) text = text[:p0] + text[p1+len(pt1)+1:] material = models.Material.objects.get(pk=mid) material.text = text material.save() print mid except: raise print 'Updating failed at ' + str(mid) <commit_msg>Add script for correcting language<commit_after>
from django.db import connection from vpr_content import models def removeDuplicatedTitleInMaterial(): cur = connection.cursor() qr0 = 'select id from vpr_content_material' qr1 = 'select text from vpr_content_material where id=%d' qr2 = 'update vpr_content_material set text=\'%s\' where id=%d' pt0 = '<div class="title">' pt1 = '</div>' cur.execute(qr0) mids = cur.fetchall() for mid in mids: try: mid = mid[0] cur.execute(qr1 % mid) text = cur.fetchone()[0] p0 = text.find(pt0) p1 = text.find(pt1, p0) text = text[:p0] + text[p1+len(pt1)+1:] material = models.Material.objects.get(pk=mid) material.text = text material.save() print mid except: raise print 'Updating failed at ' + str(mid) def determineLanguage(title, text): VI_ELEMENTS_0 = (u'\u1ea1', u'\u1ebf', u'\u01b0') VI_ELEMENTS_1 = ('&#226;', '&#432;', '&#224;') for item in VI_ELEMENTS_0: if item in title: return 'vi' for item in VI_ELEMENTS_1: if item in text: return 'vi' return 'en' def correctMaterialLanguage(material, dry=False): if type(material) == str: material = models.Material.objects.get(material_id=material) new_lang = determineLanguage(material.title, material.text) print str(material.language) + ' > ' + new_lang if not dry: material.language = new_lang material.save() def correctAllLanguages(dry=True): targets = models.Material.objects.exclude(language='vi') for material in targets: correctMaterialLanguage(material, dry)
from django.db import connection from vpr_content import models def removeDuplicatedTitleInMaterial(): cur = connection.cursor() qr0 = 'select id from vpr_content_material' qr1 = 'select text from vpr_content_material where id=%d' qr2 = 'update vpr_content_material set text=\'%s\' where id=%d' pt0 = '<div class="title">' pt1 = '</div>' cur.execute(qr0) mids = cur.fetchall() for mid in mids: try: mid = mid[0] cur.execute(qr1 % mid) text = cur.fetchone()[0] p0 = text.find(pt0) p1 = text.find(pt1, p0) text = text[:p0] + text[p1+len(pt1)+1:] material = models.Material.objects.get(pk=mid) material.text = text material.save() print mid except: raise print 'Updating failed at ' + str(mid) Add script for correcting languagefrom django.db import connection from vpr_content import models def removeDuplicatedTitleInMaterial(): cur = connection.cursor() qr0 = 'select id from vpr_content_material' qr1 = 'select text from vpr_content_material where id=%d' qr2 = 'update vpr_content_material set text=\'%s\' where id=%d' pt0 = '<div class="title">' pt1 = '</div>' cur.execute(qr0) mids = cur.fetchall() for mid in mids: try: mid = mid[0] cur.execute(qr1 % mid) text = cur.fetchone()[0] p0 = text.find(pt0) p1 = text.find(pt1, p0) text = text[:p0] + text[p1+len(pt1)+1:] material = models.Material.objects.get(pk=mid) material.text = text material.save() print mid except: raise print 'Updating failed at ' + str(mid) def determineLanguage(title, text): VI_ELEMENTS_0 = (u'\u1ea1', u'\u1ebf', u'\u01b0') VI_ELEMENTS_1 = ('&#226;', '&#432;', '&#224;') for item in VI_ELEMENTS_0: if item in title: return 'vi' for item in VI_ELEMENTS_1: if item in text: return 'vi' return 'en' def correctMaterialLanguage(material, dry=False): if type(material) == str: material = models.Material.objects.get(material_id=material) new_lang = determineLanguage(material.title, material.text) print str(material.language) + ' > ' + new_lang if not dry: material.language = new_lang material.save() def correctAllLanguages(dry=True): targets = models.Material.objects.exclude(language='vi') for material in targets: correctMaterialLanguage(material, dry)
<commit_before>from django.db import connection from vpr_content import models def removeDuplicatedTitleInMaterial(): cur = connection.cursor() qr0 = 'select id from vpr_content_material' qr1 = 'select text from vpr_content_material where id=%d' qr2 = 'update vpr_content_material set text=\'%s\' where id=%d' pt0 = '<div class="title">' pt1 = '</div>' cur.execute(qr0) mids = cur.fetchall() for mid in mids: try: mid = mid[0] cur.execute(qr1 % mid) text = cur.fetchone()[0] p0 = text.find(pt0) p1 = text.find(pt1, p0) text = text[:p0] + text[p1+len(pt1)+1:] material = models.Material.objects.get(pk=mid) material.text = text material.save() print mid except: raise print 'Updating failed at ' + str(mid) <commit_msg>Add script for correcting language<commit_after>from django.db import connection from vpr_content import models def removeDuplicatedTitleInMaterial(): cur = connection.cursor() qr0 = 'select id from vpr_content_material' qr1 = 'select text from vpr_content_material where id=%d' qr2 = 'update vpr_content_material set text=\'%s\' where id=%d' pt0 = '<div class="title">' pt1 = '</div>' cur.execute(qr0) mids = cur.fetchall() for mid in mids: try: mid = mid[0] cur.execute(qr1 % mid) text = cur.fetchone()[0] p0 = text.find(pt0) p1 = text.find(pt1, p0) text = text[:p0] + text[p1+len(pt1)+1:] material = models.Material.objects.get(pk=mid) material.text = text material.save() print mid except: raise print 'Updating failed at ' + str(mid) def determineLanguage(title, text): VI_ELEMENTS_0 = (u'\u1ea1', u'\u1ebf', u'\u01b0') VI_ELEMENTS_1 = ('&#226;', '&#432;', '&#224;') for item in VI_ELEMENTS_0: if item in title: return 'vi' for item in VI_ELEMENTS_1: if item in text: return 'vi' return 'en' def correctMaterialLanguage(material, dry=False): if type(material) == str: material = models.Material.objects.get(material_id=material) new_lang = determineLanguage(material.title, material.text) print str(material.language) + ' > ' + new_lang if not dry: material.language = new_lang material.save() def correctAllLanguages(dry=True): targets = models.Material.objects.exclude(language='vi') for material in targets: correctMaterialLanguage(material, dry)
95c5b9c139bf69ac11338a4b2eaa9b8179d27284
tests/test_async.py
tests/test_async.py
# -*- coding: utf-8 -*- from asyncio import Future, gather, new_event_loop, sleep from twisted.internet.defer import Deferred, ensureDeferred from pyee import EventEmitter def test_asyncio_emit(): """Test that event_emitters can handle wrapping coroutines as used with asyncio. """ loop = new_event_loop() ee = EventEmitter(loop=loop) should_call = Future(loop=loop) @ee.on('event') async def event_handler(): should_call.set_result(True) async def create_timeout(loop=loop): await sleep(0.1, loop=loop) if not should_call.done(): raise Exception('should_call timed out!') return should_call.cancel() timeout = create_timeout(loop=loop) @should_call.add_done_callback def _done(result): assert result ee.emit('event') loop.run_until_complete(gather(should_call, timeout, loop=loop)) loop.close() def test_twisted_emit(): """Test that event_emitters can handle wrapping coroutines when using twisted and ensureDeferred. """ ee = EventEmitter(scheduler=ensureDeferred) should_call = Deferred() @ee.on('event') async def event_handler(): should_call.callback(True) @should_call.addCallback def _done(result): assert result @should_call.addErrback def _err(exc): raise exc ee.emit('event')
# -*- coding: utf-8 -*- from asyncio import Future, gather, new_event_loop, sleep from mock import Mock from twisted.internet.defer import ensureDeferred from pyee import EventEmitter def test_asyncio_emit(): """Test that event_emitters can handle wrapping coroutines as used with asyncio. """ loop = new_event_loop() ee = EventEmitter(loop=loop) should_call = Future(loop=loop) @ee.on('event') async def event_handler(): should_call.set_result(True) async def create_timeout(loop=loop): await sleep(0.1, loop=loop) if not should_call.done(): raise Exception('should_call timed out!') return should_call.cancel() timeout = create_timeout(loop=loop) @should_call.add_done_callback def _done(result): assert result ee.emit('event') loop.run_until_complete(gather(should_call, timeout, loop=loop)) loop.close() def test_twisted_emit(): """Test that event_emitters can handle wrapping coroutines when using twisted and ensureDeferred. """ ee = EventEmitter(scheduler=ensureDeferred) should_call = Mock() @ee.on('event') async def event_handler(): should_call(True) ee.emit('event') should_call.assert_called_once()
Replace my deferred with a mock
Replace my deferred with a mock
Python
mit
jfhbrook/pyee
# -*- coding: utf-8 -*- from asyncio import Future, gather, new_event_loop, sleep from twisted.internet.defer import Deferred, ensureDeferred from pyee import EventEmitter def test_asyncio_emit(): """Test that event_emitters can handle wrapping coroutines as used with asyncio. """ loop = new_event_loop() ee = EventEmitter(loop=loop) should_call = Future(loop=loop) @ee.on('event') async def event_handler(): should_call.set_result(True) async def create_timeout(loop=loop): await sleep(0.1, loop=loop) if not should_call.done(): raise Exception('should_call timed out!') return should_call.cancel() timeout = create_timeout(loop=loop) @should_call.add_done_callback def _done(result): assert result ee.emit('event') loop.run_until_complete(gather(should_call, timeout, loop=loop)) loop.close() def test_twisted_emit(): """Test that event_emitters can handle wrapping coroutines when using twisted and ensureDeferred. """ ee = EventEmitter(scheduler=ensureDeferred) should_call = Deferred() @ee.on('event') async def event_handler(): should_call.callback(True) @should_call.addCallback def _done(result): assert result @should_call.addErrback def _err(exc): raise exc ee.emit('event') Replace my deferred with a mock
# -*- coding: utf-8 -*- from asyncio import Future, gather, new_event_loop, sleep from mock import Mock from twisted.internet.defer import ensureDeferred from pyee import EventEmitter def test_asyncio_emit(): """Test that event_emitters can handle wrapping coroutines as used with asyncio. """ loop = new_event_loop() ee = EventEmitter(loop=loop) should_call = Future(loop=loop) @ee.on('event') async def event_handler(): should_call.set_result(True) async def create_timeout(loop=loop): await sleep(0.1, loop=loop) if not should_call.done(): raise Exception('should_call timed out!') return should_call.cancel() timeout = create_timeout(loop=loop) @should_call.add_done_callback def _done(result): assert result ee.emit('event') loop.run_until_complete(gather(should_call, timeout, loop=loop)) loop.close() def test_twisted_emit(): """Test that event_emitters can handle wrapping coroutines when using twisted and ensureDeferred. """ ee = EventEmitter(scheduler=ensureDeferred) should_call = Mock() @ee.on('event') async def event_handler(): should_call(True) ee.emit('event') should_call.assert_called_once()
<commit_before># -*- coding: utf-8 -*- from asyncio import Future, gather, new_event_loop, sleep from twisted.internet.defer import Deferred, ensureDeferred from pyee import EventEmitter def test_asyncio_emit(): """Test that event_emitters can handle wrapping coroutines as used with asyncio. """ loop = new_event_loop() ee = EventEmitter(loop=loop) should_call = Future(loop=loop) @ee.on('event') async def event_handler(): should_call.set_result(True) async def create_timeout(loop=loop): await sleep(0.1, loop=loop) if not should_call.done(): raise Exception('should_call timed out!') return should_call.cancel() timeout = create_timeout(loop=loop) @should_call.add_done_callback def _done(result): assert result ee.emit('event') loop.run_until_complete(gather(should_call, timeout, loop=loop)) loop.close() def test_twisted_emit(): """Test that event_emitters can handle wrapping coroutines when using twisted and ensureDeferred. """ ee = EventEmitter(scheduler=ensureDeferred) should_call = Deferred() @ee.on('event') async def event_handler(): should_call.callback(True) @should_call.addCallback def _done(result): assert result @should_call.addErrback def _err(exc): raise exc ee.emit('event') <commit_msg>Replace my deferred with a mock<commit_after>
# -*- coding: utf-8 -*- from asyncio import Future, gather, new_event_loop, sleep from mock import Mock from twisted.internet.defer import ensureDeferred from pyee import EventEmitter def test_asyncio_emit(): """Test that event_emitters can handle wrapping coroutines as used with asyncio. """ loop = new_event_loop() ee = EventEmitter(loop=loop) should_call = Future(loop=loop) @ee.on('event') async def event_handler(): should_call.set_result(True) async def create_timeout(loop=loop): await sleep(0.1, loop=loop) if not should_call.done(): raise Exception('should_call timed out!') return should_call.cancel() timeout = create_timeout(loop=loop) @should_call.add_done_callback def _done(result): assert result ee.emit('event') loop.run_until_complete(gather(should_call, timeout, loop=loop)) loop.close() def test_twisted_emit(): """Test that event_emitters can handle wrapping coroutines when using twisted and ensureDeferred. """ ee = EventEmitter(scheduler=ensureDeferred) should_call = Mock() @ee.on('event') async def event_handler(): should_call(True) ee.emit('event') should_call.assert_called_once()
# -*- coding: utf-8 -*- from asyncio import Future, gather, new_event_loop, sleep from twisted.internet.defer import Deferred, ensureDeferred from pyee import EventEmitter def test_asyncio_emit(): """Test that event_emitters can handle wrapping coroutines as used with asyncio. """ loop = new_event_loop() ee = EventEmitter(loop=loop) should_call = Future(loop=loop) @ee.on('event') async def event_handler(): should_call.set_result(True) async def create_timeout(loop=loop): await sleep(0.1, loop=loop) if not should_call.done(): raise Exception('should_call timed out!') return should_call.cancel() timeout = create_timeout(loop=loop) @should_call.add_done_callback def _done(result): assert result ee.emit('event') loop.run_until_complete(gather(should_call, timeout, loop=loop)) loop.close() def test_twisted_emit(): """Test that event_emitters can handle wrapping coroutines when using twisted and ensureDeferred. """ ee = EventEmitter(scheduler=ensureDeferred) should_call = Deferred() @ee.on('event') async def event_handler(): should_call.callback(True) @should_call.addCallback def _done(result): assert result @should_call.addErrback def _err(exc): raise exc ee.emit('event') Replace my deferred with a mock# -*- coding: utf-8 -*- from asyncio import Future, gather, new_event_loop, sleep from mock import Mock from twisted.internet.defer import ensureDeferred from pyee import EventEmitter def test_asyncio_emit(): """Test that event_emitters can handle wrapping coroutines as used with asyncio. """ loop = new_event_loop() ee = EventEmitter(loop=loop) should_call = Future(loop=loop) @ee.on('event') async def event_handler(): should_call.set_result(True) async def create_timeout(loop=loop): await sleep(0.1, loop=loop) if not should_call.done(): raise Exception('should_call timed out!') return should_call.cancel() timeout = create_timeout(loop=loop) @should_call.add_done_callback def _done(result): assert result ee.emit('event') loop.run_until_complete(gather(should_call, timeout, loop=loop)) loop.close() def test_twisted_emit(): """Test that event_emitters can handle wrapping coroutines when using twisted and ensureDeferred. """ ee = EventEmitter(scheduler=ensureDeferred) should_call = Mock() @ee.on('event') async def event_handler(): should_call(True) ee.emit('event') should_call.assert_called_once()
<commit_before># -*- coding: utf-8 -*- from asyncio import Future, gather, new_event_loop, sleep from twisted.internet.defer import Deferred, ensureDeferred from pyee import EventEmitter def test_asyncio_emit(): """Test that event_emitters can handle wrapping coroutines as used with asyncio. """ loop = new_event_loop() ee = EventEmitter(loop=loop) should_call = Future(loop=loop) @ee.on('event') async def event_handler(): should_call.set_result(True) async def create_timeout(loop=loop): await sleep(0.1, loop=loop) if not should_call.done(): raise Exception('should_call timed out!') return should_call.cancel() timeout = create_timeout(loop=loop) @should_call.add_done_callback def _done(result): assert result ee.emit('event') loop.run_until_complete(gather(should_call, timeout, loop=loop)) loop.close() def test_twisted_emit(): """Test that event_emitters can handle wrapping coroutines when using twisted and ensureDeferred. """ ee = EventEmitter(scheduler=ensureDeferred) should_call = Deferred() @ee.on('event') async def event_handler(): should_call.callback(True) @should_call.addCallback def _done(result): assert result @should_call.addErrback def _err(exc): raise exc ee.emit('event') <commit_msg>Replace my deferred with a mock<commit_after># -*- coding: utf-8 -*- from asyncio import Future, gather, new_event_loop, sleep from mock import Mock from twisted.internet.defer import ensureDeferred from pyee import EventEmitter def test_asyncio_emit(): """Test that event_emitters can handle wrapping coroutines as used with asyncio. """ loop = new_event_loop() ee = EventEmitter(loop=loop) should_call = Future(loop=loop) @ee.on('event') async def event_handler(): should_call.set_result(True) async def create_timeout(loop=loop): await sleep(0.1, loop=loop) if not should_call.done(): raise Exception('should_call timed out!') return should_call.cancel() timeout = create_timeout(loop=loop) @should_call.add_done_callback def _done(result): assert result ee.emit('event') loop.run_until_complete(gather(should_call, timeout, loop=loop)) loop.close() def test_twisted_emit(): """Test that event_emitters can handle wrapping coroutines when using twisted and ensureDeferred. """ ee = EventEmitter(scheduler=ensureDeferred) should_call = Mock() @ee.on('event') async def event_handler(): should_call(True) ee.emit('event') should_call.assert_called_once()
7cde7a232a53138c2c199c16f55e0aecdcbf1aee
tests/test_basic.py
tests/test_basic.py
import sys import pubrunner import pubrunner.command_line def test_countwords(): #pubrunner.pubrun('examples/CountWords/',True,True) sys.argv = ['--test','examples/CountWords/'] pubrunner.command_line.main()
import sys import pubrunner import pubrunner.command_line def test_countwords(): #pubrunner.pubrun('examples/CountWords/',True,True) sys.argv = ['pubrunner', '--test','examples/CountWords/'] pubrunner.command_line.main()
Test wasn't using --test flag properly
Test wasn't using --test flag properly
Python
mit
jakelever/pubrunner,jakelever/pubrunner
import sys import pubrunner import pubrunner.command_line def test_countwords(): #pubrunner.pubrun('examples/CountWords/',True,True) sys.argv = ['--test','examples/CountWords/'] pubrunner.command_line.main() Test wasn't using --test flag properly
import sys import pubrunner import pubrunner.command_line def test_countwords(): #pubrunner.pubrun('examples/CountWords/',True,True) sys.argv = ['pubrunner', '--test','examples/CountWords/'] pubrunner.command_line.main()
<commit_before>import sys import pubrunner import pubrunner.command_line def test_countwords(): #pubrunner.pubrun('examples/CountWords/',True,True) sys.argv = ['--test','examples/CountWords/'] pubrunner.command_line.main() <commit_msg>Test wasn't using --test flag properly<commit_after>
import sys import pubrunner import pubrunner.command_line def test_countwords(): #pubrunner.pubrun('examples/CountWords/',True,True) sys.argv = ['pubrunner', '--test','examples/CountWords/'] pubrunner.command_line.main()
import sys import pubrunner import pubrunner.command_line def test_countwords(): #pubrunner.pubrun('examples/CountWords/',True,True) sys.argv = ['--test','examples/CountWords/'] pubrunner.command_line.main() Test wasn't using --test flag properlyimport sys import pubrunner import pubrunner.command_line def test_countwords(): #pubrunner.pubrun('examples/CountWords/',True,True) sys.argv = ['pubrunner', '--test','examples/CountWords/'] pubrunner.command_line.main()
<commit_before>import sys import pubrunner import pubrunner.command_line def test_countwords(): #pubrunner.pubrun('examples/CountWords/',True,True) sys.argv = ['--test','examples/CountWords/'] pubrunner.command_line.main() <commit_msg>Test wasn't using --test flag properly<commit_after>import sys import pubrunner import pubrunner.command_line def test_countwords(): #pubrunner.pubrun('examples/CountWords/',True,True) sys.argv = ['pubrunner', '--test','examples/CountWords/'] pubrunner.command_line.main()
711afeb0e01bc35b1f82588d814be4275f59f758
spdx/tv_to_rdf.py
spdx/tv_to_rdf.py
#!/usr/bin/env python # Copyright (C) 2017 BMW AG # Author: Thomas Hafner # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import sys import codecs from spdx.parsers.tagvalue import Parser from spdx.parsers.loggers import StandardLogger from spdx.parsers.tagvaluebuilders import Builder from spdx.writers.rdf import write_document def tv_to_rdf(infile_name, outfile_name): """Converts a SPDX file from tag/value format to RDF format.""" parser = Parser(Builder(), StandardLogger()) parser.build() with open(infile_name) as infile: data = infile.read() document, error = parser.parse(data) if not error: with open(outfile_name, mode='w') as outfile: write_document(document, outfile) else: print 'Errors encountered while parsing RDF file.' messages = [] document.validate(messages) print '\n'.join(messages) if __name__ == '__main__': tv_to_rdf(*sys.argv[1:])
#!/usr/bin/env python # Copyright (C) 2017 BMW AG # Author: Thomas Hafner # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import codecs import sys from spdx.parsers.tagvalue import Parser from spdx.parsers.loggers import StandardLogger from spdx.parsers.tagvaluebuilders import Builder from spdx.writers.rdf import write_document def tv_to_rdf(infile_name, outfile_name): """ Convert a SPDX file from tag/value format to RDF format. """ parser = Parser(Builder(), StandardLogger()) parser.build() with open(infile_name) as infile: data = infile.read() document, error = parser.parse(data) if not error: with open(outfile_name, mode='w') as outfile: write_document(document, outfile) else: print 'Errors encountered while parsing RDF file.' messages = [] document.validate(messages) print '\n'.join(messages) if __name__ == '__main__': tv_to_rdf(*sys.argv[1:])
Normalize license header. Update format.
Normalize license header. Update format. Signed-off-by: Philippe Ombredanne <ca95c4a6a4931f366cbdaf5878c5016609417d37@nexb.com>
Python
apache-2.0
spdx/tools-python
#!/usr/bin/env python # Copyright (C) 2017 BMW AG # Author: Thomas Hafner # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import sys import codecs from spdx.parsers.tagvalue import Parser from spdx.parsers.loggers import StandardLogger from spdx.parsers.tagvaluebuilders import Builder from spdx.writers.rdf import write_document def tv_to_rdf(infile_name, outfile_name): """Converts a SPDX file from tag/value format to RDF format.""" parser = Parser(Builder(), StandardLogger()) parser.build() with open(infile_name) as infile: data = infile.read() document, error = parser.parse(data) if not error: with open(outfile_name, mode='w') as outfile: write_document(document, outfile) else: print 'Errors encountered while parsing RDF file.' messages = [] document.validate(messages) print '\n'.join(messages) if __name__ == '__main__': tv_to_rdf(*sys.argv[1:]) Normalize license header. Update format. Signed-off-by: Philippe Ombredanne <ca95c4a6a4931f366cbdaf5878c5016609417d37@nexb.com>
#!/usr/bin/env python # Copyright (C) 2017 BMW AG # Author: Thomas Hafner # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import codecs import sys from spdx.parsers.tagvalue import Parser from spdx.parsers.loggers import StandardLogger from spdx.parsers.tagvaluebuilders import Builder from spdx.writers.rdf import write_document def tv_to_rdf(infile_name, outfile_name): """ Convert a SPDX file from tag/value format to RDF format. """ parser = Parser(Builder(), StandardLogger()) parser.build() with open(infile_name) as infile: data = infile.read() document, error = parser.parse(data) if not error: with open(outfile_name, mode='w') as outfile: write_document(document, outfile) else: print 'Errors encountered while parsing RDF file.' messages = [] document.validate(messages) print '\n'.join(messages) if __name__ == '__main__': tv_to_rdf(*sys.argv[1:])
<commit_before>#!/usr/bin/env python # Copyright (C) 2017 BMW AG # Author: Thomas Hafner # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import sys import codecs from spdx.parsers.tagvalue import Parser from spdx.parsers.loggers import StandardLogger from spdx.parsers.tagvaluebuilders import Builder from spdx.writers.rdf import write_document def tv_to_rdf(infile_name, outfile_name): """Converts a SPDX file from tag/value format to RDF format.""" parser = Parser(Builder(), StandardLogger()) parser.build() with open(infile_name) as infile: data = infile.read() document, error = parser.parse(data) if not error: with open(outfile_name, mode='w') as outfile: write_document(document, outfile) else: print 'Errors encountered while parsing RDF file.' messages = [] document.validate(messages) print '\n'.join(messages) if __name__ == '__main__': tv_to_rdf(*sys.argv[1:]) <commit_msg>Normalize license header. Update format. Signed-off-by: Philippe Ombredanne <ca95c4a6a4931f366cbdaf5878c5016609417d37@nexb.com><commit_after>
#!/usr/bin/env python # Copyright (C) 2017 BMW AG # Author: Thomas Hafner # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import codecs import sys from spdx.parsers.tagvalue import Parser from spdx.parsers.loggers import StandardLogger from spdx.parsers.tagvaluebuilders import Builder from spdx.writers.rdf import write_document def tv_to_rdf(infile_name, outfile_name): """ Convert a SPDX file from tag/value format to RDF format. """ parser = Parser(Builder(), StandardLogger()) parser.build() with open(infile_name) as infile: data = infile.read() document, error = parser.parse(data) if not error: with open(outfile_name, mode='w') as outfile: write_document(document, outfile) else: print 'Errors encountered while parsing RDF file.' messages = [] document.validate(messages) print '\n'.join(messages) if __name__ == '__main__': tv_to_rdf(*sys.argv[1:])
#!/usr/bin/env python # Copyright (C) 2017 BMW AG # Author: Thomas Hafner # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import sys import codecs from spdx.parsers.tagvalue import Parser from spdx.parsers.loggers import StandardLogger from spdx.parsers.tagvaluebuilders import Builder from spdx.writers.rdf import write_document def tv_to_rdf(infile_name, outfile_name): """Converts a SPDX file from tag/value format to RDF format.""" parser = Parser(Builder(), StandardLogger()) parser.build() with open(infile_name) as infile: data = infile.read() document, error = parser.parse(data) if not error: with open(outfile_name, mode='w') as outfile: write_document(document, outfile) else: print 'Errors encountered while parsing RDF file.' messages = [] document.validate(messages) print '\n'.join(messages) if __name__ == '__main__': tv_to_rdf(*sys.argv[1:]) Normalize license header. Update format. Signed-off-by: Philippe Ombredanne <ca95c4a6a4931f366cbdaf5878c5016609417d37@nexb.com>#!/usr/bin/env python # Copyright (C) 2017 BMW AG # Author: Thomas Hafner # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import codecs import sys from spdx.parsers.tagvalue import Parser from spdx.parsers.loggers import StandardLogger from spdx.parsers.tagvaluebuilders import Builder from spdx.writers.rdf import write_document def tv_to_rdf(infile_name, outfile_name): """ Convert a SPDX file from tag/value format to RDF format. """ parser = Parser(Builder(), StandardLogger()) parser.build() with open(infile_name) as infile: data = infile.read() document, error = parser.parse(data) if not error: with open(outfile_name, mode='w') as outfile: write_document(document, outfile) else: print 'Errors encountered while parsing RDF file.' messages = [] document.validate(messages) print '\n'.join(messages) if __name__ == '__main__': tv_to_rdf(*sys.argv[1:])
<commit_before>#!/usr/bin/env python # Copyright (C) 2017 BMW AG # Author: Thomas Hafner # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import sys import codecs from spdx.parsers.tagvalue import Parser from spdx.parsers.loggers import StandardLogger from spdx.parsers.tagvaluebuilders import Builder from spdx.writers.rdf import write_document def tv_to_rdf(infile_name, outfile_name): """Converts a SPDX file from tag/value format to RDF format.""" parser = Parser(Builder(), StandardLogger()) parser.build() with open(infile_name) as infile: data = infile.read() document, error = parser.parse(data) if not error: with open(outfile_name, mode='w') as outfile: write_document(document, outfile) else: print 'Errors encountered while parsing RDF file.' messages = [] document.validate(messages) print '\n'.join(messages) if __name__ == '__main__': tv_to_rdf(*sys.argv[1:]) <commit_msg>Normalize license header. Update format. Signed-off-by: Philippe Ombredanne <ca95c4a6a4931f366cbdaf5878c5016609417d37@nexb.com><commit_after>#!/usr/bin/env python # Copyright (C) 2017 BMW AG # Author: Thomas Hafner # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import codecs import sys from spdx.parsers.tagvalue import Parser from spdx.parsers.loggers import StandardLogger from spdx.parsers.tagvaluebuilders import Builder from spdx.writers.rdf import write_document def tv_to_rdf(infile_name, outfile_name): """ Convert a SPDX file from tag/value format to RDF format. """ parser = Parser(Builder(), StandardLogger()) parser.build() with open(infile_name) as infile: data = infile.read() document, error = parser.parse(data) if not error: with open(outfile_name, mode='w') as outfile: write_document(document, outfile) else: print 'Errors encountered while parsing RDF file.' messages = [] document.validate(messages) print '\n'.join(messages) if __name__ == '__main__': tv_to_rdf(*sys.argv[1:])
fea76a8c9c03fef5c70e2fdd93d97cd3e096de7d
tests/test_mysql.py
tests/test_mysql.py
import unittest from tests.common import load_check import time class TestMySql(unittest.TestCase): def setUp(self): # This should run on pre-2.7 python so no skiptest self.skip = False try: import pymysql except ImportError: self.skip = True def testChecks(self): if not self.skip: agentConfig = { 'version': '0.1', 'api_key': 'toto' } conf = {'init_config': {}, 'instances': [{ 'server': 'localhost', 'user': 'dog', 'pass': 'dog', 'options': {'replication': True}, }]} # Initialize the check from checks.d self.check = load_check('mysql', conf, agentConfig) self.check.run() metrics = self.check.get_metrics() self.assertTrue(len(metrics) >= 8, metrics) time.sleep(1) self.check.run() metrics = self.check.get_metrics() self.assertTrue(len(metrics) >= 16, metrics) if __name__ == '__main__': unittest.main()
import unittest from tests.common import load_check import time class TestMySql(unittest.TestCase): def setUp(self): # This should run on pre-2.7 python so no skiptest self.skip = False try: import pymysql except ImportError: self.skip = True def testChecks(self): if not self.skip: agentConfig = { 'version': '0.1', 'api_key': 'toto' } conf = {'init_config': {}, 'instances': [{ 'server': 'localhost', 'user': 'dog', 'pass': 'dog', 'options': {'replication': True}, }]} # Initialize the check from checks.d self.check = load_check('mysql', conf, agentConfig) self.check.run() metrics = self.check.get_metrics() self.assertTrue(len(metrics) >= 8, metrics) # Service checks service_checks = self.check.get_service_checks() service_checks_count = len(service_checks) self.assertTrue(type(service_checks) == type([])) self.assertTrue(service_checks_count > 0) self.assertEquals(len([sc for sc in service_checks if sc['check'] == "mysql.can_connect"]), 1, service_checks) # Assert that all service checks have the proper tags: host and port self.assertEquals(len([sc for sc in service_checks if "host:localhost" in sc['tags']]), service_checks_count, service_checks) self.assertEquals(len([sc for sc in service_checks if "port:0" in sc['tags']]), service_checks_count, service_checks) time.sleep(1) self.check.run() metrics = self.check.get_metrics() self.assertTrue(len(metrics) >= 16, metrics) if __name__ == '__main__': unittest.main()
Add test for Mysql sc
Add test for Mysql sc
Python
bsd-3-clause
brettlangdon/dd-agent,truthbk/dd-agent,jraede/dd-agent,relateiq/dd-agent,cberry777/dd-agent,tebriel/dd-agent,urosgruber/dd-agent,jraede/dd-agent,ess/dd-agent,Shopify/dd-agent,joelvanvelden/dd-agent,jyogi/purvar-agent,eeroniemi/dd-agent,citrusleaf/dd-agent,indeedops/dd-agent,eeroniemi/dd-agent,Mashape/dd-agent,AniruddhaSAtre/dd-agent,c960657/dd-agent,lookout/dd-agent,pfmooney/dd-agent,truthbk/dd-agent,PagerDuty/dd-agent,c960657/dd-agent,jvassev/dd-agent,benmccann/dd-agent,pmav99/praktoras,relateiq/dd-agent,PagerDuty/dd-agent,gphat/dd-agent,citrusleaf/dd-agent,tebriel/dd-agent,AntoCard/powerdns-recursor_check,jvassev/dd-agent,darron/dd-agent,yuecong/dd-agent,yuecong/dd-agent,zendesk/dd-agent,mderomph-coolblue/dd-agent,benmccann/dd-agent,urosgruber/dd-agent,manolama/dd-agent,citrusleaf/dd-agent,cberry777/dd-agent,ess/dd-agent,cberry777/dd-agent,a20012251/dd-agent,tebriel/dd-agent,mderomph-coolblue/dd-agent,AntoCard/powerdns-recursor_check,jshum/dd-agent,JohnLZeller/dd-agent,AntoCard/powerdns-recursor_check,jshum/dd-agent,pfmooney/dd-agent,ess/dd-agent,PagerDuty/dd-agent,joelvanvelden/dd-agent,AntoCard/powerdns-recursor_check,packetloop/dd-agent,Shopify/dd-agent,pmav99/praktoras,PagerDuty/dd-agent,Shopify/dd-agent,takus/dd-agent,packetloop/dd-agent,relateiq/dd-agent,gphat/dd-agent,jraede/dd-agent,packetloop/dd-agent,gphat/dd-agent,yuecong/dd-agent,jamesandariese/dd-agent,guruxu/dd-agent,huhongbo/dd-agent,pmav99/praktoras,huhongbo/dd-agent,zendesk/dd-agent,brettlangdon/dd-agent,lookout/dd-agent,a20012251/dd-agent,Wattpad/dd-agent,urosgruber/dd-agent,indeedops/dd-agent,takus/dd-agent,GabrielNicolasAvellaneda/dd-agent,polynomial/dd-agent,AntoCard/powerdns-recursor_check,joelvanvelden/dd-agent,JohnLZeller/dd-agent,ess/dd-agent,mderomph-coolblue/dd-agent,Wattpad/dd-agent,yuecong/dd-agent,tebriel/dd-agent,Shopify/dd-agent,jraede/dd-agent,Mashape/dd-agent,oneandoneis2/dd-agent,eeroniemi/dd-agent,amalakar/dd-agent,benmccann/dd-agent,amalakar/dd-agent,takus/dd-agent,jshum/dd-agent,Wattpad/dd-agent,remh/dd-agent,GabrielNicolasAvellaneda/dd-agent,oneandoneis2/dd-agent,remh/dd-agent,indeedops/dd-agent,oneandoneis2/dd-agent,c960657/dd-agent,zendesk/dd-agent,amalakar/dd-agent,remh/dd-agent,truthbk/dd-agent,oneandoneis2/dd-agent,manolama/dd-agent,huhongbo/dd-agent,mderomph-coolblue/dd-agent,Mashape/dd-agent,Mashape/dd-agent,mderomph-coolblue/dd-agent,c960657/dd-agent,pfmooney/dd-agent,amalakar/dd-agent,indeedops/dd-agent,eeroniemi/dd-agent,guruxu/dd-agent,relateiq/dd-agent,JohnLZeller/dd-agent,polynomial/dd-agent,pfmooney/dd-agent,jamesandariese/dd-agent,manolama/dd-agent,truthbk/dd-agent,brettlangdon/dd-agent,jvassev/dd-agent,huhongbo/dd-agent,packetloop/dd-agent,AniruddhaSAtre/dd-agent,PagerDuty/dd-agent,urosgruber/dd-agent,lookout/dd-agent,pmav99/praktoras,jraede/dd-agent,jshum/dd-agent,Shopify/dd-agent,jvassev/dd-agent,polynomial/dd-agent,manolama/dd-agent,lookout/dd-agent,pfmooney/dd-agent,jyogi/purvar-agent,takus/dd-agent,huhongbo/dd-agent,GabrielNicolasAvellaneda/dd-agent,jamesandariese/dd-agent,guruxu/dd-agent,JohnLZeller/dd-agent,jamesandariese/dd-agent,cberry777/dd-agent,jyogi/purvar-agent,gphat/dd-agent,manolama/dd-agent,darron/dd-agent,AniruddhaSAtre/dd-agent,joelvanvelden/dd-agent,indeedops/dd-agent,Wattpad/dd-agent,a20012251/dd-agent,yuecong/dd-agent,amalakar/dd-agent,cberry777/dd-agent,lookout/dd-agent,citrusleaf/dd-agent,Mashape/dd-agent,GabrielNicolasAvellaneda/dd-agent,pmav99/praktoras,remh/dd-agent,brettlangdon/dd-agent,jamesandariese/dd-agent,tebriel/dd-agent,jshum/dd-agent,JohnLZeller/dd-agent,a20012251/dd-agent,benmccann/dd-agent,darron/dd-agent,urosgruber/dd-agent,guruxu/dd-agent,jyogi/purvar-agent,truthbk/dd-agent,joelvanvelden/dd-agent,Wattpad/dd-agent,AniruddhaSAtre/dd-agent,jyogi/purvar-agent,a20012251/dd-agent,eeroniemi/dd-agent,zendesk/dd-agent,guruxu/dd-agent,gphat/dd-agent,takus/dd-agent,packetloop/dd-agent,jvassev/dd-agent,darron/dd-agent,zendesk/dd-agent,GabrielNicolasAvellaneda/dd-agent,c960657/dd-agent,ess/dd-agent,relateiq/dd-agent,polynomial/dd-agent,darron/dd-agent,citrusleaf/dd-agent,benmccann/dd-agent,remh/dd-agent,AniruddhaSAtre/dd-agent,brettlangdon/dd-agent,oneandoneis2/dd-agent,polynomial/dd-agent
import unittest from tests.common import load_check import time class TestMySql(unittest.TestCase): def setUp(self): # This should run on pre-2.7 python so no skiptest self.skip = False try: import pymysql except ImportError: self.skip = True def testChecks(self): if not self.skip: agentConfig = { 'version': '0.1', 'api_key': 'toto' } conf = {'init_config': {}, 'instances': [{ 'server': 'localhost', 'user': 'dog', 'pass': 'dog', 'options': {'replication': True}, }]} # Initialize the check from checks.d self.check = load_check('mysql', conf, agentConfig) self.check.run() metrics = self.check.get_metrics() self.assertTrue(len(metrics) >= 8, metrics) time.sleep(1) self.check.run() metrics = self.check.get_metrics() self.assertTrue(len(metrics) >= 16, metrics) if __name__ == '__main__': unittest.main() Add test for Mysql sc
import unittest from tests.common import load_check import time class TestMySql(unittest.TestCase): def setUp(self): # This should run on pre-2.7 python so no skiptest self.skip = False try: import pymysql except ImportError: self.skip = True def testChecks(self): if not self.skip: agentConfig = { 'version': '0.1', 'api_key': 'toto' } conf = {'init_config': {}, 'instances': [{ 'server': 'localhost', 'user': 'dog', 'pass': 'dog', 'options': {'replication': True}, }]} # Initialize the check from checks.d self.check = load_check('mysql', conf, agentConfig) self.check.run() metrics = self.check.get_metrics() self.assertTrue(len(metrics) >= 8, metrics) # Service checks service_checks = self.check.get_service_checks() service_checks_count = len(service_checks) self.assertTrue(type(service_checks) == type([])) self.assertTrue(service_checks_count > 0) self.assertEquals(len([sc for sc in service_checks if sc['check'] == "mysql.can_connect"]), 1, service_checks) # Assert that all service checks have the proper tags: host and port self.assertEquals(len([sc for sc in service_checks if "host:localhost" in sc['tags']]), service_checks_count, service_checks) self.assertEquals(len([sc for sc in service_checks if "port:0" in sc['tags']]), service_checks_count, service_checks) time.sleep(1) self.check.run() metrics = self.check.get_metrics() self.assertTrue(len(metrics) >= 16, metrics) if __name__ == '__main__': unittest.main()
<commit_before>import unittest from tests.common import load_check import time class TestMySql(unittest.TestCase): def setUp(self): # This should run on pre-2.7 python so no skiptest self.skip = False try: import pymysql except ImportError: self.skip = True def testChecks(self): if not self.skip: agentConfig = { 'version': '0.1', 'api_key': 'toto' } conf = {'init_config': {}, 'instances': [{ 'server': 'localhost', 'user': 'dog', 'pass': 'dog', 'options': {'replication': True}, }]} # Initialize the check from checks.d self.check = load_check('mysql', conf, agentConfig) self.check.run() metrics = self.check.get_metrics() self.assertTrue(len(metrics) >= 8, metrics) time.sleep(1) self.check.run() metrics = self.check.get_metrics() self.assertTrue(len(metrics) >= 16, metrics) if __name__ == '__main__': unittest.main() <commit_msg>Add test for Mysql sc<commit_after>
import unittest from tests.common import load_check import time class TestMySql(unittest.TestCase): def setUp(self): # This should run on pre-2.7 python so no skiptest self.skip = False try: import pymysql except ImportError: self.skip = True def testChecks(self): if not self.skip: agentConfig = { 'version': '0.1', 'api_key': 'toto' } conf = {'init_config': {}, 'instances': [{ 'server': 'localhost', 'user': 'dog', 'pass': 'dog', 'options': {'replication': True}, }]} # Initialize the check from checks.d self.check = load_check('mysql', conf, agentConfig) self.check.run() metrics = self.check.get_metrics() self.assertTrue(len(metrics) >= 8, metrics) # Service checks service_checks = self.check.get_service_checks() service_checks_count = len(service_checks) self.assertTrue(type(service_checks) == type([])) self.assertTrue(service_checks_count > 0) self.assertEquals(len([sc for sc in service_checks if sc['check'] == "mysql.can_connect"]), 1, service_checks) # Assert that all service checks have the proper tags: host and port self.assertEquals(len([sc for sc in service_checks if "host:localhost" in sc['tags']]), service_checks_count, service_checks) self.assertEquals(len([sc for sc in service_checks if "port:0" in sc['tags']]), service_checks_count, service_checks) time.sleep(1) self.check.run() metrics = self.check.get_metrics() self.assertTrue(len(metrics) >= 16, metrics) if __name__ == '__main__': unittest.main()
import unittest from tests.common import load_check import time class TestMySql(unittest.TestCase): def setUp(self): # This should run on pre-2.7 python so no skiptest self.skip = False try: import pymysql except ImportError: self.skip = True def testChecks(self): if not self.skip: agentConfig = { 'version': '0.1', 'api_key': 'toto' } conf = {'init_config': {}, 'instances': [{ 'server': 'localhost', 'user': 'dog', 'pass': 'dog', 'options': {'replication': True}, }]} # Initialize the check from checks.d self.check = load_check('mysql', conf, agentConfig) self.check.run() metrics = self.check.get_metrics() self.assertTrue(len(metrics) >= 8, metrics) time.sleep(1) self.check.run() metrics = self.check.get_metrics() self.assertTrue(len(metrics) >= 16, metrics) if __name__ == '__main__': unittest.main() Add test for Mysql scimport unittest from tests.common import load_check import time class TestMySql(unittest.TestCase): def setUp(self): # This should run on pre-2.7 python so no skiptest self.skip = False try: import pymysql except ImportError: self.skip = True def testChecks(self): if not self.skip: agentConfig = { 'version': '0.1', 'api_key': 'toto' } conf = {'init_config': {}, 'instances': [{ 'server': 'localhost', 'user': 'dog', 'pass': 'dog', 'options': {'replication': True}, }]} # Initialize the check from checks.d self.check = load_check('mysql', conf, agentConfig) self.check.run() metrics = self.check.get_metrics() self.assertTrue(len(metrics) >= 8, metrics) # Service checks service_checks = self.check.get_service_checks() service_checks_count = len(service_checks) self.assertTrue(type(service_checks) == type([])) self.assertTrue(service_checks_count > 0) self.assertEquals(len([sc for sc in service_checks if sc['check'] == "mysql.can_connect"]), 1, service_checks) # Assert that all service checks have the proper tags: host and port self.assertEquals(len([sc for sc in service_checks if "host:localhost" in sc['tags']]), service_checks_count, service_checks) self.assertEquals(len([sc for sc in service_checks if "port:0" in sc['tags']]), service_checks_count, service_checks) time.sleep(1) self.check.run() metrics = self.check.get_metrics() self.assertTrue(len(metrics) >= 16, metrics) if __name__ == '__main__': unittest.main()
<commit_before>import unittest from tests.common import load_check import time class TestMySql(unittest.TestCase): def setUp(self): # This should run on pre-2.7 python so no skiptest self.skip = False try: import pymysql except ImportError: self.skip = True def testChecks(self): if not self.skip: agentConfig = { 'version': '0.1', 'api_key': 'toto' } conf = {'init_config': {}, 'instances': [{ 'server': 'localhost', 'user': 'dog', 'pass': 'dog', 'options': {'replication': True}, }]} # Initialize the check from checks.d self.check = load_check('mysql', conf, agentConfig) self.check.run() metrics = self.check.get_metrics() self.assertTrue(len(metrics) >= 8, metrics) time.sleep(1) self.check.run() metrics = self.check.get_metrics() self.assertTrue(len(metrics) >= 16, metrics) if __name__ == '__main__': unittest.main() <commit_msg>Add test for Mysql sc<commit_after>import unittest from tests.common import load_check import time class TestMySql(unittest.TestCase): def setUp(self): # This should run on pre-2.7 python so no skiptest self.skip = False try: import pymysql except ImportError: self.skip = True def testChecks(self): if not self.skip: agentConfig = { 'version': '0.1', 'api_key': 'toto' } conf = {'init_config': {}, 'instances': [{ 'server': 'localhost', 'user': 'dog', 'pass': 'dog', 'options': {'replication': True}, }]} # Initialize the check from checks.d self.check = load_check('mysql', conf, agentConfig) self.check.run() metrics = self.check.get_metrics() self.assertTrue(len(metrics) >= 8, metrics) # Service checks service_checks = self.check.get_service_checks() service_checks_count = len(service_checks) self.assertTrue(type(service_checks) == type([])) self.assertTrue(service_checks_count > 0) self.assertEquals(len([sc for sc in service_checks if sc['check'] == "mysql.can_connect"]), 1, service_checks) # Assert that all service checks have the proper tags: host and port self.assertEquals(len([sc for sc in service_checks if "host:localhost" in sc['tags']]), service_checks_count, service_checks) self.assertEquals(len([sc for sc in service_checks if "port:0" in sc['tags']]), service_checks_count, service_checks) time.sleep(1) self.check.run() metrics = self.check.get_metrics() self.assertTrue(len(metrics) >= 16, metrics) if __name__ == '__main__': unittest.main()
42f49d19b3657c62208aacee0f2dd77081bb5aa2
zerver/migrations/0189_userprofile_add_some_emojisets.py
zerver/migrations/0189_userprofile_add_some_emojisets.py
# -*- coding: utf-8 -*- # Generated by Django 1.11.14 on 2018-08-28 19:01 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('zerver', '0188_userprofile_enable_login_emails'), ] operations = [ migrations.AlterField( model_name='userprofile', name='emojiset', field=models.CharField(choices=[('google', 'Google modern'), ('google-blob', 'Google classic'), ('twitter', 'Twitter'), ('text', 'Plain text')], default='google-blob', max_length=20), ), ]
# -*- coding: utf-8 -*- # Generated by Django 1.11.14 on 2018-08-28 19:01 from __future__ import unicode_literals from django.db import migrations, models from django.db.backends.postgresql_psycopg2.schema import DatabaseSchemaEditor from django.db.migrations.state import StateApps def change_emojiset_choice(apps: StateApps, schema_editor: DatabaseSchemaEditor) -> None: UserProfile = apps.get_model('zerver', 'UserProfile') UserProfile.objects.filter(emojiset='google').update(emojiset='google-blob') class Migration(migrations.Migration): dependencies = [ ('zerver', '0188_userprofile_enable_login_emails'), ] operations = [ migrations.AlterField( model_name='userprofile', name='emojiset', field=models.CharField(choices=[('google', 'Google modern'), ('google-blob', 'Google classic'), ('twitter', 'Twitter'), ('text', 'Plain text')], default='google-blob', max_length=20), ), migrations.RunPython( change_emojiset_choice, reverse_code=migrations.RunPython.noop), ]
Set default emoji to google-blob for existing users too.
emoji: Set default emoji to google-blob for existing users too. This fixes an inconsistent test failure with test_users.py (that depended on the ordering between this migration and the creation of test database users like hamlet).
Python
apache-2.0
brainwane/zulip,brainwane/zulip,shubhamdhama/zulip,zulip/zulip,shubhamdhama/zulip,rht/zulip,showell/zulip,eeshangarg/zulip,punchagan/zulip,hackerkid/zulip,synicalsyntax/zulip,tommyip/zulip,showell/zulip,rishig/zulip,shubhamdhama/zulip,jackrzhang/zulip,synicalsyntax/zulip,dhcrzf/zulip,kou/zulip,andersk/zulip,dhcrzf/zulip,jackrzhang/zulip,rht/zulip,rishig/zulip,jackrzhang/zulip,tommyip/zulip,zulip/zulip,zulip/zulip,rishig/zulip,hackerkid/zulip,hackerkid/zulip,rht/zulip,timabbott/zulip,punchagan/zulip,timabbott/zulip,eeshangarg/zulip,jackrzhang/zulip,showell/zulip,tommyip/zulip,synicalsyntax/zulip,punchagan/zulip,rishig/zulip,dhcrzf/zulip,kou/zulip,showell/zulip,brainwane/zulip,andersk/zulip,hackerkid/zulip,tommyip/zulip,eeshangarg/zulip,eeshangarg/zulip,jackrzhang/zulip,punchagan/zulip,brainwane/zulip,rht/zulip,timabbott/zulip,zulip/zulip,andersk/zulip,timabbott/zulip,showell/zulip,punchagan/zulip,showell/zulip,shubhamdhama/zulip,zulip/zulip,shubhamdhama/zulip,punchagan/zulip,andersk/zulip,rht/zulip,jackrzhang/zulip,rishig/zulip,tommyip/zulip,andersk/zulip,showell/zulip,hackerkid/zulip,dhcrzf/zulip,timabbott/zulip,dhcrzf/zulip,brainwane/zulip,hackerkid/zulip,eeshangarg/zulip,hackerkid/zulip,eeshangarg/zulip,brainwane/zulip,kou/zulip,kou/zulip,shubhamdhama/zulip,kou/zulip,rht/zulip,kou/zulip,synicalsyntax/zulip,rishig/zulip,zulip/zulip,andersk/zulip,rishig/zulip,dhcrzf/zulip,rht/zulip,timabbott/zulip,dhcrzf/zulip,andersk/zulip,jackrzhang/zulip,synicalsyntax/zulip,shubhamdhama/zulip,tommyip/zulip,brainwane/zulip,synicalsyntax/zulip,zulip/zulip,eeshangarg/zulip,punchagan/zulip,tommyip/zulip,timabbott/zulip,synicalsyntax/zulip,kou/zulip
# -*- coding: utf-8 -*- # Generated by Django 1.11.14 on 2018-08-28 19:01 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('zerver', '0188_userprofile_enable_login_emails'), ] operations = [ migrations.AlterField( model_name='userprofile', name='emojiset', field=models.CharField(choices=[('google', 'Google modern'), ('google-blob', 'Google classic'), ('twitter', 'Twitter'), ('text', 'Plain text')], default='google-blob', max_length=20), ), ] emoji: Set default emoji to google-blob for existing users too. This fixes an inconsistent test failure with test_users.py (that depended on the ordering between this migration and the creation of test database users like hamlet).
# -*- coding: utf-8 -*- # Generated by Django 1.11.14 on 2018-08-28 19:01 from __future__ import unicode_literals from django.db import migrations, models from django.db.backends.postgresql_psycopg2.schema import DatabaseSchemaEditor from django.db.migrations.state import StateApps def change_emojiset_choice(apps: StateApps, schema_editor: DatabaseSchemaEditor) -> None: UserProfile = apps.get_model('zerver', 'UserProfile') UserProfile.objects.filter(emojiset='google').update(emojiset='google-blob') class Migration(migrations.Migration): dependencies = [ ('zerver', '0188_userprofile_enable_login_emails'), ] operations = [ migrations.AlterField( model_name='userprofile', name='emojiset', field=models.CharField(choices=[('google', 'Google modern'), ('google-blob', 'Google classic'), ('twitter', 'Twitter'), ('text', 'Plain text')], default='google-blob', max_length=20), ), migrations.RunPython( change_emojiset_choice, reverse_code=migrations.RunPython.noop), ]
<commit_before># -*- coding: utf-8 -*- # Generated by Django 1.11.14 on 2018-08-28 19:01 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('zerver', '0188_userprofile_enable_login_emails'), ] operations = [ migrations.AlterField( model_name='userprofile', name='emojiset', field=models.CharField(choices=[('google', 'Google modern'), ('google-blob', 'Google classic'), ('twitter', 'Twitter'), ('text', 'Plain text')], default='google-blob', max_length=20), ), ] <commit_msg>emoji: Set default emoji to google-blob for existing users too. This fixes an inconsistent test failure with test_users.py (that depended on the ordering between this migration and the creation of test database users like hamlet).<commit_after>
# -*- coding: utf-8 -*- # Generated by Django 1.11.14 on 2018-08-28 19:01 from __future__ import unicode_literals from django.db import migrations, models from django.db.backends.postgresql_psycopg2.schema import DatabaseSchemaEditor from django.db.migrations.state import StateApps def change_emojiset_choice(apps: StateApps, schema_editor: DatabaseSchemaEditor) -> None: UserProfile = apps.get_model('zerver', 'UserProfile') UserProfile.objects.filter(emojiset='google').update(emojiset='google-blob') class Migration(migrations.Migration): dependencies = [ ('zerver', '0188_userprofile_enable_login_emails'), ] operations = [ migrations.AlterField( model_name='userprofile', name='emojiset', field=models.CharField(choices=[('google', 'Google modern'), ('google-blob', 'Google classic'), ('twitter', 'Twitter'), ('text', 'Plain text')], default='google-blob', max_length=20), ), migrations.RunPython( change_emojiset_choice, reverse_code=migrations.RunPython.noop), ]
# -*- coding: utf-8 -*- # Generated by Django 1.11.14 on 2018-08-28 19:01 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('zerver', '0188_userprofile_enable_login_emails'), ] operations = [ migrations.AlterField( model_name='userprofile', name='emojiset', field=models.CharField(choices=[('google', 'Google modern'), ('google-blob', 'Google classic'), ('twitter', 'Twitter'), ('text', 'Plain text')], default='google-blob', max_length=20), ), ] emoji: Set default emoji to google-blob for existing users too. This fixes an inconsistent test failure with test_users.py (that depended on the ordering between this migration and the creation of test database users like hamlet).# -*- coding: utf-8 -*- # Generated by Django 1.11.14 on 2018-08-28 19:01 from __future__ import unicode_literals from django.db import migrations, models from django.db.backends.postgresql_psycopg2.schema import DatabaseSchemaEditor from django.db.migrations.state import StateApps def change_emojiset_choice(apps: StateApps, schema_editor: DatabaseSchemaEditor) -> None: UserProfile = apps.get_model('zerver', 'UserProfile') UserProfile.objects.filter(emojiset='google').update(emojiset='google-blob') class Migration(migrations.Migration): dependencies = [ ('zerver', '0188_userprofile_enable_login_emails'), ] operations = [ migrations.AlterField( model_name='userprofile', name='emojiset', field=models.CharField(choices=[('google', 'Google modern'), ('google-blob', 'Google classic'), ('twitter', 'Twitter'), ('text', 'Plain text')], default='google-blob', max_length=20), ), migrations.RunPython( change_emojiset_choice, reverse_code=migrations.RunPython.noop), ]
<commit_before># -*- coding: utf-8 -*- # Generated by Django 1.11.14 on 2018-08-28 19:01 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('zerver', '0188_userprofile_enable_login_emails'), ] operations = [ migrations.AlterField( model_name='userprofile', name='emojiset', field=models.CharField(choices=[('google', 'Google modern'), ('google-blob', 'Google classic'), ('twitter', 'Twitter'), ('text', 'Plain text')], default='google-blob', max_length=20), ), ] <commit_msg>emoji: Set default emoji to google-blob for existing users too. This fixes an inconsistent test failure with test_users.py (that depended on the ordering between this migration and the creation of test database users like hamlet).<commit_after># -*- coding: utf-8 -*- # Generated by Django 1.11.14 on 2018-08-28 19:01 from __future__ import unicode_literals from django.db import migrations, models from django.db.backends.postgresql_psycopg2.schema import DatabaseSchemaEditor from django.db.migrations.state import StateApps def change_emojiset_choice(apps: StateApps, schema_editor: DatabaseSchemaEditor) -> None: UserProfile = apps.get_model('zerver', 'UserProfile') UserProfile.objects.filter(emojiset='google').update(emojiset='google-blob') class Migration(migrations.Migration): dependencies = [ ('zerver', '0188_userprofile_enable_login_emails'), ] operations = [ migrations.AlterField( model_name='userprofile', name='emojiset', field=models.CharField(choices=[('google', 'Google modern'), ('google-blob', 'Google classic'), ('twitter', 'Twitter'), ('text', 'Plain text')], default='google-blob', max_length=20), ), migrations.RunPython( change_emojiset_choice, reverse_code=migrations.RunPython.noop), ]
c46398091fbe591bbe79744ed4371fddcc454912
IPython/html/terminal/handlers.py
IPython/html/terminal/handlers.py
"""Tornado handlers for the terminal emulator.""" # Copyright (c) IPython Development Team. # Distributed under the terms of the Modified BSD License. from tornado import web import terminado from ..base.handlers import IPythonHandler class TerminalHandler(IPythonHandler): """Render the terminal interface.""" @web.authenticated def get(self, term_name): self.write(self.render_template('terminal.html', ws_path="terminals/websocket/%s" % term_name)) class NewTerminalHandler(IPythonHandler): """Redirect to a new terminal.""" @web.authenticated def get(self): name, _ = self.application.terminal_manager.new_named_terminal() self.redirect("/terminals/%s" % name, permanent=False) class TermSocket(terminado.TermSocket, IPythonHandler): def get(self, *args, **kwargs): if not self.get_current_user(): raise web.HTTPError(403) return super(TermSocket, self).get(*args, **kwargs)
"""Tornado handlers for the terminal emulator.""" # Copyright (c) IPython Development Team. # Distributed under the terms of the Modified BSD License. from tornado import web import terminado from ..base.handlers import IPythonHandler class TerminalHandler(IPythonHandler): """Render the terminal interface.""" @web.authenticated def get(self, term_name): self.write(self.render_template('terminal.html', ws_path="terminals/websocket/%s" % term_name)) class NewTerminalHandler(IPythonHandler): """Redirect to a new terminal.""" @web.authenticated def get(self): name, _ = self.application.terminal_manager.new_named_terminal() self.redirect(name, permanent=False) class TermSocket(terminado.TermSocket, IPythonHandler): def get(self, *args, **kwargs): if not self.get_current_user(): raise web.HTTPError(403) return super(TermSocket, self).get(*args, **kwargs)
Use relative URL for redirect in NewTerminalHandler
Use relative URL for redirect in NewTerminalHandler
Python
bsd-3-clause
ipython/ipython,ipython/ipython
"""Tornado handlers for the terminal emulator.""" # Copyright (c) IPython Development Team. # Distributed under the terms of the Modified BSD License. from tornado import web import terminado from ..base.handlers import IPythonHandler class TerminalHandler(IPythonHandler): """Render the terminal interface.""" @web.authenticated def get(self, term_name): self.write(self.render_template('terminal.html', ws_path="terminals/websocket/%s" % term_name)) class NewTerminalHandler(IPythonHandler): """Redirect to a new terminal.""" @web.authenticated def get(self): name, _ = self.application.terminal_manager.new_named_terminal() self.redirect("/terminals/%s" % name, permanent=False) class TermSocket(terminado.TermSocket, IPythonHandler): def get(self, *args, **kwargs): if not self.get_current_user(): raise web.HTTPError(403) return super(TermSocket, self).get(*args, **kwargs) Use relative URL for redirect in NewTerminalHandler
"""Tornado handlers for the terminal emulator.""" # Copyright (c) IPython Development Team. # Distributed under the terms of the Modified BSD License. from tornado import web import terminado from ..base.handlers import IPythonHandler class TerminalHandler(IPythonHandler): """Render the terminal interface.""" @web.authenticated def get(self, term_name): self.write(self.render_template('terminal.html', ws_path="terminals/websocket/%s" % term_name)) class NewTerminalHandler(IPythonHandler): """Redirect to a new terminal.""" @web.authenticated def get(self): name, _ = self.application.terminal_manager.new_named_terminal() self.redirect(name, permanent=False) class TermSocket(terminado.TermSocket, IPythonHandler): def get(self, *args, **kwargs): if not self.get_current_user(): raise web.HTTPError(403) return super(TermSocket, self).get(*args, **kwargs)
<commit_before>"""Tornado handlers for the terminal emulator.""" # Copyright (c) IPython Development Team. # Distributed under the terms of the Modified BSD License. from tornado import web import terminado from ..base.handlers import IPythonHandler class TerminalHandler(IPythonHandler): """Render the terminal interface.""" @web.authenticated def get(self, term_name): self.write(self.render_template('terminal.html', ws_path="terminals/websocket/%s" % term_name)) class NewTerminalHandler(IPythonHandler): """Redirect to a new terminal.""" @web.authenticated def get(self): name, _ = self.application.terminal_manager.new_named_terminal() self.redirect("/terminals/%s" % name, permanent=False) class TermSocket(terminado.TermSocket, IPythonHandler): def get(self, *args, **kwargs): if not self.get_current_user(): raise web.HTTPError(403) return super(TermSocket, self).get(*args, **kwargs) <commit_msg>Use relative URL for redirect in NewTerminalHandler<commit_after>
"""Tornado handlers for the terminal emulator.""" # Copyright (c) IPython Development Team. # Distributed under the terms of the Modified BSD License. from tornado import web import terminado from ..base.handlers import IPythonHandler class TerminalHandler(IPythonHandler): """Render the terminal interface.""" @web.authenticated def get(self, term_name): self.write(self.render_template('terminal.html', ws_path="terminals/websocket/%s" % term_name)) class NewTerminalHandler(IPythonHandler): """Redirect to a new terminal.""" @web.authenticated def get(self): name, _ = self.application.terminal_manager.new_named_terminal() self.redirect(name, permanent=False) class TermSocket(terminado.TermSocket, IPythonHandler): def get(self, *args, **kwargs): if not self.get_current_user(): raise web.HTTPError(403) return super(TermSocket, self).get(*args, **kwargs)
"""Tornado handlers for the terminal emulator.""" # Copyright (c) IPython Development Team. # Distributed under the terms of the Modified BSD License. from tornado import web import terminado from ..base.handlers import IPythonHandler class TerminalHandler(IPythonHandler): """Render the terminal interface.""" @web.authenticated def get(self, term_name): self.write(self.render_template('terminal.html', ws_path="terminals/websocket/%s" % term_name)) class NewTerminalHandler(IPythonHandler): """Redirect to a new terminal.""" @web.authenticated def get(self): name, _ = self.application.terminal_manager.new_named_terminal() self.redirect("/terminals/%s" % name, permanent=False) class TermSocket(terminado.TermSocket, IPythonHandler): def get(self, *args, **kwargs): if not self.get_current_user(): raise web.HTTPError(403) return super(TermSocket, self).get(*args, **kwargs) Use relative URL for redirect in NewTerminalHandler"""Tornado handlers for the terminal emulator.""" # Copyright (c) IPython Development Team. # Distributed under the terms of the Modified BSD License. from tornado import web import terminado from ..base.handlers import IPythonHandler class TerminalHandler(IPythonHandler): """Render the terminal interface.""" @web.authenticated def get(self, term_name): self.write(self.render_template('terminal.html', ws_path="terminals/websocket/%s" % term_name)) class NewTerminalHandler(IPythonHandler): """Redirect to a new terminal.""" @web.authenticated def get(self): name, _ = self.application.terminal_manager.new_named_terminal() self.redirect(name, permanent=False) class TermSocket(terminado.TermSocket, IPythonHandler): def get(self, *args, **kwargs): if not self.get_current_user(): raise web.HTTPError(403) return super(TermSocket, self).get(*args, **kwargs)
<commit_before>"""Tornado handlers for the terminal emulator.""" # Copyright (c) IPython Development Team. # Distributed under the terms of the Modified BSD License. from tornado import web import terminado from ..base.handlers import IPythonHandler class TerminalHandler(IPythonHandler): """Render the terminal interface.""" @web.authenticated def get(self, term_name): self.write(self.render_template('terminal.html', ws_path="terminals/websocket/%s" % term_name)) class NewTerminalHandler(IPythonHandler): """Redirect to a new terminal.""" @web.authenticated def get(self): name, _ = self.application.terminal_manager.new_named_terminal() self.redirect("/terminals/%s" % name, permanent=False) class TermSocket(terminado.TermSocket, IPythonHandler): def get(self, *args, **kwargs): if not self.get_current_user(): raise web.HTTPError(403) return super(TermSocket, self).get(*args, **kwargs) <commit_msg>Use relative URL for redirect in NewTerminalHandler<commit_after>"""Tornado handlers for the terminal emulator.""" # Copyright (c) IPython Development Team. # Distributed under the terms of the Modified BSD License. from tornado import web import terminado from ..base.handlers import IPythonHandler class TerminalHandler(IPythonHandler): """Render the terminal interface.""" @web.authenticated def get(self, term_name): self.write(self.render_template('terminal.html', ws_path="terminals/websocket/%s" % term_name)) class NewTerminalHandler(IPythonHandler): """Redirect to a new terminal.""" @web.authenticated def get(self): name, _ = self.application.terminal_manager.new_named_terminal() self.redirect(name, permanent=False) class TermSocket(terminado.TermSocket, IPythonHandler): def get(self, *args, **kwargs): if not self.get_current_user(): raise web.HTTPError(403) return super(TermSocket, self).get(*args, **kwargs)
d04d56bdb79d6e6ac04af1642cb8c3ceff325d75
dddp/test/__init__.py
dddp/test/__init__.py
# This file mainly exists to allow `python setup.py test` to work. import os import sys import dddp import django from django.test.utils import get_runner from django.conf import settings def run_tests(): os.environ['DJANGO_SETTINGS_MODULE'] = 'dddp.test.test_project.settings' dddp.greenify() django.setup() test_runner = get_runner(settings)() failures = test_runner.run_tests(['dddp', 'dddp.test.django_todos']) sys.exit(bool(failures))
# This file mainly exists to allow `python setup.py test` to work. import os import sys import dddp def run_tests(): os.environ['DJANGO_SETTINGS_MODULE'] = 'dddp.test.test_project.settings' dddp.greenify() import django from django.test.utils import get_runner from django.conf import settings django.setup() test_runner = get_runner(settings)() failures = test_runner.run_tests(['dddp', 'dddp.test.django_todos']) sys.exit(bool(failures))
Apply gevent monkey patching before importing django (which imports threading).
Apply gevent monkey patching before importing django (which imports threading).
Python
mit
django-ddp/django-ddp,django-ddp/django-ddp,commoncode/django-ddp,commoncode/django-ddp,django-ddp/django-ddp,commoncode/django-ddp,django-ddp/django-ddp,commoncode/django-ddp
# This file mainly exists to allow `python setup.py test` to work. import os import sys import dddp import django from django.test.utils import get_runner from django.conf import settings def run_tests(): os.environ['DJANGO_SETTINGS_MODULE'] = 'dddp.test.test_project.settings' dddp.greenify() django.setup() test_runner = get_runner(settings)() failures = test_runner.run_tests(['dddp', 'dddp.test.django_todos']) sys.exit(bool(failures)) Apply gevent monkey patching before importing django (which imports threading).
# This file mainly exists to allow `python setup.py test` to work. import os import sys import dddp def run_tests(): os.environ['DJANGO_SETTINGS_MODULE'] = 'dddp.test.test_project.settings' dddp.greenify() import django from django.test.utils import get_runner from django.conf import settings django.setup() test_runner = get_runner(settings)() failures = test_runner.run_tests(['dddp', 'dddp.test.django_todos']) sys.exit(bool(failures))
<commit_before># This file mainly exists to allow `python setup.py test` to work. import os import sys import dddp import django from django.test.utils import get_runner from django.conf import settings def run_tests(): os.environ['DJANGO_SETTINGS_MODULE'] = 'dddp.test.test_project.settings' dddp.greenify() django.setup() test_runner = get_runner(settings)() failures = test_runner.run_tests(['dddp', 'dddp.test.django_todos']) sys.exit(bool(failures)) <commit_msg>Apply gevent monkey patching before importing django (which imports threading).<commit_after>
# This file mainly exists to allow `python setup.py test` to work. import os import sys import dddp def run_tests(): os.environ['DJANGO_SETTINGS_MODULE'] = 'dddp.test.test_project.settings' dddp.greenify() import django from django.test.utils import get_runner from django.conf import settings django.setup() test_runner = get_runner(settings)() failures = test_runner.run_tests(['dddp', 'dddp.test.django_todos']) sys.exit(bool(failures))
# This file mainly exists to allow `python setup.py test` to work. import os import sys import dddp import django from django.test.utils import get_runner from django.conf import settings def run_tests(): os.environ['DJANGO_SETTINGS_MODULE'] = 'dddp.test.test_project.settings' dddp.greenify() django.setup() test_runner = get_runner(settings)() failures = test_runner.run_tests(['dddp', 'dddp.test.django_todos']) sys.exit(bool(failures)) Apply gevent monkey patching before importing django (which imports threading).# This file mainly exists to allow `python setup.py test` to work. import os import sys import dddp def run_tests(): os.environ['DJANGO_SETTINGS_MODULE'] = 'dddp.test.test_project.settings' dddp.greenify() import django from django.test.utils import get_runner from django.conf import settings django.setup() test_runner = get_runner(settings)() failures = test_runner.run_tests(['dddp', 'dddp.test.django_todos']) sys.exit(bool(failures))
<commit_before># This file mainly exists to allow `python setup.py test` to work. import os import sys import dddp import django from django.test.utils import get_runner from django.conf import settings def run_tests(): os.environ['DJANGO_SETTINGS_MODULE'] = 'dddp.test.test_project.settings' dddp.greenify() django.setup() test_runner = get_runner(settings)() failures = test_runner.run_tests(['dddp', 'dddp.test.django_todos']) sys.exit(bool(failures)) <commit_msg>Apply gevent monkey patching before importing django (which imports threading).<commit_after># This file mainly exists to allow `python setup.py test` to work. import os import sys import dddp def run_tests(): os.environ['DJANGO_SETTINGS_MODULE'] = 'dddp.test.test_project.settings' dddp.greenify() import django from django.test.utils import get_runner from django.conf import settings django.setup() test_runner = get_runner(settings)() failures = test_runner.run_tests(['dddp', 'dddp.test.django_todos']) sys.exit(bool(failures))
b2bc500c6a715493c4187e5b1c62e02717d8e31c
glue/plugins/dendro_viewer/__init__.py
glue/plugins/dendro_viewer/__init__.py
def setup(): from .qt_widget import DendroWidget from glue.config import qt_client qt_client.add(DendroWidget)
def setup(): from glue.config import qt_client from .qt_widget import DendroWidget from .data_factory import load_dendro qt_client.add(DendroWidget)
Add missing import of dendrogram data factory
Add missing import of dendrogram data factory
Python
bsd-3-clause
stscieisenhamer/glue,saimn/glue,saimn/glue,stscieisenhamer/glue
def setup(): from .qt_widget import DendroWidget from glue.config import qt_client qt_client.add(DendroWidget) Add missing import of dendrogram data factory
def setup(): from glue.config import qt_client from .qt_widget import DendroWidget from .data_factory import load_dendro qt_client.add(DendroWidget)
<commit_before>def setup(): from .qt_widget import DendroWidget from glue.config import qt_client qt_client.add(DendroWidget) <commit_msg>Add missing import of dendrogram data factory<commit_after>
def setup(): from glue.config import qt_client from .qt_widget import DendroWidget from .data_factory import load_dendro qt_client.add(DendroWidget)
def setup(): from .qt_widget import DendroWidget from glue.config import qt_client qt_client.add(DendroWidget) Add missing import of dendrogram data factorydef setup(): from glue.config import qt_client from .qt_widget import DendroWidget from .data_factory import load_dendro qt_client.add(DendroWidget)
<commit_before>def setup(): from .qt_widget import DendroWidget from glue.config import qt_client qt_client.add(DendroWidget) <commit_msg>Add missing import of dendrogram data factory<commit_after>def setup(): from glue.config import qt_client from .qt_widget import DendroWidget from .data_factory import load_dendro qt_client.add(DendroWidget)
539c7a85a84fdb2fbe8ee3f5803778baf0c66841
wmt/flask/__init__.py
wmt/flask/__init__.py
import os from flask import Flask from flask_login import LoginManager from passlib.context import CryptContext from .core import db from .blueprints import register_blueprints from .errors import ERROR_HANDLERS class User(object): def __init__(self, id): self._id = id def is_authenticated(self): return True def is_active(self): return True def is_anonymous(self): return False def get_id(self): return self._id def create_app(settings_override=None, register_security_blueprint=True): app = Flask(__name__, instance_relative_config=True) login_manager = LoginManager() login_manager.init_app(app) @login_manager.user_loader def load_user(userid): return User(userid) @app.before_first_request def create_database(): db.create_all() app.config.from_object('wmt.flask.settings') app.config.from_pyfile('settings.cfg', silent=True) app.config.from_object(settings_override) app.config['pw'] = CryptContext.from_string( app.config['CRYPT_INI_CONTENTS'], section='passlib') db.init_app(app) register_blueprints(app, __name__, __path__) for error, func in ERROR_HANDLERS: app.errorhandler(error)(func) return app
import os from flask import Flask from flask_login import LoginManager from passlib.context import CryptContext from .settings import WmtSettings from .core import db from .blueprints import register_blueprints from .errors import ERROR_HANDLERS class User(object): def __init__(self, id): self._id = id def is_authenticated(self): return True def is_active(self): return True def is_anonymous(self): return False def get_id(self): return self._id def create_app(settings_override=None, register_security_blueprint=True, wmt_root_path=None): app = Flask(__name__, instance_relative_config=True) login_manager = LoginManager() login_manager.init_app(app) @login_manager.user_loader def load_user(userid): return User(userid) @app.before_first_request def create_database(): db.create_all() app.config.from_object(WmtSettings( os.path.join(wmt_root_path or app.root_path, 'db'))) #app.config.from_object(WmtSettings('/Users/huttone/git/wmt/db')) app.config.from_pyfile('settings.cfg', silent=True) app.config.from_object(settings_override) app.config['pw'] = CryptContext.from_string( app.config['CRYPT_INI_CONTENTS'], section='passlib') db.init_app(app) register_blueprints(app, __name__, __path__) for error, func in ERROR_HANDLERS: app.errorhandler(error)(func) return app
Use settings classes and set wmt_root_path.
Use settings classes and set wmt_root_path.
Python
mit
mcflugen/wmt-rest,mcflugen/wmt-rest
import os from flask import Flask from flask_login import LoginManager from passlib.context import CryptContext from .core import db from .blueprints import register_blueprints from .errors import ERROR_HANDLERS class User(object): def __init__(self, id): self._id = id def is_authenticated(self): return True def is_active(self): return True def is_anonymous(self): return False def get_id(self): return self._id def create_app(settings_override=None, register_security_blueprint=True): app = Flask(__name__, instance_relative_config=True) login_manager = LoginManager() login_manager.init_app(app) @login_manager.user_loader def load_user(userid): return User(userid) @app.before_first_request def create_database(): db.create_all() app.config.from_object('wmt.flask.settings') app.config.from_pyfile('settings.cfg', silent=True) app.config.from_object(settings_override) app.config['pw'] = CryptContext.from_string( app.config['CRYPT_INI_CONTENTS'], section='passlib') db.init_app(app) register_blueprints(app, __name__, __path__) for error, func in ERROR_HANDLERS: app.errorhandler(error)(func) return app Use settings classes and set wmt_root_path.
import os from flask import Flask from flask_login import LoginManager from passlib.context import CryptContext from .settings import WmtSettings from .core import db from .blueprints import register_blueprints from .errors import ERROR_HANDLERS class User(object): def __init__(self, id): self._id = id def is_authenticated(self): return True def is_active(self): return True def is_anonymous(self): return False def get_id(self): return self._id def create_app(settings_override=None, register_security_blueprint=True, wmt_root_path=None): app = Flask(__name__, instance_relative_config=True) login_manager = LoginManager() login_manager.init_app(app) @login_manager.user_loader def load_user(userid): return User(userid) @app.before_first_request def create_database(): db.create_all() app.config.from_object(WmtSettings( os.path.join(wmt_root_path or app.root_path, 'db'))) #app.config.from_object(WmtSettings('/Users/huttone/git/wmt/db')) app.config.from_pyfile('settings.cfg', silent=True) app.config.from_object(settings_override) app.config['pw'] = CryptContext.from_string( app.config['CRYPT_INI_CONTENTS'], section='passlib') db.init_app(app) register_blueprints(app, __name__, __path__) for error, func in ERROR_HANDLERS: app.errorhandler(error)(func) return app
<commit_before>import os from flask import Flask from flask_login import LoginManager from passlib.context import CryptContext from .core import db from .blueprints import register_blueprints from .errors import ERROR_HANDLERS class User(object): def __init__(self, id): self._id = id def is_authenticated(self): return True def is_active(self): return True def is_anonymous(self): return False def get_id(self): return self._id def create_app(settings_override=None, register_security_blueprint=True): app = Flask(__name__, instance_relative_config=True) login_manager = LoginManager() login_manager.init_app(app) @login_manager.user_loader def load_user(userid): return User(userid) @app.before_first_request def create_database(): db.create_all() app.config.from_object('wmt.flask.settings') app.config.from_pyfile('settings.cfg', silent=True) app.config.from_object(settings_override) app.config['pw'] = CryptContext.from_string( app.config['CRYPT_INI_CONTENTS'], section='passlib') db.init_app(app) register_blueprints(app, __name__, __path__) for error, func in ERROR_HANDLERS: app.errorhandler(error)(func) return app <commit_msg>Use settings classes and set wmt_root_path.<commit_after>
import os from flask import Flask from flask_login import LoginManager from passlib.context import CryptContext from .settings import WmtSettings from .core import db from .blueprints import register_blueprints from .errors import ERROR_HANDLERS class User(object): def __init__(self, id): self._id = id def is_authenticated(self): return True def is_active(self): return True def is_anonymous(self): return False def get_id(self): return self._id def create_app(settings_override=None, register_security_blueprint=True, wmt_root_path=None): app = Flask(__name__, instance_relative_config=True) login_manager = LoginManager() login_manager.init_app(app) @login_manager.user_loader def load_user(userid): return User(userid) @app.before_first_request def create_database(): db.create_all() app.config.from_object(WmtSettings( os.path.join(wmt_root_path or app.root_path, 'db'))) #app.config.from_object(WmtSettings('/Users/huttone/git/wmt/db')) app.config.from_pyfile('settings.cfg', silent=True) app.config.from_object(settings_override) app.config['pw'] = CryptContext.from_string( app.config['CRYPT_INI_CONTENTS'], section='passlib') db.init_app(app) register_blueprints(app, __name__, __path__) for error, func in ERROR_HANDLERS: app.errorhandler(error)(func) return app
import os from flask import Flask from flask_login import LoginManager from passlib.context import CryptContext from .core import db from .blueprints import register_blueprints from .errors import ERROR_HANDLERS class User(object): def __init__(self, id): self._id = id def is_authenticated(self): return True def is_active(self): return True def is_anonymous(self): return False def get_id(self): return self._id def create_app(settings_override=None, register_security_blueprint=True): app = Flask(__name__, instance_relative_config=True) login_manager = LoginManager() login_manager.init_app(app) @login_manager.user_loader def load_user(userid): return User(userid) @app.before_first_request def create_database(): db.create_all() app.config.from_object('wmt.flask.settings') app.config.from_pyfile('settings.cfg', silent=True) app.config.from_object(settings_override) app.config['pw'] = CryptContext.from_string( app.config['CRYPT_INI_CONTENTS'], section='passlib') db.init_app(app) register_blueprints(app, __name__, __path__) for error, func in ERROR_HANDLERS: app.errorhandler(error)(func) return app Use settings classes and set wmt_root_path.import os from flask import Flask from flask_login import LoginManager from passlib.context import CryptContext from .settings import WmtSettings from .core import db from .blueprints import register_blueprints from .errors import ERROR_HANDLERS class User(object): def __init__(self, id): self._id = id def is_authenticated(self): return True def is_active(self): return True def is_anonymous(self): return False def get_id(self): return self._id def create_app(settings_override=None, register_security_blueprint=True, wmt_root_path=None): app = Flask(__name__, instance_relative_config=True) login_manager = LoginManager() login_manager.init_app(app) @login_manager.user_loader def load_user(userid): return User(userid) @app.before_first_request def create_database(): db.create_all() app.config.from_object(WmtSettings( os.path.join(wmt_root_path or app.root_path, 'db'))) #app.config.from_object(WmtSettings('/Users/huttone/git/wmt/db')) app.config.from_pyfile('settings.cfg', silent=True) app.config.from_object(settings_override) app.config['pw'] = CryptContext.from_string( app.config['CRYPT_INI_CONTENTS'], section='passlib') db.init_app(app) register_blueprints(app, __name__, __path__) for error, func in ERROR_HANDLERS: app.errorhandler(error)(func) return app
<commit_before>import os from flask import Flask from flask_login import LoginManager from passlib.context import CryptContext from .core import db from .blueprints import register_blueprints from .errors import ERROR_HANDLERS class User(object): def __init__(self, id): self._id = id def is_authenticated(self): return True def is_active(self): return True def is_anonymous(self): return False def get_id(self): return self._id def create_app(settings_override=None, register_security_blueprint=True): app = Flask(__name__, instance_relative_config=True) login_manager = LoginManager() login_manager.init_app(app) @login_manager.user_loader def load_user(userid): return User(userid) @app.before_first_request def create_database(): db.create_all() app.config.from_object('wmt.flask.settings') app.config.from_pyfile('settings.cfg', silent=True) app.config.from_object(settings_override) app.config['pw'] = CryptContext.from_string( app.config['CRYPT_INI_CONTENTS'], section='passlib') db.init_app(app) register_blueprints(app, __name__, __path__) for error, func in ERROR_HANDLERS: app.errorhandler(error)(func) return app <commit_msg>Use settings classes and set wmt_root_path.<commit_after>import os from flask import Flask from flask_login import LoginManager from passlib.context import CryptContext from .settings import WmtSettings from .core import db from .blueprints import register_blueprints from .errors import ERROR_HANDLERS class User(object): def __init__(self, id): self._id = id def is_authenticated(self): return True def is_active(self): return True def is_anonymous(self): return False def get_id(self): return self._id def create_app(settings_override=None, register_security_blueprint=True, wmt_root_path=None): app = Flask(__name__, instance_relative_config=True) login_manager = LoginManager() login_manager.init_app(app) @login_manager.user_loader def load_user(userid): return User(userid) @app.before_first_request def create_database(): db.create_all() app.config.from_object(WmtSettings( os.path.join(wmt_root_path or app.root_path, 'db'))) #app.config.from_object(WmtSettings('/Users/huttone/git/wmt/db')) app.config.from_pyfile('settings.cfg', silent=True) app.config.from_object(settings_override) app.config['pw'] = CryptContext.from_string( app.config['CRYPT_INI_CONTENTS'], section='passlib') db.init_app(app) register_blueprints(app, __name__, __path__) for error, func in ERROR_HANDLERS: app.errorhandler(error)(func) return app
a479b445c18687827e7913e8b51abd5937848fe8
anybox/buildbot/openerp/build_utils/analyze_oerp_tests.py
anybox/buildbot/openerp/build_utils/analyze_oerp_tests.py
"""Analyse the tests log file given as argument. Print a report and return status code 1 if failures are detected """ import sys import re FAILURE_REGEXPS = { 'Failure in Python block': re.compile(r'WARNING:tests[.].*AssertionError'), 'Errors during x/yml tests': re.compile(r'ERROR:tests[.]'), 'Errors or failures during unittest2 tests': re.compile( r'at least one error occurred in a test'), 'Errors loading addons': re.compile(r'ERROR.*openerp: Failed to load'), 'Critical logs': re.compile(r'CRITICAL'), 'Error init db': re.compile(r'Failed to initialize database'), 'Tests failed to excute': re.compile( r'openerp.modules.loading: Tests failed to execute'), } test_log = open(sys.argv[1], 'r') failures = {} # label -> extracted line for line in test_log.readlines(): for label, regexp in FAILURE_REGEXPS.items(): if regexp.search(line): failures.setdefault(label, []).append(line) if not failures: print "No failure detected" sys.exit(0) total = 0 print 'FAILURES DETECTED' print for label, failed_lines in failures.items(): print label + ':' for line in failed_lines: print ' ' + line print total += len(failed_lines) print "Total: %d failures " % total sys.exit(1)
"""Analyse the tests log file given as argument. Print a report and return status code 1 if failures are detected """ import sys import re FAILURE_REGEXPS = { 'Failure in Python block': re.compile(r'WARNING:tests[.].*AssertionError'), 'Errors during x/yml tests': re.compile(r'ERROR:tests[.]'), 'Errors or failures during unittest2 tests': re.compile( r'at least one error occurred in a test'), 'Errors loading addons': re.compile(r'ERROR.*openerp: Failed to load'), 'Critical logs': re.compile(r'CRITICAL'), 'Error init db': re.compile(r'Failed to initialize database'), 'Tests failed to excute': re.compile( r'openerp.modules.loading: Tests failed to execute'), 'WARNING no field in model': re.compile('No such field(s) in model'), } test_log = open(sys.argv[1], 'r') failures = {} # label -> extracted line for line in test_log.readlines(): for label, regexp in FAILURE_REGEXPS.items(): if regexp.search(line): failures.setdefault(label, []).append(line) if not failures: print "No failure detected" sys.exit(0) total = 0 print 'FAILURES DETECTED' print for label, failed_lines in failures.items(): print label + ':' for line in failed_lines: print ' ' + line print total += len(failed_lines) print "Total: %d failures " % total sys.exit(1)
Add regex to know if one field is use in a model where this field is not declared
Add regex to know if one field is use in a model where this field is not declared
Python
agpl-3.0
anybox/anybox.buildbot.odoo
"""Analyse the tests log file given as argument. Print a report and return status code 1 if failures are detected """ import sys import re FAILURE_REGEXPS = { 'Failure in Python block': re.compile(r'WARNING:tests[.].*AssertionError'), 'Errors during x/yml tests': re.compile(r'ERROR:tests[.]'), 'Errors or failures during unittest2 tests': re.compile( r'at least one error occurred in a test'), 'Errors loading addons': re.compile(r'ERROR.*openerp: Failed to load'), 'Critical logs': re.compile(r'CRITICAL'), 'Error init db': re.compile(r'Failed to initialize database'), 'Tests failed to excute': re.compile( r'openerp.modules.loading: Tests failed to execute'), } test_log = open(sys.argv[1], 'r') failures = {} # label -> extracted line for line in test_log.readlines(): for label, regexp in FAILURE_REGEXPS.items(): if regexp.search(line): failures.setdefault(label, []).append(line) if not failures: print "No failure detected" sys.exit(0) total = 0 print 'FAILURES DETECTED' print for label, failed_lines in failures.items(): print label + ':' for line in failed_lines: print ' ' + line print total += len(failed_lines) print "Total: %d failures " % total sys.exit(1) Add regex to know if one field is use in a model where this field is not declared
"""Analyse the tests log file given as argument. Print a report and return status code 1 if failures are detected """ import sys import re FAILURE_REGEXPS = { 'Failure in Python block': re.compile(r'WARNING:tests[.].*AssertionError'), 'Errors during x/yml tests': re.compile(r'ERROR:tests[.]'), 'Errors or failures during unittest2 tests': re.compile( r'at least one error occurred in a test'), 'Errors loading addons': re.compile(r'ERROR.*openerp: Failed to load'), 'Critical logs': re.compile(r'CRITICAL'), 'Error init db': re.compile(r'Failed to initialize database'), 'Tests failed to excute': re.compile( r'openerp.modules.loading: Tests failed to execute'), 'WARNING no field in model': re.compile('No such field(s) in model'), } test_log = open(sys.argv[1], 'r') failures = {} # label -> extracted line for line in test_log.readlines(): for label, regexp in FAILURE_REGEXPS.items(): if regexp.search(line): failures.setdefault(label, []).append(line) if not failures: print "No failure detected" sys.exit(0) total = 0 print 'FAILURES DETECTED' print for label, failed_lines in failures.items(): print label + ':' for line in failed_lines: print ' ' + line print total += len(failed_lines) print "Total: %d failures " % total sys.exit(1)
<commit_before>"""Analyse the tests log file given as argument. Print a report and return status code 1 if failures are detected """ import sys import re FAILURE_REGEXPS = { 'Failure in Python block': re.compile(r'WARNING:tests[.].*AssertionError'), 'Errors during x/yml tests': re.compile(r'ERROR:tests[.]'), 'Errors or failures during unittest2 tests': re.compile( r'at least one error occurred in a test'), 'Errors loading addons': re.compile(r'ERROR.*openerp: Failed to load'), 'Critical logs': re.compile(r'CRITICAL'), 'Error init db': re.compile(r'Failed to initialize database'), 'Tests failed to excute': re.compile( r'openerp.modules.loading: Tests failed to execute'), } test_log = open(sys.argv[1], 'r') failures = {} # label -> extracted line for line in test_log.readlines(): for label, regexp in FAILURE_REGEXPS.items(): if regexp.search(line): failures.setdefault(label, []).append(line) if not failures: print "No failure detected" sys.exit(0) total = 0 print 'FAILURES DETECTED' print for label, failed_lines in failures.items(): print label + ':' for line in failed_lines: print ' ' + line print total += len(failed_lines) print "Total: %d failures " % total sys.exit(1) <commit_msg>Add regex to know if one field is use in a model where this field is not declared<commit_after>
"""Analyse the tests log file given as argument. Print a report and return status code 1 if failures are detected """ import sys import re FAILURE_REGEXPS = { 'Failure in Python block': re.compile(r'WARNING:tests[.].*AssertionError'), 'Errors during x/yml tests': re.compile(r'ERROR:tests[.]'), 'Errors or failures during unittest2 tests': re.compile( r'at least one error occurred in a test'), 'Errors loading addons': re.compile(r'ERROR.*openerp: Failed to load'), 'Critical logs': re.compile(r'CRITICAL'), 'Error init db': re.compile(r'Failed to initialize database'), 'Tests failed to excute': re.compile( r'openerp.modules.loading: Tests failed to execute'), 'WARNING no field in model': re.compile('No such field(s) in model'), } test_log = open(sys.argv[1], 'r') failures = {} # label -> extracted line for line in test_log.readlines(): for label, regexp in FAILURE_REGEXPS.items(): if regexp.search(line): failures.setdefault(label, []).append(line) if not failures: print "No failure detected" sys.exit(0) total = 0 print 'FAILURES DETECTED' print for label, failed_lines in failures.items(): print label + ':' for line in failed_lines: print ' ' + line print total += len(failed_lines) print "Total: %d failures " % total sys.exit(1)
"""Analyse the tests log file given as argument. Print a report and return status code 1 if failures are detected """ import sys import re FAILURE_REGEXPS = { 'Failure in Python block': re.compile(r'WARNING:tests[.].*AssertionError'), 'Errors during x/yml tests': re.compile(r'ERROR:tests[.]'), 'Errors or failures during unittest2 tests': re.compile( r'at least one error occurred in a test'), 'Errors loading addons': re.compile(r'ERROR.*openerp: Failed to load'), 'Critical logs': re.compile(r'CRITICAL'), 'Error init db': re.compile(r'Failed to initialize database'), 'Tests failed to excute': re.compile( r'openerp.modules.loading: Tests failed to execute'), } test_log = open(sys.argv[1], 'r') failures = {} # label -> extracted line for line in test_log.readlines(): for label, regexp in FAILURE_REGEXPS.items(): if regexp.search(line): failures.setdefault(label, []).append(line) if not failures: print "No failure detected" sys.exit(0) total = 0 print 'FAILURES DETECTED' print for label, failed_lines in failures.items(): print label + ':' for line in failed_lines: print ' ' + line print total += len(failed_lines) print "Total: %d failures " % total sys.exit(1) Add regex to know if one field is use in a model where this field is not declared"""Analyse the tests log file given as argument. Print a report and return status code 1 if failures are detected """ import sys import re FAILURE_REGEXPS = { 'Failure in Python block': re.compile(r'WARNING:tests[.].*AssertionError'), 'Errors during x/yml tests': re.compile(r'ERROR:tests[.]'), 'Errors or failures during unittest2 tests': re.compile( r'at least one error occurred in a test'), 'Errors loading addons': re.compile(r'ERROR.*openerp: Failed to load'), 'Critical logs': re.compile(r'CRITICAL'), 'Error init db': re.compile(r'Failed to initialize database'), 'Tests failed to excute': re.compile( r'openerp.modules.loading: Tests failed to execute'), 'WARNING no field in model': re.compile('No such field(s) in model'), } test_log = open(sys.argv[1], 'r') failures = {} # label -> extracted line for line in test_log.readlines(): for label, regexp in FAILURE_REGEXPS.items(): if regexp.search(line): failures.setdefault(label, []).append(line) if not failures: print "No failure detected" sys.exit(0) total = 0 print 'FAILURES DETECTED' print for label, failed_lines in failures.items(): print label + ':' for line in failed_lines: print ' ' + line print total += len(failed_lines) print "Total: %d failures " % total sys.exit(1)
<commit_before>"""Analyse the tests log file given as argument. Print a report and return status code 1 if failures are detected """ import sys import re FAILURE_REGEXPS = { 'Failure in Python block': re.compile(r'WARNING:tests[.].*AssertionError'), 'Errors during x/yml tests': re.compile(r'ERROR:tests[.]'), 'Errors or failures during unittest2 tests': re.compile( r'at least one error occurred in a test'), 'Errors loading addons': re.compile(r'ERROR.*openerp: Failed to load'), 'Critical logs': re.compile(r'CRITICAL'), 'Error init db': re.compile(r'Failed to initialize database'), 'Tests failed to excute': re.compile( r'openerp.modules.loading: Tests failed to execute'), } test_log = open(sys.argv[1], 'r') failures = {} # label -> extracted line for line in test_log.readlines(): for label, regexp in FAILURE_REGEXPS.items(): if regexp.search(line): failures.setdefault(label, []).append(line) if not failures: print "No failure detected" sys.exit(0) total = 0 print 'FAILURES DETECTED' print for label, failed_lines in failures.items(): print label + ':' for line in failed_lines: print ' ' + line print total += len(failed_lines) print "Total: %d failures " % total sys.exit(1) <commit_msg>Add regex to know if one field is use in a model where this field is not declared<commit_after>"""Analyse the tests log file given as argument. Print a report and return status code 1 if failures are detected """ import sys import re FAILURE_REGEXPS = { 'Failure in Python block': re.compile(r'WARNING:tests[.].*AssertionError'), 'Errors during x/yml tests': re.compile(r'ERROR:tests[.]'), 'Errors or failures during unittest2 tests': re.compile( r'at least one error occurred in a test'), 'Errors loading addons': re.compile(r'ERROR.*openerp: Failed to load'), 'Critical logs': re.compile(r'CRITICAL'), 'Error init db': re.compile(r'Failed to initialize database'), 'Tests failed to excute': re.compile( r'openerp.modules.loading: Tests failed to execute'), 'WARNING no field in model': re.compile('No such field(s) in model'), } test_log = open(sys.argv[1], 'r') failures = {} # label -> extracted line for line in test_log.readlines(): for label, regexp in FAILURE_REGEXPS.items(): if regexp.search(line): failures.setdefault(label, []).append(line) if not failures: print "No failure detected" sys.exit(0) total = 0 print 'FAILURES DETECTED' print for label, failed_lines in failures.items(): print label + ':' for line in failed_lines: print ' ' + line print total += len(failed_lines) print "Total: %d failures " % total sys.exit(1)
24f0c0d311886571cc0c5f8badca026a6c534a52
dask/array/__init__.py
dask/array/__init__.py
from __future__ import absolute_import, division, print_function from ..utils import ignoring from .core import (Array, stack, concatenate, tensordot, transpose, from_array, choose, where, coarsen, broadcast_to, constant, fromfunction, compute, unique, store) from .core import (arccos, arcsin, arctan, arctanh, arccosh, arcsinh, arctan2, ceil, copysign, cos, cosh, degrees, exp, expm1, fabs, floor, fmod, frexp, hypot, isinf, isnan, ldexp, log, log10, log1p, modf, radians, sin, sinh, sqrt, tan, tanh, trunc, around, isnull, notnull, isclose) from .reductions import (sum, prod, mean, std, var, any, all, min, max, vnorm, argmin, argmax, nansum, nanmean, nanstd, nanvar, nanmin, nanmax, nanargmin, nanargmax) from .percentile import percentile with ignoring(ImportError): from .reductions import nanprod from . import random, linalg, ghost, creation from .wrap import ones, zeros, empty from .reblock import reblock from ..context import set_options from .optimization import optimize
from __future__ import absolute_import, division, print_function from ..utils import ignoring from .core import (Array, stack, concatenate, tensordot, transpose, from_array, choose, where, coarsen, broadcast_to, constant, fromfunction, compute, unique, store) from .core import (arccos, arcsin, arctan, arctanh, arccosh, arcsinh, arctan2, ceil, copysign, cos, cosh, degrees, exp, expm1, fabs, floor, fmod, frexp, hypot, isinf, isnan, ldexp, log, log10, log1p, modf, radians, sin, sinh, sqrt, tan, tanh, trunc, around, isnull, notnull, isclose) from .reductions import (sum, prod, mean, std, var, any, all, min, max, vnorm, argmin, argmax, nansum, nanmean, nanstd, nanvar, nanmin, nanmax, nanargmin, nanargmax) from .percentile import percentile with ignoring(ImportError): from .reductions import nanprod from . import random, linalg, ghost from .wrap import ones, zeros, empty from .reblock import reblock from ..context import set_options from .optimization import optimize from .creation import arange, linspace
Move dask.array.creation.* into dask.array.* namespace
Move dask.array.creation.* into dask.array.* namespace
Python
bsd-3-clause
mraspaud/dask,marianotepper/dask,minrk/dask,freeman-lab/dask,pombredanne/dask,ContinuumIO/dask,hainm/dask,ssanderson/dask,clarkfitzg/dask,jakirkham/dask,jakirkham/dask,PhE/dask,cowlicks/dask,ContinuumIO/dask,mikegraham/dask,cpcloud/dask,pombredanne/dask,hainm/dask,marianotepper/dask,simudream/dask,jcrist/dask,wiso/dask,blaze/dask,vikhyat/dask,blaze/dask,esc/dask,PhE/dask,minrk/dask,vikhyat/dask,wiso/dask,dask/dask,dask/dask,jayhetee/dask,freeman-lab/dask,mrocklin/dask,jcrist/dask,esc/dask,clarkfitzg/dask,ssanderson/dask,mrocklin/dask,simudream/dask,mraspaud/dask,chrisbarber/dask,jayhetee/dask,gameduell/dask
from __future__ import absolute_import, division, print_function from ..utils import ignoring from .core import (Array, stack, concatenate, tensordot, transpose, from_array, choose, where, coarsen, broadcast_to, constant, fromfunction, compute, unique, store) from .core import (arccos, arcsin, arctan, arctanh, arccosh, arcsinh, arctan2, ceil, copysign, cos, cosh, degrees, exp, expm1, fabs, floor, fmod, frexp, hypot, isinf, isnan, ldexp, log, log10, log1p, modf, radians, sin, sinh, sqrt, tan, tanh, trunc, around, isnull, notnull, isclose) from .reductions import (sum, prod, mean, std, var, any, all, min, max, vnorm, argmin, argmax, nansum, nanmean, nanstd, nanvar, nanmin, nanmax, nanargmin, nanargmax) from .percentile import percentile with ignoring(ImportError): from .reductions import nanprod from . import random, linalg, ghost, creation from .wrap import ones, zeros, empty from .reblock import reblock from ..context import set_options from .optimization import optimize Move dask.array.creation.* into dask.array.* namespace
from __future__ import absolute_import, division, print_function from ..utils import ignoring from .core import (Array, stack, concatenate, tensordot, transpose, from_array, choose, where, coarsen, broadcast_to, constant, fromfunction, compute, unique, store) from .core import (arccos, arcsin, arctan, arctanh, arccosh, arcsinh, arctan2, ceil, copysign, cos, cosh, degrees, exp, expm1, fabs, floor, fmod, frexp, hypot, isinf, isnan, ldexp, log, log10, log1p, modf, radians, sin, sinh, sqrt, tan, tanh, trunc, around, isnull, notnull, isclose) from .reductions import (sum, prod, mean, std, var, any, all, min, max, vnorm, argmin, argmax, nansum, nanmean, nanstd, nanvar, nanmin, nanmax, nanargmin, nanargmax) from .percentile import percentile with ignoring(ImportError): from .reductions import nanprod from . import random, linalg, ghost from .wrap import ones, zeros, empty from .reblock import reblock from ..context import set_options from .optimization import optimize from .creation import arange, linspace
<commit_before>from __future__ import absolute_import, division, print_function from ..utils import ignoring from .core import (Array, stack, concatenate, tensordot, transpose, from_array, choose, where, coarsen, broadcast_to, constant, fromfunction, compute, unique, store) from .core import (arccos, arcsin, arctan, arctanh, arccosh, arcsinh, arctan2, ceil, copysign, cos, cosh, degrees, exp, expm1, fabs, floor, fmod, frexp, hypot, isinf, isnan, ldexp, log, log10, log1p, modf, radians, sin, sinh, sqrt, tan, tanh, trunc, around, isnull, notnull, isclose) from .reductions import (sum, prod, mean, std, var, any, all, min, max, vnorm, argmin, argmax, nansum, nanmean, nanstd, nanvar, nanmin, nanmax, nanargmin, nanargmax) from .percentile import percentile with ignoring(ImportError): from .reductions import nanprod from . import random, linalg, ghost, creation from .wrap import ones, zeros, empty from .reblock import reblock from ..context import set_options from .optimization import optimize <commit_msg>Move dask.array.creation.* into dask.array.* namespace<commit_after>
from __future__ import absolute_import, division, print_function from ..utils import ignoring from .core import (Array, stack, concatenate, tensordot, transpose, from_array, choose, where, coarsen, broadcast_to, constant, fromfunction, compute, unique, store) from .core import (arccos, arcsin, arctan, arctanh, arccosh, arcsinh, arctan2, ceil, copysign, cos, cosh, degrees, exp, expm1, fabs, floor, fmod, frexp, hypot, isinf, isnan, ldexp, log, log10, log1p, modf, radians, sin, sinh, sqrt, tan, tanh, trunc, around, isnull, notnull, isclose) from .reductions import (sum, prod, mean, std, var, any, all, min, max, vnorm, argmin, argmax, nansum, nanmean, nanstd, nanvar, nanmin, nanmax, nanargmin, nanargmax) from .percentile import percentile with ignoring(ImportError): from .reductions import nanprod from . import random, linalg, ghost from .wrap import ones, zeros, empty from .reblock import reblock from ..context import set_options from .optimization import optimize from .creation import arange, linspace
from __future__ import absolute_import, division, print_function from ..utils import ignoring from .core import (Array, stack, concatenate, tensordot, transpose, from_array, choose, where, coarsen, broadcast_to, constant, fromfunction, compute, unique, store) from .core import (arccos, arcsin, arctan, arctanh, arccosh, arcsinh, arctan2, ceil, copysign, cos, cosh, degrees, exp, expm1, fabs, floor, fmod, frexp, hypot, isinf, isnan, ldexp, log, log10, log1p, modf, radians, sin, sinh, sqrt, tan, tanh, trunc, around, isnull, notnull, isclose) from .reductions import (sum, prod, mean, std, var, any, all, min, max, vnorm, argmin, argmax, nansum, nanmean, nanstd, nanvar, nanmin, nanmax, nanargmin, nanargmax) from .percentile import percentile with ignoring(ImportError): from .reductions import nanprod from . import random, linalg, ghost, creation from .wrap import ones, zeros, empty from .reblock import reblock from ..context import set_options from .optimization import optimize Move dask.array.creation.* into dask.array.* namespacefrom __future__ import absolute_import, division, print_function from ..utils import ignoring from .core import (Array, stack, concatenate, tensordot, transpose, from_array, choose, where, coarsen, broadcast_to, constant, fromfunction, compute, unique, store) from .core import (arccos, arcsin, arctan, arctanh, arccosh, arcsinh, arctan2, ceil, copysign, cos, cosh, degrees, exp, expm1, fabs, floor, fmod, frexp, hypot, isinf, isnan, ldexp, log, log10, log1p, modf, radians, sin, sinh, sqrt, tan, tanh, trunc, around, isnull, notnull, isclose) from .reductions import (sum, prod, mean, std, var, any, all, min, max, vnorm, argmin, argmax, nansum, nanmean, nanstd, nanvar, nanmin, nanmax, nanargmin, nanargmax) from .percentile import percentile with ignoring(ImportError): from .reductions import nanprod from . import random, linalg, ghost from .wrap import ones, zeros, empty from .reblock import reblock from ..context import set_options from .optimization import optimize from .creation import arange, linspace
<commit_before>from __future__ import absolute_import, division, print_function from ..utils import ignoring from .core import (Array, stack, concatenate, tensordot, transpose, from_array, choose, where, coarsen, broadcast_to, constant, fromfunction, compute, unique, store) from .core import (arccos, arcsin, arctan, arctanh, arccosh, arcsinh, arctan2, ceil, copysign, cos, cosh, degrees, exp, expm1, fabs, floor, fmod, frexp, hypot, isinf, isnan, ldexp, log, log10, log1p, modf, radians, sin, sinh, sqrt, tan, tanh, trunc, around, isnull, notnull, isclose) from .reductions import (sum, prod, mean, std, var, any, all, min, max, vnorm, argmin, argmax, nansum, nanmean, nanstd, nanvar, nanmin, nanmax, nanargmin, nanargmax) from .percentile import percentile with ignoring(ImportError): from .reductions import nanprod from . import random, linalg, ghost, creation from .wrap import ones, zeros, empty from .reblock import reblock from ..context import set_options from .optimization import optimize <commit_msg>Move dask.array.creation.* into dask.array.* namespace<commit_after>from __future__ import absolute_import, division, print_function from ..utils import ignoring from .core import (Array, stack, concatenate, tensordot, transpose, from_array, choose, where, coarsen, broadcast_to, constant, fromfunction, compute, unique, store) from .core import (arccos, arcsin, arctan, arctanh, arccosh, arcsinh, arctan2, ceil, copysign, cos, cosh, degrees, exp, expm1, fabs, floor, fmod, frexp, hypot, isinf, isnan, ldexp, log, log10, log1p, modf, radians, sin, sinh, sqrt, tan, tanh, trunc, around, isnull, notnull, isclose) from .reductions import (sum, prod, mean, std, var, any, all, min, max, vnorm, argmin, argmax, nansum, nanmean, nanstd, nanvar, nanmin, nanmax, nanargmin, nanargmax) from .percentile import percentile with ignoring(ImportError): from .reductions import nanprod from . import random, linalg, ghost from .wrap import ones, zeros, empty from .reblock import reblock from ..context import set_options from .optimization import optimize from .creation import arange, linspace
36550f71f4161b0b5c7af872b78dd1e7d96b788a
scripts/patches/dynamodb.py
scripts/patches/dynamodb.py
patches = [ # duplicate GlobalSecondaryIndex { "op": "move", "from": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalSecondaryIndex", "path": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalTableGlobalSecondaryIndex", }, { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::GlobalTable/Properties/GlobalSecondaryIndexes/ItemType", "value": "GlobalTableGlobalSecondaryIndex", }, # duplicate SSESpecification { "op": "move", "from": "/PropertyTypes/AWS::DynamoDB::GlobalTable.SSESpecification", "path": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalTableSSESpecification", }, { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::GlobalTable/Properties/SSESpecification/Type", "value": "GlobalTableSSESpecification", }, ]
patches = [ # duplicate GlobalSecondaryIndex { "op": "move", "from": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalSecondaryIndex", "path": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalTableGlobalSecondaryIndex", }, { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::GlobalTable/Properties/GlobalSecondaryIndexes/ItemType", "value": "GlobalTableGlobalSecondaryIndex", }, # duplicate SSESpecification { "op": "move", "from": "/PropertyTypes/AWS::DynamoDB::GlobalTable.SSESpecification", "path": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalTableSSESpecification", }, { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::GlobalTable/Properties/SSESpecification/Type", "value": "GlobalTableSSESpecification", }, # Fix issue in spec 82.0.0 that changed Type to Json { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::Table/Properties/KeySchema/Type", "value": "List", }, { "op": "add", "path": "/ResourceTypes/AWS::DynamoDB::Table/Properties/KeySchema/ItemType", "value": "KeySchema", }, ]
Fix issue in spec 82.0.0 with DynamoDB KeySchema Type
Fix issue in spec 82.0.0 with DynamoDB KeySchema Type
Python
bsd-2-clause
cloudtools/troposphere,cloudtools/troposphere
patches = [ # duplicate GlobalSecondaryIndex { "op": "move", "from": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalSecondaryIndex", "path": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalTableGlobalSecondaryIndex", }, { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::GlobalTable/Properties/GlobalSecondaryIndexes/ItemType", "value": "GlobalTableGlobalSecondaryIndex", }, # duplicate SSESpecification { "op": "move", "from": "/PropertyTypes/AWS::DynamoDB::GlobalTable.SSESpecification", "path": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalTableSSESpecification", }, { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::GlobalTable/Properties/SSESpecification/Type", "value": "GlobalTableSSESpecification", }, ] Fix issue in spec 82.0.0 with DynamoDB KeySchema Type
patches = [ # duplicate GlobalSecondaryIndex { "op": "move", "from": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalSecondaryIndex", "path": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalTableGlobalSecondaryIndex", }, { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::GlobalTable/Properties/GlobalSecondaryIndexes/ItemType", "value": "GlobalTableGlobalSecondaryIndex", }, # duplicate SSESpecification { "op": "move", "from": "/PropertyTypes/AWS::DynamoDB::GlobalTable.SSESpecification", "path": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalTableSSESpecification", }, { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::GlobalTable/Properties/SSESpecification/Type", "value": "GlobalTableSSESpecification", }, # Fix issue in spec 82.0.0 that changed Type to Json { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::Table/Properties/KeySchema/Type", "value": "List", }, { "op": "add", "path": "/ResourceTypes/AWS::DynamoDB::Table/Properties/KeySchema/ItemType", "value": "KeySchema", }, ]
<commit_before>patches = [ # duplicate GlobalSecondaryIndex { "op": "move", "from": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalSecondaryIndex", "path": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalTableGlobalSecondaryIndex", }, { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::GlobalTable/Properties/GlobalSecondaryIndexes/ItemType", "value": "GlobalTableGlobalSecondaryIndex", }, # duplicate SSESpecification { "op": "move", "from": "/PropertyTypes/AWS::DynamoDB::GlobalTable.SSESpecification", "path": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalTableSSESpecification", }, { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::GlobalTable/Properties/SSESpecification/Type", "value": "GlobalTableSSESpecification", }, ] <commit_msg>Fix issue in spec 82.0.0 with DynamoDB KeySchema Type<commit_after>
patches = [ # duplicate GlobalSecondaryIndex { "op": "move", "from": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalSecondaryIndex", "path": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalTableGlobalSecondaryIndex", }, { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::GlobalTable/Properties/GlobalSecondaryIndexes/ItemType", "value": "GlobalTableGlobalSecondaryIndex", }, # duplicate SSESpecification { "op": "move", "from": "/PropertyTypes/AWS::DynamoDB::GlobalTable.SSESpecification", "path": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalTableSSESpecification", }, { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::GlobalTable/Properties/SSESpecification/Type", "value": "GlobalTableSSESpecification", }, # Fix issue in spec 82.0.0 that changed Type to Json { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::Table/Properties/KeySchema/Type", "value": "List", }, { "op": "add", "path": "/ResourceTypes/AWS::DynamoDB::Table/Properties/KeySchema/ItemType", "value": "KeySchema", }, ]
patches = [ # duplicate GlobalSecondaryIndex { "op": "move", "from": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalSecondaryIndex", "path": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalTableGlobalSecondaryIndex", }, { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::GlobalTable/Properties/GlobalSecondaryIndexes/ItemType", "value": "GlobalTableGlobalSecondaryIndex", }, # duplicate SSESpecification { "op": "move", "from": "/PropertyTypes/AWS::DynamoDB::GlobalTable.SSESpecification", "path": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalTableSSESpecification", }, { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::GlobalTable/Properties/SSESpecification/Type", "value": "GlobalTableSSESpecification", }, ] Fix issue in spec 82.0.0 with DynamoDB KeySchema Typepatches = [ # duplicate GlobalSecondaryIndex { "op": "move", "from": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalSecondaryIndex", "path": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalTableGlobalSecondaryIndex", }, { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::GlobalTable/Properties/GlobalSecondaryIndexes/ItemType", "value": "GlobalTableGlobalSecondaryIndex", }, # duplicate SSESpecification { "op": "move", "from": "/PropertyTypes/AWS::DynamoDB::GlobalTable.SSESpecification", "path": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalTableSSESpecification", }, { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::GlobalTable/Properties/SSESpecification/Type", "value": "GlobalTableSSESpecification", }, # Fix issue in spec 82.0.0 that changed Type to Json { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::Table/Properties/KeySchema/Type", "value": "List", }, { "op": "add", "path": "/ResourceTypes/AWS::DynamoDB::Table/Properties/KeySchema/ItemType", "value": "KeySchema", }, ]
<commit_before>patches = [ # duplicate GlobalSecondaryIndex { "op": "move", "from": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalSecondaryIndex", "path": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalTableGlobalSecondaryIndex", }, { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::GlobalTable/Properties/GlobalSecondaryIndexes/ItemType", "value": "GlobalTableGlobalSecondaryIndex", }, # duplicate SSESpecification { "op": "move", "from": "/PropertyTypes/AWS::DynamoDB::GlobalTable.SSESpecification", "path": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalTableSSESpecification", }, { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::GlobalTable/Properties/SSESpecification/Type", "value": "GlobalTableSSESpecification", }, ] <commit_msg>Fix issue in spec 82.0.0 with DynamoDB KeySchema Type<commit_after>patches = [ # duplicate GlobalSecondaryIndex { "op": "move", "from": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalSecondaryIndex", "path": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalTableGlobalSecondaryIndex", }, { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::GlobalTable/Properties/GlobalSecondaryIndexes/ItemType", "value": "GlobalTableGlobalSecondaryIndex", }, # duplicate SSESpecification { "op": "move", "from": "/PropertyTypes/AWS::DynamoDB::GlobalTable.SSESpecification", "path": "/PropertyTypes/AWS::DynamoDB::GlobalTable.GlobalTableSSESpecification", }, { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::GlobalTable/Properties/SSESpecification/Type", "value": "GlobalTableSSESpecification", }, # Fix issue in spec 82.0.0 that changed Type to Json { "op": "replace", "path": "/ResourceTypes/AWS::DynamoDB::Table/Properties/KeySchema/Type", "value": "List", }, { "op": "add", "path": "/ResourceTypes/AWS::DynamoDB::Table/Properties/KeySchema/ItemType", "value": "KeySchema", }, ]
8785f602900ab3bf3e297ee8f90ecf47c059cdde
sphinxcontrib/openstreetmap.py
sphinxcontrib/openstreetmap.py
# -*- coding: utf-8 -*- """ sphinxcontrib.openstreetmap =========================== Embed OpenStreetMap on your documentation. :copyright: Copyright 2015 HAYASHI Kentaro <kenhys@gmail.com> :license: BSD, see LICENSE for details. """ from docutils import nodes from docutils.parsers.rst import directives from sphinx.util.compat import Directive class openstreetmap(nodes.General, nodes.Element): pass class OpenStreetMapDirective(Directive): """Directive for embedding OpenStreetMap""" has_content = False option_spec = { 'id': directives.unchanged, 'label': directives.unchanged } def run(self): node = openstreetmap() return [node] def visit_openstreetmap_node(self, node): self.body.append("<div id='openstreetmap' style='color:red'>OpenStreetMap directive</div>") def depart_openstreetmap_node(self, node): pass def setup(app): app.add_node(openstreetmap, html=(visit_openstreetmap_node, depart_openstreetmap_node)) app.add_directive('openstreetmap', OpenStreetMapDirective)
# -*- coding: utf-8 -*- """ sphinxcontrib.openstreetmap =========================== Embed OpenStreetMap on your documentation. :copyright: Copyright 2015 HAYASHI Kentaro <kenhys@gmail.com> :license: BSD, see LICENSE for details. """ from docutils import nodes from docutils.parsers.rst import directives from sphinx.util.compat import Directive class openstreetmap(nodes.General, nodes.Element): pass class OpenStreetMapDirective(Directive): """Directive for embedding OpenStreetMap""" has_content = False option_spec = { 'id': directives.unchanged, 'label': directives.unchanged, 'marker': directives.unchanged, } def run(self): node = openstreetmap() return [node] def visit_openstreetmap_node(self, node): self.body.append("<div id='openstreetmap' style='color:red'>OpenStreetMap directive</div>") def depart_openstreetmap_node(self, node): pass def setup(app): app.add_node(openstreetmap, html=(visit_openstreetmap_node, depart_openstreetmap_node)) app.add_directive('openstreetmap', OpenStreetMapDirective)
Add marker as required parameter
Add marker as required parameter
Python
bsd-2-clause
kenhys/sphinxcontrib-openstreetmap,kenhys/sphinxcontrib-openstreetmap
# -*- coding: utf-8 -*- """ sphinxcontrib.openstreetmap =========================== Embed OpenStreetMap on your documentation. :copyright: Copyright 2015 HAYASHI Kentaro <kenhys@gmail.com> :license: BSD, see LICENSE for details. """ from docutils import nodes from docutils.parsers.rst import directives from sphinx.util.compat import Directive class openstreetmap(nodes.General, nodes.Element): pass class OpenStreetMapDirective(Directive): """Directive for embedding OpenStreetMap""" has_content = False option_spec = { 'id': directives.unchanged, 'label': directives.unchanged } def run(self): node = openstreetmap() return [node] def visit_openstreetmap_node(self, node): self.body.append("<div id='openstreetmap' style='color:red'>OpenStreetMap directive</div>") def depart_openstreetmap_node(self, node): pass def setup(app): app.add_node(openstreetmap, html=(visit_openstreetmap_node, depart_openstreetmap_node)) app.add_directive('openstreetmap', OpenStreetMapDirective) Add marker as required parameter
# -*- coding: utf-8 -*- """ sphinxcontrib.openstreetmap =========================== Embed OpenStreetMap on your documentation. :copyright: Copyright 2015 HAYASHI Kentaro <kenhys@gmail.com> :license: BSD, see LICENSE for details. """ from docutils import nodes from docutils.parsers.rst import directives from sphinx.util.compat import Directive class openstreetmap(nodes.General, nodes.Element): pass class OpenStreetMapDirective(Directive): """Directive for embedding OpenStreetMap""" has_content = False option_spec = { 'id': directives.unchanged, 'label': directives.unchanged, 'marker': directives.unchanged, } def run(self): node = openstreetmap() return [node] def visit_openstreetmap_node(self, node): self.body.append("<div id='openstreetmap' style='color:red'>OpenStreetMap directive</div>") def depart_openstreetmap_node(self, node): pass def setup(app): app.add_node(openstreetmap, html=(visit_openstreetmap_node, depart_openstreetmap_node)) app.add_directive('openstreetmap', OpenStreetMapDirective)
<commit_before># -*- coding: utf-8 -*- """ sphinxcontrib.openstreetmap =========================== Embed OpenStreetMap on your documentation. :copyright: Copyright 2015 HAYASHI Kentaro <kenhys@gmail.com> :license: BSD, see LICENSE for details. """ from docutils import nodes from docutils.parsers.rst import directives from sphinx.util.compat import Directive class openstreetmap(nodes.General, nodes.Element): pass class OpenStreetMapDirective(Directive): """Directive for embedding OpenStreetMap""" has_content = False option_spec = { 'id': directives.unchanged, 'label': directives.unchanged } def run(self): node = openstreetmap() return [node] def visit_openstreetmap_node(self, node): self.body.append("<div id='openstreetmap' style='color:red'>OpenStreetMap directive</div>") def depart_openstreetmap_node(self, node): pass def setup(app): app.add_node(openstreetmap, html=(visit_openstreetmap_node, depart_openstreetmap_node)) app.add_directive('openstreetmap', OpenStreetMapDirective) <commit_msg>Add marker as required parameter<commit_after>
# -*- coding: utf-8 -*- """ sphinxcontrib.openstreetmap =========================== Embed OpenStreetMap on your documentation. :copyright: Copyright 2015 HAYASHI Kentaro <kenhys@gmail.com> :license: BSD, see LICENSE for details. """ from docutils import nodes from docutils.parsers.rst import directives from sphinx.util.compat import Directive class openstreetmap(nodes.General, nodes.Element): pass class OpenStreetMapDirective(Directive): """Directive for embedding OpenStreetMap""" has_content = False option_spec = { 'id': directives.unchanged, 'label': directives.unchanged, 'marker': directives.unchanged, } def run(self): node = openstreetmap() return [node] def visit_openstreetmap_node(self, node): self.body.append("<div id='openstreetmap' style='color:red'>OpenStreetMap directive</div>") def depart_openstreetmap_node(self, node): pass def setup(app): app.add_node(openstreetmap, html=(visit_openstreetmap_node, depart_openstreetmap_node)) app.add_directive('openstreetmap', OpenStreetMapDirective)
# -*- coding: utf-8 -*- """ sphinxcontrib.openstreetmap =========================== Embed OpenStreetMap on your documentation. :copyright: Copyright 2015 HAYASHI Kentaro <kenhys@gmail.com> :license: BSD, see LICENSE for details. """ from docutils import nodes from docutils.parsers.rst import directives from sphinx.util.compat import Directive class openstreetmap(nodes.General, nodes.Element): pass class OpenStreetMapDirective(Directive): """Directive for embedding OpenStreetMap""" has_content = False option_spec = { 'id': directives.unchanged, 'label': directives.unchanged } def run(self): node = openstreetmap() return [node] def visit_openstreetmap_node(self, node): self.body.append("<div id='openstreetmap' style='color:red'>OpenStreetMap directive</div>") def depart_openstreetmap_node(self, node): pass def setup(app): app.add_node(openstreetmap, html=(visit_openstreetmap_node, depart_openstreetmap_node)) app.add_directive('openstreetmap', OpenStreetMapDirective) Add marker as required parameter# -*- coding: utf-8 -*- """ sphinxcontrib.openstreetmap =========================== Embed OpenStreetMap on your documentation. :copyright: Copyright 2015 HAYASHI Kentaro <kenhys@gmail.com> :license: BSD, see LICENSE for details. """ from docutils import nodes from docutils.parsers.rst import directives from sphinx.util.compat import Directive class openstreetmap(nodes.General, nodes.Element): pass class OpenStreetMapDirective(Directive): """Directive for embedding OpenStreetMap""" has_content = False option_spec = { 'id': directives.unchanged, 'label': directives.unchanged, 'marker': directives.unchanged, } def run(self): node = openstreetmap() return [node] def visit_openstreetmap_node(self, node): self.body.append("<div id='openstreetmap' style='color:red'>OpenStreetMap directive</div>") def depart_openstreetmap_node(self, node): pass def setup(app): app.add_node(openstreetmap, html=(visit_openstreetmap_node, depart_openstreetmap_node)) app.add_directive('openstreetmap', OpenStreetMapDirective)
<commit_before># -*- coding: utf-8 -*- """ sphinxcontrib.openstreetmap =========================== Embed OpenStreetMap on your documentation. :copyright: Copyright 2015 HAYASHI Kentaro <kenhys@gmail.com> :license: BSD, see LICENSE for details. """ from docutils import nodes from docutils.parsers.rst import directives from sphinx.util.compat import Directive class openstreetmap(nodes.General, nodes.Element): pass class OpenStreetMapDirective(Directive): """Directive for embedding OpenStreetMap""" has_content = False option_spec = { 'id': directives.unchanged, 'label': directives.unchanged } def run(self): node = openstreetmap() return [node] def visit_openstreetmap_node(self, node): self.body.append("<div id='openstreetmap' style='color:red'>OpenStreetMap directive</div>") def depart_openstreetmap_node(self, node): pass def setup(app): app.add_node(openstreetmap, html=(visit_openstreetmap_node, depart_openstreetmap_node)) app.add_directive('openstreetmap', OpenStreetMapDirective) <commit_msg>Add marker as required parameter<commit_after># -*- coding: utf-8 -*- """ sphinxcontrib.openstreetmap =========================== Embed OpenStreetMap on your documentation. :copyright: Copyright 2015 HAYASHI Kentaro <kenhys@gmail.com> :license: BSD, see LICENSE for details. """ from docutils import nodes from docutils.parsers.rst import directives from sphinx.util.compat import Directive class openstreetmap(nodes.General, nodes.Element): pass class OpenStreetMapDirective(Directive): """Directive for embedding OpenStreetMap""" has_content = False option_spec = { 'id': directives.unchanged, 'label': directives.unchanged, 'marker': directives.unchanged, } def run(self): node = openstreetmap() return [node] def visit_openstreetmap_node(self, node): self.body.append("<div id='openstreetmap' style='color:red'>OpenStreetMap directive</div>") def depart_openstreetmap_node(self, node): pass def setup(app): app.add_node(openstreetmap, html=(visit_openstreetmap_node, depart_openstreetmap_node)) app.add_directive('openstreetmap', OpenStreetMapDirective)
2ba3dd9bafddd4dfd1ab712c59c6efadd58b1f46
skimage/feature/__init__.py
skimage/feature/__init__.py
from ._daisy import daisy from ._hog import hog from .texture import greycomatrix, greycoprops, local_binary_pattern from .peak import peak_local_max from .corner import (corner_kitchen_rosenfeld, corner_harris, corner_shi_tomasi, corner_foerstner, corner_subpix, corner_peaks) from .corner_cy import corner_moravec from .template import match_template
from ._daisy import daisy from ._hog import hog from .texture import greycomatrix, greycoprops, local_binary_pattern from .peak import peak_local_max from .corner import (corner_kitchen_rosenfeld, corner_harris, corner_shi_tomasi, corner_foerstner, corner_subpix, corner_peaks) from .corner_cy import corner_moravec from .template import match_template __all__ = ['daisy', 'hog', 'greycomatrix', 'greycoprops', 'local_binary_pattern', 'peak_local_max', 'corner_kitchen_rosenfeld', 'corner_harris', 'corner_shi_tomasi', 'corner_foerstner', 'corner_subpix', 'corner_peaks', 'corner_moravec', 'match_template']
Add __all__ to feature package
Add __all__ to feature package
Python
bsd-3-clause
WarrenWeckesser/scikits-image,michaelpacer/scikit-image,chintak/scikit-image,Midafi/scikit-image,bsipocz/scikit-image,almarklein/scikit-image,paalge/scikit-image,vighneshbirodkar/scikit-image,GaZ3ll3/scikit-image,ofgulban/scikit-image,GaZ3ll3/scikit-image,jwiggins/scikit-image,ClinicalGraphics/scikit-image,SamHames/scikit-image,youprofit/scikit-image,paalge/scikit-image,bsipocz/scikit-image,emon10005/scikit-image,almarklein/scikit-image,blink1073/scikit-image,chintak/scikit-image,warmspringwinds/scikit-image,chriscrosscutler/scikit-image,michaelaye/scikit-image,bennlich/scikit-image,keflavich/scikit-image,SamHames/scikit-image,chintak/scikit-image,michaelpacer/scikit-image,ajaybhat/scikit-image,blink1073/scikit-image,ofgulban/scikit-image,newville/scikit-image,newville/scikit-image,Hiyorimi/scikit-image,ClinicalGraphics/scikit-image,rjeli/scikit-image,dpshelio/scikit-image,oew1v07/scikit-image,michaelaye/scikit-image,jwiggins/scikit-image,chriscrosscutler/scikit-image,vighneshbirodkar/scikit-image,oew1v07/scikit-image,juliusbierk/scikit-image,bennlich/scikit-image,pratapvardhan/scikit-image,dpshelio/scikit-image,SamHames/scikit-image,rjeli/scikit-image,paalge/scikit-image,robintw/scikit-image,warmspringwinds/scikit-image,Britefury/scikit-image,WarrenWeckesser/scikits-image,juliusbierk/scikit-image,youprofit/scikit-image,Britefury/scikit-image,almarklein/scikit-image,Hiyorimi/scikit-image,keflavich/scikit-image,almarklein/scikit-image,SamHames/scikit-image,pratapvardhan/scikit-image,robintw/scikit-image,ofgulban/scikit-image,vighneshbirodkar/scikit-image,Midafi/scikit-image,rjeli/scikit-image,chintak/scikit-image,ajaybhat/scikit-image,emon10005/scikit-image
from ._daisy import daisy from ._hog import hog from .texture import greycomatrix, greycoprops, local_binary_pattern from .peak import peak_local_max from .corner import (corner_kitchen_rosenfeld, corner_harris, corner_shi_tomasi, corner_foerstner, corner_subpix, corner_peaks) from .corner_cy import corner_moravec from .template import match_template Add __all__ to feature package
from ._daisy import daisy from ._hog import hog from .texture import greycomatrix, greycoprops, local_binary_pattern from .peak import peak_local_max from .corner import (corner_kitchen_rosenfeld, corner_harris, corner_shi_tomasi, corner_foerstner, corner_subpix, corner_peaks) from .corner_cy import corner_moravec from .template import match_template __all__ = ['daisy', 'hog', 'greycomatrix', 'greycoprops', 'local_binary_pattern', 'peak_local_max', 'corner_kitchen_rosenfeld', 'corner_harris', 'corner_shi_tomasi', 'corner_foerstner', 'corner_subpix', 'corner_peaks', 'corner_moravec', 'match_template']
<commit_before>from ._daisy import daisy from ._hog import hog from .texture import greycomatrix, greycoprops, local_binary_pattern from .peak import peak_local_max from .corner import (corner_kitchen_rosenfeld, corner_harris, corner_shi_tomasi, corner_foerstner, corner_subpix, corner_peaks) from .corner_cy import corner_moravec from .template import match_template <commit_msg>Add __all__ to feature package<commit_after>
from ._daisy import daisy from ._hog import hog from .texture import greycomatrix, greycoprops, local_binary_pattern from .peak import peak_local_max from .corner import (corner_kitchen_rosenfeld, corner_harris, corner_shi_tomasi, corner_foerstner, corner_subpix, corner_peaks) from .corner_cy import corner_moravec from .template import match_template __all__ = ['daisy', 'hog', 'greycomatrix', 'greycoprops', 'local_binary_pattern', 'peak_local_max', 'corner_kitchen_rosenfeld', 'corner_harris', 'corner_shi_tomasi', 'corner_foerstner', 'corner_subpix', 'corner_peaks', 'corner_moravec', 'match_template']
from ._daisy import daisy from ._hog import hog from .texture import greycomatrix, greycoprops, local_binary_pattern from .peak import peak_local_max from .corner import (corner_kitchen_rosenfeld, corner_harris, corner_shi_tomasi, corner_foerstner, corner_subpix, corner_peaks) from .corner_cy import corner_moravec from .template import match_template Add __all__ to feature packagefrom ._daisy import daisy from ._hog import hog from .texture import greycomatrix, greycoprops, local_binary_pattern from .peak import peak_local_max from .corner import (corner_kitchen_rosenfeld, corner_harris, corner_shi_tomasi, corner_foerstner, corner_subpix, corner_peaks) from .corner_cy import corner_moravec from .template import match_template __all__ = ['daisy', 'hog', 'greycomatrix', 'greycoprops', 'local_binary_pattern', 'peak_local_max', 'corner_kitchen_rosenfeld', 'corner_harris', 'corner_shi_tomasi', 'corner_foerstner', 'corner_subpix', 'corner_peaks', 'corner_moravec', 'match_template']
<commit_before>from ._daisy import daisy from ._hog import hog from .texture import greycomatrix, greycoprops, local_binary_pattern from .peak import peak_local_max from .corner import (corner_kitchen_rosenfeld, corner_harris, corner_shi_tomasi, corner_foerstner, corner_subpix, corner_peaks) from .corner_cy import corner_moravec from .template import match_template <commit_msg>Add __all__ to feature package<commit_after>from ._daisy import daisy from ._hog import hog from .texture import greycomatrix, greycoprops, local_binary_pattern from .peak import peak_local_max from .corner import (corner_kitchen_rosenfeld, corner_harris, corner_shi_tomasi, corner_foerstner, corner_subpix, corner_peaks) from .corner_cy import corner_moravec from .template import match_template __all__ = ['daisy', 'hog', 'greycomatrix', 'greycoprops', 'local_binary_pattern', 'peak_local_max', 'corner_kitchen_rosenfeld', 'corner_harris', 'corner_shi_tomasi', 'corner_foerstner', 'corner_subpix', 'corner_peaks', 'corner_moravec', 'match_template']
5c89aa079d94fe70bb5627eb67404bc65b80212a
sympy/solvers/decompogen.py
sympy/solvers/decompogen.py
from sympy.core import Function, Pow, sympify from sympy.polys import Poly def decompogen(f, symbol): """ Computes General functional decomposition of ``f``. Given an expression ``f``, returns a list ``[f_1, f_2, ..., f_n]``, where:: f = f_1 o f_2 o ... f_n = f_1(f_2(... f_n)) Note: This is a General decomposition function. For Polynomial decomposition see ``decompose`` in polys. Examples ======== >>> from sympy.solvers.decompogen import decompogen >>> from sympy.abc import x >>> from sympy import sqrt, sin, cos >>> decompogen(sin(cos(x)), x) [sin(x), cos(x)] >>> decompogen(sin(x)**2 + sin(x) + 1, x) [x**2 + x + 1, sin(x)] >>> decompogen(sqrt(6*x**2 - 5), x) [sqrt(x), 6*x**2 - 5] """ f = sympify(f) # ===== Simple Functions ===== # if isinstance(f, (Function, Pow)): return [f.subs(f.args[0], symbol), f.args[0]] # ===== Convert to Polynomial ===== # fp = Poly(f) gens = fp.gens if len(gens) == 1: f1 = f.subs(gens[0], symbol) f2 = gens[0] return [f1, f2]
from sympy.core import Function, Pow, sympify from sympy.polys import Poly def decompogen(f, symbol): """ Computes General functional decomposition of ``f``. Given an expression ``f``, returns a list ``[f_1, f_2, ..., f_n]``, where:: f = f_1 o f_2 o ... f_n = f_1(f_2(... f_n)) Note: This is a General decomposition function. For Polynomial decomposition see ``decompose`` in polys. Examples ======== >>> from sympy.solvers.decompogen import decompogen >>> from sympy.abc import x >>> from sympy import sqrt, sin, cos >>> decompogen(sin(cos(x)), x) [sin(x), cos(x)] >>> decompogen(sin(x)**2 + sin(x) + 1, x) [x**2 + x + 1, sin(x)] >>> decompogen(sqrt(6*x**2 - 5), x) [sqrt(x), 6*x**2 - 5] """ f = sympify(f) # ===== Simple Functions ===== # if isinstance(f, (Function, Pow)): return [f.subs(f.args[0], symbol), f.args[0]] # ===== Convert to Polynomial ===== # fp = Poly(f) gens = fp.gens if len(gens) == 1: f1 = f.subs(gens[0], symbol) f2 = gens[0] return [f1, f2] return [f]
Return `f` if no decomposition
Return `f` if no decomposition Signed-off-by: AMiT Kumar <a9f82ad3e8c446d8f49d5b7e05ec2c64a9e15ea9@gmail.com>
Python
bsd-3-clause
drufat/sympy,ChristinaZografou/sympy,chaffra/sympy,Vishluck/sympy,skidzo/sympy,ga7g08/sympy,wanglongqi/sympy,mafiya69/sympy,ga7g08/sympy,Davidjohnwilson/sympy,yashsharan/sympy,mcdaniel67/sympy,hargup/sympy,postvakje/sympy,Davidjohnwilson/sympy,VaibhavAgarwalVA/sympy,kaushik94/sympy,ChristinaZografou/sympy,MechCoder/sympy,MechCoder/sympy,Shaswat27/sympy,emon10005/sympy,kaichogami/sympy,jaimahajan1997/sympy,Shaswat27/sympy,kaushik94/sympy,Arafatk/sympy,Titan-C/sympy,kevalds51/sympy,jaimahajan1997/sympy,oliverlee/sympy,mafiya69/sympy,maniteja123/sympy,postvakje/sympy,drufat/sympy,AkademieOlympia/sympy,wanglongqi/sympy,madan96/sympy,abhiii5459/sympy,jerli/sympy,iamutkarshtiwari/sympy,Davidjohnwilson/sympy,lindsayad/sympy,madan96/sympy,ChristinaZografou/sympy,jerli/sympy,souravsingh/sympy,postvakje/sympy,rahuldan/sympy,saurabhjn76/sympy,ahhda/sympy,aktech/sympy,Titan-C/sympy,kaushik94/sympy,chaffra/sympy,madan96/sympy,lindsayad/sympy,wanglongqi/sympy,kevalds51/sympy,iamutkarshtiwari/sympy,jbbskinny/sympy,oliverlee/sympy,aktech/sympy,souravsingh/sympy,maniteja123/sympy,ahhda/sympy,iamutkarshtiwari/sympy,Curious72/sympy,emon10005/sympy,Titan-C/sympy,skidzo/sympy,Vishluck/sympy,ahhda/sympy,saurabhjn76/sympy,aktech/sympy,sampadsaha5/sympy,ga7g08/sympy,MechCoder/sympy,sampadsaha5/sympy,mcdaniel67/sympy,emon10005/sympy,jbbskinny/sympy,farhaanbukhsh/sympy,Curious72/sympy,VaibhavAgarwalVA/sympy,kevalds51/sympy,sahmed95/sympy,saurabhjn76/sympy,sahmed95/sympy,Arafatk/sympy,rahuldan/sympy,mcdaniel67/sympy,rahuldan/sympy,Shaswat27/sympy,yashsharan/sympy,abhiii5459/sympy,mafiya69/sympy,Vishluck/sympy,Arafatk/sympy,AkademieOlympia/sympy,skidzo/sympy,Curious72/sympy,souravsingh/sympy,AkademieOlympia/sympy,VaibhavAgarwalVA/sympy,sahmed95/sympy,jbbskinny/sympy,hargup/sympy,hargup/sympy,abhiii5459/sympy,farhaanbukhsh/sympy,kaichogami/sympy,chaffra/sympy,kaichogami/sympy,sampadsaha5/sympy,farhaanbukhsh/sympy,yashsharan/sympy,jerli/sympy,maniteja123/sympy,jaimahajan1997/sympy,oliverlee/sympy,lindsayad/sympy,drufat/sympy
from sympy.core import Function, Pow, sympify from sympy.polys import Poly def decompogen(f, symbol): """ Computes General functional decomposition of ``f``. Given an expression ``f``, returns a list ``[f_1, f_2, ..., f_n]``, where:: f = f_1 o f_2 o ... f_n = f_1(f_2(... f_n)) Note: This is a General decomposition function. For Polynomial decomposition see ``decompose`` in polys. Examples ======== >>> from sympy.solvers.decompogen import decompogen >>> from sympy.abc import x >>> from sympy import sqrt, sin, cos >>> decompogen(sin(cos(x)), x) [sin(x), cos(x)] >>> decompogen(sin(x)**2 + sin(x) + 1, x) [x**2 + x + 1, sin(x)] >>> decompogen(sqrt(6*x**2 - 5), x) [sqrt(x), 6*x**2 - 5] """ f = sympify(f) # ===== Simple Functions ===== # if isinstance(f, (Function, Pow)): return [f.subs(f.args[0], symbol), f.args[0]] # ===== Convert to Polynomial ===== # fp = Poly(f) gens = fp.gens if len(gens) == 1: f1 = f.subs(gens[0], symbol) f2 = gens[0] return [f1, f2] Return `f` if no decomposition Signed-off-by: AMiT Kumar <a9f82ad3e8c446d8f49d5b7e05ec2c64a9e15ea9@gmail.com>
from sympy.core import Function, Pow, sympify from sympy.polys import Poly def decompogen(f, symbol): """ Computes General functional decomposition of ``f``. Given an expression ``f``, returns a list ``[f_1, f_2, ..., f_n]``, where:: f = f_1 o f_2 o ... f_n = f_1(f_2(... f_n)) Note: This is a General decomposition function. For Polynomial decomposition see ``decompose`` in polys. Examples ======== >>> from sympy.solvers.decompogen import decompogen >>> from sympy.abc import x >>> from sympy import sqrt, sin, cos >>> decompogen(sin(cos(x)), x) [sin(x), cos(x)] >>> decompogen(sin(x)**2 + sin(x) + 1, x) [x**2 + x + 1, sin(x)] >>> decompogen(sqrt(6*x**2 - 5), x) [sqrt(x), 6*x**2 - 5] """ f = sympify(f) # ===== Simple Functions ===== # if isinstance(f, (Function, Pow)): return [f.subs(f.args[0], symbol), f.args[0]] # ===== Convert to Polynomial ===== # fp = Poly(f) gens = fp.gens if len(gens) == 1: f1 = f.subs(gens[0], symbol) f2 = gens[0] return [f1, f2] return [f]
<commit_before>from sympy.core import Function, Pow, sympify from sympy.polys import Poly def decompogen(f, symbol): """ Computes General functional decomposition of ``f``. Given an expression ``f``, returns a list ``[f_1, f_2, ..., f_n]``, where:: f = f_1 o f_2 o ... f_n = f_1(f_2(... f_n)) Note: This is a General decomposition function. For Polynomial decomposition see ``decompose`` in polys. Examples ======== >>> from sympy.solvers.decompogen import decompogen >>> from sympy.abc import x >>> from sympy import sqrt, sin, cos >>> decompogen(sin(cos(x)), x) [sin(x), cos(x)] >>> decompogen(sin(x)**2 + sin(x) + 1, x) [x**2 + x + 1, sin(x)] >>> decompogen(sqrt(6*x**2 - 5), x) [sqrt(x), 6*x**2 - 5] """ f = sympify(f) # ===== Simple Functions ===== # if isinstance(f, (Function, Pow)): return [f.subs(f.args[0], symbol), f.args[0]] # ===== Convert to Polynomial ===== # fp = Poly(f) gens = fp.gens if len(gens) == 1: f1 = f.subs(gens[0], symbol) f2 = gens[0] return [f1, f2] <commit_msg>Return `f` if no decomposition Signed-off-by: AMiT Kumar <a9f82ad3e8c446d8f49d5b7e05ec2c64a9e15ea9@gmail.com><commit_after>
from sympy.core import Function, Pow, sympify from sympy.polys import Poly def decompogen(f, symbol): """ Computes General functional decomposition of ``f``. Given an expression ``f``, returns a list ``[f_1, f_2, ..., f_n]``, where:: f = f_1 o f_2 o ... f_n = f_1(f_2(... f_n)) Note: This is a General decomposition function. For Polynomial decomposition see ``decompose`` in polys. Examples ======== >>> from sympy.solvers.decompogen import decompogen >>> from sympy.abc import x >>> from sympy import sqrt, sin, cos >>> decompogen(sin(cos(x)), x) [sin(x), cos(x)] >>> decompogen(sin(x)**2 + sin(x) + 1, x) [x**2 + x + 1, sin(x)] >>> decompogen(sqrt(6*x**2 - 5), x) [sqrt(x), 6*x**2 - 5] """ f = sympify(f) # ===== Simple Functions ===== # if isinstance(f, (Function, Pow)): return [f.subs(f.args[0], symbol), f.args[0]] # ===== Convert to Polynomial ===== # fp = Poly(f) gens = fp.gens if len(gens) == 1: f1 = f.subs(gens[0], symbol) f2 = gens[0] return [f1, f2] return [f]
from sympy.core import Function, Pow, sympify from sympy.polys import Poly def decompogen(f, symbol): """ Computes General functional decomposition of ``f``. Given an expression ``f``, returns a list ``[f_1, f_2, ..., f_n]``, where:: f = f_1 o f_2 o ... f_n = f_1(f_2(... f_n)) Note: This is a General decomposition function. For Polynomial decomposition see ``decompose`` in polys. Examples ======== >>> from sympy.solvers.decompogen import decompogen >>> from sympy.abc import x >>> from sympy import sqrt, sin, cos >>> decompogen(sin(cos(x)), x) [sin(x), cos(x)] >>> decompogen(sin(x)**2 + sin(x) + 1, x) [x**2 + x + 1, sin(x)] >>> decompogen(sqrt(6*x**2 - 5), x) [sqrt(x), 6*x**2 - 5] """ f = sympify(f) # ===== Simple Functions ===== # if isinstance(f, (Function, Pow)): return [f.subs(f.args[0], symbol), f.args[0]] # ===== Convert to Polynomial ===== # fp = Poly(f) gens = fp.gens if len(gens) == 1: f1 = f.subs(gens[0], symbol) f2 = gens[0] return [f1, f2] Return `f` if no decomposition Signed-off-by: AMiT Kumar <a9f82ad3e8c446d8f49d5b7e05ec2c64a9e15ea9@gmail.com>from sympy.core import Function, Pow, sympify from sympy.polys import Poly def decompogen(f, symbol): """ Computes General functional decomposition of ``f``. Given an expression ``f``, returns a list ``[f_1, f_2, ..., f_n]``, where:: f = f_1 o f_2 o ... f_n = f_1(f_2(... f_n)) Note: This is a General decomposition function. For Polynomial decomposition see ``decompose`` in polys. Examples ======== >>> from sympy.solvers.decompogen import decompogen >>> from sympy.abc import x >>> from sympy import sqrt, sin, cos >>> decompogen(sin(cos(x)), x) [sin(x), cos(x)] >>> decompogen(sin(x)**2 + sin(x) + 1, x) [x**2 + x + 1, sin(x)] >>> decompogen(sqrt(6*x**2 - 5), x) [sqrt(x), 6*x**2 - 5] """ f = sympify(f) # ===== Simple Functions ===== # if isinstance(f, (Function, Pow)): return [f.subs(f.args[0], symbol), f.args[0]] # ===== Convert to Polynomial ===== # fp = Poly(f) gens = fp.gens if len(gens) == 1: f1 = f.subs(gens[0], symbol) f2 = gens[0] return [f1, f2] return [f]
<commit_before>from sympy.core import Function, Pow, sympify from sympy.polys import Poly def decompogen(f, symbol): """ Computes General functional decomposition of ``f``. Given an expression ``f``, returns a list ``[f_1, f_2, ..., f_n]``, where:: f = f_1 o f_2 o ... f_n = f_1(f_2(... f_n)) Note: This is a General decomposition function. For Polynomial decomposition see ``decompose`` in polys. Examples ======== >>> from sympy.solvers.decompogen import decompogen >>> from sympy.abc import x >>> from sympy import sqrt, sin, cos >>> decompogen(sin(cos(x)), x) [sin(x), cos(x)] >>> decompogen(sin(x)**2 + sin(x) + 1, x) [x**2 + x + 1, sin(x)] >>> decompogen(sqrt(6*x**2 - 5), x) [sqrt(x), 6*x**2 - 5] """ f = sympify(f) # ===== Simple Functions ===== # if isinstance(f, (Function, Pow)): return [f.subs(f.args[0], symbol), f.args[0]] # ===== Convert to Polynomial ===== # fp = Poly(f) gens = fp.gens if len(gens) == 1: f1 = f.subs(gens[0], symbol) f2 = gens[0] return [f1, f2] <commit_msg>Return `f` if no decomposition Signed-off-by: AMiT Kumar <a9f82ad3e8c446d8f49d5b7e05ec2c64a9e15ea9@gmail.com><commit_after>from sympy.core import Function, Pow, sympify from sympy.polys import Poly def decompogen(f, symbol): """ Computes General functional decomposition of ``f``. Given an expression ``f``, returns a list ``[f_1, f_2, ..., f_n]``, where:: f = f_1 o f_2 o ... f_n = f_1(f_2(... f_n)) Note: This is a General decomposition function. For Polynomial decomposition see ``decompose`` in polys. Examples ======== >>> from sympy.solvers.decompogen import decompogen >>> from sympy.abc import x >>> from sympy import sqrt, sin, cos >>> decompogen(sin(cos(x)), x) [sin(x), cos(x)] >>> decompogen(sin(x)**2 + sin(x) + 1, x) [x**2 + x + 1, sin(x)] >>> decompogen(sqrt(6*x**2 - 5), x) [sqrt(x), 6*x**2 - 5] """ f = sympify(f) # ===== Simple Functions ===== # if isinstance(f, (Function, Pow)): return [f.subs(f.args[0], symbol), f.args[0]] # ===== Convert to Polynomial ===== # fp = Poly(f) gens = fp.gens if len(gens) == 1: f1 = f.subs(gens[0], symbol) f2 = gens[0] return [f1, f2] return [f]
6d3b191bd35f64b097ffac2a514d2400f2e07983
tests/__main__.py
tests/__main__.py
#!/usr/bin/env python3 import os import sys import unittest sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..')) load = unittest.defaultTestLoader.loadTestsFromModule suites = [] dirname = os.path.dirname(__file__) sys.path.append(dirname) filenames = os.listdir(dirname) for filename in filenames: if filename.startswith('test_') and filename.endswith('.py'): name = filename[:-3] plugin = __import__(name) suites.append(load(plugin)) if __name__ == '__main__': suite = unittest.TestSuite(suites) runner = unittest.TextTestRunner(verbosity=2) result = runner.run(suite)
#!/usr/bin/env python3 import os import sys import unittest sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..')) load = unittest.defaultTestLoader.loadTestsFromModule suites = [] dirname = os.path.dirname(__file__) sys.path.append(dirname) filenames = os.listdir(dirname) for filename in filenames: if filename.startswith('test_') and filename.endswith('.py'): name = filename[:-3] plugin = __import__(name) suites.append(load(plugin)) if __name__ == '__main__': suite = unittest.TestSuite(suites) runner = unittest.TextTestRunner(verbosity=2) result = runner.run(suite) if result.wasSuccessful(): sys.exit(0) else: sys.exit(1)
Fix return value of tests.
Fix return value of tests.
Python
bsd-3-clause
ProgVal/Inceptoe,ProgVal/Inceptoe
#!/usr/bin/env python3 import os import sys import unittest sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..')) load = unittest.defaultTestLoader.loadTestsFromModule suites = [] dirname = os.path.dirname(__file__) sys.path.append(dirname) filenames = os.listdir(dirname) for filename in filenames: if filename.startswith('test_') and filename.endswith('.py'): name = filename[:-3] plugin = __import__(name) suites.append(load(plugin)) if __name__ == '__main__': suite = unittest.TestSuite(suites) runner = unittest.TextTestRunner(verbosity=2) result = runner.run(suite) Fix return value of tests.
#!/usr/bin/env python3 import os import sys import unittest sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..')) load = unittest.defaultTestLoader.loadTestsFromModule suites = [] dirname = os.path.dirname(__file__) sys.path.append(dirname) filenames = os.listdir(dirname) for filename in filenames: if filename.startswith('test_') and filename.endswith('.py'): name = filename[:-3] plugin = __import__(name) suites.append(load(plugin)) if __name__ == '__main__': suite = unittest.TestSuite(suites) runner = unittest.TextTestRunner(verbosity=2) result = runner.run(suite) if result.wasSuccessful(): sys.exit(0) else: sys.exit(1)
<commit_before>#!/usr/bin/env python3 import os import sys import unittest sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..')) load = unittest.defaultTestLoader.loadTestsFromModule suites = [] dirname = os.path.dirname(__file__) sys.path.append(dirname) filenames = os.listdir(dirname) for filename in filenames: if filename.startswith('test_') and filename.endswith('.py'): name = filename[:-3] plugin = __import__(name) suites.append(load(plugin)) if __name__ == '__main__': suite = unittest.TestSuite(suites) runner = unittest.TextTestRunner(verbosity=2) result = runner.run(suite) <commit_msg>Fix return value of tests.<commit_after>
#!/usr/bin/env python3 import os import sys import unittest sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..')) load = unittest.defaultTestLoader.loadTestsFromModule suites = [] dirname = os.path.dirname(__file__) sys.path.append(dirname) filenames = os.listdir(dirname) for filename in filenames: if filename.startswith('test_') and filename.endswith('.py'): name = filename[:-3] plugin = __import__(name) suites.append(load(plugin)) if __name__ == '__main__': suite = unittest.TestSuite(suites) runner = unittest.TextTestRunner(verbosity=2) result = runner.run(suite) if result.wasSuccessful(): sys.exit(0) else: sys.exit(1)
#!/usr/bin/env python3 import os import sys import unittest sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..')) load = unittest.defaultTestLoader.loadTestsFromModule suites = [] dirname = os.path.dirname(__file__) sys.path.append(dirname) filenames = os.listdir(dirname) for filename in filenames: if filename.startswith('test_') and filename.endswith('.py'): name = filename[:-3] plugin = __import__(name) suites.append(load(plugin)) if __name__ == '__main__': suite = unittest.TestSuite(suites) runner = unittest.TextTestRunner(verbosity=2) result = runner.run(suite) Fix return value of tests.#!/usr/bin/env python3 import os import sys import unittest sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..')) load = unittest.defaultTestLoader.loadTestsFromModule suites = [] dirname = os.path.dirname(__file__) sys.path.append(dirname) filenames = os.listdir(dirname) for filename in filenames: if filename.startswith('test_') and filename.endswith('.py'): name = filename[:-3] plugin = __import__(name) suites.append(load(plugin)) if __name__ == '__main__': suite = unittest.TestSuite(suites) runner = unittest.TextTestRunner(verbosity=2) result = runner.run(suite) if result.wasSuccessful(): sys.exit(0) else: sys.exit(1)
<commit_before>#!/usr/bin/env python3 import os import sys import unittest sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..')) load = unittest.defaultTestLoader.loadTestsFromModule suites = [] dirname = os.path.dirname(__file__) sys.path.append(dirname) filenames = os.listdir(dirname) for filename in filenames: if filename.startswith('test_') and filename.endswith('.py'): name = filename[:-3] plugin = __import__(name) suites.append(load(plugin)) if __name__ == '__main__': suite = unittest.TestSuite(suites) runner = unittest.TextTestRunner(verbosity=2) result = runner.run(suite) <commit_msg>Fix return value of tests.<commit_after>#!/usr/bin/env python3 import os import sys import unittest sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..')) load = unittest.defaultTestLoader.loadTestsFromModule suites = [] dirname = os.path.dirname(__file__) sys.path.append(dirname) filenames = os.listdir(dirname) for filename in filenames: if filename.startswith('test_') and filename.endswith('.py'): name = filename[:-3] plugin = __import__(name) suites.append(load(plugin)) if __name__ == '__main__': suite = unittest.TestSuite(suites) runner = unittest.TextTestRunner(verbosity=2) result = runner.run(suite) if result.wasSuccessful(): sys.exit(0) else: sys.exit(1)
969c5129ec7c742985b7a3d09eb5ddd4b44b1743
dothub/__main__.py
dothub/__main__.py
import logging from . import config, cli def main(): logging.basicConfig(levle=logging.INFO, format="%(message)s") conf = config.load_config() cli.dothub(obj={}, default_map=conf) if __name__ == '__main__': # pragma: no cover exit(main())
import logging from . import config, cli def main(): logging.basicConfig(level=logging.INFO, format="%(message)s") conf = config.load_config() cli.dothub(obj={}, default_map=conf) if __name__ == '__main__': # pragma: no cover exit(main())
Fix typo on logging configuration
Fix typo on logging configuration
Python
mit
Mariocj89/dothub
import logging from . import config, cli def main(): logging.basicConfig(levle=logging.INFO, format="%(message)s") conf = config.load_config() cli.dothub(obj={}, default_map=conf) if __name__ == '__main__': # pragma: no cover exit(main()) Fix typo on logging configuration
import logging from . import config, cli def main(): logging.basicConfig(level=logging.INFO, format="%(message)s") conf = config.load_config() cli.dothub(obj={}, default_map=conf) if __name__ == '__main__': # pragma: no cover exit(main())
<commit_before>import logging from . import config, cli def main(): logging.basicConfig(levle=logging.INFO, format="%(message)s") conf = config.load_config() cli.dothub(obj={}, default_map=conf) if __name__ == '__main__': # pragma: no cover exit(main()) <commit_msg>Fix typo on logging configuration<commit_after>
import logging from . import config, cli def main(): logging.basicConfig(level=logging.INFO, format="%(message)s") conf = config.load_config() cli.dothub(obj={}, default_map=conf) if __name__ == '__main__': # pragma: no cover exit(main())
import logging from . import config, cli def main(): logging.basicConfig(levle=logging.INFO, format="%(message)s") conf = config.load_config() cli.dothub(obj={}, default_map=conf) if __name__ == '__main__': # pragma: no cover exit(main()) Fix typo on logging configurationimport logging from . import config, cli def main(): logging.basicConfig(level=logging.INFO, format="%(message)s") conf = config.load_config() cli.dothub(obj={}, default_map=conf) if __name__ == '__main__': # pragma: no cover exit(main())
<commit_before>import logging from . import config, cli def main(): logging.basicConfig(levle=logging.INFO, format="%(message)s") conf = config.load_config() cli.dothub(obj={}, default_map=conf) if __name__ == '__main__': # pragma: no cover exit(main()) <commit_msg>Fix typo on logging configuration<commit_after>import logging from . import config, cli def main(): logging.basicConfig(level=logging.INFO, format="%(message)s") conf = config.load_config() cli.dothub(obj={}, default_map=conf) if __name__ == '__main__': # pragma: no cover exit(main())
9fdd24ed20c553638cde7c67f994ea72da0ef149
tests/conftest.py
tests/conftest.py
import shutil from pathlib import Path import pytest from pandas_profiling.utils.cache import cache_file @pytest.fixture(scope="function") def get_data_file(tmpdir): def getter(file_name, url): source_file = cache_file(file_name, url) # Move to temporary directory test_path = Path(str(tmpdir)) shutil.copy(source_file, test_path / file_name) return test_path / file_name return getter @pytest.fixture(scope="module") def test_output_dir(tmpdir_factory): test_path = Path(str(tmpdir_factory.mktemp("test"))) yield test_path shutil.rmtree(str(test_path))
import shutil from pathlib import Path import pytest from pandas_profiling.utils.cache import cache_file @pytest.fixture(scope="function") def get_data_file(tmpdir): def getter(file_name, url): source_file = cache_file(file_name, url) # Move to temporary directory test_path = Path(str(tmpdir)) shutil.copy(str(source_file), str(test_path / file_name)) return str(test_path / file_name) return getter @pytest.fixture(scope="module") def test_output_dir(tmpdir_factory): test_path = Path(str(tmpdir_factory.mktemp("test"))) yield test_path shutil.rmtree(str(test_path))
Convert Path to str for Python 3.5
Convert Path to str for Python 3.5
Python
mit
JosPolfliet/pandas-profiling,JosPolfliet/pandas-profiling
import shutil from pathlib import Path import pytest from pandas_profiling.utils.cache import cache_file @pytest.fixture(scope="function") def get_data_file(tmpdir): def getter(file_name, url): source_file = cache_file(file_name, url) # Move to temporary directory test_path = Path(str(tmpdir)) shutil.copy(source_file, test_path / file_name) return test_path / file_name return getter @pytest.fixture(scope="module") def test_output_dir(tmpdir_factory): test_path = Path(str(tmpdir_factory.mktemp("test"))) yield test_path shutil.rmtree(str(test_path)) Convert Path to str for Python 3.5
import shutil from pathlib import Path import pytest from pandas_profiling.utils.cache import cache_file @pytest.fixture(scope="function") def get_data_file(tmpdir): def getter(file_name, url): source_file = cache_file(file_name, url) # Move to temporary directory test_path = Path(str(tmpdir)) shutil.copy(str(source_file), str(test_path / file_name)) return str(test_path / file_name) return getter @pytest.fixture(scope="module") def test_output_dir(tmpdir_factory): test_path = Path(str(tmpdir_factory.mktemp("test"))) yield test_path shutil.rmtree(str(test_path))
<commit_before>import shutil from pathlib import Path import pytest from pandas_profiling.utils.cache import cache_file @pytest.fixture(scope="function") def get_data_file(tmpdir): def getter(file_name, url): source_file = cache_file(file_name, url) # Move to temporary directory test_path = Path(str(tmpdir)) shutil.copy(source_file, test_path / file_name) return test_path / file_name return getter @pytest.fixture(scope="module") def test_output_dir(tmpdir_factory): test_path = Path(str(tmpdir_factory.mktemp("test"))) yield test_path shutil.rmtree(str(test_path)) <commit_msg>Convert Path to str for Python 3.5<commit_after>
import shutil from pathlib import Path import pytest from pandas_profiling.utils.cache import cache_file @pytest.fixture(scope="function") def get_data_file(tmpdir): def getter(file_name, url): source_file = cache_file(file_name, url) # Move to temporary directory test_path = Path(str(tmpdir)) shutil.copy(str(source_file), str(test_path / file_name)) return str(test_path / file_name) return getter @pytest.fixture(scope="module") def test_output_dir(tmpdir_factory): test_path = Path(str(tmpdir_factory.mktemp("test"))) yield test_path shutil.rmtree(str(test_path))
import shutil from pathlib import Path import pytest from pandas_profiling.utils.cache import cache_file @pytest.fixture(scope="function") def get_data_file(tmpdir): def getter(file_name, url): source_file = cache_file(file_name, url) # Move to temporary directory test_path = Path(str(tmpdir)) shutil.copy(source_file, test_path / file_name) return test_path / file_name return getter @pytest.fixture(scope="module") def test_output_dir(tmpdir_factory): test_path = Path(str(tmpdir_factory.mktemp("test"))) yield test_path shutil.rmtree(str(test_path)) Convert Path to str for Python 3.5import shutil from pathlib import Path import pytest from pandas_profiling.utils.cache import cache_file @pytest.fixture(scope="function") def get_data_file(tmpdir): def getter(file_name, url): source_file = cache_file(file_name, url) # Move to temporary directory test_path = Path(str(tmpdir)) shutil.copy(str(source_file), str(test_path / file_name)) return str(test_path / file_name) return getter @pytest.fixture(scope="module") def test_output_dir(tmpdir_factory): test_path = Path(str(tmpdir_factory.mktemp("test"))) yield test_path shutil.rmtree(str(test_path))
<commit_before>import shutil from pathlib import Path import pytest from pandas_profiling.utils.cache import cache_file @pytest.fixture(scope="function") def get_data_file(tmpdir): def getter(file_name, url): source_file = cache_file(file_name, url) # Move to temporary directory test_path = Path(str(tmpdir)) shutil.copy(source_file, test_path / file_name) return test_path / file_name return getter @pytest.fixture(scope="module") def test_output_dir(tmpdir_factory): test_path = Path(str(tmpdir_factory.mktemp("test"))) yield test_path shutil.rmtree(str(test_path)) <commit_msg>Convert Path to str for Python 3.5<commit_after>import shutil from pathlib import Path import pytest from pandas_profiling.utils.cache import cache_file @pytest.fixture(scope="function") def get_data_file(tmpdir): def getter(file_name, url): source_file = cache_file(file_name, url) # Move to temporary directory test_path = Path(str(tmpdir)) shutil.copy(str(source_file), str(test_path / file_name)) return str(test_path / file_name) return getter @pytest.fixture(scope="module") def test_output_dir(tmpdir_factory): test_path = Path(str(tmpdir_factory.mktemp("test"))) yield test_path shutil.rmtree(str(test_path))
09fc5b3e37cc0bcaf764c540f32d1a3eab1b8cf2
tests/som_test.py
tests/som_test.py
import unittest from parameterized import parameterized from som.vm.universe import Universe class SomTest(unittest.TestCase): @parameterized.expand([ ("ClassStructure",), ("Array" ,), ("Block" ,), ("ClassLoading" ,), ("Closure" ,), ("Coercion" ,), ("CompilerReturn",), ("Double" ,), ("Empty" ,), ("Hash" ,), ("Integer" ,), ("ObjectSize" ,), ("Preliminary" ,), ("Reflection" ,), ("SelfBlock" ,), ("Super" ,), ("String" ,), ("Symbol" ,), ("System" ,), ("Vector" ,)]) def test_som_test(self, test_name): args = ["-cp", "Smalltalk", "TestSuite/TestHarness.som", test_name] u = Universe(True) u.interpret(args) self.assertEquals(0, u.last_exit_code()) import sys if sys.modules.has_key('pytest'): # hack to make pytest not to collect the unexpanded test method delattr(SomTest, "test_som_test")
import unittest from parameterized import parameterized from som.vm.universe import Universe class SomTest(unittest.TestCase): @parameterized.expand([ ("ClassStructure",), ("Array" ,), ("Block" ,), ("ClassLoading" ,), ("Closure" ,), ("Coercion" ,), ("CompilerReturn",), ("Double" ,), ("Empty" ,), ("Hash" ,), ("Integer" ,), ("ObjectSize" ,), ("Preliminary" ,), ("Reflection" ,), ("SelfBlock" ,), ("Super" ,), ("String" ,), ("Symbol" ,), ("System" ,), ("Vector" ,)]) def test_som_test(self, test_name): args = ["-cp", "Smalltalk", "TestSuite/TestHarness.som", test_name] u = Universe(True) u.interpret(args) self.assertEquals(0, u.last_exit_code()) import sys if 'pytest' in sys.modules: # hack to make pytest not to collect the unexpanded test method delattr(SomTest, "test_som_test")
Use in, because has_key seems to be deprecated
Use in, because has_key seems to be deprecated Signed-off-by: Stefan Marr <46f1a0bd5592a2f9244ca321b129902a06b53e03@stefan-marr.de>
Python
mit
SOM-st/RTruffleSOM,SOM-st/PySOM,SOM-st/RPySOM,SOM-st/RPySOM,SOM-st/RTruffleSOM,smarr/RTruffleSOM,smarr/PySOM,smarr/RTruffleSOM,smarr/PySOM,SOM-st/PySOM
import unittest from parameterized import parameterized from som.vm.universe import Universe class SomTest(unittest.TestCase): @parameterized.expand([ ("ClassStructure",), ("Array" ,), ("Block" ,), ("ClassLoading" ,), ("Closure" ,), ("Coercion" ,), ("CompilerReturn",), ("Double" ,), ("Empty" ,), ("Hash" ,), ("Integer" ,), ("ObjectSize" ,), ("Preliminary" ,), ("Reflection" ,), ("SelfBlock" ,), ("Super" ,), ("String" ,), ("Symbol" ,), ("System" ,), ("Vector" ,)]) def test_som_test(self, test_name): args = ["-cp", "Smalltalk", "TestSuite/TestHarness.som", test_name] u = Universe(True) u.interpret(args) self.assertEquals(0, u.last_exit_code()) import sys if sys.modules.has_key('pytest'): # hack to make pytest not to collect the unexpanded test method delattr(SomTest, "test_som_test") Use in, because has_key seems to be deprecated Signed-off-by: Stefan Marr <46f1a0bd5592a2f9244ca321b129902a06b53e03@stefan-marr.de>
import unittest from parameterized import parameterized from som.vm.universe import Universe class SomTest(unittest.TestCase): @parameterized.expand([ ("ClassStructure",), ("Array" ,), ("Block" ,), ("ClassLoading" ,), ("Closure" ,), ("Coercion" ,), ("CompilerReturn",), ("Double" ,), ("Empty" ,), ("Hash" ,), ("Integer" ,), ("ObjectSize" ,), ("Preliminary" ,), ("Reflection" ,), ("SelfBlock" ,), ("Super" ,), ("String" ,), ("Symbol" ,), ("System" ,), ("Vector" ,)]) def test_som_test(self, test_name): args = ["-cp", "Smalltalk", "TestSuite/TestHarness.som", test_name] u = Universe(True) u.interpret(args) self.assertEquals(0, u.last_exit_code()) import sys if 'pytest' in sys.modules: # hack to make pytest not to collect the unexpanded test method delattr(SomTest, "test_som_test")
<commit_before>import unittest from parameterized import parameterized from som.vm.universe import Universe class SomTest(unittest.TestCase): @parameterized.expand([ ("ClassStructure",), ("Array" ,), ("Block" ,), ("ClassLoading" ,), ("Closure" ,), ("Coercion" ,), ("CompilerReturn",), ("Double" ,), ("Empty" ,), ("Hash" ,), ("Integer" ,), ("ObjectSize" ,), ("Preliminary" ,), ("Reflection" ,), ("SelfBlock" ,), ("Super" ,), ("String" ,), ("Symbol" ,), ("System" ,), ("Vector" ,)]) def test_som_test(self, test_name): args = ["-cp", "Smalltalk", "TestSuite/TestHarness.som", test_name] u = Universe(True) u.interpret(args) self.assertEquals(0, u.last_exit_code()) import sys if sys.modules.has_key('pytest'): # hack to make pytest not to collect the unexpanded test method delattr(SomTest, "test_som_test") <commit_msg>Use in, because has_key seems to be deprecated Signed-off-by: Stefan Marr <46f1a0bd5592a2f9244ca321b129902a06b53e03@stefan-marr.de><commit_after>
import unittest from parameterized import parameterized from som.vm.universe import Universe class SomTest(unittest.TestCase): @parameterized.expand([ ("ClassStructure",), ("Array" ,), ("Block" ,), ("ClassLoading" ,), ("Closure" ,), ("Coercion" ,), ("CompilerReturn",), ("Double" ,), ("Empty" ,), ("Hash" ,), ("Integer" ,), ("ObjectSize" ,), ("Preliminary" ,), ("Reflection" ,), ("SelfBlock" ,), ("Super" ,), ("String" ,), ("Symbol" ,), ("System" ,), ("Vector" ,)]) def test_som_test(self, test_name): args = ["-cp", "Smalltalk", "TestSuite/TestHarness.som", test_name] u = Universe(True) u.interpret(args) self.assertEquals(0, u.last_exit_code()) import sys if 'pytest' in sys.modules: # hack to make pytest not to collect the unexpanded test method delattr(SomTest, "test_som_test")
import unittest from parameterized import parameterized from som.vm.universe import Universe class SomTest(unittest.TestCase): @parameterized.expand([ ("ClassStructure",), ("Array" ,), ("Block" ,), ("ClassLoading" ,), ("Closure" ,), ("Coercion" ,), ("CompilerReturn",), ("Double" ,), ("Empty" ,), ("Hash" ,), ("Integer" ,), ("ObjectSize" ,), ("Preliminary" ,), ("Reflection" ,), ("SelfBlock" ,), ("Super" ,), ("String" ,), ("Symbol" ,), ("System" ,), ("Vector" ,)]) def test_som_test(self, test_name): args = ["-cp", "Smalltalk", "TestSuite/TestHarness.som", test_name] u = Universe(True) u.interpret(args) self.assertEquals(0, u.last_exit_code()) import sys if sys.modules.has_key('pytest'): # hack to make pytest not to collect the unexpanded test method delattr(SomTest, "test_som_test") Use in, because has_key seems to be deprecated Signed-off-by: Stefan Marr <46f1a0bd5592a2f9244ca321b129902a06b53e03@stefan-marr.de>import unittest from parameterized import parameterized from som.vm.universe import Universe class SomTest(unittest.TestCase): @parameterized.expand([ ("ClassStructure",), ("Array" ,), ("Block" ,), ("ClassLoading" ,), ("Closure" ,), ("Coercion" ,), ("CompilerReturn",), ("Double" ,), ("Empty" ,), ("Hash" ,), ("Integer" ,), ("ObjectSize" ,), ("Preliminary" ,), ("Reflection" ,), ("SelfBlock" ,), ("Super" ,), ("String" ,), ("Symbol" ,), ("System" ,), ("Vector" ,)]) def test_som_test(self, test_name): args = ["-cp", "Smalltalk", "TestSuite/TestHarness.som", test_name] u = Universe(True) u.interpret(args) self.assertEquals(0, u.last_exit_code()) import sys if 'pytest' in sys.modules: # hack to make pytest not to collect the unexpanded test method delattr(SomTest, "test_som_test")
<commit_before>import unittest from parameterized import parameterized from som.vm.universe import Universe class SomTest(unittest.TestCase): @parameterized.expand([ ("ClassStructure",), ("Array" ,), ("Block" ,), ("ClassLoading" ,), ("Closure" ,), ("Coercion" ,), ("CompilerReturn",), ("Double" ,), ("Empty" ,), ("Hash" ,), ("Integer" ,), ("ObjectSize" ,), ("Preliminary" ,), ("Reflection" ,), ("SelfBlock" ,), ("Super" ,), ("String" ,), ("Symbol" ,), ("System" ,), ("Vector" ,)]) def test_som_test(self, test_name): args = ["-cp", "Smalltalk", "TestSuite/TestHarness.som", test_name] u = Universe(True) u.interpret(args) self.assertEquals(0, u.last_exit_code()) import sys if sys.modules.has_key('pytest'): # hack to make pytest not to collect the unexpanded test method delattr(SomTest, "test_som_test") <commit_msg>Use in, because has_key seems to be deprecated Signed-off-by: Stefan Marr <46f1a0bd5592a2f9244ca321b129902a06b53e03@stefan-marr.de><commit_after>import unittest from parameterized import parameterized from som.vm.universe import Universe class SomTest(unittest.TestCase): @parameterized.expand([ ("ClassStructure",), ("Array" ,), ("Block" ,), ("ClassLoading" ,), ("Closure" ,), ("Coercion" ,), ("CompilerReturn",), ("Double" ,), ("Empty" ,), ("Hash" ,), ("Integer" ,), ("ObjectSize" ,), ("Preliminary" ,), ("Reflection" ,), ("SelfBlock" ,), ("Super" ,), ("String" ,), ("Symbol" ,), ("System" ,), ("Vector" ,)]) def test_som_test(self, test_name): args = ["-cp", "Smalltalk", "TestSuite/TestHarness.som", test_name] u = Universe(True) u.interpret(args) self.assertEquals(0, u.last_exit_code()) import sys if 'pytest' in sys.modules: # hack to make pytest not to collect the unexpanded test method delattr(SomTest, "test_som_test")
e9e8428563545d00eda25a540a3621943bf34143
tests/test_now.py
tests/test_now.py
# -*- coding: utf-8 -*- import pytest from freezegun import freeze_time from jinja2 import Environment, exceptions @pytest.fixture(scope='session') def environment(): return Environment(extensions=['jinja2_time.TimeExtension']) @pytest.yield_fixture(autouse=True) def freeze(): freezer = freeze_time("2015-12-09 23:33:01") freezer.start() yield freezer.stop() def test_tz_is_required(environment): with pytest.raises(exceptions.TemplateSyntaxError): environment.from_string('{% now %}') def test_utc_default_datetime_format(environment): template = environment.from_string("{% now 'utc' %}") assert template.render() == "2015-12-09" @pytest.fixture(params=['utc', 'local', 'Europe/Berlin']) def valid_tz(request): return request.param def test_accept_valid_timezones(environment, valid_tz): template = environment.from_string( "{% now '" + valid_tz + "', '%Y-%m' %}" ) assert template.render() == '2015-12'
# -*- coding: utf-8 -*- import pytest from freezegun import freeze_time from jinja2 import Environment, exceptions @pytest.fixture def environment(): return Environment(extensions=['jinja2_time.TimeExtension']) @pytest.yield_fixture(autouse=True) def freeze(): freezer = freeze_time("2015-12-09 23:33:01") freezer.start() yield freezer.stop() def test_tz_is_required(environment): with pytest.raises(exceptions.TemplateSyntaxError): environment.from_string('{% now %}') def test_utc_default_datetime_format(environment): template = environment.from_string("{% now 'utc' %}") assert template.render() == "2015-12-09" @pytest.fixture(params=['utc', 'local', 'Europe/Berlin']) def valid_tz(request): return request.param def test_accept_valid_timezones(environment, valid_tz): template = environment.from_string( "{% now '" + valid_tz + "', '%Y-%m' %}" ) assert template.render() == '2015-12' def test_environment_datetime_format(environment): environment.datetime_format = '%a, %d %b %Y %H:%M:%S' template = environment.from_string("{% now 'utc' %}") assert template.render() == "Wed, 09 Dec 2015 23:33:01"
Implement a test for environment.datetime_format
Implement a test for environment.datetime_format
Python
mit
hackebrot/jinja2-time
# -*- coding: utf-8 -*- import pytest from freezegun import freeze_time from jinja2 import Environment, exceptions @pytest.fixture(scope='session') def environment(): return Environment(extensions=['jinja2_time.TimeExtension']) @pytest.yield_fixture(autouse=True) def freeze(): freezer = freeze_time("2015-12-09 23:33:01") freezer.start() yield freezer.stop() def test_tz_is_required(environment): with pytest.raises(exceptions.TemplateSyntaxError): environment.from_string('{% now %}') def test_utc_default_datetime_format(environment): template = environment.from_string("{% now 'utc' %}") assert template.render() == "2015-12-09" @pytest.fixture(params=['utc', 'local', 'Europe/Berlin']) def valid_tz(request): return request.param def test_accept_valid_timezones(environment, valid_tz): template = environment.from_string( "{% now '" + valid_tz + "', '%Y-%m' %}" ) assert template.render() == '2015-12' Implement a test for environment.datetime_format
# -*- coding: utf-8 -*- import pytest from freezegun import freeze_time from jinja2 import Environment, exceptions @pytest.fixture def environment(): return Environment(extensions=['jinja2_time.TimeExtension']) @pytest.yield_fixture(autouse=True) def freeze(): freezer = freeze_time("2015-12-09 23:33:01") freezer.start() yield freezer.stop() def test_tz_is_required(environment): with pytest.raises(exceptions.TemplateSyntaxError): environment.from_string('{% now %}') def test_utc_default_datetime_format(environment): template = environment.from_string("{% now 'utc' %}") assert template.render() == "2015-12-09" @pytest.fixture(params=['utc', 'local', 'Europe/Berlin']) def valid_tz(request): return request.param def test_accept_valid_timezones(environment, valid_tz): template = environment.from_string( "{% now '" + valid_tz + "', '%Y-%m' %}" ) assert template.render() == '2015-12' def test_environment_datetime_format(environment): environment.datetime_format = '%a, %d %b %Y %H:%M:%S' template = environment.from_string("{% now 'utc' %}") assert template.render() == "Wed, 09 Dec 2015 23:33:01"
<commit_before># -*- coding: utf-8 -*- import pytest from freezegun import freeze_time from jinja2 import Environment, exceptions @pytest.fixture(scope='session') def environment(): return Environment(extensions=['jinja2_time.TimeExtension']) @pytest.yield_fixture(autouse=True) def freeze(): freezer = freeze_time("2015-12-09 23:33:01") freezer.start() yield freezer.stop() def test_tz_is_required(environment): with pytest.raises(exceptions.TemplateSyntaxError): environment.from_string('{% now %}') def test_utc_default_datetime_format(environment): template = environment.from_string("{% now 'utc' %}") assert template.render() == "2015-12-09" @pytest.fixture(params=['utc', 'local', 'Europe/Berlin']) def valid_tz(request): return request.param def test_accept_valid_timezones(environment, valid_tz): template = environment.from_string( "{% now '" + valid_tz + "', '%Y-%m' %}" ) assert template.render() == '2015-12' <commit_msg>Implement a test for environment.datetime_format<commit_after>
# -*- coding: utf-8 -*- import pytest from freezegun import freeze_time from jinja2 import Environment, exceptions @pytest.fixture def environment(): return Environment(extensions=['jinja2_time.TimeExtension']) @pytest.yield_fixture(autouse=True) def freeze(): freezer = freeze_time("2015-12-09 23:33:01") freezer.start() yield freezer.stop() def test_tz_is_required(environment): with pytest.raises(exceptions.TemplateSyntaxError): environment.from_string('{% now %}') def test_utc_default_datetime_format(environment): template = environment.from_string("{% now 'utc' %}") assert template.render() == "2015-12-09" @pytest.fixture(params=['utc', 'local', 'Europe/Berlin']) def valid_tz(request): return request.param def test_accept_valid_timezones(environment, valid_tz): template = environment.from_string( "{% now '" + valid_tz + "', '%Y-%m' %}" ) assert template.render() == '2015-12' def test_environment_datetime_format(environment): environment.datetime_format = '%a, %d %b %Y %H:%M:%S' template = environment.from_string("{% now 'utc' %}") assert template.render() == "Wed, 09 Dec 2015 23:33:01"
# -*- coding: utf-8 -*- import pytest from freezegun import freeze_time from jinja2 import Environment, exceptions @pytest.fixture(scope='session') def environment(): return Environment(extensions=['jinja2_time.TimeExtension']) @pytest.yield_fixture(autouse=True) def freeze(): freezer = freeze_time("2015-12-09 23:33:01") freezer.start() yield freezer.stop() def test_tz_is_required(environment): with pytest.raises(exceptions.TemplateSyntaxError): environment.from_string('{% now %}') def test_utc_default_datetime_format(environment): template = environment.from_string("{% now 'utc' %}") assert template.render() == "2015-12-09" @pytest.fixture(params=['utc', 'local', 'Europe/Berlin']) def valid_tz(request): return request.param def test_accept_valid_timezones(environment, valid_tz): template = environment.from_string( "{% now '" + valid_tz + "', '%Y-%m' %}" ) assert template.render() == '2015-12' Implement a test for environment.datetime_format# -*- coding: utf-8 -*- import pytest from freezegun import freeze_time from jinja2 import Environment, exceptions @pytest.fixture def environment(): return Environment(extensions=['jinja2_time.TimeExtension']) @pytest.yield_fixture(autouse=True) def freeze(): freezer = freeze_time("2015-12-09 23:33:01") freezer.start() yield freezer.stop() def test_tz_is_required(environment): with pytest.raises(exceptions.TemplateSyntaxError): environment.from_string('{% now %}') def test_utc_default_datetime_format(environment): template = environment.from_string("{% now 'utc' %}") assert template.render() == "2015-12-09" @pytest.fixture(params=['utc', 'local', 'Europe/Berlin']) def valid_tz(request): return request.param def test_accept_valid_timezones(environment, valid_tz): template = environment.from_string( "{% now '" + valid_tz + "', '%Y-%m' %}" ) assert template.render() == '2015-12' def test_environment_datetime_format(environment): environment.datetime_format = '%a, %d %b %Y %H:%M:%S' template = environment.from_string("{% now 'utc' %}") assert template.render() == "Wed, 09 Dec 2015 23:33:01"
<commit_before># -*- coding: utf-8 -*- import pytest from freezegun import freeze_time from jinja2 import Environment, exceptions @pytest.fixture(scope='session') def environment(): return Environment(extensions=['jinja2_time.TimeExtension']) @pytest.yield_fixture(autouse=True) def freeze(): freezer = freeze_time("2015-12-09 23:33:01") freezer.start() yield freezer.stop() def test_tz_is_required(environment): with pytest.raises(exceptions.TemplateSyntaxError): environment.from_string('{% now %}') def test_utc_default_datetime_format(environment): template = environment.from_string("{% now 'utc' %}") assert template.render() == "2015-12-09" @pytest.fixture(params=['utc', 'local', 'Europe/Berlin']) def valid_tz(request): return request.param def test_accept_valid_timezones(environment, valid_tz): template = environment.from_string( "{% now '" + valid_tz + "', '%Y-%m' %}" ) assert template.render() == '2015-12' <commit_msg>Implement a test for environment.datetime_format<commit_after># -*- coding: utf-8 -*- import pytest from freezegun import freeze_time from jinja2 import Environment, exceptions @pytest.fixture def environment(): return Environment(extensions=['jinja2_time.TimeExtension']) @pytest.yield_fixture(autouse=True) def freeze(): freezer = freeze_time("2015-12-09 23:33:01") freezer.start() yield freezer.stop() def test_tz_is_required(environment): with pytest.raises(exceptions.TemplateSyntaxError): environment.from_string('{% now %}') def test_utc_default_datetime_format(environment): template = environment.from_string("{% now 'utc' %}") assert template.render() == "2015-12-09" @pytest.fixture(params=['utc', 'local', 'Europe/Berlin']) def valid_tz(request): return request.param def test_accept_valid_timezones(environment, valid_tz): template = environment.from_string( "{% now '" + valid_tz + "', '%Y-%m' %}" ) assert template.render() == '2015-12' def test_environment_datetime_format(environment): environment.datetime_format = '%a, %d %b %Y %H:%M:%S' template = environment.from_string("{% now 'utc' %}") assert template.render() == "Wed, 09 Dec 2015 23:33:01"
0147e9a6c9d61028781b55f4f8e068e576b653a0
manage.py
manage.py
import os from flask_script import Manager from flask_migrate import Migrate, MigrateCommand from api import create_app, db application = create_app(os.getenv('FLASK_CONFIG') or 'default') migrate = Migrate(application, db) manager = Manager(application) manager.add_command('db', MigrateCommand) @manager.command def test(): """Run the unit tests.""" import unittest tests = unittest.TestLoader().discover('tests') unittest.TextTestRunner(verbosity=2).run(tests) @manager.command def recreate_db(): """ Recreates a local database. Drops and creates a new database """ db.drop_all() db.create_all() db.session.commit() if __name__ == "__main__": manager.run()
import os from flask_script import Manager from flask_migrate import Migrate, MigrateCommand from api import create_app, db application = create_app(os.getenv('FLASK_CONFIG') or 'default') migrate = Migrate(application, db) manager = Manager(application) manager.add_command('db', MigrateCommand) @manager.command def test(): """Run the unit tests.""" import unittest tests = unittest.TestLoader().discover('tests') unittest.TextTestRunner(verbosity=2).run(tests) @manager.command def recreate_db(): """ Recreates a local database. Drops and creates a new database """ db.drop_all() db.create_all() db.session.commit() @manager.command def create_db(): db.create_all() db.session.commit() @manager.command def drop_db(): db.drop_all() db.session.commit() if __name__ == "__main__": manager.run()
Add create and drop db tasks
Add create and drop db tasks
Python
mit
EdwinKato/bucket-list,EdwinKato/bucket-list,EdwinKato/bucket-list,EdwinKato/bucket-list,EdwinKato/bucket-list
import os from flask_script import Manager from flask_migrate import Migrate, MigrateCommand from api import create_app, db application = create_app(os.getenv('FLASK_CONFIG') or 'default') migrate = Migrate(application, db) manager = Manager(application) manager.add_command('db', MigrateCommand) @manager.command def test(): """Run the unit tests.""" import unittest tests = unittest.TestLoader().discover('tests') unittest.TextTestRunner(verbosity=2).run(tests) @manager.command def recreate_db(): """ Recreates a local database. Drops and creates a new database """ db.drop_all() db.create_all() db.session.commit() if __name__ == "__main__": manager.run() Add create and drop db tasks
import os from flask_script import Manager from flask_migrate import Migrate, MigrateCommand from api import create_app, db application = create_app(os.getenv('FLASK_CONFIG') or 'default') migrate = Migrate(application, db) manager = Manager(application) manager.add_command('db', MigrateCommand) @manager.command def test(): """Run the unit tests.""" import unittest tests = unittest.TestLoader().discover('tests') unittest.TextTestRunner(verbosity=2).run(tests) @manager.command def recreate_db(): """ Recreates a local database. Drops and creates a new database """ db.drop_all() db.create_all() db.session.commit() @manager.command def create_db(): db.create_all() db.session.commit() @manager.command def drop_db(): db.drop_all() db.session.commit() if __name__ == "__main__": manager.run()
<commit_before>import os from flask_script import Manager from flask_migrate import Migrate, MigrateCommand from api import create_app, db application = create_app(os.getenv('FLASK_CONFIG') or 'default') migrate = Migrate(application, db) manager = Manager(application) manager.add_command('db', MigrateCommand) @manager.command def test(): """Run the unit tests.""" import unittest tests = unittest.TestLoader().discover('tests') unittest.TextTestRunner(verbosity=2).run(tests) @manager.command def recreate_db(): """ Recreates a local database. Drops and creates a new database """ db.drop_all() db.create_all() db.session.commit() if __name__ == "__main__": manager.run() <commit_msg>Add create and drop db tasks<commit_after>
import os from flask_script import Manager from flask_migrate import Migrate, MigrateCommand from api import create_app, db application = create_app(os.getenv('FLASK_CONFIG') or 'default') migrate = Migrate(application, db) manager = Manager(application) manager.add_command('db', MigrateCommand) @manager.command def test(): """Run the unit tests.""" import unittest tests = unittest.TestLoader().discover('tests') unittest.TextTestRunner(verbosity=2).run(tests) @manager.command def recreate_db(): """ Recreates a local database. Drops and creates a new database """ db.drop_all() db.create_all() db.session.commit() @manager.command def create_db(): db.create_all() db.session.commit() @manager.command def drop_db(): db.drop_all() db.session.commit() if __name__ == "__main__": manager.run()
import os from flask_script import Manager from flask_migrate import Migrate, MigrateCommand from api import create_app, db application = create_app(os.getenv('FLASK_CONFIG') or 'default') migrate = Migrate(application, db) manager = Manager(application) manager.add_command('db', MigrateCommand) @manager.command def test(): """Run the unit tests.""" import unittest tests = unittest.TestLoader().discover('tests') unittest.TextTestRunner(verbosity=2).run(tests) @manager.command def recreate_db(): """ Recreates a local database. Drops and creates a new database """ db.drop_all() db.create_all() db.session.commit() if __name__ == "__main__": manager.run() Add create and drop db tasksimport os from flask_script import Manager from flask_migrate import Migrate, MigrateCommand from api import create_app, db application = create_app(os.getenv('FLASK_CONFIG') or 'default') migrate = Migrate(application, db) manager = Manager(application) manager.add_command('db', MigrateCommand) @manager.command def test(): """Run the unit tests.""" import unittest tests = unittest.TestLoader().discover('tests') unittest.TextTestRunner(verbosity=2).run(tests) @manager.command def recreate_db(): """ Recreates a local database. Drops and creates a new database """ db.drop_all() db.create_all() db.session.commit() @manager.command def create_db(): db.create_all() db.session.commit() @manager.command def drop_db(): db.drop_all() db.session.commit() if __name__ == "__main__": manager.run()
<commit_before>import os from flask_script import Manager from flask_migrate import Migrate, MigrateCommand from api import create_app, db application = create_app(os.getenv('FLASK_CONFIG') or 'default') migrate = Migrate(application, db) manager = Manager(application) manager.add_command('db', MigrateCommand) @manager.command def test(): """Run the unit tests.""" import unittest tests = unittest.TestLoader().discover('tests') unittest.TextTestRunner(verbosity=2).run(tests) @manager.command def recreate_db(): """ Recreates a local database. Drops and creates a new database """ db.drop_all() db.create_all() db.session.commit() if __name__ == "__main__": manager.run() <commit_msg>Add create and drop db tasks<commit_after>import os from flask_script import Manager from flask_migrate import Migrate, MigrateCommand from api import create_app, db application = create_app(os.getenv('FLASK_CONFIG') or 'default') migrate = Migrate(application, db) manager = Manager(application) manager.add_command('db', MigrateCommand) @manager.command def test(): """Run the unit tests.""" import unittest tests = unittest.TestLoader().discover('tests') unittest.TextTestRunner(verbosity=2).run(tests) @manager.command def recreate_db(): """ Recreates a local database. Drops and creates a new database """ db.drop_all() db.create_all() db.session.commit() @manager.command def create_db(): db.create_all() db.session.commit() @manager.command def drop_db(): db.drop_all() db.session.commit() if __name__ == "__main__": manager.run()