commit
stringlengths
40
40
old_file
stringlengths
4
118
new_file
stringlengths
4
118
old_contents
stringlengths
0
2.94k
new_contents
stringlengths
1
4.43k
subject
stringlengths
15
444
message
stringlengths
16
3.45k
lang
stringclasses
1 value
license
stringclasses
13 values
repos
stringlengths
5
43.2k
prompt
stringlengths
17
4.58k
response
stringlengths
1
4.43k
prompt_tagged
stringlengths
58
4.62k
response_tagged
stringlengths
1
4.43k
text
stringlengths
132
7.29k
text_tagged
stringlengths
173
7.33k
1cfdf9b1c11da15adb1e1603c815b76a4a286b1a
searchlogger/searchlogger/settings/production.py
searchlogger/searchlogger/settings/production.py
#! /usr/bin/env python # -*- coding: utf-8 -*- from defaults import * # noqa import json DEBUG = False TEMPLATE_DEBUG = False ALLOWED_HOSTS = ['.searchlogger.tutorons.com'] # Read in the Postgres database configuration from a file DATABASE_CONFIG_FILENAME = os.path.join( os.path.abspath(os.sep), # root directory 'etc', 'django', 'searchlogger', 'database_config.json' ) with open(DATABASE_CONFIG_FILENAME) as database_config_file: database_config = json.load(databse_config_file) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.postgresql', 'NAME': database_config['name'], 'USER': database_config['user'], 'PASSWORD': database_config['password'], 'HOST': database_config['host'], 'PORT': database_config['port'], } }
#! /usr/bin/env python # -*- coding: utf-8 -*- from defaults import * # noqa import json DEBUG = False TEMPLATE_DEBUG = False ALLOWED_HOSTS = ['.searchlogger.tutorons.com'] # Read in the Postgres database configuration from a file DATABASE_CONFIG_FILENAME = os.path.join(BASE_DIR, 'database_config.json') with open(DATABASE_CONFIG_FILENAME) as database_config_file: database_config = json.load(databse_config_file) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.postgresql', 'NAME': database_config['name'], 'USER': database_config['user'], 'PASSWORD': database_config['password'], 'HOST': database_config['host'], 'PORT': database_config['port'], } }
Read database configuration from base directory
Settings: Read database configuration from base directory
Python
mit
andrewhead/Search-Task-Logger,andrewhead/Search-Task-Logger,andrewhead/Search-Task-Logger
#! /usr/bin/env python # -*- coding: utf-8 -*- from defaults import * # noqa import json DEBUG = False TEMPLATE_DEBUG = False ALLOWED_HOSTS = ['.searchlogger.tutorons.com'] # Read in the Postgres database configuration from a file DATABASE_CONFIG_FILENAME = os.path.join( os.path.abspath(os.sep), # root directory 'etc', 'django', 'searchlogger', 'database_config.json' ) with open(DATABASE_CONFIG_FILENAME) as database_config_file: database_config = json.load(databse_config_file) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.postgresql', 'NAME': database_config['name'], 'USER': database_config['user'], 'PASSWORD': database_config['password'], 'HOST': database_config['host'], 'PORT': database_config['port'], } } Settings: Read database configuration from base directory
#! /usr/bin/env python # -*- coding: utf-8 -*- from defaults import * # noqa import json DEBUG = False TEMPLATE_DEBUG = False ALLOWED_HOSTS = ['.searchlogger.tutorons.com'] # Read in the Postgres database configuration from a file DATABASE_CONFIG_FILENAME = os.path.join(BASE_DIR, 'database_config.json') with open(DATABASE_CONFIG_FILENAME) as database_config_file: database_config = json.load(databse_config_file) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.postgresql', 'NAME': database_config['name'], 'USER': database_config['user'], 'PASSWORD': database_config['password'], 'HOST': database_config['host'], 'PORT': database_config['port'], } }
<commit_before>#! /usr/bin/env python # -*- coding: utf-8 -*- from defaults import * # noqa import json DEBUG = False TEMPLATE_DEBUG = False ALLOWED_HOSTS = ['.searchlogger.tutorons.com'] # Read in the Postgres database configuration from a file DATABASE_CONFIG_FILENAME = os.path.join( os.path.abspath(os.sep), # root directory 'etc', 'django', 'searchlogger', 'database_config.json' ) with open(DATABASE_CONFIG_FILENAME) as database_config_file: database_config = json.load(databse_config_file) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.postgresql', 'NAME': database_config['name'], 'USER': database_config['user'], 'PASSWORD': database_config['password'], 'HOST': database_config['host'], 'PORT': database_config['port'], } } <commit_msg>Settings: Read database configuration from base directory<commit_after>
#! /usr/bin/env python # -*- coding: utf-8 -*- from defaults import * # noqa import json DEBUG = False TEMPLATE_DEBUG = False ALLOWED_HOSTS = ['.searchlogger.tutorons.com'] # Read in the Postgres database configuration from a file DATABASE_CONFIG_FILENAME = os.path.join(BASE_DIR, 'database_config.json') with open(DATABASE_CONFIG_FILENAME) as database_config_file: database_config = json.load(databse_config_file) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.postgresql', 'NAME': database_config['name'], 'USER': database_config['user'], 'PASSWORD': database_config['password'], 'HOST': database_config['host'], 'PORT': database_config['port'], } }
#! /usr/bin/env python # -*- coding: utf-8 -*- from defaults import * # noqa import json DEBUG = False TEMPLATE_DEBUG = False ALLOWED_HOSTS = ['.searchlogger.tutorons.com'] # Read in the Postgres database configuration from a file DATABASE_CONFIG_FILENAME = os.path.join( os.path.abspath(os.sep), # root directory 'etc', 'django', 'searchlogger', 'database_config.json' ) with open(DATABASE_CONFIG_FILENAME) as database_config_file: database_config = json.load(databse_config_file) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.postgresql', 'NAME': database_config['name'], 'USER': database_config['user'], 'PASSWORD': database_config['password'], 'HOST': database_config['host'], 'PORT': database_config['port'], } } Settings: Read database configuration from base directory#! /usr/bin/env python # -*- coding: utf-8 -*- from defaults import * # noqa import json DEBUG = False TEMPLATE_DEBUG = False ALLOWED_HOSTS = ['.searchlogger.tutorons.com'] # Read in the Postgres database configuration from a file DATABASE_CONFIG_FILENAME = os.path.join(BASE_DIR, 'database_config.json') with open(DATABASE_CONFIG_FILENAME) as database_config_file: database_config = json.load(databse_config_file) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.postgresql', 'NAME': database_config['name'], 'USER': database_config['user'], 'PASSWORD': database_config['password'], 'HOST': database_config['host'], 'PORT': database_config['port'], } }
<commit_before>#! /usr/bin/env python # -*- coding: utf-8 -*- from defaults import * # noqa import json DEBUG = False TEMPLATE_DEBUG = False ALLOWED_HOSTS = ['.searchlogger.tutorons.com'] # Read in the Postgres database configuration from a file DATABASE_CONFIG_FILENAME = os.path.join( os.path.abspath(os.sep), # root directory 'etc', 'django', 'searchlogger', 'database_config.json' ) with open(DATABASE_CONFIG_FILENAME) as database_config_file: database_config = json.load(databse_config_file) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.postgresql', 'NAME': database_config['name'], 'USER': database_config['user'], 'PASSWORD': database_config['password'], 'HOST': database_config['host'], 'PORT': database_config['port'], } } <commit_msg>Settings: Read database configuration from base directory<commit_after>#! /usr/bin/env python # -*- coding: utf-8 -*- from defaults import * # noqa import json DEBUG = False TEMPLATE_DEBUG = False ALLOWED_HOSTS = ['.searchlogger.tutorons.com'] # Read in the Postgres database configuration from a file DATABASE_CONFIG_FILENAME = os.path.join(BASE_DIR, 'database_config.json') with open(DATABASE_CONFIG_FILENAME) as database_config_file: database_config = json.load(databse_config_file) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.postgresql', 'NAME': database_config['name'], 'USER': database_config['user'], 'PASSWORD': database_config['password'], 'HOST': database_config['host'], 'PORT': database_config['port'], } }
14ff06097a72dc65a351bb6a8bf59963412d2f41
semillas_backend/users/serializers.py
semillas_backend/users/serializers.py
#from phonenumber_field.serializerfields import PhoneNumberField from rest_framework import serializers from drf_extra_fields.geo_fields import PointField from .models import User class UserSerializer(serializers.ModelSerializer): """ Usage: from rest_framework.renderers import JSONRenderer from semillas_backend.users.serializers import UserSerializer JSONRenderer().render(UserSerializer(user_instance).data) """ location = PointField() class Meta: model = User fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login') class UpdateUserSerializer(serializers.ModelSerializer): name = serializers.CharField(required=False) #phone = PhoneNumberField(required=False) email = serializers.CharField(required=False) picture = serializers.ImageField(required=False) class Meta: model = User fields = ('name', 'picture', 'phone', 'email') from wallet.serializers import WalletSerializer class FullUserSerializer(UserSerializer): wallet = WalletSerializer() class Meta: model = User fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'wallet', 'email')
#from phonenumber_field.serializerfields import PhoneNumberField from rest_framework import serializers from drf_extra_fields.geo_fields import PointField from .models import User class UserSerializer(serializers.ModelSerializer): """ Usage: from rest_framework.renderers import JSONRenderer from semillas_backend.users.serializers import UserSerializer JSONRenderer().render(UserSerializer(user_instance).data) """ location = PointField() class Meta: model = User fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'phone') class UpdateUserSerializer(serializers.ModelSerializer): name = serializers.CharField(required=False) #phone = PhoneNumberField(required=False) email = serializers.CharField(required=False) picture = serializers.ImageField(required=False) class Meta: model = User fields = ('name', 'picture', 'phone', 'email') from wallet.serializers import WalletSerializer class FullUserSerializer(UserSerializer): wallet = WalletSerializer() class Meta: model = User fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'wallet', 'email')
Add phone to user serializer
Add phone to user serializer
Python
mit
Semillas/semillas_platform,Semillas/semillas_backend,Semillas/semillas_backend,Semillas/semillas_backend,Semillas/semillas_platform,Semillas/semillas_platform,Semillas/semillas_platform,Semillas/semillas_backend
#from phonenumber_field.serializerfields import PhoneNumberField from rest_framework import serializers from drf_extra_fields.geo_fields import PointField from .models import User class UserSerializer(serializers.ModelSerializer): """ Usage: from rest_framework.renderers import JSONRenderer from semillas_backend.users.serializers import UserSerializer JSONRenderer().render(UserSerializer(user_instance).data) """ location = PointField() class Meta: model = User fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login') class UpdateUserSerializer(serializers.ModelSerializer): name = serializers.CharField(required=False) #phone = PhoneNumberField(required=False) email = serializers.CharField(required=False) picture = serializers.ImageField(required=False) class Meta: model = User fields = ('name', 'picture', 'phone', 'email') from wallet.serializers import WalletSerializer class FullUserSerializer(UserSerializer): wallet = WalletSerializer() class Meta: model = User fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'wallet', 'email') Add phone to user serializer
#from phonenumber_field.serializerfields import PhoneNumberField from rest_framework import serializers from drf_extra_fields.geo_fields import PointField from .models import User class UserSerializer(serializers.ModelSerializer): """ Usage: from rest_framework.renderers import JSONRenderer from semillas_backend.users.serializers import UserSerializer JSONRenderer().render(UserSerializer(user_instance).data) """ location = PointField() class Meta: model = User fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'phone') class UpdateUserSerializer(serializers.ModelSerializer): name = serializers.CharField(required=False) #phone = PhoneNumberField(required=False) email = serializers.CharField(required=False) picture = serializers.ImageField(required=False) class Meta: model = User fields = ('name', 'picture', 'phone', 'email') from wallet.serializers import WalletSerializer class FullUserSerializer(UserSerializer): wallet = WalletSerializer() class Meta: model = User fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'wallet', 'email')
<commit_before>#from phonenumber_field.serializerfields import PhoneNumberField from rest_framework import serializers from drf_extra_fields.geo_fields import PointField from .models import User class UserSerializer(serializers.ModelSerializer): """ Usage: from rest_framework.renderers import JSONRenderer from semillas_backend.users.serializers import UserSerializer JSONRenderer().render(UserSerializer(user_instance).data) """ location = PointField() class Meta: model = User fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login') class UpdateUserSerializer(serializers.ModelSerializer): name = serializers.CharField(required=False) #phone = PhoneNumberField(required=False) email = serializers.CharField(required=False) picture = serializers.ImageField(required=False) class Meta: model = User fields = ('name', 'picture', 'phone', 'email') from wallet.serializers import WalletSerializer class FullUserSerializer(UserSerializer): wallet = WalletSerializer() class Meta: model = User fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'wallet', 'email') <commit_msg>Add phone to user serializer<commit_after>
#from phonenumber_field.serializerfields import PhoneNumberField from rest_framework import serializers from drf_extra_fields.geo_fields import PointField from .models import User class UserSerializer(serializers.ModelSerializer): """ Usage: from rest_framework.renderers import JSONRenderer from semillas_backend.users.serializers import UserSerializer JSONRenderer().render(UserSerializer(user_instance).data) """ location = PointField() class Meta: model = User fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'phone') class UpdateUserSerializer(serializers.ModelSerializer): name = serializers.CharField(required=False) #phone = PhoneNumberField(required=False) email = serializers.CharField(required=False) picture = serializers.ImageField(required=False) class Meta: model = User fields = ('name', 'picture', 'phone', 'email') from wallet.serializers import WalletSerializer class FullUserSerializer(UserSerializer): wallet = WalletSerializer() class Meta: model = User fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'wallet', 'email')
#from phonenumber_field.serializerfields import PhoneNumberField from rest_framework import serializers from drf_extra_fields.geo_fields import PointField from .models import User class UserSerializer(serializers.ModelSerializer): """ Usage: from rest_framework.renderers import JSONRenderer from semillas_backend.users.serializers import UserSerializer JSONRenderer().render(UserSerializer(user_instance).data) """ location = PointField() class Meta: model = User fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login') class UpdateUserSerializer(serializers.ModelSerializer): name = serializers.CharField(required=False) #phone = PhoneNumberField(required=False) email = serializers.CharField(required=False) picture = serializers.ImageField(required=False) class Meta: model = User fields = ('name', 'picture', 'phone', 'email') from wallet.serializers import WalletSerializer class FullUserSerializer(UserSerializer): wallet = WalletSerializer() class Meta: model = User fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'wallet', 'email') Add phone to user serializer#from phonenumber_field.serializerfields import PhoneNumberField from rest_framework import serializers from drf_extra_fields.geo_fields import PointField from .models import User class UserSerializer(serializers.ModelSerializer): """ Usage: from rest_framework.renderers import JSONRenderer from semillas_backend.users.serializers import UserSerializer JSONRenderer().render(UserSerializer(user_instance).data) """ location = PointField() class Meta: model = User fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'phone') class UpdateUserSerializer(serializers.ModelSerializer): name = serializers.CharField(required=False) #phone = PhoneNumberField(required=False) email = serializers.CharField(required=False) picture = serializers.ImageField(required=False) class Meta: model = User fields = ('name', 'picture', 'phone', 'email') from wallet.serializers import WalletSerializer class FullUserSerializer(UserSerializer): wallet = WalletSerializer() class Meta: model = User fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'wallet', 'email')
<commit_before>#from phonenumber_field.serializerfields import PhoneNumberField from rest_framework import serializers from drf_extra_fields.geo_fields import PointField from .models import User class UserSerializer(serializers.ModelSerializer): """ Usage: from rest_framework.renderers import JSONRenderer from semillas_backend.users.serializers import UserSerializer JSONRenderer().render(UserSerializer(user_instance).data) """ location = PointField() class Meta: model = User fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login') class UpdateUserSerializer(serializers.ModelSerializer): name = serializers.CharField(required=False) #phone = PhoneNumberField(required=False) email = serializers.CharField(required=False) picture = serializers.ImageField(required=False) class Meta: model = User fields = ('name', 'picture', 'phone', 'email') from wallet.serializers import WalletSerializer class FullUserSerializer(UserSerializer): wallet = WalletSerializer() class Meta: model = User fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'wallet', 'email') <commit_msg>Add phone to user serializer<commit_after>#from phonenumber_field.serializerfields import PhoneNumberField from rest_framework import serializers from drf_extra_fields.geo_fields import PointField from .models import User class UserSerializer(serializers.ModelSerializer): """ Usage: from rest_framework.renderers import JSONRenderer from semillas_backend.users.serializers import UserSerializer JSONRenderer().render(UserSerializer(user_instance).data) """ location = PointField() class Meta: model = User fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'phone') class UpdateUserSerializer(serializers.ModelSerializer): name = serializers.CharField(required=False) #phone = PhoneNumberField(required=False) email = serializers.CharField(required=False) picture = serializers.ImageField(required=False) class Meta: model = User fields = ('name', 'picture', 'phone', 'email') from wallet.serializers import WalletSerializer class FullUserSerializer(UserSerializer): wallet = WalletSerializer() class Meta: model = User fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'wallet', 'email')
23d12b1c4b755c7d35406bf2428eefbd682ef68f
examples/xor-classifier.py
examples/xor-classifier.py
#!/usr/bin/env python # -*- coding: utf-8 -*- '''Example using the theanets package for learning the XOR relation.''' import climate import logging import numpy as np import theanets climate.enable_default_logging() X = np.array([[0.0, 0.0], [0.0, 1.0], [1.0, 0.0], [1.0, 1.0]]) Y = np.array([0, 1, 1, 0, ]) Xi = np.random.randint(0, 2, size=(256, 2)) train = [ (Xi + 0.1 * np.random.randn(*Xi.shape)).astype('f'), (Xi[:, 0] ^ Xi[:, 1]).astype('f')[:, None], ] e = theanets.Experiment(theanets.Regressor, layers=(2, 2, 1), learning_rate=0.1, momentum=0.5, patience=300) e.run(train, train) logging.info("Input:\n%s", X) logging.info("XOR output:\n%s", Y) logging.info("NN XOR predictions:\n%s", e.network(X.astype('f')))
#!/usr/bin/env python # -*- coding: utf-8 -*- '''Example using the theanets package for learning the XOR relation.''' import climate import logging import numpy as np import theanets climate.enable_default_logging() X = np.array([[0, 0], [0, 1], [1, 0], [1, 1]], dtype='f') Y = np.array([[0], [1], [1], [0]], dtype='f') e = theanets.Experiment(theanets.Regressor, layers=(2, 2, 1)) e.train([X, Y], optimize='rprop', min_improvement=0.2, patience=500) logging.info("Input:\n%s", X) logging.info("XOR output:\n%s", Y) logging.info("NN XOR predictions:\n%s", e.network(X.astype('f')).round(2))
Use rprop for xor example.
Use rprop for xor example.
Python
mit
lmjohns3/theanets,devdoer/theanets,chrinide/theanets
#!/usr/bin/env python # -*- coding: utf-8 -*- '''Example using the theanets package for learning the XOR relation.''' import climate import logging import numpy as np import theanets climate.enable_default_logging() X = np.array([[0.0, 0.0], [0.0, 1.0], [1.0, 0.0], [1.0, 1.0]]) Y = np.array([0, 1, 1, 0, ]) Xi = np.random.randint(0, 2, size=(256, 2)) train = [ (Xi + 0.1 * np.random.randn(*Xi.shape)).astype('f'), (Xi[:, 0] ^ Xi[:, 1]).astype('f')[:, None], ] e = theanets.Experiment(theanets.Regressor, layers=(2, 2, 1), learning_rate=0.1, momentum=0.5, patience=300) e.run(train, train) logging.info("Input:\n%s", X) logging.info("XOR output:\n%s", Y) logging.info("NN XOR predictions:\n%s", e.network(X.astype('f'))) Use rprop for xor example.
#!/usr/bin/env python # -*- coding: utf-8 -*- '''Example using the theanets package for learning the XOR relation.''' import climate import logging import numpy as np import theanets climate.enable_default_logging() X = np.array([[0, 0], [0, 1], [1, 0], [1, 1]], dtype='f') Y = np.array([[0], [1], [1], [0]], dtype='f') e = theanets.Experiment(theanets.Regressor, layers=(2, 2, 1)) e.train([X, Y], optimize='rprop', min_improvement=0.2, patience=500) logging.info("Input:\n%s", X) logging.info("XOR output:\n%s", Y) logging.info("NN XOR predictions:\n%s", e.network(X.astype('f')).round(2))
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- '''Example using the theanets package for learning the XOR relation.''' import climate import logging import numpy as np import theanets climate.enable_default_logging() X = np.array([[0.0, 0.0], [0.0, 1.0], [1.0, 0.0], [1.0, 1.0]]) Y = np.array([0, 1, 1, 0, ]) Xi = np.random.randint(0, 2, size=(256, 2)) train = [ (Xi + 0.1 * np.random.randn(*Xi.shape)).astype('f'), (Xi[:, 0] ^ Xi[:, 1]).astype('f')[:, None], ] e = theanets.Experiment(theanets.Regressor, layers=(2, 2, 1), learning_rate=0.1, momentum=0.5, patience=300) e.run(train, train) logging.info("Input:\n%s", X) logging.info("XOR output:\n%s", Y) logging.info("NN XOR predictions:\n%s", e.network(X.astype('f'))) <commit_msg>Use rprop for xor example.<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- '''Example using the theanets package for learning the XOR relation.''' import climate import logging import numpy as np import theanets climate.enable_default_logging() X = np.array([[0, 0], [0, 1], [1, 0], [1, 1]], dtype='f') Y = np.array([[0], [1], [1], [0]], dtype='f') e = theanets.Experiment(theanets.Regressor, layers=(2, 2, 1)) e.train([X, Y], optimize='rprop', min_improvement=0.2, patience=500) logging.info("Input:\n%s", X) logging.info("XOR output:\n%s", Y) logging.info("NN XOR predictions:\n%s", e.network(X.astype('f')).round(2))
#!/usr/bin/env python # -*- coding: utf-8 -*- '''Example using the theanets package for learning the XOR relation.''' import climate import logging import numpy as np import theanets climate.enable_default_logging() X = np.array([[0.0, 0.0], [0.0, 1.0], [1.0, 0.0], [1.0, 1.0]]) Y = np.array([0, 1, 1, 0, ]) Xi = np.random.randint(0, 2, size=(256, 2)) train = [ (Xi + 0.1 * np.random.randn(*Xi.shape)).astype('f'), (Xi[:, 0] ^ Xi[:, 1]).astype('f')[:, None], ] e = theanets.Experiment(theanets.Regressor, layers=(2, 2, 1), learning_rate=0.1, momentum=0.5, patience=300) e.run(train, train) logging.info("Input:\n%s", X) logging.info("XOR output:\n%s", Y) logging.info("NN XOR predictions:\n%s", e.network(X.astype('f'))) Use rprop for xor example.#!/usr/bin/env python # -*- coding: utf-8 -*- '''Example using the theanets package for learning the XOR relation.''' import climate import logging import numpy as np import theanets climate.enable_default_logging() X = np.array([[0, 0], [0, 1], [1, 0], [1, 1]], dtype='f') Y = np.array([[0], [1], [1], [0]], dtype='f') e = theanets.Experiment(theanets.Regressor, layers=(2, 2, 1)) e.train([X, Y], optimize='rprop', min_improvement=0.2, patience=500) logging.info("Input:\n%s", X) logging.info("XOR output:\n%s", Y) logging.info("NN XOR predictions:\n%s", e.network(X.astype('f')).round(2))
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- '''Example using the theanets package for learning the XOR relation.''' import climate import logging import numpy as np import theanets climate.enable_default_logging() X = np.array([[0.0, 0.0], [0.0, 1.0], [1.0, 0.0], [1.0, 1.0]]) Y = np.array([0, 1, 1, 0, ]) Xi = np.random.randint(0, 2, size=(256, 2)) train = [ (Xi + 0.1 * np.random.randn(*Xi.shape)).astype('f'), (Xi[:, 0] ^ Xi[:, 1]).astype('f')[:, None], ] e = theanets.Experiment(theanets.Regressor, layers=(2, 2, 1), learning_rate=0.1, momentum=0.5, patience=300) e.run(train, train) logging.info("Input:\n%s", X) logging.info("XOR output:\n%s", Y) logging.info("NN XOR predictions:\n%s", e.network(X.astype('f'))) <commit_msg>Use rprop for xor example.<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- '''Example using the theanets package for learning the XOR relation.''' import climate import logging import numpy as np import theanets climate.enable_default_logging() X = np.array([[0, 0], [0, 1], [1, 0], [1, 1]], dtype='f') Y = np.array([[0], [1], [1], [0]], dtype='f') e = theanets.Experiment(theanets.Regressor, layers=(2, 2, 1)) e.train([X, Y], optimize='rprop', min_improvement=0.2, patience=500) logging.info("Input:\n%s", X) logging.info("XOR output:\n%s", Y) logging.info("NN XOR predictions:\n%s", e.network(X.astype('f')).round(2))
94796ca0107e6c676e3905675290bbe147169717
hoppy/deploy.py
hoppy/deploy.py
from restkit import Resource from hoppy import api_key class Deploy(Resource): def __init__(self, use_ssl=False): self.api_key = api_key super(Deploy, self).__init__(self.host, follow_redirect=True) def check_configuration(self): if not self.api_key: raise HoptoadError('API Key cannot be blank') def request(self, *args, **kwargs): response = super(Deploy, self).request( api_key=self.api_key, *args, **kwargs) return response.body_string() def base_uri(self, use_ssl=False): base = 'http://hoptoadapp.com/deploys.txt' base = base.replace('http://', 'https://') if use_ssl else base return base def deploy(self, env, **kwargs): """ Optional parameters accepted by Hoptoad are: scm_revision scm_repository local_username """ params = {} params['deploy[rails_env]'] = env for key, value in kwargs: params['deploy[%s]' % key] = value return self.post(**params)
from hoppy.api import HoptoadResource class Deploy(HoptoadResource): def __init__(self, use_ssl=False): from hoppy import api_key self.api_key = api_key super(Deploy, self).__init__(use_ssl) def check_configuration(self): if not self.api_key: raise HoptoadError('API Key cannot be blank') def request(self, *args, **kwargs): response = super(Deploy, self).request( api_key=self.api_key, *args, **kwargs) return response def base_uri(self, use_ssl=False): base = 'http://hoptoadapp.com/deploys.txt' base = base.replace('http://', 'https://') if use_ssl else base return base def deploy(self, env, **kwargs): """ Optional parameters accepted by Hoptoad are: scm_revision scm_repository local_username """ params = {} params['deploy[rails_env]'] = env for key, value in kwargs.iteritems(): params['deploy[%s]' % key] = value return self.post(**params)
Test Deploy resource after reworking.
Test Deploy resource after reworking.
Python
mit
peplin/hoppy
from restkit import Resource from hoppy import api_key class Deploy(Resource): def __init__(self, use_ssl=False): self.api_key = api_key super(Deploy, self).__init__(self.host, follow_redirect=True) def check_configuration(self): if not self.api_key: raise HoptoadError('API Key cannot be blank') def request(self, *args, **kwargs): response = super(Deploy, self).request( api_key=self.api_key, *args, **kwargs) return response.body_string() def base_uri(self, use_ssl=False): base = 'http://hoptoadapp.com/deploys.txt' base = base.replace('http://', 'https://') if use_ssl else base return base def deploy(self, env, **kwargs): """ Optional parameters accepted by Hoptoad are: scm_revision scm_repository local_username """ params = {} params['deploy[rails_env]'] = env for key, value in kwargs: params['deploy[%s]' % key] = value return self.post(**params) Test Deploy resource after reworking.
from hoppy.api import HoptoadResource class Deploy(HoptoadResource): def __init__(self, use_ssl=False): from hoppy import api_key self.api_key = api_key super(Deploy, self).__init__(use_ssl) def check_configuration(self): if not self.api_key: raise HoptoadError('API Key cannot be blank') def request(self, *args, **kwargs): response = super(Deploy, self).request( api_key=self.api_key, *args, **kwargs) return response def base_uri(self, use_ssl=False): base = 'http://hoptoadapp.com/deploys.txt' base = base.replace('http://', 'https://') if use_ssl else base return base def deploy(self, env, **kwargs): """ Optional parameters accepted by Hoptoad are: scm_revision scm_repository local_username """ params = {} params['deploy[rails_env]'] = env for key, value in kwargs.iteritems(): params['deploy[%s]' % key] = value return self.post(**params)
<commit_before>from restkit import Resource from hoppy import api_key class Deploy(Resource): def __init__(self, use_ssl=False): self.api_key = api_key super(Deploy, self).__init__(self.host, follow_redirect=True) def check_configuration(self): if not self.api_key: raise HoptoadError('API Key cannot be blank') def request(self, *args, **kwargs): response = super(Deploy, self).request( api_key=self.api_key, *args, **kwargs) return response.body_string() def base_uri(self, use_ssl=False): base = 'http://hoptoadapp.com/deploys.txt' base = base.replace('http://', 'https://') if use_ssl else base return base def deploy(self, env, **kwargs): """ Optional parameters accepted by Hoptoad are: scm_revision scm_repository local_username """ params = {} params['deploy[rails_env]'] = env for key, value in kwargs: params['deploy[%s]' % key] = value return self.post(**params) <commit_msg>Test Deploy resource after reworking.<commit_after>
from hoppy.api import HoptoadResource class Deploy(HoptoadResource): def __init__(self, use_ssl=False): from hoppy import api_key self.api_key = api_key super(Deploy, self).__init__(use_ssl) def check_configuration(self): if not self.api_key: raise HoptoadError('API Key cannot be blank') def request(self, *args, **kwargs): response = super(Deploy, self).request( api_key=self.api_key, *args, **kwargs) return response def base_uri(self, use_ssl=False): base = 'http://hoptoadapp.com/deploys.txt' base = base.replace('http://', 'https://') if use_ssl else base return base def deploy(self, env, **kwargs): """ Optional parameters accepted by Hoptoad are: scm_revision scm_repository local_username """ params = {} params['deploy[rails_env]'] = env for key, value in kwargs.iteritems(): params['deploy[%s]' % key] = value return self.post(**params)
from restkit import Resource from hoppy import api_key class Deploy(Resource): def __init__(self, use_ssl=False): self.api_key = api_key super(Deploy, self).__init__(self.host, follow_redirect=True) def check_configuration(self): if not self.api_key: raise HoptoadError('API Key cannot be blank') def request(self, *args, **kwargs): response = super(Deploy, self).request( api_key=self.api_key, *args, **kwargs) return response.body_string() def base_uri(self, use_ssl=False): base = 'http://hoptoadapp.com/deploys.txt' base = base.replace('http://', 'https://') if use_ssl else base return base def deploy(self, env, **kwargs): """ Optional parameters accepted by Hoptoad are: scm_revision scm_repository local_username """ params = {} params['deploy[rails_env]'] = env for key, value in kwargs: params['deploy[%s]' % key] = value return self.post(**params) Test Deploy resource after reworking.from hoppy.api import HoptoadResource class Deploy(HoptoadResource): def __init__(self, use_ssl=False): from hoppy import api_key self.api_key = api_key super(Deploy, self).__init__(use_ssl) def check_configuration(self): if not self.api_key: raise HoptoadError('API Key cannot be blank') def request(self, *args, **kwargs): response = super(Deploy, self).request( api_key=self.api_key, *args, **kwargs) return response def base_uri(self, use_ssl=False): base = 'http://hoptoadapp.com/deploys.txt' base = base.replace('http://', 'https://') if use_ssl else base return base def deploy(self, env, **kwargs): """ Optional parameters accepted by Hoptoad are: scm_revision scm_repository local_username """ params = {} params['deploy[rails_env]'] = env for key, value in kwargs.iteritems(): params['deploy[%s]' % key] = value return self.post(**params)
<commit_before>from restkit import Resource from hoppy import api_key class Deploy(Resource): def __init__(self, use_ssl=False): self.api_key = api_key super(Deploy, self).__init__(self.host, follow_redirect=True) def check_configuration(self): if not self.api_key: raise HoptoadError('API Key cannot be blank') def request(self, *args, **kwargs): response = super(Deploy, self).request( api_key=self.api_key, *args, **kwargs) return response.body_string() def base_uri(self, use_ssl=False): base = 'http://hoptoadapp.com/deploys.txt' base = base.replace('http://', 'https://') if use_ssl else base return base def deploy(self, env, **kwargs): """ Optional parameters accepted by Hoptoad are: scm_revision scm_repository local_username """ params = {} params['deploy[rails_env]'] = env for key, value in kwargs: params['deploy[%s]' % key] = value return self.post(**params) <commit_msg>Test Deploy resource after reworking.<commit_after>from hoppy.api import HoptoadResource class Deploy(HoptoadResource): def __init__(self, use_ssl=False): from hoppy import api_key self.api_key = api_key super(Deploy, self).__init__(use_ssl) def check_configuration(self): if not self.api_key: raise HoptoadError('API Key cannot be blank') def request(self, *args, **kwargs): response = super(Deploy, self).request( api_key=self.api_key, *args, **kwargs) return response def base_uri(self, use_ssl=False): base = 'http://hoptoadapp.com/deploys.txt' base = base.replace('http://', 'https://') if use_ssl else base return base def deploy(self, env, **kwargs): """ Optional parameters accepted by Hoptoad are: scm_revision scm_repository local_username """ params = {} params['deploy[rails_env]'] = env for key, value in kwargs.iteritems(): params['deploy[%s]' % key] = value return self.post(**params)
9066d3e5bdbc95fb347b1a081d9b7db33ab68ea4
src/autobot/src/stopsign.py
src/autobot/src/stopsign.py
#!/usr/bin/env python import rospy class StopStates(object): NORMAL = 0 FULL_STOP = 1 IGNORE_STOP_SIGNS = 2 class StopSign(object): def __init__(self): self.state = StopStates.NORMAL self.stopDuration = 2 self.ignoreDuration = 2 def stopSignDetected(self): self.state = StopStates.FULL_STOP timer = rospy.Timer(rospy.Duration(self.stopDuration), self.stepStateMachine, oneshot=True) def stepStateMachine(self): if self.state is StopStates.NORMAL: self.action = StopStates.FULL_STOP elif self.state is StopStates.FULL_STOP: self.action = StopStates.IGNORE_STOP_SIGNS timer = rospy.Timer(rospy.Duration(self.ignoreDuration), self.stepStateMachine, oneshot=True) elif self.state is StopStates.IGNORE_STOP_SIGNS: self.action = StopStates.NORMAL
#!/usr/bin/env python import rospy class StopStates(object): NORMAL = 0 FULL_STOP = 1 IGNORE_STOP_SIGNS = 2 class StopSign(object): def __init__(self): self.state = StopStates.NORMAL self.stopDuration = 2 self.ignoreDuration = 4 def stopSignDetected(self): self.state = StopStates.FULL_STOP timer = rospy.Timer(rospy.Duration(self.stopDuration), self.stepStateMachine, oneshot=True) def stepStateMachine(self, event): if self.state is StopStates.NORMAL: self.state = StopStates.FULL_STOP elif self.state is StopStates.FULL_STOP: self.state = StopStates.IGNORE_STOP_SIGNS timer = rospy.Timer(rospy.Duration(self.ignoreDuration), self.stepStateMachine, oneshot=True) elif self.state is StopStates.IGNORE_STOP_SIGNS: self.state = StopStates.NORMAL
Fix state machine using wrong variable
Fix state machine using wrong variable
Python
mit
atkvo/masters-bot,atkvo/masters-bot,atkvo/masters-bot,atkvo/masters-bot,atkvo/masters-bot
#!/usr/bin/env python import rospy class StopStates(object): NORMAL = 0 FULL_STOP = 1 IGNORE_STOP_SIGNS = 2 class StopSign(object): def __init__(self): self.state = StopStates.NORMAL self.stopDuration = 2 self.ignoreDuration = 2 def stopSignDetected(self): self.state = StopStates.FULL_STOP timer = rospy.Timer(rospy.Duration(self.stopDuration), self.stepStateMachine, oneshot=True) def stepStateMachine(self): if self.state is StopStates.NORMAL: self.action = StopStates.FULL_STOP elif self.state is StopStates.FULL_STOP: self.action = StopStates.IGNORE_STOP_SIGNS timer = rospy.Timer(rospy.Duration(self.ignoreDuration), self.stepStateMachine, oneshot=True) elif self.state is StopStates.IGNORE_STOP_SIGNS: self.action = StopStates.NORMAL Fix state machine using wrong variable
#!/usr/bin/env python import rospy class StopStates(object): NORMAL = 0 FULL_STOP = 1 IGNORE_STOP_SIGNS = 2 class StopSign(object): def __init__(self): self.state = StopStates.NORMAL self.stopDuration = 2 self.ignoreDuration = 4 def stopSignDetected(self): self.state = StopStates.FULL_STOP timer = rospy.Timer(rospy.Duration(self.stopDuration), self.stepStateMachine, oneshot=True) def stepStateMachine(self, event): if self.state is StopStates.NORMAL: self.state = StopStates.FULL_STOP elif self.state is StopStates.FULL_STOP: self.state = StopStates.IGNORE_STOP_SIGNS timer = rospy.Timer(rospy.Duration(self.ignoreDuration), self.stepStateMachine, oneshot=True) elif self.state is StopStates.IGNORE_STOP_SIGNS: self.state = StopStates.NORMAL
<commit_before>#!/usr/bin/env python import rospy class StopStates(object): NORMAL = 0 FULL_STOP = 1 IGNORE_STOP_SIGNS = 2 class StopSign(object): def __init__(self): self.state = StopStates.NORMAL self.stopDuration = 2 self.ignoreDuration = 2 def stopSignDetected(self): self.state = StopStates.FULL_STOP timer = rospy.Timer(rospy.Duration(self.stopDuration), self.stepStateMachine, oneshot=True) def stepStateMachine(self): if self.state is StopStates.NORMAL: self.action = StopStates.FULL_STOP elif self.state is StopStates.FULL_STOP: self.action = StopStates.IGNORE_STOP_SIGNS timer = rospy.Timer(rospy.Duration(self.ignoreDuration), self.stepStateMachine, oneshot=True) elif self.state is StopStates.IGNORE_STOP_SIGNS: self.action = StopStates.NORMAL <commit_msg>Fix state machine using wrong variable<commit_after>
#!/usr/bin/env python import rospy class StopStates(object): NORMAL = 0 FULL_STOP = 1 IGNORE_STOP_SIGNS = 2 class StopSign(object): def __init__(self): self.state = StopStates.NORMAL self.stopDuration = 2 self.ignoreDuration = 4 def stopSignDetected(self): self.state = StopStates.FULL_STOP timer = rospy.Timer(rospy.Duration(self.stopDuration), self.stepStateMachine, oneshot=True) def stepStateMachine(self, event): if self.state is StopStates.NORMAL: self.state = StopStates.FULL_STOP elif self.state is StopStates.FULL_STOP: self.state = StopStates.IGNORE_STOP_SIGNS timer = rospy.Timer(rospy.Duration(self.ignoreDuration), self.stepStateMachine, oneshot=True) elif self.state is StopStates.IGNORE_STOP_SIGNS: self.state = StopStates.NORMAL
#!/usr/bin/env python import rospy class StopStates(object): NORMAL = 0 FULL_STOP = 1 IGNORE_STOP_SIGNS = 2 class StopSign(object): def __init__(self): self.state = StopStates.NORMAL self.stopDuration = 2 self.ignoreDuration = 2 def stopSignDetected(self): self.state = StopStates.FULL_STOP timer = rospy.Timer(rospy.Duration(self.stopDuration), self.stepStateMachine, oneshot=True) def stepStateMachine(self): if self.state is StopStates.NORMAL: self.action = StopStates.FULL_STOP elif self.state is StopStates.FULL_STOP: self.action = StopStates.IGNORE_STOP_SIGNS timer = rospy.Timer(rospy.Duration(self.ignoreDuration), self.stepStateMachine, oneshot=True) elif self.state is StopStates.IGNORE_STOP_SIGNS: self.action = StopStates.NORMAL Fix state machine using wrong variable#!/usr/bin/env python import rospy class StopStates(object): NORMAL = 0 FULL_STOP = 1 IGNORE_STOP_SIGNS = 2 class StopSign(object): def __init__(self): self.state = StopStates.NORMAL self.stopDuration = 2 self.ignoreDuration = 4 def stopSignDetected(self): self.state = StopStates.FULL_STOP timer = rospy.Timer(rospy.Duration(self.stopDuration), self.stepStateMachine, oneshot=True) def stepStateMachine(self, event): if self.state is StopStates.NORMAL: self.state = StopStates.FULL_STOP elif self.state is StopStates.FULL_STOP: self.state = StopStates.IGNORE_STOP_SIGNS timer = rospy.Timer(rospy.Duration(self.ignoreDuration), self.stepStateMachine, oneshot=True) elif self.state is StopStates.IGNORE_STOP_SIGNS: self.state = StopStates.NORMAL
<commit_before>#!/usr/bin/env python import rospy class StopStates(object): NORMAL = 0 FULL_STOP = 1 IGNORE_STOP_SIGNS = 2 class StopSign(object): def __init__(self): self.state = StopStates.NORMAL self.stopDuration = 2 self.ignoreDuration = 2 def stopSignDetected(self): self.state = StopStates.FULL_STOP timer = rospy.Timer(rospy.Duration(self.stopDuration), self.stepStateMachine, oneshot=True) def stepStateMachine(self): if self.state is StopStates.NORMAL: self.action = StopStates.FULL_STOP elif self.state is StopStates.FULL_STOP: self.action = StopStates.IGNORE_STOP_SIGNS timer = rospy.Timer(rospy.Duration(self.ignoreDuration), self.stepStateMachine, oneshot=True) elif self.state is StopStates.IGNORE_STOP_SIGNS: self.action = StopStates.NORMAL <commit_msg>Fix state machine using wrong variable<commit_after>#!/usr/bin/env python import rospy class StopStates(object): NORMAL = 0 FULL_STOP = 1 IGNORE_STOP_SIGNS = 2 class StopSign(object): def __init__(self): self.state = StopStates.NORMAL self.stopDuration = 2 self.ignoreDuration = 4 def stopSignDetected(self): self.state = StopStates.FULL_STOP timer = rospy.Timer(rospy.Duration(self.stopDuration), self.stepStateMachine, oneshot=True) def stepStateMachine(self, event): if self.state is StopStates.NORMAL: self.state = StopStates.FULL_STOP elif self.state is StopStates.FULL_STOP: self.state = StopStates.IGNORE_STOP_SIGNS timer = rospy.Timer(rospy.Duration(self.ignoreDuration), self.stepStateMachine, oneshot=True) elif self.state is StopStates.IGNORE_STOP_SIGNS: self.state = StopStates.NORMAL
2c5a1bebf805c9bf5208fc75c32d8998b865eb32
designate/objects/zone_transfer_request.py
designate/objects/zone_transfer_request.py
# Copyright 2014 Hewlett-Packard Development Company, L.P. # # Author: Graham Hayes <graham.hayes@hp.com> # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from designate.objects import base class ZoneTransferRequest(base.DictObjectMixin, base.DesignateObject, base.PersistentObjectMixin): FIELDS = { 'domain_id': {}, 'key': {}, 'description': {}, 'tenant_id': {}, 'target_tenant_id': {}, 'status': {}, 'id': {}, 'created_at': {}, 'domain_name': {}, 'updated_at': {}, 'version': {}, } class ZoneTransferRequestList(base.ListObjectMixin, base.DesignateObject): LIST_ITEM_TYPE = ZoneTransferRequest
# Copyright 2014 Hewlett-Packard Development Company, L.P. # # Author: Graham Hayes <graham.hayes@hp.com> # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from designate.objects import base class ZoneTransferRequest(base.DictObjectMixin, base.PersistentObjectMixin, base.DesignateObject,): FIELDS = { 'domain_id': {}, 'key': {}, 'description': {}, 'tenant_id': {}, 'target_tenant_id': {}, 'status': {}, 'domain_name': {}, } class ZoneTransferRequestList(base.ListObjectMixin, base.DesignateObject): LIST_ITEM_TYPE = ZoneTransferRequest
Remove duplicate fields from ZoneTransferRequest object
Remove duplicate fields from ZoneTransferRequest object The fields id, version, created_at, updated_at are defined in the PersistentObjectMixin which ZoneTransferRequest extends, so this patch removes them from ZoneTransferRequest. Change-Id: Iff20a31b4a208bff0bc879677a9901fedc43226b Closes-Bug: #1403274
Python
apache-2.0
kiall/designate-py3,muraliselva10/designate,openstack/designate,kiall/designate-py3,ramsateesh/designate,kiall/designate-py3,openstack/designate,muraliselva10/designate,cneill/designate,tonyli71/designate,cneill/designate,cneill/designate-testing,cneill/designate-testing,cneill/designate,tonyli71/designate,muraliselva10/designate,ionrock/designate,grahamhayes/designate,cneill/designate,ramsateesh/designate,tonyli71/designate,ionrock/designate,openstack/designate,cneill/designate-testing,kiall/designate-py3,cneill/designate,ramsateesh/designate,kiall/designate-py3,ionrock/designate,grahamhayes/designate,grahamhayes/designate
# Copyright 2014 Hewlett-Packard Development Company, L.P. # # Author: Graham Hayes <graham.hayes@hp.com> # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from designate.objects import base class ZoneTransferRequest(base.DictObjectMixin, base.DesignateObject, base.PersistentObjectMixin): FIELDS = { 'domain_id': {}, 'key': {}, 'description': {}, 'tenant_id': {}, 'target_tenant_id': {}, 'status': {}, 'id': {}, 'created_at': {}, 'domain_name': {}, 'updated_at': {}, 'version': {}, } class ZoneTransferRequestList(base.ListObjectMixin, base.DesignateObject): LIST_ITEM_TYPE = ZoneTransferRequest Remove duplicate fields from ZoneTransferRequest object The fields id, version, created_at, updated_at are defined in the PersistentObjectMixin which ZoneTransferRequest extends, so this patch removes them from ZoneTransferRequest. Change-Id: Iff20a31b4a208bff0bc879677a9901fedc43226b Closes-Bug: #1403274
# Copyright 2014 Hewlett-Packard Development Company, L.P. # # Author: Graham Hayes <graham.hayes@hp.com> # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from designate.objects import base class ZoneTransferRequest(base.DictObjectMixin, base.PersistentObjectMixin, base.DesignateObject,): FIELDS = { 'domain_id': {}, 'key': {}, 'description': {}, 'tenant_id': {}, 'target_tenant_id': {}, 'status': {}, 'domain_name': {}, } class ZoneTransferRequestList(base.ListObjectMixin, base.DesignateObject): LIST_ITEM_TYPE = ZoneTransferRequest
<commit_before># Copyright 2014 Hewlett-Packard Development Company, L.P. # # Author: Graham Hayes <graham.hayes@hp.com> # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from designate.objects import base class ZoneTransferRequest(base.DictObjectMixin, base.DesignateObject, base.PersistentObjectMixin): FIELDS = { 'domain_id': {}, 'key': {}, 'description': {}, 'tenant_id': {}, 'target_tenant_id': {}, 'status': {}, 'id': {}, 'created_at': {}, 'domain_name': {}, 'updated_at': {}, 'version': {}, } class ZoneTransferRequestList(base.ListObjectMixin, base.DesignateObject): LIST_ITEM_TYPE = ZoneTransferRequest <commit_msg>Remove duplicate fields from ZoneTransferRequest object The fields id, version, created_at, updated_at are defined in the PersistentObjectMixin which ZoneTransferRequest extends, so this patch removes them from ZoneTransferRequest. Change-Id: Iff20a31b4a208bff0bc879677a9901fedc43226b Closes-Bug: #1403274<commit_after>
# Copyright 2014 Hewlett-Packard Development Company, L.P. # # Author: Graham Hayes <graham.hayes@hp.com> # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from designate.objects import base class ZoneTransferRequest(base.DictObjectMixin, base.PersistentObjectMixin, base.DesignateObject,): FIELDS = { 'domain_id': {}, 'key': {}, 'description': {}, 'tenant_id': {}, 'target_tenant_id': {}, 'status': {}, 'domain_name': {}, } class ZoneTransferRequestList(base.ListObjectMixin, base.DesignateObject): LIST_ITEM_TYPE = ZoneTransferRequest
# Copyright 2014 Hewlett-Packard Development Company, L.P. # # Author: Graham Hayes <graham.hayes@hp.com> # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from designate.objects import base class ZoneTransferRequest(base.DictObjectMixin, base.DesignateObject, base.PersistentObjectMixin): FIELDS = { 'domain_id': {}, 'key': {}, 'description': {}, 'tenant_id': {}, 'target_tenant_id': {}, 'status': {}, 'id': {}, 'created_at': {}, 'domain_name': {}, 'updated_at': {}, 'version': {}, } class ZoneTransferRequestList(base.ListObjectMixin, base.DesignateObject): LIST_ITEM_TYPE = ZoneTransferRequest Remove duplicate fields from ZoneTransferRequest object The fields id, version, created_at, updated_at are defined in the PersistentObjectMixin which ZoneTransferRequest extends, so this patch removes them from ZoneTransferRequest. Change-Id: Iff20a31b4a208bff0bc879677a9901fedc43226b Closes-Bug: #1403274# Copyright 2014 Hewlett-Packard Development Company, L.P. # # Author: Graham Hayes <graham.hayes@hp.com> # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from designate.objects import base class ZoneTransferRequest(base.DictObjectMixin, base.PersistentObjectMixin, base.DesignateObject,): FIELDS = { 'domain_id': {}, 'key': {}, 'description': {}, 'tenant_id': {}, 'target_tenant_id': {}, 'status': {}, 'domain_name': {}, } class ZoneTransferRequestList(base.ListObjectMixin, base.DesignateObject): LIST_ITEM_TYPE = ZoneTransferRequest
<commit_before># Copyright 2014 Hewlett-Packard Development Company, L.P. # # Author: Graham Hayes <graham.hayes@hp.com> # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from designate.objects import base class ZoneTransferRequest(base.DictObjectMixin, base.DesignateObject, base.PersistentObjectMixin): FIELDS = { 'domain_id': {}, 'key': {}, 'description': {}, 'tenant_id': {}, 'target_tenant_id': {}, 'status': {}, 'id': {}, 'created_at': {}, 'domain_name': {}, 'updated_at': {}, 'version': {}, } class ZoneTransferRequestList(base.ListObjectMixin, base.DesignateObject): LIST_ITEM_TYPE = ZoneTransferRequest <commit_msg>Remove duplicate fields from ZoneTransferRequest object The fields id, version, created_at, updated_at are defined in the PersistentObjectMixin which ZoneTransferRequest extends, so this patch removes them from ZoneTransferRequest. Change-Id: Iff20a31b4a208bff0bc879677a9901fedc43226b Closes-Bug: #1403274<commit_after># Copyright 2014 Hewlett-Packard Development Company, L.P. # # Author: Graham Hayes <graham.hayes@hp.com> # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from designate.objects import base class ZoneTransferRequest(base.DictObjectMixin, base.PersistentObjectMixin, base.DesignateObject,): FIELDS = { 'domain_id': {}, 'key': {}, 'description': {}, 'tenant_id': {}, 'target_tenant_id': {}, 'status': {}, 'domain_name': {}, } class ZoneTransferRequestList(base.ListObjectMixin, base.DesignateObject): LIST_ITEM_TYPE = ZoneTransferRequest
8441acfd5071e8b63fde816f67e167997045d510
Lib/misc/setup.py
Lib/misc/setup.py
import os from numpy.distutils.misc_util import Configuration def configuration(parent_package='',top_path=None): config = Configuration('misc',parent_package, top_path) config.add_data_files('lena.dat') print "########", config return config if __name__ == '__main__': from numpy.distutils.core import setup setup(**configuration())
import os from numpy.distutils.misc_util import Configuration def configuration(parent_package='',top_path=None): config = Configuration('misc',parent_package, top_path) config.add_data_files('lena.dat') return config if __name__ == '__main__': from numpy.distutils.core import setup setup(**configuration())
Remove extra noise on install.
Remove extra noise on install.
Python
bsd-3-clause
jseabold/scipy,richardotis/scipy,anntzer/scipy,fredrikw/scipy,behzadnouri/scipy,aman-iitj/scipy,mortada/scipy,njwilson23/scipy,trankmichael/scipy,trankmichael/scipy,apbard/scipy,niknow/scipy,aman-iitj/scipy,behzadnouri/scipy,FRidh/scipy,vanpact/scipy,Eric89GXL/scipy,rmcgibbo/scipy,larsmans/scipy,Shaswat27/scipy,ogrisel/scipy,niknow/scipy,chatcannon/scipy,mhogg/scipy,petebachant/scipy,jor-/scipy,nonhermitian/scipy,mgaitan/scipy,chatcannon/scipy,Eric89GXL/scipy,gdooper/scipy,jonycgn/scipy,FRidh/scipy,fernand/scipy,juliantaylor/scipy,apbard/scipy,vigna/scipy,sargas/scipy,pschella/scipy,mhogg/scipy,chatcannon/scipy,Kamp9/scipy,juliantaylor/scipy,behzadnouri/scipy,surhudm/scipy,ales-erjavec/scipy,kalvdans/scipy,gertingold/scipy,piyush0609/scipy,surhudm/scipy,nonhermitian/scipy,mgaitan/scipy,pbrod/scipy,mortonjt/scipy,rmcgibbo/scipy,matthew-brett/scipy,aeklant/scipy,fredrikw/scipy,lukauskas/scipy,sonnyhu/scipy,raoulbq/scipy,sonnyhu/scipy,njwilson23/scipy,haudren/scipy,pyramania/scipy,zerothi/scipy,arokem/scipy,mikebenfield/scipy,minhlongdo/scipy,grlee77/scipy,gdooper/scipy,bkendzior/scipy,anntzer/scipy,dch312/scipy,grlee77/scipy,fredrikw/scipy,haudren/scipy,zerothi/scipy,witcxc/scipy,mtrbean/scipy,sriki18/scipy,jakevdp/scipy,sonnyhu/scipy,pbrod/scipy,giorgiop/scipy,vhaasteren/scipy,ogrisel/scipy,gertingold/scipy,nvoron23/scipy,andyfaff/scipy,andim/scipy,jamestwebber/scipy,maniteja123/scipy,FRidh/scipy,Kamp9/scipy,jjhelmus/scipy,WillieMaddox/scipy,nmayorov/scipy,Gillu13/scipy,jjhelmus/scipy,Eric89GXL/scipy,woodscn/scipy,sriki18/scipy,pizzathief/scipy,newemailjdm/scipy,ChanderG/scipy,felipebetancur/scipy,Stefan-Endres/scipy,gfyoung/scipy,endolith/scipy,ales-erjavec/scipy,giorgiop/scipy,WarrenWeckesser/scipy,surhudm/scipy,sriki18/scipy,Eric89GXL/scipy,anntzer/scipy,kleskjr/scipy,vberaudi/scipy,rgommers/scipy,nvoron23/scipy,pyramania/scipy,kalvdans/scipy,rgommers/scipy,felipebetancur/scipy,Newman101/scipy,ortylp/scipy,dominicelse/scipy,aarchiba/scipy,vhaasteren/scipy,endolith/scipy,nonhermitian/scipy,pizzathief/scipy,surhudm/scipy,Dapid/scipy,jseabold/scipy,kleskjr/scipy,Newman101/scipy,andyfaff/scipy,aman-iitj/scipy,tylerjereddy/scipy,mikebenfield/scipy,futurulus/scipy,kalvdans/scipy,teoliphant/scipy,maciejkula/scipy,mingwpy/scipy,jor-/scipy,ChanderG/scipy,mtrbean/scipy,endolith/scipy,cpaulik/scipy,surhudm/scipy,rgommers/scipy,mortonjt/scipy,perimosocordiae/scipy,rmcgibbo/scipy,mdhaber/scipy,Shaswat27/scipy,woodscn/scipy,jakevdp/scipy,maciejkula/scipy,andim/scipy,WarrenWeckesser/scipy,rgommers/scipy,Shaswat27/scipy,rmcgibbo/scipy,argriffing/scipy,jonycgn/scipy,gef756/scipy,befelix/scipy,andim/scipy,Srisai85/scipy,minhlongdo/scipy,ChanderG/scipy,sauliusl/scipy,sauliusl/scipy,trankmichael/scipy,ilayn/scipy,kleskjr/scipy,newemailjdm/scipy,gdooper/scipy,ortylp/scipy,vhaasteren/scipy,juliantaylor/scipy,josephcslater/scipy,lukauskas/scipy,teoliphant/scipy,Gillu13/scipy,vigna/scipy,zxsted/scipy,jamestwebber/scipy,apbard/scipy,pnedunuri/scipy,dominicelse/scipy,grlee77/scipy,witcxc/scipy,raoulbq/scipy,zaxliu/scipy,petebachant/scipy,perimosocordiae/scipy,Gillu13/scipy,mdhaber/scipy,jseabold/scipy,jamestwebber/scipy,hainm/scipy,zxsted/scipy,FRidh/scipy,cpaulik/scipy,gef756/scipy,FRidh/scipy,grlee77/scipy,trankmichael/scipy,e-q/scipy,maciejkula/scipy,richardotis/scipy,dominicelse/scipy,haudren/scipy,sriki18/scipy,dch312/scipy,efiring/scipy,piyush0609/scipy,jsilter/scipy,sargas/scipy,matthewalbani/scipy,Gillu13/scipy,maniteja123/scipy,gef756/scipy,lhilt/scipy,andim/scipy,Stefan-Endres/scipy,ilayn/scipy,niknow/scipy,mhogg/scipy,petebachant/scipy,giorgiop/scipy,anntzer/scipy,larsmans/scipy,minhlongdo/scipy,aman-iitj/scipy,kalvdans/scipy,zerothi/scipy,efiring/scipy,pnedunuri/scipy,perimosocordiae/scipy,sauliusl/scipy,jonycgn/scipy,Newman101/scipy,Gillu13/scipy,vhaasteren/scipy,andyfaff/scipy,apbard/scipy,jor-/scipy,sargas/scipy,ales-erjavec/scipy,raoulbq/scipy,jonycgn/scipy,nvoron23/scipy,richardotis/scipy,lukauskas/scipy,ChanderG/scipy,minhlongdo/scipy,Shaswat27/scipy,woodscn/scipy,niknow/scipy,aeklant/scipy,mingwpy/scipy,vhaasteren/scipy,anielsen001/scipy,njwilson23/scipy,ogrisel/scipy,fernand/scipy,anntzer/scipy,jakevdp/scipy,person142/scipy,pbrod/scipy,pbrod/scipy,aarchiba/scipy,vberaudi/scipy,tylerjereddy/scipy,Kamp9/scipy,dch312/scipy,njwilson23/scipy,hainm/scipy,endolith/scipy,zerothi/scipy,nvoron23/scipy,minhlongdo/scipy,sriki18/scipy,scipy/scipy,gertingold/scipy,petebachant/scipy,pizzathief/scipy,nmayorov/scipy,mhogg/scipy,mtrbean/scipy,kalvdans/scipy,mgaitan/scipy,mhogg/scipy,matthew-brett/scipy,scipy/scipy,jamestwebber/scipy,aarchiba/scipy,gfyoung/scipy,ales-erjavec/scipy,zaxliu/scipy,vberaudi/scipy,pschella/scipy,jseabold/scipy,perimosocordiae/scipy,zxsted/scipy,behzadnouri/scipy,gfyoung/scipy,woodscn/scipy,mingwpy/scipy,maniteja123/scipy,mdhaber/scipy,josephcslater/scipy,Kamp9/scipy,ndchorley/scipy,sriki18/scipy,trankmichael/scipy,Stefan-Endres/scipy,niknow/scipy,witcxc/scipy,argriffing/scipy,vanpact/scipy,person142/scipy,lhilt/scipy,mtrbean/scipy,haudren/scipy,larsmans/scipy,efiring/scipy,mtrbean/scipy,hainm/scipy,zxsted/scipy,ChanderG/scipy,mortada/scipy,giorgiop/scipy,aman-iitj/scipy,josephcslater/scipy,vanpact/scipy,matthew-brett/scipy,teoliphant/scipy,matthewalbani/scipy,newemailjdm/scipy,woodscn/scipy,sonnyhu/scipy,aeklant/scipy,anielsen001/scipy,andyfaff/scipy,cpaulik/scipy,perimosocordiae/scipy,pizzathief/scipy,njwilson23/scipy,e-q/scipy,trankmichael/scipy,lhilt/scipy,tylerjereddy/scipy,teoliphant/scipy,vigna/scipy,mdhaber/scipy,ogrisel/scipy,petebachant/scipy,raoulbq/scipy,felipebetancur/scipy,nmayorov/scipy,njwilson23/scipy,sargas/scipy,larsmans/scipy,jamestwebber/scipy,newemailjdm/scipy,rgommers/scipy,Stefan-Endres/scipy,rmcgibbo/scipy,mgaitan/scipy,anielsen001/scipy,arokem/scipy,jjhelmus/scipy,zaxliu/scipy,nvoron23/scipy,chatcannon/scipy,mikebenfield/scipy,richardotis/scipy,mortonjt/scipy,nonhermitian/scipy,befelix/scipy,mikebenfield/scipy,gertingold/scipy,gdooper/scipy,pnedunuri/scipy,befelix/scipy,scipy/scipy,nonhermitian/scipy,WillieMaddox/scipy,WillieMaddox/scipy,behzadnouri/scipy,haudren/scipy,piyush0609/scipy,dominicelse/scipy,Shaswat27/scipy,richardotis/scipy,WarrenWeckesser/scipy,gfyoung/scipy,mortada/scipy,efiring/scipy,ales-erjavec/scipy,WarrenWeckesser/scipy,mingwpy/scipy,aeklant/scipy,piyush0609/scipy,felipebetancur/scipy,mhogg/scipy,pbrod/scipy,mdhaber/scipy,maniteja123/scipy,anielsen001/scipy,Srisai85/scipy,mdhaber/scipy,pschella/scipy,pnedunuri/scipy,mortada/scipy,mingwpy/scipy,piyush0609/scipy,endolith/scipy,ortylp/scipy,mgaitan/scipy,surhudm/scipy,WillieMaddox/scipy,giorgiop/scipy,Srisai85/scipy,Newman101/scipy,maciejkula/scipy,perimosocordiae/scipy,person142/scipy,josephcslater/scipy,anntzer/scipy,gfyoung/scipy,pyramania/scipy,fredrikw/scipy,grlee77/scipy,kleskjr/scipy,Stefan-Endres/scipy,sonnyhu/scipy,cpaulik/scipy,Kamp9/scipy,ndchorley/scipy,nmayorov/scipy,gef756/scipy,vanpact/scipy,arokem/scipy,ndchorley/scipy,jonycgn/scipy,e-q/scipy,Dapid/scipy,ales-erjavec/scipy,jsilter/scipy,argriffing/scipy,zxsted/scipy,pyramania/scipy,jseabold/scipy,witcxc/scipy,fernand/scipy,sauliusl/scipy,zaxliu/scipy,lhilt/scipy,scipy/scipy,lhilt/scipy,juliantaylor/scipy,bkendzior/scipy,aman-iitj/scipy,Eric89GXL/scipy,cpaulik/scipy,aarchiba/scipy,fernand/scipy,fredrikw/scipy,Kamp9/scipy,zerothi/scipy,pizzathief/scipy,fernand/scipy,jonycgn/scipy,zxsted/scipy,andyfaff/scipy,kleskjr/scipy,mortonjt/scipy,jseabold/scipy,vigna/scipy,e-q/scipy,jakevdp/scipy,vanpact/scipy,mingwpy/scipy,futurulus/scipy,andim/scipy,pnedunuri/scipy,sonnyhu/scipy,matthew-brett/scipy,lukauskas/scipy,vanpact/scipy,jjhelmus/scipy,Srisai85/scipy,jsilter/scipy,argriffing/scipy,bkendzior/scipy,felipebetancur/scipy,hainm/scipy,pyramania/scipy,vberaudi/scipy,apbard/scipy,befelix/scipy,giorgiop/scipy,hainm/scipy,WarrenWeckesser/scipy,fredrikw/scipy,jor-/scipy,larsmans/scipy,ilayn/scipy,josephcslater/scipy,vigna/scipy,matthew-brett/scipy,person142/scipy,futurulus/scipy,zerothi/scipy,vberaudi/scipy,gertingold/scipy,ndchorley/scipy,arokem/scipy,arokem/scipy,raoulbq/scipy,matthewalbani/scipy,ndchorley/scipy,anielsen001/scipy,matthewalbani/scipy,pschella/scipy,cpaulik/scipy,bkendzior/scipy,gdooper/scipy,Shaswat27/scipy,sargas/scipy,Eric89GXL/scipy,aarchiba/scipy,ChanderG/scipy,ortylp/scipy,efiring/scipy,teoliphant/scipy,richardotis/scipy,andyfaff/scipy,jor-/scipy,ilayn/scipy,WarrenWeckesser/scipy,argriffing/scipy,maciejkula/scipy,mtrbean/scipy,futurulus/scipy,scipy/scipy,futurulus/scipy,person142/scipy,ilayn/scipy,aeklant/scipy,Newman101/scipy,nvoron23/scipy,felipebetancur/scipy,dch312/scipy,futurulus/scipy,raoulbq/scipy,lukauskas/scipy,anielsen001/scipy,zaxliu/scipy,efiring/scipy,lukauskas/scipy,minhlongdo/scipy,gef756/scipy,argriffing/scipy,ilayn/scipy,tylerjereddy/scipy,Newman101/scipy,witcxc/scipy,rmcgibbo/scipy,scipy/scipy,haudren/scipy,behzadnouri/scipy,chatcannon/scipy,larsmans/scipy,endolith/scipy,maniteja123/scipy,piyush0609/scipy,mortada/scipy,ortylp/scipy,jakevdp/scipy,Dapid/scipy,woodscn/scipy,newemailjdm/scipy,sauliusl/scipy,vhaasteren/scipy,vberaudi/scipy,FRidh/scipy,tylerjereddy/scipy,pschella/scipy,juliantaylor/scipy,newemailjdm/scipy,petebachant/scipy,chatcannon/scipy,Srisai85/scipy,sauliusl/scipy,mortonjt/scipy,kleskjr/scipy,dch312/scipy,Stefan-Endres/scipy,befelix/scipy,hainm/scipy,ndchorley/scipy,Gillu13/scipy,fernand/scipy,andim/scipy,gef756/scipy,pnedunuri/scipy,Dapid/scipy,jsilter/scipy,dominicelse/scipy,matthewalbani/scipy,mikebenfield/scipy,bkendzior/scipy,WillieMaddox/scipy,Dapid/scipy,jsilter/scipy,niknow/scipy,ortylp/scipy,mortonjt/scipy,pbrod/scipy,jjhelmus/scipy,mortada/scipy,WillieMaddox/scipy,e-q/scipy,Dapid/scipy,mgaitan/scipy,zaxliu/scipy,nmayorov/scipy,maniteja123/scipy,Srisai85/scipy,ogrisel/scipy
import os from numpy.distutils.misc_util import Configuration def configuration(parent_package='',top_path=None): config = Configuration('misc',parent_package, top_path) config.add_data_files('lena.dat') print "########", config return config if __name__ == '__main__': from numpy.distutils.core import setup setup(**configuration()) Remove extra noise on install.
import os from numpy.distutils.misc_util import Configuration def configuration(parent_package='',top_path=None): config = Configuration('misc',parent_package, top_path) config.add_data_files('lena.dat') return config if __name__ == '__main__': from numpy.distutils.core import setup setup(**configuration())
<commit_before> import os from numpy.distutils.misc_util import Configuration def configuration(parent_package='',top_path=None): config = Configuration('misc',parent_package, top_path) config.add_data_files('lena.dat') print "########", config return config if __name__ == '__main__': from numpy.distutils.core import setup setup(**configuration()) <commit_msg>Remove extra noise on install.<commit_after>
import os from numpy.distutils.misc_util import Configuration def configuration(parent_package='',top_path=None): config = Configuration('misc',parent_package, top_path) config.add_data_files('lena.dat') return config if __name__ == '__main__': from numpy.distutils.core import setup setup(**configuration())
import os from numpy.distutils.misc_util import Configuration def configuration(parent_package='',top_path=None): config = Configuration('misc',parent_package, top_path) config.add_data_files('lena.dat') print "########", config return config if __name__ == '__main__': from numpy.distutils.core import setup setup(**configuration()) Remove extra noise on install. import os from numpy.distutils.misc_util import Configuration def configuration(parent_package='',top_path=None): config = Configuration('misc',parent_package, top_path) config.add_data_files('lena.dat') return config if __name__ == '__main__': from numpy.distutils.core import setup setup(**configuration())
<commit_before> import os from numpy.distutils.misc_util import Configuration def configuration(parent_package='',top_path=None): config = Configuration('misc',parent_package, top_path) config.add_data_files('lena.dat') print "########", config return config if __name__ == '__main__': from numpy.distutils.core import setup setup(**configuration()) <commit_msg>Remove extra noise on install.<commit_after> import os from numpy.distutils.misc_util import Configuration def configuration(parent_package='',top_path=None): config = Configuration('misc',parent_package, top_path) config.add_data_files('lena.dat') return config if __name__ == '__main__': from numpy.distutils.core import setup setup(**configuration())
758f73e1ecc34f52929595dfcf5db4a3a24fcbc6
Python/views.py
Python/views.py
import requests from django.shortcuts import render from django.conf import settings def oauthtest(request): return render(request, 'oauthtest.html', { 'link': '{}o/authorize/?response_type=code&client_id={}&redirect_uri={}{}/oauthdone/'.format( settings.API_URL, settings.OAUTH_CLIENT_ID, 'https://' if request.is_secure() else 'http://', request.META['HTTP_HOST'], ) }) def oauthdone(request): # Get token r = requests.post(settings.API_URL + 'o/token/', data={ 'grant_type': 'authorization_code', 'code': request.GET['code'], 'redirect_uri': 'http://localhost:8000/oauthdone/', 'client_id': settings.OAUTH_CLIENT_ID, 'client_secret': settings.OAUTH_CLIENT_SECRET, }) oauth = r.json() # Get user object + accounts headers = { 'Authorization': oauth['token_type'] + ' ' + oauth['access_token'], } r = requests.get(settings.API_URL + 'api/users/me/?expand_accounts=True', headers=headers) return render(request, 'accounts.html', { 'user': r.json(), })
import requests from django.shortcuts import render from django.conf import settings def oauthtest(request): return render(request, 'oauthtest.html', { 'link': '{}o/authorize/?response_type=code&client_id={}&redirect_uri={}{}/oauthdone/'.format( settings.API_URL, settings.OAUTH_CLIENT_ID, 'https://' if request.is_secure() else 'http://', request.META['HTTP_HOST'], ) }) def oauthdone(request): # Get token r = requests.post(settings.API_URL + 'o/token/', data={ 'grant_type': 'authorization_code', 'code': request.GET['code'], 'redirect_uri': '{}{}/oauthdone/'.format('https://' if request.is_secure() else 'http://', request.META['HTTP_HOST']), 'client_id': settings.OAUTH_CLIENT_ID, 'client_secret': settings.OAUTH_CLIENT_SECRET, }) oauth = r.json() # Get user object + accounts headers = { 'Authorization': oauth['token_type'] + ' ' + oauth['access_token'], } r = requests.get(settings.API_URL + 'api/users/me/?expand_accounts=True', headers=headers) return render(request, 'accounts.html', { 'user': r.json(), })
Fix redirect URI in oauthdone
Fix redirect URI in oauthdone
Python
apache-2.0
SchoolIdolTomodachi/SchoolIdolAPIOAuthExample,SchoolIdolTomodachi/SchoolIdolAPIOAuthExample,SchoolIdolTomodachi/SchoolIdolAPIOAuthExample
import requests from django.shortcuts import render from django.conf import settings def oauthtest(request): return render(request, 'oauthtest.html', { 'link': '{}o/authorize/?response_type=code&client_id={}&redirect_uri={}{}/oauthdone/'.format( settings.API_URL, settings.OAUTH_CLIENT_ID, 'https://' if request.is_secure() else 'http://', request.META['HTTP_HOST'], ) }) def oauthdone(request): # Get token r = requests.post(settings.API_URL + 'o/token/', data={ 'grant_type': 'authorization_code', 'code': request.GET['code'], 'redirect_uri': 'http://localhost:8000/oauthdone/', 'client_id': settings.OAUTH_CLIENT_ID, 'client_secret': settings.OAUTH_CLIENT_SECRET, }) oauth = r.json() # Get user object + accounts headers = { 'Authorization': oauth['token_type'] + ' ' + oauth['access_token'], } r = requests.get(settings.API_URL + 'api/users/me/?expand_accounts=True', headers=headers) return render(request, 'accounts.html', { 'user': r.json(), }) Fix redirect URI in oauthdone
import requests from django.shortcuts import render from django.conf import settings def oauthtest(request): return render(request, 'oauthtest.html', { 'link': '{}o/authorize/?response_type=code&client_id={}&redirect_uri={}{}/oauthdone/'.format( settings.API_URL, settings.OAUTH_CLIENT_ID, 'https://' if request.is_secure() else 'http://', request.META['HTTP_HOST'], ) }) def oauthdone(request): # Get token r = requests.post(settings.API_URL + 'o/token/', data={ 'grant_type': 'authorization_code', 'code': request.GET['code'], 'redirect_uri': '{}{}/oauthdone/'.format('https://' if request.is_secure() else 'http://', request.META['HTTP_HOST']), 'client_id': settings.OAUTH_CLIENT_ID, 'client_secret': settings.OAUTH_CLIENT_SECRET, }) oauth = r.json() # Get user object + accounts headers = { 'Authorization': oauth['token_type'] + ' ' + oauth['access_token'], } r = requests.get(settings.API_URL + 'api/users/me/?expand_accounts=True', headers=headers) return render(request, 'accounts.html', { 'user': r.json(), })
<commit_before>import requests from django.shortcuts import render from django.conf import settings def oauthtest(request): return render(request, 'oauthtest.html', { 'link': '{}o/authorize/?response_type=code&client_id={}&redirect_uri={}{}/oauthdone/'.format( settings.API_URL, settings.OAUTH_CLIENT_ID, 'https://' if request.is_secure() else 'http://', request.META['HTTP_HOST'], ) }) def oauthdone(request): # Get token r = requests.post(settings.API_URL + 'o/token/', data={ 'grant_type': 'authorization_code', 'code': request.GET['code'], 'redirect_uri': 'http://localhost:8000/oauthdone/', 'client_id': settings.OAUTH_CLIENT_ID, 'client_secret': settings.OAUTH_CLIENT_SECRET, }) oauth = r.json() # Get user object + accounts headers = { 'Authorization': oauth['token_type'] + ' ' + oauth['access_token'], } r = requests.get(settings.API_URL + 'api/users/me/?expand_accounts=True', headers=headers) return render(request, 'accounts.html', { 'user': r.json(), }) <commit_msg>Fix redirect URI in oauthdone<commit_after>
import requests from django.shortcuts import render from django.conf import settings def oauthtest(request): return render(request, 'oauthtest.html', { 'link': '{}o/authorize/?response_type=code&client_id={}&redirect_uri={}{}/oauthdone/'.format( settings.API_URL, settings.OAUTH_CLIENT_ID, 'https://' if request.is_secure() else 'http://', request.META['HTTP_HOST'], ) }) def oauthdone(request): # Get token r = requests.post(settings.API_URL + 'o/token/', data={ 'grant_type': 'authorization_code', 'code': request.GET['code'], 'redirect_uri': '{}{}/oauthdone/'.format('https://' if request.is_secure() else 'http://', request.META['HTTP_HOST']), 'client_id': settings.OAUTH_CLIENT_ID, 'client_secret': settings.OAUTH_CLIENT_SECRET, }) oauth = r.json() # Get user object + accounts headers = { 'Authorization': oauth['token_type'] + ' ' + oauth['access_token'], } r = requests.get(settings.API_URL + 'api/users/me/?expand_accounts=True', headers=headers) return render(request, 'accounts.html', { 'user': r.json(), })
import requests from django.shortcuts import render from django.conf import settings def oauthtest(request): return render(request, 'oauthtest.html', { 'link': '{}o/authorize/?response_type=code&client_id={}&redirect_uri={}{}/oauthdone/'.format( settings.API_URL, settings.OAUTH_CLIENT_ID, 'https://' if request.is_secure() else 'http://', request.META['HTTP_HOST'], ) }) def oauthdone(request): # Get token r = requests.post(settings.API_URL + 'o/token/', data={ 'grant_type': 'authorization_code', 'code': request.GET['code'], 'redirect_uri': 'http://localhost:8000/oauthdone/', 'client_id': settings.OAUTH_CLIENT_ID, 'client_secret': settings.OAUTH_CLIENT_SECRET, }) oauth = r.json() # Get user object + accounts headers = { 'Authorization': oauth['token_type'] + ' ' + oauth['access_token'], } r = requests.get(settings.API_URL + 'api/users/me/?expand_accounts=True', headers=headers) return render(request, 'accounts.html', { 'user': r.json(), }) Fix redirect URI in oauthdoneimport requests from django.shortcuts import render from django.conf import settings def oauthtest(request): return render(request, 'oauthtest.html', { 'link': '{}o/authorize/?response_type=code&client_id={}&redirect_uri={}{}/oauthdone/'.format( settings.API_URL, settings.OAUTH_CLIENT_ID, 'https://' if request.is_secure() else 'http://', request.META['HTTP_HOST'], ) }) def oauthdone(request): # Get token r = requests.post(settings.API_URL + 'o/token/', data={ 'grant_type': 'authorization_code', 'code': request.GET['code'], 'redirect_uri': '{}{}/oauthdone/'.format('https://' if request.is_secure() else 'http://', request.META['HTTP_HOST']), 'client_id': settings.OAUTH_CLIENT_ID, 'client_secret': settings.OAUTH_CLIENT_SECRET, }) oauth = r.json() # Get user object + accounts headers = { 'Authorization': oauth['token_type'] + ' ' + oauth['access_token'], } r = requests.get(settings.API_URL + 'api/users/me/?expand_accounts=True', headers=headers) return render(request, 'accounts.html', { 'user': r.json(), })
<commit_before>import requests from django.shortcuts import render from django.conf import settings def oauthtest(request): return render(request, 'oauthtest.html', { 'link': '{}o/authorize/?response_type=code&client_id={}&redirect_uri={}{}/oauthdone/'.format( settings.API_URL, settings.OAUTH_CLIENT_ID, 'https://' if request.is_secure() else 'http://', request.META['HTTP_HOST'], ) }) def oauthdone(request): # Get token r = requests.post(settings.API_URL + 'o/token/', data={ 'grant_type': 'authorization_code', 'code': request.GET['code'], 'redirect_uri': 'http://localhost:8000/oauthdone/', 'client_id': settings.OAUTH_CLIENT_ID, 'client_secret': settings.OAUTH_CLIENT_SECRET, }) oauth = r.json() # Get user object + accounts headers = { 'Authorization': oauth['token_type'] + ' ' + oauth['access_token'], } r = requests.get(settings.API_URL + 'api/users/me/?expand_accounts=True', headers=headers) return render(request, 'accounts.html', { 'user': r.json(), }) <commit_msg>Fix redirect URI in oauthdone<commit_after>import requests from django.shortcuts import render from django.conf import settings def oauthtest(request): return render(request, 'oauthtest.html', { 'link': '{}o/authorize/?response_type=code&client_id={}&redirect_uri={}{}/oauthdone/'.format( settings.API_URL, settings.OAUTH_CLIENT_ID, 'https://' if request.is_secure() else 'http://', request.META['HTTP_HOST'], ) }) def oauthdone(request): # Get token r = requests.post(settings.API_URL + 'o/token/', data={ 'grant_type': 'authorization_code', 'code': request.GET['code'], 'redirect_uri': '{}{}/oauthdone/'.format('https://' if request.is_secure() else 'http://', request.META['HTTP_HOST']), 'client_id': settings.OAUTH_CLIENT_ID, 'client_secret': settings.OAUTH_CLIENT_SECRET, }) oauth = r.json() # Get user object + accounts headers = { 'Authorization': oauth['token_type'] + ' ' + oauth['access_token'], } r = requests.get(settings.API_URL + 'api/users/me/?expand_accounts=True', headers=headers) return render(request, 'accounts.html', { 'user': r.json(), })
68a1877bcd4511008aeff977cb45fa9edb5e9a8b
fusekiutils/__init__.py
fusekiutils/__init__.py
__author__ = 'adam' import time from subprocess import Popen import shlex import os import urllib def LaunchFuseki(): fuseki_url = "http://localhost:3030" fuseki_dir = os.getcwd() + "/jena-fuseki" fuseki_executable = fuseki_dir + "/fuseki-server" f_log = open("fuseki.log","w") fuseki = Popen( args=shlex.split("-q --update --loc=../fuseki-data /qudt4dt"), executable=fuseki_executable, cwd=fuseki_dir, stdout=f_log) f_log.close() PollFusekiLaunch("http://localhost:3030") return fuseki def PollFusekiLaunch(fuseki_url): while True: if IsFusekiRunning(fuseki_url): return else: print "polling..." time.sleep(1) def IsFusekiRunning(fuseki_url): try: urllib.urlopen(fuseki_url) return True except IOError: return False
__author__ = 'adam' import time from subprocess import Popen import shlex import os import urllib import sys def LaunchFuseki(): fuseki_dir = os.path.join(os.path.abspath(os.getcwd()), 'jena-fuseki') if sys.platform == 'win32': fuseki_executable = os.path.join(fuseki_dir, 'fuseki-server.bat') else: fuseki_executable = os.path.join(fuseki_dir, 'fuseki-server') fuseki_data = os.path.join('..', 'fuseki-data') f_log = open("fuseki.log", "w") args = '%(exec)s -q --update --loc=%(data)s /qudt4dt' % {'exec': fuseki_executable, 'data': fuseki_data} fuseki = Popen(args=args, cwd=fuseki_dir, stdout=f_log) f_log.close() PollFusekiLaunch("http://localhost:3030") return fuseki def PollFusekiLaunch(fuseki_url): while True: if IsFusekiRunning(fuseki_url): return else: print "polling..." time.sleep(1) def IsFusekiRunning(fuseki_url): try: urllib.urlopen(fuseki_url) return True except IOError: return False
Support both windows and shell environments when launching fuseki
Support both windows and shell environments when launching fuseki
Python
lgpl-2.1
adamnagel/qudt-for-domain-tools,adamnagel/qudt-for-domain-tools,adamnagel/qudt-for-domain-tools
__author__ = 'adam' import time from subprocess import Popen import shlex import os import urllib def LaunchFuseki(): fuseki_url = "http://localhost:3030" fuseki_dir = os.getcwd() + "/jena-fuseki" fuseki_executable = fuseki_dir + "/fuseki-server" f_log = open("fuseki.log","w") fuseki = Popen( args=shlex.split("-q --update --loc=../fuseki-data /qudt4dt"), executable=fuseki_executable, cwd=fuseki_dir, stdout=f_log) f_log.close() PollFusekiLaunch("http://localhost:3030") return fuseki def PollFusekiLaunch(fuseki_url): while True: if IsFusekiRunning(fuseki_url): return else: print "polling..." time.sleep(1) def IsFusekiRunning(fuseki_url): try: urllib.urlopen(fuseki_url) return True except IOError: return FalseSupport both windows and shell environments when launching fuseki
__author__ = 'adam' import time from subprocess import Popen import shlex import os import urllib import sys def LaunchFuseki(): fuseki_dir = os.path.join(os.path.abspath(os.getcwd()), 'jena-fuseki') if sys.platform == 'win32': fuseki_executable = os.path.join(fuseki_dir, 'fuseki-server.bat') else: fuseki_executable = os.path.join(fuseki_dir, 'fuseki-server') fuseki_data = os.path.join('..', 'fuseki-data') f_log = open("fuseki.log", "w") args = '%(exec)s -q --update --loc=%(data)s /qudt4dt' % {'exec': fuseki_executable, 'data': fuseki_data} fuseki = Popen(args=args, cwd=fuseki_dir, stdout=f_log) f_log.close() PollFusekiLaunch("http://localhost:3030") return fuseki def PollFusekiLaunch(fuseki_url): while True: if IsFusekiRunning(fuseki_url): return else: print "polling..." time.sleep(1) def IsFusekiRunning(fuseki_url): try: urllib.urlopen(fuseki_url) return True except IOError: return False
<commit_before>__author__ = 'adam' import time from subprocess import Popen import shlex import os import urllib def LaunchFuseki(): fuseki_url = "http://localhost:3030" fuseki_dir = os.getcwd() + "/jena-fuseki" fuseki_executable = fuseki_dir + "/fuseki-server" f_log = open("fuseki.log","w") fuseki = Popen( args=shlex.split("-q --update --loc=../fuseki-data /qudt4dt"), executable=fuseki_executable, cwd=fuseki_dir, stdout=f_log) f_log.close() PollFusekiLaunch("http://localhost:3030") return fuseki def PollFusekiLaunch(fuseki_url): while True: if IsFusekiRunning(fuseki_url): return else: print "polling..." time.sleep(1) def IsFusekiRunning(fuseki_url): try: urllib.urlopen(fuseki_url) return True except IOError: return False<commit_msg>Support both windows and shell environments when launching fuseki<commit_after>
__author__ = 'adam' import time from subprocess import Popen import shlex import os import urllib import sys def LaunchFuseki(): fuseki_dir = os.path.join(os.path.abspath(os.getcwd()), 'jena-fuseki') if sys.platform == 'win32': fuseki_executable = os.path.join(fuseki_dir, 'fuseki-server.bat') else: fuseki_executable = os.path.join(fuseki_dir, 'fuseki-server') fuseki_data = os.path.join('..', 'fuseki-data') f_log = open("fuseki.log", "w") args = '%(exec)s -q --update --loc=%(data)s /qudt4dt' % {'exec': fuseki_executable, 'data': fuseki_data} fuseki = Popen(args=args, cwd=fuseki_dir, stdout=f_log) f_log.close() PollFusekiLaunch("http://localhost:3030") return fuseki def PollFusekiLaunch(fuseki_url): while True: if IsFusekiRunning(fuseki_url): return else: print "polling..." time.sleep(1) def IsFusekiRunning(fuseki_url): try: urllib.urlopen(fuseki_url) return True except IOError: return False
__author__ = 'adam' import time from subprocess import Popen import shlex import os import urllib def LaunchFuseki(): fuseki_url = "http://localhost:3030" fuseki_dir = os.getcwd() + "/jena-fuseki" fuseki_executable = fuseki_dir + "/fuseki-server" f_log = open("fuseki.log","w") fuseki = Popen( args=shlex.split("-q --update --loc=../fuseki-data /qudt4dt"), executable=fuseki_executable, cwd=fuseki_dir, stdout=f_log) f_log.close() PollFusekiLaunch("http://localhost:3030") return fuseki def PollFusekiLaunch(fuseki_url): while True: if IsFusekiRunning(fuseki_url): return else: print "polling..." time.sleep(1) def IsFusekiRunning(fuseki_url): try: urllib.urlopen(fuseki_url) return True except IOError: return FalseSupport both windows and shell environments when launching fuseki__author__ = 'adam' import time from subprocess import Popen import shlex import os import urllib import sys def LaunchFuseki(): fuseki_dir = os.path.join(os.path.abspath(os.getcwd()), 'jena-fuseki') if sys.platform == 'win32': fuseki_executable = os.path.join(fuseki_dir, 'fuseki-server.bat') else: fuseki_executable = os.path.join(fuseki_dir, 'fuseki-server') fuseki_data = os.path.join('..', 'fuseki-data') f_log = open("fuseki.log", "w") args = '%(exec)s -q --update --loc=%(data)s /qudt4dt' % {'exec': fuseki_executable, 'data': fuseki_data} fuseki = Popen(args=args, cwd=fuseki_dir, stdout=f_log) f_log.close() PollFusekiLaunch("http://localhost:3030") return fuseki def PollFusekiLaunch(fuseki_url): while True: if IsFusekiRunning(fuseki_url): return else: print "polling..." time.sleep(1) def IsFusekiRunning(fuseki_url): try: urllib.urlopen(fuseki_url) return True except IOError: return False
<commit_before>__author__ = 'adam' import time from subprocess import Popen import shlex import os import urllib def LaunchFuseki(): fuseki_url = "http://localhost:3030" fuseki_dir = os.getcwd() + "/jena-fuseki" fuseki_executable = fuseki_dir + "/fuseki-server" f_log = open("fuseki.log","w") fuseki = Popen( args=shlex.split("-q --update --loc=../fuseki-data /qudt4dt"), executable=fuseki_executable, cwd=fuseki_dir, stdout=f_log) f_log.close() PollFusekiLaunch("http://localhost:3030") return fuseki def PollFusekiLaunch(fuseki_url): while True: if IsFusekiRunning(fuseki_url): return else: print "polling..." time.sleep(1) def IsFusekiRunning(fuseki_url): try: urllib.urlopen(fuseki_url) return True except IOError: return False<commit_msg>Support both windows and shell environments when launching fuseki<commit_after>__author__ = 'adam' import time from subprocess import Popen import shlex import os import urllib import sys def LaunchFuseki(): fuseki_dir = os.path.join(os.path.abspath(os.getcwd()), 'jena-fuseki') if sys.platform == 'win32': fuseki_executable = os.path.join(fuseki_dir, 'fuseki-server.bat') else: fuseki_executable = os.path.join(fuseki_dir, 'fuseki-server') fuseki_data = os.path.join('..', 'fuseki-data') f_log = open("fuseki.log", "w") args = '%(exec)s -q --update --loc=%(data)s /qudt4dt' % {'exec': fuseki_executable, 'data': fuseki_data} fuseki = Popen(args=args, cwd=fuseki_dir, stdout=f_log) f_log.close() PollFusekiLaunch("http://localhost:3030") return fuseki def PollFusekiLaunch(fuseki_url): while True: if IsFusekiRunning(fuseki_url): return else: print "polling..." time.sleep(1) def IsFusekiRunning(fuseki_url): try: urllib.urlopen(fuseki_url) return True except IOError: return False
94ff1527fb16c7a3557112f6e30cded4de99dda8
fabtastic/fabric/commands/c_supervisord.py
fabtastic/fabric/commands/c_supervisord.py
from fabric.api import * from fabtastic.fabric.util import _current_host_has_role def supervisord_restart_all(roles='webapp_servers'): """ Restarts all of supervisord's managed programs. """ if _current_host_has_role(roles): print("=== RESTARTING SUPERVISORD PROGRAMS ===") with cd(env.REMOTE_CODEBASE_PATH): run("workon %s && ./manage.py ft_supervisord_restart_prog" % env.REMOTE_VIRTUALENV_NAME) def supervisord_restart_prog(program, roles='webapp_servers'): """ Restarts all of supervisord's managed programs. :arg str program: The name of the program to restart (as per supervisor's conf.d/ contents). """ if _current_host_has_role(roles): print("=== RESTARTING SUPERVISORD PROGRAMS ===") with cd(env.REMOTE_CODEBASE_PATH): run("workon %s && ./manage.py ft_supervisord_restart_prog %s" % ( program, env.REMOTE_VIRTUALENV_NAME))
from fabric.api import * from fabtastic.fabric.util import _current_host_has_role def supervisord_restart_all(roles='webapp_servers'): """ Restarts all of supervisord's managed programs. """ if _current_host_has_role(roles): print("=== RESTARTING SUPERVISORD PROGRAMS ===") with cd(env.REMOTE_CODEBASE_PATH): run("workon %s && ./manage.py ft_supervisord_restart_prog" % env.REMOTE_VIRTUALENV_NAME) def supervisord_restart_prog(program, roles='webapp_servers'): """ Restarts all of supervisord's managed programs. :arg str program: The name of the program to restart (as per supervisor's conf.d/ contents). """ if _current_host_has_role(roles): print("=== RESTARTING SUPERVISORD PROGRAMS ===") with cd(env.REMOTE_CODEBASE_PATH): run("workon %s && ./manage.py ft_supervisord_restart_prog %s" % ( env.REMOTE_VIRTUALENV_NAME, program))
Fix arg order for supervisord_restart_prog
Fix arg order for supervisord_restart_prog
Python
bsd-3-clause
duointeractive/django-fabtastic
from fabric.api import * from fabtastic.fabric.util import _current_host_has_role def supervisord_restart_all(roles='webapp_servers'): """ Restarts all of supervisord's managed programs. """ if _current_host_has_role(roles): print("=== RESTARTING SUPERVISORD PROGRAMS ===") with cd(env.REMOTE_CODEBASE_PATH): run("workon %s && ./manage.py ft_supervisord_restart_prog" % env.REMOTE_VIRTUALENV_NAME) def supervisord_restart_prog(program, roles='webapp_servers'): """ Restarts all of supervisord's managed programs. :arg str program: The name of the program to restart (as per supervisor's conf.d/ contents). """ if _current_host_has_role(roles): print("=== RESTARTING SUPERVISORD PROGRAMS ===") with cd(env.REMOTE_CODEBASE_PATH): run("workon %s && ./manage.py ft_supervisord_restart_prog %s" % ( program, env.REMOTE_VIRTUALENV_NAME)) Fix arg order for supervisord_restart_prog
from fabric.api import * from fabtastic.fabric.util import _current_host_has_role def supervisord_restart_all(roles='webapp_servers'): """ Restarts all of supervisord's managed programs. """ if _current_host_has_role(roles): print("=== RESTARTING SUPERVISORD PROGRAMS ===") with cd(env.REMOTE_CODEBASE_PATH): run("workon %s && ./manage.py ft_supervisord_restart_prog" % env.REMOTE_VIRTUALENV_NAME) def supervisord_restart_prog(program, roles='webapp_servers'): """ Restarts all of supervisord's managed programs. :arg str program: The name of the program to restart (as per supervisor's conf.d/ contents). """ if _current_host_has_role(roles): print("=== RESTARTING SUPERVISORD PROGRAMS ===") with cd(env.REMOTE_CODEBASE_PATH): run("workon %s && ./manage.py ft_supervisord_restart_prog %s" % ( env.REMOTE_VIRTUALENV_NAME, program))
<commit_before>from fabric.api import * from fabtastic.fabric.util import _current_host_has_role def supervisord_restart_all(roles='webapp_servers'): """ Restarts all of supervisord's managed programs. """ if _current_host_has_role(roles): print("=== RESTARTING SUPERVISORD PROGRAMS ===") with cd(env.REMOTE_CODEBASE_PATH): run("workon %s && ./manage.py ft_supervisord_restart_prog" % env.REMOTE_VIRTUALENV_NAME) def supervisord_restart_prog(program, roles='webapp_servers'): """ Restarts all of supervisord's managed programs. :arg str program: The name of the program to restart (as per supervisor's conf.d/ contents). """ if _current_host_has_role(roles): print("=== RESTARTING SUPERVISORD PROGRAMS ===") with cd(env.REMOTE_CODEBASE_PATH): run("workon %s && ./manage.py ft_supervisord_restart_prog %s" % ( program, env.REMOTE_VIRTUALENV_NAME)) <commit_msg>Fix arg order for supervisord_restart_prog<commit_after>
from fabric.api import * from fabtastic.fabric.util import _current_host_has_role def supervisord_restart_all(roles='webapp_servers'): """ Restarts all of supervisord's managed programs. """ if _current_host_has_role(roles): print("=== RESTARTING SUPERVISORD PROGRAMS ===") with cd(env.REMOTE_CODEBASE_PATH): run("workon %s && ./manage.py ft_supervisord_restart_prog" % env.REMOTE_VIRTUALENV_NAME) def supervisord_restart_prog(program, roles='webapp_servers'): """ Restarts all of supervisord's managed programs. :arg str program: The name of the program to restart (as per supervisor's conf.d/ contents). """ if _current_host_has_role(roles): print("=== RESTARTING SUPERVISORD PROGRAMS ===") with cd(env.REMOTE_CODEBASE_PATH): run("workon %s && ./manage.py ft_supervisord_restart_prog %s" % ( env.REMOTE_VIRTUALENV_NAME, program))
from fabric.api import * from fabtastic.fabric.util import _current_host_has_role def supervisord_restart_all(roles='webapp_servers'): """ Restarts all of supervisord's managed programs. """ if _current_host_has_role(roles): print("=== RESTARTING SUPERVISORD PROGRAMS ===") with cd(env.REMOTE_CODEBASE_PATH): run("workon %s && ./manage.py ft_supervisord_restart_prog" % env.REMOTE_VIRTUALENV_NAME) def supervisord_restart_prog(program, roles='webapp_servers'): """ Restarts all of supervisord's managed programs. :arg str program: The name of the program to restart (as per supervisor's conf.d/ contents). """ if _current_host_has_role(roles): print("=== RESTARTING SUPERVISORD PROGRAMS ===") with cd(env.REMOTE_CODEBASE_PATH): run("workon %s && ./manage.py ft_supervisord_restart_prog %s" % ( program, env.REMOTE_VIRTUALENV_NAME)) Fix arg order for supervisord_restart_progfrom fabric.api import * from fabtastic.fabric.util import _current_host_has_role def supervisord_restart_all(roles='webapp_servers'): """ Restarts all of supervisord's managed programs. """ if _current_host_has_role(roles): print("=== RESTARTING SUPERVISORD PROGRAMS ===") with cd(env.REMOTE_CODEBASE_PATH): run("workon %s && ./manage.py ft_supervisord_restart_prog" % env.REMOTE_VIRTUALENV_NAME) def supervisord_restart_prog(program, roles='webapp_servers'): """ Restarts all of supervisord's managed programs. :arg str program: The name of the program to restart (as per supervisor's conf.d/ contents). """ if _current_host_has_role(roles): print("=== RESTARTING SUPERVISORD PROGRAMS ===") with cd(env.REMOTE_CODEBASE_PATH): run("workon %s && ./manage.py ft_supervisord_restart_prog %s" % ( env.REMOTE_VIRTUALENV_NAME, program))
<commit_before>from fabric.api import * from fabtastic.fabric.util import _current_host_has_role def supervisord_restart_all(roles='webapp_servers'): """ Restarts all of supervisord's managed programs. """ if _current_host_has_role(roles): print("=== RESTARTING SUPERVISORD PROGRAMS ===") with cd(env.REMOTE_CODEBASE_PATH): run("workon %s && ./manage.py ft_supervisord_restart_prog" % env.REMOTE_VIRTUALENV_NAME) def supervisord_restart_prog(program, roles='webapp_servers'): """ Restarts all of supervisord's managed programs. :arg str program: The name of the program to restart (as per supervisor's conf.d/ contents). """ if _current_host_has_role(roles): print("=== RESTARTING SUPERVISORD PROGRAMS ===") with cd(env.REMOTE_CODEBASE_PATH): run("workon %s && ./manage.py ft_supervisord_restart_prog %s" % ( program, env.REMOTE_VIRTUALENV_NAME)) <commit_msg>Fix arg order for supervisord_restart_prog<commit_after>from fabric.api import * from fabtastic.fabric.util import _current_host_has_role def supervisord_restart_all(roles='webapp_servers'): """ Restarts all of supervisord's managed programs. """ if _current_host_has_role(roles): print("=== RESTARTING SUPERVISORD PROGRAMS ===") with cd(env.REMOTE_CODEBASE_PATH): run("workon %s && ./manage.py ft_supervisord_restart_prog" % env.REMOTE_VIRTUALENV_NAME) def supervisord_restart_prog(program, roles='webapp_servers'): """ Restarts all of supervisord's managed programs. :arg str program: The name of the program to restart (as per supervisor's conf.d/ contents). """ if _current_host_has_role(roles): print("=== RESTARTING SUPERVISORD PROGRAMS ===") with cd(env.REMOTE_CODEBASE_PATH): run("workon %s && ./manage.py ft_supervisord_restart_prog %s" % ( env.REMOTE_VIRTUALENV_NAME, program))
2422e0eb14bc9ae0b79b88f9b02b7e9c7f6ee4fd
tests/window/window_util.py
tests/window/window_util.py
#!/usr/bin/python # $Id:$ from pyglet.gl import * def draw_client_border(window): glClearColor(0, 0, 0, 1) glClear(GL_COLOR_BUFFER_BIT) glMatrixMode(GL_PROJECTION) glLoadIdentity() glOrtho(0, window.width, 0, window.height, -1, 1) glMatrixMode(GL_MODELVIEW) glLoadIdentity() def rect(x1, y1, x2, y2): glBegin(GL_LINE_LOOP) glVertex2f(x1, y1) glVertex2f(x2, y1) glVertex2f(x2, y2) glVertex2f(x1, y2) glEnd() glColor3f(1, 0, 0) rect(-1, -1, window.width + 1, window.height - 1) glColor3f(0, 1, 0) rect(0, 0, window.width - 1, window.height - 1)
#!/usr/bin/python # $Id:$ from pyglet.gl import * def draw_client_border(window): glClearColor(0, 0, 0, 1) glClear(GL_COLOR_BUFFER_BIT) glMatrixMode(GL_PROJECTION) glLoadIdentity() glOrtho(0, window.width, 0, window.height, -1, 1) glMatrixMode(GL_MODELVIEW) glLoadIdentity() def rect(x1, y1, x2, y2): glBegin(GL_LINE_LOOP) glVertex2f(x1, y1) glVertex2f(x2, y1) glVertex2f(x2, y2) glVertex2f(x1, y2) glEnd() glColor3f(1, 0, 0) rect(-1, -1, window.width, window.height) glColor3f(0, 1, 0) rect(0, 0, window.width - 1, window.height - 1)
Fix window test border _again_ (more fixed).
Fix window test border _again_ (more fixed). git-svn-id: d4fdfcd4de20a449196f78acc655f735742cd30d@1383 14d46d22-621c-0410-bb3d-6f67920f7d95
Python
bsd-3-clause
regular/pyglet-avbin-optimizations,regular/pyglet-avbin-optimizations,regular/pyglet-avbin-optimizations,regular/pyglet-avbin-optimizations
#!/usr/bin/python # $Id:$ from pyglet.gl import * def draw_client_border(window): glClearColor(0, 0, 0, 1) glClear(GL_COLOR_BUFFER_BIT) glMatrixMode(GL_PROJECTION) glLoadIdentity() glOrtho(0, window.width, 0, window.height, -1, 1) glMatrixMode(GL_MODELVIEW) glLoadIdentity() def rect(x1, y1, x2, y2): glBegin(GL_LINE_LOOP) glVertex2f(x1, y1) glVertex2f(x2, y1) glVertex2f(x2, y2) glVertex2f(x1, y2) glEnd() glColor3f(1, 0, 0) rect(-1, -1, window.width + 1, window.height - 1) glColor3f(0, 1, 0) rect(0, 0, window.width - 1, window.height - 1) Fix window test border _again_ (more fixed). git-svn-id: d4fdfcd4de20a449196f78acc655f735742cd30d@1383 14d46d22-621c-0410-bb3d-6f67920f7d95
#!/usr/bin/python # $Id:$ from pyglet.gl import * def draw_client_border(window): glClearColor(0, 0, 0, 1) glClear(GL_COLOR_BUFFER_BIT) glMatrixMode(GL_PROJECTION) glLoadIdentity() glOrtho(0, window.width, 0, window.height, -1, 1) glMatrixMode(GL_MODELVIEW) glLoadIdentity() def rect(x1, y1, x2, y2): glBegin(GL_LINE_LOOP) glVertex2f(x1, y1) glVertex2f(x2, y1) glVertex2f(x2, y2) glVertex2f(x1, y2) glEnd() glColor3f(1, 0, 0) rect(-1, -1, window.width, window.height) glColor3f(0, 1, 0) rect(0, 0, window.width - 1, window.height - 1)
<commit_before>#!/usr/bin/python # $Id:$ from pyglet.gl import * def draw_client_border(window): glClearColor(0, 0, 0, 1) glClear(GL_COLOR_BUFFER_BIT) glMatrixMode(GL_PROJECTION) glLoadIdentity() glOrtho(0, window.width, 0, window.height, -1, 1) glMatrixMode(GL_MODELVIEW) glLoadIdentity() def rect(x1, y1, x2, y2): glBegin(GL_LINE_LOOP) glVertex2f(x1, y1) glVertex2f(x2, y1) glVertex2f(x2, y2) glVertex2f(x1, y2) glEnd() glColor3f(1, 0, 0) rect(-1, -1, window.width + 1, window.height - 1) glColor3f(0, 1, 0) rect(0, 0, window.width - 1, window.height - 1) <commit_msg>Fix window test border _again_ (more fixed). git-svn-id: d4fdfcd4de20a449196f78acc655f735742cd30d@1383 14d46d22-621c-0410-bb3d-6f67920f7d95<commit_after>
#!/usr/bin/python # $Id:$ from pyglet.gl import * def draw_client_border(window): glClearColor(0, 0, 0, 1) glClear(GL_COLOR_BUFFER_BIT) glMatrixMode(GL_PROJECTION) glLoadIdentity() glOrtho(0, window.width, 0, window.height, -1, 1) glMatrixMode(GL_MODELVIEW) glLoadIdentity() def rect(x1, y1, x2, y2): glBegin(GL_LINE_LOOP) glVertex2f(x1, y1) glVertex2f(x2, y1) glVertex2f(x2, y2) glVertex2f(x1, y2) glEnd() glColor3f(1, 0, 0) rect(-1, -1, window.width, window.height) glColor3f(0, 1, 0) rect(0, 0, window.width - 1, window.height - 1)
#!/usr/bin/python # $Id:$ from pyglet.gl import * def draw_client_border(window): glClearColor(0, 0, 0, 1) glClear(GL_COLOR_BUFFER_BIT) glMatrixMode(GL_PROJECTION) glLoadIdentity() glOrtho(0, window.width, 0, window.height, -1, 1) glMatrixMode(GL_MODELVIEW) glLoadIdentity() def rect(x1, y1, x2, y2): glBegin(GL_LINE_LOOP) glVertex2f(x1, y1) glVertex2f(x2, y1) glVertex2f(x2, y2) glVertex2f(x1, y2) glEnd() glColor3f(1, 0, 0) rect(-1, -1, window.width + 1, window.height - 1) glColor3f(0, 1, 0) rect(0, 0, window.width - 1, window.height - 1) Fix window test border _again_ (more fixed). git-svn-id: d4fdfcd4de20a449196f78acc655f735742cd30d@1383 14d46d22-621c-0410-bb3d-6f67920f7d95#!/usr/bin/python # $Id:$ from pyglet.gl import * def draw_client_border(window): glClearColor(0, 0, 0, 1) glClear(GL_COLOR_BUFFER_BIT) glMatrixMode(GL_PROJECTION) glLoadIdentity() glOrtho(0, window.width, 0, window.height, -1, 1) glMatrixMode(GL_MODELVIEW) glLoadIdentity() def rect(x1, y1, x2, y2): glBegin(GL_LINE_LOOP) glVertex2f(x1, y1) glVertex2f(x2, y1) glVertex2f(x2, y2) glVertex2f(x1, y2) glEnd() glColor3f(1, 0, 0) rect(-1, -1, window.width, window.height) glColor3f(0, 1, 0) rect(0, 0, window.width - 1, window.height - 1)
<commit_before>#!/usr/bin/python # $Id:$ from pyglet.gl import * def draw_client_border(window): glClearColor(0, 0, 0, 1) glClear(GL_COLOR_BUFFER_BIT) glMatrixMode(GL_PROJECTION) glLoadIdentity() glOrtho(0, window.width, 0, window.height, -1, 1) glMatrixMode(GL_MODELVIEW) glLoadIdentity() def rect(x1, y1, x2, y2): glBegin(GL_LINE_LOOP) glVertex2f(x1, y1) glVertex2f(x2, y1) glVertex2f(x2, y2) glVertex2f(x1, y2) glEnd() glColor3f(1, 0, 0) rect(-1, -1, window.width + 1, window.height - 1) glColor3f(0, 1, 0) rect(0, 0, window.width - 1, window.height - 1) <commit_msg>Fix window test border _again_ (more fixed). git-svn-id: d4fdfcd4de20a449196f78acc655f735742cd30d@1383 14d46d22-621c-0410-bb3d-6f67920f7d95<commit_after>#!/usr/bin/python # $Id:$ from pyglet.gl import * def draw_client_border(window): glClearColor(0, 0, 0, 1) glClear(GL_COLOR_BUFFER_BIT) glMatrixMode(GL_PROJECTION) glLoadIdentity() glOrtho(0, window.width, 0, window.height, -1, 1) glMatrixMode(GL_MODELVIEW) glLoadIdentity() def rect(x1, y1, x2, y2): glBegin(GL_LINE_LOOP) glVertex2f(x1, y1) glVertex2f(x2, y1) glVertex2f(x2, y2) glVertex2f(x1, y2) glEnd() glColor3f(1, 0, 0) rect(-1, -1, window.width, window.height) glColor3f(0, 1, 0) rect(0, 0, window.width - 1, window.height - 1)
86008628f7bff187c956273fbf6f15376ab861d1
src/sgeparse/query.py
src/sgeparse/query.py
#!/usr/bin/env python # -*- coding: utf-8 -*- import subprocess as sp from .parser import JobsParser def get_jobs(): xml_text = fetch_xml() parser = JobsParser(xml_text) return parser.jobs def fetch_xml(user=None): cmd = ['qstat', '-xml'] if user is not None: cmd.extend(['-u', user]) return sp.check_output(list(map(str, cmd)))
#!/usr/bin/env python # -*- coding: utf-8 -*- import subprocess as sp from .parser import JobsParser def get_jobs(user=None): xml_text = fetch_xml(user=user) parser = JobsParser(xml_text) return parser.jobs def fetch_xml(user=None): cmd = ['qstat', '-xml'] if user is not None: cmd.extend(['-u', user]) return sp.check_output(list(map(str, cmd)))
Add user argument to get_jobs
Add user argument to get_jobs
Python
mit
mindriot101/sgeparse
#!/usr/bin/env python # -*- coding: utf-8 -*- import subprocess as sp from .parser import JobsParser def get_jobs(): xml_text = fetch_xml() parser = JobsParser(xml_text) return parser.jobs def fetch_xml(user=None): cmd = ['qstat', '-xml'] if user is not None: cmd.extend(['-u', user]) return sp.check_output(list(map(str, cmd))) Add user argument to get_jobs
#!/usr/bin/env python # -*- coding: utf-8 -*- import subprocess as sp from .parser import JobsParser def get_jobs(user=None): xml_text = fetch_xml(user=user) parser = JobsParser(xml_text) return parser.jobs def fetch_xml(user=None): cmd = ['qstat', '-xml'] if user is not None: cmd.extend(['-u', user]) return sp.check_output(list(map(str, cmd)))
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- import subprocess as sp from .parser import JobsParser def get_jobs(): xml_text = fetch_xml() parser = JobsParser(xml_text) return parser.jobs def fetch_xml(user=None): cmd = ['qstat', '-xml'] if user is not None: cmd.extend(['-u', user]) return sp.check_output(list(map(str, cmd))) <commit_msg>Add user argument to get_jobs<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- import subprocess as sp from .parser import JobsParser def get_jobs(user=None): xml_text = fetch_xml(user=user) parser = JobsParser(xml_text) return parser.jobs def fetch_xml(user=None): cmd = ['qstat', '-xml'] if user is not None: cmd.extend(['-u', user]) return sp.check_output(list(map(str, cmd)))
#!/usr/bin/env python # -*- coding: utf-8 -*- import subprocess as sp from .parser import JobsParser def get_jobs(): xml_text = fetch_xml() parser = JobsParser(xml_text) return parser.jobs def fetch_xml(user=None): cmd = ['qstat', '-xml'] if user is not None: cmd.extend(['-u', user]) return sp.check_output(list(map(str, cmd))) Add user argument to get_jobs#!/usr/bin/env python # -*- coding: utf-8 -*- import subprocess as sp from .parser import JobsParser def get_jobs(user=None): xml_text = fetch_xml(user=user) parser = JobsParser(xml_text) return parser.jobs def fetch_xml(user=None): cmd = ['qstat', '-xml'] if user is not None: cmd.extend(['-u', user]) return sp.check_output(list(map(str, cmd)))
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- import subprocess as sp from .parser import JobsParser def get_jobs(): xml_text = fetch_xml() parser = JobsParser(xml_text) return parser.jobs def fetch_xml(user=None): cmd = ['qstat', '-xml'] if user is not None: cmd.extend(['-u', user]) return sp.check_output(list(map(str, cmd))) <commit_msg>Add user argument to get_jobs<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- import subprocess as sp from .parser import JobsParser def get_jobs(user=None): xml_text = fetch_xml(user=user) parser = JobsParser(xml_text) return parser.jobs def fetch_xml(user=None): cmd = ['qstat', '-xml'] if user is not None: cmd.extend(['-u', user]) return sp.check_output(list(map(str, cmd)))
43a209bd122329d5a70e5f0bdc2066e952676c6a
tests/unit/output/yaml_out_test.py
tests/unit/output/yaml_out_test.py
# -*- coding: utf-8 -*- ''' unittests for yaml outputter ''' # Import Python Libs from __future__ import absolute_import from StringIO import StringIO import sys # Import Salt Testing Libs from salttesting import TestCase from salttesting.helpers import ensure_in_syspath ensure_in_syspath('../../') # Import Salt Libs from salt.output import yaml_out as yaml class YamlTestCase(TestCase): ''' Test cases for salt.output.json_out ''' def setUp(self): # reset to default behavior yaml.__opts__ = {} self.data = {'test': 'two', 'example': 'one'} def test_default_output(self): ret = yaml.output(self.data) expect = 'example: one\ntest: two\n' self.assertEqual(expect, ret) def test_negative_int_output(self): yaml.__opts__['output_indent'] = -1 ret = yaml.output(self.data) expect = '{example: one, test: two}\n' self.assertEqual(expect, ret) if __name__ == '__main__': from integration import run_tests run_tests(YamlTestCase, needs_daemon=False)
# -*- coding: utf-8 -*- ''' unittests for yaml outputter ''' # Import Python Libs from __future__ import absolute_import # Import Salt Testing Libs from salttesting import TestCase from salttesting.helpers import ensure_in_syspath ensure_in_syspath('../../') # Import Salt Libs from salt.output import yaml_out as yaml class YamlTestCase(TestCase): ''' Test cases for salt.output.json_out ''' def setUp(self): # reset to default behavior yaml.__opts__ = {} self.data = {'test': 'two', 'example': 'one'} def test_default_output(self): ret = yaml.output(self.data) expect = 'example: one\ntest: two\n' self.assertEqual(expect, ret) def test_negative_int_output(self): yaml.__opts__['output_indent'] = -1 ret = yaml.output(self.data) expect = '{example: one, test: two}\n' self.assertEqual(expect, ret) if __name__ == '__main__': from integration import run_tests run_tests(YamlTestCase, needs_daemon=False)
Remove unused imports for lint
Remove unused imports for lint
Python
apache-2.0
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
# -*- coding: utf-8 -*- ''' unittests for yaml outputter ''' # Import Python Libs from __future__ import absolute_import from StringIO import StringIO import sys # Import Salt Testing Libs from salttesting import TestCase from salttesting.helpers import ensure_in_syspath ensure_in_syspath('../../') # Import Salt Libs from salt.output import yaml_out as yaml class YamlTestCase(TestCase): ''' Test cases for salt.output.json_out ''' def setUp(self): # reset to default behavior yaml.__opts__ = {} self.data = {'test': 'two', 'example': 'one'} def test_default_output(self): ret = yaml.output(self.data) expect = 'example: one\ntest: two\n' self.assertEqual(expect, ret) def test_negative_int_output(self): yaml.__opts__['output_indent'] = -1 ret = yaml.output(self.data) expect = '{example: one, test: two}\n' self.assertEqual(expect, ret) if __name__ == '__main__': from integration import run_tests run_tests(YamlTestCase, needs_daemon=False) Remove unused imports for lint
# -*- coding: utf-8 -*- ''' unittests for yaml outputter ''' # Import Python Libs from __future__ import absolute_import # Import Salt Testing Libs from salttesting import TestCase from salttesting.helpers import ensure_in_syspath ensure_in_syspath('../../') # Import Salt Libs from salt.output import yaml_out as yaml class YamlTestCase(TestCase): ''' Test cases for salt.output.json_out ''' def setUp(self): # reset to default behavior yaml.__opts__ = {} self.data = {'test': 'two', 'example': 'one'} def test_default_output(self): ret = yaml.output(self.data) expect = 'example: one\ntest: two\n' self.assertEqual(expect, ret) def test_negative_int_output(self): yaml.__opts__['output_indent'] = -1 ret = yaml.output(self.data) expect = '{example: one, test: two}\n' self.assertEqual(expect, ret) if __name__ == '__main__': from integration import run_tests run_tests(YamlTestCase, needs_daemon=False)
<commit_before># -*- coding: utf-8 -*- ''' unittests for yaml outputter ''' # Import Python Libs from __future__ import absolute_import from StringIO import StringIO import sys # Import Salt Testing Libs from salttesting import TestCase from salttesting.helpers import ensure_in_syspath ensure_in_syspath('../../') # Import Salt Libs from salt.output import yaml_out as yaml class YamlTestCase(TestCase): ''' Test cases for salt.output.json_out ''' def setUp(self): # reset to default behavior yaml.__opts__ = {} self.data = {'test': 'two', 'example': 'one'} def test_default_output(self): ret = yaml.output(self.data) expect = 'example: one\ntest: two\n' self.assertEqual(expect, ret) def test_negative_int_output(self): yaml.__opts__['output_indent'] = -1 ret = yaml.output(self.data) expect = '{example: one, test: two}\n' self.assertEqual(expect, ret) if __name__ == '__main__': from integration import run_tests run_tests(YamlTestCase, needs_daemon=False) <commit_msg>Remove unused imports for lint<commit_after>
# -*- coding: utf-8 -*- ''' unittests for yaml outputter ''' # Import Python Libs from __future__ import absolute_import # Import Salt Testing Libs from salttesting import TestCase from salttesting.helpers import ensure_in_syspath ensure_in_syspath('../../') # Import Salt Libs from salt.output import yaml_out as yaml class YamlTestCase(TestCase): ''' Test cases for salt.output.json_out ''' def setUp(self): # reset to default behavior yaml.__opts__ = {} self.data = {'test': 'two', 'example': 'one'} def test_default_output(self): ret = yaml.output(self.data) expect = 'example: one\ntest: two\n' self.assertEqual(expect, ret) def test_negative_int_output(self): yaml.__opts__['output_indent'] = -1 ret = yaml.output(self.data) expect = '{example: one, test: two}\n' self.assertEqual(expect, ret) if __name__ == '__main__': from integration import run_tests run_tests(YamlTestCase, needs_daemon=False)
# -*- coding: utf-8 -*- ''' unittests for yaml outputter ''' # Import Python Libs from __future__ import absolute_import from StringIO import StringIO import sys # Import Salt Testing Libs from salttesting import TestCase from salttesting.helpers import ensure_in_syspath ensure_in_syspath('../../') # Import Salt Libs from salt.output import yaml_out as yaml class YamlTestCase(TestCase): ''' Test cases for salt.output.json_out ''' def setUp(self): # reset to default behavior yaml.__opts__ = {} self.data = {'test': 'two', 'example': 'one'} def test_default_output(self): ret = yaml.output(self.data) expect = 'example: one\ntest: two\n' self.assertEqual(expect, ret) def test_negative_int_output(self): yaml.__opts__['output_indent'] = -1 ret = yaml.output(self.data) expect = '{example: one, test: two}\n' self.assertEqual(expect, ret) if __name__ == '__main__': from integration import run_tests run_tests(YamlTestCase, needs_daemon=False) Remove unused imports for lint# -*- coding: utf-8 -*- ''' unittests for yaml outputter ''' # Import Python Libs from __future__ import absolute_import # Import Salt Testing Libs from salttesting import TestCase from salttesting.helpers import ensure_in_syspath ensure_in_syspath('../../') # Import Salt Libs from salt.output import yaml_out as yaml class YamlTestCase(TestCase): ''' Test cases for salt.output.json_out ''' def setUp(self): # reset to default behavior yaml.__opts__ = {} self.data = {'test': 'two', 'example': 'one'} def test_default_output(self): ret = yaml.output(self.data) expect = 'example: one\ntest: two\n' self.assertEqual(expect, ret) def test_negative_int_output(self): yaml.__opts__['output_indent'] = -1 ret = yaml.output(self.data) expect = '{example: one, test: two}\n' self.assertEqual(expect, ret) if __name__ == '__main__': from integration import run_tests run_tests(YamlTestCase, needs_daemon=False)
<commit_before># -*- coding: utf-8 -*- ''' unittests for yaml outputter ''' # Import Python Libs from __future__ import absolute_import from StringIO import StringIO import sys # Import Salt Testing Libs from salttesting import TestCase from salttesting.helpers import ensure_in_syspath ensure_in_syspath('../../') # Import Salt Libs from salt.output import yaml_out as yaml class YamlTestCase(TestCase): ''' Test cases for salt.output.json_out ''' def setUp(self): # reset to default behavior yaml.__opts__ = {} self.data = {'test': 'two', 'example': 'one'} def test_default_output(self): ret = yaml.output(self.data) expect = 'example: one\ntest: two\n' self.assertEqual(expect, ret) def test_negative_int_output(self): yaml.__opts__['output_indent'] = -1 ret = yaml.output(self.data) expect = '{example: one, test: two}\n' self.assertEqual(expect, ret) if __name__ == '__main__': from integration import run_tests run_tests(YamlTestCase, needs_daemon=False) <commit_msg>Remove unused imports for lint<commit_after># -*- coding: utf-8 -*- ''' unittests for yaml outputter ''' # Import Python Libs from __future__ import absolute_import # Import Salt Testing Libs from salttesting import TestCase from salttesting.helpers import ensure_in_syspath ensure_in_syspath('../../') # Import Salt Libs from salt.output import yaml_out as yaml class YamlTestCase(TestCase): ''' Test cases for salt.output.json_out ''' def setUp(self): # reset to default behavior yaml.__opts__ = {} self.data = {'test': 'two', 'example': 'one'} def test_default_output(self): ret = yaml.output(self.data) expect = 'example: one\ntest: two\n' self.assertEqual(expect, ret) def test_negative_int_output(self): yaml.__opts__['output_indent'] = -1 ret = yaml.output(self.data) expect = '{example: one, test: two}\n' self.assertEqual(expect, ret) if __name__ == '__main__': from integration import run_tests run_tests(YamlTestCase, needs_daemon=False)
b970f230864b40eaddb8e5faa76538c9f8e5c59c
txircd/modules/rfc/cmd_userhost.py
txircd/modules/rfc/cmd_userhost.py
from twisted.plugin import IPlugin from twisted.words.protocols import irc from txircd.module_interface import Command, ICommand, IModuleData, ModuleData from zope.interface import implements class UserhostCommand(ModuleData, Command): implements(IPlugin, IModuleData, ICommand) name = "UserhostCommand" core = True def userCommands(self): return [ ("USERHOST", 1, self) ] def parseParams(self, user, params, prefix, tags): if not params: user.sendSingleError("UserhostParams", irc.ERR_NEEDMOREPARAMS, "USERHOST", "Not enough parameters") return None return { "nicks": params[:5] } def execute(self, user, data): userHosts = [] for nick in data["nicks"]: if nick not in self.ircd.userNicks: continue targetUser = self.ircd.users[self.ircd.userNicks[nick]] output = targetUser.nick if self.ircd.runActionUntilValue("userhasoperpermission", targetUser, ""): output += "*" output += "=" if user.metadataKeyExists("away"): output += "-" else: output += "+" output += "{}@{}".format(targetUser.ident, targetUser.host()) userHosts.append(output) user.sendMessage(irc.RPL_USERHOST, " ".join(userHosts)) return True userhostCmd = UserhostCommand()
from twisted.plugin import IPlugin from twisted.words.protocols import irc from txircd.module_interface import Command, ICommand, IModuleData, ModuleData from zope.interface import implements class UserhostCommand(ModuleData, Command): implements(IPlugin, IModuleData, ICommand) name = "UserhostCommand" core = True def userCommands(self): return [ ("USERHOST", 1, self) ] def parseParams(self, user, params, prefix, tags): if not params: user.sendSingleError("UserhostParams", irc.ERR_NEEDMOREPARAMS, "USERHOST", "Not enough parameters") return None return { "nicks": params[:5] } def execute(self, user, data): userHosts = [] for nick in data["nicks"]: if nick not in self.ircd.userNicks: continue targetUser = self.ircd.users[self.ircd.userNicks[nick]] output = targetUser.nick if self.ircd.runActionUntilValue("userhasoperpermission", targetUser, "", users=[targetUser]): output += "*" output += "=" if user.metadataKeyExists("away"): output += "-" else: output += "+" output += "{}@{}".format(targetUser.ident, targetUser.host()) userHosts.append(output) user.sendMessage(irc.RPL_USERHOST, " ".join(userHosts)) return True userhostCmd = UserhostCommand()
Add affected users to userhasoperpermission call in USERHOST
Add affected users to userhasoperpermission call in USERHOST
Python
bsd-3-clause
Heufneutje/txircd,ElementalAlchemist/txircd
from twisted.plugin import IPlugin from twisted.words.protocols import irc from txircd.module_interface import Command, ICommand, IModuleData, ModuleData from zope.interface import implements class UserhostCommand(ModuleData, Command): implements(IPlugin, IModuleData, ICommand) name = "UserhostCommand" core = True def userCommands(self): return [ ("USERHOST", 1, self) ] def parseParams(self, user, params, prefix, tags): if not params: user.sendSingleError("UserhostParams", irc.ERR_NEEDMOREPARAMS, "USERHOST", "Not enough parameters") return None return { "nicks": params[:5] } def execute(self, user, data): userHosts = [] for nick in data["nicks"]: if nick not in self.ircd.userNicks: continue targetUser = self.ircd.users[self.ircd.userNicks[nick]] output = targetUser.nick if self.ircd.runActionUntilValue("userhasoperpermission", targetUser, ""): output += "*" output += "=" if user.metadataKeyExists("away"): output += "-" else: output += "+" output += "{}@{}".format(targetUser.ident, targetUser.host()) userHosts.append(output) user.sendMessage(irc.RPL_USERHOST, " ".join(userHosts)) return True userhostCmd = UserhostCommand()Add affected users to userhasoperpermission call in USERHOST
from twisted.plugin import IPlugin from twisted.words.protocols import irc from txircd.module_interface import Command, ICommand, IModuleData, ModuleData from zope.interface import implements class UserhostCommand(ModuleData, Command): implements(IPlugin, IModuleData, ICommand) name = "UserhostCommand" core = True def userCommands(self): return [ ("USERHOST", 1, self) ] def parseParams(self, user, params, prefix, tags): if not params: user.sendSingleError("UserhostParams", irc.ERR_NEEDMOREPARAMS, "USERHOST", "Not enough parameters") return None return { "nicks": params[:5] } def execute(self, user, data): userHosts = [] for nick in data["nicks"]: if nick not in self.ircd.userNicks: continue targetUser = self.ircd.users[self.ircd.userNicks[nick]] output = targetUser.nick if self.ircd.runActionUntilValue("userhasoperpermission", targetUser, "", users=[targetUser]): output += "*" output += "=" if user.metadataKeyExists("away"): output += "-" else: output += "+" output += "{}@{}".format(targetUser.ident, targetUser.host()) userHosts.append(output) user.sendMessage(irc.RPL_USERHOST, " ".join(userHosts)) return True userhostCmd = UserhostCommand()
<commit_before>from twisted.plugin import IPlugin from twisted.words.protocols import irc from txircd.module_interface import Command, ICommand, IModuleData, ModuleData from zope.interface import implements class UserhostCommand(ModuleData, Command): implements(IPlugin, IModuleData, ICommand) name = "UserhostCommand" core = True def userCommands(self): return [ ("USERHOST", 1, self) ] def parseParams(self, user, params, prefix, tags): if not params: user.sendSingleError("UserhostParams", irc.ERR_NEEDMOREPARAMS, "USERHOST", "Not enough parameters") return None return { "nicks": params[:5] } def execute(self, user, data): userHosts = [] for nick in data["nicks"]: if nick not in self.ircd.userNicks: continue targetUser = self.ircd.users[self.ircd.userNicks[nick]] output = targetUser.nick if self.ircd.runActionUntilValue("userhasoperpermission", targetUser, ""): output += "*" output += "=" if user.metadataKeyExists("away"): output += "-" else: output += "+" output += "{}@{}".format(targetUser.ident, targetUser.host()) userHosts.append(output) user.sendMessage(irc.RPL_USERHOST, " ".join(userHosts)) return True userhostCmd = UserhostCommand()<commit_msg>Add affected users to userhasoperpermission call in USERHOST<commit_after>
from twisted.plugin import IPlugin from twisted.words.protocols import irc from txircd.module_interface import Command, ICommand, IModuleData, ModuleData from zope.interface import implements class UserhostCommand(ModuleData, Command): implements(IPlugin, IModuleData, ICommand) name = "UserhostCommand" core = True def userCommands(self): return [ ("USERHOST", 1, self) ] def parseParams(self, user, params, prefix, tags): if not params: user.sendSingleError("UserhostParams", irc.ERR_NEEDMOREPARAMS, "USERHOST", "Not enough parameters") return None return { "nicks": params[:5] } def execute(self, user, data): userHosts = [] for nick in data["nicks"]: if nick not in self.ircd.userNicks: continue targetUser = self.ircd.users[self.ircd.userNicks[nick]] output = targetUser.nick if self.ircd.runActionUntilValue("userhasoperpermission", targetUser, "", users=[targetUser]): output += "*" output += "=" if user.metadataKeyExists("away"): output += "-" else: output += "+" output += "{}@{}".format(targetUser.ident, targetUser.host()) userHosts.append(output) user.sendMessage(irc.RPL_USERHOST, " ".join(userHosts)) return True userhostCmd = UserhostCommand()
from twisted.plugin import IPlugin from twisted.words.protocols import irc from txircd.module_interface import Command, ICommand, IModuleData, ModuleData from zope.interface import implements class UserhostCommand(ModuleData, Command): implements(IPlugin, IModuleData, ICommand) name = "UserhostCommand" core = True def userCommands(self): return [ ("USERHOST", 1, self) ] def parseParams(self, user, params, prefix, tags): if not params: user.sendSingleError("UserhostParams", irc.ERR_NEEDMOREPARAMS, "USERHOST", "Not enough parameters") return None return { "nicks": params[:5] } def execute(self, user, data): userHosts = [] for nick in data["nicks"]: if nick not in self.ircd.userNicks: continue targetUser = self.ircd.users[self.ircd.userNicks[nick]] output = targetUser.nick if self.ircd.runActionUntilValue("userhasoperpermission", targetUser, ""): output += "*" output += "=" if user.metadataKeyExists("away"): output += "-" else: output += "+" output += "{}@{}".format(targetUser.ident, targetUser.host()) userHosts.append(output) user.sendMessage(irc.RPL_USERHOST, " ".join(userHosts)) return True userhostCmd = UserhostCommand()Add affected users to userhasoperpermission call in USERHOSTfrom twisted.plugin import IPlugin from twisted.words.protocols import irc from txircd.module_interface import Command, ICommand, IModuleData, ModuleData from zope.interface import implements class UserhostCommand(ModuleData, Command): implements(IPlugin, IModuleData, ICommand) name = "UserhostCommand" core = True def userCommands(self): return [ ("USERHOST", 1, self) ] def parseParams(self, user, params, prefix, tags): if not params: user.sendSingleError("UserhostParams", irc.ERR_NEEDMOREPARAMS, "USERHOST", "Not enough parameters") return None return { "nicks": params[:5] } def execute(self, user, data): userHosts = [] for nick in data["nicks"]: if nick not in self.ircd.userNicks: continue targetUser = self.ircd.users[self.ircd.userNicks[nick]] output = targetUser.nick if self.ircd.runActionUntilValue("userhasoperpermission", targetUser, "", users=[targetUser]): output += "*" output += "=" if user.metadataKeyExists("away"): output += "-" else: output += "+" output += "{}@{}".format(targetUser.ident, targetUser.host()) userHosts.append(output) user.sendMessage(irc.RPL_USERHOST, " ".join(userHosts)) return True userhostCmd = UserhostCommand()
<commit_before>from twisted.plugin import IPlugin from twisted.words.protocols import irc from txircd.module_interface import Command, ICommand, IModuleData, ModuleData from zope.interface import implements class UserhostCommand(ModuleData, Command): implements(IPlugin, IModuleData, ICommand) name = "UserhostCommand" core = True def userCommands(self): return [ ("USERHOST", 1, self) ] def parseParams(self, user, params, prefix, tags): if not params: user.sendSingleError("UserhostParams", irc.ERR_NEEDMOREPARAMS, "USERHOST", "Not enough parameters") return None return { "nicks": params[:5] } def execute(self, user, data): userHosts = [] for nick in data["nicks"]: if nick not in self.ircd.userNicks: continue targetUser = self.ircd.users[self.ircd.userNicks[nick]] output = targetUser.nick if self.ircd.runActionUntilValue("userhasoperpermission", targetUser, ""): output += "*" output += "=" if user.metadataKeyExists("away"): output += "-" else: output += "+" output += "{}@{}".format(targetUser.ident, targetUser.host()) userHosts.append(output) user.sendMessage(irc.RPL_USERHOST, " ".join(userHosts)) return True userhostCmd = UserhostCommand()<commit_msg>Add affected users to userhasoperpermission call in USERHOST<commit_after>from twisted.plugin import IPlugin from twisted.words.protocols import irc from txircd.module_interface import Command, ICommand, IModuleData, ModuleData from zope.interface import implements class UserhostCommand(ModuleData, Command): implements(IPlugin, IModuleData, ICommand) name = "UserhostCommand" core = True def userCommands(self): return [ ("USERHOST", 1, self) ] def parseParams(self, user, params, prefix, tags): if not params: user.sendSingleError("UserhostParams", irc.ERR_NEEDMOREPARAMS, "USERHOST", "Not enough parameters") return None return { "nicks": params[:5] } def execute(self, user, data): userHosts = [] for nick in data["nicks"]: if nick not in self.ircd.userNicks: continue targetUser = self.ircd.users[self.ircd.userNicks[nick]] output = targetUser.nick if self.ircd.runActionUntilValue("userhasoperpermission", targetUser, "", users=[targetUser]): output += "*" output += "=" if user.metadataKeyExists("away"): output += "-" else: output += "+" output += "{}@{}".format(targetUser.ident, targetUser.host()) userHosts.append(output) user.sendMessage(irc.RPL_USERHOST, " ".join(userHosts)) return True userhostCmd = UserhostCommand()
4c58426a88ba056841b1d1b44536f2f85de120cc
pythonx/completers/javascript/__init__.py
pythonx/completers/javascript/__init__.py
# -*- coding: utf-8 -*- import json import os.path import re from completor import Completor from completor.compat import to_unicode dirname = os.path.dirname(__file__) class Tern(Completor): filetype = 'javascript' daemon = True ident = re.compile(r"""(\w+)|(('|").+)""", re.U) trigger = r"""\w+$|[\w\)\]\}\'\"]+\.\w*$|('|").*$""" def format_cmd(self): binary = self.get_option('node_binary') or 'node' tern_config = self.find_config_file('.tern-project') cmd = [binary, os.path.join(dirname, 'tern_wrapper.js')] if tern_config: cmd.append(os.path.dirname(tern_config)) return cmd def parse(self, data): try: data = to_unicode(data[0], 'utf-8') return [i for i in json.loads(data) if not self.input_data.endswith(i['word'])] except Exception: return []
# -*- coding: utf-8 -*- import json import os.path import re from completor import Completor from completor.compat import to_unicode dirname = os.path.dirname(__file__) class Tern(Completor): filetype = 'javascript' daemon = True ident = re.compile(r"""(\w+)|(["'][^"']*)""", re.U) trigger = r"""\w+$|[\w\)\]\}\'\"]+\.\w*$|["']\w*$""" def format_cmd(self): binary = self.get_option('node_binary') or 'node' tern_config = self.find_config_file('.tern-project') cmd = [binary, os.path.join(dirname, 'tern_wrapper.js')] if tern_config: cmd.append(os.path.dirname(tern_config)) return cmd def parse(self, data): try: data = to_unicode(data[0], 'utf-8') return [i for i in json.loads(data) if not self.input_data.endswith(i['word'])] except Exception: return []
Fix regex for tern complete_strings plugin
Fix regex for tern complete_strings plugin
Python
mit
maralla/completor.vim,maralla/completor.vim
# -*- coding: utf-8 -*- import json import os.path import re from completor import Completor from completor.compat import to_unicode dirname = os.path.dirname(__file__) class Tern(Completor): filetype = 'javascript' daemon = True ident = re.compile(r"""(\w+)|(('|").+)""", re.U) trigger = r"""\w+$|[\w\)\]\}\'\"]+\.\w*$|('|").*$""" def format_cmd(self): binary = self.get_option('node_binary') or 'node' tern_config = self.find_config_file('.tern-project') cmd = [binary, os.path.join(dirname, 'tern_wrapper.js')] if tern_config: cmd.append(os.path.dirname(tern_config)) return cmd def parse(self, data): try: data = to_unicode(data[0], 'utf-8') return [i for i in json.loads(data) if not self.input_data.endswith(i['word'])] except Exception: return [] Fix regex for tern complete_strings plugin
# -*- coding: utf-8 -*- import json import os.path import re from completor import Completor from completor.compat import to_unicode dirname = os.path.dirname(__file__) class Tern(Completor): filetype = 'javascript' daemon = True ident = re.compile(r"""(\w+)|(["'][^"']*)""", re.U) trigger = r"""\w+$|[\w\)\]\}\'\"]+\.\w*$|["']\w*$""" def format_cmd(self): binary = self.get_option('node_binary') or 'node' tern_config = self.find_config_file('.tern-project') cmd = [binary, os.path.join(dirname, 'tern_wrapper.js')] if tern_config: cmd.append(os.path.dirname(tern_config)) return cmd def parse(self, data): try: data = to_unicode(data[0], 'utf-8') return [i for i in json.loads(data) if not self.input_data.endswith(i['word'])] except Exception: return []
<commit_before># -*- coding: utf-8 -*- import json import os.path import re from completor import Completor from completor.compat import to_unicode dirname = os.path.dirname(__file__) class Tern(Completor): filetype = 'javascript' daemon = True ident = re.compile(r"""(\w+)|(('|").+)""", re.U) trigger = r"""\w+$|[\w\)\]\}\'\"]+\.\w*$|('|").*$""" def format_cmd(self): binary = self.get_option('node_binary') or 'node' tern_config = self.find_config_file('.tern-project') cmd = [binary, os.path.join(dirname, 'tern_wrapper.js')] if tern_config: cmd.append(os.path.dirname(tern_config)) return cmd def parse(self, data): try: data = to_unicode(data[0], 'utf-8') return [i for i in json.loads(data) if not self.input_data.endswith(i['word'])] except Exception: return [] <commit_msg>Fix regex for tern complete_strings plugin<commit_after>
# -*- coding: utf-8 -*- import json import os.path import re from completor import Completor from completor.compat import to_unicode dirname = os.path.dirname(__file__) class Tern(Completor): filetype = 'javascript' daemon = True ident = re.compile(r"""(\w+)|(["'][^"']*)""", re.U) trigger = r"""\w+$|[\w\)\]\}\'\"]+\.\w*$|["']\w*$""" def format_cmd(self): binary = self.get_option('node_binary') or 'node' tern_config = self.find_config_file('.tern-project') cmd = [binary, os.path.join(dirname, 'tern_wrapper.js')] if tern_config: cmd.append(os.path.dirname(tern_config)) return cmd def parse(self, data): try: data = to_unicode(data[0], 'utf-8') return [i for i in json.loads(data) if not self.input_data.endswith(i['word'])] except Exception: return []
# -*- coding: utf-8 -*- import json import os.path import re from completor import Completor from completor.compat import to_unicode dirname = os.path.dirname(__file__) class Tern(Completor): filetype = 'javascript' daemon = True ident = re.compile(r"""(\w+)|(('|").+)""", re.U) trigger = r"""\w+$|[\w\)\]\}\'\"]+\.\w*$|('|").*$""" def format_cmd(self): binary = self.get_option('node_binary') or 'node' tern_config = self.find_config_file('.tern-project') cmd = [binary, os.path.join(dirname, 'tern_wrapper.js')] if tern_config: cmd.append(os.path.dirname(tern_config)) return cmd def parse(self, data): try: data = to_unicode(data[0], 'utf-8') return [i for i in json.loads(data) if not self.input_data.endswith(i['word'])] except Exception: return [] Fix regex for tern complete_strings plugin# -*- coding: utf-8 -*- import json import os.path import re from completor import Completor from completor.compat import to_unicode dirname = os.path.dirname(__file__) class Tern(Completor): filetype = 'javascript' daemon = True ident = re.compile(r"""(\w+)|(["'][^"']*)""", re.U) trigger = r"""\w+$|[\w\)\]\}\'\"]+\.\w*$|["']\w*$""" def format_cmd(self): binary = self.get_option('node_binary') or 'node' tern_config = self.find_config_file('.tern-project') cmd = [binary, os.path.join(dirname, 'tern_wrapper.js')] if tern_config: cmd.append(os.path.dirname(tern_config)) return cmd def parse(self, data): try: data = to_unicode(data[0], 'utf-8') return [i for i in json.loads(data) if not self.input_data.endswith(i['word'])] except Exception: return []
<commit_before># -*- coding: utf-8 -*- import json import os.path import re from completor import Completor from completor.compat import to_unicode dirname = os.path.dirname(__file__) class Tern(Completor): filetype = 'javascript' daemon = True ident = re.compile(r"""(\w+)|(('|").+)""", re.U) trigger = r"""\w+$|[\w\)\]\}\'\"]+\.\w*$|('|").*$""" def format_cmd(self): binary = self.get_option('node_binary') or 'node' tern_config = self.find_config_file('.tern-project') cmd = [binary, os.path.join(dirname, 'tern_wrapper.js')] if tern_config: cmd.append(os.path.dirname(tern_config)) return cmd def parse(self, data): try: data = to_unicode(data[0], 'utf-8') return [i for i in json.loads(data) if not self.input_data.endswith(i['word'])] except Exception: return [] <commit_msg>Fix regex for tern complete_strings plugin<commit_after># -*- coding: utf-8 -*- import json import os.path import re from completor import Completor from completor.compat import to_unicode dirname = os.path.dirname(__file__) class Tern(Completor): filetype = 'javascript' daemon = True ident = re.compile(r"""(\w+)|(["'][^"']*)""", re.U) trigger = r"""\w+$|[\w\)\]\}\'\"]+\.\w*$|["']\w*$""" def format_cmd(self): binary = self.get_option('node_binary') or 'node' tern_config = self.find_config_file('.tern-project') cmd = [binary, os.path.join(dirname, 'tern_wrapper.js')] if tern_config: cmd.append(os.path.dirname(tern_config)) return cmd def parse(self, data): try: data = to_unicode(data[0], 'utf-8') return [i for i in json.loads(data) if not self.input_data.endswith(i['word'])] except Exception: return []
2890660ee3e87eb9af2c81caac0dc3131a264310
app.py
app.py
from flask import Flask from flask import jsonify from red_api import get_suspect_mentions, get_tweet_count, get_suspect_tweets from jsonp_flask import support_jsonp app = Flask(__name__) @app.route('/api/tweets/suspects/count', methods=['GET']) @support_jsonp def get_mentions(): return jsonify(results=get_suspect_mentions()) @app.route('/api/tweets/count', methods=['GET']) @support_jsonp def get_total_tweet_count(): return jsonify(results=get_tweet_count()) @app.route('/api/tweets/<suspect>/<limit>') @support_jsonp def get_tweets(suspect, limit): return jsonify(results=get_suspect_tweets(suspect, limit)) if __name__ == '__main__': app.run(debug=True, host='0.0.0.0')
from flask import Flask from flask import jsonify from red_api import get_suspect_mentions, get_tweet_count, get_suspect_tweets from jsonp_flask import support_jsonp app = Flask(__name__) @app.route('/api/tweets/suspects/count', methods=['GET']) @support_jsonp def get_mentions(): return jsonify(results=get_suspect_mentions()) @app.route('/api/tweets/count', methods=['GET']) @support_jsonp def get_total_tweet_count(): return jsonify(results=get_tweet_count()) @app.route('/api/tweets/<suspect>/<limit>') @support_jsonp def get_tweets(suspect, limit): return jsonify(results=get_suspect_tweets(suspect, int(limit))) if __name__ == '__main__': app.run(debug=True, host='0.0.0.0')
Make sure the limit is an int
Make sure the limit is an int
Python
mit
AnSavvides/redjohn,AnSavvides/redjohn
from flask import Flask from flask import jsonify from red_api import get_suspect_mentions, get_tweet_count, get_suspect_tweets from jsonp_flask import support_jsonp app = Flask(__name__) @app.route('/api/tweets/suspects/count', methods=['GET']) @support_jsonp def get_mentions(): return jsonify(results=get_suspect_mentions()) @app.route('/api/tweets/count', methods=['GET']) @support_jsonp def get_total_tweet_count(): return jsonify(results=get_tweet_count()) @app.route('/api/tweets/<suspect>/<limit>') @support_jsonp def get_tweets(suspect, limit): return jsonify(results=get_suspect_tweets(suspect, limit)) if __name__ == '__main__': app.run(debug=True, host='0.0.0.0')Make sure the limit is an int
from flask import Flask from flask import jsonify from red_api import get_suspect_mentions, get_tweet_count, get_suspect_tweets from jsonp_flask import support_jsonp app = Flask(__name__) @app.route('/api/tweets/suspects/count', methods=['GET']) @support_jsonp def get_mentions(): return jsonify(results=get_suspect_mentions()) @app.route('/api/tweets/count', methods=['GET']) @support_jsonp def get_total_tweet_count(): return jsonify(results=get_tweet_count()) @app.route('/api/tweets/<suspect>/<limit>') @support_jsonp def get_tweets(suspect, limit): return jsonify(results=get_suspect_tweets(suspect, int(limit))) if __name__ == '__main__': app.run(debug=True, host='0.0.0.0')
<commit_before>from flask import Flask from flask import jsonify from red_api import get_suspect_mentions, get_tweet_count, get_suspect_tweets from jsonp_flask import support_jsonp app = Flask(__name__) @app.route('/api/tweets/suspects/count', methods=['GET']) @support_jsonp def get_mentions(): return jsonify(results=get_suspect_mentions()) @app.route('/api/tweets/count', methods=['GET']) @support_jsonp def get_total_tweet_count(): return jsonify(results=get_tweet_count()) @app.route('/api/tweets/<suspect>/<limit>') @support_jsonp def get_tweets(suspect, limit): return jsonify(results=get_suspect_tweets(suspect, limit)) if __name__ == '__main__': app.run(debug=True, host='0.0.0.0')<commit_msg>Make sure the limit is an int<commit_after>
from flask import Flask from flask import jsonify from red_api import get_suspect_mentions, get_tweet_count, get_suspect_tweets from jsonp_flask import support_jsonp app = Flask(__name__) @app.route('/api/tweets/suspects/count', methods=['GET']) @support_jsonp def get_mentions(): return jsonify(results=get_suspect_mentions()) @app.route('/api/tweets/count', methods=['GET']) @support_jsonp def get_total_tweet_count(): return jsonify(results=get_tweet_count()) @app.route('/api/tweets/<suspect>/<limit>') @support_jsonp def get_tweets(suspect, limit): return jsonify(results=get_suspect_tweets(suspect, int(limit))) if __name__ == '__main__': app.run(debug=True, host='0.0.0.0')
from flask import Flask from flask import jsonify from red_api import get_suspect_mentions, get_tweet_count, get_suspect_tweets from jsonp_flask import support_jsonp app = Flask(__name__) @app.route('/api/tweets/suspects/count', methods=['GET']) @support_jsonp def get_mentions(): return jsonify(results=get_suspect_mentions()) @app.route('/api/tweets/count', methods=['GET']) @support_jsonp def get_total_tweet_count(): return jsonify(results=get_tweet_count()) @app.route('/api/tweets/<suspect>/<limit>') @support_jsonp def get_tweets(suspect, limit): return jsonify(results=get_suspect_tweets(suspect, limit)) if __name__ == '__main__': app.run(debug=True, host='0.0.0.0')Make sure the limit is an intfrom flask import Flask from flask import jsonify from red_api import get_suspect_mentions, get_tweet_count, get_suspect_tweets from jsonp_flask import support_jsonp app = Flask(__name__) @app.route('/api/tweets/suspects/count', methods=['GET']) @support_jsonp def get_mentions(): return jsonify(results=get_suspect_mentions()) @app.route('/api/tweets/count', methods=['GET']) @support_jsonp def get_total_tweet_count(): return jsonify(results=get_tweet_count()) @app.route('/api/tweets/<suspect>/<limit>') @support_jsonp def get_tweets(suspect, limit): return jsonify(results=get_suspect_tweets(suspect, int(limit))) if __name__ == '__main__': app.run(debug=True, host='0.0.0.0')
<commit_before>from flask import Flask from flask import jsonify from red_api import get_suspect_mentions, get_tweet_count, get_suspect_tweets from jsonp_flask import support_jsonp app = Flask(__name__) @app.route('/api/tweets/suspects/count', methods=['GET']) @support_jsonp def get_mentions(): return jsonify(results=get_suspect_mentions()) @app.route('/api/tweets/count', methods=['GET']) @support_jsonp def get_total_tweet_count(): return jsonify(results=get_tweet_count()) @app.route('/api/tweets/<suspect>/<limit>') @support_jsonp def get_tweets(suspect, limit): return jsonify(results=get_suspect_tweets(suspect, limit)) if __name__ == '__main__': app.run(debug=True, host='0.0.0.0')<commit_msg>Make sure the limit is an int<commit_after>from flask import Flask from flask import jsonify from red_api import get_suspect_mentions, get_tweet_count, get_suspect_tweets from jsonp_flask import support_jsonp app = Flask(__name__) @app.route('/api/tweets/suspects/count', methods=['GET']) @support_jsonp def get_mentions(): return jsonify(results=get_suspect_mentions()) @app.route('/api/tweets/count', methods=['GET']) @support_jsonp def get_total_tweet_count(): return jsonify(results=get_tweet_count()) @app.route('/api/tweets/<suspect>/<limit>') @support_jsonp def get_tweets(suspect, limit): return jsonify(results=get_suspect_tweets(suspect, int(limit))) if __name__ == '__main__': app.run(debug=True, host='0.0.0.0')
7ac8ae993a30ce6ea221e2474df4a8eb7eada1ef
scrapy/trunk/scrapy/conf/core_settings.py
scrapy/trunk/scrapy/conf/core_settings.py
import scrapy # Scrapy core settings BOT_NAME = 'scrapy' BOT_VERSION = scrapy.__version__ ENGINE_DEBUG = False # Download configuration options USER_AGENT = '%s/%s' % (BOT_NAME, BOT_VERSION) DOWNLOAD_TIMEOUT = 180 # 3mins CONCURRENT_DOMAINS = 8 # number of domains to scrape in parallel REQUESTS_PER_DOMAIN = 8 # max simultaneous requests per domain CACHE2_EXPIRATION_SECS = 48 * 60 * 60 # seconds while cached response is still valid (a negative value means "never expires") LOG_ENABLED = True # LOGLEVEL = 'DEBUG' # default loglevel LOGFILE = None # None means sys.stderr by default LOG_STDOUT = False # DEFAULT_ITEM_CLASS = 'scrapy.item.ScrapedItem' SCHEDULER = 'scrapy.core.scheduler.Scheduler' MEMORYSTORE = 'scrapy.core.scheduler.MemoryStore' PRIORITIZER = 'scrapy.core.prioritizers.RandomPrioritizer' EXTENSIONS = [] # contrib.middleware.retry.RetryMiddleware default settings RETRY_TIMES = 3 RETRY_HTTP_CODES = ['500', '503', '504', '400', '408', '200'] # amazon s3 images backend sometimes fails to authenticate or validate http protocol RETRY_HTTP_CODES = RETRY_HTTP_CODES + ['505', '403']
import scrapy # Scrapy core settings BOT_NAME = 'scrapy' BOT_VERSION = scrapy.__version__ ENGINE_DEBUG = False # Download configuration options USER_AGENT = '%s/%s' % (BOT_NAME, BOT_VERSION) DOWNLOAD_TIMEOUT = 180 # 3mins CONCURRENT_DOMAINS = 8 # number of domains to scrape in parallel REQUESTS_PER_DOMAIN = 8 # max simultaneous requests per domain CACHE2_EXPIRATION_SECS = 48 * 60 * 60 # seconds while cached response is still valid (a negative value means "never expires") LOG_ENABLED = True # LOGLEVEL = 'DEBUG' # default loglevel LOGFILE = None # None means sys.stderr by default LOG_STDOUT = False # DEFAULT_ITEM_CLASS = 'scrapy.item.ScrapedItem' SCHEDULER = 'scrapy.core.scheduler.Scheduler' MEMORYSTORE = 'scrapy.core.scheduler.MemoryStore' PRIORITIZER = 'scrapy.core.prioritizers.RandomPrioritizer' EXTENSIONS = [] # contrib.middleware.retry.RetryMiddleware default settings RETRY_TIMES = 3 RETRY_HTTP_CODES = ['500', '503', '504', '400', '408', '200']
Revert "add 505 and 403 to retry status codes due to amazon s3 random fails while uploading images"
Revert "add 505 and 403 to retry status codes due to amazon s3 random fails while uploading images" This reverts changeset r457 --HG-- extra : convert_revision : svn%3Ab85faa78-f9eb-468e-a121-7cced6da292c%40461
Python
bsd-3-clause
AaronTao1990/scrapy,yidongliu/scrapy,ArturGaspar/scrapy,tagatac/scrapy,URXtech/scrapy,olafdietsche/scrapy,w495/scrapy,agusc/scrapy,Preetwinder/scrapy,foromer4/scrapy,zorojean/scrapy,xiao26/scrapy,kmike/scrapy,w495/scrapy,ENjOyAbLE1991/scrapy,Digenis/scrapy,haiiiiiyun/scrapy,eLRuLL/scrapy,wujuguang/scrapy,scrapy/scrapy,mlyundin/scrapy,jiezhu2007/scrapy,rolando-contrib/scrapy,taito/scrapy,umrashrf/scrapy,heamon7/scrapy,fpy171/scrapy,irwinlove/scrapy,kmike/scrapy,fontenele/scrapy,carlosp420/scrapy,redapple/scrapy,github-account-because-they-want-it/scrapy,webmakin/scrapy,nfunato/scrapy,rahulsharma1991/scrapy,cursesun/scrapy,foromer4/scrapy,rahul-c1/scrapy,tntC4stl3/scrapy,jc0n/scrapy,johnardavies/scrapy,ylcolala/scrapy,curita/scrapy,CodeJuan/scrapy,ashishnerkar1/scrapy,pombredanne/scrapy,agreen/scrapy,aivarsk/scrapy,zackslash/scrapy,finfish/scrapy,nett55/scrapy,elacuesta/scrapy,Cnfc19932/scrapy,scorphus/scrapy,elijah513/scrapy,beni55/scrapy,amboxer21/scrapy,liyy7/scrapy,taito/scrapy,cursesun/scrapy,Chenmxs/scrapy,Zephor5/scrapy,lacrazyboy/scrapy,redapple/scrapy,scrapy/scrapy,agusc/scrapy,AaronTao1990/scrapy,Parlin-Galanodel/scrapy,jeffreyjinfeng/scrapy,stenskjaer/scrapy,olafdietsche/scrapy,fafaman/scrapy,Geeglee/scrapy,Preetwinder/scrapy,rolando/scrapy,hwsyy/scrapy,famorted/scrapy,zhangtao11/scrapy,dracony/scrapy,rklabs/scrapy,Bourneer/scrapy,wzyuliyang/scrapy,livepy/scrapy,codebhendi/scrapy,fafaman/scrapy,rolando/scrapy,sigma-random/scrapy,Geeglee/scrapy,aivarsk/scrapy,ArturGaspar/scrapy,elacuesta/scrapy,Djlavoy/scrapy,github-account-because-they-want-it/scrapy,smaty1/scrapy,wujuguang/scrapy,Timeship/scrapy,mouadino/scrapy,curita/scrapy,codebhendi/scrapy,Timeship/scrapy,AaronTao1990/scrapy,ramiro/scrapy,CodeJuan/scrapy,wzyuliyang/scrapy,nguyenhongson03/scrapy,hyrole/scrapy,kalessin/scrapy,JacobStevenR/scrapy,famorted/scrapy,dgillis/scrapy,hansenDise/scrapy,wangjun/scrapy,TarasRudnyk/scrapy,dracony/scrapy,Cnfc19932/scrapy,rahulsharma1991/scrapy,nguyenhongson03/scrapy,CodeJuan/scrapy,Chenmxs/scrapy,z-fork/scrapy,ENjOyAbLE1991/scrapy,umrashrf/scrapy,finfish/scrapy,farhan0581/scrapy,yusofm/scrapy,rootAvish/scrapy,mgedmin/scrapy,KublaikhanGeek/scrapy,devGregA/scrapy,crasker/scrapy,xiao26/scrapy,pranjalpatil/scrapy,IvanGavran/scrapy,liyy7/scrapy,elacuesta/scrapy,kimimj/scrapy,emschorsch/scrapy,jiezhu2007/scrapy,barraponto/scrapy,eliasdorneles/scrapy,pfctdayelise/scrapy,KublaikhanGeek/scrapy,rdowinton/scrapy,godfreyy/scrapy,nfunato/scrapy,beni55/scrapy,z-fork/scrapy,heamon7/scrapy,smaty1/scrapy,livepy/scrapy,jorik041/scrapy,fontenele/scrapy,jc0n/scrapy,Djlavoy/scrapy,songfj/scrapy,Geeglee/scrapy,tntC4stl3/scrapy,famorted/scrapy,fpy171/scrapy,irwinlove/scrapy,z-fork/scrapy,amboxer21/scrapy,kimimj/scrapy,fqul/scrapy,xiao26/scrapy,coderabhishek/scrapy,johnardavies/scrapy,Partoo/scrapy,ssteo/scrapy,yarikoptic/scrapy,wujuguang/scrapy,hyrole/scrapy,cyrixhero/scrapy,jorik041/scrapy,cleydson/scrapy,taito/scrapy,huoxudong125/scrapy,IvanGavran/scrapy,crasker/scrapy,nfunato/scrapy,Cnfc19932/scrapy,olorz/scrapy,Slater-Victoroff/scrapy,avtoritet/scrapy,hyrole/scrapy,CENDARI/scrapy,OpenWhere/scrapy,ssh-odoo/scrapy,rolando-contrib/scrapy,ssh-odoo/scrapy,csalazar/scrapy,Lucifer-Kim/scrapy,YeelerG/scrapy,gbirke/scrapy,CENDARI/scrapy,profjrr/scrapy,farhan0581/scrapy,cyrixhero/scrapy,devGregA/scrapy,ylcolala/scrapy,arush0311/scrapy,fontenele/scrapy,hbwzhsh/scrapy,kalessin/scrapy,JacobStevenR/scrapy,pombredanne/scrapy,Slater-Victoroff/scrapy,dacjames/scrapy,ashishnerkar1/scrapy,zackslash/scrapy,dangra/scrapy,Digenis/scrapy,rdowinton/scrapy,emschorsch/scrapy,fafaman/scrapy,csalazar/scrapy,rootAvish/scrapy,beni55/scrapy,hansenDise/scrapy,darkrho/scrapy-scrapy,GregoryVigoTorres/scrapy,rolando-contrib/scrapy,wenyu1001/scrapy,darkrho/scrapy-scrapy,nowopen/scrapy,agusc/scrapy,Slater-Victoroff/scrapy,huoxudong125/scrapy,barraponto/scrapy,yusofm/scrapy,curita/scrapy,webmakin/scrapy,devGregA/scrapy,joshlk/scrapy,rdowinton/scrapy,zjuwangg/scrapy,shaform/scrapy,joshlk/scrapy,olorz/scrapy,zhangtao11/scrapy,eLRuLL/scrapy,sigma-random/scrapy,haiiiiiyun/scrapy,irwinlove/scrapy,chekunkov/scrapy,jeffreyjinfeng/scrapy,cursesun/scrapy,jdemaeyer/scrapy,sardok/scrapy,TarasRudnyk/scrapy,songfj/scrapy,zjuwangg/scrapy,dhenyjarasandy/scrapy,ramiro/scrapy,eliasdorneles/scrapy,huoxudong125/scrapy,yusofm/scrapy,heamon7/scrapy,pranjalpatil/scrapy,elijah513/scrapy,rahul-c1/scrapy,dgillis/scrapy,tliber/scrapy,YeelerG/scrapy,zjuwangg/scrapy,nett55/scrapy,WilliamKinaan/scrapy,raphaelfruneaux/scrapy,haiiiiiyun/scrapy,dhenyjarasandy/scrapy,github-account-because-they-want-it/scrapy,wangjun/scrapy,WilliamKinaan/scrapy,cleydson/scrapy,pawelmhm/scrapy,farhan0581/scrapy,stenskjaer/scrapy,Ryezhang/scrapy,godfreyy/scrapy,Bourneer/scrapy,kimimj/scrapy,eliasdorneles/scrapy,tntC4stl3/scrapy,bmess/scrapy,1yvT0s/scrapy,moraesnicol/scrapy,snowdream1314/scrapy,Partoo/scrapy,nowopen/scrapy,jc0n/scrapy,eLRuLL/scrapy,dgillis/scrapy,kmike/scrapy,dacjames/scrapy,rolando/scrapy,moraesnicol/scrapy,rklabs/scrapy,JacobStevenR/scrapy,Allianzcortex/scrapy,Djlavoy/scrapy,Digenis/scrapy,liyy7/scrapy,starrify/scrapy,olorz/scrapy,finfish/scrapy,hectoruelo/scrapy,chekunkov/scrapy,yidongliu/scrapy,sardok/scrapy,mlyundin/scrapy,ssteo/scrapy,WilliamKinaan/scrapy,legendtkl/scrapy,snowdream1314/scrapy,nguyenhongson03/scrapy,jamesblunt/scrapy,legendtkl/scrapy,wenyu1001/scrapy,kashyap32/scrapy,moraesnicol/scrapy,jiezhu2007/scrapy,livepy/scrapy,tagatac/scrapy,wangjun/scrapy,codebhendi/scrapy,avtoritet/scrapy,raphaelfruneaux/scrapy,pombredanne/scrapy,csalazar/scrapy,tagatac/scrapy,kazitanvirahsan/scrapy,smaty1/scrapy,fqul/scrapy,Parlin-Galanodel/scrapy,lacrazyboy/scrapy,kashyap32/scrapy,shaform/scrapy,arush0311/scrapy,bmess/scrapy,Partoo/scrapy,YeelerG/scrapy,cyberplant/scrapy,wenyu1001/scrapy,cyberplant/scrapy,hwsyy/scrapy,gbirke/scrapy,johnardavies/scrapy,darkrho/scrapy-scrapy,mgedmin/scrapy,GregoryVigoTorres/scrapy,fpy171/scrapy,olafdietsche/scrapy,starrify/scrapy,mgedmin/scrapy,agreen/scrapy,yidongliu/scrapy,URXtech/scrapy,Lucifer-Kim/scrapy,kazitanvirahsan/scrapy,webmakin/scrapy,kashyap32/scrapy,ylcolala/scrapy,kalessin/scrapy,dangra/scrapy,zhangtao11/scrapy,KublaikhanGeek/scrapy,ENjOyAbLE1991/scrapy,jorik041/scrapy,nowopen/scrapy,scorphus/scrapy,yarikoptic/scrapy,Chenmxs/scrapy,mlyundin/scrapy,rklabs/scrapy,hectoruelo/scrapy,nett55/scrapy,chekunkov/scrapy,1yvT0s/scrapy,zorojean/scrapy,Allianzcortex/scrapy,stenskjaer/scrapy,cleydson/scrapy,dracony/scrapy,TarasRudnyk/scrapy,hbwzhsh/scrapy,Ryezhang/scrapy,ArturGaspar/scrapy,avtoritet/scrapy,starrify/scrapy,Adai0808/scrapy-1,Lucifer-Kim/scrapy,rootAvish/scrapy,nikgr95/scrapy,pablohoffman/scrapy,ssteo/scrapy,ndemir/scrapy,profjrr/scrapy,pablohoffman/scrapy,gnemoug/scrapy,scrapy/scrapy,profjrr/scrapy,scorphus/scrapy,joshlk/scrapy,rahul-c1/scrapy,jeffreyjinfeng/scrapy,cyrixhero/scrapy,shaform/scrapy,raphaelfruneaux/scrapy,arush0311/scrapy,1yvT0s/scrapy,tliber/scrapy,yarikoptic/scrapy,pablohoffman/scrapy,godfreyy/scrapy,jamesblunt/scrapy,Zephor5/scrapy,kazitanvirahsan/scrapy,Parlin-Galanodel/scrapy,Timeship/scrapy,agreen/scrapy,cyberplant/scrapy,coderabhishek/scrapy,carlosp420/scrapy,zackslash/scrapy,hwsyy/scrapy,OpenWhere/scrapy,redapple/scrapy,Allianzcortex/scrapy,fqul/scrapy,Adai0808/scrapy-1,dhenyjarasandy/scrapy,wzyuliyang/scrapy,w495/scrapy,snowdream1314/scrapy,ramiro/scrapy,bmess/scrapy,dangra/scrapy,IvanGavran/scrapy,pawelmhm/scrapy,URXtech/scrapy,elijah513/scrapy,gnemoug/scrapy,dacjames/scrapy,jdemaeyer/scrapy,rahulsharma1991/scrapy,Zephor5/scrapy,nikgr95/scrapy,CENDARI/scrapy,songfj/scrapy,nikgr95/scrapy,ndemir/scrapy,lacrazyboy/scrapy,ssh-odoo/scrapy,hectoruelo/scrapy,barraponto/scrapy,Preetwinder/scrapy,GregoryVigoTorres/scrapy,Adai0808/scrapy-1,OpenWhere/scrapy,umrashrf/scrapy,zorojean/scrapy,legendtkl/scrapy,amboxer21/scrapy,hansenDise/scrapy,carlosp420/scrapy,hbwzhsh/scrapy,pfctdayelise/scrapy,crasker/scrapy,jdemaeyer/scrapy,pfctdayelise/scrapy,Bourneer/scrapy,aivarsk/scrapy,mouadino/scrapy,tliber/scrapy,foromer4/scrapy,pranjalpatil/scrapy,pawelmhm/scrapy,emschorsch/scrapy,Ryezhang/scrapy,coderabhishek/scrapy
import scrapy # Scrapy core settings BOT_NAME = 'scrapy' BOT_VERSION = scrapy.__version__ ENGINE_DEBUG = False # Download configuration options USER_AGENT = '%s/%s' % (BOT_NAME, BOT_VERSION) DOWNLOAD_TIMEOUT = 180 # 3mins CONCURRENT_DOMAINS = 8 # number of domains to scrape in parallel REQUESTS_PER_DOMAIN = 8 # max simultaneous requests per domain CACHE2_EXPIRATION_SECS = 48 * 60 * 60 # seconds while cached response is still valid (a negative value means "never expires") LOG_ENABLED = True # LOGLEVEL = 'DEBUG' # default loglevel LOGFILE = None # None means sys.stderr by default LOG_STDOUT = False # DEFAULT_ITEM_CLASS = 'scrapy.item.ScrapedItem' SCHEDULER = 'scrapy.core.scheduler.Scheduler' MEMORYSTORE = 'scrapy.core.scheduler.MemoryStore' PRIORITIZER = 'scrapy.core.prioritizers.RandomPrioritizer' EXTENSIONS = [] # contrib.middleware.retry.RetryMiddleware default settings RETRY_TIMES = 3 RETRY_HTTP_CODES = ['500', '503', '504', '400', '408', '200'] # amazon s3 images backend sometimes fails to authenticate or validate http protocol RETRY_HTTP_CODES = RETRY_HTTP_CODES + ['505', '403'] Revert "add 505 and 403 to retry status codes due to amazon s3 random fails while uploading images" This reverts changeset r457 --HG-- extra : convert_revision : svn%3Ab85faa78-f9eb-468e-a121-7cced6da292c%40461
import scrapy # Scrapy core settings BOT_NAME = 'scrapy' BOT_VERSION = scrapy.__version__ ENGINE_DEBUG = False # Download configuration options USER_AGENT = '%s/%s' % (BOT_NAME, BOT_VERSION) DOWNLOAD_TIMEOUT = 180 # 3mins CONCURRENT_DOMAINS = 8 # number of domains to scrape in parallel REQUESTS_PER_DOMAIN = 8 # max simultaneous requests per domain CACHE2_EXPIRATION_SECS = 48 * 60 * 60 # seconds while cached response is still valid (a negative value means "never expires") LOG_ENABLED = True # LOGLEVEL = 'DEBUG' # default loglevel LOGFILE = None # None means sys.stderr by default LOG_STDOUT = False # DEFAULT_ITEM_CLASS = 'scrapy.item.ScrapedItem' SCHEDULER = 'scrapy.core.scheduler.Scheduler' MEMORYSTORE = 'scrapy.core.scheduler.MemoryStore' PRIORITIZER = 'scrapy.core.prioritizers.RandomPrioritizer' EXTENSIONS = [] # contrib.middleware.retry.RetryMiddleware default settings RETRY_TIMES = 3 RETRY_HTTP_CODES = ['500', '503', '504', '400', '408', '200']
<commit_before>import scrapy # Scrapy core settings BOT_NAME = 'scrapy' BOT_VERSION = scrapy.__version__ ENGINE_DEBUG = False # Download configuration options USER_AGENT = '%s/%s' % (BOT_NAME, BOT_VERSION) DOWNLOAD_TIMEOUT = 180 # 3mins CONCURRENT_DOMAINS = 8 # number of domains to scrape in parallel REQUESTS_PER_DOMAIN = 8 # max simultaneous requests per domain CACHE2_EXPIRATION_SECS = 48 * 60 * 60 # seconds while cached response is still valid (a negative value means "never expires") LOG_ENABLED = True # LOGLEVEL = 'DEBUG' # default loglevel LOGFILE = None # None means sys.stderr by default LOG_STDOUT = False # DEFAULT_ITEM_CLASS = 'scrapy.item.ScrapedItem' SCHEDULER = 'scrapy.core.scheduler.Scheduler' MEMORYSTORE = 'scrapy.core.scheduler.MemoryStore' PRIORITIZER = 'scrapy.core.prioritizers.RandomPrioritizer' EXTENSIONS = [] # contrib.middleware.retry.RetryMiddleware default settings RETRY_TIMES = 3 RETRY_HTTP_CODES = ['500', '503', '504', '400', '408', '200'] # amazon s3 images backend sometimes fails to authenticate or validate http protocol RETRY_HTTP_CODES = RETRY_HTTP_CODES + ['505', '403'] <commit_msg>Revert "add 505 and 403 to retry status codes due to amazon s3 random fails while uploading images" This reverts changeset r457 --HG-- extra : convert_revision : svn%3Ab85faa78-f9eb-468e-a121-7cced6da292c%40461<commit_after>
import scrapy # Scrapy core settings BOT_NAME = 'scrapy' BOT_VERSION = scrapy.__version__ ENGINE_DEBUG = False # Download configuration options USER_AGENT = '%s/%s' % (BOT_NAME, BOT_VERSION) DOWNLOAD_TIMEOUT = 180 # 3mins CONCURRENT_DOMAINS = 8 # number of domains to scrape in parallel REQUESTS_PER_DOMAIN = 8 # max simultaneous requests per domain CACHE2_EXPIRATION_SECS = 48 * 60 * 60 # seconds while cached response is still valid (a negative value means "never expires") LOG_ENABLED = True # LOGLEVEL = 'DEBUG' # default loglevel LOGFILE = None # None means sys.stderr by default LOG_STDOUT = False # DEFAULT_ITEM_CLASS = 'scrapy.item.ScrapedItem' SCHEDULER = 'scrapy.core.scheduler.Scheduler' MEMORYSTORE = 'scrapy.core.scheduler.MemoryStore' PRIORITIZER = 'scrapy.core.prioritizers.RandomPrioritizer' EXTENSIONS = [] # contrib.middleware.retry.RetryMiddleware default settings RETRY_TIMES = 3 RETRY_HTTP_CODES = ['500', '503', '504', '400', '408', '200']
import scrapy # Scrapy core settings BOT_NAME = 'scrapy' BOT_VERSION = scrapy.__version__ ENGINE_DEBUG = False # Download configuration options USER_AGENT = '%s/%s' % (BOT_NAME, BOT_VERSION) DOWNLOAD_TIMEOUT = 180 # 3mins CONCURRENT_DOMAINS = 8 # number of domains to scrape in parallel REQUESTS_PER_DOMAIN = 8 # max simultaneous requests per domain CACHE2_EXPIRATION_SECS = 48 * 60 * 60 # seconds while cached response is still valid (a negative value means "never expires") LOG_ENABLED = True # LOGLEVEL = 'DEBUG' # default loglevel LOGFILE = None # None means sys.stderr by default LOG_STDOUT = False # DEFAULT_ITEM_CLASS = 'scrapy.item.ScrapedItem' SCHEDULER = 'scrapy.core.scheduler.Scheduler' MEMORYSTORE = 'scrapy.core.scheduler.MemoryStore' PRIORITIZER = 'scrapy.core.prioritizers.RandomPrioritizer' EXTENSIONS = [] # contrib.middleware.retry.RetryMiddleware default settings RETRY_TIMES = 3 RETRY_HTTP_CODES = ['500', '503', '504', '400', '408', '200'] # amazon s3 images backend sometimes fails to authenticate or validate http protocol RETRY_HTTP_CODES = RETRY_HTTP_CODES + ['505', '403'] Revert "add 505 and 403 to retry status codes due to amazon s3 random fails while uploading images" This reverts changeset r457 --HG-- extra : convert_revision : svn%3Ab85faa78-f9eb-468e-a121-7cced6da292c%40461import scrapy # Scrapy core settings BOT_NAME = 'scrapy' BOT_VERSION = scrapy.__version__ ENGINE_DEBUG = False # Download configuration options USER_AGENT = '%s/%s' % (BOT_NAME, BOT_VERSION) DOWNLOAD_TIMEOUT = 180 # 3mins CONCURRENT_DOMAINS = 8 # number of domains to scrape in parallel REQUESTS_PER_DOMAIN = 8 # max simultaneous requests per domain CACHE2_EXPIRATION_SECS = 48 * 60 * 60 # seconds while cached response is still valid (a negative value means "never expires") LOG_ENABLED = True # LOGLEVEL = 'DEBUG' # default loglevel LOGFILE = None # None means sys.stderr by default LOG_STDOUT = False # DEFAULT_ITEM_CLASS = 'scrapy.item.ScrapedItem' SCHEDULER = 'scrapy.core.scheduler.Scheduler' MEMORYSTORE = 'scrapy.core.scheduler.MemoryStore' PRIORITIZER = 'scrapy.core.prioritizers.RandomPrioritizer' EXTENSIONS = [] # contrib.middleware.retry.RetryMiddleware default settings RETRY_TIMES = 3 RETRY_HTTP_CODES = ['500', '503', '504', '400', '408', '200']
<commit_before>import scrapy # Scrapy core settings BOT_NAME = 'scrapy' BOT_VERSION = scrapy.__version__ ENGINE_DEBUG = False # Download configuration options USER_AGENT = '%s/%s' % (BOT_NAME, BOT_VERSION) DOWNLOAD_TIMEOUT = 180 # 3mins CONCURRENT_DOMAINS = 8 # number of domains to scrape in parallel REQUESTS_PER_DOMAIN = 8 # max simultaneous requests per domain CACHE2_EXPIRATION_SECS = 48 * 60 * 60 # seconds while cached response is still valid (a negative value means "never expires") LOG_ENABLED = True # LOGLEVEL = 'DEBUG' # default loglevel LOGFILE = None # None means sys.stderr by default LOG_STDOUT = False # DEFAULT_ITEM_CLASS = 'scrapy.item.ScrapedItem' SCHEDULER = 'scrapy.core.scheduler.Scheduler' MEMORYSTORE = 'scrapy.core.scheduler.MemoryStore' PRIORITIZER = 'scrapy.core.prioritizers.RandomPrioritizer' EXTENSIONS = [] # contrib.middleware.retry.RetryMiddleware default settings RETRY_TIMES = 3 RETRY_HTTP_CODES = ['500', '503', '504', '400', '408', '200'] # amazon s3 images backend sometimes fails to authenticate or validate http protocol RETRY_HTTP_CODES = RETRY_HTTP_CODES + ['505', '403'] <commit_msg>Revert "add 505 and 403 to retry status codes due to amazon s3 random fails while uploading images" This reverts changeset r457 --HG-- extra : convert_revision : svn%3Ab85faa78-f9eb-468e-a121-7cced6da292c%40461<commit_after>import scrapy # Scrapy core settings BOT_NAME = 'scrapy' BOT_VERSION = scrapy.__version__ ENGINE_DEBUG = False # Download configuration options USER_AGENT = '%s/%s' % (BOT_NAME, BOT_VERSION) DOWNLOAD_TIMEOUT = 180 # 3mins CONCURRENT_DOMAINS = 8 # number of domains to scrape in parallel REQUESTS_PER_DOMAIN = 8 # max simultaneous requests per domain CACHE2_EXPIRATION_SECS = 48 * 60 * 60 # seconds while cached response is still valid (a negative value means "never expires") LOG_ENABLED = True # LOGLEVEL = 'DEBUG' # default loglevel LOGFILE = None # None means sys.stderr by default LOG_STDOUT = False # DEFAULT_ITEM_CLASS = 'scrapy.item.ScrapedItem' SCHEDULER = 'scrapy.core.scheduler.Scheduler' MEMORYSTORE = 'scrapy.core.scheduler.MemoryStore' PRIORITIZER = 'scrapy.core.prioritizers.RandomPrioritizer' EXTENSIONS = [] # contrib.middleware.retry.RetryMiddleware default settings RETRY_TIMES = 3 RETRY_HTTP_CODES = ['500', '503', '504', '400', '408', '200']
bad65df528da18293d38b0f50dbbb16390af465e
sphinx/source/docs/user_guide/source_examples/plotting_label.py
sphinx/source/docs/user_guide/source_examples/plotting_label.py
from bokeh.plotting import figure, show, output_file from bokeh.models import ColumnDataSource, Range1d, Label output_file("label.html", title="label.py example") source = ColumnDataSource(data=dict(height=[66, 71, 72, 68, 58, 62], weight=[165, 189, 220, 141, 260, 174], names=['Mark', 'Amir', 'Matt', 'Greg', 'Owen', 'Juan'])) p = figure(title='Dist. of 10th Grade Students at Lee High', x_range=Range1d(140, 275)) p.scatter(x='weight', y='height', size=8, source=source) p.xaxis[0].axis_label = 'Weight (lbs)' p.yaxis[0].axis_label = 'Height (in)' label = Label(x='weight', y='height', text='names', level='glyph', x_offset=5, y_offset=-5, source=source) p.add_annotation(label) show(p)
from bokeh.plotting import figure, show, output_file from bokeh.models import ColumnDataSource, Range1d, Label output_file("label.html", title="label.py example") source = ColumnDataSource(data=dict(height=[66, 71, 72, 68, 58, 62], weight=[165, 189, 220, 141, 260, 174], names=['Mark', 'Amir', 'Matt', 'Greg', 'Owen', 'Juan'])) p = figure(title='Dist. of 10th Grade Students at Lee High', x_range=Range1d(140, 275)) p.scatter(x='weight', y='height', size=8, source=source) p.xaxis[0].axis_label = 'Weight (lbs)' p.yaxis[0].axis_label = 'Height (in)' labels = Label(x='weight', y='height', text='names', level='glyph', x_offset=5, y_offset=-5, source=source, render_mode='canvas') citation = Label(x=70, y=70, x_units='screen', y_units='screen', text=['Collected by Luke C. 2016-04-01'], render_mode='css', border_line_color='black', border_line_alpha=1.0, background_fill_color='white', background_fill_alpha=1.0) p.add_annotation(labels) p.add_annotation(citation) show(p)
Include example of css render_mode
Include example of css render_mode
Python
bsd-3-clause
clairetang6/bokeh,Karel-van-de-Plassche/bokeh,mindriot101/bokeh,aiguofer/bokeh,rs2/bokeh,Karel-van-de-Plassche/bokeh,KasperPRasmussen/bokeh,dennisobrien/bokeh,draperjames/bokeh,bokeh/bokeh,quasiben/bokeh,KasperPRasmussen/bokeh,philippjfr/bokeh,stonebig/bokeh,justacec/bokeh,KasperPRasmussen/bokeh,phobson/bokeh,phobson/bokeh,jakirkham/bokeh,rs2/bokeh,ericmjl/bokeh,jakirkham/bokeh,bokeh/bokeh,phobson/bokeh,bokeh/bokeh,aavanian/bokeh,dennisobrien/bokeh,rs2/bokeh,timsnyder/bokeh,ptitjano/bokeh,schoolie/bokeh,DuCorey/bokeh,dennisobrien/bokeh,quasiben/bokeh,mindriot101/bokeh,draperjames/bokeh,stonebig/bokeh,dennisobrien/bokeh,philippjfr/bokeh,rs2/bokeh,Karel-van-de-Plassche/bokeh,rs2/bokeh,azjps/bokeh,ptitjano/bokeh,jakirkham/bokeh,aavanian/bokeh,mindriot101/bokeh,KasperPRasmussen/bokeh,percyfal/bokeh,clairetang6/bokeh,ericmjl/bokeh,philippjfr/bokeh,clairetang6/bokeh,ericmjl/bokeh,mindriot101/bokeh,aiguofer/bokeh,draperjames/bokeh,bokeh/bokeh,bokeh/bokeh,Karel-van-de-Plassche/bokeh,philippjfr/bokeh,ericmjl/bokeh,phobson/bokeh,aavanian/bokeh,jakirkham/bokeh,timsnyder/bokeh,azjps/bokeh,schoolie/bokeh,DuCorey/bokeh,azjps/bokeh,quasiben/bokeh,DuCorey/bokeh,justacec/bokeh,draperjames/bokeh,justacec/bokeh,jakirkham/bokeh,schoolie/bokeh,justacec/bokeh,stonebig/bokeh,philippjfr/bokeh,Karel-van-de-Plassche/bokeh,aiguofer/bokeh,aavanian/bokeh,ptitjano/bokeh,ptitjano/bokeh,DuCorey/bokeh,azjps/bokeh,percyfal/bokeh,timsnyder/bokeh,aiguofer/bokeh,KasperPRasmussen/bokeh,timsnyder/bokeh,schoolie/bokeh,ptitjano/bokeh,stonebig/bokeh,phobson/bokeh,timsnyder/bokeh,dennisobrien/bokeh,draperjames/bokeh,schoolie/bokeh,DuCorey/bokeh,azjps/bokeh,percyfal/bokeh,percyfal/bokeh,clairetang6/bokeh,percyfal/bokeh,ericmjl/bokeh,aiguofer/bokeh,aavanian/bokeh
from bokeh.plotting import figure, show, output_file from bokeh.models import ColumnDataSource, Range1d, Label output_file("label.html", title="label.py example") source = ColumnDataSource(data=dict(height=[66, 71, 72, 68, 58, 62], weight=[165, 189, 220, 141, 260, 174], names=['Mark', 'Amir', 'Matt', 'Greg', 'Owen', 'Juan'])) p = figure(title='Dist. of 10th Grade Students at Lee High', x_range=Range1d(140, 275)) p.scatter(x='weight', y='height', size=8, source=source) p.xaxis[0].axis_label = 'Weight (lbs)' p.yaxis[0].axis_label = 'Height (in)' label = Label(x='weight', y='height', text='names', level='glyph', x_offset=5, y_offset=-5, source=source) p.add_annotation(label) show(p) Include example of css render_mode
from bokeh.plotting import figure, show, output_file from bokeh.models import ColumnDataSource, Range1d, Label output_file("label.html", title="label.py example") source = ColumnDataSource(data=dict(height=[66, 71, 72, 68, 58, 62], weight=[165, 189, 220, 141, 260, 174], names=['Mark', 'Amir', 'Matt', 'Greg', 'Owen', 'Juan'])) p = figure(title='Dist. of 10th Grade Students at Lee High', x_range=Range1d(140, 275)) p.scatter(x='weight', y='height', size=8, source=source) p.xaxis[0].axis_label = 'Weight (lbs)' p.yaxis[0].axis_label = 'Height (in)' labels = Label(x='weight', y='height', text='names', level='glyph', x_offset=5, y_offset=-5, source=source, render_mode='canvas') citation = Label(x=70, y=70, x_units='screen', y_units='screen', text=['Collected by Luke C. 2016-04-01'], render_mode='css', border_line_color='black', border_line_alpha=1.0, background_fill_color='white', background_fill_alpha=1.0) p.add_annotation(labels) p.add_annotation(citation) show(p)
<commit_before>from bokeh.plotting import figure, show, output_file from bokeh.models import ColumnDataSource, Range1d, Label output_file("label.html", title="label.py example") source = ColumnDataSource(data=dict(height=[66, 71, 72, 68, 58, 62], weight=[165, 189, 220, 141, 260, 174], names=['Mark', 'Amir', 'Matt', 'Greg', 'Owen', 'Juan'])) p = figure(title='Dist. of 10th Grade Students at Lee High', x_range=Range1d(140, 275)) p.scatter(x='weight', y='height', size=8, source=source) p.xaxis[0].axis_label = 'Weight (lbs)' p.yaxis[0].axis_label = 'Height (in)' label = Label(x='weight', y='height', text='names', level='glyph', x_offset=5, y_offset=-5, source=source) p.add_annotation(label) show(p) <commit_msg>Include example of css render_mode<commit_after>
from bokeh.plotting import figure, show, output_file from bokeh.models import ColumnDataSource, Range1d, Label output_file("label.html", title="label.py example") source = ColumnDataSource(data=dict(height=[66, 71, 72, 68, 58, 62], weight=[165, 189, 220, 141, 260, 174], names=['Mark', 'Amir', 'Matt', 'Greg', 'Owen', 'Juan'])) p = figure(title='Dist. of 10th Grade Students at Lee High', x_range=Range1d(140, 275)) p.scatter(x='weight', y='height', size=8, source=source) p.xaxis[0].axis_label = 'Weight (lbs)' p.yaxis[0].axis_label = 'Height (in)' labels = Label(x='weight', y='height', text='names', level='glyph', x_offset=5, y_offset=-5, source=source, render_mode='canvas') citation = Label(x=70, y=70, x_units='screen', y_units='screen', text=['Collected by Luke C. 2016-04-01'], render_mode='css', border_line_color='black', border_line_alpha=1.0, background_fill_color='white', background_fill_alpha=1.0) p.add_annotation(labels) p.add_annotation(citation) show(p)
from bokeh.plotting import figure, show, output_file from bokeh.models import ColumnDataSource, Range1d, Label output_file("label.html", title="label.py example") source = ColumnDataSource(data=dict(height=[66, 71, 72, 68, 58, 62], weight=[165, 189, 220, 141, 260, 174], names=['Mark', 'Amir', 'Matt', 'Greg', 'Owen', 'Juan'])) p = figure(title='Dist. of 10th Grade Students at Lee High', x_range=Range1d(140, 275)) p.scatter(x='weight', y='height', size=8, source=source) p.xaxis[0].axis_label = 'Weight (lbs)' p.yaxis[0].axis_label = 'Height (in)' label = Label(x='weight', y='height', text='names', level='glyph', x_offset=5, y_offset=-5, source=source) p.add_annotation(label) show(p) Include example of css render_modefrom bokeh.plotting import figure, show, output_file from bokeh.models import ColumnDataSource, Range1d, Label output_file("label.html", title="label.py example") source = ColumnDataSource(data=dict(height=[66, 71, 72, 68, 58, 62], weight=[165, 189, 220, 141, 260, 174], names=['Mark', 'Amir', 'Matt', 'Greg', 'Owen', 'Juan'])) p = figure(title='Dist. of 10th Grade Students at Lee High', x_range=Range1d(140, 275)) p.scatter(x='weight', y='height', size=8, source=source) p.xaxis[0].axis_label = 'Weight (lbs)' p.yaxis[0].axis_label = 'Height (in)' labels = Label(x='weight', y='height', text='names', level='glyph', x_offset=5, y_offset=-5, source=source, render_mode='canvas') citation = Label(x=70, y=70, x_units='screen', y_units='screen', text=['Collected by Luke C. 2016-04-01'], render_mode='css', border_line_color='black', border_line_alpha=1.0, background_fill_color='white', background_fill_alpha=1.0) p.add_annotation(labels) p.add_annotation(citation) show(p)
<commit_before>from bokeh.plotting import figure, show, output_file from bokeh.models import ColumnDataSource, Range1d, Label output_file("label.html", title="label.py example") source = ColumnDataSource(data=dict(height=[66, 71, 72, 68, 58, 62], weight=[165, 189, 220, 141, 260, 174], names=['Mark', 'Amir', 'Matt', 'Greg', 'Owen', 'Juan'])) p = figure(title='Dist. of 10th Grade Students at Lee High', x_range=Range1d(140, 275)) p.scatter(x='weight', y='height', size=8, source=source) p.xaxis[0].axis_label = 'Weight (lbs)' p.yaxis[0].axis_label = 'Height (in)' label = Label(x='weight', y='height', text='names', level='glyph', x_offset=5, y_offset=-5, source=source) p.add_annotation(label) show(p) <commit_msg>Include example of css render_mode<commit_after>from bokeh.plotting import figure, show, output_file from bokeh.models import ColumnDataSource, Range1d, Label output_file("label.html", title="label.py example") source = ColumnDataSource(data=dict(height=[66, 71, 72, 68, 58, 62], weight=[165, 189, 220, 141, 260, 174], names=['Mark', 'Amir', 'Matt', 'Greg', 'Owen', 'Juan'])) p = figure(title='Dist. of 10th Grade Students at Lee High', x_range=Range1d(140, 275)) p.scatter(x='weight', y='height', size=8, source=source) p.xaxis[0].axis_label = 'Weight (lbs)' p.yaxis[0].axis_label = 'Height (in)' labels = Label(x='weight', y='height', text='names', level='glyph', x_offset=5, y_offset=-5, source=source, render_mode='canvas') citation = Label(x=70, y=70, x_units='screen', y_units='screen', text=['Collected by Luke C. 2016-04-01'], render_mode='css', border_line_color='black', border_line_alpha=1.0, background_fill_color='white', background_fill_alpha=1.0) p.add_annotation(labels) p.add_annotation(citation) show(p)
c6cdf543f6bfd0049594eeb530551371bf21bae4
test/test_scraping.py
test/test_scraping.py
from datetime import datetime import sys import unittest import btceapi class TestScraping(unittest.TestCase): def test_scrape_main_page(self): mainPage = btceapi.scrapeMainPage() for message in mainPage.messages: msgId, user, time, text = message self.assertIs(type(time), datetime) if sys.version_info[0] == 2: # python2.x assert type(msgId) in (str, unicode) assert type(user) in (str, unicode) assert type(text) in (str, unicode) else: # python3.x self.assertIs(type(msgId), str) self.assertIs(type(user), str) self.assertIs(type(text), str) if __name__ == '__main__': unittest.main()
from datetime import datetime import sys import unittest import btceapi class TestScraping(unittest.TestCase): def test_scrape_main_page(self): mainPage = btceapi.scrapeMainPage() for message in mainPage.messages: msgId, user, time, text = message assert type(time) is datetime if sys.version_info[0] == 2: # python2.x assert type(msgId) in (str, unicode) assert type(user) in (str, unicode) assert type(text) in (str, unicode) else: # python3.x self.assertIs(type(msgId), str) self.assertIs(type(user), str) self.assertIs(type(text), str) if __name__ == '__main__': unittest.main()
Fix for assertIs method not being present in Python 2.6.
Fix for assertIs method not being present in Python 2.6.
Python
mit
lromanov/tidex-api,CodeReclaimers/btce-api,alanmcintyre/btce-api
from datetime import datetime import sys import unittest import btceapi class TestScraping(unittest.TestCase): def test_scrape_main_page(self): mainPage = btceapi.scrapeMainPage() for message in mainPage.messages: msgId, user, time, text = message self.assertIs(type(time), datetime) if sys.version_info[0] == 2: # python2.x assert type(msgId) in (str, unicode) assert type(user) in (str, unicode) assert type(text) in (str, unicode) else: # python3.x self.assertIs(type(msgId), str) self.assertIs(type(user), str) self.assertIs(type(text), str) if __name__ == '__main__': unittest.main() Fix for assertIs method not being present in Python 2.6.
from datetime import datetime import sys import unittest import btceapi class TestScraping(unittest.TestCase): def test_scrape_main_page(self): mainPage = btceapi.scrapeMainPage() for message in mainPage.messages: msgId, user, time, text = message assert type(time) is datetime if sys.version_info[0] == 2: # python2.x assert type(msgId) in (str, unicode) assert type(user) in (str, unicode) assert type(text) in (str, unicode) else: # python3.x self.assertIs(type(msgId), str) self.assertIs(type(user), str) self.assertIs(type(text), str) if __name__ == '__main__': unittest.main()
<commit_before>from datetime import datetime import sys import unittest import btceapi class TestScraping(unittest.TestCase): def test_scrape_main_page(self): mainPage = btceapi.scrapeMainPage() for message in mainPage.messages: msgId, user, time, text = message self.assertIs(type(time), datetime) if sys.version_info[0] == 2: # python2.x assert type(msgId) in (str, unicode) assert type(user) in (str, unicode) assert type(text) in (str, unicode) else: # python3.x self.assertIs(type(msgId), str) self.assertIs(type(user), str) self.assertIs(type(text), str) if __name__ == '__main__': unittest.main() <commit_msg>Fix for assertIs method not being present in Python 2.6.<commit_after>
from datetime import datetime import sys import unittest import btceapi class TestScraping(unittest.TestCase): def test_scrape_main_page(self): mainPage = btceapi.scrapeMainPage() for message in mainPage.messages: msgId, user, time, text = message assert type(time) is datetime if sys.version_info[0] == 2: # python2.x assert type(msgId) in (str, unicode) assert type(user) in (str, unicode) assert type(text) in (str, unicode) else: # python3.x self.assertIs(type(msgId), str) self.assertIs(type(user), str) self.assertIs(type(text), str) if __name__ == '__main__': unittest.main()
from datetime import datetime import sys import unittest import btceapi class TestScraping(unittest.TestCase): def test_scrape_main_page(self): mainPage = btceapi.scrapeMainPage() for message in mainPage.messages: msgId, user, time, text = message self.assertIs(type(time), datetime) if sys.version_info[0] == 2: # python2.x assert type(msgId) in (str, unicode) assert type(user) in (str, unicode) assert type(text) in (str, unicode) else: # python3.x self.assertIs(type(msgId), str) self.assertIs(type(user), str) self.assertIs(type(text), str) if __name__ == '__main__': unittest.main() Fix for assertIs method not being present in Python 2.6.from datetime import datetime import sys import unittest import btceapi class TestScraping(unittest.TestCase): def test_scrape_main_page(self): mainPage = btceapi.scrapeMainPage() for message in mainPage.messages: msgId, user, time, text = message assert type(time) is datetime if sys.version_info[0] == 2: # python2.x assert type(msgId) in (str, unicode) assert type(user) in (str, unicode) assert type(text) in (str, unicode) else: # python3.x self.assertIs(type(msgId), str) self.assertIs(type(user), str) self.assertIs(type(text), str) if __name__ == '__main__': unittest.main()
<commit_before>from datetime import datetime import sys import unittest import btceapi class TestScraping(unittest.TestCase): def test_scrape_main_page(self): mainPage = btceapi.scrapeMainPage() for message in mainPage.messages: msgId, user, time, text = message self.assertIs(type(time), datetime) if sys.version_info[0] == 2: # python2.x assert type(msgId) in (str, unicode) assert type(user) in (str, unicode) assert type(text) in (str, unicode) else: # python3.x self.assertIs(type(msgId), str) self.assertIs(type(user), str) self.assertIs(type(text), str) if __name__ == '__main__': unittest.main() <commit_msg>Fix for assertIs method not being present in Python 2.6.<commit_after>from datetime import datetime import sys import unittest import btceapi class TestScraping(unittest.TestCase): def test_scrape_main_page(self): mainPage = btceapi.scrapeMainPage() for message in mainPage.messages: msgId, user, time, text = message assert type(time) is datetime if sys.version_info[0] == 2: # python2.x assert type(msgId) in (str, unicode) assert type(user) in (str, unicode) assert type(text) in (str, unicode) else: # python3.x self.assertIs(type(msgId), str) self.assertIs(type(user), str) self.assertIs(type(text), str) if __name__ == '__main__': unittest.main()
96076567bac3329cba55b61c59781c7670c7a02b
anybox/recipe/odoo/runtime/patch_odoo.py
anybox/recipe/odoo/runtime/patch_odoo.py
"""Necessary monkey patches to make Odoo work in the buildout context. """ import subprocess def do_patch(gevent_script_path): """ Patch odoo prefork so that --workers execute the correct gevent script. This monkey patch could be safer, if the script path determination could be isolated from the actual process management logic in the original. """ from odoo.service.server import PreforkServer, stripped_sys_argv def long_polling_spawn(server): nargs = stripped_sys_argv() nargs[0] = gevent_script_path popen = subprocess.Popen(nargs) server.long_polling_pid = popen.pid PreforkServer.long_polling_spawn = long_polling_spawn
"""Necessary monkey patches to make Odoo work in the buildout context. """ import subprocess def do_patch(gevent_script_path): """ Patch odoo prefork so that --workers execute the correct gevent script. This monkey patch could be safer, if the script path determination could be isolated from the actual process management logic in the original. """ try: from odoo.service.server import PreforkServer, stripped_sys_argv except ImportError: from openerp.service.server import PreforkServer, stripped_sys_argv def long_polling_spawn(server): nargs = stripped_sys_argv() nargs[0] = gevent_script_path popen = subprocess.Popen(nargs) server.long_polling_pid = popen.pid PreforkServer.long_polling_spawn = long_polling_spawn
Maintain compatilbility with <10 version
Maintain compatilbility with <10 version
Python
agpl-3.0
anybox/anybox.recipe.odoo
"""Necessary monkey patches to make Odoo work in the buildout context. """ import subprocess def do_patch(gevent_script_path): """ Patch odoo prefork so that --workers execute the correct gevent script. This monkey patch could be safer, if the script path determination could be isolated from the actual process management logic in the original. """ from odoo.service.server import PreforkServer, stripped_sys_argv def long_polling_spawn(server): nargs = stripped_sys_argv() nargs[0] = gevent_script_path popen = subprocess.Popen(nargs) server.long_polling_pid = popen.pid PreforkServer.long_polling_spawn = long_polling_spawn Maintain compatilbility with <10 version
"""Necessary monkey patches to make Odoo work in the buildout context. """ import subprocess def do_patch(gevent_script_path): """ Patch odoo prefork so that --workers execute the correct gevent script. This monkey patch could be safer, if the script path determination could be isolated from the actual process management logic in the original. """ try: from odoo.service.server import PreforkServer, stripped_sys_argv except ImportError: from openerp.service.server import PreforkServer, stripped_sys_argv def long_polling_spawn(server): nargs = stripped_sys_argv() nargs[0] = gevent_script_path popen = subprocess.Popen(nargs) server.long_polling_pid = popen.pid PreforkServer.long_polling_spawn = long_polling_spawn
<commit_before>"""Necessary monkey patches to make Odoo work in the buildout context. """ import subprocess def do_patch(gevent_script_path): """ Patch odoo prefork so that --workers execute the correct gevent script. This monkey patch could be safer, if the script path determination could be isolated from the actual process management logic in the original. """ from odoo.service.server import PreforkServer, stripped_sys_argv def long_polling_spawn(server): nargs = stripped_sys_argv() nargs[0] = gevent_script_path popen = subprocess.Popen(nargs) server.long_polling_pid = popen.pid PreforkServer.long_polling_spawn = long_polling_spawn <commit_msg>Maintain compatilbility with <10 version<commit_after>
"""Necessary monkey patches to make Odoo work in the buildout context. """ import subprocess def do_patch(gevent_script_path): """ Patch odoo prefork so that --workers execute the correct gevent script. This monkey patch could be safer, if the script path determination could be isolated from the actual process management logic in the original. """ try: from odoo.service.server import PreforkServer, stripped_sys_argv except ImportError: from openerp.service.server import PreforkServer, stripped_sys_argv def long_polling_spawn(server): nargs = stripped_sys_argv() nargs[0] = gevent_script_path popen = subprocess.Popen(nargs) server.long_polling_pid = popen.pid PreforkServer.long_polling_spawn = long_polling_spawn
"""Necessary monkey patches to make Odoo work in the buildout context. """ import subprocess def do_patch(gevent_script_path): """ Patch odoo prefork so that --workers execute the correct gevent script. This monkey patch could be safer, if the script path determination could be isolated from the actual process management logic in the original. """ from odoo.service.server import PreforkServer, stripped_sys_argv def long_polling_spawn(server): nargs = stripped_sys_argv() nargs[0] = gevent_script_path popen = subprocess.Popen(nargs) server.long_polling_pid = popen.pid PreforkServer.long_polling_spawn = long_polling_spawn Maintain compatilbility with <10 version"""Necessary monkey patches to make Odoo work in the buildout context. """ import subprocess def do_patch(gevent_script_path): """ Patch odoo prefork so that --workers execute the correct gevent script. This monkey patch could be safer, if the script path determination could be isolated from the actual process management logic in the original. """ try: from odoo.service.server import PreforkServer, stripped_sys_argv except ImportError: from openerp.service.server import PreforkServer, stripped_sys_argv def long_polling_spawn(server): nargs = stripped_sys_argv() nargs[0] = gevent_script_path popen = subprocess.Popen(nargs) server.long_polling_pid = popen.pid PreforkServer.long_polling_spawn = long_polling_spawn
<commit_before>"""Necessary monkey patches to make Odoo work in the buildout context. """ import subprocess def do_patch(gevent_script_path): """ Patch odoo prefork so that --workers execute the correct gevent script. This monkey patch could be safer, if the script path determination could be isolated from the actual process management logic in the original. """ from odoo.service.server import PreforkServer, stripped_sys_argv def long_polling_spawn(server): nargs = stripped_sys_argv() nargs[0] = gevent_script_path popen = subprocess.Popen(nargs) server.long_polling_pid = popen.pid PreforkServer.long_polling_spawn = long_polling_spawn <commit_msg>Maintain compatilbility with <10 version<commit_after>"""Necessary monkey patches to make Odoo work in the buildout context. """ import subprocess def do_patch(gevent_script_path): """ Patch odoo prefork so that --workers execute the correct gevent script. This monkey patch could be safer, if the script path determination could be isolated from the actual process management logic in the original. """ try: from odoo.service.server import PreforkServer, stripped_sys_argv except ImportError: from openerp.service.server import PreforkServer, stripped_sys_argv def long_polling_spawn(server): nargs = stripped_sys_argv() nargs[0] = gevent_script_path popen = subprocess.Popen(nargs) server.long_polling_pid = popen.pid PreforkServer.long_polling_spawn = long_polling_spawn
3cc3c0b90714bbf7a2638b16faec69aba82a4050
op_robot_tests/tests_files/brokers/openprocurement_client_helper.py
op_robot_tests/tests_files/brokers/openprocurement_client_helper.py
from openprocurement_client.client import Client import sys def prepare_api_wrapper(key='', host_url="https://api-sandbox.openprocurement.org", api_version='0.8' ): return Client(key, host_url, api_version ) def get_internal_id(get_tenders_function, date): result = get_tenders_function({"offset": date, "opt_fields": 'tenderID', }) #import pdb; pdb.Pdb(stdout=sys.__stdout__).set_trace() return result
from openprocurement_client.client import Client import sys def prepare_api_wrapper(key='', host_url="https://api-sandbox.openprocurement.org", api_version='0.8'): return Client(key, host_url, api_version) def get_internal_id(get_tenders_function, date): result = get_tenders_function({"offset": date, "opt_fields": 'tenderID'}) return result
Improve PEP8 compliance in op_client_helper.py
Improve PEP8 compliance in op_client_helper.py
Python
apache-2.0
SlaOne/robot_tests,kosaniak/robot_tests,selurvedu/robot_tests,Leits/robot_tests,cleardevice/robot_tests,VadimShurhal/robot_tests.broker.aps,mykhaly/robot_tests,Rzaporozhets/robot_tests,bubanoid/robot_tests,openprocurement/robot_tests
from openprocurement_client.client import Client import sys def prepare_api_wrapper(key='', host_url="https://api-sandbox.openprocurement.org", api_version='0.8' ): return Client(key, host_url, api_version ) def get_internal_id(get_tenders_function, date): result = get_tenders_function({"offset": date, "opt_fields": 'tenderID', }) #import pdb; pdb.Pdb(stdout=sys.__stdout__).set_trace() return result Improve PEP8 compliance in op_client_helper.py
from openprocurement_client.client import Client import sys def prepare_api_wrapper(key='', host_url="https://api-sandbox.openprocurement.org", api_version='0.8'): return Client(key, host_url, api_version) def get_internal_id(get_tenders_function, date): result = get_tenders_function({"offset": date, "opt_fields": 'tenderID'}) return result
<commit_before>from openprocurement_client.client import Client import sys def prepare_api_wrapper(key='', host_url="https://api-sandbox.openprocurement.org", api_version='0.8' ): return Client(key, host_url, api_version ) def get_internal_id(get_tenders_function, date): result = get_tenders_function({"offset": date, "opt_fields": 'tenderID', }) #import pdb; pdb.Pdb(stdout=sys.__stdout__).set_trace() return result <commit_msg>Improve PEP8 compliance in op_client_helper.py<commit_after>
from openprocurement_client.client import Client import sys def prepare_api_wrapper(key='', host_url="https://api-sandbox.openprocurement.org", api_version='0.8'): return Client(key, host_url, api_version) def get_internal_id(get_tenders_function, date): result = get_tenders_function({"offset": date, "opt_fields": 'tenderID'}) return result
from openprocurement_client.client import Client import sys def prepare_api_wrapper(key='', host_url="https://api-sandbox.openprocurement.org", api_version='0.8' ): return Client(key, host_url, api_version ) def get_internal_id(get_tenders_function, date): result = get_tenders_function({"offset": date, "opt_fields": 'tenderID', }) #import pdb; pdb.Pdb(stdout=sys.__stdout__).set_trace() return result Improve PEP8 compliance in op_client_helper.pyfrom openprocurement_client.client import Client import sys def prepare_api_wrapper(key='', host_url="https://api-sandbox.openprocurement.org", api_version='0.8'): return Client(key, host_url, api_version) def get_internal_id(get_tenders_function, date): result = get_tenders_function({"offset": date, "opt_fields": 'tenderID'}) return result
<commit_before>from openprocurement_client.client import Client import sys def prepare_api_wrapper(key='', host_url="https://api-sandbox.openprocurement.org", api_version='0.8' ): return Client(key, host_url, api_version ) def get_internal_id(get_tenders_function, date): result = get_tenders_function({"offset": date, "opt_fields": 'tenderID', }) #import pdb; pdb.Pdb(stdout=sys.__stdout__).set_trace() return result <commit_msg>Improve PEP8 compliance in op_client_helper.py<commit_after>from openprocurement_client.client import Client import sys def prepare_api_wrapper(key='', host_url="https://api-sandbox.openprocurement.org", api_version='0.8'): return Client(key, host_url, api_version) def get_internal_id(get_tenders_function, date): result = get_tenders_function({"offset": date, "opt_fields": 'tenderID'}) return result
27c54cfd5eaf180595e671c80bd7c39406c8a24c
databroker/__init__.py
databroker/__init__.py
# Import intake to run driver discovery first and avoid circular import issues. import intake del intake import warnings import logging logger = logging.getLogger(__name__) from ._core import (Broker, BrokerES, Header, ALL, lookup_config, list_configs, describe_configs, temp_config, wrap_in_doct, DeprecatedDoct, wrap_in_deprecated_doct) from .discovery import MergedCatalog, EntrypointsCatalog, V0Catalog # A catalog created from discovered entrypoints and v0 catalogs. catalog = MergedCatalog([EntrypointsCatalog(), V0Catalog()]) # set version string using versioneer from ._version import get_versions __version__ = get_versions()['version'] del get_versions ### Legacy imports ### try: from .databroker import DataBroker except ImportError: pass else: from .databroker import (DataBroker, DataBroker as db, get_events, get_table, stream, get_fields, restream, process) from .pims_readers import get_images
# Import intake to run driver discovery first and avoid circular import issues. import intake del intake import warnings import logging logger = logging.getLogger(__name__) from .v1 import Broker, Header, ALL, temp, temp_config from .utils import (lookup_config, list_configs, describe_configs, wrap_in_doct, DeprecatedDoct, wrap_in_deprecated_doct) from .discovery import MergedCatalog, EntrypointsCatalog, V0Catalog # A catalog created from discovered entrypoints and v0 catalogs. catalog = MergedCatalog([EntrypointsCatalog(), V0Catalog()]) # set version string using versioneer from ._version import get_versions __version__ = get_versions()['version'] del get_versions ### Legacy imports ### try: from .databroker import DataBroker except ImportError: pass else: from .databroker import (DataBroker, DataBroker as db, get_events, get_table, stream, get_fields, restream, process) from .pims_readers import get_images
Move top-level imports from v0 to v1.
Move top-level imports from v0 to v1.
Python
bsd-3-clause
ericdill/databroker,ericdill/databroker
# Import intake to run driver discovery first and avoid circular import issues. import intake del intake import warnings import logging logger = logging.getLogger(__name__) from ._core import (Broker, BrokerES, Header, ALL, lookup_config, list_configs, describe_configs, temp_config, wrap_in_doct, DeprecatedDoct, wrap_in_deprecated_doct) from .discovery import MergedCatalog, EntrypointsCatalog, V0Catalog # A catalog created from discovered entrypoints and v0 catalogs. catalog = MergedCatalog([EntrypointsCatalog(), V0Catalog()]) # set version string using versioneer from ._version import get_versions __version__ = get_versions()['version'] del get_versions ### Legacy imports ### try: from .databroker import DataBroker except ImportError: pass else: from .databroker import (DataBroker, DataBroker as db, get_events, get_table, stream, get_fields, restream, process) from .pims_readers import get_images Move top-level imports from v0 to v1.
# Import intake to run driver discovery first and avoid circular import issues. import intake del intake import warnings import logging logger = logging.getLogger(__name__) from .v1 import Broker, Header, ALL, temp, temp_config from .utils import (lookup_config, list_configs, describe_configs, wrap_in_doct, DeprecatedDoct, wrap_in_deprecated_doct) from .discovery import MergedCatalog, EntrypointsCatalog, V0Catalog # A catalog created from discovered entrypoints and v0 catalogs. catalog = MergedCatalog([EntrypointsCatalog(), V0Catalog()]) # set version string using versioneer from ._version import get_versions __version__ = get_versions()['version'] del get_versions ### Legacy imports ### try: from .databroker import DataBroker except ImportError: pass else: from .databroker import (DataBroker, DataBroker as db, get_events, get_table, stream, get_fields, restream, process) from .pims_readers import get_images
<commit_before># Import intake to run driver discovery first and avoid circular import issues. import intake del intake import warnings import logging logger = logging.getLogger(__name__) from ._core import (Broker, BrokerES, Header, ALL, lookup_config, list_configs, describe_configs, temp_config, wrap_in_doct, DeprecatedDoct, wrap_in_deprecated_doct) from .discovery import MergedCatalog, EntrypointsCatalog, V0Catalog # A catalog created from discovered entrypoints and v0 catalogs. catalog = MergedCatalog([EntrypointsCatalog(), V0Catalog()]) # set version string using versioneer from ._version import get_versions __version__ = get_versions()['version'] del get_versions ### Legacy imports ### try: from .databroker import DataBroker except ImportError: pass else: from .databroker import (DataBroker, DataBroker as db, get_events, get_table, stream, get_fields, restream, process) from .pims_readers import get_images <commit_msg>Move top-level imports from v0 to v1.<commit_after>
# Import intake to run driver discovery first and avoid circular import issues. import intake del intake import warnings import logging logger = logging.getLogger(__name__) from .v1 import Broker, Header, ALL, temp, temp_config from .utils import (lookup_config, list_configs, describe_configs, wrap_in_doct, DeprecatedDoct, wrap_in_deprecated_doct) from .discovery import MergedCatalog, EntrypointsCatalog, V0Catalog # A catalog created from discovered entrypoints and v0 catalogs. catalog = MergedCatalog([EntrypointsCatalog(), V0Catalog()]) # set version string using versioneer from ._version import get_versions __version__ = get_versions()['version'] del get_versions ### Legacy imports ### try: from .databroker import DataBroker except ImportError: pass else: from .databroker import (DataBroker, DataBroker as db, get_events, get_table, stream, get_fields, restream, process) from .pims_readers import get_images
# Import intake to run driver discovery first and avoid circular import issues. import intake del intake import warnings import logging logger = logging.getLogger(__name__) from ._core import (Broker, BrokerES, Header, ALL, lookup_config, list_configs, describe_configs, temp_config, wrap_in_doct, DeprecatedDoct, wrap_in_deprecated_doct) from .discovery import MergedCatalog, EntrypointsCatalog, V0Catalog # A catalog created from discovered entrypoints and v0 catalogs. catalog = MergedCatalog([EntrypointsCatalog(), V0Catalog()]) # set version string using versioneer from ._version import get_versions __version__ = get_versions()['version'] del get_versions ### Legacy imports ### try: from .databroker import DataBroker except ImportError: pass else: from .databroker import (DataBroker, DataBroker as db, get_events, get_table, stream, get_fields, restream, process) from .pims_readers import get_images Move top-level imports from v0 to v1.# Import intake to run driver discovery first and avoid circular import issues. import intake del intake import warnings import logging logger = logging.getLogger(__name__) from .v1 import Broker, Header, ALL, temp, temp_config from .utils import (lookup_config, list_configs, describe_configs, wrap_in_doct, DeprecatedDoct, wrap_in_deprecated_doct) from .discovery import MergedCatalog, EntrypointsCatalog, V0Catalog # A catalog created from discovered entrypoints and v0 catalogs. catalog = MergedCatalog([EntrypointsCatalog(), V0Catalog()]) # set version string using versioneer from ._version import get_versions __version__ = get_versions()['version'] del get_versions ### Legacy imports ### try: from .databroker import DataBroker except ImportError: pass else: from .databroker import (DataBroker, DataBroker as db, get_events, get_table, stream, get_fields, restream, process) from .pims_readers import get_images
<commit_before># Import intake to run driver discovery first and avoid circular import issues. import intake del intake import warnings import logging logger = logging.getLogger(__name__) from ._core import (Broker, BrokerES, Header, ALL, lookup_config, list_configs, describe_configs, temp_config, wrap_in_doct, DeprecatedDoct, wrap_in_deprecated_doct) from .discovery import MergedCatalog, EntrypointsCatalog, V0Catalog # A catalog created from discovered entrypoints and v0 catalogs. catalog = MergedCatalog([EntrypointsCatalog(), V0Catalog()]) # set version string using versioneer from ._version import get_versions __version__ = get_versions()['version'] del get_versions ### Legacy imports ### try: from .databroker import DataBroker except ImportError: pass else: from .databroker import (DataBroker, DataBroker as db, get_events, get_table, stream, get_fields, restream, process) from .pims_readers import get_images <commit_msg>Move top-level imports from v0 to v1.<commit_after># Import intake to run driver discovery first and avoid circular import issues. import intake del intake import warnings import logging logger = logging.getLogger(__name__) from .v1 import Broker, Header, ALL, temp, temp_config from .utils import (lookup_config, list_configs, describe_configs, wrap_in_doct, DeprecatedDoct, wrap_in_deprecated_doct) from .discovery import MergedCatalog, EntrypointsCatalog, V0Catalog # A catalog created from discovered entrypoints and v0 catalogs. catalog = MergedCatalog([EntrypointsCatalog(), V0Catalog()]) # set version string using versioneer from ._version import get_versions __version__ = get_versions()['version'] del get_versions ### Legacy imports ### try: from .databroker import DataBroker except ImportError: pass else: from .databroker import (DataBroker, DataBroker as db, get_events, get_table, stream, get_fields, restream, process) from .pims_readers import get_images
09f86488096880870bbd3363e0a4c018f11e935d
lingcod/layers/urls.py
lingcod/layers/urls.py
from django.conf.urls.defaults import * urlpatterns = patterns('lingcod.layers.views', url(r'^public/', 'get_public_layers', name='public-data-layers'), url(r'^kml_file/(?P<session_key>\w+)/(?P<uid>[\w_]+).kml', 'get_kml_file', name='kml-file'), url(r'^privatekml/(?P<session_key>\w+)/$', 'get_privatekml_list', name='layers-privatekml-list'), url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/$', 'get_privatekml', name='layers-privatekml'), url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/(?P<path>[^\z]+)$', 'get_relative_to_privatekml', name='layers-privatekml-relative'), )
from django.conf.urls.defaults import * import time urlpatterns = patterns('lingcod.layers.views', url(r'^public/$', 'get_public_layers', name='public-data-layers'), # Useful for debugging, avoids GE caching interference url(r'^public/cachebuster/%s' % str(time.time()), 'get_public_layers', name='public-data-layers-cachebuster'), url(r'^kml_file/(?P<session_key>\w+)/(?P<uid>[\w_]+).kml', 'get_kml_file', name='kml-file'), url(r'^privatekml/(?P<session_key>\w+)/$', 'get_privatekml_list', name='layers-privatekml-list'), url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/$', 'get_privatekml', name='layers-privatekml'), url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/(?P<path>[^\z]+)$', 'get_relative_to_privatekml', name='layers-privatekml-relative'), )
Add another url pattern for debugging public layers
Add another url pattern for debugging public layers
Python
bsd-3-clause
Alwnikrotikz/marinemap,google-code-export/marinemap,Alwnikrotikz/marinemap,google-code-export/marinemap,Alwnikrotikz/marinemap,google-code-export/marinemap,google-code-export/marinemap,Alwnikrotikz/marinemap
from django.conf.urls.defaults import * urlpatterns = patterns('lingcod.layers.views', url(r'^public/', 'get_public_layers', name='public-data-layers'), url(r'^kml_file/(?P<session_key>\w+)/(?P<uid>[\w_]+).kml', 'get_kml_file', name='kml-file'), url(r'^privatekml/(?P<session_key>\w+)/$', 'get_privatekml_list', name='layers-privatekml-list'), url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/$', 'get_privatekml', name='layers-privatekml'), url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/(?P<path>[^\z]+)$', 'get_relative_to_privatekml', name='layers-privatekml-relative'), ) Add another url pattern for debugging public layers
from django.conf.urls.defaults import * import time urlpatterns = patterns('lingcod.layers.views', url(r'^public/$', 'get_public_layers', name='public-data-layers'), # Useful for debugging, avoids GE caching interference url(r'^public/cachebuster/%s' % str(time.time()), 'get_public_layers', name='public-data-layers-cachebuster'), url(r'^kml_file/(?P<session_key>\w+)/(?P<uid>[\w_]+).kml', 'get_kml_file', name='kml-file'), url(r'^privatekml/(?P<session_key>\w+)/$', 'get_privatekml_list', name='layers-privatekml-list'), url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/$', 'get_privatekml', name='layers-privatekml'), url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/(?P<path>[^\z]+)$', 'get_relative_to_privatekml', name='layers-privatekml-relative'), )
<commit_before>from django.conf.urls.defaults import * urlpatterns = patterns('lingcod.layers.views', url(r'^public/', 'get_public_layers', name='public-data-layers'), url(r'^kml_file/(?P<session_key>\w+)/(?P<uid>[\w_]+).kml', 'get_kml_file', name='kml-file'), url(r'^privatekml/(?P<session_key>\w+)/$', 'get_privatekml_list', name='layers-privatekml-list'), url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/$', 'get_privatekml', name='layers-privatekml'), url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/(?P<path>[^\z]+)$', 'get_relative_to_privatekml', name='layers-privatekml-relative'), ) <commit_msg>Add another url pattern for debugging public layers<commit_after>
from django.conf.urls.defaults import * import time urlpatterns = patterns('lingcod.layers.views', url(r'^public/$', 'get_public_layers', name='public-data-layers'), # Useful for debugging, avoids GE caching interference url(r'^public/cachebuster/%s' % str(time.time()), 'get_public_layers', name='public-data-layers-cachebuster'), url(r'^kml_file/(?P<session_key>\w+)/(?P<uid>[\w_]+).kml', 'get_kml_file', name='kml-file'), url(r'^privatekml/(?P<session_key>\w+)/$', 'get_privatekml_list', name='layers-privatekml-list'), url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/$', 'get_privatekml', name='layers-privatekml'), url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/(?P<path>[^\z]+)$', 'get_relative_to_privatekml', name='layers-privatekml-relative'), )
from django.conf.urls.defaults import * urlpatterns = patterns('lingcod.layers.views', url(r'^public/', 'get_public_layers', name='public-data-layers'), url(r'^kml_file/(?P<session_key>\w+)/(?P<uid>[\w_]+).kml', 'get_kml_file', name='kml-file'), url(r'^privatekml/(?P<session_key>\w+)/$', 'get_privatekml_list', name='layers-privatekml-list'), url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/$', 'get_privatekml', name='layers-privatekml'), url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/(?P<path>[^\z]+)$', 'get_relative_to_privatekml', name='layers-privatekml-relative'), ) Add another url pattern for debugging public layersfrom django.conf.urls.defaults import * import time urlpatterns = patterns('lingcod.layers.views', url(r'^public/$', 'get_public_layers', name='public-data-layers'), # Useful for debugging, avoids GE caching interference url(r'^public/cachebuster/%s' % str(time.time()), 'get_public_layers', name='public-data-layers-cachebuster'), url(r'^kml_file/(?P<session_key>\w+)/(?P<uid>[\w_]+).kml', 'get_kml_file', name='kml-file'), url(r'^privatekml/(?P<session_key>\w+)/$', 'get_privatekml_list', name='layers-privatekml-list'), url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/$', 'get_privatekml', name='layers-privatekml'), url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/(?P<path>[^\z]+)$', 'get_relative_to_privatekml', name='layers-privatekml-relative'), )
<commit_before>from django.conf.urls.defaults import * urlpatterns = patterns('lingcod.layers.views', url(r'^public/', 'get_public_layers', name='public-data-layers'), url(r'^kml_file/(?P<session_key>\w+)/(?P<uid>[\w_]+).kml', 'get_kml_file', name='kml-file'), url(r'^privatekml/(?P<session_key>\w+)/$', 'get_privatekml_list', name='layers-privatekml-list'), url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/$', 'get_privatekml', name='layers-privatekml'), url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/(?P<path>[^\z]+)$', 'get_relative_to_privatekml', name='layers-privatekml-relative'), ) <commit_msg>Add another url pattern for debugging public layers<commit_after>from django.conf.urls.defaults import * import time urlpatterns = patterns('lingcod.layers.views', url(r'^public/$', 'get_public_layers', name='public-data-layers'), # Useful for debugging, avoids GE caching interference url(r'^public/cachebuster/%s' % str(time.time()), 'get_public_layers', name='public-data-layers-cachebuster'), url(r'^kml_file/(?P<session_key>\w+)/(?P<uid>[\w_]+).kml', 'get_kml_file', name='kml-file'), url(r'^privatekml/(?P<session_key>\w+)/$', 'get_privatekml_list', name='layers-privatekml-list'), url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/$', 'get_privatekml', name='layers-privatekml'), url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/(?P<path>[^\z]+)$', 'get_relative_to_privatekml', name='layers-privatekml-relative'), )
a2e7642034bf89bf1d7d513ef155da3375482373
virtool/user_permissions.py
virtool/user_permissions.py
#: A list of the permission strings used by Virtool. PERMISSIONS = [ "add_virus", "modify_virus", "remove_virus", "modify_hmm", "add_sample", "add_host", "remove_host", "cancel_job", "remove_job", "archive_job", "rebuild_index", "modify_options", "manage_users" ]
#: A list of the permission strings used by Virtool. PERMISSIONS = [ "add_virus", "modify_virus", "remove_virus", "modify_hmm", "create_sample", "add_host", "remove_host", "cancel_job", "remove_job", "archive_job", "rebuild_index", "modify_options", "manage_users" ]
Change 'add_sample' permission to 'create_sample'
Change 'add_sample' permission to 'create_sample'
Python
mit
igboyes/virtool,igboyes/virtool,virtool/virtool,virtool/virtool
#: A list of the permission strings used by Virtool. PERMISSIONS = [ "add_virus", "modify_virus", "remove_virus", "modify_hmm", "add_sample", "add_host", "remove_host", "cancel_job", "remove_job", "archive_job", "rebuild_index", "modify_options", "manage_users" ] Change 'add_sample' permission to 'create_sample'
#: A list of the permission strings used by Virtool. PERMISSIONS = [ "add_virus", "modify_virus", "remove_virus", "modify_hmm", "create_sample", "add_host", "remove_host", "cancel_job", "remove_job", "archive_job", "rebuild_index", "modify_options", "manage_users" ]
<commit_before>#: A list of the permission strings used by Virtool. PERMISSIONS = [ "add_virus", "modify_virus", "remove_virus", "modify_hmm", "add_sample", "add_host", "remove_host", "cancel_job", "remove_job", "archive_job", "rebuild_index", "modify_options", "manage_users" ] <commit_msg>Change 'add_sample' permission to 'create_sample'<commit_after>
#: A list of the permission strings used by Virtool. PERMISSIONS = [ "add_virus", "modify_virus", "remove_virus", "modify_hmm", "create_sample", "add_host", "remove_host", "cancel_job", "remove_job", "archive_job", "rebuild_index", "modify_options", "manage_users" ]
#: A list of the permission strings used by Virtool. PERMISSIONS = [ "add_virus", "modify_virus", "remove_virus", "modify_hmm", "add_sample", "add_host", "remove_host", "cancel_job", "remove_job", "archive_job", "rebuild_index", "modify_options", "manage_users" ] Change 'add_sample' permission to 'create_sample'#: A list of the permission strings used by Virtool. PERMISSIONS = [ "add_virus", "modify_virus", "remove_virus", "modify_hmm", "create_sample", "add_host", "remove_host", "cancel_job", "remove_job", "archive_job", "rebuild_index", "modify_options", "manage_users" ]
<commit_before>#: A list of the permission strings used by Virtool. PERMISSIONS = [ "add_virus", "modify_virus", "remove_virus", "modify_hmm", "add_sample", "add_host", "remove_host", "cancel_job", "remove_job", "archive_job", "rebuild_index", "modify_options", "manage_users" ] <commit_msg>Change 'add_sample' permission to 'create_sample'<commit_after>#: A list of the permission strings used by Virtool. PERMISSIONS = [ "add_virus", "modify_virus", "remove_virus", "modify_hmm", "create_sample", "add_host", "remove_host", "cancel_job", "remove_job", "archive_job", "rebuild_index", "modify_options", "manage_users" ]
685116d1a2799399819ed780679403e7576e67b5
keystone/tests/unit/common/test_manager.py
keystone/tests/unit/common/test_manager.py
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import mock from keystone import catalog from keystone.common import manager from keystone.tests import unit class TestCreateLegacyDriver(unit.BaseTestCase): @mock.patch('oslo_log.versionutils.report_deprecated_feature') def test_class_is_properly_deprecated(self, mock_reporter): Driver = manager.create_legacy_driver(catalog.CatalogDriverV8) # NOTE(dstanek): I want to subvert the requirement for this # class to implement all of the abstract methods. Driver.__abstractmethods__ = set() impl = Driver() details = { 'as_of': 'Liberty', 'what': 'keystone.catalog.core.Driver', 'in_favor_of': 'keystone.catalog.core.CatalogDriverV8', 'remove_in': 'N', } mock_reporter.assert_called_with(mock.ANY, mock.ANY, details) self.assertIsInstance(impl, catalog.CatalogDriverV8)
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import mock from keystone import catalog from keystone.common import manager from keystone.tests import unit class TestCreateLegacyDriver(unit.BaseTestCase): @mock.patch('oslo_log.versionutils.report_deprecated_feature') def test_class_is_properly_deprecated(self, mock_reporter): Driver = manager.create_legacy_driver(catalog.CatalogDriverV8) # NOTE(dstanek): I want to subvert the requirement for this # class to implement all of the abstract methods. Driver.__abstractmethods__ = set() impl = Driver() details = { 'as_of': 'Liberty', 'what': 'keystone.catalog.core.Driver', 'in_favor_of': 'keystone.catalog.core.CatalogDriverV8', 'remove_in': mock.ANY, } mock_reporter.assert_called_with(mock.ANY, mock.ANY, details) self.assertEqual('N', mock_reporter.call_args[0][2]['remove_in'][0]) self.assertIsInstance(impl, catalog.CatalogDriverV8)
Correct test to support changing N release name
Correct test to support changing N release name oslo.log is going to change to use Newton rather than N so this test should not make an assumption about the way that versionutils.deprecated is calling report_deprecated_feature. Change-Id: I06aa6d085232376811f73597b2d84b5174bc7a8d Closes-Bug: 1561121 (cherry picked from commit 8556437ee02de028ec5de3b867abaab82533cb91)
Python
apache-2.0
openstack/keystone,openstack/keystone,cernops/keystone,klmitch/keystone,mahak/keystone,mahak/keystone,rajalokan/keystone,ilay09/keystone,openstack/keystone,ilay09/keystone,ilay09/keystone,cernops/keystone,klmitch/keystone,mahak/keystone,rajalokan/keystone,rajalokan/keystone
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import mock from keystone import catalog from keystone.common import manager from keystone.tests import unit class TestCreateLegacyDriver(unit.BaseTestCase): @mock.patch('oslo_log.versionutils.report_deprecated_feature') def test_class_is_properly_deprecated(self, mock_reporter): Driver = manager.create_legacy_driver(catalog.CatalogDriverV8) # NOTE(dstanek): I want to subvert the requirement for this # class to implement all of the abstract methods. Driver.__abstractmethods__ = set() impl = Driver() details = { 'as_of': 'Liberty', 'what': 'keystone.catalog.core.Driver', 'in_favor_of': 'keystone.catalog.core.CatalogDriverV8', 'remove_in': 'N', } mock_reporter.assert_called_with(mock.ANY, mock.ANY, details) self.assertIsInstance(impl, catalog.CatalogDriverV8) Correct test to support changing N release name oslo.log is going to change to use Newton rather than N so this test should not make an assumption about the way that versionutils.deprecated is calling report_deprecated_feature. Change-Id: I06aa6d085232376811f73597b2d84b5174bc7a8d Closes-Bug: 1561121 (cherry picked from commit 8556437ee02de028ec5de3b867abaab82533cb91)
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import mock from keystone import catalog from keystone.common import manager from keystone.tests import unit class TestCreateLegacyDriver(unit.BaseTestCase): @mock.patch('oslo_log.versionutils.report_deprecated_feature') def test_class_is_properly_deprecated(self, mock_reporter): Driver = manager.create_legacy_driver(catalog.CatalogDriverV8) # NOTE(dstanek): I want to subvert the requirement for this # class to implement all of the abstract methods. Driver.__abstractmethods__ = set() impl = Driver() details = { 'as_of': 'Liberty', 'what': 'keystone.catalog.core.Driver', 'in_favor_of': 'keystone.catalog.core.CatalogDriverV8', 'remove_in': mock.ANY, } mock_reporter.assert_called_with(mock.ANY, mock.ANY, details) self.assertEqual('N', mock_reporter.call_args[0][2]['remove_in'][0]) self.assertIsInstance(impl, catalog.CatalogDriverV8)
<commit_before># Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import mock from keystone import catalog from keystone.common import manager from keystone.tests import unit class TestCreateLegacyDriver(unit.BaseTestCase): @mock.patch('oslo_log.versionutils.report_deprecated_feature') def test_class_is_properly_deprecated(self, mock_reporter): Driver = manager.create_legacy_driver(catalog.CatalogDriverV8) # NOTE(dstanek): I want to subvert the requirement for this # class to implement all of the abstract methods. Driver.__abstractmethods__ = set() impl = Driver() details = { 'as_of': 'Liberty', 'what': 'keystone.catalog.core.Driver', 'in_favor_of': 'keystone.catalog.core.CatalogDriverV8', 'remove_in': 'N', } mock_reporter.assert_called_with(mock.ANY, mock.ANY, details) self.assertIsInstance(impl, catalog.CatalogDriverV8) <commit_msg>Correct test to support changing N release name oslo.log is going to change to use Newton rather than N so this test should not make an assumption about the way that versionutils.deprecated is calling report_deprecated_feature. Change-Id: I06aa6d085232376811f73597b2d84b5174bc7a8d Closes-Bug: 1561121 (cherry picked from commit 8556437ee02de028ec5de3b867abaab82533cb91)<commit_after>
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import mock from keystone import catalog from keystone.common import manager from keystone.tests import unit class TestCreateLegacyDriver(unit.BaseTestCase): @mock.patch('oslo_log.versionutils.report_deprecated_feature') def test_class_is_properly_deprecated(self, mock_reporter): Driver = manager.create_legacy_driver(catalog.CatalogDriverV8) # NOTE(dstanek): I want to subvert the requirement for this # class to implement all of the abstract methods. Driver.__abstractmethods__ = set() impl = Driver() details = { 'as_of': 'Liberty', 'what': 'keystone.catalog.core.Driver', 'in_favor_of': 'keystone.catalog.core.CatalogDriverV8', 'remove_in': mock.ANY, } mock_reporter.assert_called_with(mock.ANY, mock.ANY, details) self.assertEqual('N', mock_reporter.call_args[0][2]['remove_in'][0]) self.assertIsInstance(impl, catalog.CatalogDriverV8)
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import mock from keystone import catalog from keystone.common import manager from keystone.tests import unit class TestCreateLegacyDriver(unit.BaseTestCase): @mock.patch('oslo_log.versionutils.report_deprecated_feature') def test_class_is_properly_deprecated(self, mock_reporter): Driver = manager.create_legacy_driver(catalog.CatalogDriverV8) # NOTE(dstanek): I want to subvert the requirement for this # class to implement all of the abstract methods. Driver.__abstractmethods__ = set() impl = Driver() details = { 'as_of': 'Liberty', 'what': 'keystone.catalog.core.Driver', 'in_favor_of': 'keystone.catalog.core.CatalogDriverV8', 'remove_in': 'N', } mock_reporter.assert_called_with(mock.ANY, mock.ANY, details) self.assertIsInstance(impl, catalog.CatalogDriverV8) Correct test to support changing N release name oslo.log is going to change to use Newton rather than N so this test should not make an assumption about the way that versionutils.deprecated is calling report_deprecated_feature. Change-Id: I06aa6d085232376811f73597b2d84b5174bc7a8d Closes-Bug: 1561121 (cherry picked from commit 8556437ee02de028ec5de3b867abaab82533cb91)# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import mock from keystone import catalog from keystone.common import manager from keystone.tests import unit class TestCreateLegacyDriver(unit.BaseTestCase): @mock.patch('oslo_log.versionutils.report_deprecated_feature') def test_class_is_properly_deprecated(self, mock_reporter): Driver = manager.create_legacy_driver(catalog.CatalogDriverV8) # NOTE(dstanek): I want to subvert the requirement for this # class to implement all of the abstract methods. Driver.__abstractmethods__ = set() impl = Driver() details = { 'as_of': 'Liberty', 'what': 'keystone.catalog.core.Driver', 'in_favor_of': 'keystone.catalog.core.CatalogDriverV8', 'remove_in': mock.ANY, } mock_reporter.assert_called_with(mock.ANY, mock.ANY, details) self.assertEqual('N', mock_reporter.call_args[0][2]['remove_in'][0]) self.assertIsInstance(impl, catalog.CatalogDriverV8)
<commit_before># Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import mock from keystone import catalog from keystone.common import manager from keystone.tests import unit class TestCreateLegacyDriver(unit.BaseTestCase): @mock.patch('oslo_log.versionutils.report_deprecated_feature') def test_class_is_properly_deprecated(self, mock_reporter): Driver = manager.create_legacy_driver(catalog.CatalogDriverV8) # NOTE(dstanek): I want to subvert the requirement for this # class to implement all of the abstract methods. Driver.__abstractmethods__ = set() impl = Driver() details = { 'as_of': 'Liberty', 'what': 'keystone.catalog.core.Driver', 'in_favor_of': 'keystone.catalog.core.CatalogDriverV8', 'remove_in': 'N', } mock_reporter.assert_called_with(mock.ANY, mock.ANY, details) self.assertIsInstance(impl, catalog.CatalogDriverV8) <commit_msg>Correct test to support changing N release name oslo.log is going to change to use Newton rather than N so this test should not make an assumption about the way that versionutils.deprecated is calling report_deprecated_feature. Change-Id: I06aa6d085232376811f73597b2d84b5174bc7a8d Closes-Bug: 1561121 (cherry picked from commit 8556437ee02de028ec5de3b867abaab82533cb91)<commit_after># Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import mock from keystone import catalog from keystone.common import manager from keystone.tests import unit class TestCreateLegacyDriver(unit.BaseTestCase): @mock.patch('oslo_log.versionutils.report_deprecated_feature') def test_class_is_properly_deprecated(self, mock_reporter): Driver = manager.create_legacy_driver(catalog.CatalogDriverV8) # NOTE(dstanek): I want to subvert the requirement for this # class to implement all of the abstract methods. Driver.__abstractmethods__ = set() impl = Driver() details = { 'as_of': 'Liberty', 'what': 'keystone.catalog.core.Driver', 'in_favor_of': 'keystone.catalog.core.CatalogDriverV8', 'remove_in': mock.ANY, } mock_reporter.assert_called_with(mock.ANY, mock.ANY, details) self.assertEqual('N', mock_reporter.call_args[0][2]['remove_in'][0]) self.assertIsInstance(impl, catalog.CatalogDriverV8)
74bd9ffd412f22671232cb301b3762660a73d912
lot/landmapper/urls.py
lot/landmapper/urls.py
from django.urls import include, re_path, path from landmapper.views import * urlpatterns = [ # What is difference between re_path and path? # re_path(r'', # home, name='landmapper-home'), path('', home, name="home"), path('/identify/', identify, name="identify"), path('/report/', report, name="report"), path('/pages/', include('django.contrib.flatpages.urls')), url(r'^get_taxlot_json', views.get_taxlot_json, name='get taxlot json'), ]
from django.urls import include, re_path, path from landmapper.views import * urlpatterns = [ # What is difference between re_path and path? # re_path(r'', # home, name='landmapper-home'), path('', home, name="home"), path('/identify', identify, name="identify"), path('/report', report, name="report"), path('/get_taxlot_json', get_taxlot_json, name='get taxlot json'), ]
Fix get taxlot url and remove trailing slashes
Fix get taxlot url and remove trailing slashes
Python
bsd-3-clause
Ecotrust/forestplanner,Ecotrust/forestplanner,Ecotrust/forestplanner,Ecotrust/forestplanner,Ecotrust/forestplanner,Ecotrust/forestplanner,Ecotrust/forestplanner,Ecotrust/forestplanner
from django.urls import include, re_path, path from landmapper.views import * urlpatterns = [ # What is difference between re_path and path? # re_path(r'', # home, name='landmapper-home'), path('', home, name="home"), path('/identify/', identify, name="identify"), path('/report/', report, name="report"), path('/pages/', include('django.contrib.flatpages.urls')), url(r'^get_taxlot_json', views.get_taxlot_json, name='get taxlot json'), ] Fix get taxlot url and remove trailing slashes
from django.urls import include, re_path, path from landmapper.views import * urlpatterns = [ # What is difference between re_path and path? # re_path(r'', # home, name='landmapper-home'), path('', home, name="home"), path('/identify', identify, name="identify"), path('/report', report, name="report"), path('/get_taxlot_json', get_taxlot_json, name='get taxlot json'), ]
<commit_before>from django.urls import include, re_path, path from landmapper.views import * urlpatterns = [ # What is difference between re_path and path? # re_path(r'', # home, name='landmapper-home'), path('', home, name="home"), path('/identify/', identify, name="identify"), path('/report/', report, name="report"), path('/pages/', include('django.contrib.flatpages.urls')), url(r'^get_taxlot_json', views.get_taxlot_json, name='get taxlot json'), ] <commit_msg>Fix get taxlot url and remove trailing slashes<commit_after>
from django.urls import include, re_path, path from landmapper.views import * urlpatterns = [ # What is difference between re_path and path? # re_path(r'', # home, name='landmapper-home'), path('', home, name="home"), path('/identify', identify, name="identify"), path('/report', report, name="report"), path('/get_taxlot_json', get_taxlot_json, name='get taxlot json'), ]
from django.urls import include, re_path, path from landmapper.views import * urlpatterns = [ # What is difference between re_path and path? # re_path(r'', # home, name='landmapper-home'), path('', home, name="home"), path('/identify/', identify, name="identify"), path('/report/', report, name="report"), path('/pages/', include('django.contrib.flatpages.urls')), url(r'^get_taxlot_json', views.get_taxlot_json, name='get taxlot json'), ] Fix get taxlot url and remove trailing slashesfrom django.urls import include, re_path, path from landmapper.views import * urlpatterns = [ # What is difference between re_path and path? # re_path(r'', # home, name='landmapper-home'), path('', home, name="home"), path('/identify', identify, name="identify"), path('/report', report, name="report"), path('/get_taxlot_json', get_taxlot_json, name='get taxlot json'), ]
<commit_before>from django.urls import include, re_path, path from landmapper.views import * urlpatterns = [ # What is difference between re_path and path? # re_path(r'', # home, name='landmapper-home'), path('', home, name="home"), path('/identify/', identify, name="identify"), path('/report/', report, name="report"), path('/pages/', include('django.contrib.flatpages.urls')), url(r'^get_taxlot_json', views.get_taxlot_json, name='get taxlot json'), ] <commit_msg>Fix get taxlot url and remove trailing slashes<commit_after>from django.urls import include, re_path, path from landmapper.views import * urlpatterns = [ # What is difference between re_path and path? # re_path(r'', # home, name='landmapper-home'), path('', home, name="home"), path('/identify', identify, name="identify"), path('/report', report, name="report"), path('/get_taxlot_json', get_taxlot_json, name='get taxlot json'), ]
789b33f8c6d4ddad4c46e7a3815d9f9543485caa
usb/blueprints/api.py
usb/blueprints/api.py
from flask import Blueprint, jsonify, request from usb.models import db, Redirect, DeviceType from usb.shortener import get_short_id, get_short_url api = Blueprint('api', __name__) @api.route('/links') def get_links(): return jsonify({}), 200 @api.route('/links', methods=['POST']) def shorten_url(): short_id = get_short_id() long_url = request.json['url'] for device_type in DeviceType: db.session.add(Redirect(short_id, device_type, long_url)) db.session.commit() short_url = get_short_url(short_id) return jsonify(url=short_url), 200
from flask import Blueprint, jsonify, request from usb.models import db, Redirect, DeviceType from usb.shortener import get_short_id, get_short_url api = Blueprint('api', __name__) @api.route('/links') def get_links(): return jsonify({}), 200 @api.route('/links', methods=['POST']) def shorten_url(): short_id = get_short_id() long_url = request.json['url'] redirect = Redirect.query.filter_by(url=long_url).first() if redirect: short_url = get_short_url(redirect.short) return jsonify(url=short_url), 409 for device_type in DeviceType: db.session.add(Redirect(short_id, device_type, long_url)) db.session.commit() short_url = get_short_url(short_id) return jsonify(url=short_url), 200
Return short URL if it's already exists
Return short URL if it's already exists
Python
mit
dizpers/usb
from flask import Blueprint, jsonify, request from usb.models import db, Redirect, DeviceType from usb.shortener import get_short_id, get_short_url api = Blueprint('api', __name__) @api.route('/links') def get_links(): return jsonify({}), 200 @api.route('/links', methods=['POST']) def shorten_url(): short_id = get_short_id() long_url = request.json['url'] for device_type in DeviceType: db.session.add(Redirect(short_id, device_type, long_url)) db.session.commit() short_url = get_short_url(short_id) return jsonify(url=short_url), 200 Return short URL if it's already exists
from flask import Blueprint, jsonify, request from usb.models import db, Redirect, DeviceType from usb.shortener import get_short_id, get_short_url api = Blueprint('api', __name__) @api.route('/links') def get_links(): return jsonify({}), 200 @api.route('/links', methods=['POST']) def shorten_url(): short_id = get_short_id() long_url = request.json['url'] redirect = Redirect.query.filter_by(url=long_url).first() if redirect: short_url = get_short_url(redirect.short) return jsonify(url=short_url), 409 for device_type in DeviceType: db.session.add(Redirect(short_id, device_type, long_url)) db.session.commit() short_url = get_short_url(short_id) return jsonify(url=short_url), 200
<commit_before>from flask import Blueprint, jsonify, request from usb.models import db, Redirect, DeviceType from usb.shortener import get_short_id, get_short_url api = Blueprint('api', __name__) @api.route('/links') def get_links(): return jsonify({}), 200 @api.route('/links', methods=['POST']) def shorten_url(): short_id = get_short_id() long_url = request.json['url'] for device_type in DeviceType: db.session.add(Redirect(short_id, device_type, long_url)) db.session.commit() short_url = get_short_url(short_id) return jsonify(url=short_url), 200 <commit_msg>Return short URL if it's already exists<commit_after>
from flask import Blueprint, jsonify, request from usb.models import db, Redirect, DeviceType from usb.shortener import get_short_id, get_short_url api = Blueprint('api', __name__) @api.route('/links') def get_links(): return jsonify({}), 200 @api.route('/links', methods=['POST']) def shorten_url(): short_id = get_short_id() long_url = request.json['url'] redirect = Redirect.query.filter_by(url=long_url).first() if redirect: short_url = get_short_url(redirect.short) return jsonify(url=short_url), 409 for device_type in DeviceType: db.session.add(Redirect(short_id, device_type, long_url)) db.session.commit() short_url = get_short_url(short_id) return jsonify(url=short_url), 200
from flask import Blueprint, jsonify, request from usb.models import db, Redirect, DeviceType from usb.shortener import get_short_id, get_short_url api = Blueprint('api', __name__) @api.route('/links') def get_links(): return jsonify({}), 200 @api.route('/links', methods=['POST']) def shorten_url(): short_id = get_short_id() long_url = request.json['url'] for device_type in DeviceType: db.session.add(Redirect(short_id, device_type, long_url)) db.session.commit() short_url = get_short_url(short_id) return jsonify(url=short_url), 200 Return short URL if it's already existsfrom flask import Blueprint, jsonify, request from usb.models import db, Redirect, DeviceType from usb.shortener import get_short_id, get_short_url api = Blueprint('api', __name__) @api.route('/links') def get_links(): return jsonify({}), 200 @api.route('/links', methods=['POST']) def shorten_url(): short_id = get_short_id() long_url = request.json['url'] redirect = Redirect.query.filter_by(url=long_url).first() if redirect: short_url = get_short_url(redirect.short) return jsonify(url=short_url), 409 for device_type in DeviceType: db.session.add(Redirect(short_id, device_type, long_url)) db.session.commit() short_url = get_short_url(short_id) return jsonify(url=short_url), 200
<commit_before>from flask import Blueprint, jsonify, request from usb.models import db, Redirect, DeviceType from usb.shortener import get_short_id, get_short_url api = Blueprint('api', __name__) @api.route('/links') def get_links(): return jsonify({}), 200 @api.route('/links', methods=['POST']) def shorten_url(): short_id = get_short_id() long_url = request.json['url'] for device_type in DeviceType: db.session.add(Redirect(short_id, device_type, long_url)) db.session.commit() short_url = get_short_url(short_id) return jsonify(url=short_url), 200 <commit_msg>Return short URL if it's already exists<commit_after>from flask import Blueprint, jsonify, request from usb.models import db, Redirect, DeviceType from usb.shortener import get_short_id, get_short_url api = Blueprint('api', __name__) @api.route('/links') def get_links(): return jsonify({}), 200 @api.route('/links', methods=['POST']) def shorten_url(): short_id = get_short_id() long_url = request.json['url'] redirect = Redirect.query.filter_by(url=long_url).first() if redirect: short_url = get_short_url(redirect.short) return jsonify(url=short_url), 409 for device_type in DeviceType: db.session.add(Redirect(short_id, device_type, long_url)) db.session.commit() short_url = get_short_url(short_id) return jsonify(url=short_url), 200
9b4e7a06932d6ed6a5a9032619fa433629187d69
utilkit/stringutil.py
utilkit/stringutil.py
""" String/unicode helper functions """ def safe_unicode(obj, *args): """ return the unicode representation of obj """ try: return unicode(obj, *args) # pylint:disable=undefined-variable except UnicodeDecodeError: # obj is byte string ascii_text = str(obj).encode('string_escape') return unicode(ascii_text) # pylint:disable=undefined-variable def safe_str(obj): """ return the byte string representation of obj """ try: return str(obj) except UnicodeEncodeError: # obj is unicode return unicode(obj).encode('unicode_escape') # pylint:disable=undefined-variable
""" String/unicode helper functions """ def safe_unicode(obj, *args): """ return the unicode representation of obj """ try: return unicode(obj, *args) # noqa for undefined-variable except UnicodeDecodeError: # obj is byte string ascii_text = str(obj).encode('string_escape') return unicode(ascii_text) # noqa for undefined-variable def safe_str(obj): """ return the byte string representation of obj """ try: return str(obj) except UnicodeEncodeError: # obj is unicode return unicode(obj).encode('unicode_escape') # noqa for undefined-variable
Disable error-checking that assumes Python 3 for these Python 2 helpers, landscape.io style
Disable error-checking that assumes Python 3 for these Python 2 helpers, landscape.io style
Python
mit
aquatix/python-utilkit
""" String/unicode helper functions """ def safe_unicode(obj, *args): """ return the unicode representation of obj """ try: return unicode(obj, *args) # pylint:disable=undefined-variable except UnicodeDecodeError: # obj is byte string ascii_text = str(obj).encode('string_escape') return unicode(ascii_text) # pylint:disable=undefined-variable def safe_str(obj): """ return the byte string representation of obj """ try: return str(obj) except UnicodeEncodeError: # obj is unicode return unicode(obj).encode('unicode_escape') # pylint:disable=undefined-variable Disable error-checking that assumes Python 3 for these Python 2 helpers, landscape.io style
""" String/unicode helper functions """ def safe_unicode(obj, *args): """ return the unicode representation of obj """ try: return unicode(obj, *args) # noqa for undefined-variable except UnicodeDecodeError: # obj is byte string ascii_text = str(obj).encode('string_escape') return unicode(ascii_text) # noqa for undefined-variable def safe_str(obj): """ return the byte string representation of obj """ try: return str(obj) except UnicodeEncodeError: # obj is unicode return unicode(obj).encode('unicode_escape') # noqa for undefined-variable
<commit_before>""" String/unicode helper functions """ def safe_unicode(obj, *args): """ return the unicode representation of obj """ try: return unicode(obj, *args) # pylint:disable=undefined-variable except UnicodeDecodeError: # obj is byte string ascii_text = str(obj).encode('string_escape') return unicode(ascii_text) # pylint:disable=undefined-variable def safe_str(obj): """ return the byte string representation of obj """ try: return str(obj) except UnicodeEncodeError: # obj is unicode return unicode(obj).encode('unicode_escape') # pylint:disable=undefined-variable <commit_msg>Disable error-checking that assumes Python 3 for these Python 2 helpers, landscape.io style<commit_after>
""" String/unicode helper functions """ def safe_unicode(obj, *args): """ return the unicode representation of obj """ try: return unicode(obj, *args) # noqa for undefined-variable except UnicodeDecodeError: # obj is byte string ascii_text = str(obj).encode('string_escape') return unicode(ascii_text) # noqa for undefined-variable def safe_str(obj): """ return the byte string representation of obj """ try: return str(obj) except UnicodeEncodeError: # obj is unicode return unicode(obj).encode('unicode_escape') # noqa for undefined-variable
""" String/unicode helper functions """ def safe_unicode(obj, *args): """ return the unicode representation of obj """ try: return unicode(obj, *args) # pylint:disable=undefined-variable except UnicodeDecodeError: # obj is byte string ascii_text = str(obj).encode('string_escape') return unicode(ascii_text) # pylint:disable=undefined-variable def safe_str(obj): """ return the byte string representation of obj """ try: return str(obj) except UnicodeEncodeError: # obj is unicode return unicode(obj).encode('unicode_escape') # pylint:disable=undefined-variable Disable error-checking that assumes Python 3 for these Python 2 helpers, landscape.io style""" String/unicode helper functions """ def safe_unicode(obj, *args): """ return the unicode representation of obj """ try: return unicode(obj, *args) # noqa for undefined-variable except UnicodeDecodeError: # obj is byte string ascii_text = str(obj).encode('string_escape') return unicode(ascii_text) # noqa for undefined-variable def safe_str(obj): """ return the byte string representation of obj """ try: return str(obj) except UnicodeEncodeError: # obj is unicode return unicode(obj).encode('unicode_escape') # noqa for undefined-variable
<commit_before>""" String/unicode helper functions """ def safe_unicode(obj, *args): """ return the unicode representation of obj """ try: return unicode(obj, *args) # pylint:disable=undefined-variable except UnicodeDecodeError: # obj is byte string ascii_text = str(obj).encode('string_escape') return unicode(ascii_text) # pylint:disable=undefined-variable def safe_str(obj): """ return the byte string representation of obj """ try: return str(obj) except UnicodeEncodeError: # obj is unicode return unicode(obj).encode('unicode_escape') # pylint:disable=undefined-variable <commit_msg>Disable error-checking that assumes Python 3 for these Python 2 helpers, landscape.io style<commit_after>""" String/unicode helper functions """ def safe_unicode(obj, *args): """ return the unicode representation of obj """ try: return unicode(obj, *args) # noqa for undefined-variable except UnicodeDecodeError: # obj is byte string ascii_text = str(obj).encode('string_escape') return unicode(ascii_text) # noqa for undefined-variable def safe_str(obj): """ return the byte string representation of obj """ try: return str(obj) except UnicodeEncodeError: # obj is unicode return unicode(obj).encode('unicode_escape') # noqa for undefined-variable
8995cbf71454e3424e15913661ee659c48f7b8fa
volunteer_planner/settings/local_mysql.py
volunteer_planner/settings/local_mysql.py
# coding: utf-8 from volunteer_planner.settings.local import * DATABASES = { 'default': { 'ENGINE': 'django.db.backends.mysql', 'NAME': 'volunteer_planner', 'PASSWORD': os.environ.get('DATABASE_PW', 'volunteer_planner'), 'USER': os.environ.get('DB_USER', 'vp') } }
# coding: utf-8 from volunteer_planner.settings.local import * DATABASES = { 'default': { 'ENGINE': 'django.db.backends.mysql', 'NAME': os.environ.get('DATABASE_NAME', 'volunteer_planner'), 'PASSWORD': os.environ.get('DATABASE_PW', 'volunteer_planner'), 'USER': os.environ.get('DB_USER', 'vp') } }
Make local mysql db name overridable with DATABASE_NAME environment variable
Make local mysql db name overridable with DATABASE_NAME environment variable
Python
agpl-3.0
christophmeissner/volunteer_planner,christophmeissner/volunteer_planner,coders4help/volunteer_planner,klinger/volunteer_planner,klinger/volunteer_planner,pitpalme/volunteer_planner,pitpalme/volunteer_planner,pitpalme/volunteer_planner,coders4help/volunteer_planner,pitpalme/volunteer_planner,coders4help/volunteer_planner,christophmeissner/volunteer_planner,alper/volunteer_planner,klinger/volunteer_planner,coders4help/volunteer_planner,alper/volunteer_planner,volunteer-planner/volunteer_planner,christophmeissner/volunteer_planner,volunteer-planner/volunteer_planner,klinger/volunteer_planner,volunteer-planner/volunteer_planner,volunteer-planner/volunteer_planner,alper/volunteer_planner
# coding: utf-8 from volunteer_planner.settings.local import * DATABASES = { 'default': { 'ENGINE': 'django.db.backends.mysql', 'NAME': 'volunteer_planner', 'PASSWORD': os.environ.get('DATABASE_PW', 'volunteer_planner'), 'USER': os.environ.get('DB_USER', 'vp') } } Make local mysql db name overridable with DATABASE_NAME environment variable
# coding: utf-8 from volunteer_planner.settings.local import * DATABASES = { 'default': { 'ENGINE': 'django.db.backends.mysql', 'NAME': os.environ.get('DATABASE_NAME', 'volunteer_planner'), 'PASSWORD': os.environ.get('DATABASE_PW', 'volunteer_planner'), 'USER': os.environ.get('DB_USER', 'vp') } }
<commit_before># coding: utf-8 from volunteer_planner.settings.local import * DATABASES = { 'default': { 'ENGINE': 'django.db.backends.mysql', 'NAME': 'volunteer_planner', 'PASSWORD': os.environ.get('DATABASE_PW', 'volunteer_planner'), 'USER': os.environ.get('DB_USER', 'vp') } } <commit_msg>Make local mysql db name overridable with DATABASE_NAME environment variable<commit_after>
# coding: utf-8 from volunteer_planner.settings.local import * DATABASES = { 'default': { 'ENGINE': 'django.db.backends.mysql', 'NAME': os.environ.get('DATABASE_NAME', 'volunteer_planner'), 'PASSWORD': os.environ.get('DATABASE_PW', 'volunteer_planner'), 'USER': os.environ.get('DB_USER', 'vp') } }
# coding: utf-8 from volunteer_planner.settings.local import * DATABASES = { 'default': { 'ENGINE': 'django.db.backends.mysql', 'NAME': 'volunteer_planner', 'PASSWORD': os.environ.get('DATABASE_PW', 'volunteer_planner'), 'USER': os.environ.get('DB_USER', 'vp') } } Make local mysql db name overridable with DATABASE_NAME environment variable# coding: utf-8 from volunteer_planner.settings.local import * DATABASES = { 'default': { 'ENGINE': 'django.db.backends.mysql', 'NAME': os.environ.get('DATABASE_NAME', 'volunteer_planner'), 'PASSWORD': os.environ.get('DATABASE_PW', 'volunteer_planner'), 'USER': os.environ.get('DB_USER', 'vp') } }
<commit_before># coding: utf-8 from volunteer_planner.settings.local import * DATABASES = { 'default': { 'ENGINE': 'django.db.backends.mysql', 'NAME': 'volunteer_planner', 'PASSWORD': os.environ.get('DATABASE_PW', 'volunteer_planner'), 'USER': os.environ.get('DB_USER', 'vp') } } <commit_msg>Make local mysql db name overridable with DATABASE_NAME environment variable<commit_after># coding: utf-8 from volunteer_planner.settings.local import * DATABASES = { 'default': { 'ENGINE': 'django.db.backends.mysql', 'NAME': os.environ.get('DATABASE_NAME', 'volunteer_planner'), 'PASSWORD': os.environ.get('DATABASE_PW', 'volunteer_planner'), 'USER': os.environ.get('DB_USER', 'vp') } }
7fb1b95205de32ec27b4e5428928b1bba417c9c8
build/fbcode_builder/specs/fbthrift.py
build/fbcode_builder/specs/fbthrift.py
#!/usr/bin/env python # Copyright (c) Facebook, Inc. and its affiliates. from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import specs.folly as folly import specs.fizz as fizz import specs.rsocket as rsocket import specs.sodium as sodium import specs.wangle as wangle import specs.zstd as zstd from shell_quoting import ShellQuoted def fbcode_builder_spec(builder): # This API should change rarely, so build the latest tag instead of master. builder.add_option( 'no1msd/mstch:git_hash', ShellQuoted('$(git describe --abbrev=0 --tags)') ) builder.add_option('krb5/krb5:git_hash', 'krb5-1.16.1-final') return { 'depends_on': [folly, fizz, sodium, rsocket, wangle, zstd], 'steps': [ # This isn't a separete spec, since only fbthrift uses mstch. builder.github_project_workdir('no1msd/mstch', 'build'), builder.cmake_install('no1msd/mstch'), builder.github_project_workdir('krb5/krb5', 'src'), builder.autoconf_install('krb5/krb5'), builder.fb_github_cmake_install('fbthrift/thrift'), ], }
#!/usr/bin/env python # Copyright (c) Facebook, Inc. and its affiliates. from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import specs.folly as folly import specs.fizz as fizz import specs.rsocket as rsocket import specs.sodium as sodium import specs.wangle as wangle import specs.zstd as zstd from shell_quoting import ShellQuoted def fbcode_builder_spec(builder): # This API should change rarely, so build the latest tag instead of master. builder.add_option( 'no1msd/mstch:git_hash', ShellQuoted('$(git describe --abbrev=0 --tags)') ) return { 'depends_on': [folly, fizz, sodium, rsocket, wangle, zstd], 'steps': [ # This isn't a separete spec, since only fbthrift uses mstch. builder.github_project_workdir('no1msd/mstch', 'build'), builder.cmake_install('no1msd/mstch'), builder.fb_github_cmake_install('fbthrift/thrift'), ], }
Cut fbcode_builder dep for thrift on krb5
Cut fbcode_builder dep for thrift on krb5 Summary: [Thrift] Cut `fbcode_builder` dep for `thrift` on `krb5`. In the past, Thrift depended on Kerberos and the `krb5` implementation for its transport-layer security. However, Thrift has since migrated fully to Transport Layer Security for its transport-layer security and no longer has any build-time dependency on `krb5`. Clean this up. Reviewed By: stevegury, vitaut Differential Revision: D14814205 fbshipit-source-id: dca469d22098e34573674194facaaac6c4c6aa32
Python
unknown
ReactiveSocket/reactivesocket-cpp,ReactiveSocket/reactivesocket-cpp,phoad/rsocket-cpp,phoad/rsocket-cpp,rsocket/rsocket-cpp,phoad/rsocket-cpp,rsocket/rsocket-cpp,rsocket/rsocket-cpp,ReactiveSocket/reactivesocket-cpp,rsocket/rsocket-cpp,phoad/rsocket-cpp,phoad/rsocket-cpp
#!/usr/bin/env python # Copyright (c) Facebook, Inc. and its affiliates. from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import specs.folly as folly import specs.fizz as fizz import specs.rsocket as rsocket import specs.sodium as sodium import specs.wangle as wangle import specs.zstd as zstd from shell_quoting import ShellQuoted def fbcode_builder_spec(builder): # This API should change rarely, so build the latest tag instead of master. builder.add_option( 'no1msd/mstch:git_hash', ShellQuoted('$(git describe --abbrev=0 --tags)') ) builder.add_option('krb5/krb5:git_hash', 'krb5-1.16.1-final') return { 'depends_on': [folly, fizz, sodium, rsocket, wangle, zstd], 'steps': [ # This isn't a separete spec, since only fbthrift uses mstch. builder.github_project_workdir('no1msd/mstch', 'build'), builder.cmake_install('no1msd/mstch'), builder.github_project_workdir('krb5/krb5', 'src'), builder.autoconf_install('krb5/krb5'), builder.fb_github_cmake_install('fbthrift/thrift'), ], } Cut fbcode_builder dep for thrift on krb5 Summary: [Thrift] Cut `fbcode_builder` dep for `thrift` on `krb5`. In the past, Thrift depended on Kerberos and the `krb5` implementation for its transport-layer security. However, Thrift has since migrated fully to Transport Layer Security for its transport-layer security and no longer has any build-time dependency on `krb5`. Clean this up. Reviewed By: stevegury, vitaut Differential Revision: D14814205 fbshipit-source-id: dca469d22098e34573674194facaaac6c4c6aa32
#!/usr/bin/env python # Copyright (c) Facebook, Inc. and its affiliates. from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import specs.folly as folly import specs.fizz as fizz import specs.rsocket as rsocket import specs.sodium as sodium import specs.wangle as wangle import specs.zstd as zstd from shell_quoting import ShellQuoted def fbcode_builder_spec(builder): # This API should change rarely, so build the latest tag instead of master. builder.add_option( 'no1msd/mstch:git_hash', ShellQuoted('$(git describe --abbrev=0 --tags)') ) return { 'depends_on': [folly, fizz, sodium, rsocket, wangle, zstd], 'steps': [ # This isn't a separete spec, since only fbthrift uses mstch. builder.github_project_workdir('no1msd/mstch', 'build'), builder.cmake_install('no1msd/mstch'), builder.fb_github_cmake_install('fbthrift/thrift'), ], }
<commit_before>#!/usr/bin/env python # Copyright (c) Facebook, Inc. and its affiliates. from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import specs.folly as folly import specs.fizz as fizz import specs.rsocket as rsocket import specs.sodium as sodium import specs.wangle as wangle import specs.zstd as zstd from shell_quoting import ShellQuoted def fbcode_builder_spec(builder): # This API should change rarely, so build the latest tag instead of master. builder.add_option( 'no1msd/mstch:git_hash', ShellQuoted('$(git describe --abbrev=0 --tags)') ) builder.add_option('krb5/krb5:git_hash', 'krb5-1.16.1-final') return { 'depends_on': [folly, fizz, sodium, rsocket, wangle, zstd], 'steps': [ # This isn't a separete spec, since only fbthrift uses mstch. builder.github_project_workdir('no1msd/mstch', 'build'), builder.cmake_install('no1msd/mstch'), builder.github_project_workdir('krb5/krb5', 'src'), builder.autoconf_install('krb5/krb5'), builder.fb_github_cmake_install('fbthrift/thrift'), ], } <commit_msg>Cut fbcode_builder dep for thrift on krb5 Summary: [Thrift] Cut `fbcode_builder` dep for `thrift` on `krb5`. In the past, Thrift depended on Kerberos and the `krb5` implementation for its transport-layer security. However, Thrift has since migrated fully to Transport Layer Security for its transport-layer security and no longer has any build-time dependency on `krb5`. Clean this up. Reviewed By: stevegury, vitaut Differential Revision: D14814205 fbshipit-source-id: dca469d22098e34573674194facaaac6c4c6aa32<commit_after>
#!/usr/bin/env python # Copyright (c) Facebook, Inc. and its affiliates. from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import specs.folly as folly import specs.fizz as fizz import specs.rsocket as rsocket import specs.sodium as sodium import specs.wangle as wangle import specs.zstd as zstd from shell_quoting import ShellQuoted def fbcode_builder_spec(builder): # This API should change rarely, so build the latest tag instead of master. builder.add_option( 'no1msd/mstch:git_hash', ShellQuoted('$(git describe --abbrev=0 --tags)') ) return { 'depends_on': [folly, fizz, sodium, rsocket, wangle, zstd], 'steps': [ # This isn't a separete spec, since only fbthrift uses mstch. builder.github_project_workdir('no1msd/mstch', 'build'), builder.cmake_install('no1msd/mstch'), builder.fb_github_cmake_install('fbthrift/thrift'), ], }
#!/usr/bin/env python # Copyright (c) Facebook, Inc. and its affiliates. from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import specs.folly as folly import specs.fizz as fizz import specs.rsocket as rsocket import specs.sodium as sodium import specs.wangle as wangle import specs.zstd as zstd from shell_quoting import ShellQuoted def fbcode_builder_spec(builder): # This API should change rarely, so build the latest tag instead of master. builder.add_option( 'no1msd/mstch:git_hash', ShellQuoted('$(git describe --abbrev=0 --tags)') ) builder.add_option('krb5/krb5:git_hash', 'krb5-1.16.1-final') return { 'depends_on': [folly, fizz, sodium, rsocket, wangle, zstd], 'steps': [ # This isn't a separete spec, since only fbthrift uses mstch. builder.github_project_workdir('no1msd/mstch', 'build'), builder.cmake_install('no1msd/mstch'), builder.github_project_workdir('krb5/krb5', 'src'), builder.autoconf_install('krb5/krb5'), builder.fb_github_cmake_install('fbthrift/thrift'), ], } Cut fbcode_builder dep for thrift on krb5 Summary: [Thrift] Cut `fbcode_builder` dep for `thrift` on `krb5`. In the past, Thrift depended on Kerberos and the `krb5` implementation for its transport-layer security. However, Thrift has since migrated fully to Transport Layer Security for its transport-layer security and no longer has any build-time dependency on `krb5`. Clean this up. Reviewed By: stevegury, vitaut Differential Revision: D14814205 fbshipit-source-id: dca469d22098e34573674194facaaac6c4c6aa32#!/usr/bin/env python # Copyright (c) Facebook, Inc. and its affiliates. from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import specs.folly as folly import specs.fizz as fizz import specs.rsocket as rsocket import specs.sodium as sodium import specs.wangle as wangle import specs.zstd as zstd from shell_quoting import ShellQuoted def fbcode_builder_spec(builder): # This API should change rarely, so build the latest tag instead of master. builder.add_option( 'no1msd/mstch:git_hash', ShellQuoted('$(git describe --abbrev=0 --tags)') ) return { 'depends_on': [folly, fizz, sodium, rsocket, wangle, zstd], 'steps': [ # This isn't a separete spec, since only fbthrift uses mstch. builder.github_project_workdir('no1msd/mstch', 'build'), builder.cmake_install('no1msd/mstch'), builder.fb_github_cmake_install('fbthrift/thrift'), ], }
<commit_before>#!/usr/bin/env python # Copyright (c) Facebook, Inc. and its affiliates. from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import specs.folly as folly import specs.fizz as fizz import specs.rsocket as rsocket import specs.sodium as sodium import specs.wangle as wangle import specs.zstd as zstd from shell_quoting import ShellQuoted def fbcode_builder_spec(builder): # This API should change rarely, so build the latest tag instead of master. builder.add_option( 'no1msd/mstch:git_hash', ShellQuoted('$(git describe --abbrev=0 --tags)') ) builder.add_option('krb5/krb5:git_hash', 'krb5-1.16.1-final') return { 'depends_on': [folly, fizz, sodium, rsocket, wangle, zstd], 'steps': [ # This isn't a separete spec, since only fbthrift uses mstch. builder.github_project_workdir('no1msd/mstch', 'build'), builder.cmake_install('no1msd/mstch'), builder.github_project_workdir('krb5/krb5', 'src'), builder.autoconf_install('krb5/krb5'), builder.fb_github_cmake_install('fbthrift/thrift'), ], } <commit_msg>Cut fbcode_builder dep for thrift on krb5 Summary: [Thrift] Cut `fbcode_builder` dep for `thrift` on `krb5`. In the past, Thrift depended on Kerberos and the `krb5` implementation for its transport-layer security. However, Thrift has since migrated fully to Transport Layer Security for its transport-layer security and no longer has any build-time dependency on `krb5`. Clean this up. Reviewed By: stevegury, vitaut Differential Revision: D14814205 fbshipit-source-id: dca469d22098e34573674194facaaac6c4c6aa32<commit_after>#!/usr/bin/env python # Copyright (c) Facebook, Inc. and its affiliates. from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import specs.folly as folly import specs.fizz as fizz import specs.rsocket as rsocket import specs.sodium as sodium import specs.wangle as wangle import specs.zstd as zstd from shell_quoting import ShellQuoted def fbcode_builder_spec(builder): # This API should change rarely, so build the latest tag instead of master. builder.add_option( 'no1msd/mstch:git_hash', ShellQuoted('$(git describe --abbrev=0 --tags)') ) return { 'depends_on': [folly, fizz, sodium, rsocket, wangle, zstd], 'steps': [ # This isn't a separete spec, since only fbthrift uses mstch. builder.github_project_workdir('no1msd/mstch', 'build'), builder.cmake_install('no1msd/mstch'), builder.fb_github_cmake_install('fbthrift/thrift'), ], }
6ec13485a475aeabf8a7fc461b160bbc4a453a00
windmill/server/__init__.py
windmill/server/__init__.py
# Copyright (c) 2006-2007 Open Source Applications Foundation # Copyright (c) 2008-2009 Mikeal Rogers <mikeal.rogers@gmail.com> # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import wsgi, convergence forwarding_conditions = [ lambda e : 'google.com/safebrowsing/downloads' not in e['reconstructed_url'], lambda e : 'mozilla.org/en-US/firefox/livebookmarks.html' not in e['reconstructed_url'], ] def add_forward_condition(condition): forwarding_conditions.append(condition) def remove_forward_condition(condition): while condition in forwarding_conditions: forwarding_conditions.remove(condition)
# Copyright (c) 2006-2007 Open Source Applications Foundation # Copyright (c) 2008-2009 Mikeal Rogers <mikeal.rogers@gmail.com> # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import wsgi, convergence forwarding_conditions = [ lambda e : 'google.com/safebrowsing/downloads' not in e['reconstructed_url'], lambda e : 'mozilla.org/en-US/firefox/livebookmarks.html' not in e['reconstructed_url'], lambda e : e.get('CONTENT_TYPE') != 'application/x-shockwave-flash', ] def add_forward_condition(condition): forwarding_conditions.append(condition) def remove_forward_condition(condition): while condition in forwarding_conditions: forwarding_conditions.remove(condition)
Stop forwarding flash by default, it breaks more than it doesn't.
Stop forwarding flash by default, it breaks more than it doesn't. git-svn-id: 87d19257dd11500985d055ec4730e446075a5f07@1279 78c7df6f-8922-0410-bcd3-9426b1ad491b
Python
apache-2.0
ept/windmill,ept/windmill,ept/windmill
# Copyright (c) 2006-2007 Open Source Applications Foundation # Copyright (c) 2008-2009 Mikeal Rogers <mikeal.rogers@gmail.com> # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import wsgi, convergence forwarding_conditions = [ lambda e : 'google.com/safebrowsing/downloads' not in e['reconstructed_url'], lambda e : 'mozilla.org/en-US/firefox/livebookmarks.html' not in e['reconstructed_url'], ] def add_forward_condition(condition): forwarding_conditions.append(condition) def remove_forward_condition(condition): while condition in forwarding_conditions: forwarding_conditions.remove(condition) Stop forwarding flash by default, it breaks more than it doesn't. git-svn-id: 87d19257dd11500985d055ec4730e446075a5f07@1279 78c7df6f-8922-0410-bcd3-9426b1ad491b
# Copyright (c) 2006-2007 Open Source Applications Foundation # Copyright (c) 2008-2009 Mikeal Rogers <mikeal.rogers@gmail.com> # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import wsgi, convergence forwarding_conditions = [ lambda e : 'google.com/safebrowsing/downloads' not in e['reconstructed_url'], lambda e : 'mozilla.org/en-US/firefox/livebookmarks.html' not in e['reconstructed_url'], lambda e : e.get('CONTENT_TYPE') != 'application/x-shockwave-flash', ] def add_forward_condition(condition): forwarding_conditions.append(condition) def remove_forward_condition(condition): while condition in forwarding_conditions: forwarding_conditions.remove(condition)
<commit_before># Copyright (c) 2006-2007 Open Source Applications Foundation # Copyright (c) 2008-2009 Mikeal Rogers <mikeal.rogers@gmail.com> # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import wsgi, convergence forwarding_conditions = [ lambda e : 'google.com/safebrowsing/downloads' not in e['reconstructed_url'], lambda e : 'mozilla.org/en-US/firefox/livebookmarks.html' not in e['reconstructed_url'], ] def add_forward_condition(condition): forwarding_conditions.append(condition) def remove_forward_condition(condition): while condition in forwarding_conditions: forwarding_conditions.remove(condition) <commit_msg>Stop forwarding flash by default, it breaks more than it doesn't. git-svn-id: 87d19257dd11500985d055ec4730e446075a5f07@1279 78c7df6f-8922-0410-bcd3-9426b1ad491b<commit_after>
# Copyright (c) 2006-2007 Open Source Applications Foundation # Copyright (c) 2008-2009 Mikeal Rogers <mikeal.rogers@gmail.com> # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import wsgi, convergence forwarding_conditions = [ lambda e : 'google.com/safebrowsing/downloads' not in e['reconstructed_url'], lambda e : 'mozilla.org/en-US/firefox/livebookmarks.html' not in e['reconstructed_url'], lambda e : e.get('CONTENT_TYPE') != 'application/x-shockwave-flash', ] def add_forward_condition(condition): forwarding_conditions.append(condition) def remove_forward_condition(condition): while condition in forwarding_conditions: forwarding_conditions.remove(condition)
# Copyright (c) 2006-2007 Open Source Applications Foundation # Copyright (c) 2008-2009 Mikeal Rogers <mikeal.rogers@gmail.com> # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import wsgi, convergence forwarding_conditions = [ lambda e : 'google.com/safebrowsing/downloads' not in e['reconstructed_url'], lambda e : 'mozilla.org/en-US/firefox/livebookmarks.html' not in e['reconstructed_url'], ] def add_forward_condition(condition): forwarding_conditions.append(condition) def remove_forward_condition(condition): while condition in forwarding_conditions: forwarding_conditions.remove(condition) Stop forwarding flash by default, it breaks more than it doesn't. git-svn-id: 87d19257dd11500985d055ec4730e446075a5f07@1279 78c7df6f-8922-0410-bcd3-9426b1ad491b# Copyright (c) 2006-2007 Open Source Applications Foundation # Copyright (c) 2008-2009 Mikeal Rogers <mikeal.rogers@gmail.com> # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import wsgi, convergence forwarding_conditions = [ lambda e : 'google.com/safebrowsing/downloads' not in e['reconstructed_url'], lambda e : 'mozilla.org/en-US/firefox/livebookmarks.html' not in e['reconstructed_url'], lambda e : e.get('CONTENT_TYPE') != 'application/x-shockwave-flash', ] def add_forward_condition(condition): forwarding_conditions.append(condition) def remove_forward_condition(condition): while condition in forwarding_conditions: forwarding_conditions.remove(condition)
<commit_before># Copyright (c) 2006-2007 Open Source Applications Foundation # Copyright (c) 2008-2009 Mikeal Rogers <mikeal.rogers@gmail.com> # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import wsgi, convergence forwarding_conditions = [ lambda e : 'google.com/safebrowsing/downloads' not in e['reconstructed_url'], lambda e : 'mozilla.org/en-US/firefox/livebookmarks.html' not in e['reconstructed_url'], ] def add_forward_condition(condition): forwarding_conditions.append(condition) def remove_forward_condition(condition): while condition in forwarding_conditions: forwarding_conditions.remove(condition) <commit_msg>Stop forwarding flash by default, it breaks more than it doesn't. git-svn-id: 87d19257dd11500985d055ec4730e446075a5f07@1279 78c7df6f-8922-0410-bcd3-9426b1ad491b<commit_after># Copyright (c) 2006-2007 Open Source Applications Foundation # Copyright (c) 2008-2009 Mikeal Rogers <mikeal.rogers@gmail.com> # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import wsgi, convergence forwarding_conditions = [ lambda e : 'google.com/safebrowsing/downloads' not in e['reconstructed_url'], lambda e : 'mozilla.org/en-US/firefox/livebookmarks.html' not in e['reconstructed_url'], lambda e : e.get('CONTENT_TYPE') != 'application/x-shockwave-flash', ] def add_forward_condition(condition): forwarding_conditions.append(condition) def remove_forward_condition(condition): while condition in forwarding_conditions: forwarding_conditions.remove(condition)
9b676c6a4945540a6b23333b43e75c3f539862ae
propertyfrontend/__init__.py
propertyfrontend/__init__.py
import os, logging from flask import Flask from flask.ext.basicauth import BasicAuth from raven.contrib.flask import Sentry app = Flask(__name__) app.config.from_object(os.environ.get('SETTINGS')) if app.config.get('BASIC_AUTH_USERNAME'): app.config['BASIC_AUTH_FORCE'] = True basic_auth = BasicAuth(app) # Sentry exception reporting if 'SENTRY_DSN' in os.environ: sentry = Sentry(app, dsn=os.environ['SENTRY_DSN']) if not app.debug: app.logger.addHandler(logging.StreamHandler()) app.logger.setLevel(logging.INFO) app.logger.info("\nConfiguration\n%s\n" % app.config) @app.context_processor def asset_path_context_processor(): return { 'asset_path': '/static/build/', 'landregistry_asset_path': '/static/build/' } @app.context_processor def address_processor(): from lrutils import build_address def process_address_json(address_json): return build_address(address_json) return dict(formatted=process_address_json)
import os, logging from flask import Flask from flask.ext.basicauth import BasicAuth from raven.contrib.flask import Sentry app = Flask(__name__) app.config.from_object(os.environ.get('SETTINGS')) if app.config.get('BASIC_AUTH_USERNAME'): app.config['BASIC_AUTH_FORCE'] = True basic_auth = BasicAuth(app) # Sentry exception reporting if 'SENTRY_DSN' in os.environ: sentry = Sentry(app, dsn=os.environ['SENTRY_DSN']) if not app.debug: app.logger.addHandler(logging.StreamHandler()) app.logger.setLevel(logging.INFO) app.logger.debug("\nConfiguration\n%s\n" % app.config) @app.context_processor def asset_path_context_processor(): return { 'asset_path': '/static/build/', 'landregistry_asset_path': '/static/build/' } @app.context_processor def address_processor(): from lrutils import build_address def process_address_json(address_json): return build_address(address_json) return dict(formatted=process_address_json)
Set config logging in init to debug
Set config logging in init to debug
Python
mit
LandRegistry/property-frontend-alpha,LandRegistry/property-frontend-alpha,LandRegistry/property-frontend-alpha,LandRegistry/property-frontend-alpha
import os, logging from flask import Flask from flask.ext.basicauth import BasicAuth from raven.contrib.flask import Sentry app = Flask(__name__) app.config.from_object(os.environ.get('SETTINGS')) if app.config.get('BASIC_AUTH_USERNAME'): app.config['BASIC_AUTH_FORCE'] = True basic_auth = BasicAuth(app) # Sentry exception reporting if 'SENTRY_DSN' in os.environ: sentry = Sentry(app, dsn=os.environ['SENTRY_DSN']) if not app.debug: app.logger.addHandler(logging.StreamHandler()) app.logger.setLevel(logging.INFO) app.logger.info("\nConfiguration\n%s\n" % app.config) @app.context_processor def asset_path_context_processor(): return { 'asset_path': '/static/build/', 'landregistry_asset_path': '/static/build/' } @app.context_processor def address_processor(): from lrutils import build_address def process_address_json(address_json): return build_address(address_json) return dict(formatted=process_address_json) Set config logging in init to debug
import os, logging from flask import Flask from flask.ext.basicauth import BasicAuth from raven.contrib.flask import Sentry app = Flask(__name__) app.config.from_object(os.environ.get('SETTINGS')) if app.config.get('BASIC_AUTH_USERNAME'): app.config['BASIC_AUTH_FORCE'] = True basic_auth = BasicAuth(app) # Sentry exception reporting if 'SENTRY_DSN' in os.environ: sentry = Sentry(app, dsn=os.environ['SENTRY_DSN']) if not app.debug: app.logger.addHandler(logging.StreamHandler()) app.logger.setLevel(logging.INFO) app.logger.debug("\nConfiguration\n%s\n" % app.config) @app.context_processor def asset_path_context_processor(): return { 'asset_path': '/static/build/', 'landregistry_asset_path': '/static/build/' } @app.context_processor def address_processor(): from lrutils import build_address def process_address_json(address_json): return build_address(address_json) return dict(formatted=process_address_json)
<commit_before>import os, logging from flask import Flask from flask.ext.basicauth import BasicAuth from raven.contrib.flask import Sentry app = Flask(__name__) app.config.from_object(os.environ.get('SETTINGS')) if app.config.get('BASIC_AUTH_USERNAME'): app.config['BASIC_AUTH_FORCE'] = True basic_auth = BasicAuth(app) # Sentry exception reporting if 'SENTRY_DSN' in os.environ: sentry = Sentry(app, dsn=os.environ['SENTRY_DSN']) if not app.debug: app.logger.addHandler(logging.StreamHandler()) app.logger.setLevel(logging.INFO) app.logger.info("\nConfiguration\n%s\n" % app.config) @app.context_processor def asset_path_context_processor(): return { 'asset_path': '/static/build/', 'landregistry_asset_path': '/static/build/' } @app.context_processor def address_processor(): from lrutils import build_address def process_address_json(address_json): return build_address(address_json) return dict(formatted=process_address_json) <commit_msg>Set config logging in init to debug<commit_after>
import os, logging from flask import Flask from flask.ext.basicauth import BasicAuth from raven.contrib.flask import Sentry app = Flask(__name__) app.config.from_object(os.environ.get('SETTINGS')) if app.config.get('BASIC_AUTH_USERNAME'): app.config['BASIC_AUTH_FORCE'] = True basic_auth = BasicAuth(app) # Sentry exception reporting if 'SENTRY_DSN' in os.environ: sentry = Sentry(app, dsn=os.environ['SENTRY_DSN']) if not app.debug: app.logger.addHandler(logging.StreamHandler()) app.logger.setLevel(logging.INFO) app.logger.debug("\nConfiguration\n%s\n" % app.config) @app.context_processor def asset_path_context_processor(): return { 'asset_path': '/static/build/', 'landregistry_asset_path': '/static/build/' } @app.context_processor def address_processor(): from lrutils import build_address def process_address_json(address_json): return build_address(address_json) return dict(formatted=process_address_json)
import os, logging from flask import Flask from flask.ext.basicauth import BasicAuth from raven.contrib.flask import Sentry app = Flask(__name__) app.config.from_object(os.environ.get('SETTINGS')) if app.config.get('BASIC_AUTH_USERNAME'): app.config['BASIC_AUTH_FORCE'] = True basic_auth = BasicAuth(app) # Sentry exception reporting if 'SENTRY_DSN' in os.environ: sentry = Sentry(app, dsn=os.environ['SENTRY_DSN']) if not app.debug: app.logger.addHandler(logging.StreamHandler()) app.logger.setLevel(logging.INFO) app.logger.info("\nConfiguration\n%s\n" % app.config) @app.context_processor def asset_path_context_processor(): return { 'asset_path': '/static/build/', 'landregistry_asset_path': '/static/build/' } @app.context_processor def address_processor(): from lrutils import build_address def process_address_json(address_json): return build_address(address_json) return dict(formatted=process_address_json) Set config logging in init to debugimport os, logging from flask import Flask from flask.ext.basicauth import BasicAuth from raven.contrib.flask import Sentry app = Flask(__name__) app.config.from_object(os.environ.get('SETTINGS')) if app.config.get('BASIC_AUTH_USERNAME'): app.config['BASIC_AUTH_FORCE'] = True basic_auth = BasicAuth(app) # Sentry exception reporting if 'SENTRY_DSN' in os.environ: sentry = Sentry(app, dsn=os.environ['SENTRY_DSN']) if not app.debug: app.logger.addHandler(logging.StreamHandler()) app.logger.setLevel(logging.INFO) app.logger.debug("\nConfiguration\n%s\n" % app.config) @app.context_processor def asset_path_context_processor(): return { 'asset_path': '/static/build/', 'landregistry_asset_path': '/static/build/' } @app.context_processor def address_processor(): from lrutils import build_address def process_address_json(address_json): return build_address(address_json) return dict(formatted=process_address_json)
<commit_before>import os, logging from flask import Flask from flask.ext.basicauth import BasicAuth from raven.contrib.flask import Sentry app = Flask(__name__) app.config.from_object(os.environ.get('SETTINGS')) if app.config.get('BASIC_AUTH_USERNAME'): app.config['BASIC_AUTH_FORCE'] = True basic_auth = BasicAuth(app) # Sentry exception reporting if 'SENTRY_DSN' in os.environ: sentry = Sentry(app, dsn=os.environ['SENTRY_DSN']) if not app.debug: app.logger.addHandler(logging.StreamHandler()) app.logger.setLevel(logging.INFO) app.logger.info("\nConfiguration\n%s\n" % app.config) @app.context_processor def asset_path_context_processor(): return { 'asset_path': '/static/build/', 'landregistry_asset_path': '/static/build/' } @app.context_processor def address_processor(): from lrutils import build_address def process_address_json(address_json): return build_address(address_json) return dict(formatted=process_address_json) <commit_msg>Set config logging in init to debug<commit_after>import os, logging from flask import Flask from flask.ext.basicauth import BasicAuth from raven.contrib.flask import Sentry app = Flask(__name__) app.config.from_object(os.environ.get('SETTINGS')) if app.config.get('BASIC_AUTH_USERNAME'): app.config['BASIC_AUTH_FORCE'] = True basic_auth = BasicAuth(app) # Sentry exception reporting if 'SENTRY_DSN' in os.environ: sentry = Sentry(app, dsn=os.environ['SENTRY_DSN']) if not app.debug: app.logger.addHandler(logging.StreamHandler()) app.logger.setLevel(logging.INFO) app.logger.debug("\nConfiguration\n%s\n" % app.config) @app.context_processor def asset_path_context_processor(): return { 'asset_path': '/static/build/', 'landregistry_asset_path': '/static/build/' } @app.context_processor def address_processor(): from lrutils import build_address def process_address_json(address_json): return build_address(address_json) return dict(formatted=process_address_json)
c57910adc6e907881a99e092837fc35e5f45518b
survey_creation/config/de_17.py
survey_creation/config/de_17.py
""" Config file specific to uk to create automated survey """ class config: # To modify, just add the keys of the dictionary header_to_modify = [{'class': 'S', 'name': 'sid', 'text': '421498'}, {'class': 'S', 'name': 'admin_email', 'text': 'olivier.philippe@soton.ac.uk'}] # Same as header_to_modify description_to_modify = [] # Add header and description. Tuple of a dictionary + the position where it is supposed # to be inserted header_to_add = [] # The index positions starts at 0 description_to_add = [({'class': 'SL', 'name': 'additional_languages', 'text': 'nl'}, 12)]
""" Config file specific to uk to create automated survey """ class config: # To modify, just add the keys of the dictionary header_to_modify = [{'class': 'S', 'name': 'sid', 'text': '421498'}, {'class': 'S', 'name': 'admin_email', 'text': 'olivier.philippe@soton.ac.uk'}] # Same as header_to_modify description_to_modify = [] # Add header and description. Tuple of a dictionary + the position where it is supposed # to be inserted header_to_add = [({'class': 'S', 'name': 'additional_languages', 'text': 'nl'}, 12)] # The index positions starts at 0 description_to_add = []
Fix issue with headers about additional language in description rather than header
Fix issue with headers about additional language in description rather than header
Python
bsd-3-clause
softwaresaved/international-survey
""" Config file specific to uk to create automated survey """ class config: # To modify, just add the keys of the dictionary header_to_modify = [{'class': 'S', 'name': 'sid', 'text': '421498'}, {'class': 'S', 'name': 'admin_email', 'text': 'olivier.philippe@soton.ac.uk'}] # Same as header_to_modify description_to_modify = [] # Add header and description. Tuple of a dictionary + the position where it is supposed # to be inserted header_to_add = [] # The index positions starts at 0 description_to_add = [({'class': 'SL', 'name': 'additional_languages', 'text': 'nl'}, 12)] Fix issue with headers about additional language in description rather than header
""" Config file specific to uk to create automated survey """ class config: # To modify, just add the keys of the dictionary header_to_modify = [{'class': 'S', 'name': 'sid', 'text': '421498'}, {'class': 'S', 'name': 'admin_email', 'text': 'olivier.philippe@soton.ac.uk'}] # Same as header_to_modify description_to_modify = [] # Add header and description. Tuple of a dictionary + the position where it is supposed # to be inserted header_to_add = [({'class': 'S', 'name': 'additional_languages', 'text': 'nl'}, 12)] # The index positions starts at 0 description_to_add = []
<commit_before>""" Config file specific to uk to create automated survey """ class config: # To modify, just add the keys of the dictionary header_to_modify = [{'class': 'S', 'name': 'sid', 'text': '421498'}, {'class': 'S', 'name': 'admin_email', 'text': 'olivier.philippe@soton.ac.uk'}] # Same as header_to_modify description_to_modify = [] # Add header and description. Tuple of a dictionary + the position where it is supposed # to be inserted header_to_add = [] # The index positions starts at 0 description_to_add = [({'class': 'SL', 'name': 'additional_languages', 'text': 'nl'}, 12)] <commit_msg>Fix issue with headers about additional language in description rather than header<commit_after>
""" Config file specific to uk to create automated survey """ class config: # To modify, just add the keys of the dictionary header_to_modify = [{'class': 'S', 'name': 'sid', 'text': '421498'}, {'class': 'S', 'name': 'admin_email', 'text': 'olivier.philippe@soton.ac.uk'}] # Same as header_to_modify description_to_modify = [] # Add header and description. Tuple of a dictionary + the position where it is supposed # to be inserted header_to_add = [({'class': 'S', 'name': 'additional_languages', 'text': 'nl'}, 12)] # The index positions starts at 0 description_to_add = []
""" Config file specific to uk to create automated survey """ class config: # To modify, just add the keys of the dictionary header_to_modify = [{'class': 'S', 'name': 'sid', 'text': '421498'}, {'class': 'S', 'name': 'admin_email', 'text': 'olivier.philippe@soton.ac.uk'}] # Same as header_to_modify description_to_modify = [] # Add header and description. Tuple of a dictionary + the position where it is supposed # to be inserted header_to_add = [] # The index positions starts at 0 description_to_add = [({'class': 'SL', 'name': 'additional_languages', 'text': 'nl'}, 12)] Fix issue with headers about additional language in description rather than header""" Config file specific to uk to create automated survey """ class config: # To modify, just add the keys of the dictionary header_to_modify = [{'class': 'S', 'name': 'sid', 'text': '421498'}, {'class': 'S', 'name': 'admin_email', 'text': 'olivier.philippe@soton.ac.uk'}] # Same as header_to_modify description_to_modify = [] # Add header and description. Tuple of a dictionary + the position where it is supposed # to be inserted header_to_add = [({'class': 'S', 'name': 'additional_languages', 'text': 'nl'}, 12)] # The index positions starts at 0 description_to_add = []
<commit_before>""" Config file specific to uk to create automated survey """ class config: # To modify, just add the keys of the dictionary header_to_modify = [{'class': 'S', 'name': 'sid', 'text': '421498'}, {'class': 'S', 'name': 'admin_email', 'text': 'olivier.philippe@soton.ac.uk'}] # Same as header_to_modify description_to_modify = [] # Add header and description. Tuple of a dictionary + the position where it is supposed # to be inserted header_to_add = [] # The index positions starts at 0 description_to_add = [({'class': 'SL', 'name': 'additional_languages', 'text': 'nl'}, 12)] <commit_msg>Fix issue with headers about additional language in description rather than header<commit_after>""" Config file specific to uk to create automated survey """ class config: # To modify, just add the keys of the dictionary header_to_modify = [{'class': 'S', 'name': 'sid', 'text': '421498'}, {'class': 'S', 'name': 'admin_email', 'text': 'olivier.philippe@soton.ac.uk'}] # Same as header_to_modify description_to_modify = [] # Add header and description. Tuple of a dictionary + the position where it is supposed # to be inserted header_to_add = [({'class': 'S', 'name': 'additional_languages', 'text': 'nl'}, 12)] # The index positions starts at 0 description_to_add = []
4de89e1d1cf258e903b469deff9d2a7df34a1db9
dotfiles/.ipython/profile_default/startup/bytes.py
dotfiles/.ipython/profile_default/startup/bytes.py
def bytes(b): """ Print bytes in a humanized way """ def humanize(b, base, suffices=[]): bb = int(b) for suffix in suffices: if bb < base: break bb /= float(base) return "%.2f %s" % (bb, suffix) print "Base 1024: ", humanize( b, 1024, ['B', 'KiB', 'MiB', 'GiB', 'TiB', 'PiB', 'EiB']) print "Base 1000: ", humanize( b, 1000, ['B', 'KB', 'MB', 'GB', 'TB', 'PB', 'EB'])
def bytes(b): """ Print bytes in a humanized way """ def humanize(b, base, suffices=[]): bb = int(b) for suffix in suffices: if bb < base: break bb /= float(base) return "%.2f %s" % (bb, suffix) print("Base 1024: ", humanize( b, 1024, ['B', 'KiB', 'MiB', 'GiB', 'TiB', 'PiB', 'EiB'])) print("Base 1000: ", humanize( b, 1000, ['B', 'KB', 'MB', 'GB', 'TB', 'PB', 'EB']))
Make ipython profile python3 compliant
Make ipython profile python3 compliant
Python
mit
izidormatusov/dotfiles,izidormatusov/dotfiles
def bytes(b): """ Print bytes in a humanized way """ def humanize(b, base, suffices=[]): bb = int(b) for suffix in suffices: if bb < base: break bb /= float(base) return "%.2f %s" % (bb, suffix) print "Base 1024: ", humanize( b, 1024, ['B', 'KiB', 'MiB', 'GiB', 'TiB', 'PiB', 'EiB']) print "Base 1000: ", humanize( b, 1000, ['B', 'KB', 'MB', 'GB', 'TB', 'PB', 'EB']) Make ipython profile python3 compliant
def bytes(b): """ Print bytes in a humanized way """ def humanize(b, base, suffices=[]): bb = int(b) for suffix in suffices: if bb < base: break bb /= float(base) return "%.2f %s" % (bb, suffix) print("Base 1024: ", humanize( b, 1024, ['B', 'KiB', 'MiB', 'GiB', 'TiB', 'PiB', 'EiB'])) print("Base 1000: ", humanize( b, 1000, ['B', 'KB', 'MB', 'GB', 'TB', 'PB', 'EB']))
<commit_before>def bytes(b): """ Print bytes in a humanized way """ def humanize(b, base, suffices=[]): bb = int(b) for suffix in suffices: if bb < base: break bb /= float(base) return "%.2f %s" % (bb, suffix) print "Base 1024: ", humanize( b, 1024, ['B', 'KiB', 'MiB', 'GiB', 'TiB', 'PiB', 'EiB']) print "Base 1000: ", humanize( b, 1000, ['B', 'KB', 'MB', 'GB', 'TB', 'PB', 'EB']) <commit_msg>Make ipython profile python3 compliant<commit_after>
def bytes(b): """ Print bytes in a humanized way """ def humanize(b, base, suffices=[]): bb = int(b) for suffix in suffices: if bb < base: break bb /= float(base) return "%.2f %s" % (bb, suffix) print("Base 1024: ", humanize( b, 1024, ['B', 'KiB', 'MiB', 'GiB', 'TiB', 'PiB', 'EiB'])) print("Base 1000: ", humanize( b, 1000, ['B', 'KB', 'MB', 'GB', 'TB', 'PB', 'EB']))
def bytes(b): """ Print bytes in a humanized way """ def humanize(b, base, suffices=[]): bb = int(b) for suffix in suffices: if bb < base: break bb /= float(base) return "%.2f %s" % (bb, suffix) print "Base 1024: ", humanize( b, 1024, ['B', 'KiB', 'MiB', 'GiB', 'TiB', 'PiB', 'EiB']) print "Base 1000: ", humanize( b, 1000, ['B', 'KB', 'MB', 'GB', 'TB', 'PB', 'EB']) Make ipython profile python3 compliantdef bytes(b): """ Print bytes in a humanized way """ def humanize(b, base, suffices=[]): bb = int(b) for suffix in suffices: if bb < base: break bb /= float(base) return "%.2f %s" % (bb, suffix) print("Base 1024: ", humanize( b, 1024, ['B', 'KiB', 'MiB', 'GiB', 'TiB', 'PiB', 'EiB'])) print("Base 1000: ", humanize( b, 1000, ['B', 'KB', 'MB', 'GB', 'TB', 'PB', 'EB']))
<commit_before>def bytes(b): """ Print bytes in a humanized way """ def humanize(b, base, suffices=[]): bb = int(b) for suffix in suffices: if bb < base: break bb /= float(base) return "%.2f %s" % (bb, suffix) print "Base 1024: ", humanize( b, 1024, ['B', 'KiB', 'MiB', 'GiB', 'TiB', 'PiB', 'EiB']) print "Base 1000: ", humanize( b, 1000, ['B', 'KB', 'MB', 'GB', 'TB', 'PB', 'EB']) <commit_msg>Make ipython profile python3 compliant<commit_after>def bytes(b): """ Print bytes in a humanized way """ def humanize(b, base, suffices=[]): bb = int(b) for suffix in suffices: if bb < base: break bb /= float(base) return "%.2f %s" % (bb, suffix) print("Base 1024: ", humanize( b, 1024, ['B', 'KiB', 'MiB', 'GiB', 'TiB', 'PiB', 'EiB'])) print("Base 1000: ", humanize( b, 1000, ['B', 'KB', 'MB', 'GB', 'TB', 'PB', 'EB']))
bed671bdd7dc221e55b5f60c4f9daca3c338a737
artists/views.py
artists/views.py
from django.shortcuts import get_object_or_404 from rest_framework import permissions, viewsets from similarities.utils import get_similar from .models import Artist from similarities.models import UserSimilarity from .serializers import ArtistSerializer, SimilaritySerializer class ArtistViewSet(viewsets.ModelViewSet): """API endpoint that allows artists to be viewed or edited""" queryset = Artist.objects.all() serializer_class = ArtistSerializer permission_classes = (permissions.IsAuthenticatedOrReadOnly,) def get_queryset(self): name = self.request.GET.get('name', "") if name: qs = get_similar(name) else: qs = super().get_queryset() return qs[:100] class SimilarViewSet(viewsets.ModelViewSet): queryset = UserSimilarity.objects.all() serializer_class = SimilaritySerializer permission_classes = (permissions.IsAuthenticated,) http_method_names = ['get', 'post', 'put', 'delete'] filter_fields = ['cc_artist'] def get_queryset(self): return super().get_queryset().filter(user=self.request.user) def pre_save(self, obj): obj.user = self.request.user
from django.shortcuts import get_object_or_404 from rest_framework import permissions, viewsets from similarities.utils import get_similar from .models import Artist from similarities.models import UserSimilarity, Similarity, update_similarities from .serializers import ArtistSerializer, SimilaritySerializer class ArtistViewSet(viewsets.ModelViewSet): """API endpoint that allows artists to be viewed or edited""" queryset = Artist.objects.all() serializer_class = ArtistSerializer permission_classes = (permissions.IsAuthenticatedOrReadOnly,) def get_queryset(self): name = self.request.GET.get('name', "") if name: qs = get_similar(name) else: qs = super().get_queryset() return qs[:100] class SimilarViewSet(viewsets.ModelViewSet): queryset = UserSimilarity.objects.all() serializer_class = SimilaritySerializer permission_classes = (permissions.IsAuthenticated,) http_method_names = ['get', 'post', 'put', 'delete'] filter_fields = ['cc_artist'] def get_queryset(self): return super().get_queryset().filter(user=self.request.user) def pre_save(self, obj): obj.user = self.request.user def post_save(self, obj, created=False): cumulative_similarity, _ = Similarity.objects.get_or_create( other_artist=obj.other_artist, cc_artist=obj.cc_artist, ) update_similarities([cumulative_similarity])
Update cumulative similarities on save
Update cumulative similarities on save
Python
bsd-3-clause
FreeMusicNinja/api.freemusic.ninja
from django.shortcuts import get_object_or_404 from rest_framework import permissions, viewsets from similarities.utils import get_similar from .models import Artist from similarities.models import UserSimilarity from .serializers import ArtistSerializer, SimilaritySerializer class ArtistViewSet(viewsets.ModelViewSet): """API endpoint that allows artists to be viewed or edited""" queryset = Artist.objects.all() serializer_class = ArtistSerializer permission_classes = (permissions.IsAuthenticatedOrReadOnly,) def get_queryset(self): name = self.request.GET.get('name', "") if name: qs = get_similar(name) else: qs = super().get_queryset() return qs[:100] class SimilarViewSet(viewsets.ModelViewSet): queryset = UserSimilarity.objects.all() serializer_class = SimilaritySerializer permission_classes = (permissions.IsAuthenticated,) http_method_names = ['get', 'post', 'put', 'delete'] filter_fields = ['cc_artist'] def get_queryset(self): return super().get_queryset().filter(user=self.request.user) def pre_save(self, obj): obj.user = self.request.user Update cumulative similarities on save
from django.shortcuts import get_object_or_404 from rest_framework import permissions, viewsets from similarities.utils import get_similar from .models import Artist from similarities.models import UserSimilarity, Similarity, update_similarities from .serializers import ArtistSerializer, SimilaritySerializer class ArtistViewSet(viewsets.ModelViewSet): """API endpoint that allows artists to be viewed or edited""" queryset = Artist.objects.all() serializer_class = ArtistSerializer permission_classes = (permissions.IsAuthenticatedOrReadOnly,) def get_queryset(self): name = self.request.GET.get('name', "") if name: qs = get_similar(name) else: qs = super().get_queryset() return qs[:100] class SimilarViewSet(viewsets.ModelViewSet): queryset = UserSimilarity.objects.all() serializer_class = SimilaritySerializer permission_classes = (permissions.IsAuthenticated,) http_method_names = ['get', 'post', 'put', 'delete'] filter_fields = ['cc_artist'] def get_queryset(self): return super().get_queryset().filter(user=self.request.user) def pre_save(self, obj): obj.user = self.request.user def post_save(self, obj, created=False): cumulative_similarity, _ = Similarity.objects.get_or_create( other_artist=obj.other_artist, cc_artist=obj.cc_artist, ) update_similarities([cumulative_similarity])
<commit_before>from django.shortcuts import get_object_or_404 from rest_framework import permissions, viewsets from similarities.utils import get_similar from .models import Artist from similarities.models import UserSimilarity from .serializers import ArtistSerializer, SimilaritySerializer class ArtistViewSet(viewsets.ModelViewSet): """API endpoint that allows artists to be viewed or edited""" queryset = Artist.objects.all() serializer_class = ArtistSerializer permission_classes = (permissions.IsAuthenticatedOrReadOnly,) def get_queryset(self): name = self.request.GET.get('name', "") if name: qs = get_similar(name) else: qs = super().get_queryset() return qs[:100] class SimilarViewSet(viewsets.ModelViewSet): queryset = UserSimilarity.objects.all() serializer_class = SimilaritySerializer permission_classes = (permissions.IsAuthenticated,) http_method_names = ['get', 'post', 'put', 'delete'] filter_fields = ['cc_artist'] def get_queryset(self): return super().get_queryset().filter(user=self.request.user) def pre_save(self, obj): obj.user = self.request.user <commit_msg>Update cumulative similarities on save<commit_after>
from django.shortcuts import get_object_or_404 from rest_framework import permissions, viewsets from similarities.utils import get_similar from .models import Artist from similarities.models import UserSimilarity, Similarity, update_similarities from .serializers import ArtistSerializer, SimilaritySerializer class ArtistViewSet(viewsets.ModelViewSet): """API endpoint that allows artists to be viewed or edited""" queryset = Artist.objects.all() serializer_class = ArtistSerializer permission_classes = (permissions.IsAuthenticatedOrReadOnly,) def get_queryset(self): name = self.request.GET.get('name', "") if name: qs = get_similar(name) else: qs = super().get_queryset() return qs[:100] class SimilarViewSet(viewsets.ModelViewSet): queryset = UserSimilarity.objects.all() serializer_class = SimilaritySerializer permission_classes = (permissions.IsAuthenticated,) http_method_names = ['get', 'post', 'put', 'delete'] filter_fields = ['cc_artist'] def get_queryset(self): return super().get_queryset().filter(user=self.request.user) def pre_save(self, obj): obj.user = self.request.user def post_save(self, obj, created=False): cumulative_similarity, _ = Similarity.objects.get_or_create( other_artist=obj.other_artist, cc_artist=obj.cc_artist, ) update_similarities([cumulative_similarity])
from django.shortcuts import get_object_or_404 from rest_framework import permissions, viewsets from similarities.utils import get_similar from .models import Artist from similarities.models import UserSimilarity from .serializers import ArtistSerializer, SimilaritySerializer class ArtistViewSet(viewsets.ModelViewSet): """API endpoint that allows artists to be viewed or edited""" queryset = Artist.objects.all() serializer_class = ArtistSerializer permission_classes = (permissions.IsAuthenticatedOrReadOnly,) def get_queryset(self): name = self.request.GET.get('name', "") if name: qs = get_similar(name) else: qs = super().get_queryset() return qs[:100] class SimilarViewSet(viewsets.ModelViewSet): queryset = UserSimilarity.objects.all() serializer_class = SimilaritySerializer permission_classes = (permissions.IsAuthenticated,) http_method_names = ['get', 'post', 'put', 'delete'] filter_fields = ['cc_artist'] def get_queryset(self): return super().get_queryset().filter(user=self.request.user) def pre_save(self, obj): obj.user = self.request.user Update cumulative similarities on savefrom django.shortcuts import get_object_or_404 from rest_framework import permissions, viewsets from similarities.utils import get_similar from .models import Artist from similarities.models import UserSimilarity, Similarity, update_similarities from .serializers import ArtistSerializer, SimilaritySerializer class ArtistViewSet(viewsets.ModelViewSet): """API endpoint that allows artists to be viewed or edited""" queryset = Artist.objects.all() serializer_class = ArtistSerializer permission_classes = (permissions.IsAuthenticatedOrReadOnly,) def get_queryset(self): name = self.request.GET.get('name', "") if name: qs = get_similar(name) else: qs = super().get_queryset() return qs[:100] class SimilarViewSet(viewsets.ModelViewSet): queryset = UserSimilarity.objects.all() serializer_class = SimilaritySerializer permission_classes = (permissions.IsAuthenticated,) http_method_names = ['get', 'post', 'put', 'delete'] filter_fields = ['cc_artist'] def get_queryset(self): return super().get_queryset().filter(user=self.request.user) def pre_save(self, obj): obj.user = self.request.user def post_save(self, obj, created=False): cumulative_similarity, _ = Similarity.objects.get_or_create( other_artist=obj.other_artist, cc_artist=obj.cc_artist, ) update_similarities([cumulative_similarity])
<commit_before>from django.shortcuts import get_object_or_404 from rest_framework import permissions, viewsets from similarities.utils import get_similar from .models import Artist from similarities.models import UserSimilarity from .serializers import ArtistSerializer, SimilaritySerializer class ArtistViewSet(viewsets.ModelViewSet): """API endpoint that allows artists to be viewed or edited""" queryset = Artist.objects.all() serializer_class = ArtistSerializer permission_classes = (permissions.IsAuthenticatedOrReadOnly,) def get_queryset(self): name = self.request.GET.get('name', "") if name: qs = get_similar(name) else: qs = super().get_queryset() return qs[:100] class SimilarViewSet(viewsets.ModelViewSet): queryset = UserSimilarity.objects.all() serializer_class = SimilaritySerializer permission_classes = (permissions.IsAuthenticated,) http_method_names = ['get', 'post', 'put', 'delete'] filter_fields = ['cc_artist'] def get_queryset(self): return super().get_queryset().filter(user=self.request.user) def pre_save(self, obj): obj.user = self.request.user <commit_msg>Update cumulative similarities on save<commit_after>from django.shortcuts import get_object_or_404 from rest_framework import permissions, viewsets from similarities.utils import get_similar from .models import Artist from similarities.models import UserSimilarity, Similarity, update_similarities from .serializers import ArtistSerializer, SimilaritySerializer class ArtistViewSet(viewsets.ModelViewSet): """API endpoint that allows artists to be viewed or edited""" queryset = Artist.objects.all() serializer_class = ArtistSerializer permission_classes = (permissions.IsAuthenticatedOrReadOnly,) def get_queryset(self): name = self.request.GET.get('name', "") if name: qs = get_similar(name) else: qs = super().get_queryset() return qs[:100] class SimilarViewSet(viewsets.ModelViewSet): queryset = UserSimilarity.objects.all() serializer_class = SimilaritySerializer permission_classes = (permissions.IsAuthenticated,) http_method_names = ['get', 'post', 'put', 'delete'] filter_fields = ['cc_artist'] def get_queryset(self): return super().get_queryset().filter(user=self.request.user) def pre_save(self, obj): obj.user = self.request.user def post_save(self, obj, created=False): cumulative_similarity, _ = Similarity.objects.get_or_create( other_artist=obj.other_artist, cc_artist=obj.cc_artist, ) update_similarities([cumulative_similarity])
42560625d8f83a60320e111503521a9a17d8ae09
mollie/api/objects/list.py
mollie/api/objects/list.py
from .base import Base class List(Base): def __init__(self, result, object_type): Base.__init__(self, result) self.object_type = object_type def get_resource_name(self): return self.object_type.__name__.lower() + 's' def __iter__(self): for item in self['_embedded'][self.get_resource_name()]: yield self.object_type(item) @property def count(self): if 'count' not in self: return None return int(self['count']) def get_offset(self): if 'offset' not in self: return None return self['offset']
from .base import Base class List(Base): def __init__(self, result, object_type): Base.__init__(self, result) self.object_type = object_type def get_object_name(self): return self.object_type.__name__.lower() + 's' def __iter__(self): for item in self['_embedded'][self.get_object_name()]: yield self.object_type(item) @property def count(self): if 'count' not in self: return None return int(self['count']) def get_offset(self): if 'offset' not in self: return None return self['offset']
Rename method to be more logical
Rename method to be more logical
Python
bsd-2-clause
mollie/mollie-api-python
from .base import Base class List(Base): def __init__(self, result, object_type): Base.__init__(self, result) self.object_type = object_type def get_resource_name(self): return self.object_type.__name__.lower() + 's' def __iter__(self): for item in self['_embedded'][self.get_resource_name()]: yield self.object_type(item) @property def count(self): if 'count' not in self: return None return int(self['count']) def get_offset(self): if 'offset' not in self: return None return self['offset'] Rename method to be more logical
from .base import Base class List(Base): def __init__(self, result, object_type): Base.__init__(self, result) self.object_type = object_type def get_object_name(self): return self.object_type.__name__.lower() + 's' def __iter__(self): for item in self['_embedded'][self.get_object_name()]: yield self.object_type(item) @property def count(self): if 'count' not in self: return None return int(self['count']) def get_offset(self): if 'offset' not in self: return None return self['offset']
<commit_before>from .base import Base class List(Base): def __init__(self, result, object_type): Base.__init__(self, result) self.object_type = object_type def get_resource_name(self): return self.object_type.__name__.lower() + 's' def __iter__(self): for item in self['_embedded'][self.get_resource_name()]: yield self.object_type(item) @property def count(self): if 'count' not in self: return None return int(self['count']) def get_offset(self): if 'offset' not in self: return None return self['offset'] <commit_msg>Rename method to be more logical<commit_after>
from .base import Base class List(Base): def __init__(self, result, object_type): Base.__init__(self, result) self.object_type = object_type def get_object_name(self): return self.object_type.__name__.lower() + 's' def __iter__(self): for item in self['_embedded'][self.get_object_name()]: yield self.object_type(item) @property def count(self): if 'count' not in self: return None return int(self['count']) def get_offset(self): if 'offset' not in self: return None return self['offset']
from .base import Base class List(Base): def __init__(self, result, object_type): Base.__init__(self, result) self.object_type = object_type def get_resource_name(self): return self.object_type.__name__.lower() + 's' def __iter__(self): for item in self['_embedded'][self.get_resource_name()]: yield self.object_type(item) @property def count(self): if 'count' not in self: return None return int(self['count']) def get_offset(self): if 'offset' not in self: return None return self['offset'] Rename method to be more logicalfrom .base import Base class List(Base): def __init__(self, result, object_type): Base.__init__(self, result) self.object_type = object_type def get_object_name(self): return self.object_type.__name__.lower() + 's' def __iter__(self): for item in self['_embedded'][self.get_object_name()]: yield self.object_type(item) @property def count(self): if 'count' not in self: return None return int(self['count']) def get_offset(self): if 'offset' not in self: return None return self['offset']
<commit_before>from .base import Base class List(Base): def __init__(self, result, object_type): Base.__init__(self, result) self.object_type = object_type def get_resource_name(self): return self.object_type.__name__.lower() + 's' def __iter__(self): for item in self['_embedded'][self.get_resource_name()]: yield self.object_type(item) @property def count(self): if 'count' not in self: return None return int(self['count']) def get_offset(self): if 'offset' not in self: return None return self['offset'] <commit_msg>Rename method to be more logical<commit_after>from .base import Base class List(Base): def __init__(self, result, object_type): Base.__init__(self, result) self.object_type = object_type def get_object_name(self): return self.object_type.__name__.lower() + 's' def __iter__(self): for item in self['_embedded'][self.get_object_name()]: yield self.object_type(item) @property def count(self): if 'count' not in self: return None return int(self['count']) def get_offset(self): if 'offset' not in self: return None return self['offset']
616d92fed79bbfe6ea70ed7e053622819d99088d
python/getmonotime.py
python/getmonotime.py
import getopt, sys if __name__ == '__main__': sippy_path = None try: opts, args = getopt.getopt(sys.argv[1:], 's:S:i:o:b') except getopt.GetoptError: usage() for o, a in opts: if o == '-S': sippy_path = a.strip() continue if sippy_path != None: sys.path.insert(0, sippy_path) from sippy.Time.clock_dtime import clock_getdtime, CLOCK_MONOTONIC print clock_getdtime(CLOCK_MONOTONIC)
import getopt, sys if __name__ == '__main__': sippy_path = None try: opts, args = getopt.getopt(sys.argv[1:], 'rS:') except getopt.GetoptError: usage() out_realtime = False for o, a in opts: if o == '-S': sippy_path = a.strip() continue if o == '-r': out_realtime = True if sippy_path != None: sys.path.insert(0, sippy_path) from sippy.Time.clock_dtime import clock_getdtime, CLOCK_MONOTONIC if not out_realtime: print(clock_getdtime(CLOCK_MONOTONIC)) else: from sippy.Time.clock_dtime import CLOCK_REALTIME print("%f %f" % (clock_getdtime(CLOCK_MONOTONIC), clock_getdtime(CLOCK_REALTIME)))
Add an option to also output realtime along with monotime.
Add an option to also output realtime along with monotime.
Python
bsd-2-clause
sippy/rtp_cluster,sippy/rtp_cluster
import getopt, sys if __name__ == '__main__': sippy_path = None try: opts, args = getopt.getopt(sys.argv[1:], 's:S:i:o:b') except getopt.GetoptError: usage() for o, a in opts: if o == '-S': sippy_path = a.strip() continue if sippy_path != None: sys.path.insert(0, sippy_path) from sippy.Time.clock_dtime import clock_getdtime, CLOCK_MONOTONIC print clock_getdtime(CLOCK_MONOTONIC) Add an option to also output realtime along with monotime.
import getopt, sys if __name__ == '__main__': sippy_path = None try: opts, args = getopt.getopt(sys.argv[1:], 'rS:') except getopt.GetoptError: usage() out_realtime = False for o, a in opts: if o == '-S': sippy_path = a.strip() continue if o == '-r': out_realtime = True if sippy_path != None: sys.path.insert(0, sippy_path) from sippy.Time.clock_dtime import clock_getdtime, CLOCK_MONOTONIC if not out_realtime: print(clock_getdtime(CLOCK_MONOTONIC)) else: from sippy.Time.clock_dtime import CLOCK_REALTIME print("%f %f" % (clock_getdtime(CLOCK_MONOTONIC), clock_getdtime(CLOCK_REALTIME)))
<commit_before>import getopt, sys if __name__ == '__main__': sippy_path = None try: opts, args = getopt.getopt(sys.argv[1:], 's:S:i:o:b') except getopt.GetoptError: usage() for o, a in opts: if o == '-S': sippy_path = a.strip() continue if sippy_path != None: sys.path.insert(0, sippy_path) from sippy.Time.clock_dtime import clock_getdtime, CLOCK_MONOTONIC print clock_getdtime(CLOCK_MONOTONIC) <commit_msg>Add an option to also output realtime along with monotime.<commit_after>
import getopt, sys if __name__ == '__main__': sippy_path = None try: opts, args = getopt.getopt(sys.argv[1:], 'rS:') except getopt.GetoptError: usage() out_realtime = False for o, a in opts: if o == '-S': sippy_path = a.strip() continue if o == '-r': out_realtime = True if sippy_path != None: sys.path.insert(0, sippy_path) from sippy.Time.clock_dtime import clock_getdtime, CLOCK_MONOTONIC if not out_realtime: print(clock_getdtime(CLOCK_MONOTONIC)) else: from sippy.Time.clock_dtime import CLOCK_REALTIME print("%f %f" % (clock_getdtime(CLOCK_MONOTONIC), clock_getdtime(CLOCK_REALTIME)))
import getopt, sys if __name__ == '__main__': sippy_path = None try: opts, args = getopt.getopt(sys.argv[1:], 's:S:i:o:b') except getopt.GetoptError: usage() for o, a in opts: if o == '-S': sippy_path = a.strip() continue if sippy_path != None: sys.path.insert(0, sippy_path) from sippy.Time.clock_dtime import clock_getdtime, CLOCK_MONOTONIC print clock_getdtime(CLOCK_MONOTONIC) Add an option to also output realtime along with monotime.import getopt, sys if __name__ == '__main__': sippy_path = None try: opts, args = getopt.getopt(sys.argv[1:], 'rS:') except getopt.GetoptError: usage() out_realtime = False for o, a in opts: if o == '-S': sippy_path = a.strip() continue if o == '-r': out_realtime = True if sippy_path != None: sys.path.insert(0, sippy_path) from sippy.Time.clock_dtime import clock_getdtime, CLOCK_MONOTONIC if not out_realtime: print(clock_getdtime(CLOCK_MONOTONIC)) else: from sippy.Time.clock_dtime import CLOCK_REALTIME print("%f %f" % (clock_getdtime(CLOCK_MONOTONIC), clock_getdtime(CLOCK_REALTIME)))
<commit_before>import getopt, sys if __name__ == '__main__': sippy_path = None try: opts, args = getopt.getopt(sys.argv[1:], 's:S:i:o:b') except getopt.GetoptError: usage() for o, a in opts: if o == '-S': sippy_path = a.strip() continue if sippy_path != None: sys.path.insert(0, sippy_path) from sippy.Time.clock_dtime import clock_getdtime, CLOCK_MONOTONIC print clock_getdtime(CLOCK_MONOTONIC) <commit_msg>Add an option to also output realtime along with monotime.<commit_after>import getopt, sys if __name__ == '__main__': sippy_path = None try: opts, args = getopt.getopt(sys.argv[1:], 'rS:') except getopt.GetoptError: usage() out_realtime = False for o, a in opts: if o == '-S': sippy_path = a.strip() continue if o == '-r': out_realtime = True if sippy_path != None: sys.path.insert(0, sippy_path) from sippy.Time.clock_dtime import clock_getdtime, CLOCK_MONOTONIC if not out_realtime: print(clock_getdtime(CLOCK_MONOTONIC)) else: from sippy.Time.clock_dtime import CLOCK_REALTIME print("%f %f" % (clock_getdtime(CLOCK_MONOTONIC), clock_getdtime(CLOCK_REALTIME)))
ff45b8c21f99b20ed044e8b194bc84f21f4f15d7
httpserver_with_post.py
httpserver_with_post.py
# Adapted from http://stackoverflow.com/questions/10017859/how-to-build-a-simple-http-post-server # Thank you! import sys import BaseHTTPServer import cgi class MyHandler(BaseHTTPServer.BaseHTTPRequestHandler): def do_POST(self): ctype, pdict = cgi.parse_header(self.headers.getheader('content-type')) postvars = {} try: if ctype == 'application/x-www-form-urlencoded': length = int(self.headers.getheader('content-length')) postvars = cgi.parse_qs(self.rfile.read(length), keep_blank_values=1) self.send_response(200) self.send_header("Content-type", "text") self.send_header("Content-length", str(len(body))) self.end_headers() self.wfile.write(body) except: print "Error" def httpd(handler_class=MyHandler, server_address = ('127.0.0.1', 8000)): try: print "Server started" srvr = BaseHTTPServer.HTTPServer(server_address, handler_class) srvr.serve_forever() # serve_forever except KeyboardInterrupt: srvr.socket.close() if __name__ == "__main__": httpd(server_address = (sys.argv[1], int(sys.argv[2])))
# Adapted from http://stackoverflow.com/questions/10017859/how-to-build-a-simple-http-post-server # Thank you! import sys import BaseHTTPServer import cgi class MyHandler(BaseHTTPServer.BaseHTTPRequestHandler): def do_POST(self): ctype, pdict = cgi.parse_header(self.headers.getheader('content-type')) postvars = {} try: if ctype == 'application/x-www-form-urlencoded': length = int(self.headers.getheader('content-length')) postvars = cgi.parse_qs(self.rfile.read(length), keep_blank_values=1) print "Client posted", postvars self.send_response(200) """ self.send_header("Content-type", "text") self.send_header("Content-length", str(len(body))) """ self.end_headers() # self.wfile.write(body) except Exception, e: print "Error", repr(e) def httpd(handler_class=MyHandler, server_address = ('127.0.0.1', 8000)): try: print "Server started" srvr = BaseHTTPServer.HTTPServer(server_address, handler_class) srvr.serve_forever() # serve_forever except KeyboardInterrupt: srvr.socket.close() if __name__ == "__main__": httpd(server_address = (sys.argv[1], int(sys.argv[2])))
Print client-POSTed data, more verbose error handling
Print client-POSTed data, more verbose error handling And less fiddling with the returned header. For the time being, I don't care about correcting the bugs in that part of the code.
Python
unlicense
aaaaalbert/repy-doodles
# Adapted from http://stackoverflow.com/questions/10017859/how-to-build-a-simple-http-post-server # Thank you! import sys import BaseHTTPServer import cgi class MyHandler(BaseHTTPServer.BaseHTTPRequestHandler): def do_POST(self): ctype, pdict = cgi.parse_header(self.headers.getheader('content-type')) postvars = {} try: if ctype == 'application/x-www-form-urlencoded': length = int(self.headers.getheader('content-length')) postvars = cgi.parse_qs(self.rfile.read(length), keep_blank_values=1) self.send_response(200) self.send_header("Content-type", "text") self.send_header("Content-length", str(len(body))) self.end_headers() self.wfile.write(body) except: print "Error" def httpd(handler_class=MyHandler, server_address = ('127.0.0.1', 8000)): try: print "Server started" srvr = BaseHTTPServer.HTTPServer(server_address, handler_class) srvr.serve_forever() # serve_forever except KeyboardInterrupt: srvr.socket.close() if __name__ == "__main__": httpd(server_address = (sys.argv[1], int(sys.argv[2]))) Print client-POSTed data, more verbose error handling And less fiddling with the returned header. For the time being, I don't care about correcting the bugs in that part of the code.
# Adapted from http://stackoverflow.com/questions/10017859/how-to-build-a-simple-http-post-server # Thank you! import sys import BaseHTTPServer import cgi class MyHandler(BaseHTTPServer.BaseHTTPRequestHandler): def do_POST(self): ctype, pdict = cgi.parse_header(self.headers.getheader('content-type')) postvars = {} try: if ctype == 'application/x-www-form-urlencoded': length = int(self.headers.getheader('content-length')) postvars = cgi.parse_qs(self.rfile.read(length), keep_blank_values=1) print "Client posted", postvars self.send_response(200) """ self.send_header("Content-type", "text") self.send_header("Content-length", str(len(body))) """ self.end_headers() # self.wfile.write(body) except Exception, e: print "Error", repr(e) def httpd(handler_class=MyHandler, server_address = ('127.0.0.1', 8000)): try: print "Server started" srvr = BaseHTTPServer.HTTPServer(server_address, handler_class) srvr.serve_forever() # serve_forever except KeyboardInterrupt: srvr.socket.close() if __name__ == "__main__": httpd(server_address = (sys.argv[1], int(sys.argv[2])))
<commit_before># Adapted from http://stackoverflow.com/questions/10017859/how-to-build-a-simple-http-post-server # Thank you! import sys import BaseHTTPServer import cgi class MyHandler(BaseHTTPServer.BaseHTTPRequestHandler): def do_POST(self): ctype, pdict = cgi.parse_header(self.headers.getheader('content-type')) postvars = {} try: if ctype == 'application/x-www-form-urlencoded': length = int(self.headers.getheader('content-length')) postvars = cgi.parse_qs(self.rfile.read(length), keep_blank_values=1) self.send_response(200) self.send_header("Content-type", "text") self.send_header("Content-length", str(len(body))) self.end_headers() self.wfile.write(body) except: print "Error" def httpd(handler_class=MyHandler, server_address = ('127.0.0.1', 8000)): try: print "Server started" srvr = BaseHTTPServer.HTTPServer(server_address, handler_class) srvr.serve_forever() # serve_forever except KeyboardInterrupt: srvr.socket.close() if __name__ == "__main__": httpd(server_address = (sys.argv[1], int(sys.argv[2]))) <commit_msg>Print client-POSTed data, more verbose error handling And less fiddling with the returned header. For the time being, I don't care about correcting the bugs in that part of the code.<commit_after>
# Adapted from http://stackoverflow.com/questions/10017859/how-to-build-a-simple-http-post-server # Thank you! import sys import BaseHTTPServer import cgi class MyHandler(BaseHTTPServer.BaseHTTPRequestHandler): def do_POST(self): ctype, pdict = cgi.parse_header(self.headers.getheader('content-type')) postvars = {} try: if ctype == 'application/x-www-form-urlencoded': length = int(self.headers.getheader('content-length')) postvars = cgi.parse_qs(self.rfile.read(length), keep_blank_values=1) print "Client posted", postvars self.send_response(200) """ self.send_header("Content-type", "text") self.send_header("Content-length", str(len(body))) """ self.end_headers() # self.wfile.write(body) except Exception, e: print "Error", repr(e) def httpd(handler_class=MyHandler, server_address = ('127.0.0.1', 8000)): try: print "Server started" srvr = BaseHTTPServer.HTTPServer(server_address, handler_class) srvr.serve_forever() # serve_forever except KeyboardInterrupt: srvr.socket.close() if __name__ == "__main__": httpd(server_address = (sys.argv[1], int(sys.argv[2])))
# Adapted from http://stackoverflow.com/questions/10017859/how-to-build-a-simple-http-post-server # Thank you! import sys import BaseHTTPServer import cgi class MyHandler(BaseHTTPServer.BaseHTTPRequestHandler): def do_POST(self): ctype, pdict = cgi.parse_header(self.headers.getheader('content-type')) postvars = {} try: if ctype == 'application/x-www-form-urlencoded': length = int(self.headers.getheader('content-length')) postvars = cgi.parse_qs(self.rfile.read(length), keep_blank_values=1) self.send_response(200) self.send_header("Content-type", "text") self.send_header("Content-length", str(len(body))) self.end_headers() self.wfile.write(body) except: print "Error" def httpd(handler_class=MyHandler, server_address = ('127.0.0.1', 8000)): try: print "Server started" srvr = BaseHTTPServer.HTTPServer(server_address, handler_class) srvr.serve_forever() # serve_forever except KeyboardInterrupt: srvr.socket.close() if __name__ == "__main__": httpd(server_address = (sys.argv[1], int(sys.argv[2]))) Print client-POSTed data, more verbose error handling And less fiddling with the returned header. For the time being, I don't care about correcting the bugs in that part of the code.# Adapted from http://stackoverflow.com/questions/10017859/how-to-build-a-simple-http-post-server # Thank you! import sys import BaseHTTPServer import cgi class MyHandler(BaseHTTPServer.BaseHTTPRequestHandler): def do_POST(self): ctype, pdict = cgi.parse_header(self.headers.getheader('content-type')) postvars = {} try: if ctype == 'application/x-www-form-urlencoded': length = int(self.headers.getheader('content-length')) postvars = cgi.parse_qs(self.rfile.read(length), keep_blank_values=1) print "Client posted", postvars self.send_response(200) """ self.send_header("Content-type", "text") self.send_header("Content-length", str(len(body))) """ self.end_headers() # self.wfile.write(body) except Exception, e: print "Error", repr(e) def httpd(handler_class=MyHandler, server_address = ('127.0.0.1', 8000)): try: print "Server started" srvr = BaseHTTPServer.HTTPServer(server_address, handler_class) srvr.serve_forever() # serve_forever except KeyboardInterrupt: srvr.socket.close() if __name__ == "__main__": httpd(server_address = (sys.argv[1], int(sys.argv[2])))
<commit_before># Adapted from http://stackoverflow.com/questions/10017859/how-to-build-a-simple-http-post-server # Thank you! import sys import BaseHTTPServer import cgi class MyHandler(BaseHTTPServer.BaseHTTPRequestHandler): def do_POST(self): ctype, pdict = cgi.parse_header(self.headers.getheader('content-type')) postvars = {} try: if ctype == 'application/x-www-form-urlencoded': length = int(self.headers.getheader('content-length')) postvars = cgi.parse_qs(self.rfile.read(length), keep_blank_values=1) self.send_response(200) self.send_header("Content-type", "text") self.send_header("Content-length", str(len(body))) self.end_headers() self.wfile.write(body) except: print "Error" def httpd(handler_class=MyHandler, server_address = ('127.0.0.1', 8000)): try: print "Server started" srvr = BaseHTTPServer.HTTPServer(server_address, handler_class) srvr.serve_forever() # serve_forever except KeyboardInterrupt: srvr.socket.close() if __name__ == "__main__": httpd(server_address = (sys.argv[1], int(sys.argv[2]))) <commit_msg>Print client-POSTed data, more verbose error handling And less fiddling with the returned header. For the time being, I don't care about correcting the bugs in that part of the code.<commit_after># Adapted from http://stackoverflow.com/questions/10017859/how-to-build-a-simple-http-post-server # Thank you! import sys import BaseHTTPServer import cgi class MyHandler(BaseHTTPServer.BaseHTTPRequestHandler): def do_POST(self): ctype, pdict = cgi.parse_header(self.headers.getheader('content-type')) postvars = {} try: if ctype == 'application/x-www-form-urlencoded': length = int(self.headers.getheader('content-length')) postvars = cgi.parse_qs(self.rfile.read(length), keep_blank_values=1) print "Client posted", postvars self.send_response(200) """ self.send_header("Content-type", "text") self.send_header("Content-length", str(len(body))) """ self.end_headers() # self.wfile.write(body) except Exception, e: print "Error", repr(e) def httpd(handler_class=MyHandler, server_address = ('127.0.0.1', 8000)): try: print "Server started" srvr = BaseHTTPServer.HTTPServer(server_address, handler_class) srvr.serve_forever() # serve_forever except KeyboardInterrupt: srvr.socket.close() if __name__ == "__main__": httpd(server_address = (sys.argv[1], int(sys.argv[2])))
5b54df50752b3f661ad43f2086734f90a8d1a11e
src/ggrc/migrations/versions/20150205020509_5254f4f31427_system_editable_object_state.py
src/ggrc/migrations/versions/20150205020509_5254f4f31427_system_editable_object_state.py
"""System editable object state Revision ID: 5254f4f31427 Revises: 512c71e4d93b Create Date: 2015-02-05 02:05:09.351265 """ # revision identifiers, used by Alembic. revision = '5254f4f31427' down_revision = '512c71e4d93b' import sqlalchemy as sa from sqlalchemy.sql import table, column from alembic import op from ggrc.models.track_object_state import ObjectStates, ObjectStateTables def upgrade(): for table_name in ObjectStateTables.table_names: op.add_column(table_name, sa.Column('os_state', sa.String(length=16), nullable=True)) # Set the value into all existing records object_table = table(table_name, column('os_state', sa.String(length=16))) connection = op.get_bind() connection.execute( object_table.update().values( { 'os_state': ObjectStates.DRAFT } ) ) # Make the field not-nullable op.alter_column(table_name, 'os_state',existing_type=sa.String(length=16),nullable=False) def downgrade(): for table_name in ObjectStateTables: op.drop_column(table_name, 'os_state')
"""System editable object state Revision ID: 5254f4f31427 Revises: 512c71e4d93b Create Date: 2015-02-05 02:05:09.351265 """ # revision identifiers, used by Alembic. revision = '5254f4f31427' down_revision = '512c71e4d93b' import sqlalchemy as sa from sqlalchemy.sql import table, column from alembic import op from ggrc.models.track_object_state import ObjectStates, ObjectStateTables def upgrade(): for table_name in ObjectStateTables.table_names: op.add_column(table_name, sa.Column('os_state', sa.String(length=16), nullable=True)) # Set the value into all existing records object_table = table(table_name, column('os_state', sa.String(length=16))) connection = op.get_bind() connection.execute( object_table.update().values( { 'os_state': ObjectStates.DRAFT } ) ) # Make the field not-nullable op.alter_column(table_name, 'os_state',existing_type=sa.String(length=16),nullable=False) def downgrade(): for table_name in ObjectStateTables.table_names: op.drop_column(table_name, 'os_state')
Fix db_downgrade for "System editable object state"
Fix db_downgrade for "System editable object state"
Python
apache-2.0
jmakov/ggrc-core,edofic/ggrc-core,uskudnik/ggrc-core,plamut/ggrc-core,vladan-m/ggrc-core,j0gurt/ggrc-core,selahssea/ggrc-core,andrei-karalionak/ggrc-core,edofic/ggrc-core,AleksNeStu/ggrc-core,selahssea/ggrc-core,j0gurt/ggrc-core,NejcZupec/ggrc-core,VinnieJohns/ggrc-core,hasanalom/ggrc-core,uskudnik/ggrc-core,j0gurt/ggrc-core,kr41/ggrc-core,josthkko/ggrc-core,jmakov/ggrc-core,andrei-karalionak/ggrc-core,prasannav7/ggrc-core,plamut/ggrc-core,hyperNURb/ggrc-core,selahssea/ggrc-core,jmakov/ggrc-core,kr41/ggrc-core,hyperNURb/ggrc-core,andrei-karalionak/ggrc-core,hasanalom/ggrc-core,hasanalom/ggrc-core,uskudnik/ggrc-core,vladan-m/ggrc-core,AleksNeStu/ggrc-core,VinnieJohns/ggrc-core,AleksNeStu/ggrc-core,uskudnik/ggrc-core,kr41/ggrc-core,VinnieJohns/ggrc-core,hasanalom/ggrc-core,josthkko/ggrc-core,vladan-m/ggrc-core,edofic/ggrc-core,AleksNeStu/ggrc-core,NejcZupec/ggrc-core,vladan-m/ggrc-core,hyperNURb/ggrc-core,prasannav7/ggrc-core,NejcZupec/ggrc-core,VinnieJohns/ggrc-core,andrei-karalionak/ggrc-core,j0gurt/ggrc-core,uskudnik/ggrc-core,kr41/ggrc-core,plamut/ggrc-core,selahssea/ggrc-core,vladan-m/ggrc-core,hasanalom/ggrc-core,jmakov/ggrc-core,hyperNURb/ggrc-core,prasannav7/ggrc-core,josthkko/ggrc-core,prasannav7/ggrc-core,josthkko/ggrc-core,hyperNURb/ggrc-core,NejcZupec/ggrc-core,edofic/ggrc-core,jmakov/ggrc-core,plamut/ggrc-core
"""System editable object state Revision ID: 5254f4f31427 Revises: 512c71e4d93b Create Date: 2015-02-05 02:05:09.351265 """ # revision identifiers, used by Alembic. revision = '5254f4f31427' down_revision = '512c71e4d93b' import sqlalchemy as sa from sqlalchemy.sql import table, column from alembic import op from ggrc.models.track_object_state import ObjectStates, ObjectStateTables def upgrade(): for table_name in ObjectStateTables.table_names: op.add_column(table_name, sa.Column('os_state', sa.String(length=16), nullable=True)) # Set the value into all existing records object_table = table(table_name, column('os_state', sa.String(length=16))) connection = op.get_bind() connection.execute( object_table.update().values( { 'os_state': ObjectStates.DRAFT } ) ) # Make the field not-nullable op.alter_column(table_name, 'os_state',existing_type=sa.String(length=16),nullable=False) def downgrade(): for table_name in ObjectStateTables: op.drop_column(table_name, 'os_state') Fix db_downgrade for "System editable object state"
"""System editable object state Revision ID: 5254f4f31427 Revises: 512c71e4d93b Create Date: 2015-02-05 02:05:09.351265 """ # revision identifiers, used by Alembic. revision = '5254f4f31427' down_revision = '512c71e4d93b' import sqlalchemy as sa from sqlalchemy.sql import table, column from alembic import op from ggrc.models.track_object_state import ObjectStates, ObjectStateTables def upgrade(): for table_name in ObjectStateTables.table_names: op.add_column(table_name, sa.Column('os_state', sa.String(length=16), nullable=True)) # Set the value into all existing records object_table = table(table_name, column('os_state', sa.String(length=16))) connection = op.get_bind() connection.execute( object_table.update().values( { 'os_state': ObjectStates.DRAFT } ) ) # Make the field not-nullable op.alter_column(table_name, 'os_state',existing_type=sa.String(length=16),nullable=False) def downgrade(): for table_name in ObjectStateTables.table_names: op.drop_column(table_name, 'os_state')
<commit_before> """System editable object state Revision ID: 5254f4f31427 Revises: 512c71e4d93b Create Date: 2015-02-05 02:05:09.351265 """ # revision identifiers, used by Alembic. revision = '5254f4f31427' down_revision = '512c71e4d93b' import sqlalchemy as sa from sqlalchemy.sql import table, column from alembic import op from ggrc.models.track_object_state import ObjectStates, ObjectStateTables def upgrade(): for table_name in ObjectStateTables.table_names: op.add_column(table_name, sa.Column('os_state', sa.String(length=16), nullable=True)) # Set the value into all existing records object_table = table(table_name, column('os_state', sa.String(length=16))) connection = op.get_bind() connection.execute( object_table.update().values( { 'os_state': ObjectStates.DRAFT } ) ) # Make the field not-nullable op.alter_column(table_name, 'os_state',existing_type=sa.String(length=16),nullable=False) def downgrade(): for table_name in ObjectStateTables: op.drop_column(table_name, 'os_state') <commit_msg>Fix db_downgrade for "System editable object state"<commit_after>
"""System editable object state Revision ID: 5254f4f31427 Revises: 512c71e4d93b Create Date: 2015-02-05 02:05:09.351265 """ # revision identifiers, used by Alembic. revision = '5254f4f31427' down_revision = '512c71e4d93b' import sqlalchemy as sa from sqlalchemy.sql import table, column from alembic import op from ggrc.models.track_object_state import ObjectStates, ObjectStateTables def upgrade(): for table_name in ObjectStateTables.table_names: op.add_column(table_name, sa.Column('os_state', sa.String(length=16), nullable=True)) # Set the value into all existing records object_table = table(table_name, column('os_state', sa.String(length=16))) connection = op.get_bind() connection.execute( object_table.update().values( { 'os_state': ObjectStates.DRAFT } ) ) # Make the field not-nullable op.alter_column(table_name, 'os_state',existing_type=sa.String(length=16),nullable=False) def downgrade(): for table_name in ObjectStateTables.table_names: op.drop_column(table_name, 'os_state')
"""System editable object state Revision ID: 5254f4f31427 Revises: 512c71e4d93b Create Date: 2015-02-05 02:05:09.351265 """ # revision identifiers, used by Alembic. revision = '5254f4f31427' down_revision = '512c71e4d93b' import sqlalchemy as sa from sqlalchemy.sql import table, column from alembic import op from ggrc.models.track_object_state import ObjectStates, ObjectStateTables def upgrade(): for table_name in ObjectStateTables.table_names: op.add_column(table_name, sa.Column('os_state', sa.String(length=16), nullable=True)) # Set the value into all existing records object_table = table(table_name, column('os_state', sa.String(length=16))) connection = op.get_bind() connection.execute( object_table.update().values( { 'os_state': ObjectStates.DRAFT } ) ) # Make the field not-nullable op.alter_column(table_name, 'os_state',existing_type=sa.String(length=16),nullable=False) def downgrade(): for table_name in ObjectStateTables: op.drop_column(table_name, 'os_state') Fix db_downgrade for "System editable object state" """System editable object state Revision ID: 5254f4f31427 Revises: 512c71e4d93b Create Date: 2015-02-05 02:05:09.351265 """ # revision identifiers, used by Alembic. revision = '5254f4f31427' down_revision = '512c71e4d93b' import sqlalchemy as sa from sqlalchemy.sql import table, column from alembic import op from ggrc.models.track_object_state import ObjectStates, ObjectStateTables def upgrade(): for table_name in ObjectStateTables.table_names: op.add_column(table_name, sa.Column('os_state', sa.String(length=16), nullable=True)) # Set the value into all existing records object_table = table(table_name, column('os_state', sa.String(length=16))) connection = op.get_bind() connection.execute( object_table.update().values( { 'os_state': ObjectStates.DRAFT } ) ) # Make the field not-nullable op.alter_column(table_name, 'os_state',existing_type=sa.String(length=16),nullable=False) def downgrade(): for table_name in ObjectStateTables.table_names: op.drop_column(table_name, 'os_state')
<commit_before> """System editable object state Revision ID: 5254f4f31427 Revises: 512c71e4d93b Create Date: 2015-02-05 02:05:09.351265 """ # revision identifiers, used by Alembic. revision = '5254f4f31427' down_revision = '512c71e4d93b' import sqlalchemy as sa from sqlalchemy.sql import table, column from alembic import op from ggrc.models.track_object_state import ObjectStates, ObjectStateTables def upgrade(): for table_name in ObjectStateTables.table_names: op.add_column(table_name, sa.Column('os_state', sa.String(length=16), nullable=True)) # Set the value into all existing records object_table = table(table_name, column('os_state', sa.String(length=16))) connection = op.get_bind() connection.execute( object_table.update().values( { 'os_state': ObjectStates.DRAFT } ) ) # Make the field not-nullable op.alter_column(table_name, 'os_state',existing_type=sa.String(length=16),nullable=False) def downgrade(): for table_name in ObjectStateTables: op.drop_column(table_name, 'os_state') <commit_msg>Fix db_downgrade for "System editable object state"<commit_after> """System editable object state Revision ID: 5254f4f31427 Revises: 512c71e4d93b Create Date: 2015-02-05 02:05:09.351265 """ # revision identifiers, used by Alembic. revision = '5254f4f31427' down_revision = '512c71e4d93b' import sqlalchemy as sa from sqlalchemy.sql import table, column from alembic import op from ggrc.models.track_object_state import ObjectStates, ObjectStateTables def upgrade(): for table_name in ObjectStateTables.table_names: op.add_column(table_name, sa.Column('os_state', sa.String(length=16), nullable=True)) # Set the value into all existing records object_table = table(table_name, column('os_state', sa.String(length=16))) connection = op.get_bind() connection.execute( object_table.update().values( { 'os_state': ObjectStates.DRAFT } ) ) # Make the field not-nullable op.alter_column(table_name, 'os_state',existing_type=sa.String(length=16),nullable=False) def downgrade(): for table_name in ObjectStateTables.table_names: op.drop_column(table_name, 'os_state')
877a3470044c98d3a938633479d38df6df6d26bd
boltiot/urls.py
boltiot/urls.py
#Creating a key value store for all the urls BASE_URL = 'http://cloud.boltiot.com/remote/' url_list = { 'digitalWrite' : '{}/digitalWrite?pin={}&state={}&deviceName={}', 'digitalRead' : '{}/digitalRead?pin={}&deviceName={}', 'analogWrite' : '{}/analogWrite?pin=1&value={}&state={}&deviceName={}', 'analogRead' : '{}/analogRead?pin={}&deviceName={}', 'serialBegin' : '{}/serialBegin?baud={}&deviceName={}', 'serialWrite' : '{}/serialWrite?data={}&deviceName={}', 'serialRead' : '{}/serialRead?till={}&deviceName={}', 'version' : '{}/version?&deviceName={}', 'restart' : '{}/restart?&deviceName={}', 'isAlive' : '{}/isAlive?&deviceName={}', } def url(operation): return BASE_URL+url_list[operation]
#Creating a key value store for all the urls BASE_URL = 'http://cloud.boltiot.com/remote/' url_list = { 'digitalWrite' : '{}/digitalWrite?pin={}&state={}&deviceName={}', 'digitalRead' : '{}/digitalRead?pin={}&deviceName={}', 'analogWrite' : '{}/analogWrite?pin={}&value={}&deviceName={}', 'analogRead' : '{}/analogRead?pin={}&deviceName={}', 'serialBegin' : '{}/serialBegin?baud={}&deviceName={}', 'serialWrite' : '{}/serialWrite?data={}&deviceName={}', 'serialRead' : '{}/serialRead?till={}&deviceName={}', 'version' : '{}/version?&deviceName={}', 'restart' : '{}/restart?&deviceName={}', 'isAlive' : '{}/isAlive?&deviceName={}', } def url(operation): return BASE_URL+url_list[operation]
Remove the static pin fir analog read
Remove the static pin fir analog read
Python
mit
Inventrom/bolt-api-python
#Creating a key value store for all the urls BASE_URL = 'http://cloud.boltiot.com/remote/' url_list = { 'digitalWrite' : '{}/digitalWrite?pin={}&state={}&deviceName={}', 'digitalRead' : '{}/digitalRead?pin={}&deviceName={}', 'analogWrite' : '{}/analogWrite?pin=1&value={}&state={}&deviceName={}', 'analogRead' : '{}/analogRead?pin={}&deviceName={}', 'serialBegin' : '{}/serialBegin?baud={}&deviceName={}', 'serialWrite' : '{}/serialWrite?data={}&deviceName={}', 'serialRead' : '{}/serialRead?till={}&deviceName={}', 'version' : '{}/version?&deviceName={}', 'restart' : '{}/restart?&deviceName={}', 'isAlive' : '{}/isAlive?&deviceName={}', } def url(operation): return BASE_URL+url_list[operation] Remove the static pin fir analog read
#Creating a key value store for all the urls BASE_URL = 'http://cloud.boltiot.com/remote/' url_list = { 'digitalWrite' : '{}/digitalWrite?pin={}&state={}&deviceName={}', 'digitalRead' : '{}/digitalRead?pin={}&deviceName={}', 'analogWrite' : '{}/analogWrite?pin={}&value={}&deviceName={}', 'analogRead' : '{}/analogRead?pin={}&deviceName={}', 'serialBegin' : '{}/serialBegin?baud={}&deviceName={}', 'serialWrite' : '{}/serialWrite?data={}&deviceName={}', 'serialRead' : '{}/serialRead?till={}&deviceName={}', 'version' : '{}/version?&deviceName={}', 'restart' : '{}/restart?&deviceName={}', 'isAlive' : '{}/isAlive?&deviceName={}', } def url(operation): return BASE_URL+url_list[operation]
<commit_before>#Creating a key value store for all the urls BASE_URL = 'http://cloud.boltiot.com/remote/' url_list = { 'digitalWrite' : '{}/digitalWrite?pin={}&state={}&deviceName={}', 'digitalRead' : '{}/digitalRead?pin={}&deviceName={}', 'analogWrite' : '{}/analogWrite?pin=1&value={}&state={}&deviceName={}', 'analogRead' : '{}/analogRead?pin={}&deviceName={}', 'serialBegin' : '{}/serialBegin?baud={}&deviceName={}', 'serialWrite' : '{}/serialWrite?data={}&deviceName={}', 'serialRead' : '{}/serialRead?till={}&deviceName={}', 'version' : '{}/version?&deviceName={}', 'restart' : '{}/restart?&deviceName={}', 'isAlive' : '{}/isAlive?&deviceName={}', } def url(operation): return BASE_URL+url_list[operation] <commit_msg>Remove the static pin fir analog read<commit_after>
#Creating a key value store for all the urls BASE_URL = 'http://cloud.boltiot.com/remote/' url_list = { 'digitalWrite' : '{}/digitalWrite?pin={}&state={}&deviceName={}', 'digitalRead' : '{}/digitalRead?pin={}&deviceName={}', 'analogWrite' : '{}/analogWrite?pin={}&value={}&deviceName={}', 'analogRead' : '{}/analogRead?pin={}&deviceName={}', 'serialBegin' : '{}/serialBegin?baud={}&deviceName={}', 'serialWrite' : '{}/serialWrite?data={}&deviceName={}', 'serialRead' : '{}/serialRead?till={}&deviceName={}', 'version' : '{}/version?&deviceName={}', 'restart' : '{}/restart?&deviceName={}', 'isAlive' : '{}/isAlive?&deviceName={}', } def url(operation): return BASE_URL+url_list[operation]
#Creating a key value store for all the urls BASE_URL = 'http://cloud.boltiot.com/remote/' url_list = { 'digitalWrite' : '{}/digitalWrite?pin={}&state={}&deviceName={}', 'digitalRead' : '{}/digitalRead?pin={}&deviceName={}', 'analogWrite' : '{}/analogWrite?pin=1&value={}&state={}&deviceName={}', 'analogRead' : '{}/analogRead?pin={}&deviceName={}', 'serialBegin' : '{}/serialBegin?baud={}&deviceName={}', 'serialWrite' : '{}/serialWrite?data={}&deviceName={}', 'serialRead' : '{}/serialRead?till={}&deviceName={}', 'version' : '{}/version?&deviceName={}', 'restart' : '{}/restart?&deviceName={}', 'isAlive' : '{}/isAlive?&deviceName={}', } def url(operation): return BASE_URL+url_list[operation] Remove the static pin fir analog read#Creating a key value store for all the urls BASE_URL = 'http://cloud.boltiot.com/remote/' url_list = { 'digitalWrite' : '{}/digitalWrite?pin={}&state={}&deviceName={}', 'digitalRead' : '{}/digitalRead?pin={}&deviceName={}', 'analogWrite' : '{}/analogWrite?pin={}&value={}&deviceName={}', 'analogRead' : '{}/analogRead?pin={}&deviceName={}', 'serialBegin' : '{}/serialBegin?baud={}&deviceName={}', 'serialWrite' : '{}/serialWrite?data={}&deviceName={}', 'serialRead' : '{}/serialRead?till={}&deviceName={}', 'version' : '{}/version?&deviceName={}', 'restart' : '{}/restart?&deviceName={}', 'isAlive' : '{}/isAlive?&deviceName={}', } def url(operation): return BASE_URL+url_list[operation]
<commit_before>#Creating a key value store for all the urls BASE_URL = 'http://cloud.boltiot.com/remote/' url_list = { 'digitalWrite' : '{}/digitalWrite?pin={}&state={}&deviceName={}', 'digitalRead' : '{}/digitalRead?pin={}&deviceName={}', 'analogWrite' : '{}/analogWrite?pin=1&value={}&state={}&deviceName={}', 'analogRead' : '{}/analogRead?pin={}&deviceName={}', 'serialBegin' : '{}/serialBegin?baud={}&deviceName={}', 'serialWrite' : '{}/serialWrite?data={}&deviceName={}', 'serialRead' : '{}/serialRead?till={}&deviceName={}', 'version' : '{}/version?&deviceName={}', 'restart' : '{}/restart?&deviceName={}', 'isAlive' : '{}/isAlive?&deviceName={}', } def url(operation): return BASE_URL+url_list[operation] <commit_msg>Remove the static pin fir analog read<commit_after>#Creating a key value store for all the urls BASE_URL = 'http://cloud.boltiot.com/remote/' url_list = { 'digitalWrite' : '{}/digitalWrite?pin={}&state={}&deviceName={}', 'digitalRead' : '{}/digitalRead?pin={}&deviceName={}', 'analogWrite' : '{}/analogWrite?pin={}&value={}&deviceName={}', 'analogRead' : '{}/analogRead?pin={}&deviceName={}', 'serialBegin' : '{}/serialBegin?baud={}&deviceName={}', 'serialWrite' : '{}/serialWrite?data={}&deviceName={}', 'serialRead' : '{}/serialRead?till={}&deviceName={}', 'version' : '{}/version?&deviceName={}', 'restart' : '{}/restart?&deviceName={}', 'isAlive' : '{}/isAlive?&deviceName={}', } def url(operation): return BASE_URL+url_list[operation]
ab5d570b92aca2c598d12fcdb0b063782ad4c871
templates/root/appfiles/urls.py
templates/root/appfiles/urls.py
"""template URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.8/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Add an import: from blog import urls as blog_urls 2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls)) """ from django.conf.urls import include, url, patterns from django.contrib import admin from <%= appName %>.forms import AuthenticationForm urlpatterns = patterns('', url(r'^$', '<%= appName %>.views.index', name="index"), url(r'^', include('<%= appName %>.urls')), url(r'^admin/', include(admin.site.urls)), ) <% if (includeLoginPage == true) { %> urlpatterns += patterns( 'django.contrib.auth', url(r'^accounts/login/$', 'views.login', { 'template_name': 'login.html', 'authentication_form': AuthenticationForm }, name='login'), url(r'^accounts/logout/$', 'views.logout', { 'next_page': '/' }, name='logout'), ) <% } %>
"""template URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.8/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Add an import: from blog import urls as blog_urls 2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls)) """ from django.conf.urls import include, url, patterns from django.contrib import admin <% if (includeLoginPage == true) { %> from <%= appName %>.forms import AuthenticationForm <% } %> urlpatterns = patterns('', url(r'^$', '<%= appName %>.views.index', name="index"), url(r'^', include('<%= appName %>.urls')), url(r'^admin/', include(admin.site.urls)), ) <% if (includeLoginPage == true) { %> urlpatterns += patterns( 'django.contrib.auth', url(r'^accounts/login/$', 'views.login', { 'template_name': 'login.html', 'authentication_form': AuthenticationForm }, name='login'), url(r'^accounts/logout/$', 'views.logout', { 'next_page': '/' }, name='logout'), ) <% } %>
Fix Import error as a result of answering No to include Login
Fix Import error as a result of answering No to include Login
Python
mit
dfurtado/generator-djangospa,dfurtado/generator-djangospa,dfurtado/generator-djangospa
"""template URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.8/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Add an import: from blog import urls as blog_urls 2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls)) """ from django.conf.urls import include, url, patterns from django.contrib import admin from <%= appName %>.forms import AuthenticationForm urlpatterns = patterns('', url(r'^$', '<%= appName %>.views.index', name="index"), url(r'^', include('<%= appName %>.urls')), url(r'^admin/', include(admin.site.urls)), ) <% if (includeLoginPage == true) { %> urlpatterns += patterns( 'django.contrib.auth', url(r'^accounts/login/$', 'views.login', { 'template_name': 'login.html', 'authentication_form': AuthenticationForm }, name='login'), url(r'^accounts/logout/$', 'views.logout', { 'next_page': '/' }, name='logout'), ) <% } %> Fix Import error as a result of answering No to include Login
"""template URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.8/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Add an import: from blog import urls as blog_urls 2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls)) """ from django.conf.urls import include, url, patterns from django.contrib import admin <% if (includeLoginPage == true) { %> from <%= appName %>.forms import AuthenticationForm <% } %> urlpatterns = patterns('', url(r'^$', '<%= appName %>.views.index', name="index"), url(r'^', include('<%= appName %>.urls')), url(r'^admin/', include(admin.site.urls)), ) <% if (includeLoginPage == true) { %> urlpatterns += patterns( 'django.contrib.auth', url(r'^accounts/login/$', 'views.login', { 'template_name': 'login.html', 'authentication_form': AuthenticationForm }, name='login'), url(r'^accounts/logout/$', 'views.logout', { 'next_page': '/' }, name='logout'), ) <% } %>
<commit_before>"""template URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.8/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Add an import: from blog import urls as blog_urls 2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls)) """ from django.conf.urls import include, url, patterns from django.contrib import admin from <%= appName %>.forms import AuthenticationForm urlpatterns = patterns('', url(r'^$', '<%= appName %>.views.index', name="index"), url(r'^', include('<%= appName %>.urls')), url(r'^admin/', include(admin.site.urls)), ) <% if (includeLoginPage == true) { %> urlpatterns += patterns( 'django.contrib.auth', url(r'^accounts/login/$', 'views.login', { 'template_name': 'login.html', 'authentication_form': AuthenticationForm }, name='login'), url(r'^accounts/logout/$', 'views.logout', { 'next_page': '/' }, name='logout'), ) <% } %> <commit_msg>Fix Import error as a result of answering No to include Login<commit_after>
"""template URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.8/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Add an import: from blog import urls as blog_urls 2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls)) """ from django.conf.urls import include, url, patterns from django.contrib import admin <% if (includeLoginPage == true) { %> from <%= appName %>.forms import AuthenticationForm <% } %> urlpatterns = patterns('', url(r'^$', '<%= appName %>.views.index', name="index"), url(r'^', include('<%= appName %>.urls')), url(r'^admin/', include(admin.site.urls)), ) <% if (includeLoginPage == true) { %> urlpatterns += patterns( 'django.contrib.auth', url(r'^accounts/login/$', 'views.login', { 'template_name': 'login.html', 'authentication_form': AuthenticationForm }, name='login'), url(r'^accounts/logout/$', 'views.logout', { 'next_page': '/' }, name='logout'), ) <% } %>
"""template URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.8/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Add an import: from blog import urls as blog_urls 2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls)) """ from django.conf.urls import include, url, patterns from django.contrib import admin from <%= appName %>.forms import AuthenticationForm urlpatterns = patterns('', url(r'^$', '<%= appName %>.views.index', name="index"), url(r'^', include('<%= appName %>.urls')), url(r'^admin/', include(admin.site.urls)), ) <% if (includeLoginPage == true) { %> urlpatterns += patterns( 'django.contrib.auth', url(r'^accounts/login/$', 'views.login', { 'template_name': 'login.html', 'authentication_form': AuthenticationForm }, name='login'), url(r'^accounts/logout/$', 'views.logout', { 'next_page': '/' }, name='logout'), ) <% } %> Fix Import error as a result of answering No to include Login"""template URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.8/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Add an import: from blog import urls as blog_urls 2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls)) """ from django.conf.urls import include, url, patterns from django.contrib import admin <% if (includeLoginPage == true) { %> from <%= appName %>.forms import AuthenticationForm <% } %> urlpatterns = patterns('', url(r'^$', '<%= appName %>.views.index', name="index"), url(r'^', include('<%= appName %>.urls')), url(r'^admin/', include(admin.site.urls)), ) <% if (includeLoginPage == true) { %> urlpatterns += patterns( 'django.contrib.auth', url(r'^accounts/login/$', 'views.login', { 'template_name': 'login.html', 'authentication_form': AuthenticationForm }, name='login'), url(r'^accounts/logout/$', 'views.logout', { 'next_page': '/' }, name='logout'), ) <% } %>
<commit_before>"""template URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.8/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Add an import: from blog import urls as blog_urls 2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls)) """ from django.conf.urls import include, url, patterns from django.contrib import admin from <%= appName %>.forms import AuthenticationForm urlpatterns = patterns('', url(r'^$', '<%= appName %>.views.index', name="index"), url(r'^', include('<%= appName %>.urls')), url(r'^admin/', include(admin.site.urls)), ) <% if (includeLoginPage == true) { %> urlpatterns += patterns( 'django.contrib.auth', url(r'^accounts/login/$', 'views.login', { 'template_name': 'login.html', 'authentication_form': AuthenticationForm }, name='login'), url(r'^accounts/logout/$', 'views.logout', { 'next_page': '/' }, name='logout'), ) <% } %> <commit_msg>Fix Import error as a result of answering No to include Login<commit_after>"""template URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.8/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Add an import: from blog import urls as blog_urls 2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls)) """ from django.conf.urls import include, url, patterns from django.contrib import admin <% if (includeLoginPage == true) { %> from <%= appName %>.forms import AuthenticationForm <% } %> urlpatterns = patterns('', url(r'^$', '<%= appName %>.views.index', name="index"), url(r'^', include('<%= appName %>.urls')), url(r'^admin/', include(admin.site.urls)), ) <% if (includeLoginPage == true) { %> urlpatterns += patterns( 'django.contrib.auth', url(r'^accounts/login/$', 'views.login', { 'template_name': 'login.html', 'authentication_form': AuthenticationForm }, name='login'), url(r'^accounts/logout/$', 'views.logout', { 'next_page': '/' }, name='logout'), ) <% } %>
115615a2a183684eed4f11e98a7da12190059fb1
armstrong/core/arm_layout/utils.py
armstrong/core/arm_layout/utils.py
# Here for backwards compatibility (deprecated) from django.utils.safestring import mark_safe from django.template.loader import render_to_string from armstrong.utils.backends import GenericBackend render_model = (GenericBackend("ARMSTRONG_RENDER_MODEL_BACKEND", defaults="armstrong.core.arm_layout.backends.BasicRenderModelBackend") .get_backend()) # Here for backwards compatibility (deprecated) get_layout_template_name = render_model.get_layout_template_name
import warnings from armstrong.utils.backends import GenericBackend render_model = (GenericBackend("ARMSTRONG_RENDER_MODEL_BACKEND", defaults="armstrong.core.arm_layout.backends.BasicRenderModelBackend") .get_backend()) # DEPRECATED: To be removed in ArmLayout 1.4. Here for backwards compatibility from django.utils.safestring import mark_safe from django.template.loader import render_to_string def deprecate(func): def wrapper(*args, **kwargs): msg = "Importing `{}` from this module is deprecated and will be removed in ArmLayout 1.4" warnings.warn(msg.format(func.__name__), DeprecationWarning) return func(*args, **kwargs) return wrapper mark_safe = deprecate(mark_safe) render_to_string = deprecate(render_to_string) get_layout_template_name = deprecate(render_model.get_layout_template_name)
Throw deprecation warnings for these imports, which will be removed in the next version. They've been deprecated for two years so we can make it happen.
Throw deprecation warnings for these imports, which will be removed in the next version. They've been deprecated for two years so we can make it happen.
Python
apache-2.0
armstrong/armstrong.core.arm_layout,armstrong/armstrong.core.arm_layout
# Here for backwards compatibility (deprecated) from django.utils.safestring import mark_safe from django.template.loader import render_to_string from armstrong.utils.backends import GenericBackend render_model = (GenericBackend("ARMSTRONG_RENDER_MODEL_BACKEND", defaults="armstrong.core.arm_layout.backends.BasicRenderModelBackend") .get_backend()) # Here for backwards compatibility (deprecated) get_layout_template_name = render_model.get_layout_template_name Throw deprecation warnings for these imports, which will be removed in the next version. They've been deprecated for two years so we can make it happen.
import warnings from armstrong.utils.backends import GenericBackend render_model = (GenericBackend("ARMSTRONG_RENDER_MODEL_BACKEND", defaults="armstrong.core.arm_layout.backends.BasicRenderModelBackend") .get_backend()) # DEPRECATED: To be removed in ArmLayout 1.4. Here for backwards compatibility from django.utils.safestring import mark_safe from django.template.loader import render_to_string def deprecate(func): def wrapper(*args, **kwargs): msg = "Importing `{}` from this module is deprecated and will be removed in ArmLayout 1.4" warnings.warn(msg.format(func.__name__), DeprecationWarning) return func(*args, **kwargs) return wrapper mark_safe = deprecate(mark_safe) render_to_string = deprecate(render_to_string) get_layout_template_name = deprecate(render_model.get_layout_template_name)
<commit_before># Here for backwards compatibility (deprecated) from django.utils.safestring import mark_safe from django.template.loader import render_to_string from armstrong.utils.backends import GenericBackend render_model = (GenericBackend("ARMSTRONG_RENDER_MODEL_BACKEND", defaults="armstrong.core.arm_layout.backends.BasicRenderModelBackend") .get_backend()) # Here for backwards compatibility (deprecated) get_layout_template_name = render_model.get_layout_template_name <commit_msg>Throw deprecation warnings for these imports, which will be removed in the next version. They've been deprecated for two years so we can make it happen.<commit_after>
import warnings from armstrong.utils.backends import GenericBackend render_model = (GenericBackend("ARMSTRONG_RENDER_MODEL_BACKEND", defaults="armstrong.core.arm_layout.backends.BasicRenderModelBackend") .get_backend()) # DEPRECATED: To be removed in ArmLayout 1.4. Here for backwards compatibility from django.utils.safestring import mark_safe from django.template.loader import render_to_string def deprecate(func): def wrapper(*args, **kwargs): msg = "Importing `{}` from this module is deprecated and will be removed in ArmLayout 1.4" warnings.warn(msg.format(func.__name__), DeprecationWarning) return func(*args, **kwargs) return wrapper mark_safe = deprecate(mark_safe) render_to_string = deprecate(render_to_string) get_layout_template_name = deprecate(render_model.get_layout_template_name)
# Here for backwards compatibility (deprecated) from django.utils.safestring import mark_safe from django.template.loader import render_to_string from armstrong.utils.backends import GenericBackend render_model = (GenericBackend("ARMSTRONG_RENDER_MODEL_BACKEND", defaults="armstrong.core.arm_layout.backends.BasicRenderModelBackend") .get_backend()) # Here for backwards compatibility (deprecated) get_layout_template_name = render_model.get_layout_template_name Throw deprecation warnings for these imports, which will be removed in the next version. They've been deprecated for two years so we can make it happen.import warnings from armstrong.utils.backends import GenericBackend render_model = (GenericBackend("ARMSTRONG_RENDER_MODEL_BACKEND", defaults="armstrong.core.arm_layout.backends.BasicRenderModelBackend") .get_backend()) # DEPRECATED: To be removed in ArmLayout 1.4. Here for backwards compatibility from django.utils.safestring import mark_safe from django.template.loader import render_to_string def deprecate(func): def wrapper(*args, **kwargs): msg = "Importing `{}` from this module is deprecated and will be removed in ArmLayout 1.4" warnings.warn(msg.format(func.__name__), DeprecationWarning) return func(*args, **kwargs) return wrapper mark_safe = deprecate(mark_safe) render_to_string = deprecate(render_to_string) get_layout_template_name = deprecate(render_model.get_layout_template_name)
<commit_before># Here for backwards compatibility (deprecated) from django.utils.safestring import mark_safe from django.template.loader import render_to_string from armstrong.utils.backends import GenericBackend render_model = (GenericBackend("ARMSTRONG_RENDER_MODEL_BACKEND", defaults="armstrong.core.arm_layout.backends.BasicRenderModelBackend") .get_backend()) # Here for backwards compatibility (deprecated) get_layout_template_name = render_model.get_layout_template_name <commit_msg>Throw deprecation warnings for these imports, which will be removed in the next version. They've been deprecated for two years so we can make it happen.<commit_after>import warnings from armstrong.utils.backends import GenericBackend render_model = (GenericBackend("ARMSTRONG_RENDER_MODEL_BACKEND", defaults="armstrong.core.arm_layout.backends.BasicRenderModelBackend") .get_backend()) # DEPRECATED: To be removed in ArmLayout 1.4. Here for backwards compatibility from django.utils.safestring import mark_safe from django.template.loader import render_to_string def deprecate(func): def wrapper(*args, **kwargs): msg = "Importing `{}` from this module is deprecated and will be removed in ArmLayout 1.4" warnings.warn(msg.format(func.__name__), DeprecationWarning) return func(*args, **kwargs) return wrapper mark_safe = deprecate(mark_safe) render_to_string = deprecate(render_to_string) get_layout_template_name = deprecate(render_model.get_layout_template_name)
4753dffc6a1672dfa99a5a5da8f082d6554bbb8f
http_request_translator/templates/bash_template.py
http_request_translator/templates/bash_template.py
begin_code = """ #!/usr/bin/env bash curl -s --request """ request_header = """ --header "{header} : {header_value}" """ code_search = " | egrep --color ' {search_string} |$' " code_simple = "{method} {url} {headers} --include " proxy_code = "-x {proxy}" body_code = " --data '{body}' "
begin_code = """ #!/usr/bin/env bash curl""" request_header = """ --header "{header} : {header_value}" """ code_search = " | egrep --color ' {search_string} |$'" code_simple = " -s --request {method} {url} {headers} --include" proxy_code = " -x {proxy}" body_code = " --data '{body}'"
Fix whitespace in bash script code template
Fix whitespace in bash script code template Signed-off-by: Arun Sori <e3bf7af6e125f7de61de92cd66a64411bed42bee@gmail.com>
Python
bsd-3-clause
owtf/http-request-translator,dhruvagarwal/http-request-translator
begin_code = """ #!/usr/bin/env bash curl -s --request """ request_header = """ --header "{header} : {header_value}" """ code_search = " | egrep --color ' {search_string} |$' " code_simple = "{method} {url} {headers} --include " proxy_code = "-x {proxy}" body_code = " --data '{body}' " Fix whitespace in bash script code template Signed-off-by: Arun Sori <e3bf7af6e125f7de61de92cd66a64411bed42bee@gmail.com>
begin_code = """ #!/usr/bin/env bash curl""" request_header = """ --header "{header} : {header_value}" """ code_search = " | egrep --color ' {search_string} |$'" code_simple = " -s --request {method} {url} {headers} --include" proxy_code = " -x {proxy}" body_code = " --data '{body}'"
<commit_before>begin_code = """ #!/usr/bin/env bash curl -s --request """ request_header = """ --header "{header} : {header_value}" """ code_search = " | egrep --color ' {search_string} |$' " code_simple = "{method} {url} {headers} --include " proxy_code = "-x {proxy}" body_code = " --data '{body}' " <commit_msg>Fix whitespace in bash script code template Signed-off-by: Arun Sori <e3bf7af6e125f7de61de92cd66a64411bed42bee@gmail.com><commit_after>
begin_code = """ #!/usr/bin/env bash curl""" request_header = """ --header "{header} : {header_value}" """ code_search = " | egrep --color ' {search_string} |$'" code_simple = " -s --request {method} {url} {headers} --include" proxy_code = " -x {proxy}" body_code = " --data '{body}'"
begin_code = """ #!/usr/bin/env bash curl -s --request """ request_header = """ --header "{header} : {header_value}" """ code_search = " | egrep --color ' {search_string} |$' " code_simple = "{method} {url} {headers} --include " proxy_code = "-x {proxy}" body_code = " --data '{body}' " Fix whitespace in bash script code template Signed-off-by: Arun Sori <e3bf7af6e125f7de61de92cd66a64411bed42bee@gmail.com>begin_code = """ #!/usr/bin/env bash curl""" request_header = """ --header "{header} : {header_value}" """ code_search = " | egrep --color ' {search_string} |$'" code_simple = " -s --request {method} {url} {headers} --include" proxy_code = " -x {proxy}" body_code = " --data '{body}'"
<commit_before>begin_code = """ #!/usr/bin/env bash curl -s --request """ request_header = """ --header "{header} : {header_value}" """ code_search = " | egrep --color ' {search_string} |$' " code_simple = "{method} {url} {headers} --include " proxy_code = "-x {proxy}" body_code = " --data '{body}' " <commit_msg>Fix whitespace in bash script code template Signed-off-by: Arun Sori <e3bf7af6e125f7de61de92cd66a64411bed42bee@gmail.com><commit_after>begin_code = """ #!/usr/bin/env bash curl""" request_header = """ --header "{header} : {header_value}" """ code_search = " | egrep --color ' {search_string} |$'" code_simple = " -s --request {method} {url} {headers} --include" proxy_code = " -x {proxy}" body_code = " --data '{body}'"
f1760fe01ae82289d8de2bb9323271edb80d4c08
f8a_jobs/graph_sync.py
f8a_jobs/graph_sync.py
"""Functions to retrieve pending list and invoke Graph Sync.""" import f8a_jobs.defaults as configuration import requests import traceback import logging logger = logging.getLogger(__name__) def _api_call(url, params={}): url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending") try: logger.info("API Call for url: %s, params: %s" % (url, params)) r = requests.get(url, params=params) if r is None: logger.error("Returned response is: %s" % r) raise Exception("Empty response found") result = {"data": r.json()} except Exception: logger.error(traceback.format_exc()) result = {"error": "Failed to retrieve data from Data Model Importer backend"} return result def fetch_pending(params={}): """Invoke Pending Graph Sync APIs for given parameters.""" url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending") return _api_call(url, params) def invoke_sync(params={}): """Invoke Graph Sync APIs to sync for given parameters.""" url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/sync_all") return _api_call(url, params)
"""Functions to retrieve pending list and invoke Graph Sync.""" import f8a_jobs.defaults as configuration import requests import traceback import logging logger = logging.getLogger(__name__) def _api_call(url, params={}): try: logger.info("API Call for url: %s, params: %s" % (url, params)) r = requests.get(url, params=params) if r is None: logger.error("Returned response is: %s" % r) raise Exception("Empty response found") result = {"data": r.json()} except Exception: logger.error(traceback.format_exc()) result = {"error": "Failed to retrieve data from Data Model Importer backend"} return result def fetch_pending(params={}): """Invoke Pending Graph Sync APIs for given parameters.""" url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending") return _api_call(url, params) def invoke_sync(params={}): """Invoke Graph Sync APIs to sync for given parameters.""" url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/sync_all") return _api_call(url, params)
Use url from the parameters
Use url from the parameters
Python
apache-2.0
fabric8-analytics/fabric8-analytics-jobs,fabric8-analytics/fabric8-analytics-jobs
"""Functions to retrieve pending list and invoke Graph Sync.""" import f8a_jobs.defaults as configuration import requests import traceback import logging logger = logging.getLogger(__name__) def _api_call(url, params={}): url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending") try: logger.info("API Call for url: %s, params: %s" % (url, params)) r = requests.get(url, params=params) if r is None: logger.error("Returned response is: %s" % r) raise Exception("Empty response found") result = {"data": r.json()} except Exception: logger.error(traceback.format_exc()) result = {"error": "Failed to retrieve data from Data Model Importer backend"} return result def fetch_pending(params={}): """Invoke Pending Graph Sync APIs for given parameters.""" url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending") return _api_call(url, params) def invoke_sync(params={}): """Invoke Graph Sync APIs to sync for given parameters.""" url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/sync_all") return _api_call(url, params) Use url from the parameters
"""Functions to retrieve pending list and invoke Graph Sync.""" import f8a_jobs.defaults as configuration import requests import traceback import logging logger = logging.getLogger(__name__) def _api_call(url, params={}): try: logger.info("API Call for url: %s, params: %s" % (url, params)) r = requests.get(url, params=params) if r is None: logger.error("Returned response is: %s" % r) raise Exception("Empty response found") result = {"data": r.json()} except Exception: logger.error(traceback.format_exc()) result = {"error": "Failed to retrieve data from Data Model Importer backend"} return result def fetch_pending(params={}): """Invoke Pending Graph Sync APIs for given parameters.""" url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending") return _api_call(url, params) def invoke_sync(params={}): """Invoke Graph Sync APIs to sync for given parameters.""" url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/sync_all") return _api_call(url, params)
<commit_before>"""Functions to retrieve pending list and invoke Graph Sync.""" import f8a_jobs.defaults as configuration import requests import traceback import logging logger = logging.getLogger(__name__) def _api_call(url, params={}): url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending") try: logger.info("API Call for url: %s, params: %s" % (url, params)) r = requests.get(url, params=params) if r is None: logger.error("Returned response is: %s" % r) raise Exception("Empty response found") result = {"data": r.json()} except Exception: logger.error(traceback.format_exc()) result = {"error": "Failed to retrieve data from Data Model Importer backend"} return result def fetch_pending(params={}): """Invoke Pending Graph Sync APIs for given parameters.""" url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending") return _api_call(url, params) def invoke_sync(params={}): """Invoke Graph Sync APIs to sync for given parameters.""" url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/sync_all") return _api_call(url, params) <commit_msg>Use url from the parameters<commit_after>
"""Functions to retrieve pending list and invoke Graph Sync.""" import f8a_jobs.defaults as configuration import requests import traceback import logging logger = logging.getLogger(__name__) def _api_call(url, params={}): try: logger.info("API Call for url: %s, params: %s" % (url, params)) r = requests.get(url, params=params) if r is None: logger.error("Returned response is: %s" % r) raise Exception("Empty response found") result = {"data": r.json()} except Exception: logger.error(traceback.format_exc()) result = {"error": "Failed to retrieve data from Data Model Importer backend"} return result def fetch_pending(params={}): """Invoke Pending Graph Sync APIs for given parameters.""" url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending") return _api_call(url, params) def invoke_sync(params={}): """Invoke Graph Sync APIs to sync for given parameters.""" url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/sync_all") return _api_call(url, params)
"""Functions to retrieve pending list and invoke Graph Sync.""" import f8a_jobs.defaults as configuration import requests import traceback import logging logger = logging.getLogger(__name__) def _api_call(url, params={}): url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending") try: logger.info("API Call for url: %s, params: %s" % (url, params)) r = requests.get(url, params=params) if r is None: logger.error("Returned response is: %s" % r) raise Exception("Empty response found") result = {"data": r.json()} except Exception: logger.error(traceback.format_exc()) result = {"error": "Failed to retrieve data from Data Model Importer backend"} return result def fetch_pending(params={}): """Invoke Pending Graph Sync APIs for given parameters.""" url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending") return _api_call(url, params) def invoke_sync(params={}): """Invoke Graph Sync APIs to sync for given parameters.""" url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/sync_all") return _api_call(url, params) Use url from the parameters"""Functions to retrieve pending list and invoke Graph Sync.""" import f8a_jobs.defaults as configuration import requests import traceback import logging logger = logging.getLogger(__name__) def _api_call(url, params={}): try: logger.info("API Call for url: %s, params: %s" % (url, params)) r = requests.get(url, params=params) if r is None: logger.error("Returned response is: %s" % r) raise Exception("Empty response found") result = {"data": r.json()} except Exception: logger.error(traceback.format_exc()) result = {"error": "Failed to retrieve data from Data Model Importer backend"} return result def fetch_pending(params={}): """Invoke Pending Graph Sync APIs for given parameters.""" url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending") return _api_call(url, params) def invoke_sync(params={}): """Invoke Graph Sync APIs to sync for given parameters.""" url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/sync_all") return _api_call(url, params)
<commit_before>"""Functions to retrieve pending list and invoke Graph Sync.""" import f8a_jobs.defaults as configuration import requests import traceback import logging logger = logging.getLogger(__name__) def _api_call(url, params={}): url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending") try: logger.info("API Call for url: %s, params: %s" % (url, params)) r = requests.get(url, params=params) if r is None: logger.error("Returned response is: %s" % r) raise Exception("Empty response found") result = {"data": r.json()} except Exception: logger.error(traceback.format_exc()) result = {"error": "Failed to retrieve data from Data Model Importer backend"} return result def fetch_pending(params={}): """Invoke Pending Graph Sync APIs for given parameters.""" url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending") return _api_call(url, params) def invoke_sync(params={}): """Invoke Graph Sync APIs to sync for given parameters.""" url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/sync_all") return _api_call(url, params) <commit_msg>Use url from the parameters<commit_after>"""Functions to retrieve pending list and invoke Graph Sync.""" import f8a_jobs.defaults as configuration import requests import traceback import logging logger = logging.getLogger(__name__) def _api_call(url, params={}): try: logger.info("API Call for url: %s, params: %s" % (url, params)) r = requests.get(url, params=params) if r is None: logger.error("Returned response is: %s" % r) raise Exception("Empty response found") result = {"data": r.json()} except Exception: logger.error(traceback.format_exc()) result = {"error": "Failed to retrieve data from Data Model Importer backend"} return result def fetch_pending(params={}): """Invoke Pending Graph Sync APIs for given parameters.""" url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending") return _api_call(url, params) def invoke_sync(params={}): """Invoke Graph Sync APIs to sync for given parameters.""" url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/sync_all") return _api_call(url, params)
d3af229c5c692fdb52c211cd8785bcb7c869090b
reobject/query.py
reobject/query.py
from reobject.utils import signed_attrgetter class QuerySet(list): def __init__(self, *args, **kwargs): super(QuerySet, self).__init__(*args, **kwargs) def count(self): return len(self) def delete(self): for item in self: item.delete() def exists(self): return bool(self) def order_by(self, *args): return type(self)( sorted(self, key=signed_attrgetter(*args)) )
from reobject.utils import signed_attrgetter class QuerySet(list): def __init__(self, *args, **kwargs): super(QuerySet, self).__init__(*args, **kwargs) def count(self): return len(self) def delete(self): for item in self: item.delete() def exists(self): return bool(self) def order_by(self, *args): return type(self)( sorted(self, key=signed_attrgetter(*args)) ) def reverse(self): return type(self)( reversed(self) )
Allow QuerySet objects to be reversed
Allow QuerySet objects to be reversed
Python
apache-2.0
onyb/reobject,onyb/reobject
from reobject.utils import signed_attrgetter class QuerySet(list): def __init__(self, *args, **kwargs): super(QuerySet, self).__init__(*args, **kwargs) def count(self): return len(self) def delete(self): for item in self: item.delete() def exists(self): return bool(self) def order_by(self, *args): return type(self)( sorted(self, key=signed_attrgetter(*args)) ) Allow QuerySet objects to be reversed
from reobject.utils import signed_attrgetter class QuerySet(list): def __init__(self, *args, **kwargs): super(QuerySet, self).__init__(*args, **kwargs) def count(self): return len(self) def delete(self): for item in self: item.delete() def exists(self): return bool(self) def order_by(self, *args): return type(self)( sorted(self, key=signed_attrgetter(*args)) ) def reverse(self): return type(self)( reversed(self) )
<commit_before>from reobject.utils import signed_attrgetter class QuerySet(list): def __init__(self, *args, **kwargs): super(QuerySet, self).__init__(*args, **kwargs) def count(self): return len(self) def delete(self): for item in self: item.delete() def exists(self): return bool(self) def order_by(self, *args): return type(self)( sorted(self, key=signed_attrgetter(*args)) ) <commit_msg>Allow QuerySet objects to be reversed<commit_after>
from reobject.utils import signed_attrgetter class QuerySet(list): def __init__(self, *args, **kwargs): super(QuerySet, self).__init__(*args, **kwargs) def count(self): return len(self) def delete(self): for item in self: item.delete() def exists(self): return bool(self) def order_by(self, *args): return type(self)( sorted(self, key=signed_attrgetter(*args)) ) def reverse(self): return type(self)( reversed(self) )
from reobject.utils import signed_attrgetter class QuerySet(list): def __init__(self, *args, **kwargs): super(QuerySet, self).__init__(*args, **kwargs) def count(self): return len(self) def delete(self): for item in self: item.delete() def exists(self): return bool(self) def order_by(self, *args): return type(self)( sorted(self, key=signed_attrgetter(*args)) ) Allow QuerySet objects to be reversedfrom reobject.utils import signed_attrgetter class QuerySet(list): def __init__(self, *args, **kwargs): super(QuerySet, self).__init__(*args, **kwargs) def count(self): return len(self) def delete(self): for item in self: item.delete() def exists(self): return bool(self) def order_by(self, *args): return type(self)( sorted(self, key=signed_attrgetter(*args)) ) def reverse(self): return type(self)( reversed(self) )
<commit_before>from reobject.utils import signed_attrgetter class QuerySet(list): def __init__(self, *args, **kwargs): super(QuerySet, self).__init__(*args, **kwargs) def count(self): return len(self) def delete(self): for item in self: item.delete() def exists(self): return bool(self) def order_by(self, *args): return type(self)( sorted(self, key=signed_attrgetter(*args)) ) <commit_msg>Allow QuerySet objects to be reversed<commit_after>from reobject.utils import signed_attrgetter class QuerySet(list): def __init__(self, *args, **kwargs): super(QuerySet, self).__init__(*args, **kwargs) def count(self): return len(self) def delete(self): for item in self: item.delete() def exists(self): return bool(self) def order_by(self, *args): return type(self)( sorted(self, key=signed_attrgetter(*args)) ) def reverse(self): return type(self)( reversed(self) )
06c5f27c04de9fa62f6ac4834e0a920349c27084
rules/binutils.py
rules/binutils.py
import xyz import os import shutil class Binutils(xyz.BuildProtocol): pkg_name = 'binutils' supported_targets = ['arm-none-eabi'] def check(self, builder): if builder.target not in self.supported_targets: raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name)) def configure(self, builder, config): builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib', config=config) def install(self, builder, config): super().install(builder, config) # For some reason binutils plonks libiberty.a in the output directory libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config) if os.path.exists(libdir): shutil.rmtree(libdir) rules = Binutils()
import xyz import os import shutil class Binutils(xyz.BuildProtocol): pkg_name = 'binutils' supported_targets = ['arm-none-eabi'] def check(self, builder): if builder.target not in self.supported_targets: raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name)) def configure(self, builder, config): builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib', config=config) def install(self, builder, config): super().install(builder, config) # For some reason binutils plonks libiberty.a in the output directory libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config) if os.path.exists(libdir): shutil.rmtree(libdir) # For now we strip the man pages. # man pages created on different systems are (for no good reason) different! man_dir = builder.j('{install_dir}', config['prefix'][1:], 'share', 'man', config=config) shutil.rmtree(man_dir) rules = Binutils()
Remove man pages post-install (for now)
Remove man pages post-install (for now)
Python
mit
BreakawayConsulting/xyz
import xyz import os import shutil class Binutils(xyz.BuildProtocol): pkg_name = 'binutils' supported_targets = ['arm-none-eabi'] def check(self, builder): if builder.target not in self.supported_targets: raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name)) def configure(self, builder, config): builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib', config=config) def install(self, builder, config): super().install(builder, config) # For some reason binutils plonks libiberty.a in the output directory libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config) if os.path.exists(libdir): shutil.rmtree(libdir) rules = Binutils() Remove man pages post-install (for now)
import xyz import os import shutil class Binutils(xyz.BuildProtocol): pkg_name = 'binutils' supported_targets = ['arm-none-eabi'] def check(self, builder): if builder.target not in self.supported_targets: raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name)) def configure(self, builder, config): builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib', config=config) def install(self, builder, config): super().install(builder, config) # For some reason binutils plonks libiberty.a in the output directory libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config) if os.path.exists(libdir): shutil.rmtree(libdir) # For now we strip the man pages. # man pages created on different systems are (for no good reason) different! man_dir = builder.j('{install_dir}', config['prefix'][1:], 'share', 'man', config=config) shutil.rmtree(man_dir) rules = Binutils()
<commit_before>import xyz import os import shutil class Binutils(xyz.BuildProtocol): pkg_name = 'binutils' supported_targets = ['arm-none-eabi'] def check(self, builder): if builder.target not in self.supported_targets: raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name)) def configure(self, builder, config): builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib', config=config) def install(self, builder, config): super().install(builder, config) # For some reason binutils plonks libiberty.a in the output directory libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config) if os.path.exists(libdir): shutil.rmtree(libdir) rules = Binutils() <commit_msg>Remove man pages post-install (for now)<commit_after>
import xyz import os import shutil class Binutils(xyz.BuildProtocol): pkg_name = 'binutils' supported_targets = ['arm-none-eabi'] def check(self, builder): if builder.target not in self.supported_targets: raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name)) def configure(self, builder, config): builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib', config=config) def install(self, builder, config): super().install(builder, config) # For some reason binutils plonks libiberty.a in the output directory libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config) if os.path.exists(libdir): shutil.rmtree(libdir) # For now we strip the man pages. # man pages created on different systems are (for no good reason) different! man_dir = builder.j('{install_dir}', config['prefix'][1:], 'share', 'man', config=config) shutil.rmtree(man_dir) rules = Binutils()
import xyz import os import shutil class Binutils(xyz.BuildProtocol): pkg_name = 'binutils' supported_targets = ['arm-none-eabi'] def check(self, builder): if builder.target not in self.supported_targets: raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name)) def configure(self, builder, config): builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib', config=config) def install(self, builder, config): super().install(builder, config) # For some reason binutils plonks libiberty.a in the output directory libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config) if os.path.exists(libdir): shutil.rmtree(libdir) rules = Binutils() Remove man pages post-install (for now)import xyz import os import shutil class Binutils(xyz.BuildProtocol): pkg_name = 'binutils' supported_targets = ['arm-none-eabi'] def check(self, builder): if builder.target not in self.supported_targets: raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name)) def configure(self, builder, config): builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib', config=config) def install(self, builder, config): super().install(builder, config) # For some reason binutils plonks libiberty.a in the output directory libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config) if os.path.exists(libdir): shutil.rmtree(libdir) # For now we strip the man pages. # man pages created on different systems are (for no good reason) different! man_dir = builder.j('{install_dir}', config['prefix'][1:], 'share', 'man', config=config) shutil.rmtree(man_dir) rules = Binutils()
<commit_before>import xyz import os import shutil class Binutils(xyz.BuildProtocol): pkg_name = 'binutils' supported_targets = ['arm-none-eabi'] def check(self, builder): if builder.target not in self.supported_targets: raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name)) def configure(self, builder, config): builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib', config=config) def install(self, builder, config): super().install(builder, config) # For some reason binutils plonks libiberty.a in the output directory libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config) if os.path.exists(libdir): shutil.rmtree(libdir) rules = Binutils() <commit_msg>Remove man pages post-install (for now)<commit_after>import xyz import os import shutil class Binutils(xyz.BuildProtocol): pkg_name = 'binutils' supported_targets = ['arm-none-eabi'] def check(self, builder): if builder.target not in self.supported_targets: raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name)) def configure(self, builder, config): builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib', config=config) def install(self, builder, config): super().install(builder, config) # For some reason binutils plonks libiberty.a in the output directory libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config) if os.path.exists(libdir): shutil.rmtree(libdir) # For now we strip the man pages. # man pages created on different systems are (for no good reason) different! man_dir = builder.j('{install_dir}', config['prefix'][1:], 'share', 'man', config=config) shutil.rmtree(man_dir) rules = Binutils()
7a5f2d0397f8ecda1c1b0517e844eec9d0e3e9d4
geotrek/common/urls.py
geotrek/common/urls.py
from django.conf.urls import patterns, url from .views import settings_json urlpatterns += patterns('', url(r'^api/settings.json', settings_json, name='settings_json'), )
from django.conf.urls import patterns, url from .views import settings_json urlpatterns = patterns('', url(r'^api/settings.json', settings_json, name='settings_json'), )
Fix URL pattern after removing
Fix URL pattern after removing
Python
bsd-2-clause
makinacorpus/Geotrek,mabhub/Geotrek,GeotrekCE/Geotrek-admin,makinacorpus/Geotrek,Anaethelion/Geotrek,johan--/Geotrek,Anaethelion/Geotrek,Anaethelion/Geotrek,GeotrekCE/Geotrek-admin,mabhub/Geotrek,GeotrekCE/Geotrek-admin,camillemonchicourt/Geotrek,makinacorpus/Geotrek,makinacorpus/Geotrek,mabhub/Geotrek,johan--/Geotrek,Anaethelion/Geotrek,johan--/Geotrek,johan--/Geotrek,camillemonchicourt/Geotrek,mabhub/Geotrek,camillemonchicourt/Geotrek,GeotrekCE/Geotrek-admin
from django.conf.urls import patterns, url from .views import settings_json urlpatterns += patterns('', url(r'^api/settings.json', settings_json, name='settings_json'), ) Fix URL pattern after removing
from django.conf.urls import patterns, url from .views import settings_json urlpatterns = patterns('', url(r'^api/settings.json', settings_json, name='settings_json'), )
<commit_before>from django.conf.urls import patterns, url from .views import settings_json urlpatterns += patterns('', url(r'^api/settings.json', settings_json, name='settings_json'), ) <commit_msg>Fix URL pattern after removing<commit_after>
from django.conf.urls import patterns, url from .views import settings_json urlpatterns = patterns('', url(r'^api/settings.json', settings_json, name='settings_json'), )
from django.conf.urls import patterns, url from .views import settings_json urlpatterns += patterns('', url(r'^api/settings.json', settings_json, name='settings_json'), ) Fix URL pattern after removingfrom django.conf.urls import patterns, url from .views import settings_json urlpatterns = patterns('', url(r'^api/settings.json', settings_json, name='settings_json'), )
<commit_before>from django.conf.urls import patterns, url from .views import settings_json urlpatterns += patterns('', url(r'^api/settings.json', settings_json, name='settings_json'), ) <commit_msg>Fix URL pattern after removing<commit_after>from django.conf.urls import patterns, url from .views import settings_json urlpatterns = patterns('', url(r'^api/settings.json', settings_json, name='settings_json'), )
9c0d88ba1681949c02f2cd136efc0de1c23d170d
simuvex/procedures/libc___so___6/fileno.py
simuvex/procedures/libc___so___6/fileno.py
import simuvex from simuvex.s_type import SimTypeFd import logging l = logging.getLogger("simuvex.procedures.fileno") ###################################### # memset ###################################### class fileno(simuvex.SimProcedure): #pylint:disable=arguments-differ def run(self, f): self.argument_types = {0: SimTypeFd()} self.return_type = SimTypeFd() return f
import simuvex from simuvex.s_type import SimTypeFd, SimTypeTop from . import io_file_data_for_arch import logging l = logging.getLogger("simuvex.procedures.fileno") ###################################### # fileno ###################################### class fileno(simuvex.SimProcedure): #pylint:disable=arguments-differ def run(self, f): self.argument_types = {0: self.ty_ptr(SimTypeTop())} self.return_type = SimTypeFd() # Get FILE struct io_file_data = io_file_data_for_arch(self.state.arch) # Get the file descriptor from FILE struct fd = self.state.se.any_int(self.state.memory.load(f + io_file_data['fd'], 4 * 8, # int endness=self.state.arch.memory_endness)) return fd
Add logic for grabbing file descriptor from FILE struct
Add logic for grabbing file descriptor from FILE struct
Python
bsd-2-clause
chubbymaggie/angr,angr/angr,schieb/angr,tyb0807/angr,axt/angr,f-prettyland/angr,axt/angr,iamahuman/angr,angr/angr,iamahuman/angr,axt/angr,iamahuman/angr,f-prettyland/angr,chubbymaggie/angr,schieb/angr,angr/angr,tyb0807/angr,tyb0807/angr,chubbymaggie/angr,f-prettyland/angr,angr/simuvex,schieb/angr
import simuvex from simuvex.s_type import SimTypeFd import logging l = logging.getLogger("simuvex.procedures.fileno") ###################################### # memset ###################################### class fileno(simuvex.SimProcedure): #pylint:disable=arguments-differ def run(self, f): self.argument_types = {0: SimTypeFd()} self.return_type = SimTypeFd() return f Add logic for grabbing file descriptor from FILE struct
import simuvex from simuvex.s_type import SimTypeFd, SimTypeTop from . import io_file_data_for_arch import logging l = logging.getLogger("simuvex.procedures.fileno") ###################################### # fileno ###################################### class fileno(simuvex.SimProcedure): #pylint:disable=arguments-differ def run(self, f): self.argument_types = {0: self.ty_ptr(SimTypeTop())} self.return_type = SimTypeFd() # Get FILE struct io_file_data = io_file_data_for_arch(self.state.arch) # Get the file descriptor from FILE struct fd = self.state.se.any_int(self.state.memory.load(f + io_file_data['fd'], 4 * 8, # int endness=self.state.arch.memory_endness)) return fd
<commit_before>import simuvex from simuvex.s_type import SimTypeFd import logging l = logging.getLogger("simuvex.procedures.fileno") ###################################### # memset ###################################### class fileno(simuvex.SimProcedure): #pylint:disable=arguments-differ def run(self, f): self.argument_types = {0: SimTypeFd()} self.return_type = SimTypeFd() return f <commit_msg>Add logic for grabbing file descriptor from FILE struct<commit_after>
import simuvex from simuvex.s_type import SimTypeFd, SimTypeTop from . import io_file_data_for_arch import logging l = logging.getLogger("simuvex.procedures.fileno") ###################################### # fileno ###################################### class fileno(simuvex.SimProcedure): #pylint:disable=arguments-differ def run(self, f): self.argument_types = {0: self.ty_ptr(SimTypeTop())} self.return_type = SimTypeFd() # Get FILE struct io_file_data = io_file_data_for_arch(self.state.arch) # Get the file descriptor from FILE struct fd = self.state.se.any_int(self.state.memory.load(f + io_file_data['fd'], 4 * 8, # int endness=self.state.arch.memory_endness)) return fd
import simuvex from simuvex.s_type import SimTypeFd import logging l = logging.getLogger("simuvex.procedures.fileno") ###################################### # memset ###################################### class fileno(simuvex.SimProcedure): #pylint:disable=arguments-differ def run(self, f): self.argument_types = {0: SimTypeFd()} self.return_type = SimTypeFd() return f Add logic for grabbing file descriptor from FILE structimport simuvex from simuvex.s_type import SimTypeFd, SimTypeTop from . import io_file_data_for_arch import logging l = logging.getLogger("simuvex.procedures.fileno") ###################################### # fileno ###################################### class fileno(simuvex.SimProcedure): #pylint:disable=arguments-differ def run(self, f): self.argument_types = {0: self.ty_ptr(SimTypeTop())} self.return_type = SimTypeFd() # Get FILE struct io_file_data = io_file_data_for_arch(self.state.arch) # Get the file descriptor from FILE struct fd = self.state.se.any_int(self.state.memory.load(f + io_file_data['fd'], 4 * 8, # int endness=self.state.arch.memory_endness)) return fd
<commit_before>import simuvex from simuvex.s_type import SimTypeFd import logging l = logging.getLogger("simuvex.procedures.fileno") ###################################### # memset ###################################### class fileno(simuvex.SimProcedure): #pylint:disable=arguments-differ def run(self, f): self.argument_types = {0: SimTypeFd()} self.return_type = SimTypeFd() return f <commit_msg>Add logic for grabbing file descriptor from FILE struct<commit_after>import simuvex from simuvex.s_type import SimTypeFd, SimTypeTop from . import io_file_data_for_arch import logging l = logging.getLogger("simuvex.procedures.fileno") ###################################### # fileno ###################################### class fileno(simuvex.SimProcedure): #pylint:disable=arguments-differ def run(self, f): self.argument_types = {0: self.ty_ptr(SimTypeTop())} self.return_type = SimTypeFd() # Get FILE struct io_file_data = io_file_data_for_arch(self.state.arch) # Get the file descriptor from FILE struct fd = self.state.se.any_int(self.state.memory.load(f + io_file_data['fd'], 4 * 8, # int endness=self.state.arch.memory_endness)) return fd
d387ab236634f91186805dd114ee85455d1244f8
pywikibot/echo.py
pywikibot/echo.py
# -*- coding: utf-8 -*- """Classes and functions for working with the Echo extension.""" from __future__ import absolute_import, unicode_literals import pywikibot class Notification(object): """A notification issued by the Echo extension.""" def __init__(self, site): """Construct an empty Notification object.""" self.site = site @classmethod def fromJSON(cls, site, data): """ Construct a Notification object from JSON data returned by the API. @rtype: Notification """ notif = cls(site) notif.id = data['id'] # TODO: use numeric id ? notif.type = data['type'] notif.category = data['category'] notif.timestamp = pywikibot.Timestamp.fromtimestampformat(data['timestamp']['mw']) # TODO: use 'namespace-key' + 'text' ? notif.page = pywikibot.Page(site, data['title']['full']) if 'agent' in data and 'name' in data['agent']: notif.agent = pywikibot.User(site, data['agent']['name']) else: notif.agent = None if 'read' in data: notif.read = pywikibot.Timestamp.fromtimestampformat(data['read']) else: notif.read = False notif.content = data.get('*', None) return notif def mark_as_read(self): """Mark the notification as read.""" return self.site.notifications_mark_read(list=self.id)
# -*- coding: utf-8 -*- """Classes and functions for working with the Echo extension.""" from __future__ import absolute_import, unicode_literals import pywikibot class Notification(object): """A notification issued by the Echo extension.""" def __init__(self, site): """Construct an empty Notification object.""" self.site = site @classmethod def fromJSON(cls, site, data): """ Construct a Notification object from JSON data returned by the API. @rtype: Notification """ notif = cls(site) notif.id = data['id'] # TODO: use numeric id ? notif.type = data['type'] notif.category = data['category'] notif.timestamp = pywikibot.Timestamp.fromtimestampformat(data['timestamp']['mw']) if 'title' in data and 'full' in data['title']: notif.page = pywikibot.Page(site, data['title']['full']) else: notif.page = None if 'agent' in data and 'name' in data['agent']: notif.agent = pywikibot.User(site, data['agent']['name']) else: notif.agent = None if 'read' in data: notif.read = pywikibot.Timestamp.fromtimestampformat(data['read']) else: notif.read = False notif.content = data.get('*', None) return notif def mark_as_read(self): """Mark the notification as read.""" return self.site.notifications_mark_read(list=self.id)
Fix notifications building from JSON
Fix notifications building from JSON Sometimes (like in welcome messages), notifications don't have a 'title' property, so we shouldn't assume there is one. Bug: T139015 Change-Id: I83e480d04e8e09aa9bcb5edef4f56b47d150e199
Python
mit
magul/pywikibot-core,hasteur/g13bot_tools_new,PersianWikipedia/pywikibot-core,npdoty/pywikibot,npdoty/pywikibot,wikimedia/pywikibot-core,happy5214/pywikibot-core,happy5214/pywikibot-core,wikimedia/pywikibot-core,Darkdadaah/pywikibot-core,jayvdb/pywikibot-core,hasteur/g13bot_tools_new,jayvdb/pywikibot-core,hasteur/g13bot_tools_new,magul/pywikibot-core,Darkdadaah/pywikibot-core
# -*- coding: utf-8 -*- """Classes and functions for working with the Echo extension.""" from __future__ import absolute_import, unicode_literals import pywikibot class Notification(object): """A notification issued by the Echo extension.""" def __init__(self, site): """Construct an empty Notification object.""" self.site = site @classmethod def fromJSON(cls, site, data): """ Construct a Notification object from JSON data returned by the API. @rtype: Notification """ notif = cls(site) notif.id = data['id'] # TODO: use numeric id ? notif.type = data['type'] notif.category = data['category'] notif.timestamp = pywikibot.Timestamp.fromtimestampformat(data['timestamp']['mw']) # TODO: use 'namespace-key' + 'text' ? notif.page = pywikibot.Page(site, data['title']['full']) if 'agent' in data and 'name' in data['agent']: notif.agent = pywikibot.User(site, data['agent']['name']) else: notif.agent = None if 'read' in data: notif.read = pywikibot.Timestamp.fromtimestampformat(data['read']) else: notif.read = False notif.content = data.get('*', None) return notif def mark_as_read(self): """Mark the notification as read.""" return self.site.notifications_mark_read(list=self.id) Fix notifications building from JSON Sometimes (like in welcome messages), notifications don't have a 'title' property, so we shouldn't assume there is one. Bug: T139015 Change-Id: I83e480d04e8e09aa9bcb5edef4f56b47d150e199
# -*- coding: utf-8 -*- """Classes and functions for working with the Echo extension.""" from __future__ import absolute_import, unicode_literals import pywikibot class Notification(object): """A notification issued by the Echo extension.""" def __init__(self, site): """Construct an empty Notification object.""" self.site = site @classmethod def fromJSON(cls, site, data): """ Construct a Notification object from JSON data returned by the API. @rtype: Notification """ notif = cls(site) notif.id = data['id'] # TODO: use numeric id ? notif.type = data['type'] notif.category = data['category'] notif.timestamp = pywikibot.Timestamp.fromtimestampformat(data['timestamp']['mw']) if 'title' in data and 'full' in data['title']: notif.page = pywikibot.Page(site, data['title']['full']) else: notif.page = None if 'agent' in data and 'name' in data['agent']: notif.agent = pywikibot.User(site, data['agent']['name']) else: notif.agent = None if 'read' in data: notif.read = pywikibot.Timestamp.fromtimestampformat(data['read']) else: notif.read = False notif.content = data.get('*', None) return notif def mark_as_read(self): """Mark the notification as read.""" return self.site.notifications_mark_read(list=self.id)
<commit_before># -*- coding: utf-8 -*- """Classes and functions for working with the Echo extension.""" from __future__ import absolute_import, unicode_literals import pywikibot class Notification(object): """A notification issued by the Echo extension.""" def __init__(self, site): """Construct an empty Notification object.""" self.site = site @classmethod def fromJSON(cls, site, data): """ Construct a Notification object from JSON data returned by the API. @rtype: Notification """ notif = cls(site) notif.id = data['id'] # TODO: use numeric id ? notif.type = data['type'] notif.category = data['category'] notif.timestamp = pywikibot.Timestamp.fromtimestampformat(data['timestamp']['mw']) # TODO: use 'namespace-key' + 'text' ? notif.page = pywikibot.Page(site, data['title']['full']) if 'agent' in data and 'name' in data['agent']: notif.agent = pywikibot.User(site, data['agent']['name']) else: notif.agent = None if 'read' in data: notif.read = pywikibot.Timestamp.fromtimestampformat(data['read']) else: notif.read = False notif.content = data.get('*', None) return notif def mark_as_read(self): """Mark the notification as read.""" return self.site.notifications_mark_read(list=self.id) <commit_msg>Fix notifications building from JSON Sometimes (like in welcome messages), notifications don't have a 'title' property, so we shouldn't assume there is one. Bug: T139015 Change-Id: I83e480d04e8e09aa9bcb5edef4f56b47d150e199<commit_after>
# -*- coding: utf-8 -*- """Classes and functions for working with the Echo extension.""" from __future__ import absolute_import, unicode_literals import pywikibot class Notification(object): """A notification issued by the Echo extension.""" def __init__(self, site): """Construct an empty Notification object.""" self.site = site @classmethod def fromJSON(cls, site, data): """ Construct a Notification object from JSON data returned by the API. @rtype: Notification """ notif = cls(site) notif.id = data['id'] # TODO: use numeric id ? notif.type = data['type'] notif.category = data['category'] notif.timestamp = pywikibot.Timestamp.fromtimestampformat(data['timestamp']['mw']) if 'title' in data and 'full' in data['title']: notif.page = pywikibot.Page(site, data['title']['full']) else: notif.page = None if 'agent' in data and 'name' in data['agent']: notif.agent = pywikibot.User(site, data['agent']['name']) else: notif.agent = None if 'read' in data: notif.read = pywikibot.Timestamp.fromtimestampformat(data['read']) else: notif.read = False notif.content = data.get('*', None) return notif def mark_as_read(self): """Mark the notification as read.""" return self.site.notifications_mark_read(list=self.id)
# -*- coding: utf-8 -*- """Classes and functions for working with the Echo extension.""" from __future__ import absolute_import, unicode_literals import pywikibot class Notification(object): """A notification issued by the Echo extension.""" def __init__(self, site): """Construct an empty Notification object.""" self.site = site @classmethod def fromJSON(cls, site, data): """ Construct a Notification object from JSON data returned by the API. @rtype: Notification """ notif = cls(site) notif.id = data['id'] # TODO: use numeric id ? notif.type = data['type'] notif.category = data['category'] notif.timestamp = pywikibot.Timestamp.fromtimestampformat(data['timestamp']['mw']) # TODO: use 'namespace-key' + 'text' ? notif.page = pywikibot.Page(site, data['title']['full']) if 'agent' in data and 'name' in data['agent']: notif.agent = pywikibot.User(site, data['agent']['name']) else: notif.agent = None if 'read' in data: notif.read = pywikibot.Timestamp.fromtimestampformat(data['read']) else: notif.read = False notif.content = data.get('*', None) return notif def mark_as_read(self): """Mark the notification as read.""" return self.site.notifications_mark_read(list=self.id) Fix notifications building from JSON Sometimes (like in welcome messages), notifications don't have a 'title' property, so we shouldn't assume there is one. Bug: T139015 Change-Id: I83e480d04e8e09aa9bcb5edef4f56b47d150e199# -*- coding: utf-8 -*- """Classes and functions for working with the Echo extension.""" from __future__ import absolute_import, unicode_literals import pywikibot class Notification(object): """A notification issued by the Echo extension.""" def __init__(self, site): """Construct an empty Notification object.""" self.site = site @classmethod def fromJSON(cls, site, data): """ Construct a Notification object from JSON data returned by the API. @rtype: Notification """ notif = cls(site) notif.id = data['id'] # TODO: use numeric id ? notif.type = data['type'] notif.category = data['category'] notif.timestamp = pywikibot.Timestamp.fromtimestampformat(data['timestamp']['mw']) if 'title' in data and 'full' in data['title']: notif.page = pywikibot.Page(site, data['title']['full']) else: notif.page = None if 'agent' in data and 'name' in data['agent']: notif.agent = pywikibot.User(site, data['agent']['name']) else: notif.agent = None if 'read' in data: notif.read = pywikibot.Timestamp.fromtimestampformat(data['read']) else: notif.read = False notif.content = data.get('*', None) return notif def mark_as_read(self): """Mark the notification as read.""" return self.site.notifications_mark_read(list=self.id)
<commit_before># -*- coding: utf-8 -*- """Classes and functions for working with the Echo extension.""" from __future__ import absolute_import, unicode_literals import pywikibot class Notification(object): """A notification issued by the Echo extension.""" def __init__(self, site): """Construct an empty Notification object.""" self.site = site @classmethod def fromJSON(cls, site, data): """ Construct a Notification object from JSON data returned by the API. @rtype: Notification """ notif = cls(site) notif.id = data['id'] # TODO: use numeric id ? notif.type = data['type'] notif.category = data['category'] notif.timestamp = pywikibot.Timestamp.fromtimestampformat(data['timestamp']['mw']) # TODO: use 'namespace-key' + 'text' ? notif.page = pywikibot.Page(site, data['title']['full']) if 'agent' in data and 'name' in data['agent']: notif.agent = pywikibot.User(site, data['agent']['name']) else: notif.agent = None if 'read' in data: notif.read = pywikibot.Timestamp.fromtimestampformat(data['read']) else: notif.read = False notif.content = data.get('*', None) return notif def mark_as_read(self): """Mark the notification as read.""" return self.site.notifications_mark_read(list=self.id) <commit_msg>Fix notifications building from JSON Sometimes (like in welcome messages), notifications don't have a 'title' property, so we shouldn't assume there is one. Bug: T139015 Change-Id: I83e480d04e8e09aa9bcb5edef4f56b47d150e199<commit_after># -*- coding: utf-8 -*- """Classes and functions for working with the Echo extension.""" from __future__ import absolute_import, unicode_literals import pywikibot class Notification(object): """A notification issued by the Echo extension.""" def __init__(self, site): """Construct an empty Notification object.""" self.site = site @classmethod def fromJSON(cls, site, data): """ Construct a Notification object from JSON data returned by the API. @rtype: Notification """ notif = cls(site) notif.id = data['id'] # TODO: use numeric id ? notif.type = data['type'] notif.category = data['category'] notif.timestamp = pywikibot.Timestamp.fromtimestampformat(data['timestamp']['mw']) if 'title' in data and 'full' in data['title']: notif.page = pywikibot.Page(site, data['title']['full']) else: notif.page = None if 'agent' in data and 'name' in data['agent']: notif.agent = pywikibot.User(site, data['agent']['name']) else: notif.agent = None if 'read' in data: notif.read = pywikibot.Timestamp.fromtimestampformat(data['read']) else: notif.read = False notif.content = data.get('*', None) return notif def mark_as_read(self): """Mark the notification as read.""" return self.site.notifications_mark_read(list=self.id)
6664d075b4037ae40a91267afaca5731aa73ed3c
bluebottle/utils/widgets.py
bluebottle/utils/widgets.py
from __future__ import unicode_literals from urlparse import urlparse from django.contrib.admin.widgets import AdminURLFieldWidget from django.forms.widgets import CheckboxFieldRenderer, CheckboxSelectMultiple, CheckboxChoiceInput from django.utils.html import format_html class NiceCheckboxChoiceInput(CheckboxChoiceInput): def render(self, name=None, value=None, attrs=None): if self.id_for_label: label_for = format_html(' for="{}"', self.id_for_label) else: label_for = '' attrs = dict(self.attrs, **attrs) if attrs else self.attrs return format_html( '{} <label{}>{}</label>', self.tag(attrs), label_for, self.choice_label ) def is_checked(self): return self.choice_value in self.value class MultiCheckboxRenderer(CheckboxFieldRenderer): choice_input_class = NiceCheckboxChoiceInput class CheckboxSelectMultipleWidget(CheckboxSelectMultiple): renderer = MultiCheckboxRenderer class SecureAdminURLFieldWidget(AdminURLFieldWidget): def render(self, name, value, attrs=None): if urlparse(value).scheme not in ('http', 'https', ): return super(AdminURLFieldWidget, self).render(name, value, attrs) else: return super(SecureAdminURLFieldWidget, self).render(name, value, attrs)
from __future__ import unicode_literals from urlparse import urlparse from django.contrib.admin.widgets import AdminURLFieldWidget from django.forms.widgets import CheckboxFieldRenderer, CheckboxSelectMultiple, CheckboxChoiceInput from django.utils.html import format_html class NiceCheckboxChoiceInput(CheckboxChoiceInput): def render(self, name=None, value=None, attrs=None): if self.id_for_label: label_for = format_html(' for="{}"', self.id_for_label) else: label_for = '' attrs = dict(self.attrs, **attrs) if attrs else self.attrs return format_html( '{} <label{}>{}</label>', self.tag(attrs), label_for, self.choice_label ) def is_checked(self): return self.choice_value in self.value class MultiCheckboxRenderer(CheckboxFieldRenderer): choice_input_class = NiceCheckboxChoiceInput class CheckboxSelectMultipleWidget(CheckboxSelectMultiple): renderer = MultiCheckboxRenderer class SecureAdminURLFieldWidget(AdminURLFieldWidget): def render(self, name, value, attrs=None): if value and urlparse(value).scheme not in ('http', 'https', ): return super(AdminURLFieldWidget, self).render(name, value, attrs) else: return super(SecureAdminURLFieldWidget, self).render(name, value, attrs)
Fix url fields when no value is set
Fix url fields when no value is set
Python
bsd-3-clause
onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle
from __future__ import unicode_literals from urlparse import urlparse from django.contrib.admin.widgets import AdminURLFieldWidget from django.forms.widgets import CheckboxFieldRenderer, CheckboxSelectMultiple, CheckboxChoiceInput from django.utils.html import format_html class NiceCheckboxChoiceInput(CheckboxChoiceInput): def render(self, name=None, value=None, attrs=None): if self.id_for_label: label_for = format_html(' for="{}"', self.id_for_label) else: label_for = '' attrs = dict(self.attrs, **attrs) if attrs else self.attrs return format_html( '{} <label{}>{}</label>', self.tag(attrs), label_for, self.choice_label ) def is_checked(self): return self.choice_value in self.value class MultiCheckboxRenderer(CheckboxFieldRenderer): choice_input_class = NiceCheckboxChoiceInput class CheckboxSelectMultipleWidget(CheckboxSelectMultiple): renderer = MultiCheckboxRenderer class SecureAdminURLFieldWidget(AdminURLFieldWidget): def render(self, name, value, attrs=None): if urlparse(value).scheme not in ('http', 'https', ): return super(AdminURLFieldWidget, self).render(name, value, attrs) else: return super(SecureAdminURLFieldWidget, self).render(name, value, attrs) Fix url fields when no value is set
from __future__ import unicode_literals from urlparse import urlparse from django.contrib.admin.widgets import AdminURLFieldWidget from django.forms.widgets import CheckboxFieldRenderer, CheckboxSelectMultiple, CheckboxChoiceInput from django.utils.html import format_html class NiceCheckboxChoiceInput(CheckboxChoiceInput): def render(self, name=None, value=None, attrs=None): if self.id_for_label: label_for = format_html(' for="{}"', self.id_for_label) else: label_for = '' attrs = dict(self.attrs, **attrs) if attrs else self.attrs return format_html( '{} <label{}>{}</label>', self.tag(attrs), label_for, self.choice_label ) def is_checked(self): return self.choice_value in self.value class MultiCheckboxRenderer(CheckboxFieldRenderer): choice_input_class = NiceCheckboxChoiceInput class CheckboxSelectMultipleWidget(CheckboxSelectMultiple): renderer = MultiCheckboxRenderer class SecureAdminURLFieldWidget(AdminURLFieldWidget): def render(self, name, value, attrs=None): if value and urlparse(value).scheme not in ('http', 'https', ): return super(AdminURLFieldWidget, self).render(name, value, attrs) else: return super(SecureAdminURLFieldWidget, self).render(name, value, attrs)
<commit_before>from __future__ import unicode_literals from urlparse import urlparse from django.contrib.admin.widgets import AdminURLFieldWidget from django.forms.widgets import CheckboxFieldRenderer, CheckboxSelectMultiple, CheckboxChoiceInput from django.utils.html import format_html class NiceCheckboxChoiceInput(CheckboxChoiceInput): def render(self, name=None, value=None, attrs=None): if self.id_for_label: label_for = format_html(' for="{}"', self.id_for_label) else: label_for = '' attrs = dict(self.attrs, **attrs) if attrs else self.attrs return format_html( '{} <label{}>{}</label>', self.tag(attrs), label_for, self.choice_label ) def is_checked(self): return self.choice_value in self.value class MultiCheckboxRenderer(CheckboxFieldRenderer): choice_input_class = NiceCheckboxChoiceInput class CheckboxSelectMultipleWidget(CheckboxSelectMultiple): renderer = MultiCheckboxRenderer class SecureAdminURLFieldWidget(AdminURLFieldWidget): def render(self, name, value, attrs=None): if urlparse(value).scheme not in ('http', 'https', ): return super(AdminURLFieldWidget, self).render(name, value, attrs) else: return super(SecureAdminURLFieldWidget, self).render(name, value, attrs) <commit_msg>Fix url fields when no value is set<commit_after>
from __future__ import unicode_literals from urlparse import urlparse from django.contrib.admin.widgets import AdminURLFieldWidget from django.forms.widgets import CheckboxFieldRenderer, CheckboxSelectMultiple, CheckboxChoiceInput from django.utils.html import format_html class NiceCheckboxChoiceInput(CheckboxChoiceInput): def render(self, name=None, value=None, attrs=None): if self.id_for_label: label_for = format_html(' for="{}"', self.id_for_label) else: label_for = '' attrs = dict(self.attrs, **attrs) if attrs else self.attrs return format_html( '{} <label{}>{}</label>', self.tag(attrs), label_for, self.choice_label ) def is_checked(self): return self.choice_value in self.value class MultiCheckboxRenderer(CheckboxFieldRenderer): choice_input_class = NiceCheckboxChoiceInput class CheckboxSelectMultipleWidget(CheckboxSelectMultiple): renderer = MultiCheckboxRenderer class SecureAdminURLFieldWidget(AdminURLFieldWidget): def render(self, name, value, attrs=None): if value and urlparse(value).scheme not in ('http', 'https', ): return super(AdminURLFieldWidget, self).render(name, value, attrs) else: return super(SecureAdminURLFieldWidget, self).render(name, value, attrs)
from __future__ import unicode_literals from urlparse import urlparse from django.contrib.admin.widgets import AdminURLFieldWidget from django.forms.widgets import CheckboxFieldRenderer, CheckboxSelectMultiple, CheckboxChoiceInput from django.utils.html import format_html class NiceCheckboxChoiceInput(CheckboxChoiceInput): def render(self, name=None, value=None, attrs=None): if self.id_for_label: label_for = format_html(' for="{}"', self.id_for_label) else: label_for = '' attrs = dict(self.attrs, **attrs) if attrs else self.attrs return format_html( '{} <label{}>{}</label>', self.tag(attrs), label_for, self.choice_label ) def is_checked(self): return self.choice_value in self.value class MultiCheckboxRenderer(CheckboxFieldRenderer): choice_input_class = NiceCheckboxChoiceInput class CheckboxSelectMultipleWidget(CheckboxSelectMultiple): renderer = MultiCheckboxRenderer class SecureAdminURLFieldWidget(AdminURLFieldWidget): def render(self, name, value, attrs=None): if urlparse(value).scheme not in ('http', 'https', ): return super(AdminURLFieldWidget, self).render(name, value, attrs) else: return super(SecureAdminURLFieldWidget, self).render(name, value, attrs) Fix url fields when no value is setfrom __future__ import unicode_literals from urlparse import urlparse from django.contrib.admin.widgets import AdminURLFieldWidget from django.forms.widgets import CheckboxFieldRenderer, CheckboxSelectMultiple, CheckboxChoiceInput from django.utils.html import format_html class NiceCheckboxChoiceInput(CheckboxChoiceInput): def render(self, name=None, value=None, attrs=None): if self.id_for_label: label_for = format_html(' for="{}"', self.id_for_label) else: label_for = '' attrs = dict(self.attrs, **attrs) if attrs else self.attrs return format_html( '{} <label{}>{}</label>', self.tag(attrs), label_for, self.choice_label ) def is_checked(self): return self.choice_value in self.value class MultiCheckboxRenderer(CheckboxFieldRenderer): choice_input_class = NiceCheckboxChoiceInput class CheckboxSelectMultipleWidget(CheckboxSelectMultiple): renderer = MultiCheckboxRenderer class SecureAdminURLFieldWidget(AdminURLFieldWidget): def render(self, name, value, attrs=None): if value and urlparse(value).scheme not in ('http', 'https', ): return super(AdminURLFieldWidget, self).render(name, value, attrs) else: return super(SecureAdminURLFieldWidget, self).render(name, value, attrs)
<commit_before>from __future__ import unicode_literals from urlparse import urlparse from django.contrib.admin.widgets import AdminURLFieldWidget from django.forms.widgets import CheckboxFieldRenderer, CheckboxSelectMultiple, CheckboxChoiceInput from django.utils.html import format_html class NiceCheckboxChoiceInput(CheckboxChoiceInput): def render(self, name=None, value=None, attrs=None): if self.id_for_label: label_for = format_html(' for="{}"', self.id_for_label) else: label_for = '' attrs = dict(self.attrs, **attrs) if attrs else self.attrs return format_html( '{} <label{}>{}</label>', self.tag(attrs), label_for, self.choice_label ) def is_checked(self): return self.choice_value in self.value class MultiCheckboxRenderer(CheckboxFieldRenderer): choice_input_class = NiceCheckboxChoiceInput class CheckboxSelectMultipleWidget(CheckboxSelectMultiple): renderer = MultiCheckboxRenderer class SecureAdminURLFieldWidget(AdminURLFieldWidget): def render(self, name, value, attrs=None): if urlparse(value).scheme not in ('http', 'https', ): return super(AdminURLFieldWidget, self).render(name, value, attrs) else: return super(SecureAdminURLFieldWidget, self).render(name, value, attrs) <commit_msg>Fix url fields when no value is set<commit_after>from __future__ import unicode_literals from urlparse import urlparse from django.contrib.admin.widgets import AdminURLFieldWidget from django.forms.widgets import CheckboxFieldRenderer, CheckboxSelectMultiple, CheckboxChoiceInput from django.utils.html import format_html class NiceCheckboxChoiceInput(CheckboxChoiceInput): def render(self, name=None, value=None, attrs=None): if self.id_for_label: label_for = format_html(' for="{}"', self.id_for_label) else: label_for = '' attrs = dict(self.attrs, **attrs) if attrs else self.attrs return format_html( '{} <label{}>{}</label>', self.tag(attrs), label_for, self.choice_label ) def is_checked(self): return self.choice_value in self.value class MultiCheckboxRenderer(CheckboxFieldRenderer): choice_input_class = NiceCheckboxChoiceInput class CheckboxSelectMultipleWidget(CheckboxSelectMultiple): renderer = MultiCheckboxRenderer class SecureAdminURLFieldWidget(AdminURLFieldWidget): def render(self, name, value, attrs=None): if value and urlparse(value).scheme not in ('http', 'https', ): return super(AdminURLFieldWidget, self).render(name, value, attrs) else: return super(SecureAdminURLFieldWidget, self).render(name, value, attrs)
bfe45a24800817e7445fa12e7cd859679e6452c3
porchlightapi/views.py
porchlightapi/views.py
# -*- coding: utf-8 -*- from django.shortcuts import render # Create your views here. import django_filters from rest_framework import viewsets from rest_framework import filters from porchlightapi.models import Repository, ValueDataPoint from porchlightapi.serializers import RepositorySerializer, ValueDataPointSerializer class RepositoryFilter(django_filters.FilterSet): """ Provide filtering of repository objects based on name or project. This is 'icontains' filtering, so a repo with the name "Porchlight" will match 'por', 'Por', etc. """ name = django_filters.CharFilter(name="name", lookup_type='icontains') project = django_filters.CharFilter(name="project", lookup_type='icontains') class Meta: model = Repository fields = ['name', 'project',] class RepositoryViewSet(viewsets.ReadOnlyModelViewSet): """ A REST view of repositories. Provides GET listing of Repositories, filtering on name and project, and individual repsositories on id. """ queryset = Repository.objects.all() serializer_class = RepositorySerializer filter_backends = (filters.DjangoFilterBackend,) filter_class = RepositoryFilter # filter_fields = ('name', 'project') class ValueDataPointViewSet(viewsets.ReadOnlyModelViewSet): """ A REST view of data points. """ queryset = ValueDataPoint.objects.all() serializer_class = ValueDataPointSerializer
# -*- coding: utf-8 -*- from django.shortcuts import render # Create your views here. from rest_framework import viewsets from rest_framework import filters from porchlightapi.models import Repository, ValueDataPoint from porchlightapi.serializers import RepositorySerializer, ValueDataPointSerializer class RepositoryViewSet(viewsets.ReadOnlyModelViewSet): """ A REST view of repositories. Provides GET listing of Repositories, filtering on name and project, and individual repsositories on id. """ queryset = Repository.objects.all() serializer_class = RepositorySerializer filter_backends = (filters.SearchFilter,) search_fields = ('name', 'project', 'url') class ValueDataPointViewSet(viewsets.ReadOnlyModelViewSet): """ A REST view of data points. """ queryset = ValueDataPoint.objects.all() serializer_class = ValueDataPointSerializer
Use DRF's built-in search filter
Use DRF's built-in search filter
Python
cc0-1.0
cfpb/porchlight,cfpb/porchlight,cfpb/porchlight
# -*- coding: utf-8 -*- from django.shortcuts import render # Create your views here. import django_filters from rest_framework import viewsets from rest_framework import filters from porchlightapi.models import Repository, ValueDataPoint from porchlightapi.serializers import RepositorySerializer, ValueDataPointSerializer class RepositoryFilter(django_filters.FilterSet): """ Provide filtering of repository objects based on name or project. This is 'icontains' filtering, so a repo with the name "Porchlight" will match 'por', 'Por', etc. """ name = django_filters.CharFilter(name="name", lookup_type='icontains') project = django_filters.CharFilter(name="project", lookup_type='icontains') class Meta: model = Repository fields = ['name', 'project',] class RepositoryViewSet(viewsets.ReadOnlyModelViewSet): """ A REST view of repositories. Provides GET listing of Repositories, filtering on name and project, and individual repsositories on id. """ queryset = Repository.objects.all() serializer_class = RepositorySerializer filter_backends = (filters.DjangoFilterBackend,) filter_class = RepositoryFilter # filter_fields = ('name', 'project') class ValueDataPointViewSet(viewsets.ReadOnlyModelViewSet): """ A REST view of data points. """ queryset = ValueDataPoint.objects.all() serializer_class = ValueDataPointSerializer Use DRF's built-in search filter
# -*- coding: utf-8 -*- from django.shortcuts import render # Create your views here. from rest_framework import viewsets from rest_framework import filters from porchlightapi.models import Repository, ValueDataPoint from porchlightapi.serializers import RepositorySerializer, ValueDataPointSerializer class RepositoryViewSet(viewsets.ReadOnlyModelViewSet): """ A REST view of repositories. Provides GET listing of Repositories, filtering on name and project, and individual repsositories on id. """ queryset = Repository.objects.all() serializer_class = RepositorySerializer filter_backends = (filters.SearchFilter,) search_fields = ('name', 'project', 'url') class ValueDataPointViewSet(viewsets.ReadOnlyModelViewSet): """ A REST view of data points. """ queryset = ValueDataPoint.objects.all() serializer_class = ValueDataPointSerializer
<commit_before># -*- coding: utf-8 -*- from django.shortcuts import render # Create your views here. import django_filters from rest_framework import viewsets from rest_framework import filters from porchlightapi.models import Repository, ValueDataPoint from porchlightapi.serializers import RepositorySerializer, ValueDataPointSerializer class RepositoryFilter(django_filters.FilterSet): """ Provide filtering of repository objects based on name or project. This is 'icontains' filtering, so a repo with the name "Porchlight" will match 'por', 'Por', etc. """ name = django_filters.CharFilter(name="name", lookup_type='icontains') project = django_filters.CharFilter(name="project", lookup_type='icontains') class Meta: model = Repository fields = ['name', 'project',] class RepositoryViewSet(viewsets.ReadOnlyModelViewSet): """ A REST view of repositories. Provides GET listing of Repositories, filtering on name and project, and individual repsositories on id. """ queryset = Repository.objects.all() serializer_class = RepositorySerializer filter_backends = (filters.DjangoFilterBackend,) filter_class = RepositoryFilter # filter_fields = ('name', 'project') class ValueDataPointViewSet(viewsets.ReadOnlyModelViewSet): """ A REST view of data points. """ queryset = ValueDataPoint.objects.all() serializer_class = ValueDataPointSerializer <commit_msg>Use DRF's built-in search filter<commit_after>
# -*- coding: utf-8 -*- from django.shortcuts import render # Create your views here. from rest_framework import viewsets from rest_framework import filters from porchlightapi.models import Repository, ValueDataPoint from porchlightapi.serializers import RepositorySerializer, ValueDataPointSerializer class RepositoryViewSet(viewsets.ReadOnlyModelViewSet): """ A REST view of repositories. Provides GET listing of Repositories, filtering on name and project, and individual repsositories on id. """ queryset = Repository.objects.all() serializer_class = RepositorySerializer filter_backends = (filters.SearchFilter,) search_fields = ('name', 'project', 'url') class ValueDataPointViewSet(viewsets.ReadOnlyModelViewSet): """ A REST view of data points. """ queryset = ValueDataPoint.objects.all() serializer_class = ValueDataPointSerializer
# -*- coding: utf-8 -*- from django.shortcuts import render # Create your views here. import django_filters from rest_framework import viewsets from rest_framework import filters from porchlightapi.models import Repository, ValueDataPoint from porchlightapi.serializers import RepositorySerializer, ValueDataPointSerializer class RepositoryFilter(django_filters.FilterSet): """ Provide filtering of repository objects based on name or project. This is 'icontains' filtering, so a repo with the name "Porchlight" will match 'por', 'Por', etc. """ name = django_filters.CharFilter(name="name", lookup_type='icontains') project = django_filters.CharFilter(name="project", lookup_type='icontains') class Meta: model = Repository fields = ['name', 'project',] class RepositoryViewSet(viewsets.ReadOnlyModelViewSet): """ A REST view of repositories. Provides GET listing of Repositories, filtering on name and project, and individual repsositories on id. """ queryset = Repository.objects.all() serializer_class = RepositorySerializer filter_backends = (filters.DjangoFilterBackend,) filter_class = RepositoryFilter # filter_fields = ('name', 'project') class ValueDataPointViewSet(viewsets.ReadOnlyModelViewSet): """ A REST view of data points. """ queryset = ValueDataPoint.objects.all() serializer_class = ValueDataPointSerializer Use DRF's built-in search filter# -*- coding: utf-8 -*- from django.shortcuts import render # Create your views here. from rest_framework import viewsets from rest_framework import filters from porchlightapi.models import Repository, ValueDataPoint from porchlightapi.serializers import RepositorySerializer, ValueDataPointSerializer class RepositoryViewSet(viewsets.ReadOnlyModelViewSet): """ A REST view of repositories. Provides GET listing of Repositories, filtering on name and project, and individual repsositories on id. """ queryset = Repository.objects.all() serializer_class = RepositorySerializer filter_backends = (filters.SearchFilter,) search_fields = ('name', 'project', 'url') class ValueDataPointViewSet(viewsets.ReadOnlyModelViewSet): """ A REST view of data points. """ queryset = ValueDataPoint.objects.all() serializer_class = ValueDataPointSerializer
<commit_before># -*- coding: utf-8 -*- from django.shortcuts import render # Create your views here. import django_filters from rest_framework import viewsets from rest_framework import filters from porchlightapi.models import Repository, ValueDataPoint from porchlightapi.serializers import RepositorySerializer, ValueDataPointSerializer class RepositoryFilter(django_filters.FilterSet): """ Provide filtering of repository objects based on name or project. This is 'icontains' filtering, so a repo with the name "Porchlight" will match 'por', 'Por', etc. """ name = django_filters.CharFilter(name="name", lookup_type='icontains') project = django_filters.CharFilter(name="project", lookup_type='icontains') class Meta: model = Repository fields = ['name', 'project',] class RepositoryViewSet(viewsets.ReadOnlyModelViewSet): """ A REST view of repositories. Provides GET listing of Repositories, filtering on name and project, and individual repsositories on id. """ queryset = Repository.objects.all() serializer_class = RepositorySerializer filter_backends = (filters.DjangoFilterBackend,) filter_class = RepositoryFilter # filter_fields = ('name', 'project') class ValueDataPointViewSet(viewsets.ReadOnlyModelViewSet): """ A REST view of data points. """ queryset = ValueDataPoint.objects.all() serializer_class = ValueDataPointSerializer <commit_msg>Use DRF's built-in search filter<commit_after># -*- coding: utf-8 -*- from django.shortcuts import render # Create your views here. from rest_framework import viewsets from rest_framework import filters from porchlightapi.models import Repository, ValueDataPoint from porchlightapi.serializers import RepositorySerializer, ValueDataPointSerializer class RepositoryViewSet(viewsets.ReadOnlyModelViewSet): """ A REST view of repositories. Provides GET listing of Repositories, filtering on name and project, and individual repsositories on id. """ queryset = Repository.objects.all() serializer_class = RepositorySerializer filter_backends = (filters.SearchFilter,) search_fields = ('name', 'project', 'url') class ValueDataPointViewSet(viewsets.ReadOnlyModelViewSet): """ A REST view of data points. """ queryset = ValueDataPoint.objects.all() serializer_class = ValueDataPointSerializer
1ecbd06083ac65a9520bcf0f87c5f5f1b4a4e532
helloworld.py
helloworld.py
#This is my hello world program str1='Hello' str2='Tarun' print str1 +' '+ str2 # this is my hello world program print 'Hello World!' #This is my Hello world program str1='Hello' str2='Akash' print str1 + ' ' + str2 + '!' #this is a comment str1='Hello' str2='Priyanka' print str1+' '+str2
print "helloworld"
Add strings to print hello world
Add strings to print hello world
Python
apache-2.0
ctsit/J.O.B-Training-Repo-1
#This is my hello world program str1='Hello' str2='Tarun' print str1 +' '+ str2 # this is my hello world program print 'Hello World!' #This is my Hello world program str1='Hello' str2='Akash' print str1 + ' ' + str2 + '!' #this is a comment str1='Hello' str2='Priyanka' print str1+' '+str2Add strings to print hello world
print "helloworld"
<commit_before> #This is my hello world program str1='Hello' str2='Tarun' print str1 +' '+ str2 # this is my hello world program print 'Hello World!' #This is my Hello world program str1='Hello' str2='Akash' print str1 + ' ' + str2 + '!' #this is a comment str1='Hello' str2='Priyanka' print str1+' '+str2<commit_msg>Add strings to print hello world<commit_after>
print "helloworld"
#This is my hello world program str1='Hello' str2='Tarun' print str1 +' '+ str2 # this is my hello world program print 'Hello World!' #This is my Hello world program str1='Hello' str2='Akash' print str1 + ' ' + str2 + '!' #this is a comment str1='Hello' str2='Priyanka' print str1+' '+str2Add strings to print hello world print "helloworld"
<commit_before> #This is my hello world program str1='Hello' str2='Tarun' print str1 +' '+ str2 # this is my hello world program print 'Hello World!' #This is my Hello world program str1='Hello' str2='Akash' print str1 + ' ' + str2 + '!' #this is a comment str1='Hello' str2='Priyanka' print str1+' '+str2<commit_msg>Add strings to print hello world<commit_after> print "helloworld"
14c473b8bef44ee5b521ce365ad89249c7f6e39e
linter.py
linter.py
# # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Jon Surrell # Copyright (c) 2014 Jon Surrell # # License: MIT # """This module exports the Gotype plugin class.""" from SublimeLinter.lint import Linter, util class Gotype(Linter): """Provides an interface to gotype.""" syntax = 'go' cmd = ('gotype', '-e') regex = r'^.+:(?P<line>\d+):(?P<col>\d+):\s+(?P<message>.+)$' tempfile_suffix = 'go' error_stream = util.STREAM_STDERR
# # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Jon Surrell # Copyright (c) 2014 Jon Surrell # # License: MIT # """This module exports the Gotype plugin class.""" from SublimeLinter.lint import Linter, util class Gotype(Linter): """Provides an interface to gotype.""" syntax = 'go' cmd = ('gotype', '-e') regex = r'^.+:(?P<line>\d+):(?P<col>\d+):\s+(?P<message>.+)' tempfile_suffix = 'go' error_stream = util.STREAM_STDERR
Remove trailing $ from regex
Remove trailing $ from regex
Python
mit
sirreal/SublimeLinter-contrib-gotype
# # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Jon Surrell # Copyright (c) 2014 Jon Surrell # # License: MIT # """This module exports the Gotype plugin class.""" from SublimeLinter.lint import Linter, util class Gotype(Linter): """Provides an interface to gotype.""" syntax = 'go' cmd = ('gotype', '-e') regex = r'^.+:(?P<line>\d+):(?P<col>\d+):\s+(?P<message>.+)$' tempfile_suffix = 'go' error_stream = util.STREAM_STDERR Remove trailing $ from regex
# # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Jon Surrell # Copyright (c) 2014 Jon Surrell # # License: MIT # """This module exports the Gotype plugin class.""" from SublimeLinter.lint import Linter, util class Gotype(Linter): """Provides an interface to gotype.""" syntax = 'go' cmd = ('gotype', '-e') regex = r'^.+:(?P<line>\d+):(?P<col>\d+):\s+(?P<message>.+)' tempfile_suffix = 'go' error_stream = util.STREAM_STDERR
<commit_before># # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Jon Surrell # Copyright (c) 2014 Jon Surrell # # License: MIT # """This module exports the Gotype plugin class.""" from SublimeLinter.lint import Linter, util class Gotype(Linter): """Provides an interface to gotype.""" syntax = 'go' cmd = ('gotype', '-e') regex = r'^.+:(?P<line>\d+):(?P<col>\d+):\s+(?P<message>.+)$' tempfile_suffix = 'go' error_stream = util.STREAM_STDERR <commit_msg>Remove trailing $ from regex<commit_after>
# # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Jon Surrell # Copyright (c) 2014 Jon Surrell # # License: MIT # """This module exports the Gotype plugin class.""" from SublimeLinter.lint import Linter, util class Gotype(Linter): """Provides an interface to gotype.""" syntax = 'go' cmd = ('gotype', '-e') regex = r'^.+:(?P<line>\d+):(?P<col>\d+):\s+(?P<message>.+)' tempfile_suffix = 'go' error_stream = util.STREAM_STDERR
# # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Jon Surrell # Copyright (c) 2014 Jon Surrell # # License: MIT # """This module exports the Gotype plugin class.""" from SublimeLinter.lint import Linter, util class Gotype(Linter): """Provides an interface to gotype.""" syntax = 'go' cmd = ('gotype', '-e') regex = r'^.+:(?P<line>\d+):(?P<col>\d+):\s+(?P<message>.+)$' tempfile_suffix = 'go' error_stream = util.STREAM_STDERR Remove trailing $ from regex# # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Jon Surrell # Copyright (c) 2014 Jon Surrell # # License: MIT # """This module exports the Gotype plugin class.""" from SublimeLinter.lint import Linter, util class Gotype(Linter): """Provides an interface to gotype.""" syntax = 'go' cmd = ('gotype', '-e') regex = r'^.+:(?P<line>\d+):(?P<col>\d+):\s+(?P<message>.+)' tempfile_suffix = 'go' error_stream = util.STREAM_STDERR
<commit_before># # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Jon Surrell # Copyright (c) 2014 Jon Surrell # # License: MIT # """This module exports the Gotype plugin class.""" from SublimeLinter.lint import Linter, util class Gotype(Linter): """Provides an interface to gotype.""" syntax = 'go' cmd = ('gotype', '-e') regex = r'^.+:(?P<line>\d+):(?P<col>\d+):\s+(?P<message>.+)$' tempfile_suffix = 'go' error_stream = util.STREAM_STDERR <commit_msg>Remove trailing $ from regex<commit_after># # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Jon Surrell # Copyright (c) 2014 Jon Surrell # # License: MIT # """This module exports the Gotype plugin class.""" from SublimeLinter.lint import Linter, util class Gotype(Linter): """Provides an interface to gotype.""" syntax = 'go' cmd = ('gotype', '-e') regex = r'^.+:(?P<line>\d+):(?P<col>\d+):\s+(?P<message>.+)' tempfile_suffix = 'go' error_stream = util.STREAM_STDERR
d1928f0b1c98093b977ae208613c2b7eeb9a3ce5
carepoint/tests/models/cph/test_address.py
carepoint/tests/models/cph/test_address.py
# -*- coding: utf-8 -*- ############################################################################## # # Author: Dave Lasley <dave@laslabs.com> # Copyright: 2015 LasLabs, Inc [https://laslabs.com] # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # ############################################################################## import os import unittest import mock from carepoint.tests.db.db import DatabaseTest from carepoint.models.cph.address import Address class ModelCphAddressTest(DatabaseTest): def test_primary_key(self, ): print Address.__table__ if __name__ == '__main__': unittest.main()
# -*- coding: utf-8 -*- ############################################################################## # # Author: Dave Lasley <dave@laslabs.com> # Copyright: 2015 LasLabs, Inc [https://laslabs.com] # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # ############################################################################## import os import unittest import mock from sqlalchemy.schema import Table from carepoint.tests.db.db import DatabaseTest from carepoint.models.cph.address import Address class ModelCphAddressTest(DatabaseTest): def test_table_initialization(self, ): self.assertIsInstance(Address.__table__, Table) if __name__ == '__main__': unittest.main()
Add instance assertion to table
Add instance assertion to table
Python
mit
laslabs/Python-Carepoint
# -*- coding: utf-8 -*- ############################################################################## # # Author: Dave Lasley <dave@laslabs.com> # Copyright: 2015 LasLabs, Inc [https://laslabs.com] # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # ############################################################################## import os import unittest import mock from carepoint.tests.db.db import DatabaseTest from carepoint.models.cph.address import Address class ModelCphAddressTest(DatabaseTest): def test_primary_key(self, ): print Address.__table__ if __name__ == '__main__': unittest.main() Add instance assertion to table
# -*- coding: utf-8 -*- ############################################################################## # # Author: Dave Lasley <dave@laslabs.com> # Copyright: 2015 LasLabs, Inc [https://laslabs.com] # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # ############################################################################## import os import unittest import mock from sqlalchemy.schema import Table from carepoint.tests.db.db import DatabaseTest from carepoint.models.cph.address import Address class ModelCphAddressTest(DatabaseTest): def test_table_initialization(self, ): self.assertIsInstance(Address.__table__, Table) if __name__ == '__main__': unittest.main()
<commit_before># -*- coding: utf-8 -*- ############################################################################## # # Author: Dave Lasley <dave@laslabs.com> # Copyright: 2015 LasLabs, Inc [https://laslabs.com] # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # ############################################################################## import os import unittest import mock from carepoint.tests.db.db import DatabaseTest from carepoint.models.cph.address import Address class ModelCphAddressTest(DatabaseTest): def test_primary_key(self, ): print Address.__table__ if __name__ == '__main__': unittest.main() <commit_msg>Add instance assertion to table<commit_after>
# -*- coding: utf-8 -*- ############################################################################## # # Author: Dave Lasley <dave@laslabs.com> # Copyright: 2015 LasLabs, Inc [https://laslabs.com] # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # ############################################################################## import os import unittest import mock from sqlalchemy.schema import Table from carepoint.tests.db.db import DatabaseTest from carepoint.models.cph.address import Address class ModelCphAddressTest(DatabaseTest): def test_table_initialization(self, ): self.assertIsInstance(Address.__table__, Table) if __name__ == '__main__': unittest.main()
# -*- coding: utf-8 -*- ############################################################################## # # Author: Dave Lasley <dave@laslabs.com> # Copyright: 2015 LasLabs, Inc [https://laslabs.com] # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # ############################################################################## import os import unittest import mock from carepoint.tests.db.db import DatabaseTest from carepoint.models.cph.address import Address class ModelCphAddressTest(DatabaseTest): def test_primary_key(self, ): print Address.__table__ if __name__ == '__main__': unittest.main() Add instance assertion to table# -*- coding: utf-8 -*- ############################################################################## # # Author: Dave Lasley <dave@laslabs.com> # Copyright: 2015 LasLabs, Inc [https://laslabs.com] # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # ############################################################################## import os import unittest import mock from sqlalchemy.schema import Table from carepoint.tests.db.db import DatabaseTest from carepoint.models.cph.address import Address class ModelCphAddressTest(DatabaseTest): def test_table_initialization(self, ): self.assertIsInstance(Address.__table__, Table) if __name__ == '__main__': unittest.main()
<commit_before># -*- coding: utf-8 -*- ############################################################################## # # Author: Dave Lasley <dave@laslabs.com> # Copyright: 2015 LasLabs, Inc [https://laslabs.com] # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # ############################################################################## import os import unittest import mock from carepoint.tests.db.db import DatabaseTest from carepoint.models.cph.address import Address class ModelCphAddressTest(DatabaseTest): def test_primary_key(self, ): print Address.__table__ if __name__ == '__main__': unittest.main() <commit_msg>Add instance assertion to table<commit_after># -*- coding: utf-8 -*- ############################################################################## # # Author: Dave Lasley <dave@laslabs.com> # Copyright: 2015 LasLabs, Inc [https://laslabs.com] # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # ############################################################################## import os import unittest import mock from sqlalchemy.schema import Table from carepoint.tests.db.db import DatabaseTest from carepoint.models.cph.address import Address class ModelCphAddressTest(DatabaseTest): def test_table_initialization(self, ): self.assertIsInstance(Address.__table__, Table) if __name__ == '__main__': unittest.main()
60951f30d8b5e2a450c13aa2b146be14ceb53c4d
rolldembones.py
rolldembones.py
#!/usr/bin/python import argparse import dice def main(): roller = dice.Roller(args) for repeat in range(args.repeats): roller.do_roll() for result in roller: if isinstance(result, list): print(' '.join(map(str, result))) else: print(result) if __name__ == '__main__': parser = argparse.ArgumentParser(description="Roll some dice.") parser.add_argument("-r, --repeat", dest="repeats", type=int, metavar="N", default=1, help="Repeat these rolls N times.") parser.add_argument("-e, --explode", dest="explode", metavar="E", type=int, default=None, help="Any die whose roll matches or exceeds E is counted and rolled again. Set to 1 or lower to disable this behavior on special dice.") parser.add_argument("dice", nargs='*', help="Dice to roll, given in pairs of the number of dice to roll, and the sides those dice have.") args = parser.parse_args() # some basic error checking if len(args.dice)%2 != 0: parser.error("Incorrect number of arguments: Rolls and faces must be paired") main()
#!/usr/bin/python3 import argparse import dice def main(): roller = dice.Roller(args) for repeat in range(args.repeats): roller.do_roll() for result in roller: if isinstance(result, list): print(' '.join(map(str, result))) else: print(result) if __name__ == '__main__': parser = argparse.ArgumentParser(description="Roll some dice.") parser.add_argument("-r, --repeat", dest="repeats", type=int, metavar="N", default=1, help="Repeat these rolls N times.") parser.add_argument("-e, --explode", dest="explode", metavar="E", type=int, default=None, help="Any die whose roll matches or exceeds E is counted and rolled again. Set to 1 or lower to disable this behavior on special dice.") parser.add_argument("dice", nargs='*', help="Dice to roll, given in pairs of the number of dice to roll, and the sides those dice have.") args = parser.parse_args() # some basic error checking if len(args.dice)%2 != 0: parser.error("Incorrect number of arguments: Rolls and faces must be paired") main()
Update shebang to request python 3
Update shebang to request python 3
Python
mit
aurule/rolldembones
#!/usr/bin/python import argparse import dice def main(): roller = dice.Roller(args) for repeat in range(args.repeats): roller.do_roll() for result in roller: if isinstance(result, list): print(' '.join(map(str, result))) else: print(result) if __name__ == '__main__': parser = argparse.ArgumentParser(description="Roll some dice.") parser.add_argument("-r, --repeat", dest="repeats", type=int, metavar="N", default=1, help="Repeat these rolls N times.") parser.add_argument("-e, --explode", dest="explode", metavar="E", type=int, default=None, help="Any die whose roll matches or exceeds E is counted and rolled again. Set to 1 or lower to disable this behavior on special dice.") parser.add_argument("dice", nargs='*', help="Dice to roll, given in pairs of the number of dice to roll, and the sides those dice have.") args = parser.parse_args() # some basic error checking if len(args.dice)%2 != 0: parser.error("Incorrect number of arguments: Rolls and faces must be paired") main()Update shebang to request python 3
#!/usr/bin/python3 import argparse import dice def main(): roller = dice.Roller(args) for repeat in range(args.repeats): roller.do_roll() for result in roller: if isinstance(result, list): print(' '.join(map(str, result))) else: print(result) if __name__ == '__main__': parser = argparse.ArgumentParser(description="Roll some dice.") parser.add_argument("-r, --repeat", dest="repeats", type=int, metavar="N", default=1, help="Repeat these rolls N times.") parser.add_argument("-e, --explode", dest="explode", metavar="E", type=int, default=None, help="Any die whose roll matches or exceeds E is counted and rolled again. Set to 1 or lower to disable this behavior on special dice.") parser.add_argument("dice", nargs='*', help="Dice to roll, given in pairs of the number of dice to roll, and the sides those dice have.") args = parser.parse_args() # some basic error checking if len(args.dice)%2 != 0: parser.error("Incorrect number of arguments: Rolls and faces must be paired") main()
<commit_before>#!/usr/bin/python import argparse import dice def main(): roller = dice.Roller(args) for repeat in range(args.repeats): roller.do_roll() for result in roller: if isinstance(result, list): print(' '.join(map(str, result))) else: print(result) if __name__ == '__main__': parser = argparse.ArgumentParser(description="Roll some dice.") parser.add_argument("-r, --repeat", dest="repeats", type=int, metavar="N", default=1, help="Repeat these rolls N times.") parser.add_argument("-e, --explode", dest="explode", metavar="E", type=int, default=None, help="Any die whose roll matches or exceeds E is counted and rolled again. Set to 1 or lower to disable this behavior on special dice.") parser.add_argument("dice", nargs='*', help="Dice to roll, given in pairs of the number of dice to roll, and the sides those dice have.") args = parser.parse_args() # some basic error checking if len(args.dice)%2 != 0: parser.error("Incorrect number of arguments: Rolls and faces must be paired") main()<commit_msg>Update shebang to request python 3<commit_after>
#!/usr/bin/python3 import argparse import dice def main(): roller = dice.Roller(args) for repeat in range(args.repeats): roller.do_roll() for result in roller: if isinstance(result, list): print(' '.join(map(str, result))) else: print(result) if __name__ == '__main__': parser = argparse.ArgumentParser(description="Roll some dice.") parser.add_argument("-r, --repeat", dest="repeats", type=int, metavar="N", default=1, help="Repeat these rolls N times.") parser.add_argument("-e, --explode", dest="explode", metavar="E", type=int, default=None, help="Any die whose roll matches or exceeds E is counted and rolled again. Set to 1 or lower to disable this behavior on special dice.") parser.add_argument("dice", nargs='*', help="Dice to roll, given in pairs of the number of dice to roll, and the sides those dice have.") args = parser.parse_args() # some basic error checking if len(args.dice)%2 != 0: parser.error("Incorrect number of arguments: Rolls and faces must be paired") main()
#!/usr/bin/python import argparse import dice def main(): roller = dice.Roller(args) for repeat in range(args.repeats): roller.do_roll() for result in roller: if isinstance(result, list): print(' '.join(map(str, result))) else: print(result) if __name__ == '__main__': parser = argparse.ArgumentParser(description="Roll some dice.") parser.add_argument("-r, --repeat", dest="repeats", type=int, metavar="N", default=1, help="Repeat these rolls N times.") parser.add_argument("-e, --explode", dest="explode", metavar="E", type=int, default=None, help="Any die whose roll matches or exceeds E is counted and rolled again. Set to 1 or lower to disable this behavior on special dice.") parser.add_argument("dice", nargs='*', help="Dice to roll, given in pairs of the number of dice to roll, and the sides those dice have.") args = parser.parse_args() # some basic error checking if len(args.dice)%2 != 0: parser.error("Incorrect number of arguments: Rolls and faces must be paired") main()Update shebang to request python 3#!/usr/bin/python3 import argparse import dice def main(): roller = dice.Roller(args) for repeat in range(args.repeats): roller.do_roll() for result in roller: if isinstance(result, list): print(' '.join(map(str, result))) else: print(result) if __name__ == '__main__': parser = argparse.ArgumentParser(description="Roll some dice.") parser.add_argument("-r, --repeat", dest="repeats", type=int, metavar="N", default=1, help="Repeat these rolls N times.") parser.add_argument("-e, --explode", dest="explode", metavar="E", type=int, default=None, help="Any die whose roll matches or exceeds E is counted and rolled again. Set to 1 or lower to disable this behavior on special dice.") parser.add_argument("dice", nargs='*', help="Dice to roll, given in pairs of the number of dice to roll, and the sides those dice have.") args = parser.parse_args() # some basic error checking if len(args.dice)%2 != 0: parser.error("Incorrect number of arguments: Rolls and faces must be paired") main()
<commit_before>#!/usr/bin/python import argparse import dice def main(): roller = dice.Roller(args) for repeat in range(args.repeats): roller.do_roll() for result in roller: if isinstance(result, list): print(' '.join(map(str, result))) else: print(result) if __name__ == '__main__': parser = argparse.ArgumentParser(description="Roll some dice.") parser.add_argument("-r, --repeat", dest="repeats", type=int, metavar="N", default=1, help="Repeat these rolls N times.") parser.add_argument("-e, --explode", dest="explode", metavar="E", type=int, default=None, help="Any die whose roll matches or exceeds E is counted and rolled again. Set to 1 or lower to disable this behavior on special dice.") parser.add_argument("dice", nargs='*', help="Dice to roll, given in pairs of the number of dice to roll, and the sides those dice have.") args = parser.parse_args() # some basic error checking if len(args.dice)%2 != 0: parser.error("Incorrect number of arguments: Rolls and faces must be paired") main()<commit_msg>Update shebang to request python 3<commit_after>#!/usr/bin/python3 import argparse import dice def main(): roller = dice.Roller(args) for repeat in range(args.repeats): roller.do_roll() for result in roller: if isinstance(result, list): print(' '.join(map(str, result))) else: print(result) if __name__ == '__main__': parser = argparse.ArgumentParser(description="Roll some dice.") parser.add_argument("-r, --repeat", dest="repeats", type=int, metavar="N", default=1, help="Repeat these rolls N times.") parser.add_argument("-e, --explode", dest="explode", metavar="E", type=int, default=None, help="Any die whose roll matches or exceeds E is counted and rolled again. Set to 1 or lower to disable this behavior on special dice.") parser.add_argument("dice", nargs='*', help="Dice to roll, given in pairs of the number of dice to roll, and the sides those dice have.") args = parser.parse_args() # some basic error checking if len(args.dice)%2 != 0: parser.error("Incorrect number of arguments: Rolls and faces must be paired") main()
0f67d19a2cc38d8781946e20f6cd17b5287848a4
common/djangoapps/track/backends/logger.py
common/djangoapps/track/backends/logger.py
"""Event tracker backend that saves events to a python logger.""" from __future__ import absolute_import import logging import json from django.conf import settings from track.backends import BaseBackend from track.utils import DateTimeJSONEncoder log = logging.getLogger('track.backends.logger') class LoggerBackend(BaseBackend): """Event tracker backend that uses a python logger. Events are logged to the INFO level as JSON strings. """ def __init__(self, name, **kwargs): """Event tracker backend that uses a python logger. :Parameters: - `name`: identifier of the logger, which should have been configured using the default python mechanisms. """ super(LoggerBackend, self).__init__(**kwargs) self.event_logger = logging.getLogger(name) def send(self, event): event_str = json.dumps(event, cls=DateTimeJSONEncoder) # TODO: remove trucation of the serialized event, either at a # higher level during the emittion of the event, or by # providing warnings when the events exceed certain size. event_str = event_str[:settings.TRACK_MAX_EVENT] self.event_logger.info(event_str)
"""Event tracker backend that saves events to a python logger.""" from __future__ import absolute_import import logging import json from django.conf import settings from track.backends import BaseBackend from track.utils import DateTimeJSONEncoder log = logging.getLogger('track.backends.logger') application_log = logging.getLogger('track.backends.application_log') # pylint: disable=invalid-name class LoggerBackend(BaseBackend): """Event tracker backend that uses a python logger. Events are logged to the INFO level as JSON strings. """ def __init__(self, name, **kwargs): """Event tracker backend that uses a python logger. :Parameters: - `name`: identifier of the logger, which should have been configured using the default python mechanisms. """ super(LoggerBackend, self).__init__(**kwargs) self.event_logger = logging.getLogger(name) def send(self, event): try: event_str = json.dumps(event, cls=DateTimeJSONEncoder) except UnicodeDecodeError: application_log.exception( "UnicodeDecodeError Event_type: %r, Event_source: %r, Page: %r, Referer: %r", event.get('event_type'), event.get('event_source'), event.get('page'), event.get('referer') ) # TODO: remove trucation of the serialized event, either at a # higher level during the emittion of the event, or by # providing warnings when the events exceed certain size. event_str = event_str[:settings.TRACK_MAX_EVENT] self.event_logger.info(event_str)
Add logging for UnicodeDecodeError excpetion LoggerBackend
Add logging for UnicodeDecodeError excpetion LoggerBackend
Python
agpl-3.0
cognitiveclass/edx-platform,jolyonb/edx-platform,miptliot/edx-platform,jjmiranda/edx-platform,ZLLab-Mooc/edx-platform,procangroup/edx-platform,edx-solutions/edx-platform,jjmiranda/edx-platform,synergeticsedx/deployment-wipro,edx/edx-platform,UOMx/edx-platform,defance/edx-platform,EDUlib/edx-platform,raccoongang/edx-platform,antoviaque/edx-platform,alu042/edx-platform,marcore/edx-platform,simbs/edx-platform,jzoldak/edx-platform,philanthropy-u/edx-platform,waheedahmed/edx-platform,appsembler/edx-platform,shurihell/testasia,edx/edx-platform,kmoocdev2/edx-platform,10clouds/edx-platform,pabloborrego93/edx-platform,lduarte1991/edx-platform,ZLLab-Mooc/edx-platform,longmen21/edx-platform,Livit/Livit.Learn.EdX,caesar2164/edx-platform,fintech-circle/edx-platform,tanmaykm/edx-platform,raccoongang/edx-platform,defance/edx-platform,wwj718/edx-platform,jolyonb/edx-platform,RPI-OPENEDX/edx-platform,cpennington/edx-platform,Ayub-Khan/edx-platform,Stanford-Online/edx-platform,Edraak/circleci-edx-platform,chrisndodge/edx-platform,msegado/edx-platform,raccoongang/edx-platform,halvertoluke/edx-platform,Stanford-Online/edx-platform,procangroup/edx-platform,prarthitm/edxplatform,IndonesiaX/edx-platform,cpennington/edx-platform,devs1991/test_edx_docmode,a-parhom/edx-platform,10clouds/edx-platform,deepsrijit1105/edx-platform,devs1991/test_edx_docmode,eduNEXT/edunext-platform,ESOedX/edx-platform,JioEducation/edx-platform,ampax/edx-platform,prarthitm/edxplatform,solashirai/edx-platform,nttks/edx-platform,ampax/edx-platform,EDUlib/edx-platform,lduarte1991/edx-platform,franosincic/edx-platform,pomegranited/edx-platform,Endika/edx-platform,Edraak/edx-platform,waheedahmed/edx-platform,jjmiranda/edx-platform,nttks/edx-platform,TeachAtTUM/edx-platform,Livit/Livit.Learn.EdX,ahmedaljazzar/edx-platform,doganov/edx-platform,edx/edx-platform,deepsrijit1105/edx-platform,RPI-OPENEDX/edx-platform,10clouds/edx-platform,zhenzhai/edx-platform,Edraak/edx-platform,zhenzhai/edx-platform,wwj718/edx-platform,JioEducation/edx-platform,amir-qayyum-khan/edx-platform,Lektorium-LLC/edx-platform,cognitiveclass/edx-platform,kursitet/edx-platform,pomegranited/edx-platform,MakeHer/edx-platform,doganov/edx-platform,mbareta/edx-platform-ft,bigdatauniversity/edx-platform,BehavioralInsightsTeam/edx-platform,ESOedX/edx-platform,cecep-edu/edx-platform,simbs/edx-platform,IndonesiaX/edx-platform,CredoReference/edx-platform,raccoongang/edx-platform,bigdatauniversity/edx-platform,romain-li/edx-platform,appsembler/edx-platform,simbs/edx-platform,devs1991/test_edx_docmode,kursitet/edx-platform,eduNEXT/edunext-platform,analyseuc3m/ANALYSE-v1,BehavioralInsightsTeam/edx-platform,shurihell/testasia,jzoldak/edx-platform,mitocw/edx-platform,ovnicraft/edx-platform,defance/edx-platform,eduNEXT/edx-platform,proversity-org/edx-platform,marcore/edx-platform,stvstnfrd/edx-platform,jjmiranda/edx-platform,Edraak/circleci-edx-platform,edx-solutions/edx-platform,IndonesiaX/edx-platform,MakeHer/edx-platform,halvertoluke/edx-platform,Edraak/edraak-platform,itsjeyd/edx-platform,edx/edx-platform,shabab12/edx-platform,cognitiveclass/edx-platform,waheedahmed/edx-platform,mbareta/edx-platform-ft,msegado/edx-platform,appsembler/edx-platform,proversity-org/edx-platform,angelapper/edx-platform,marcore/edx-platform,zhenzhai/edx-platform,stvstnfrd/edx-platform,gymnasium/edx-platform,shurihell/testasia,halvertoluke/edx-platform,fintech-circle/edx-platform,Ayub-Khan/edx-platform,gymnasium/edx-platform,pabloborrego93/edx-platform,hastexo/edx-platform,romain-li/edx-platform,caesar2164/edx-platform,pepeportela/edx-platform,hastexo/edx-platform,wwj718/edx-platform,Ayub-Khan/edx-platform,naresh21/synergetics-edx-platform,hastexo/edx-platform,franosincic/edx-platform,pepeportela/edx-platform,marcore/edx-platform,synergeticsedx/deployment-wipro,IndonesiaX/edx-platform,kmoocdev2/edx-platform,romain-li/edx-platform,jzoldak/edx-platform,naresh21/synergetics-edx-platform,cecep-edu/edx-platform,franosincic/edx-platform,Stanford-Online/edx-platform,analyseuc3m/ANALYSE-v1,caesar2164/edx-platform,angelapper/edx-platform,msegado/edx-platform,msegado/edx-platform,waheedahmed/edx-platform,MakeHer/edx-platform,ahmedaljazzar/edx-platform,itsjeyd/edx-platform,kmoocdev2/edx-platform,synergeticsedx/deployment-wipro,antoviaque/edx-platform,proversity-org/edx-platform,nttks/edx-platform,kursitet/edx-platform,antoviaque/edx-platform,RPI-OPENEDX/edx-platform,kursitet/edx-platform,Livit/Livit.Learn.EdX,procangroup/edx-platform,shurihell/testasia,edx-solutions/edx-platform,kmoocdev2/edx-platform,teltek/edx-platform,shabab12/edx-platform,shabab12/edx-platform,shurihell/testasia,Edraak/circleci-edx-platform,BehavioralInsightsTeam/edx-platform,halvertoluke/edx-platform,analyseuc3m/ANALYSE-v1,cecep-edu/edx-platform,CourseTalk/edx-platform,deepsrijit1105/edx-platform,pepeportela/edx-platform,eduNEXT/edunext-platform,pabloborrego93/edx-platform,appsembler/edx-platform,UOMx/edx-platform,hastexo/edx-platform,Endika/edx-platform,JioEducation/edx-platform,chrisndodge/edx-platform,ESOedX/edx-platform,JioEducation/edx-platform,synergeticsedx/deployment-wipro,chrisndodge/edx-platform,halvertoluke/edx-platform,romain-li/edx-platform,philanthropy-u/edx-platform,defance/edx-platform,CredoReference/edx-platform,EDUlib/edx-platform,Lektorium-LLC/edx-platform,cognitiveclass/edx-platform,philanthropy-u/edx-platform,naresh21/synergetics-edx-platform,MakeHer/edx-platform,teltek/edx-platform,angelapper/edx-platform,jbzdak/edx-platform,jzoldak/edx-platform,alu042/edx-platform,bigdatauniversity/edx-platform,solashirai/edx-platform,RPI-OPENEDX/edx-platform,alu042/edx-platform,zhenzhai/edx-platform,UOMx/edx-platform,fintech-circle/edx-platform,pepeportela/edx-platform,angelapper/edx-platform,longmen21/edx-platform,solashirai/edx-platform,doganov/edx-platform,TeachAtTUM/edx-platform,chrisndodge/edx-platform,simbs/edx-platform,simbs/edx-platform,pomegranited/edx-platform,Edraak/edx-platform,procangroup/edx-platform,itsjeyd/edx-platform,antoviaque/edx-platform,tanmaykm/edx-platform,UOMx/edx-platform,eduNEXT/edx-platform,BehavioralInsightsTeam/edx-platform,ZLLab-Mooc/edx-platform,louyihua/edx-platform,eduNEXT/edunext-platform,Stanford-Online/edx-platform,CredoReference/edx-platform,jolyonb/edx-platform,Edraak/edraak-platform,10clouds/edx-platform,Edraak/circleci-edx-platform,devs1991/test_edx_docmode,zhenzhai/edx-platform,Ayub-Khan/edx-platform,devs1991/test_edx_docmode,pomegranited/edx-platform,ampax/edx-platform,arbrandes/edx-platform,shabab12/edx-platform,proversity-org/edx-platform,kursitet/edx-platform,ZLLab-Mooc/edx-platform,mitocw/edx-platform,miptliot/edx-platform,stvstnfrd/edx-platform,cpennington/edx-platform,louyihua/edx-platform,louyihua/edx-platform,Endika/edx-platform,fintech-circle/edx-platform,analyseuc3m/ANALYSE-v1,solashirai/edx-platform,ovnicraft/edx-platform,waheedahmed/edx-platform,lduarte1991/edx-platform,msegado/edx-platform,Lektorium-LLC/edx-platform,RPI-OPENEDX/edx-platform,jbzdak/edx-platform,Edraak/edraak-platform,pomegranited/edx-platform,tanmaykm/edx-platform,eduNEXT/edx-platform,miptliot/edx-platform,doganov/edx-platform,amir-qayyum-khan/edx-platform,stvstnfrd/edx-platform,philanthropy-u/edx-platform,arbrandes/edx-platform,tanmaykm/edx-platform,nttks/edx-platform,EDUlib/edx-platform,amir-qayyum-khan/edx-platform,ovnicraft/edx-platform,cecep-edu/edx-platform,wwj718/edx-platform,ovnicraft/edx-platform,longmen21/edx-platform,longmen21/edx-platform,TeachAtTUM/edx-platform,deepsrijit1105/edx-platform,prarthitm/edxplatform,pabloborrego93/edx-platform,CourseTalk/edx-platform,naresh21/synergetics-edx-platform,gymnasium/edx-platform,gsehub/edx-platform,a-parhom/edx-platform,CourseTalk/edx-platform,louyihua/edx-platform,jbzdak/edx-platform,mitocw/edx-platform,CourseTalk/edx-platform,ESOedX/edx-platform,arbrandes/edx-platform,eduNEXT/edx-platform,devs1991/test_edx_docmode,alu042/edx-platform,jbzdak/edx-platform,ZLLab-Mooc/edx-platform,Endika/edx-platform,prarthitm/edxplatform,teltek/edx-platform,Edraak/edx-platform,doganov/edx-platform,Edraak/edx-platform,a-parhom/edx-platform,TeachAtTUM/edx-platform,gsehub/edx-platform,jolyonb/edx-platform,kmoocdev2/edx-platform,Edraak/circleci-edx-platform,gymnasium/edx-platform,amir-qayyum-khan/edx-platform,edx-solutions/edx-platform,teltek/edx-platform,caesar2164/edx-platform,mbareta/edx-platform-ft,Livit/Livit.Learn.EdX,ampax/edx-platform,Edraak/edraak-platform,longmen21/edx-platform,lduarte1991/edx-platform,nttks/edx-platform,bigdatauniversity/edx-platform,cognitiveclass/edx-platform,devs1991/test_edx_docmode,ahmedaljazzar/edx-platform,mbareta/edx-platform-ft,Lektorium-LLC/edx-platform,miptliot/edx-platform,gsehub/edx-platform,CredoReference/edx-platform,devs1991/test_edx_docmode,ahmedaljazzar/edx-platform,ovnicraft/edx-platform,wwj718/edx-platform,gsehub/edx-platform,cpennington/edx-platform,Ayub-Khan/edx-platform,franosincic/edx-platform,solashirai/edx-platform,jbzdak/edx-platform,a-parhom/edx-platform,IndonesiaX/edx-platform,MakeHer/edx-platform,itsjeyd/edx-platform,franosincic/edx-platform,mitocw/edx-platform,romain-li/edx-platform,arbrandes/edx-platform,cecep-edu/edx-platform,bigdatauniversity/edx-platform
"""Event tracker backend that saves events to a python logger.""" from __future__ import absolute_import import logging import json from django.conf import settings from track.backends import BaseBackend from track.utils import DateTimeJSONEncoder log = logging.getLogger('track.backends.logger') class LoggerBackend(BaseBackend): """Event tracker backend that uses a python logger. Events are logged to the INFO level as JSON strings. """ def __init__(self, name, **kwargs): """Event tracker backend that uses a python logger. :Parameters: - `name`: identifier of the logger, which should have been configured using the default python mechanisms. """ super(LoggerBackend, self).__init__(**kwargs) self.event_logger = logging.getLogger(name) def send(self, event): event_str = json.dumps(event, cls=DateTimeJSONEncoder) # TODO: remove trucation of the serialized event, either at a # higher level during the emittion of the event, or by # providing warnings when the events exceed certain size. event_str = event_str[:settings.TRACK_MAX_EVENT] self.event_logger.info(event_str) Add logging for UnicodeDecodeError excpetion LoggerBackend
"""Event tracker backend that saves events to a python logger.""" from __future__ import absolute_import import logging import json from django.conf import settings from track.backends import BaseBackend from track.utils import DateTimeJSONEncoder log = logging.getLogger('track.backends.logger') application_log = logging.getLogger('track.backends.application_log') # pylint: disable=invalid-name class LoggerBackend(BaseBackend): """Event tracker backend that uses a python logger. Events are logged to the INFO level as JSON strings. """ def __init__(self, name, **kwargs): """Event tracker backend that uses a python logger. :Parameters: - `name`: identifier of the logger, which should have been configured using the default python mechanisms. """ super(LoggerBackend, self).__init__(**kwargs) self.event_logger = logging.getLogger(name) def send(self, event): try: event_str = json.dumps(event, cls=DateTimeJSONEncoder) except UnicodeDecodeError: application_log.exception( "UnicodeDecodeError Event_type: %r, Event_source: %r, Page: %r, Referer: %r", event.get('event_type'), event.get('event_source'), event.get('page'), event.get('referer') ) # TODO: remove trucation of the serialized event, either at a # higher level during the emittion of the event, or by # providing warnings when the events exceed certain size. event_str = event_str[:settings.TRACK_MAX_EVENT] self.event_logger.info(event_str)
<commit_before>"""Event tracker backend that saves events to a python logger.""" from __future__ import absolute_import import logging import json from django.conf import settings from track.backends import BaseBackend from track.utils import DateTimeJSONEncoder log = logging.getLogger('track.backends.logger') class LoggerBackend(BaseBackend): """Event tracker backend that uses a python logger. Events are logged to the INFO level as JSON strings. """ def __init__(self, name, **kwargs): """Event tracker backend that uses a python logger. :Parameters: - `name`: identifier of the logger, which should have been configured using the default python mechanisms. """ super(LoggerBackend, self).__init__(**kwargs) self.event_logger = logging.getLogger(name) def send(self, event): event_str = json.dumps(event, cls=DateTimeJSONEncoder) # TODO: remove trucation of the serialized event, either at a # higher level during the emittion of the event, or by # providing warnings when the events exceed certain size. event_str = event_str[:settings.TRACK_MAX_EVENT] self.event_logger.info(event_str) <commit_msg>Add logging for UnicodeDecodeError excpetion LoggerBackend<commit_after>
"""Event tracker backend that saves events to a python logger.""" from __future__ import absolute_import import logging import json from django.conf import settings from track.backends import BaseBackend from track.utils import DateTimeJSONEncoder log = logging.getLogger('track.backends.logger') application_log = logging.getLogger('track.backends.application_log') # pylint: disable=invalid-name class LoggerBackend(BaseBackend): """Event tracker backend that uses a python logger. Events are logged to the INFO level as JSON strings. """ def __init__(self, name, **kwargs): """Event tracker backend that uses a python logger. :Parameters: - `name`: identifier of the logger, which should have been configured using the default python mechanisms. """ super(LoggerBackend, self).__init__(**kwargs) self.event_logger = logging.getLogger(name) def send(self, event): try: event_str = json.dumps(event, cls=DateTimeJSONEncoder) except UnicodeDecodeError: application_log.exception( "UnicodeDecodeError Event_type: %r, Event_source: %r, Page: %r, Referer: %r", event.get('event_type'), event.get('event_source'), event.get('page'), event.get('referer') ) # TODO: remove trucation of the serialized event, either at a # higher level during the emittion of the event, or by # providing warnings when the events exceed certain size. event_str = event_str[:settings.TRACK_MAX_EVENT] self.event_logger.info(event_str)
"""Event tracker backend that saves events to a python logger.""" from __future__ import absolute_import import logging import json from django.conf import settings from track.backends import BaseBackend from track.utils import DateTimeJSONEncoder log = logging.getLogger('track.backends.logger') class LoggerBackend(BaseBackend): """Event tracker backend that uses a python logger. Events are logged to the INFO level as JSON strings. """ def __init__(self, name, **kwargs): """Event tracker backend that uses a python logger. :Parameters: - `name`: identifier of the logger, which should have been configured using the default python mechanisms. """ super(LoggerBackend, self).__init__(**kwargs) self.event_logger = logging.getLogger(name) def send(self, event): event_str = json.dumps(event, cls=DateTimeJSONEncoder) # TODO: remove trucation of the serialized event, either at a # higher level during the emittion of the event, or by # providing warnings when the events exceed certain size. event_str = event_str[:settings.TRACK_MAX_EVENT] self.event_logger.info(event_str) Add logging for UnicodeDecodeError excpetion LoggerBackend"""Event tracker backend that saves events to a python logger.""" from __future__ import absolute_import import logging import json from django.conf import settings from track.backends import BaseBackend from track.utils import DateTimeJSONEncoder log = logging.getLogger('track.backends.logger') application_log = logging.getLogger('track.backends.application_log') # pylint: disable=invalid-name class LoggerBackend(BaseBackend): """Event tracker backend that uses a python logger. Events are logged to the INFO level as JSON strings. """ def __init__(self, name, **kwargs): """Event tracker backend that uses a python logger. :Parameters: - `name`: identifier of the logger, which should have been configured using the default python mechanisms. """ super(LoggerBackend, self).__init__(**kwargs) self.event_logger = logging.getLogger(name) def send(self, event): try: event_str = json.dumps(event, cls=DateTimeJSONEncoder) except UnicodeDecodeError: application_log.exception( "UnicodeDecodeError Event_type: %r, Event_source: %r, Page: %r, Referer: %r", event.get('event_type'), event.get('event_source'), event.get('page'), event.get('referer') ) # TODO: remove trucation of the serialized event, either at a # higher level during the emittion of the event, or by # providing warnings when the events exceed certain size. event_str = event_str[:settings.TRACK_MAX_EVENT] self.event_logger.info(event_str)
<commit_before>"""Event tracker backend that saves events to a python logger.""" from __future__ import absolute_import import logging import json from django.conf import settings from track.backends import BaseBackend from track.utils import DateTimeJSONEncoder log = logging.getLogger('track.backends.logger') class LoggerBackend(BaseBackend): """Event tracker backend that uses a python logger. Events are logged to the INFO level as JSON strings. """ def __init__(self, name, **kwargs): """Event tracker backend that uses a python logger. :Parameters: - `name`: identifier of the logger, which should have been configured using the default python mechanisms. """ super(LoggerBackend, self).__init__(**kwargs) self.event_logger = logging.getLogger(name) def send(self, event): event_str = json.dumps(event, cls=DateTimeJSONEncoder) # TODO: remove trucation of the serialized event, either at a # higher level during the emittion of the event, or by # providing warnings when the events exceed certain size. event_str = event_str[:settings.TRACK_MAX_EVENT] self.event_logger.info(event_str) <commit_msg>Add logging for UnicodeDecodeError excpetion LoggerBackend<commit_after>"""Event tracker backend that saves events to a python logger.""" from __future__ import absolute_import import logging import json from django.conf import settings from track.backends import BaseBackend from track.utils import DateTimeJSONEncoder log = logging.getLogger('track.backends.logger') application_log = logging.getLogger('track.backends.application_log') # pylint: disable=invalid-name class LoggerBackend(BaseBackend): """Event tracker backend that uses a python logger. Events are logged to the INFO level as JSON strings. """ def __init__(self, name, **kwargs): """Event tracker backend that uses a python logger. :Parameters: - `name`: identifier of the logger, which should have been configured using the default python mechanisms. """ super(LoggerBackend, self).__init__(**kwargs) self.event_logger = logging.getLogger(name) def send(self, event): try: event_str = json.dumps(event, cls=DateTimeJSONEncoder) except UnicodeDecodeError: application_log.exception( "UnicodeDecodeError Event_type: %r, Event_source: %r, Page: %r, Referer: %r", event.get('event_type'), event.get('event_source'), event.get('page'), event.get('referer') ) # TODO: remove trucation of the serialized event, either at a # higher level during the emittion of the event, or by # providing warnings when the events exceed certain size. event_str = event_str[:settings.TRACK_MAX_EVENT] self.event_logger.info(event_str)
8fc274021a8c0813f3fc3568d1d7984112952b9c
pytilemap/qtsupport.py
pytilemap/qtsupport.py
import sys import sip import qtpy __all__ = [ 'getQVariantValue', 'wheelAngleDelta', ] try: if qtpy.PYQT5: QVARIANT_API = 2 else: QVARIANT_API = sip.getapi('QVariant') except ValueError: QVARIANT_API = 1 if QVARIANT_API == 1: def getQVariantValue(variant): return variant.toPyObject() else: def getQVariantValue(variant): return variant if qtpy.PYQT5: def wheelAngleDelta(wheelEvent): return wheelEvent.angleDelta().y() else: def wheelAngleDelta(wheelEvent): return wheelEvent.delta() if qtpy.PYQT5: from qtpy.QtCore import QStandardPaths def getTemporaryFolder(): return QStandardPaths.writableLocation(QStandardPaths.TempLocation) else: from qtpy.QtGui import QDesktopServices def getTemporaryFolder(): return QDesktopServices.storageLocation(QDesktopServices.TempLocation)
import sys import sip import qtpy __all__ = [ 'getQVariantValue', 'wheelAngleDelta', ] try: if qtpy.PYQT5: QVARIANT_API = 2 else: QVARIANT_API = sip.getapi('QVariant') except ValueError: QVARIANT_API = 1 if QVARIANT_API == 1: def getQVariantValue(variant): return variant.toPyObject() else: def getQVariantValue(variant): return variant if qtpy.PYQT5: def wheelAngleDelta(wheelEvent): return wheelEvent.angleDelta().y() else: def wheelAngleDelta(wheelEvent): return wheelEvent.delta() if qtpy.PYQT5: from qtpy.QtCore import QStandardPaths def getTemporaryFolder(): return QStandardPaths.writableLocation(QStandardPaths.CacheLocation) else: from qtpy.QtGui import QDesktopServices def getTemporaryFolder(): return QDesktopServices.storageLocation(QDesktopServices.CacheLocation)
Use Cache location instead of temp folder
Use Cache location instead of temp folder
Python
mit
allebacco/PyTileMap
import sys import sip import qtpy __all__ = [ 'getQVariantValue', 'wheelAngleDelta', ] try: if qtpy.PYQT5: QVARIANT_API = 2 else: QVARIANT_API = sip.getapi('QVariant') except ValueError: QVARIANT_API = 1 if QVARIANT_API == 1: def getQVariantValue(variant): return variant.toPyObject() else: def getQVariantValue(variant): return variant if qtpy.PYQT5: def wheelAngleDelta(wheelEvent): return wheelEvent.angleDelta().y() else: def wheelAngleDelta(wheelEvent): return wheelEvent.delta() if qtpy.PYQT5: from qtpy.QtCore import QStandardPaths def getTemporaryFolder(): return QStandardPaths.writableLocation(QStandardPaths.TempLocation) else: from qtpy.QtGui import QDesktopServices def getTemporaryFolder(): return QDesktopServices.storageLocation(QDesktopServices.TempLocation) Use Cache location instead of temp folder
import sys import sip import qtpy __all__ = [ 'getQVariantValue', 'wheelAngleDelta', ] try: if qtpy.PYQT5: QVARIANT_API = 2 else: QVARIANT_API = sip.getapi('QVariant') except ValueError: QVARIANT_API = 1 if QVARIANT_API == 1: def getQVariantValue(variant): return variant.toPyObject() else: def getQVariantValue(variant): return variant if qtpy.PYQT5: def wheelAngleDelta(wheelEvent): return wheelEvent.angleDelta().y() else: def wheelAngleDelta(wheelEvent): return wheelEvent.delta() if qtpy.PYQT5: from qtpy.QtCore import QStandardPaths def getTemporaryFolder(): return QStandardPaths.writableLocation(QStandardPaths.CacheLocation) else: from qtpy.QtGui import QDesktopServices def getTemporaryFolder(): return QDesktopServices.storageLocation(QDesktopServices.CacheLocation)
<commit_before> import sys import sip import qtpy __all__ = [ 'getQVariantValue', 'wheelAngleDelta', ] try: if qtpy.PYQT5: QVARIANT_API = 2 else: QVARIANT_API = sip.getapi('QVariant') except ValueError: QVARIANT_API = 1 if QVARIANT_API == 1: def getQVariantValue(variant): return variant.toPyObject() else: def getQVariantValue(variant): return variant if qtpy.PYQT5: def wheelAngleDelta(wheelEvent): return wheelEvent.angleDelta().y() else: def wheelAngleDelta(wheelEvent): return wheelEvent.delta() if qtpy.PYQT5: from qtpy.QtCore import QStandardPaths def getTemporaryFolder(): return QStandardPaths.writableLocation(QStandardPaths.TempLocation) else: from qtpy.QtGui import QDesktopServices def getTemporaryFolder(): return QDesktopServices.storageLocation(QDesktopServices.TempLocation) <commit_msg>Use Cache location instead of temp folder<commit_after>
import sys import sip import qtpy __all__ = [ 'getQVariantValue', 'wheelAngleDelta', ] try: if qtpy.PYQT5: QVARIANT_API = 2 else: QVARIANT_API = sip.getapi('QVariant') except ValueError: QVARIANT_API = 1 if QVARIANT_API == 1: def getQVariantValue(variant): return variant.toPyObject() else: def getQVariantValue(variant): return variant if qtpy.PYQT5: def wheelAngleDelta(wheelEvent): return wheelEvent.angleDelta().y() else: def wheelAngleDelta(wheelEvent): return wheelEvent.delta() if qtpy.PYQT5: from qtpy.QtCore import QStandardPaths def getTemporaryFolder(): return QStandardPaths.writableLocation(QStandardPaths.CacheLocation) else: from qtpy.QtGui import QDesktopServices def getTemporaryFolder(): return QDesktopServices.storageLocation(QDesktopServices.CacheLocation)
import sys import sip import qtpy __all__ = [ 'getQVariantValue', 'wheelAngleDelta', ] try: if qtpy.PYQT5: QVARIANT_API = 2 else: QVARIANT_API = sip.getapi('QVariant') except ValueError: QVARIANT_API = 1 if QVARIANT_API == 1: def getQVariantValue(variant): return variant.toPyObject() else: def getQVariantValue(variant): return variant if qtpy.PYQT5: def wheelAngleDelta(wheelEvent): return wheelEvent.angleDelta().y() else: def wheelAngleDelta(wheelEvent): return wheelEvent.delta() if qtpy.PYQT5: from qtpy.QtCore import QStandardPaths def getTemporaryFolder(): return QStandardPaths.writableLocation(QStandardPaths.TempLocation) else: from qtpy.QtGui import QDesktopServices def getTemporaryFolder(): return QDesktopServices.storageLocation(QDesktopServices.TempLocation) Use Cache location instead of temp folder import sys import sip import qtpy __all__ = [ 'getQVariantValue', 'wheelAngleDelta', ] try: if qtpy.PYQT5: QVARIANT_API = 2 else: QVARIANT_API = sip.getapi('QVariant') except ValueError: QVARIANT_API = 1 if QVARIANT_API == 1: def getQVariantValue(variant): return variant.toPyObject() else: def getQVariantValue(variant): return variant if qtpy.PYQT5: def wheelAngleDelta(wheelEvent): return wheelEvent.angleDelta().y() else: def wheelAngleDelta(wheelEvent): return wheelEvent.delta() if qtpy.PYQT5: from qtpy.QtCore import QStandardPaths def getTemporaryFolder(): return QStandardPaths.writableLocation(QStandardPaths.CacheLocation) else: from qtpy.QtGui import QDesktopServices def getTemporaryFolder(): return QDesktopServices.storageLocation(QDesktopServices.CacheLocation)
<commit_before> import sys import sip import qtpy __all__ = [ 'getQVariantValue', 'wheelAngleDelta', ] try: if qtpy.PYQT5: QVARIANT_API = 2 else: QVARIANT_API = sip.getapi('QVariant') except ValueError: QVARIANT_API = 1 if QVARIANT_API == 1: def getQVariantValue(variant): return variant.toPyObject() else: def getQVariantValue(variant): return variant if qtpy.PYQT5: def wheelAngleDelta(wheelEvent): return wheelEvent.angleDelta().y() else: def wheelAngleDelta(wheelEvent): return wheelEvent.delta() if qtpy.PYQT5: from qtpy.QtCore import QStandardPaths def getTemporaryFolder(): return QStandardPaths.writableLocation(QStandardPaths.TempLocation) else: from qtpy.QtGui import QDesktopServices def getTemporaryFolder(): return QDesktopServices.storageLocation(QDesktopServices.TempLocation) <commit_msg>Use Cache location instead of temp folder<commit_after> import sys import sip import qtpy __all__ = [ 'getQVariantValue', 'wheelAngleDelta', ] try: if qtpy.PYQT5: QVARIANT_API = 2 else: QVARIANT_API = sip.getapi('QVariant') except ValueError: QVARIANT_API = 1 if QVARIANT_API == 1: def getQVariantValue(variant): return variant.toPyObject() else: def getQVariantValue(variant): return variant if qtpy.PYQT5: def wheelAngleDelta(wheelEvent): return wheelEvent.angleDelta().y() else: def wheelAngleDelta(wheelEvent): return wheelEvent.delta() if qtpy.PYQT5: from qtpy.QtCore import QStandardPaths def getTemporaryFolder(): return QStandardPaths.writableLocation(QStandardPaths.CacheLocation) else: from qtpy.QtGui import QDesktopServices def getTemporaryFolder(): return QDesktopServices.storageLocation(QDesktopServices.CacheLocation)
4fd6abddcc3457e53046f5a1c1bcc277083a8b15
entrypoint.py
entrypoint.py
#!/usr/bin/python3 # # Define containerized environment for running Diosix on Qemu # # On Google Cloud Run: Creates HTTP server on port 8080 # or whatever was specified using the PORT system variable. # Use this to signal the build was successful and the container\ # can be run via the command line. # # On all other environments: Log to stdout # # syntax: entrypoint.py <command> # # Author: Chris Williams <diodesign@tuta.io> # import os import sys global command_result from flask import Flask app = Flask(__name__) # for Google Cloud Run @app.route('/') def ContainerService(): return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\n' if __name__ == "__main__": if (os.environ.get('K_SERVICE')) != '': app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080))) else: stream = os.popen('. $HOME/.cargo/env && cd /build/diosix && {}'.format(' '.join(sys.argv[1:]))) output = stream.read() output
#!/usr/bin/python3 # # Define containerized environment for running Diosix on Qemu # # On Google Cloud Run: Creates HTTP server on port 8080 # or whatever was specified using the PORT system variable. # Outputs via the HTTP port. This requires K_SERVICE to be set. # # On all other environments: Log to stdout # # syntax: entrypoint.py <command> # # Author: Chris Williams <diodesign@tuta.io> # import os import sys global command_result from flask import Flask app = Flask(__name__) # for Google Cloud Run @app.route('/') def ContainerService(): return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\n' if __name__ == "__main__": if (os.environ.get('K_SERVICE')) != '': print('Running HTTP service for Google Cloud') app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080))) else: print('Running locally') stream = os.popen('. $HOME/.cargo/env && cd /build/diosix && {}'.format(' '.join(sys.argv[1:]))) output = stream.read() output
Debug Google Cloud Run support
Debug Google Cloud Run support
Python
mit
diodesign/diosix
#!/usr/bin/python3 # # Define containerized environment for running Diosix on Qemu # # On Google Cloud Run: Creates HTTP server on port 8080 # or whatever was specified using the PORT system variable. # Use this to signal the build was successful and the container\ # can be run via the command line. # # On all other environments: Log to stdout # # syntax: entrypoint.py <command> # # Author: Chris Williams <diodesign@tuta.io> # import os import sys global command_result from flask import Flask app = Flask(__name__) # for Google Cloud Run @app.route('/') def ContainerService(): return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\n' if __name__ == "__main__": if (os.environ.get('K_SERVICE')) != '': app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080))) else: stream = os.popen('. $HOME/.cargo/env && cd /build/diosix && {}'.format(' '.join(sys.argv[1:]))) output = stream.read() output Debug Google Cloud Run support
#!/usr/bin/python3 # # Define containerized environment for running Diosix on Qemu # # On Google Cloud Run: Creates HTTP server on port 8080 # or whatever was specified using the PORT system variable. # Outputs via the HTTP port. This requires K_SERVICE to be set. # # On all other environments: Log to stdout # # syntax: entrypoint.py <command> # # Author: Chris Williams <diodesign@tuta.io> # import os import sys global command_result from flask import Flask app = Flask(__name__) # for Google Cloud Run @app.route('/') def ContainerService(): return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\n' if __name__ == "__main__": if (os.environ.get('K_SERVICE')) != '': print('Running HTTP service for Google Cloud') app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080))) else: print('Running locally') stream = os.popen('. $HOME/.cargo/env && cd /build/diosix && {}'.format(' '.join(sys.argv[1:]))) output = stream.read() output
<commit_before>#!/usr/bin/python3 # # Define containerized environment for running Diosix on Qemu # # On Google Cloud Run: Creates HTTP server on port 8080 # or whatever was specified using the PORT system variable. # Use this to signal the build was successful and the container\ # can be run via the command line. # # On all other environments: Log to stdout # # syntax: entrypoint.py <command> # # Author: Chris Williams <diodesign@tuta.io> # import os import sys global command_result from flask import Flask app = Flask(__name__) # for Google Cloud Run @app.route('/') def ContainerService(): return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\n' if __name__ == "__main__": if (os.environ.get('K_SERVICE')) != '': app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080))) else: stream = os.popen('. $HOME/.cargo/env && cd /build/diosix && {}'.format(' '.join(sys.argv[1:]))) output = stream.read() output <commit_msg>Debug Google Cloud Run support<commit_after>
#!/usr/bin/python3 # # Define containerized environment for running Diosix on Qemu # # On Google Cloud Run: Creates HTTP server on port 8080 # or whatever was specified using the PORT system variable. # Outputs via the HTTP port. This requires K_SERVICE to be set. # # On all other environments: Log to stdout # # syntax: entrypoint.py <command> # # Author: Chris Williams <diodesign@tuta.io> # import os import sys global command_result from flask import Flask app = Flask(__name__) # for Google Cloud Run @app.route('/') def ContainerService(): return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\n' if __name__ == "__main__": if (os.environ.get('K_SERVICE')) != '': print('Running HTTP service for Google Cloud') app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080))) else: print('Running locally') stream = os.popen('. $HOME/.cargo/env && cd /build/diosix && {}'.format(' '.join(sys.argv[1:]))) output = stream.read() output
#!/usr/bin/python3 # # Define containerized environment for running Diosix on Qemu # # On Google Cloud Run: Creates HTTP server on port 8080 # or whatever was specified using the PORT system variable. # Use this to signal the build was successful and the container\ # can be run via the command line. # # On all other environments: Log to stdout # # syntax: entrypoint.py <command> # # Author: Chris Williams <diodesign@tuta.io> # import os import sys global command_result from flask import Flask app = Flask(__name__) # for Google Cloud Run @app.route('/') def ContainerService(): return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\n' if __name__ == "__main__": if (os.environ.get('K_SERVICE')) != '': app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080))) else: stream = os.popen('. $HOME/.cargo/env && cd /build/diosix && {}'.format(' '.join(sys.argv[1:]))) output = stream.read() output Debug Google Cloud Run support#!/usr/bin/python3 # # Define containerized environment for running Diosix on Qemu # # On Google Cloud Run: Creates HTTP server on port 8080 # or whatever was specified using the PORT system variable. # Outputs via the HTTP port. This requires K_SERVICE to be set. # # On all other environments: Log to stdout # # syntax: entrypoint.py <command> # # Author: Chris Williams <diodesign@tuta.io> # import os import sys global command_result from flask import Flask app = Flask(__name__) # for Google Cloud Run @app.route('/') def ContainerService(): return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\n' if __name__ == "__main__": if (os.environ.get('K_SERVICE')) != '': print('Running HTTP service for Google Cloud') app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080))) else: print('Running locally') stream = os.popen('. $HOME/.cargo/env && cd /build/diosix && {}'.format(' '.join(sys.argv[1:]))) output = stream.read() output
<commit_before>#!/usr/bin/python3 # # Define containerized environment for running Diosix on Qemu # # On Google Cloud Run: Creates HTTP server on port 8080 # or whatever was specified using the PORT system variable. # Use this to signal the build was successful and the container\ # can be run via the command line. # # On all other environments: Log to stdout # # syntax: entrypoint.py <command> # # Author: Chris Williams <diodesign@tuta.io> # import os import sys global command_result from flask import Flask app = Flask(__name__) # for Google Cloud Run @app.route('/') def ContainerService(): return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\n' if __name__ == "__main__": if (os.environ.get('K_SERVICE')) != '': app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080))) else: stream = os.popen('. $HOME/.cargo/env && cd /build/diosix && {}'.format(' '.join(sys.argv[1:]))) output = stream.read() output <commit_msg>Debug Google Cloud Run support<commit_after>#!/usr/bin/python3 # # Define containerized environment for running Diosix on Qemu # # On Google Cloud Run: Creates HTTP server on port 8080 # or whatever was specified using the PORT system variable. # Outputs via the HTTP port. This requires K_SERVICE to be set. # # On all other environments: Log to stdout # # syntax: entrypoint.py <command> # # Author: Chris Williams <diodesign@tuta.io> # import os import sys global command_result from flask import Flask app = Flask(__name__) # for Google Cloud Run @app.route('/') def ContainerService(): return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\n' if __name__ == "__main__": if (os.environ.get('K_SERVICE')) != '': print('Running HTTP service for Google Cloud') app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080))) else: print('Running locally') stream = os.popen('. $HOME/.cargo/env && cd /build/diosix && {}'.format(' '.join(sys.argv[1:]))) output = stream.read() output
739cf9a93afd9c742675e24cc637634e67d2c3b9
src/lavatory/utils/get_artifactory_info.py
src/lavatory/utils/get_artifactory_info.py
"""Helper method for getting artifactory information.""" import logging from .artifactory import Artifactory def get_artifactory_info(repo_names=None, repo_type='local'): """Get storage info from Artifactory. Args: repo_names (tuple, optional): Name of artifactory repo. repo_type (str): Type of artifactory repo. Returns: keys (dict, optional): Dictionary of repo data. storage_info (dict): Storage information api call. """ artifactory = Artifactory(repo_name=repo_names) storage_info = artifactory.repos(repo_type=repo_type) if repo_names: keys = repo_names else: keys = storage_info.keys() logging.debug('Storage info: %s', storage_info) logging.debug('Keys: %s', keys) return storage_info, keys
"""Helper method for getting artifactory information.""" import logging import requests from .artifactory import Artifactory def _artifactory(artifactory=None, repo_names=None): if not artifactory: artifactory = Artifactory(repo_name=repo_names) return artifactory def get_storage(repo_names=None, repo_type=None): artifactory = _artifactory(repo_names=repo_names) storage_info = [] try: storage_info = artifactory.repos(repo_type=repo_type) except requests.exceptions.HTTPError: logging.warning('Account is not an admin and may not be able to get storage details.') logging.debug('Storage info: %s', storage_info) return storage_info def get_repos(repo_names=None, repo_type='local'): repos = [] if repo_names: repos = repo_names else: repos = get_storage(repo_names=repo_names, repo_type=repo_type) return repos def get_artifactory_info(repo_names=None, repo_type='local'): """Get storage info from Artifactory. Args: repo_names (tuple, optional): Name of artifactory repo. repo_type (str): Type of artifactory repo. Returns: keys (dict, optional): Dictionary of repo data. storage_info (dict): Storage information api call. """ artifactory = Artifactory(repo_name=repo_names) storage_info = artifactory.repos(repo_type=repo_type) if repo_names: keys = repo_names else: keys = storage_info.keys() logging.debug('Storage info: %s', storage_info) logging.debug('Keys: %s', keys) return storage_info, keys
Add storage and repo helper functions
feat: Add storage and repo helper functions
Python
apache-2.0
gogoair/lavatory
"""Helper method for getting artifactory information.""" import logging from .artifactory import Artifactory def get_artifactory_info(repo_names=None, repo_type='local'): """Get storage info from Artifactory. Args: repo_names (tuple, optional): Name of artifactory repo. repo_type (str): Type of artifactory repo. Returns: keys (dict, optional): Dictionary of repo data. storage_info (dict): Storage information api call. """ artifactory = Artifactory(repo_name=repo_names) storage_info = artifactory.repos(repo_type=repo_type) if repo_names: keys = repo_names else: keys = storage_info.keys() logging.debug('Storage info: %s', storage_info) logging.debug('Keys: %s', keys) return storage_info, keys feat: Add storage and repo helper functions
"""Helper method for getting artifactory information.""" import logging import requests from .artifactory import Artifactory def _artifactory(artifactory=None, repo_names=None): if not artifactory: artifactory = Artifactory(repo_name=repo_names) return artifactory def get_storage(repo_names=None, repo_type=None): artifactory = _artifactory(repo_names=repo_names) storage_info = [] try: storage_info = artifactory.repos(repo_type=repo_type) except requests.exceptions.HTTPError: logging.warning('Account is not an admin and may not be able to get storage details.') logging.debug('Storage info: %s', storage_info) return storage_info def get_repos(repo_names=None, repo_type='local'): repos = [] if repo_names: repos = repo_names else: repos = get_storage(repo_names=repo_names, repo_type=repo_type) return repos def get_artifactory_info(repo_names=None, repo_type='local'): """Get storage info from Artifactory. Args: repo_names (tuple, optional): Name of artifactory repo. repo_type (str): Type of artifactory repo. Returns: keys (dict, optional): Dictionary of repo data. storage_info (dict): Storage information api call. """ artifactory = Artifactory(repo_name=repo_names) storage_info = artifactory.repos(repo_type=repo_type) if repo_names: keys = repo_names else: keys = storage_info.keys() logging.debug('Storage info: %s', storage_info) logging.debug('Keys: %s', keys) return storage_info, keys
<commit_before>"""Helper method for getting artifactory information.""" import logging from .artifactory import Artifactory def get_artifactory_info(repo_names=None, repo_type='local'): """Get storage info from Artifactory. Args: repo_names (tuple, optional): Name of artifactory repo. repo_type (str): Type of artifactory repo. Returns: keys (dict, optional): Dictionary of repo data. storage_info (dict): Storage information api call. """ artifactory = Artifactory(repo_name=repo_names) storage_info = artifactory.repos(repo_type=repo_type) if repo_names: keys = repo_names else: keys = storage_info.keys() logging.debug('Storage info: %s', storage_info) logging.debug('Keys: %s', keys) return storage_info, keys <commit_msg>feat: Add storage and repo helper functions<commit_after>
"""Helper method for getting artifactory information.""" import logging import requests from .artifactory import Artifactory def _artifactory(artifactory=None, repo_names=None): if not artifactory: artifactory = Artifactory(repo_name=repo_names) return artifactory def get_storage(repo_names=None, repo_type=None): artifactory = _artifactory(repo_names=repo_names) storage_info = [] try: storage_info = artifactory.repos(repo_type=repo_type) except requests.exceptions.HTTPError: logging.warning('Account is not an admin and may not be able to get storage details.') logging.debug('Storage info: %s', storage_info) return storage_info def get_repos(repo_names=None, repo_type='local'): repos = [] if repo_names: repos = repo_names else: repos = get_storage(repo_names=repo_names, repo_type=repo_type) return repos def get_artifactory_info(repo_names=None, repo_type='local'): """Get storage info from Artifactory. Args: repo_names (tuple, optional): Name of artifactory repo. repo_type (str): Type of artifactory repo. Returns: keys (dict, optional): Dictionary of repo data. storage_info (dict): Storage information api call. """ artifactory = Artifactory(repo_name=repo_names) storage_info = artifactory.repos(repo_type=repo_type) if repo_names: keys = repo_names else: keys = storage_info.keys() logging.debug('Storage info: %s', storage_info) logging.debug('Keys: %s', keys) return storage_info, keys
"""Helper method for getting artifactory information.""" import logging from .artifactory import Artifactory def get_artifactory_info(repo_names=None, repo_type='local'): """Get storage info from Artifactory. Args: repo_names (tuple, optional): Name of artifactory repo. repo_type (str): Type of artifactory repo. Returns: keys (dict, optional): Dictionary of repo data. storage_info (dict): Storage information api call. """ artifactory = Artifactory(repo_name=repo_names) storage_info = artifactory.repos(repo_type=repo_type) if repo_names: keys = repo_names else: keys = storage_info.keys() logging.debug('Storage info: %s', storage_info) logging.debug('Keys: %s', keys) return storage_info, keys feat: Add storage and repo helper functions"""Helper method for getting artifactory information.""" import logging import requests from .artifactory import Artifactory def _artifactory(artifactory=None, repo_names=None): if not artifactory: artifactory = Artifactory(repo_name=repo_names) return artifactory def get_storage(repo_names=None, repo_type=None): artifactory = _artifactory(repo_names=repo_names) storage_info = [] try: storage_info = artifactory.repos(repo_type=repo_type) except requests.exceptions.HTTPError: logging.warning('Account is not an admin and may not be able to get storage details.') logging.debug('Storage info: %s', storage_info) return storage_info def get_repos(repo_names=None, repo_type='local'): repos = [] if repo_names: repos = repo_names else: repos = get_storage(repo_names=repo_names, repo_type=repo_type) return repos def get_artifactory_info(repo_names=None, repo_type='local'): """Get storage info from Artifactory. Args: repo_names (tuple, optional): Name of artifactory repo. repo_type (str): Type of artifactory repo. Returns: keys (dict, optional): Dictionary of repo data. storage_info (dict): Storage information api call. """ artifactory = Artifactory(repo_name=repo_names) storage_info = artifactory.repos(repo_type=repo_type) if repo_names: keys = repo_names else: keys = storage_info.keys() logging.debug('Storage info: %s', storage_info) logging.debug('Keys: %s', keys) return storage_info, keys
<commit_before>"""Helper method for getting artifactory information.""" import logging from .artifactory import Artifactory def get_artifactory_info(repo_names=None, repo_type='local'): """Get storage info from Artifactory. Args: repo_names (tuple, optional): Name of artifactory repo. repo_type (str): Type of artifactory repo. Returns: keys (dict, optional): Dictionary of repo data. storage_info (dict): Storage information api call. """ artifactory = Artifactory(repo_name=repo_names) storage_info = artifactory.repos(repo_type=repo_type) if repo_names: keys = repo_names else: keys = storage_info.keys() logging.debug('Storage info: %s', storage_info) logging.debug('Keys: %s', keys) return storage_info, keys <commit_msg>feat: Add storage and repo helper functions<commit_after>"""Helper method for getting artifactory information.""" import logging import requests from .artifactory import Artifactory def _artifactory(artifactory=None, repo_names=None): if not artifactory: artifactory = Artifactory(repo_name=repo_names) return artifactory def get_storage(repo_names=None, repo_type=None): artifactory = _artifactory(repo_names=repo_names) storage_info = [] try: storage_info = artifactory.repos(repo_type=repo_type) except requests.exceptions.HTTPError: logging.warning('Account is not an admin and may not be able to get storage details.') logging.debug('Storage info: %s', storage_info) return storage_info def get_repos(repo_names=None, repo_type='local'): repos = [] if repo_names: repos = repo_names else: repos = get_storage(repo_names=repo_names, repo_type=repo_type) return repos def get_artifactory_info(repo_names=None, repo_type='local'): """Get storage info from Artifactory. Args: repo_names (tuple, optional): Name of artifactory repo. repo_type (str): Type of artifactory repo. Returns: keys (dict, optional): Dictionary of repo data. storage_info (dict): Storage information api call. """ artifactory = Artifactory(repo_name=repo_names) storage_info = artifactory.repos(repo_type=repo_type) if repo_names: keys = repo_names else: keys = storage_info.keys() logging.debug('Storage info: %s', storage_info) logging.debug('Keys: %s', keys) return storage_info, keys
06b99c4415a6605cbd6123271d44af96585fbb9d
conda_env/exceptions.py
conda_env/exceptions.py
class CondaEnvException(Exception): pass class EnvironmentFileNotFound(CondaEnvException): def __init__(self, filename, *args, **kwargs): msg = '{} file not found'.format(filename) self.filename = filename super(EnvironmentFileNotFound, self).__init__(msg, *args, **kwargs)
class CondaEnvException(Exception): pass class EnvironmentFileNotFound(CondaEnvException): def __init__(self, filename, *args, **kwargs): msg = '{} file not found'.format(filename) self.filename = filename super(EnvironmentFileNotFound, self).__init__(msg, *args, **kwargs) class EnvironmentFileNotDownloaded(CondaEnvException): def __init__(self, filename, *args, **kwargs): msg = '{} file not downloaded'.format(filename) self.filename = filename super(EnvironmentFileNotDownloaded, self).__init__(msg, *args, **kwargs)
Add environment not found exception
Add environment not found exception
Python
bsd-3-clause
nicoddemus/conda-env,mikecroucher/conda-env,dan-blanchard/conda-env,conda/conda-env,dan-blanchard/conda-env,phobson/conda-env,conda/conda-env,ESSS/conda-env,mikecroucher/conda-env,asmeurer/conda-env,nicoddemus/conda-env,ESSS/conda-env,phobson/conda-env,isaac-kit/conda-env,asmeurer/conda-env,isaac-kit/conda-env
class CondaEnvException(Exception): pass class EnvironmentFileNotFound(CondaEnvException): def __init__(self, filename, *args, **kwargs): msg = '{} file not found'.format(filename) self.filename = filename super(EnvironmentFileNotFound, self).__init__(msg, *args, **kwargs) Add environment not found exception
class CondaEnvException(Exception): pass class EnvironmentFileNotFound(CondaEnvException): def __init__(self, filename, *args, **kwargs): msg = '{} file not found'.format(filename) self.filename = filename super(EnvironmentFileNotFound, self).__init__(msg, *args, **kwargs) class EnvironmentFileNotDownloaded(CondaEnvException): def __init__(self, filename, *args, **kwargs): msg = '{} file not downloaded'.format(filename) self.filename = filename super(EnvironmentFileNotDownloaded, self).__init__(msg, *args, **kwargs)
<commit_before>class CondaEnvException(Exception): pass class EnvironmentFileNotFound(CondaEnvException): def __init__(self, filename, *args, **kwargs): msg = '{} file not found'.format(filename) self.filename = filename super(EnvironmentFileNotFound, self).__init__(msg, *args, **kwargs) <commit_msg>Add environment not found exception<commit_after>
class CondaEnvException(Exception): pass class EnvironmentFileNotFound(CondaEnvException): def __init__(self, filename, *args, **kwargs): msg = '{} file not found'.format(filename) self.filename = filename super(EnvironmentFileNotFound, self).__init__(msg, *args, **kwargs) class EnvironmentFileNotDownloaded(CondaEnvException): def __init__(self, filename, *args, **kwargs): msg = '{} file not downloaded'.format(filename) self.filename = filename super(EnvironmentFileNotDownloaded, self).__init__(msg, *args, **kwargs)
class CondaEnvException(Exception): pass class EnvironmentFileNotFound(CondaEnvException): def __init__(self, filename, *args, **kwargs): msg = '{} file not found'.format(filename) self.filename = filename super(EnvironmentFileNotFound, self).__init__(msg, *args, **kwargs) Add environment not found exceptionclass CondaEnvException(Exception): pass class EnvironmentFileNotFound(CondaEnvException): def __init__(self, filename, *args, **kwargs): msg = '{} file not found'.format(filename) self.filename = filename super(EnvironmentFileNotFound, self).__init__(msg, *args, **kwargs) class EnvironmentFileNotDownloaded(CondaEnvException): def __init__(self, filename, *args, **kwargs): msg = '{} file not downloaded'.format(filename) self.filename = filename super(EnvironmentFileNotDownloaded, self).__init__(msg, *args, **kwargs)
<commit_before>class CondaEnvException(Exception): pass class EnvironmentFileNotFound(CondaEnvException): def __init__(self, filename, *args, **kwargs): msg = '{} file not found'.format(filename) self.filename = filename super(EnvironmentFileNotFound, self).__init__(msg, *args, **kwargs) <commit_msg>Add environment not found exception<commit_after>class CondaEnvException(Exception): pass class EnvironmentFileNotFound(CondaEnvException): def __init__(self, filename, *args, **kwargs): msg = '{} file not found'.format(filename) self.filename = filename super(EnvironmentFileNotFound, self).__init__(msg, *args, **kwargs) class EnvironmentFileNotDownloaded(CondaEnvException): def __init__(self, filename, *args, **kwargs): msg = '{} file not downloaded'.format(filename) self.filename = filename super(EnvironmentFileNotDownloaded, self).__init__(msg, *args, **kwargs)
484f42b6fc1a8129a53480bc6e7913c5c7d58f46
froide/foirequest/search_indexes.py
froide/foirequest/search_indexes.py
from haystack import indexes from celery_haystack.indexes import CelerySearchIndex from .models import FoiRequest class FoiRequestIndex(CelerySearchIndex, indexes.Indexable): text = indexes.EdgeNgramField(document=True, use_template=True) title = indexes.CharField(model_attr='title') description = indexes.CharField(model_attr='description') resolution = indexes.CharField(model_attr='resolution', default="") status = indexes.CharField(model_attr='status') readable_status = indexes.CharField(model_attr='readable_status') first_message = indexes.DateTimeField(model_attr='first_message') last_message = indexes.DateTimeField(model_attr='last_message') url = indexes.CharField(model_attr='get_absolute_url') public_body_name = indexes.CharField(model_attr='public_body__name', default="") def get_model(self): return FoiRequest def index_queryset(self): """Used when the entire index for model is updated.""" return self.get_model().published.get_for_search_index() def should_update(self, instance, **kwargs): return instance.visibility > 1
from haystack import indexes from celery_haystack.indexes import CelerySearchIndex from .models import FoiRequest class FoiRequestIndex(CelerySearchIndex, indexes.Indexable): text = indexes.EdgeNgramField(document=True, use_template=True) title = indexes.CharField(model_attr='title') description = indexes.CharField(model_attr='description') resolution = indexes.CharField(model_attr='resolution', default="") status = indexes.CharField(model_attr='status') readable_status = indexes.CharField(model_attr='readable_status') first_message = indexes.DateTimeField(model_attr='first_message') last_message = indexes.DateTimeField(model_attr='last_message') url = indexes.CharField(model_attr='get_absolute_url') public_body_name = indexes.CharField(model_attr='public_body__name', default="") def get_model(self): return FoiRequest def index_queryset(self): """Used when the entire index for model is updated.""" return self.get_model().published.get_for_search_index() def should_update(self, instance, **kwargs): return instance.visibility > 1 and instance.is_foi
Index only FoiRequests marked is_foi
Index only FoiRequests marked is_foi
Python
mit
fin/froide,CodeforHawaii/froide,catcosmo/froide,okfse/froide,LilithWittmann/froide,stefanw/froide,catcosmo/froide,catcosmo/froide,okfse/froide,fin/froide,ryankanno/froide,stefanw/froide,CodeforHawaii/froide,okfse/froide,LilithWittmann/froide,CodeforHawaii/froide,ryankanno/froide,ryankanno/froide,stefanw/froide,LilithWittmann/froide,okfse/froide,ryankanno/froide,stefanw/froide,okfse/froide,LilithWittmann/froide,LilithWittmann/froide,stefanw/froide,fin/froide,catcosmo/froide,CodeforHawaii/froide,catcosmo/froide,CodeforHawaii/froide,fin/froide,ryankanno/froide
from haystack import indexes from celery_haystack.indexes import CelerySearchIndex from .models import FoiRequest class FoiRequestIndex(CelerySearchIndex, indexes.Indexable): text = indexes.EdgeNgramField(document=True, use_template=True) title = indexes.CharField(model_attr='title') description = indexes.CharField(model_attr='description') resolution = indexes.CharField(model_attr='resolution', default="") status = indexes.CharField(model_attr='status') readable_status = indexes.CharField(model_attr='readable_status') first_message = indexes.DateTimeField(model_attr='first_message') last_message = indexes.DateTimeField(model_attr='last_message') url = indexes.CharField(model_attr='get_absolute_url') public_body_name = indexes.CharField(model_attr='public_body__name', default="") def get_model(self): return FoiRequest def index_queryset(self): """Used when the entire index for model is updated.""" return self.get_model().published.get_for_search_index() def should_update(self, instance, **kwargs): return instance.visibility > 1 Index only FoiRequests marked is_foi
from haystack import indexes from celery_haystack.indexes import CelerySearchIndex from .models import FoiRequest class FoiRequestIndex(CelerySearchIndex, indexes.Indexable): text = indexes.EdgeNgramField(document=True, use_template=True) title = indexes.CharField(model_attr='title') description = indexes.CharField(model_attr='description') resolution = indexes.CharField(model_attr='resolution', default="") status = indexes.CharField(model_attr='status') readable_status = indexes.CharField(model_attr='readable_status') first_message = indexes.DateTimeField(model_attr='first_message') last_message = indexes.DateTimeField(model_attr='last_message') url = indexes.CharField(model_attr='get_absolute_url') public_body_name = indexes.CharField(model_attr='public_body__name', default="") def get_model(self): return FoiRequest def index_queryset(self): """Used when the entire index for model is updated.""" return self.get_model().published.get_for_search_index() def should_update(self, instance, **kwargs): return instance.visibility > 1 and instance.is_foi
<commit_before>from haystack import indexes from celery_haystack.indexes import CelerySearchIndex from .models import FoiRequest class FoiRequestIndex(CelerySearchIndex, indexes.Indexable): text = indexes.EdgeNgramField(document=True, use_template=True) title = indexes.CharField(model_attr='title') description = indexes.CharField(model_attr='description') resolution = indexes.CharField(model_attr='resolution', default="") status = indexes.CharField(model_attr='status') readable_status = indexes.CharField(model_attr='readable_status') first_message = indexes.DateTimeField(model_attr='first_message') last_message = indexes.DateTimeField(model_attr='last_message') url = indexes.CharField(model_attr='get_absolute_url') public_body_name = indexes.CharField(model_attr='public_body__name', default="") def get_model(self): return FoiRequest def index_queryset(self): """Used when the entire index for model is updated.""" return self.get_model().published.get_for_search_index() def should_update(self, instance, **kwargs): return instance.visibility > 1 <commit_msg>Index only FoiRequests marked is_foi<commit_after>
from haystack import indexes from celery_haystack.indexes import CelerySearchIndex from .models import FoiRequest class FoiRequestIndex(CelerySearchIndex, indexes.Indexable): text = indexes.EdgeNgramField(document=True, use_template=True) title = indexes.CharField(model_attr='title') description = indexes.CharField(model_attr='description') resolution = indexes.CharField(model_attr='resolution', default="") status = indexes.CharField(model_attr='status') readable_status = indexes.CharField(model_attr='readable_status') first_message = indexes.DateTimeField(model_attr='first_message') last_message = indexes.DateTimeField(model_attr='last_message') url = indexes.CharField(model_attr='get_absolute_url') public_body_name = indexes.CharField(model_attr='public_body__name', default="") def get_model(self): return FoiRequest def index_queryset(self): """Used when the entire index for model is updated.""" return self.get_model().published.get_for_search_index() def should_update(self, instance, **kwargs): return instance.visibility > 1 and instance.is_foi
from haystack import indexes from celery_haystack.indexes import CelerySearchIndex from .models import FoiRequest class FoiRequestIndex(CelerySearchIndex, indexes.Indexable): text = indexes.EdgeNgramField(document=True, use_template=True) title = indexes.CharField(model_attr='title') description = indexes.CharField(model_attr='description') resolution = indexes.CharField(model_attr='resolution', default="") status = indexes.CharField(model_attr='status') readable_status = indexes.CharField(model_attr='readable_status') first_message = indexes.DateTimeField(model_attr='first_message') last_message = indexes.DateTimeField(model_attr='last_message') url = indexes.CharField(model_attr='get_absolute_url') public_body_name = indexes.CharField(model_attr='public_body__name', default="") def get_model(self): return FoiRequest def index_queryset(self): """Used when the entire index for model is updated.""" return self.get_model().published.get_for_search_index() def should_update(self, instance, **kwargs): return instance.visibility > 1 Index only FoiRequests marked is_foifrom haystack import indexes from celery_haystack.indexes import CelerySearchIndex from .models import FoiRequest class FoiRequestIndex(CelerySearchIndex, indexes.Indexable): text = indexes.EdgeNgramField(document=True, use_template=True) title = indexes.CharField(model_attr='title') description = indexes.CharField(model_attr='description') resolution = indexes.CharField(model_attr='resolution', default="") status = indexes.CharField(model_attr='status') readable_status = indexes.CharField(model_attr='readable_status') first_message = indexes.DateTimeField(model_attr='first_message') last_message = indexes.DateTimeField(model_attr='last_message') url = indexes.CharField(model_attr='get_absolute_url') public_body_name = indexes.CharField(model_attr='public_body__name', default="") def get_model(self): return FoiRequest def index_queryset(self): """Used when the entire index for model is updated.""" return self.get_model().published.get_for_search_index() def should_update(self, instance, **kwargs): return instance.visibility > 1 and instance.is_foi
<commit_before>from haystack import indexes from celery_haystack.indexes import CelerySearchIndex from .models import FoiRequest class FoiRequestIndex(CelerySearchIndex, indexes.Indexable): text = indexes.EdgeNgramField(document=True, use_template=True) title = indexes.CharField(model_attr='title') description = indexes.CharField(model_attr='description') resolution = indexes.CharField(model_attr='resolution', default="") status = indexes.CharField(model_attr='status') readable_status = indexes.CharField(model_attr='readable_status') first_message = indexes.DateTimeField(model_attr='first_message') last_message = indexes.DateTimeField(model_attr='last_message') url = indexes.CharField(model_attr='get_absolute_url') public_body_name = indexes.CharField(model_attr='public_body__name', default="") def get_model(self): return FoiRequest def index_queryset(self): """Used when the entire index for model is updated.""" return self.get_model().published.get_for_search_index() def should_update(self, instance, **kwargs): return instance.visibility > 1 <commit_msg>Index only FoiRequests marked is_foi<commit_after>from haystack import indexes from celery_haystack.indexes import CelerySearchIndex from .models import FoiRequest class FoiRequestIndex(CelerySearchIndex, indexes.Indexable): text = indexes.EdgeNgramField(document=True, use_template=True) title = indexes.CharField(model_attr='title') description = indexes.CharField(model_attr='description') resolution = indexes.CharField(model_attr='resolution', default="") status = indexes.CharField(model_attr='status') readable_status = indexes.CharField(model_attr='readable_status') first_message = indexes.DateTimeField(model_attr='first_message') last_message = indexes.DateTimeField(model_attr='last_message') url = indexes.CharField(model_attr='get_absolute_url') public_body_name = indexes.CharField(model_attr='public_body__name', default="") def get_model(self): return FoiRequest def index_queryset(self): """Used when the entire index for model is updated.""" return self.get_model().published.get_for_search_index() def should_update(self, instance, **kwargs): return instance.visibility > 1 and instance.is_foi
359595413071ff706b484a875a23a4a7d1508f50
bindings/python/llvm/tests/base.py
bindings/python/llvm/tests/base.py
import os.path import unittest POSSIBLE_TEST_BINARIES = [ 'libreadline.so.5', 'libreadline.so.6', ] POSSIBLE_TEST_BINARY_PATHS = [ '/lib', '/usr/lib', '/usr/local/lib', ] class TestBase(unittest.TestCase): def get_test_binary(self): """Helper to obtain a test binary for object file testing. FIXME Support additional, highly-likely targets or create one ourselves. """ for d in POSSIBLE_TEST_BINARY_PATHS: for lib in POSSIBLE_TEST_BINARIES: path = os.path.join(d, lib) if os.path.exists(path): return path raise Exception('No suitable test binaries available!')
import os.path import unittest POSSIBLE_TEST_BINARIES = [ 'libreadline.so.5', 'libreadline.so.6', ] POSSIBLE_TEST_BINARY_PATHS = [ '/lib', '/usr/lib', '/usr/local/lib', ] class TestBase(unittest.TestCase): def get_test_binary(self): """Helper to obtain a test binary for object file testing. FIXME Support additional, highly-likely targets or create one ourselves. """ for d in POSSIBLE_TEST_BINARY_PATHS: for lib in POSSIBLE_TEST_BINARIES: path = os.path.join(d, lib) if os.path.exists(path): return path raise Exception('No suitable test binaries available!') get_test_binary.__test__ = False
Mark get_test_binary as not being a test
[python] Mark get_test_binary as not being a test get_test_binary is a helper method, not a test, make sure nosetests doesn't pick it up as a test. git-svn-id: 0ff597fd157e6f4fc38580e8d64ab130330d2411@153173 91177308-0d34-0410-b5e6-96231b3b80d8
Python
apache-2.0
llvm-mirror/llvm,GPUOpen-Drivers/llvm,GPUOpen-Drivers/llvm,llvm-mirror/llvm,apple/swift-llvm,llvm-mirror/llvm,apple/swift-llvm,GPUOpen-Drivers/llvm,GPUOpen-Drivers/llvm,chubbymaggie/asap,dslab-epfl/asap,llvm-mirror/llvm,llvm-mirror/llvm,apple/swift-llvm,llvm-mirror/llvm,apple/swift-llvm,GPUOpen-Drivers/llvm,apple/swift-llvm,chubbymaggie/asap,chubbymaggie/asap,dslab-epfl/asap,dslab-epfl/asap,llvm-mirror/llvm,chubbymaggie/asap,llvm-mirror/llvm,dslab-epfl/asap,llvm-mirror/llvm,apple/swift-llvm,dslab-epfl/asap,apple/swift-llvm,apple/swift-llvm,dslab-epfl/asap,dslab-epfl/asap,GPUOpen-Drivers/llvm,chubbymaggie/asap,chubbymaggie/asap,GPUOpen-Drivers/llvm,GPUOpen-Drivers/llvm
import os.path import unittest POSSIBLE_TEST_BINARIES = [ 'libreadline.so.5', 'libreadline.so.6', ] POSSIBLE_TEST_BINARY_PATHS = [ '/lib', '/usr/lib', '/usr/local/lib', ] class TestBase(unittest.TestCase): def get_test_binary(self): """Helper to obtain a test binary for object file testing. FIXME Support additional, highly-likely targets or create one ourselves. """ for d in POSSIBLE_TEST_BINARY_PATHS: for lib in POSSIBLE_TEST_BINARIES: path = os.path.join(d, lib) if os.path.exists(path): return path raise Exception('No suitable test binaries available!') [python] Mark get_test_binary as not being a test get_test_binary is a helper method, not a test, make sure nosetests doesn't pick it up as a test. git-svn-id: 0ff597fd157e6f4fc38580e8d64ab130330d2411@153173 91177308-0d34-0410-b5e6-96231b3b80d8
import os.path import unittest POSSIBLE_TEST_BINARIES = [ 'libreadline.so.5', 'libreadline.so.6', ] POSSIBLE_TEST_BINARY_PATHS = [ '/lib', '/usr/lib', '/usr/local/lib', ] class TestBase(unittest.TestCase): def get_test_binary(self): """Helper to obtain a test binary for object file testing. FIXME Support additional, highly-likely targets or create one ourselves. """ for d in POSSIBLE_TEST_BINARY_PATHS: for lib in POSSIBLE_TEST_BINARIES: path = os.path.join(d, lib) if os.path.exists(path): return path raise Exception('No suitable test binaries available!') get_test_binary.__test__ = False
<commit_before>import os.path import unittest POSSIBLE_TEST_BINARIES = [ 'libreadline.so.5', 'libreadline.so.6', ] POSSIBLE_TEST_BINARY_PATHS = [ '/lib', '/usr/lib', '/usr/local/lib', ] class TestBase(unittest.TestCase): def get_test_binary(self): """Helper to obtain a test binary for object file testing. FIXME Support additional, highly-likely targets or create one ourselves. """ for d in POSSIBLE_TEST_BINARY_PATHS: for lib in POSSIBLE_TEST_BINARIES: path = os.path.join(d, lib) if os.path.exists(path): return path raise Exception('No suitable test binaries available!') <commit_msg>[python] Mark get_test_binary as not being a test get_test_binary is a helper method, not a test, make sure nosetests doesn't pick it up as a test. git-svn-id: 0ff597fd157e6f4fc38580e8d64ab130330d2411@153173 91177308-0d34-0410-b5e6-96231b3b80d8<commit_after>
import os.path import unittest POSSIBLE_TEST_BINARIES = [ 'libreadline.so.5', 'libreadline.so.6', ] POSSIBLE_TEST_BINARY_PATHS = [ '/lib', '/usr/lib', '/usr/local/lib', ] class TestBase(unittest.TestCase): def get_test_binary(self): """Helper to obtain a test binary for object file testing. FIXME Support additional, highly-likely targets or create one ourselves. """ for d in POSSIBLE_TEST_BINARY_PATHS: for lib in POSSIBLE_TEST_BINARIES: path = os.path.join(d, lib) if os.path.exists(path): return path raise Exception('No suitable test binaries available!') get_test_binary.__test__ = False
import os.path import unittest POSSIBLE_TEST_BINARIES = [ 'libreadline.so.5', 'libreadline.so.6', ] POSSIBLE_TEST_BINARY_PATHS = [ '/lib', '/usr/lib', '/usr/local/lib', ] class TestBase(unittest.TestCase): def get_test_binary(self): """Helper to obtain a test binary for object file testing. FIXME Support additional, highly-likely targets or create one ourselves. """ for d in POSSIBLE_TEST_BINARY_PATHS: for lib in POSSIBLE_TEST_BINARIES: path = os.path.join(d, lib) if os.path.exists(path): return path raise Exception('No suitable test binaries available!') [python] Mark get_test_binary as not being a test get_test_binary is a helper method, not a test, make sure nosetests doesn't pick it up as a test. git-svn-id: 0ff597fd157e6f4fc38580e8d64ab130330d2411@153173 91177308-0d34-0410-b5e6-96231b3b80d8import os.path import unittest POSSIBLE_TEST_BINARIES = [ 'libreadline.so.5', 'libreadline.so.6', ] POSSIBLE_TEST_BINARY_PATHS = [ '/lib', '/usr/lib', '/usr/local/lib', ] class TestBase(unittest.TestCase): def get_test_binary(self): """Helper to obtain a test binary for object file testing. FIXME Support additional, highly-likely targets or create one ourselves. """ for d in POSSIBLE_TEST_BINARY_PATHS: for lib in POSSIBLE_TEST_BINARIES: path = os.path.join(d, lib) if os.path.exists(path): return path raise Exception('No suitable test binaries available!') get_test_binary.__test__ = False
<commit_before>import os.path import unittest POSSIBLE_TEST_BINARIES = [ 'libreadline.so.5', 'libreadline.so.6', ] POSSIBLE_TEST_BINARY_PATHS = [ '/lib', '/usr/lib', '/usr/local/lib', ] class TestBase(unittest.TestCase): def get_test_binary(self): """Helper to obtain a test binary for object file testing. FIXME Support additional, highly-likely targets or create one ourselves. """ for d in POSSIBLE_TEST_BINARY_PATHS: for lib in POSSIBLE_TEST_BINARIES: path = os.path.join(d, lib) if os.path.exists(path): return path raise Exception('No suitable test binaries available!') <commit_msg>[python] Mark get_test_binary as not being a test get_test_binary is a helper method, not a test, make sure nosetests doesn't pick it up as a test. git-svn-id: 0ff597fd157e6f4fc38580e8d64ab130330d2411@153173 91177308-0d34-0410-b5e6-96231b3b80d8<commit_after>import os.path import unittest POSSIBLE_TEST_BINARIES = [ 'libreadline.so.5', 'libreadline.so.6', ] POSSIBLE_TEST_BINARY_PATHS = [ '/lib', '/usr/lib', '/usr/local/lib', ] class TestBase(unittest.TestCase): def get_test_binary(self): """Helper to obtain a test binary for object file testing. FIXME Support additional, highly-likely targets or create one ourselves. """ for d in POSSIBLE_TEST_BINARY_PATHS: for lib in POSSIBLE_TEST_BINARIES: path = os.path.join(d, lib) if os.path.exists(path): return path raise Exception('No suitable test binaries available!') get_test_binary.__test__ = False
50442966938b532cc759089692ffb52e94c6e89b
config_example.py
config_example.py
"""Example config.py""" webhook_urls = ["DISCORD WEBHOOK", "DISCORD WEBHOOK"] # Used to update webhooks on Discord key_path = "/path/to/key/in/format/of/file.pem" # Private key to sign the file file_path = "/path/to/folder" # Path to save the file to lzss_path = "/path/to/lzss" # Path to lzss production = None # Use production mode cachet_url = "http://status.domain.tld/api/v1" # URL for Cachet cachet_key = "api_key" # API Key for Cachet sentry_url = "http://status.domain.tld/" # URL for Sentry """News Channel only""" force_all = False # Force the script to replace all news files google_maps_api_key = "api_key" # API Key for Google Maps geocoding API geoparser_keys = ["GEOPARSER KEY", "GEOPARSER KEY"] # API Key for Geoparser """Forecast Channel only""" import forecastlists useVerbose = None # Print more verbose messages useMultithreaded = None # Use multithreading weathercities = ["CITY LIST", "CITY LIST"] # Lists of cities to use cachet_elapsed_time = None # ID of the Cachet point to log elapsed time. """Everybody Votes Channel only""" mysql_user = "user" # MySQL username mysql_password = "password" # MySQL password mysql_database = "database" # MySQL database
"""Example config.py""" webhook_urls = ["DISCORD WEBHOOK", "DISCORD WEBHOOK"] # Used to update webhooks on Discord key_path = "/path/to/key/in/format/of/file.pem" # Private key to sign the file file_path = "/path/to/folder" # Path to save the file to lzss_path = "/path/to/lzss" # Path to lzss production = None # Use production mode cachet_url = "http://status.domain.tld/api/v1" # URL for Cachet cachet_key = "api_key" # API Key for Cachet sentry_url = "http://status.domain.tld/" # URL for Sentry """News Channel only""" force_all = False # Force the script to replace all news files google_maps_api_key = "api_key" # API Key for Google Maps geocoding API geoparser_keys = ["GEOPARSER KEY", "GEOPARSER KEY"] # API Key for Geoparser """Forecast Channel only""" import forecastlists useVerbose = None # Print more verbose messages useMultithreaded = None # Use multithreading weathercities = ["CITY LIST", "CITY LIST"] # Lists of cities to use cachet_elapsed_time = None # ID of the Cachet point to log elapsed time. """Everybody Votes Channel only""" mysql_user = "user" # MySQL username mysql_password = "password" # MySQL password mysql_database = "database" # MySQL database
Fix PEP 8 coding violations
Fix PEP 8 coding violations
Python
agpl-3.0
RiiConnect24/File-Maker,RiiConnect24/File-Maker
"""Example config.py""" webhook_urls = ["DISCORD WEBHOOK", "DISCORD WEBHOOK"] # Used to update webhooks on Discord key_path = "/path/to/key/in/format/of/file.pem" # Private key to sign the file file_path = "/path/to/folder" # Path to save the file to lzss_path = "/path/to/lzss" # Path to lzss production = None # Use production mode cachet_url = "http://status.domain.tld/api/v1" # URL for Cachet cachet_key = "api_key" # API Key for Cachet sentry_url = "http://status.domain.tld/" # URL for Sentry """News Channel only""" force_all = False # Force the script to replace all news files google_maps_api_key = "api_key" # API Key for Google Maps geocoding API geoparser_keys = ["GEOPARSER KEY", "GEOPARSER KEY"] # API Key for Geoparser """Forecast Channel only""" import forecastlists useVerbose = None # Print more verbose messages useMultithreaded = None # Use multithreading weathercities = ["CITY LIST", "CITY LIST"] # Lists of cities to use cachet_elapsed_time = None # ID of the Cachet point to log elapsed time. """Everybody Votes Channel only""" mysql_user = "user" # MySQL username mysql_password = "password" # MySQL password mysql_database = "database" # MySQL database Fix PEP 8 coding violations
"""Example config.py""" webhook_urls = ["DISCORD WEBHOOK", "DISCORD WEBHOOK"] # Used to update webhooks on Discord key_path = "/path/to/key/in/format/of/file.pem" # Private key to sign the file file_path = "/path/to/folder" # Path to save the file to lzss_path = "/path/to/lzss" # Path to lzss production = None # Use production mode cachet_url = "http://status.domain.tld/api/v1" # URL for Cachet cachet_key = "api_key" # API Key for Cachet sentry_url = "http://status.domain.tld/" # URL for Sentry """News Channel only""" force_all = False # Force the script to replace all news files google_maps_api_key = "api_key" # API Key for Google Maps geocoding API geoparser_keys = ["GEOPARSER KEY", "GEOPARSER KEY"] # API Key for Geoparser """Forecast Channel only""" import forecastlists useVerbose = None # Print more verbose messages useMultithreaded = None # Use multithreading weathercities = ["CITY LIST", "CITY LIST"] # Lists of cities to use cachet_elapsed_time = None # ID of the Cachet point to log elapsed time. """Everybody Votes Channel only""" mysql_user = "user" # MySQL username mysql_password = "password" # MySQL password mysql_database = "database" # MySQL database
<commit_before>"""Example config.py""" webhook_urls = ["DISCORD WEBHOOK", "DISCORD WEBHOOK"] # Used to update webhooks on Discord key_path = "/path/to/key/in/format/of/file.pem" # Private key to sign the file file_path = "/path/to/folder" # Path to save the file to lzss_path = "/path/to/lzss" # Path to lzss production = None # Use production mode cachet_url = "http://status.domain.tld/api/v1" # URL for Cachet cachet_key = "api_key" # API Key for Cachet sentry_url = "http://status.domain.tld/" # URL for Sentry """News Channel only""" force_all = False # Force the script to replace all news files google_maps_api_key = "api_key" # API Key for Google Maps geocoding API geoparser_keys = ["GEOPARSER KEY", "GEOPARSER KEY"] # API Key for Geoparser """Forecast Channel only""" import forecastlists useVerbose = None # Print more verbose messages useMultithreaded = None # Use multithreading weathercities = ["CITY LIST", "CITY LIST"] # Lists of cities to use cachet_elapsed_time = None # ID of the Cachet point to log elapsed time. """Everybody Votes Channel only""" mysql_user = "user" # MySQL username mysql_password = "password" # MySQL password mysql_database = "database" # MySQL database <commit_msg>Fix PEP 8 coding violations<commit_after>
"""Example config.py""" webhook_urls = ["DISCORD WEBHOOK", "DISCORD WEBHOOK"] # Used to update webhooks on Discord key_path = "/path/to/key/in/format/of/file.pem" # Private key to sign the file file_path = "/path/to/folder" # Path to save the file to lzss_path = "/path/to/lzss" # Path to lzss production = None # Use production mode cachet_url = "http://status.domain.tld/api/v1" # URL for Cachet cachet_key = "api_key" # API Key for Cachet sentry_url = "http://status.domain.tld/" # URL for Sentry """News Channel only""" force_all = False # Force the script to replace all news files google_maps_api_key = "api_key" # API Key for Google Maps geocoding API geoparser_keys = ["GEOPARSER KEY", "GEOPARSER KEY"] # API Key for Geoparser """Forecast Channel only""" import forecastlists useVerbose = None # Print more verbose messages useMultithreaded = None # Use multithreading weathercities = ["CITY LIST", "CITY LIST"] # Lists of cities to use cachet_elapsed_time = None # ID of the Cachet point to log elapsed time. """Everybody Votes Channel only""" mysql_user = "user" # MySQL username mysql_password = "password" # MySQL password mysql_database = "database" # MySQL database
"""Example config.py""" webhook_urls = ["DISCORD WEBHOOK", "DISCORD WEBHOOK"] # Used to update webhooks on Discord key_path = "/path/to/key/in/format/of/file.pem" # Private key to sign the file file_path = "/path/to/folder" # Path to save the file to lzss_path = "/path/to/lzss" # Path to lzss production = None # Use production mode cachet_url = "http://status.domain.tld/api/v1" # URL for Cachet cachet_key = "api_key" # API Key for Cachet sentry_url = "http://status.domain.tld/" # URL for Sentry """News Channel only""" force_all = False # Force the script to replace all news files google_maps_api_key = "api_key" # API Key for Google Maps geocoding API geoparser_keys = ["GEOPARSER KEY", "GEOPARSER KEY"] # API Key for Geoparser """Forecast Channel only""" import forecastlists useVerbose = None # Print more verbose messages useMultithreaded = None # Use multithreading weathercities = ["CITY LIST", "CITY LIST"] # Lists of cities to use cachet_elapsed_time = None # ID of the Cachet point to log elapsed time. """Everybody Votes Channel only""" mysql_user = "user" # MySQL username mysql_password = "password" # MySQL password mysql_database = "database" # MySQL database Fix PEP 8 coding violations"""Example config.py""" webhook_urls = ["DISCORD WEBHOOK", "DISCORD WEBHOOK"] # Used to update webhooks on Discord key_path = "/path/to/key/in/format/of/file.pem" # Private key to sign the file file_path = "/path/to/folder" # Path to save the file to lzss_path = "/path/to/lzss" # Path to lzss production = None # Use production mode cachet_url = "http://status.domain.tld/api/v1" # URL for Cachet cachet_key = "api_key" # API Key for Cachet sentry_url = "http://status.domain.tld/" # URL for Sentry """News Channel only""" force_all = False # Force the script to replace all news files google_maps_api_key = "api_key" # API Key for Google Maps geocoding API geoparser_keys = ["GEOPARSER KEY", "GEOPARSER KEY"] # API Key for Geoparser """Forecast Channel only""" import forecastlists useVerbose = None # Print more verbose messages useMultithreaded = None # Use multithreading weathercities = ["CITY LIST", "CITY LIST"] # Lists of cities to use cachet_elapsed_time = None # ID of the Cachet point to log elapsed time. """Everybody Votes Channel only""" mysql_user = "user" # MySQL username mysql_password = "password" # MySQL password mysql_database = "database" # MySQL database
<commit_before>"""Example config.py""" webhook_urls = ["DISCORD WEBHOOK", "DISCORD WEBHOOK"] # Used to update webhooks on Discord key_path = "/path/to/key/in/format/of/file.pem" # Private key to sign the file file_path = "/path/to/folder" # Path to save the file to lzss_path = "/path/to/lzss" # Path to lzss production = None # Use production mode cachet_url = "http://status.domain.tld/api/v1" # URL for Cachet cachet_key = "api_key" # API Key for Cachet sentry_url = "http://status.domain.tld/" # URL for Sentry """News Channel only""" force_all = False # Force the script to replace all news files google_maps_api_key = "api_key" # API Key for Google Maps geocoding API geoparser_keys = ["GEOPARSER KEY", "GEOPARSER KEY"] # API Key for Geoparser """Forecast Channel only""" import forecastlists useVerbose = None # Print more verbose messages useMultithreaded = None # Use multithreading weathercities = ["CITY LIST", "CITY LIST"] # Lists of cities to use cachet_elapsed_time = None # ID of the Cachet point to log elapsed time. """Everybody Votes Channel only""" mysql_user = "user" # MySQL username mysql_password = "password" # MySQL password mysql_database = "database" # MySQL database <commit_msg>Fix PEP 8 coding violations<commit_after>"""Example config.py""" webhook_urls = ["DISCORD WEBHOOK", "DISCORD WEBHOOK"] # Used to update webhooks on Discord key_path = "/path/to/key/in/format/of/file.pem" # Private key to sign the file file_path = "/path/to/folder" # Path to save the file to lzss_path = "/path/to/lzss" # Path to lzss production = None # Use production mode cachet_url = "http://status.domain.tld/api/v1" # URL for Cachet cachet_key = "api_key" # API Key for Cachet sentry_url = "http://status.domain.tld/" # URL for Sentry """News Channel only""" force_all = False # Force the script to replace all news files google_maps_api_key = "api_key" # API Key for Google Maps geocoding API geoparser_keys = ["GEOPARSER KEY", "GEOPARSER KEY"] # API Key for Geoparser """Forecast Channel only""" import forecastlists useVerbose = None # Print more verbose messages useMultithreaded = None # Use multithreading weathercities = ["CITY LIST", "CITY LIST"] # Lists of cities to use cachet_elapsed_time = None # ID of the Cachet point to log elapsed time. """Everybody Votes Channel only""" mysql_user = "user" # MySQL username mysql_password = "password" # MySQL password mysql_database = "database" # MySQL database
b085d519da9869be8c4bc4f56cb0e040a6b1525b
build/combine.py
build/combine.py
import os, sys, re from simplejson import load as json from simplejson import dumps as dump from glob import glob VERSION = 0.1 all = [] for p in glob("../plugins/*.json"): fp = open(p, "r") x = json(fp, "utf-8") x['date'] = int(os.path.getmtime(p) * 1000) fp.close() all += x, fp = open("../modules/plugins.json", "w") fp.write(dump(all))
import os, sys, re try: from simplejson import load as json from simplejson import dumps as dump except: from json import load as json from json import dumps as dump from glob import glob VERSION = 0.1 all = [] for p in glob("../plugins/*.json"): fp = open(p, "r") x = json(fp, "utf-8") x['date'] = int(os.path.getmtime(p) * 1000) fp.close() all += x, fp = open("../modules/plugins.json", "w") fp.write(dump(all))
Use json standard module if simplejson is not present
Use json standard module if simplejson is not present
Python
mpl-2.0
marianocarrazana/anticontainer,downthemall/anticontainer,downthemall/anticontainer,marianocarrazana/anticontainer,downthemall/anticontainer,marianocarrazana/anticontainer
import os, sys, re from simplejson import load as json from simplejson import dumps as dump from glob import glob VERSION = 0.1 all = [] for p in glob("../plugins/*.json"): fp = open(p, "r") x = json(fp, "utf-8") x['date'] = int(os.path.getmtime(p) * 1000) fp.close() all += x, fp = open("../modules/plugins.json", "w") fp.write(dump(all))Use json standard module if simplejson is not present
import os, sys, re try: from simplejson import load as json from simplejson import dumps as dump except: from json import load as json from json import dumps as dump from glob import glob VERSION = 0.1 all = [] for p in glob("../plugins/*.json"): fp = open(p, "r") x = json(fp, "utf-8") x['date'] = int(os.path.getmtime(p) * 1000) fp.close() all += x, fp = open("../modules/plugins.json", "w") fp.write(dump(all))
<commit_before>import os, sys, re from simplejson import load as json from simplejson import dumps as dump from glob import glob VERSION = 0.1 all = [] for p in glob("../plugins/*.json"): fp = open(p, "r") x = json(fp, "utf-8") x['date'] = int(os.path.getmtime(p) * 1000) fp.close() all += x, fp = open("../modules/plugins.json", "w") fp.write(dump(all))<commit_msg>Use json standard module if simplejson is not present<commit_after>
import os, sys, re try: from simplejson import load as json from simplejson import dumps as dump except: from json import load as json from json import dumps as dump from glob import glob VERSION = 0.1 all = [] for p in glob("../plugins/*.json"): fp = open(p, "r") x = json(fp, "utf-8") x['date'] = int(os.path.getmtime(p) * 1000) fp.close() all += x, fp = open("../modules/plugins.json", "w") fp.write(dump(all))
import os, sys, re from simplejson import load as json from simplejson import dumps as dump from glob import glob VERSION = 0.1 all = [] for p in glob("../plugins/*.json"): fp = open(p, "r") x = json(fp, "utf-8") x['date'] = int(os.path.getmtime(p) * 1000) fp.close() all += x, fp = open("../modules/plugins.json", "w") fp.write(dump(all))Use json standard module if simplejson is not presentimport os, sys, re try: from simplejson import load as json from simplejson import dumps as dump except: from json import load as json from json import dumps as dump from glob import glob VERSION = 0.1 all = [] for p in glob("../plugins/*.json"): fp = open(p, "r") x = json(fp, "utf-8") x['date'] = int(os.path.getmtime(p) * 1000) fp.close() all += x, fp = open("../modules/plugins.json", "w") fp.write(dump(all))
<commit_before>import os, sys, re from simplejson import load as json from simplejson import dumps as dump from glob import glob VERSION = 0.1 all = [] for p in glob("../plugins/*.json"): fp = open(p, "r") x = json(fp, "utf-8") x['date'] = int(os.path.getmtime(p) * 1000) fp.close() all += x, fp = open("../modules/plugins.json", "w") fp.write(dump(all))<commit_msg>Use json standard module if simplejson is not present<commit_after>import os, sys, re try: from simplejson import load as json from simplejson import dumps as dump except: from json import load as json from json import dumps as dump from glob import glob VERSION = 0.1 all = [] for p in glob("../plugins/*.json"): fp = open(p, "r") x = json(fp, "utf-8") x['date'] = int(os.path.getmtime(p) * 1000) fp.close() all += x, fp = open("../modules/plugins.json", "w") fp.write(dump(all))
0b77033563ab85c98ca5ea9c190bcee4da5c6aef
sanic_sentry.py
sanic_sentry.py
import logging import sanic import raven import raven_aiohttp from raven.handlers.logging import SentryHandler try: from sanic.log import logger except ImportError: logger = logging.getLogger('sanic') class SanicSentry: def __init__(self, app=None): self.app = None self.handler = None self.client = None if app is not None: self.init_app(app) def init_app(self, app: sanic.Sanic): self.client = raven.Client( dsn=app.config['SENTRY_DSN'], transport=raven_aiohttp.AioHttpTransport, **app.config.get('SENTRY_PARAMS') ) self.handler = SentryHandler(client=self.client, level=app.config.get('SENTRY_LEVEL', logging.ERROR)) logger.addHandler(self.handler) self.app = app self.app.sentry = self
import logging import sanic import raven import raven_aiohttp from raven.handlers.logging import SentryHandler try: from sanic.log import logger except ImportError: logger = logging.getLogger('sanic') class SanicSentry: def __init__(self, app=None): self.app = None self.handler = None self.client = None if app is not None: self.init_app(app) def init_app(self, app: sanic.Sanic): self.client = raven.Client( dsn=app.config['SENTRY_DSN'], transport=raven_aiohttp.AioHttpTransport, **app.config.get('SENTRY_PARAMS', {}) ) self.handler = SentryHandler(client=self.client, level=app.config.get('SENTRY_LEVEL', logging.ERROR)) logger.addHandler(self.handler) self.app = app self.app.sentry = self
Add a default value for SENTRY_PARAMS
Add a default value for SENTRY_PARAMS
Python
mit
serathius/sanic-sentry
import logging import sanic import raven import raven_aiohttp from raven.handlers.logging import SentryHandler try: from sanic.log import logger except ImportError: logger = logging.getLogger('sanic') class SanicSentry: def __init__(self, app=None): self.app = None self.handler = None self.client = None if app is not None: self.init_app(app) def init_app(self, app: sanic.Sanic): self.client = raven.Client( dsn=app.config['SENTRY_DSN'], transport=raven_aiohttp.AioHttpTransport, **app.config.get('SENTRY_PARAMS') ) self.handler = SentryHandler(client=self.client, level=app.config.get('SENTRY_LEVEL', logging.ERROR)) logger.addHandler(self.handler) self.app = app self.app.sentry = self Add a default value for SENTRY_PARAMS
import logging import sanic import raven import raven_aiohttp from raven.handlers.logging import SentryHandler try: from sanic.log import logger except ImportError: logger = logging.getLogger('sanic') class SanicSentry: def __init__(self, app=None): self.app = None self.handler = None self.client = None if app is not None: self.init_app(app) def init_app(self, app: sanic.Sanic): self.client = raven.Client( dsn=app.config['SENTRY_DSN'], transport=raven_aiohttp.AioHttpTransport, **app.config.get('SENTRY_PARAMS', {}) ) self.handler = SentryHandler(client=self.client, level=app.config.get('SENTRY_LEVEL', logging.ERROR)) logger.addHandler(self.handler) self.app = app self.app.sentry = self
<commit_before>import logging import sanic import raven import raven_aiohttp from raven.handlers.logging import SentryHandler try: from sanic.log import logger except ImportError: logger = logging.getLogger('sanic') class SanicSentry: def __init__(self, app=None): self.app = None self.handler = None self.client = None if app is not None: self.init_app(app) def init_app(self, app: sanic.Sanic): self.client = raven.Client( dsn=app.config['SENTRY_DSN'], transport=raven_aiohttp.AioHttpTransport, **app.config.get('SENTRY_PARAMS') ) self.handler = SentryHandler(client=self.client, level=app.config.get('SENTRY_LEVEL', logging.ERROR)) logger.addHandler(self.handler) self.app = app self.app.sentry = self <commit_msg>Add a default value for SENTRY_PARAMS<commit_after>
import logging import sanic import raven import raven_aiohttp from raven.handlers.logging import SentryHandler try: from sanic.log import logger except ImportError: logger = logging.getLogger('sanic') class SanicSentry: def __init__(self, app=None): self.app = None self.handler = None self.client = None if app is not None: self.init_app(app) def init_app(self, app: sanic.Sanic): self.client = raven.Client( dsn=app.config['SENTRY_DSN'], transport=raven_aiohttp.AioHttpTransport, **app.config.get('SENTRY_PARAMS', {}) ) self.handler = SentryHandler(client=self.client, level=app.config.get('SENTRY_LEVEL', logging.ERROR)) logger.addHandler(self.handler) self.app = app self.app.sentry = self
import logging import sanic import raven import raven_aiohttp from raven.handlers.logging import SentryHandler try: from sanic.log import logger except ImportError: logger = logging.getLogger('sanic') class SanicSentry: def __init__(self, app=None): self.app = None self.handler = None self.client = None if app is not None: self.init_app(app) def init_app(self, app: sanic.Sanic): self.client = raven.Client( dsn=app.config['SENTRY_DSN'], transport=raven_aiohttp.AioHttpTransport, **app.config.get('SENTRY_PARAMS') ) self.handler = SentryHandler(client=self.client, level=app.config.get('SENTRY_LEVEL', logging.ERROR)) logger.addHandler(self.handler) self.app = app self.app.sentry = self Add a default value for SENTRY_PARAMSimport logging import sanic import raven import raven_aiohttp from raven.handlers.logging import SentryHandler try: from sanic.log import logger except ImportError: logger = logging.getLogger('sanic') class SanicSentry: def __init__(self, app=None): self.app = None self.handler = None self.client = None if app is not None: self.init_app(app) def init_app(self, app: sanic.Sanic): self.client = raven.Client( dsn=app.config['SENTRY_DSN'], transport=raven_aiohttp.AioHttpTransport, **app.config.get('SENTRY_PARAMS', {}) ) self.handler = SentryHandler(client=self.client, level=app.config.get('SENTRY_LEVEL', logging.ERROR)) logger.addHandler(self.handler) self.app = app self.app.sentry = self
<commit_before>import logging import sanic import raven import raven_aiohttp from raven.handlers.logging import SentryHandler try: from sanic.log import logger except ImportError: logger = logging.getLogger('sanic') class SanicSentry: def __init__(self, app=None): self.app = None self.handler = None self.client = None if app is not None: self.init_app(app) def init_app(self, app: sanic.Sanic): self.client = raven.Client( dsn=app.config['SENTRY_DSN'], transport=raven_aiohttp.AioHttpTransport, **app.config.get('SENTRY_PARAMS') ) self.handler = SentryHandler(client=self.client, level=app.config.get('SENTRY_LEVEL', logging.ERROR)) logger.addHandler(self.handler) self.app = app self.app.sentry = self <commit_msg>Add a default value for SENTRY_PARAMS<commit_after>import logging import sanic import raven import raven_aiohttp from raven.handlers.logging import SentryHandler try: from sanic.log import logger except ImportError: logger = logging.getLogger('sanic') class SanicSentry: def __init__(self, app=None): self.app = None self.handler = None self.client = None if app is not None: self.init_app(app) def init_app(self, app: sanic.Sanic): self.client = raven.Client( dsn=app.config['SENTRY_DSN'], transport=raven_aiohttp.AioHttpTransport, **app.config.get('SENTRY_PARAMS', {}) ) self.handler = SentryHandler(client=self.client, level=app.config.get('SENTRY_LEVEL', logging.ERROR)) logger.addHandler(self.handler) self.app = app self.app.sentry = self
aaa7da2b43ab08758456c972cd2bd727082c835d
build/release.py
build/release.py
#!/usr/bin/env python import os import sys import shutil import subprocess from zipfile import ZipFile if len(sys.argv) != 2: print 'Usage: release.py version-number' sys.exit(1) version = sys.argv[1] work_dir = 'minified' name = 'goo-' + version # Root directory inside zip file zip_root = name + '/' print 'Creating release', name if os.path.isdir(work_dir): shutil.rmtree(work_dir) if os.name == 'nt': command = 'cake.cmd' else: command = 'cake' subprocess.check_call([command, 'minify']) zipfile = ZipFile(name + '.zip', 'w') zipfile.write('COPYING', zip_root + 'COPYING') goo_root = work_dir + '/goo' for root, dirs, files in os.walk(goo_root): for f in files: filename = root[len(goo_root) + 1:] + '/' + f zipfile.write(root + '/' + f, zip_root + filename) zipfile.close()
#!/usr/bin/env python import os import sys import shutil import subprocess from zipfile import ZipFile def prepend(filename, to_prepend): """Prepends a string to a file """ with open(filename, 'r') as stream: content = stream.read() with open(filename, 'w') as stream: stream.write(to_prepend) stream.write(content) if len(sys.argv) != 2: print 'Usage: release.py version-number' sys.exit(1) version = sys.argv[1] work_dir = 'minified' name = 'goo-' + version # Root directory inside zip file zip_root = name + '/' print 'Creating release', name if os.path.isdir(work_dir): shutil.rmtree(work_dir) if os.name == 'nt': command = 'cake.cmd' else: command = 'cake' subprocess.check_call([command, 'minify']) zipfile = ZipFile(name + '.zip', 'w') zipfile.write('COPYING', zip_root + 'COPYING') goo_root = work_dir + '/goo' prepend(goo_root + '/goo.js', '/*\n' + ' * Goo Engine ' + version + '\n' + ' * Copyright 2013 Goo Technologies AB\n' + ' */\n' ) for root, dirs, files in os.walk(goo_root): for f in files: filename = root[len(goo_root) + 1:] + '/' + f zipfile.write(root + '/' + f, zip_root + filename) zipfile.close()
Add version number and copyright to goo.js
Add version number and copyright to goo.js This is useful to keep track of which engine version the tool uses, story #294
Python
mit
GooTechnologies/goojs,GooTechnologies/goojs,GooTechnologies/goojs
#!/usr/bin/env python import os import sys import shutil import subprocess from zipfile import ZipFile if len(sys.argv) != 2: print 'Usage: release.py version-number' sys.exit(1) version = sys.argv[1] work_dir = 'minified' name = 'goo-' + version # Root directory inside zip file zip_root = name + '/' print 'Creating release', name if os.path.isdir(work_dir): shutil.rmtree(work_dir) if os.name == 'nt': command = 'cake.cmd' else: command = 'cake' subprocess.check_call([command, 'minify']) zipfile = ZipFile(name + '.zip', 'w') zipfile.write('COPYING', zip_root + 'COPYING') goo_root = work_dir + '/goo' for root, dirs, files in os.walk(goo_root): for f in files: filename = root[len(goo_root) + 1:] + '/' + f zipfile.write(root + '/' + f, zip_root + filename) zipfile.close()Add version number and copyright to goo.js This is useful to keep track of which engine version the tool uses, story #294
#!/usr/bin/env python import os import sys import shutil import subprocess from zipfile import ZipFile def prepend(filename, to_prepend): """Prepends a string to a file """ with open(filename, 'r') as stream: content = stream.read() with open(filename, 'w') as stream: stream.write(to_prepend) stream.write(content) if len(sys.argv) != 2: print 'Usage: release.py version-number' sys.exit(1) version = sys.argv[1] work_dir = 'minified' name = 'goo-' + version # Root directory inside zip file zip_root = name + '/' print 'Creating release', name if os.path.isdir(work_dir): shutil.rmtree(work_dir) if os.name == 'nt': command = 'cake.cmd' else: command = 'cake' subprocess.check_call([command, 'minify']) zipfile = ZipFile(name + '.zip', 'w') zipfile.write('COPYING', zip_root + 'COPYING') goo_root = work_dir + '/goo' prepend(goo_root + '/goo.js', '/*\n' + ' * Goo Engine ' + version + '\n' + ' * Copyright 2013 Goo Technologies AB\n' + ' */\n' ) for root, dirs, files in os.walk(goo_root): for f in files: filename = root[len(goo_root) + 1:] + '/' + f zipfile.write(root + '/' + f, zip_root + filename) zipfile.close()
<commit_before>#!/usr/bin/env python import os import sys import shutil import subprocess from zipfile import ZipFile if len(sys.argv) != 2: print 'Usage: release.py version-number' sys.exit(1) version = sys.argv[1] work_dir = 'minified' name = 'goo-' + version # Root directory inside zip file zip_root = name + '/' print 'Creating release', name if os.path.isdir(work_dir): shutil.rmtree(work_dir) if os.name == 'nt': command = 'cake.cmd' else: command = 'cake' subprocess.check_call([command, 'minify']) zipfile = ZipFile(name + '.zip', 'w') zipfile.write('COPYING', zip_root + 'COPYING') goo_root = work_dir + '/goo' for root, dirs, files in os.walk(goo_root): for f in files: filename = root[len(goo_root) + 1:] + '/' + f zipfile.write(root + '/' + f, zip_root + filename) zipfile.close()<commit_msg>Add version number and copyright to goo.js This is useful to keep track of which engine version the tool uses, story #294<commit_after>
#!/usr/bin/env python import os import sys import shutil import subprocess from zipfile import ZipFile def prepend(filename, to_prepend): """Prepends a string to a file """ with open(filename, 'r') as stream: content = stream.read() with open(filename, 'w') as stream: stream.write(to_prepend) stream.write(content) if len(sys.argv) != 2: print 'Usage: release.py version-number' sys.exit(1) version = sys.argv[1] work_dir = 'minified' name = 'goo-' + version # Root directory inside zip file zip_root = name + '/' print 'Creating release', name if os.path.isdir(work_dir): shutil.rmtree(work_dir) if os.name == 'nt': command = 'cake.cmd' else: command = 'cake' subprocess.check_call([command, 'minify']) zipfile = ZipFile(name + '.zip', 'w') zipfile.write('COPYING', zip_root + 'COPYING') goo_root = work_dir + '/goo' prepend(goo_root + '/goo.js', '/*\n' + ' * Goo Engine ' + version + '\n' + ' * Copyright 2013 Goo Technologies AB\n' + ' */\n' ) for root, dirs, files in os.walk(goo_root): for f in files: filename = root[len(goo_root) + 1:] + '/' + f zipfile.write(root + '/' + f, zip_root + filename) zipfile.close()
#!/usr/bin/env python import os import sys import shutil import subprocess from zipfile import ZipFile if len(sys.argv) != 2: print 'Usage: release.py version-number' sys.exit(1) version = sys.argv[1] work_dir = 'minified' name = 'goo-' + version # Root directory inside zip file zip_root = name + '/' print 'Creating release', name if os.path.isdir(work_dir): shutil.rmtree(work_dir) if os.name == 'nt': command = 'cake.cmd' else: command = 'cake' subprocess.check_call([command, 'minify']) zipfile = ZipFile(name + '.zip', 'w') zipfile.write('COPYING', zip_root + 'COPYING') goo_root = work_dir + '/goo' for root, dirs, files in os.walk(goo_root): for f in files: filename = root[len(goo_root) + 1:] + '/' + f zipfile.write(root + '/' + f, zip_root + filename) zipfile.close()Add version number and copyright to goo.js This is useful to keep track of which engine version the tool uses, story #294#!/usr/bin/env python import os import sys import shutil import subprocess from zipfile import ZipFile def prepend(filename, to_prepend): """Prepends a string to a file """ with open(filename, 'r') as stream: content = stream.read() with open(filename, 'w') as stream: stream.write(to_prepend) stream.write(content) if len(sys.argv) != 2: print 'Usage: release.py version-number' sys.exit(1) version = sys.argv[1] work_dir = 'minified' name = 'goo-' + version # Root directory inside zip file zip_root = name + '/' print 'Creating release', name if os.path.isdir(work_dir): shutil.rmtree(work_dir) if os.name == 'nt': command = 'cake.cmd' else: command = 'cake' subprocess.check_call([command, 'minify']) zipfile = ZipFile(name + '.zip', 'w') zipfile.write('COPYING', zip_root + 'COPYING') goo_root = work_dir + '/goo' prepend(goo_root + '/goo.js', '/*\n' + ' * Goo Engine ' + version + '\n' + ' * Copyright 2013 Goo Technologies AB\n' + ' */\n' ) for root, dirs, files in os.walk(goo_root): for f in files: filename = root[len(goo_root) + 1:] + '/' + f zipfile.write(root + '/' + f, zip_root + filename) zipfile.close()
<commit_before>#!/usr/bin/env python import os import sys import shutil import subprocess from zipfile import ZipFile if len(sys.argv) != 2: print 'Usage: release.py version-number' sys.exit(1) version = sys.argv[1] work_dir = 'minified' name = 'goo-' + version # Root directory inside zip file zip_root = name + '/' print 'Creating release', name if os.path.isdir(work_dir): shutil.rmtree(work_dir) if os.name == 'nt': command = 'cake.cmd' else: command = 'cake' subprocess.check_call([command, 'minify']) zipfile = ZipFile(name + '.zip', 'w') zipfile.write('COPYING', zip_root + 'COPYING') goo_root = work_dir + '/goo' for root, dirs, files in os.walk(goo_root): for f in files: filename = root[len(goo_root) + 1:] + '/' + f zipfile.write(root + '/' + f, zip_root + filename) zipfile.close()<commit_msg>Add version number and copyright to goo.js This is useful to keep track of which engine version the tool uses, story #294<commit_after>#!/usr/bin/env python import os import sys import shutil import subprocess from zipfile import ZipFile def prepend(filename, to_prepend): """Prepends a string to a file """ with open(filename, 'r') as stream: content = stream.read() with open(filename, 'w') as stream: stream.write(to_prepend) stream.write(content) if len(sys.argv) != 2: print 'Usage: release.py version-number' sys.exit(1) version = sys.argv[1] work_dir = 'minified' name = 'goo-' + version # Root directory inside zip file zip_root = name + '/' print 'Creating release', name if os.path.isdir(work_dir): shutil.rmtree(work_dir) if os.name == 'nt': command = 'cake.cmd' else: command = 'cake' subprocess.check_call([command, 'minify']) zipfile = ZipFile(name + '.zip', 'w') zipfile.write('COPYING', zip_root + 'COPYING') goo_root = work_dir + '/goo' prepend(goo_root + '/goo.js', '/*\n' + ' * Goo Engine ' + version + '\n' + ' * Copyright 2013 Goo Technologies AB\n' + ' */\n' ) for root, dirs, files in os.walk(goo_root): for f in files: filename = root[len(goo_root) + 1:] + '/' + f zipfile.write(root + '/' + f, zip_root + filename) zipfile.close()
1c3f89110ede8998b63831c181c44e92709481b6
demo/widgy.py
demo/widgy.py
from __future__ import absolute_import from widgy.site import WidgySite class DemoWidgySite(WidgySite): def valid_parent_of(self, parent, child_class, obj=None): if isinstance(parent, I18NLayout): return True else: return super(DemoWidgySite, self).valid_parent_of(parent, child_class, obj) widgy_site = DemoWidgySite() from widgy.contrib.widgy_i18n.models import I18NLayout
from __future__ import absolute_import from widgy.site import ReviewedWidgySite class DemoWidgySite(ReviewedWidgySite): def valid_parent_of(self, parent, child_class, obj=None): if isinstance(parent, I18NLayout): return True else: return super(DemoWidgySite, self).valid_parent_of(parent, child_class, obj) widgy_site = DemoWidgySite() from widgy.contrib.widgy_i18n.models import I18NLayout
Enable the review queue on the demo site
Enable the review queue on the demo site
Python
apache-2.0
j00bar/django-widgy,j00bar/django-widgy,j00bar/django-widgy
from __future__ import absolute_import from widgy.site import WidgySite class DemoWidgySite(WidgySite): def valid_parent_of(self, parent, child_class, obj=None): if isinstance(parent, I18NLayout): return True else: return super(DemoWidgySite, self).valid_parent_of(parent, child_class, obj) widgy_site = DemoWidgySite() from widgy.contrib.widgy_i18n.models import I18NLayout Enable the review queue on the demo site
from __future__ import absolute_import from widgy.site import ReviewedWidgySite class DemoWidgySite(ReviewedWidgySite): def valid_parent_of(self, parent, child_class, obj=None): if isinstance(parent, I18NLayout): return True else: return super(DemoWidgySite, self).valid_parent_of(parent, child_class, obj) widgy_site = DemoWidgySite() from widgy.contrib.widgy_i18n.models import I18NLayout
<commit_before>from __future__ import absolute_import from widgy.site import WidgySite class DemoWidgySite(WidgySite): def valid_parent_of(self, parent, child_class, obj=None): if isinstance(parent, I18NLayout): return True else: return super(DemoWidgySite, self).valid_parent_of(parent, child_class, obj) widgy_site = DemoWidgySite() from widgy.contrib.widgy_i18n.models import I18NLayout <commit_msg>Enable the review queue on the demo site<commit_after>
from __future__ import absolute_import from widgy.site import ReviewedWidgySite class DemoWidgySite(ReviewedWidgySite): def valid_parent_of(self, parent, child_class, obj=None): if isinstance(parent, I18NLayout): return True else: return super(DemoWidgySite, self).valid_parent_of(parent, child_class, obj) widgy_site = DemoWidgySite() from widgy.contrib.widgy_i18n.models import I18NLayout
from __future__ import absolute_import from widgy.site import WidgySite class DemoWidgySite(WidgySite): def valid_parent_of(self, parent, child_class, obj=None): if isinstance(parent, I18NLayout): return True else: return super(DemoWidgySite, self).valid_parent_of(parent, child_class, obj) widgy_site = DemoWidgySite() from widgy.contrib.widgy_i18n.models import I18NLayout Enable the review queue on the demo sitefrom __future__ import absolute_import from widgy.site import ReviewedWidgySite class DemoWidgySite(ReviewedWidgySite): def valid_parent_of(self, parent, child_class, obj=None): if isinstance(parent, I18NLayout): return True else: return super(DemoWidgySite, self).valid_parent_of(parent, child_class, obj) widgy_site = DemoWidgySite() from widgy.contrib.widgy_i18n.models import I18NLayout
<commit_before>from __future__ import absolute_import from widgy.site import WidgySite class DemoWidgySite(WidgySite): def valid_parent_of(self, parent, child_class, obj=None): if isinstance(parent, I18NLayout): return True else: return super(DemoWidgySite, self).valid_parent_of(parent, child_class, obj) widgy_site = DemoWidgySite() from widgy.contrib.widgy_i18n.models import I18NLayout <commit_msg>Enable the review queue on the demo site<commit_after>from __future__ import absolute_import from widgy.site import ReviewedWidgySite class DemoWidgySite(ReviewedWidgySite): def valid_parent_of(self, parent, child_class, obj=None): if isinstance(parent, I18NLayout): return True else: return super(DemoWidgySite, self).valid_parent_of(parent, child_class, obj) widgy_site = DemoWidgySite() from widgy.contrib.widgy_i18n.models import I18NLayout
ea09470ebdd69af2fa1d7d07d7b04fe3ff857987
raffle.py
raffle.py
""" St. George Game raffle.py Sage Berg Created: 9 Dec 2014 """ from random import randint class Raffle(object): """ Raffle contains a list of action objects, one of which will be chosen and shown to the player. """ def __init__(self): self.options = dict() # Maps options to weights def add(self, option, weight=1): """ """ if option in self.options: self.options[option] += weight else: self.options[option] = weight def get(self): """ chooses one action from the bag and returns it. """ total_weights = 0 for weight in self.options.values(): total_weights += weight roll = randint(0, total_weights) for option, weight in self.options.items(): if roll <= weight: return option else: roll -= weight def merge(self, other): """ Merge the contents of another Raffle with this Raffle. """ for option, weight in other.options.items(): self.add(option, weight)
""" St. George Game raffle.py Sage Berg Created: 9 Dec 2014 """ from random import randint class Raffle(object): """ Raffle contains a list of action objects, one of which will be chosen and shown to the player. """ def __init__(self): self.options = dict() # Maps options to weights def add(self, option, weight=1): """ """ if option in self.options: self.options[option] += weight else: self.options[option] = weight def get(self): """ chooses one action from the bag and returns it. """ total_weights = 0 for weight in self.options.values(): total_weights += weight roll = randint(0, total_weights) for option, weight in self.options.items(): if roll <= weight: return option else: roll -= weight def merge(self, other): """ Merge the contents of another Raffle with this Raffle. """ for option, weight in other.options.items(): self.add(option, weight) def __len__(self): """ Return the combined weight of all options in the Raffle. """ total = 0 for _, weight in self.options.items(): total += weight return total
Add length method to Raffle
Add length method to Raffle
Python
apache-2.0
SageBerg/St.GeorgeGame,SageBerg/St.GeorgeGame,SageBerg/St.GeorgeGame,SageBerg/St.GeorgeGame
""" St. George Game raffle.py Sage Berg Created: 9 Dec 2014 """ from random import randint class Raffle(object): """ Raffle contains a list of action objects, one of which will be chosen and shown to the player. """ def __init__(self): self.options = dict() # Maps options to weights def add(self, option, weight=1): """ """ if option in self.options: self.options[option] += weight else: self.options[option] = weight def get(self): """ chooses one action from the bag and returns it. """ total_weights = 0 for weight in self.options.values(): total_weights += weight roll = randint(0, total_weights) for option, weight in self.options.items(): if roll <= weight: return option else: roll -= weight def merge(self, other): """ Merge the contents of another Raffle with this Raffle. """ for option, weight in other.options.items(): self.add(option, weight) Add length method to Raffle
""" St. George Game raffle.py Sage Berg Created: 9 Dec 2014 """ from random import randint class Raffle(object): """ Raffle contains a list of action objects, one of which will be chosen and shown to the player. """ def __init__(self): self.options = dict() # Maps options to weights def add(self, option, weight=1): """ """ if option in self.options: self.options[option] += weight else: self.options[option] = weight def get(self): """ chooses one action from the bag and returns it. """ total_weights = 0 for weight in self.options.values(): total_weights += weight roll = randint(0, total_weights) for option, weight in self.options.items(): if roll <= weight: return option else: roll -= weight def merge(self, other): """ Merge the contents of another Raffle with this Raffle. """ for option, weight in other.options.items(): self.add(option, weight) def __len__(self): """ Return the combined weight of all options in the Raffle. """ total = 0 for _, weight in self.options.items(): total += weight return total
<commit_before>""" St. George Game raffle.py Sage Berg Created: 9 Dec 2014 """ from random import randint class Raffle(object): """ Raffle contains a list of action objects, one of which will be chosen and shown to the player. """ def __init__(self): self.options = dict() # Maps options to weights def add(self, option, weight=1): """ """ if option in self.options: self.options[option] += weight else: self.options[option] = weight def get(self): """ chooses one action from the bag and returns it. """ total_weights = 0 for weight in self.options.values(): total_weights += weight roll = randint(0, total_weights) for option, weight in self.options.items(): if roll <= weight: return option else: roll -= weight def merge(self, other): """ Merge the contents of another Raffle with this Raffle. """ for option, weight in other.options.items(): self.add(option, weight) <commit_msg>Add length method to Raffle<commit_after>
""" St. George Game raffle.py Sage Berg Created: 9 Dec 2014 """ from random import randint class Raffle(object): """ Raffle contains a list of action objects, one of which will be chosen and shown to the player. """ def __init__(self): self.options = dict() # Maps options to weights def add(self, option, weight=1): """ """ if option in self.options: self.options[option] += weight else: self.options[option] = weight def get(self): """ chooses one action from the bag and returns it. """ total_weights = 0 for weight in self.options.values(): total_weights += weight roll = randint(0, total_weights) for option, weight in self.options.items(): if roll <= weight: return option else: roll -= weight def merge(self, other): """ Merge the contents of another Raffle with this Raffle. """ for option, weight in other.options.items(): self.add(option, weight) def __len__(self): """ Return the combined weight of all options in the Raffle. """ total = 0 for _, weight in self.options.items(): total += weight return total
""" St. George Game raffle.py Sage Berg Created: 9 Dec 2014 """ from random import randint class Raffle(object): """ Raffle contains a list of action objects, one of which will be chosen and shown to the player. """ def __init__(self): self.options = dict() # Maps options to weights def add(self, option, weight=1): """ """ if option in self.options: self.options[option] += weight else: self.options[option] = weight def get(self): """ chooses one action from the bag and returns it. """ total_weights = 0 for weight in self.options.values(): total_weights += weight roll = randint(0, total_weights) for option, weight in self.options.items(): if roll <= weight: return option else: roll -= weight def merge(self, other): """ Merge the contents of another Raffle with this Raffle. """ for option, weight in other.options.items(): self.add(option, weight) Add length method to Raffle""" St. George Game raffle.py Sage Berg Created: 9 Dec 2014 """ from random import randint class Raffle(object): """ Raffle contains a list of action objects, one of which will be chosen and shown to the player. """ def __init__(self): self.options = dict() # Maps options to weights def add(self, option, weight=1): """ """ if option in self.options: self.options[option] += weight else: self.options[option] = weight def get(self): """ chooses one action from the bag and returns it. """ total_weights = 0 for weight in self.options.values(): total_weights += weight roll = randint(0, total_weights) for option, weight in self.options.items(): if roll <= weight: return option else: roll -= weight def merge(self, other): """ Merge the contents of another Raffle with this Raffle. """ for option, weight in other.options.items(): self.add(option, weight) def __len__(self): """ Return the combined weight of all options in the Raffle. """ total = 0 for _, weight in self.options.items(): total += weight return total
<commit_before>""" St. George Game raffle.py Sage Berg Created: 9 Dec 2014 """ from random import randint class Raffle(object): """ Raffle contains a list of action objects, one of which will be chosen and shown to the player. """ def __init__(self): self.options = dict() # Maps options to weights def add(self, option, weight=1): """ """ if option in self.options: self.options[option] += weight else: self.options[option] = weight def get(self): """ chooses one action from the bag and returns it. """ total_weights = 0 for weight in self.options.values(): total_weights += weight roll = randint(0, total_weights) for option, weight in self.options.items(): if roll <= weight: return option else: roll -= weight def merge(self, other): """ Merge the contents of another Raffle with this Raffle. """ for option, weight in other.options.items(): self.add(option, weight) <commit_msg>Add length method to Raffle<commit_after>""" St. George Game raffle.py Sage Berg Created: 9 Dec 2014 """ from random import randint class Raffle(object): """ Raffle contains a list of action objects, one of which will be chosen and shown to the player. """ def __init__(self): self.options = dict() # Maps options to weights def add(self, option, weight=1): """ """ if option in self.options: self.options[option] += weight else: self.options[option] = weight def get(self): """ chooses one action from the bag and returns it. """ total_weights = 0 for weight in self.options.values(): total_weights += weight roll = randint(0, total_weights) for option, weight in self.options.items(): if roll <= weight: return option else: roll -= weight def merge(self, other): """ Merge the contents of another Raffle with this Raffle. """ for option, weight in other.options.items(): self.add(option, weight) def __len__(self): """ Return the combined weight of all options in the Raffle. """ total = 0 for _, weight in self.options.items(): total += weight return total
616e9727397853e8d8f8de5b2c040c99c91e4a50
gen_settings.py
gen_settings.py
import os settings = os.path.join(os.path.dirname(__file__),'lib','mapnik_settings.js') # this goes into a mapnik_settings.js file beside the C++ _mapnik.node settings_template = """ module.exports.paths = { 'fonts': %s, 'input_plugins': %s }; """ def write_mapnik_settings(fonts='undefined',input_plugins='undefined'): global settings_template if '__dirname' in fonts or '__dirname' in input_plugins: settings_template = "var path = require('path');\n" + settings_template open(settings,'w').write(settings_template % (fonts,input_plugins)) if __name__ == '__main__': settings_dict = {} # settings for fonts and input plugins settings_dict['input_plugins'] = '\'%s\'' % os.popen("mapnik-config --input-plugins").readline().strip() settings_dict['fonts'] = '\'%s\'' % os.popen("mapnik-config --fonts").readline().strip() write_mapnik_settings(**settings_dict)
import os settings = os.path.join(os.path.dirname(__file__),'lib','mapnik_settings.js') # this goes into a mapnik_settings.js file beside the C++ _mapnik.node settings_template = """ module.exports.paths = { 'fonts': %s, 'input_plugins': %s }; """ def write_mapnik_settings(fonts='undefined',input_plugins='undefined'): global settings_template if '__dirname' in fonts or '__dirname' in input_plugins: settings_template = "var path = require('path');\n" + settings_template open(settings,'w').write(settings_template % (fonts,input_plugins)) if __name__ == '__main__': settings_dict = {} # settings for fonts and input plugins if os.environ.has_key('MAPNIK_INPUT_PLUGINS_DIRECTORY'): settings_dict['input_plugins'] = os.environ['MAPNIK_INPUT_PLUGINS_DIRECTORY'] else: settings_dict['input_plugins'] = '\'%s\'' % os.popen("mapnik-config --input-plugins").readline().strip() if os.environ.has_key('MAPNIK_FONT_DIRECTORY'): settings_dict['fonts'] = os.environ['MAPNIK_FONT_DIRECTORY'] else: settings_dict['fonts'] = '\'%s\'' % os.popen("mapnik-config --fonts").readline().strip() write_mapnik_settings(**settings_dict)
Revert "stop reading fonts/input plugins from environ as we now have a working mapnik-config.bat on windows"
Revert "stop reading fonts/input plugins from environ as we now have a working mapnik-config.bat on windows" This reverts commit d87c71142ba7bcc0d99d84886f3534dea7617b0c.
Python
bsd-3-clause
mapnik/node-mapnik,langateam/node-mapnik,mojodna/node-mapnik,CartoDB/node-mapnik,CartoDB/node-mapnik,MaxSem/node-mapnik,gravitystorm/node-mapnik,tomhughes/node-mapnik,mojodna/node-mapnik,tomhughes/node-mapnik,CartoDB/node-mapnik,Uli1/node-mapnik,mojodna/node-mapnik,stefanklug/node-mapnik,CartoDB/node-mapnik,langateam/node-mapnik,gravitystorm/node-mapnik,Uli1/node-mapnik,Uli1/node-mapnik,MaxSem/node-mapnik,CartoDB/node-mapnik,mapnik/node-mapnik,tomhughes/node-mapnik,stefanklug/node-mapnik,mapnik/node-mapnik,mojodna/node-mapnik,langateam/node-mapnik,MaxSem/node-mapnik,tomhughes/node-mapnik,gravitystorm/node-mapnik,langateam/node-mapnik,gravitystorm/node-mapnik,Uli1/node-mapnik,stefanklug/node-mapnik,tomhughes/node-mapnik,mapnik/node-mapnik,stefanklug/node-mapnik,MaxSem/node-mapnik,mapnik/node-mapnik,langateam/node-mapnik
import os settings = os.path.join(os.path.dirname(__file__),'lib','mapnik_settings.js') # this goes into a mapnik_settings.js file beside the C++ _mapnik.node settings_template = """ module.exports.paths = { 'fonts': %s, 'input_plugins': %s }; """ def write_mapnik_settings(fonts='undefined',input_plugins='undefined'): global settings_template if '__dirname' in fonts or '__dirname' in input_plugins: settings_template = "var path = require('path');\n" + settings_template open(settings,'w').write(settings_template % (fonts,input_plugins)) if __name__ == '__main__': settings_dict = {} # settings for fonts and input plugins settings_dict['input_plugins'] = '\'%s\'' % os.popen("mapnik-config --input-plugins").readline().strip() settings_dict['fonts'] = '\'%s\'' % os.popen("mapnik-config --fonts").readline().strip() write_mapnik_settings(**settings_dict)Revert "stop reading fonts/input plugins from environ as we now have a working mapnik-config.bat on windows" This reverts commit d87c71142ba7bcc0d99d84886f3534dea7617b0c.
import os settings = os.path.join(os.path.dirname(__file__),'lib','mapnik_settings.js') # this goes into a mapnik_settings.js file beside the C++ _mapnik.node settings_template = """ module.exports.paths = { 'fonts': %s, 'input_plugins': %s }; """ def write_mapnik_settings(fonts='undefined',input_plugins='undefined'): global settings_template if '__dirname' in fonts or '__dirname' in input_plugins: settings_template = "var path = require('path');\n" + settings_template open(settings,'w').write(settings_template % (fonts,input_plugins)) if __name__ == '__main__': settings_dict = {} # settings for fonts and input plugins if os.environ.has_key('MAPNIK_INPUT_PLUGINS_DIRECTORY'): settings_dict['input_plugins'] = os.environ['MAPNIK_INPUT_PLUGINS_DIRECTORY'] else: settings_dict['input_plugins'] = '\'%s\'' % os.popen("mapnik-config --input-plugins").readline().strip() if os.environ.has_key('MAPNIK_FONT_DIRECTORY'): settings_dict['fonts'] = os.environ['MAPNIK_FONT_DIRECTORY'] else: settings_dict['fonts'] = '\'%s\'' % os.popen("mapnik-config --fonts").readline().strip() write_mapnik_settings(**settings_dict)
<commit_before>import os settings = os.path.join(os.path.dirname(__file__),'lib','mapnik_settings.js') # this goes into a mapnik_settings.js file beside the C++ _mapnik.node settings_template = """ module.exports.paths = { 'fonts': %s, 'input_plugins': %s }; """ def write_mapnik_settings(fonts='undefined',input_plugins='undefined'): global settings_template if '__dirname' in fonts or '__dirname' in input_plugins: settings_template = "var path = require('path');\n" + settings_template open(settings,'w').write(settings_template % (fonts,input_plugins)) if __name__ == '__main__': settings_dict = {} # settings for fonts and input plugins settings_dict['input_plugins'] = '\'%s\'' % os.popen("mapnik-config --input-plugins").readline().strip() settings_dict['fonts'] = '\'%s\'' % os.popen("mapnik-config --fonts").readline().strip() write_mapnik_settings(**settings_dict)<commit_msg>Revert "stop reading fonts/input plugins from environ as we now have a working mapnik-config.bat on windows" This reverts commit d87c71142ba7bcc0d99d84886f3534dea7617b0c.<commit_after>
import os settings = os.path.join(os.path.dirname(__file__),'lib','mapnik_settings.js') # this goes into a mapnik_settings.js file beside the C++ _mapnik.node settings_template = """ module.exports.paths = { 'fonts': %s, 'input_plugins': %s }; """ def write_mapnik_settings(fonts='undefined',input_plugins='undefined'): global settings_template if '__dirname' in fonts or '__dirname' in input_plugins: settings_template = "var path = require('path');\n" + settings_template open(settings,'w').write(settings_template % (fonts,input_plugins)) if __name__ == '__main__': settings_dict = {} # settings for fonts and input plugins if os.environ.has_key('MAPNIK_INPUT_PLUGINS_DIRECTORY'): settings_dict['input_plugins'] = os.environ['MAPNIK_INPUT_PLUGINS_DIRECTORY'] else: settings_dict['input_plugins'] = '\'%s\'' % os.popen("mapnik-config --input-plugins").readline().strip() if os.environ.has_key('MAPNIK_FONT_DIRECTORY'): settings_dict['fonts'] = os.environ['MAPNIK_FONT_DIRECTORY'] else: settings_dict['fonts'] = '\'%s\'' % os.popen("mapnik-config --fonts").readline().strip() write_mapnik_settings(**settings_dict)
import os settings = os.path.join(os.path.dirname(__file__),'lib','mapnik_settings.js') # this goes into a mapnik_settings.js file beside the C++ _mapnik.node settings_template = """ module.exports.paths = { 'fonts': %s, 'input_plugins': %s }; """ def write_mapnik_settings(fonts='undefined',input_plugins='undefined'): global settings_template if '__dirname' in fonts or '__dirname' in input_plugins: settings_template = "var path = require('path');\n" + settings_template open(settings,'w').write(settings_template % (fonts,input_plugins)) if __name__ == '__main__': settings_dict = {} # settings for fonts and input plugins settings_dict['input_plugins'] = '\'%s\'' % os.popen("mapnik-config --input-plugins").readline().strip() settings_dict['fonts'] = '\'%s\'' % os.popen("mapnik-config --fonts").readline().strip() write_mapnik_settings(**settings_dict)Revert "stop reading fonts/input plugins from environ as we now have a working mapnik-config.bat on windows" This reverts commit d87c71142ba7bcc0d99d84886f3534dea7617b0c.import os settings = os.path.join(os.path.dirname(__file__),'lib','mapnik_settings.js') # this goes into a mapnik_settings.js file beside the C++ _mapnik.node settings_template = """ module.exports.paths = { 'fonts': %s, 'input_plugins': %s }; """ def write_mapnik_settings(fonts='undefined',input_plugins='undefined'): global settings_template if '__dirname' in fonts or '__dirname' in input_plugins: settings_template = "var path = require('path');\n" + settings_template open(settings,'w').write(settings_template % (fonts,input_plugins)) if __name__ == '__main__': settings_dict = {} # settings for fonts and input plugins if os.environ.has_key('MAPNIK_INPUT_PLUGINS_DIRECTORY'): settings_dict['input_plugins'] = os.environ['MAPNIK_INPUT_PLUGINS_DIRECTORY'] else: settings_dict['input_plugins'] = '\'%s\'' % os.popen("mapnik-config --input-plugins").readline().strip() if os.environ.has_key('MAPNIK_FONT_DIRECTORY'): settings_dict['fonts'] = os.environ['MAPNIK_FONT_DIRECTORY'] else: settings_dict['fonts'] = '\'%s\'' % os.popen("mapnik-config --fonts").readline().strip() write_mapnik_settings(**settings_dict)
<commit_before>import os settings = os.path.join(os.path.dirname(__file__),'lib','mapnik_settings.js') # this goes into a mapnik_settings.js file beside the C++ _mapnik.node settings_template = """ module.exports.paths = { 'fonts': %s, 'input_plugins': %s }; """ def write_mapnik_settings(fonts='undefined',input_plugins='undefined'): global settings_template if '__dirname' in fonts or '__dirname' in input_plugins: settings_template = "var path = require('path');\n" + settings_template open(settings,'w').write(settings_template % (fonts,input_plugins)) if __name__ == '__main__': settings_dict = {} # settings for fonts and input plugins settings_dict['input_plugins'] = '\'%s\'' % os.popen("mapnik-config --input-plugins").readline().strip() settings_dict['fonts'] = '\'%s\'' % os.popen("mapnik-config --fonts").readline().strip() write_mapnik_settings(**settings_dict)<commit_msg>Revert "stop reading fonts/input plugins from environ as we now have a working mapnik-config.bat on windows" This reverts commit d87c71142ba7bcc0d99d84886f3534dea7617b0c.<commit_after>import os settings = os.path.join(os.path.dirname(__file__),'lib','mapnik_settings.js') # this goes into a mapnik_settings.js file beside the C++ _mapnik.node settings_template = """ module.exports.paths = { 'fonts': %s, 'input_plugins': %s }; """ def write_mapnik_settings(fonts='undefined',input_plugins='undefined'): global settings_template if '__dirname' in fonts or '__dirname' in input_plugins: settings_template = "var path = require('path');\n" + settings_template open(settings,'w').write(settings_template % (fonts,input_plugins)) if __name__ == '__main__': settings_dict = {} # settings for fonts and input plugins if os.environ.has_key('MAPNIK_INPUT_PLUGINS_DIRECTORY'): settings_dict['input_plugins'] = os.environ['MAPNIK_INPUT_PLUGINS_DIRECTORY'] else: settings_dict['input_plugins'] = '\'%s\'' % os.popen("mapnik-config --input-plugins").readline().strip() if os.environ.has_key('MAPNIK_FONT_DIRECTORY'): settings_dict['fonts'] = os.environ['MAPNIK_FONT_DIRECTORY'] else: settings_dict['fonts'] = '\'%s\'' % os.popen("mapnik-config --fonts").readline().strip() write_mapnik_settings(**settings_dict)
bd3473a8514e6d323dd03174ce65ecf278fa3772
groups/admin.py
groups/admin.py
from django.contrib import admin from .models import Discussion, Group class GroupAdmin(admin.ModelAdmin): filter_horizontal = ('moderators', 'watchers', 'members_if_private') class Meta: model = Group class DiscussionAdmin(admin.ModelAdmin): filter_horizontal = ('subscribers', 'ignorers') class Meta: model = Discussion admin.site.register(Group, GroupAdmin) admin.site.register(Discussion, DiscussionAdmin)
from django.contrib import admin from .models import Discussion, Group @admin.register(Group) class GroupAdmin(admin.ModelAdmin): filter_horizontal = ('moderators', 'watchers', 'members_if_private') class Meta: model = Group @admin.register(Discussion) class DiscussionAdmin(admin.ModelAdmin): filter_horizontal = ('subscribers', 'ignorers') class Meta: model = Discussion
Use a decorator for slickness.
Use a decorator for slickness.
Python
bsd-2-clause
incuna/incuna-groups,incuna/incuna-groups
from django.contrib import admin from .models import Discussion, Group class GroupAdmin(admin.ModelAdmin): filter_horizontal = ('moderators', 'watchers', 'members_if_private') class Meta: model = Group class DiscussionAdmin(admin.ModelAdmin): filter_horizontal = ('subscribers', 'ignorers') class Meta: model = Discussion admin.site.register(Group, GroupAdmin) admin.site.register(Discussion, DiscussionAdmin) Use a decorator for slickness.
from django.contrib import admin from .models import Discussion, Group @admin.register(Group) class GroupAdmin(admin.ModelAdmin): filter_horizontal = ('moderators', 'watchers', 'members_if_private') class Meta: model = Group @admin.register(Discussion) class DiscussionAdmin(admin.ModelAdmin): filter_horizontal = ('subscribers', 'ignorers') class Meta: model = Discussion
<commit_before>from django.contrib import admin from .models import Discussion, Group class GroupAdmin(admin.ModelAdmin): filter_horizontal = ('moderators', 'watchers', 'members_if_private') class Meta: model = Group class DiscussionAdmin(admin.ModelAdmin): filter_horizontal = ('subscribers', 'ignorers') class Meta: model = Discussion admin.site.register(Group, GroupAdmin) admin.site.register(Discussion, DiscussionAdmin) <commit_msg>Use a decorator for slickness.<commit_after>
from django.contrib import admin from .models import Discussion, Group @admin.register(Group) class GroupAdmin(admin.ModelAdmin): filter_horizontal = ('moderators', 'watchers', 'members_if_private') class Meta: model = Group @admin.register(Discussion) class DiscussionAdmin(admin.ModelAdmin): filter_horizontal = ('subscribers', 'ignorers') class Meta: model = Discussion
from django.contrib import admin from .models import Discussion, Group class GroupAdmin(admin.ModelAdmin): filter_horizontal = ('moderators', 'watchers', 'members_if_private') class Meta: model = Group class DiscussionAdmin(admin.ModelAdmin): filter_horizontal = ('subscribers', 'ignorers') class Meta: model = Discussion admin.site.register(Group, GroupAdmin) admin.site.register(Discussion, DiscussionAdmin) Use a decorator for slickness.from django.contrib import admin from .models import Discussion, Group @admin.register(Group) class GroupAdmin(admin.ModelAdmin): filter_horizontal = ('moderators', 'watchers', 'members_if_private') class Meta: model = Group @admin.register(Discussion) class DiscussionAdmin(admin.ModelAdmin): filter_horizontal = ('subscribers', 'ignorers') class Meta: model = Discussion
<commit_before>from django.contrib import admin from .models import Discussion, Group class GroupAdmin(admin.ModelAdmin): filter_horizontal = ('moderators', 'watchers', 'members_if_private') class Meta: model = Group class DiscussionAdmin(admin.ModelAdmin): filter_horizontal = ('subscribers', 'ignorers') class Meta: model = Discussion admin.site.register(Group, GroupAdmin) admin.site.register(Discussion, DiscussionAdmin) <commit_msg>Use a decorator for slickness.<commit_after>from django.contrib import admin from .models import Discussion, Group @admin.register(Group) class GroupAdmin(admin.ModelAdmin): filter_horizontal = ('moderators', 'watchers', 'members_if_private') class Meta: model = Group @admin.register(Discussion) class DiscussionAdmin(admin.ModelAdmin): filter_horizontal = ('subscribers', 'ignorers') class Meta: model = Discussion
b9cf2145097f8d1c702183a09bf2d54f669e2218
skimage/filter/__init__.py
skimage/filter/__init__.py
from .lpi_filter import inverse, wiener, LPIFilter2D from .ctmf import median_filter from ._canny import canny from .edges import (sobel, hsobel, vsobel, scharr, hscharr, vscharr, prewitt, hprewitt, vprewitt, roberts , roberts_positive_diagonal, roberts_negative_diagonal) from ._denoise import denoise_tv_chambolle, tv_denoise from ._denoise_cy import denoise_bilateral, denoise_tv_bregman from ._rank_order import rank_order from ._gabor import gabor_kernel, gabor_filter from .thresholding import threshold_otsu, threshold_adaptive __all__ = ['inverse', 'wiener', 'LPIFilter2D', 'median_filter', 'canny', 'sobel', 'hsobel', 'vsobel', 'scharr', 'hscharr', 'vscharr', 'prewitt', 'hprewitt', 'vprewitt', 'roberts', 'roberts_positive_diagonal', 'roberts_negative_diagonal', 'denoise_tv_chambolle', 'tv_denoise', 'denoise_bilateral', 'denoise_tv_bregman', 'rank_order', 'gabor_kernel', 'gabor_filter', 'threshold_otsu', 'threshold_adaptive']
from .lpi_filter import inverse, wiener, LPIFilter2D from .ctmf import median_filter from ._canny import canny from .edges import (sobel, hsobel, vsobel, scharr, hscharr, vscharr, prewitt, hprewitt, vprewitt, roberts , roberts_positive_diagonal, roberts_negative_diagonal) from ._denoise import denoise_tv_chambolle, tv_denoise from ._denoise_cy import denoise_bilateral, denoise_tv_bregman from ._rank_order import rank_order from ._gabor import gabor_kernel, gabor_filter from .thresholding import threshold_otsu, threshold_adaptive from . import rank __all__ = ['inverse', 'wiener', 'LPIFilter2D', 'median_filter', 'canny', 'sobel', 'hsobel', 'vsobel', 'scharr', 'hscharr', 'vscharr', 'prewitt', 'hprewitt', 'vprewitt', 'roberts', 'roberts_positive_diagonal', 'roberts_negative_diagonal', 'denoise_tv_chambolle', 'tv_denoise', 'denoise_bilateral', 'denoise_tv_bregman', 'rank_order', 'gabor_kernel', 'gabor_filter', 'threshold_otsu', 'threshold_adaptive', 'rank']
Add filter.rank to __all__ of filter package
Add filter.rank to __all__ of filter package
Python
bsd-3-clause
michaelpacer/scikit-image,oew1v07/scikit-image,vighneshbirodkar/scikit-image,michaelpacer/scikit-image,chriscrosscutler/scikit-image,juliusbierk/scikit-image,chintak/scikit-image,GaZ3ll3/scikit-image,warmspringwinds/scikit-image,ajaybhat/scikit-image,robintw/scikit-image,keflavich/scikit-image,chintak/scikit-image,jwiggins/scikit-image,rjeli/scikit-image,Britefury/scikit-image,bennlich/scikit-image,ofgulban/scikit-image,dpshelio/scikit-image,dpshelio/scikit-image,almarklein/scikit-image,keflavich/scikit-image,Midafi/scikit-image,pratapvardhan/scikit-image,emon10005/scikit-image,juliusbierk/scikit-image,chintak/scikit-image,pratapvardhan/scikit-image,bsipocz/scikit-image,jwiggins/scikit-image,almarklein/scikit-image,youprofit/scikit-image,SamHames/scikit-image,GaZ3ll3/scikit-image,Midafi/scikit-image,oew1v07/scikit-image,bennlich/scikit-image,SamHames/scikit-image,blink1073/scikit-image,warmspringwinds/scikit-image,Hiyorimi/scikit-image,chintak/scikit-image,newville/scikit-image,youprofit/scikit-image,Hiyorimi/scikit-image,SamHames/scikit-image,blink1073/scikit-image,rjeli/scikit-image,newville/scikit-image,paalge/scikit-image,paalge/scikit-image,almarklein/scikit-image,bsipocz/scikit-image,paalge/scikit-image,emon10005/scikit-image,ofgulban/scikit-image,michaelaye/scikit-image,ajaybhat/scikit-image,WarrenWeckesser/scikits-image,ClinicalGraphics/scikit-image,michaelaye/scikit-image,Britefury/scikit-image,vighneshbirodkar/scikit-image,SamHames/scikit-image,robintw/scikit-image,ClinicalGraphics/scikit-image,vighneshbirodkar/scikit-image,rjeli/scikit-image,almarklein/scikit-image,ofgulban/scikit-image,WarrenWeckesser/scikits-image,chriscrosscutler/scikit-image
from .lpi_filter import inverse, wiener, LPIFilter2D from .ctmf import median_filter from ._canny import canny from .edges import (sobel, hsobel, vsobel, scharr, hscharr, vscharr, prewitt, hprewitt, vprewitt, roberts , roberts_positive_diagonal, roberts_negative_diagonal) from ._denoise import denoise_tv_chambolle, tv_denoise from ._denoise_cy import denoise_bilateral, denoise_tv_bregman from ._rank_order import rank_order from ._gabor import gabor_kernel, gabor_filter from .thresholding import threshold_otsu, threshold_adaptive __all__ = ['inverse', 'wiener', 'LPIFilter2D', 'median_filter', 'canny', 'sobel', 'hsobel', 'vsobel', 'scharr', 'hscharr', 'vscharr', 'prewitt', 'hprewitt', 'vprewitt', 'roberts', 'roberts_positive_diagonal', 'roberts_negative_diagonal', 'denoise_tv_chambolle', 'tv_denoise', 'denoise_bilateral', 'denoise_tv_bregman', 'rank_order', 'gabor_kernel', 'gabor_filter', 'threshold_otsu', 'threshold_adaptive'] Add filter.rank to __all__ of filter package
from .lpi_filter import inverse, wiener, LPIFilter2D from .ctmf import median_filter from ._canny import canny from .edges import (sobel, hsobel, vsobel, scharr, hscharr, vscharr, prewitt, hprewitt, vprewitt, roberts , roberts_positive_diagonal, roberts_negative_diagonal) from ._denoise import denoise_tv_chambolle, tv_denoise from ._denoise_cy import denoise_bilateral, denoise_tv_bregman from ._rank_order import rank_order from ._gabor import gabor_kernel, gabor_filter from .thresholding import threshold_otsu, threshold_adaptive from . import rank __all__ = ['inverse', 'wiener', 'LPIFilter2D', 'median_filter', 'canny', 'sobel', 'hsobel', 'vsobel', 'scharr', 'hscharr', 'vscharr', 'prewitt', 'hprewitt', 'vprewitt', 'roberts', 'roberts_positive_diagonal', 'roberts_negative_diagonal', 'denoise_tv_chambolle', 'tv_denoise', 'denoise_bilateral', 'denoise_tv_bregman', 'rank_order', 'gabor_kernel', 'gabor_filter', 'threshold_otsu', 'threshold_adaptive', 'rank']
<commit_before>from .lpi_filter import inverse, wiener, LPIFilter2D from .ctmf import median_filter from ._canny import canny from .edges import (sobel, hsobel, vsobel, scharr, hscharr, vscharr, prewitt, hprewitt, vprewitt, roberts , roberts_positive_diagonal, roberts_negative_diagonal) from ._denoise import denoise_tv_chambolle, tv_denoise from ._denoise_cy import denoise_bilateral, denoise_tv_bregman from ._rank_order import rank_order from ._gabor import gabor_kernel, gabor_filter from .thresholding import threshold_otsu, threshold_adaptive __all__ = ['inverse', 'wiener', 'LPIFilter2D', 'median_filter', 'canny', 'sobel', 'hsobel', 'vsobel', 'scharr', 'hscharr', 'vscharr', 'prewitt', 'hprewitt', 'vprewitt', 'roberts', 'roberts_positive_diagonal', 'roberts_negative_diagonal', 'denoise_tv_chambolle', 'tv_denoise', 'denoise_bilateral', 'denoise_tv_bregman', 'rank_order', 'gabor_kernel', 'gabor_filter', 'threshold_otsu', 'threshold_adaptive'] <commit_msg>Add filter.rank to __all__ of filter package<commit_after>
from .lpi_filter import inverse, wiener, LPIFilter2D from .ctmf import median_filter from ._canny import canny from .edges import (sobel, hsobel, vsobel, scharr, hscharr, vscharr, prewitt, hprewitt, vprewitt, roberts , roberts_positive_diagonal, roberts_negative_diagonal) from ._denoise import denoise_tv_chambolle, tv_denoise from ._denoise_cy import denoise_bilateral, denoise_tv_bregman from ._rank_order import rank_order from ._gabor import gabor_kernel, gabor_filter from .thresholding import threshold_otsu, threshold_adaptive from . import rank __all__ = ['inverse', 'wiener', 'LPIFilter2D', 'median_filter', 'canny', 'sobel', 'hsobel', 'vsobel', 'scharr', 'hscharr', 'vscharr', 'prewitt', 'hprewitt', 'vprewitt', 'roberts', 'roberts_positive_diagonal', 'roberts_negative_diagonal', 'denoise_tv_chambolle', 'tv_denoise', 'denoise_bilateral', 'denoise_tv_bregman', 'rank_order', 'gabor_kernel', 'gabor_filter', 'threshold_otsu', 'threshold_adaptive', 'rank']
from .lpi_filter import inverse, wiener, LPIFilter2D from .ctmf import median_filter from ._canny import canny from .edges import (sobel, hsobel, vsobel, scharr, hscharr, vscharr, prewitt, hprewitt, vprewitt, roberts , roberts_positive_diagonal, roberts_negative_diagonal) from ._denoise import denoise_tv_chambolle, tv_denoise from ._denoise_cy import denoise_bilateral, denoise_tv_bregman from ._rank_order import rank_order from ._gabor import gabor_kernel, gabor_filter from .thresholding import threshold_otsu, threshold_adaptive __all__ = ['inverse', 'wiener', 'LPIFilter2D', 'median_filter', 'canny', 'sobel', 'hsobel', 'vsobel', 'scharr', 'hscharr', 'vscharr', 'prewitt', 'hprewitt', 'vprewitt', 'roberts', 'roberts_positive_diagonal', 'roberts_negative_diagonal', 'denoise_tv_chambolle', 'tv_denoise', 'denoise_bilateral', 'denoise_tv_bregman', 'rank_order', 'gabor_kernel', 'gabor_filter', 'threshold_otsu', 'threshold_adaptive'] Add filter.rank to __all__ of filter packagefrom .lpi_filter import inverse, wiener, LPIFilter2D from .ctmf import median_filter from ._canny import canny from .edges import (sobel, hsobel, vsobel, scharr, hscharr, vscharr, prewitt, hprewitt, vprewitt, roberts , roberts_positive_diagonal, roberts_negative_diagonal) from ._denoise import denoise_tv_chambolle, tv_denoise from ._denoise_cy import denoise_bilateral, denoise_tv_bregman from ._rank_order import rank_order from ._gabor import gabor_kernel, gabor_filter from .thresholding import threshold_otsu, threshold_adaptive from . import rank __all__ = ['inverse', 'wiener', 'LPIFilter2D', 'median_filter', 'canny', 'sobel', 'hsobel', 'vsobel', 'scharr', 'hscharr', 'vscharr', 'prewitt', 'hprewitt', 'vprewitt', 'roberts', 'roberts_positive_diagonal', 'roberts_negative_diagonal', 'denoise_tv_chambolle', 'tv_denoise', 'denoise_bilateral', 'denoise_tv_bregman', 'rank_order', 'gabor_kernel', 'gabor_filter', 'threshold_otsu', 'threshold_adaptive', 'rank']
<commit_before>from .lpi_filter import inverse, wiener, LPIFilter2D from .ctmf import median_filter from ._canny import canny from .edges import (sobel, hsobel, vsobel, scharr, hscharr, vscharr, prewitt, hprewitt, vprewitt, roberts , roberts_positive_diagonal, roberts_negative_diagonal) from ._denoise import denoise_tv_chambolle, tv_denoise from ._denoise_cy import denoise_bilateral, denoise_tv_bregman from ._rank_order import rank_order from ._gabor import gabor_kernel, gabor_filter from .thresholding import threshold_otsu, threshold_adaptive __all__ = ['inverse', 'wiener', 'LPIFilter2D', 'median_filter', 'canny', 'sobel', 'hsobel', 'vsobel', 'scharr', 'hscharr', 'vscharr', 'prewitt', 'hprewitt', 'vprewitt', 'roberts', 'roberts_positive_diagonal', 'roberts_negative_diagonal', 'denoise_tv_chambolle', 'tv_denoise', 'denoise_bilateral', 'denoise_tv_bregman', 'rank_order', 'gabor_kernel', 'gabor_filter', 'threshold_otsu', 'threshold_adaptive'] <commit_msg>Add filter.rank to __all__ of filter package<commit_after>from .lpi_filter import inverse, wiener, LPIFilter2D from .ctmf import median_filter from ._canny import canny from .edges import (sobel, hsobel, vsobel, scharr, hscharr, vscharr, prewitt, hprewitt, vprewitt, roberts , roberts_positive_diagonal, roberts_negative_diagonal) from ._denoise import denoise_tv_chambolle, tv_denoise from ._denoise_cy import denoise_bilateral, denoise_tv_bregman from ._rank_order import rank_order from ._gabor import gabor_kernel, gabor_filter from .thresholding import threshold_otsu, threshold_adaptive from . import rank __all__ = ['inverse', 'wiener', 'LPIFilter2D', 'median_filter', 'canny', 'sobel', 'hsobel', 'vsobel', 'scharr', 'hscharr', 'vscharr', 'prewitt', 'hprewitt', 'vprewitt', 'roberts', 'roberts_positive_diagonal', 'roberts_negative_diagonal', 'denoise_tv_chambolle', 'tv_denoise', 'denoise_bilateral', 'denoise_tv_bregman', 'rank_order', 'gabor_kernel', 'gabor_filter', 'threshold_otsu', 'threshold_adaptive', 'rank']
98a2b7e11eb3e0d5ddc89a4d40c3d10586e400ab
website/filters/__init__.py
website/filters/__init__.py
import hashlib import urllib # Adapted from https://github.com/zzzsochi/Flask-Gravatar/blob/master/flaskext/gravatar.py def gravatar(user, use_ssl=False, d=None, r=None, size=None): if use_ssl: base_url = 'https://secure.gravatar.com/avatar/' else: base_url = 'http://www.gravatar.com/avatar/' # user can be a User instance or a username string username = user.username if hasattr(user, 'username') else user hash_code = hashlib.md5(unicode(username).encode('utf-8')).hexdigest() url = base_url + '?' # Order of query params matters, due to a quirk with gravatar params = [ ('s', size), ('d', 'identicon'), ] if r: params.append(('r', r)) url = base_url + hash_code + '?' + urllib.urlencode(params) return url
import hashlib import urllib # Adapted from https://github.com/zzzsochi/Flask-Gravatar/blob/master/flaskext/gravatar.py def gravatar(user, use_ssl=False, d=None, r=None, size=None): if use_ssl: base_url = 'https://secure.gravatar.com/avatar/' else: base_url = 'http://www.gravatar.com/avatar/' # user can be a User instance or a username string username = user.username if hasattr(user, 'username') else user hash_code = hashlib.md5(unicode(username).encode('utf-8')).hexdigest() url = base_url + '?' # Order of query params matters, due to a quirk with gravatar params = [ ('d', 'identicon'), ('s', size), ] if r: params.append(('r', r)) url = base_url + hash_code + '?' + urllib.urlencode(params) return url
Fix ordering of query params
Fix ordering of query params 3rd time's a charm
Python
apache-2.0
mluke93/osf.io,binoculars/osf.io,leb2dg/osf.io,caseyrygt/osf.io,samanehsan/osf.io,cwisecarver/osf.io,petermalcolm/osf.io,kwierman/osf.io,emetsger/osf.io,mluo613/osf.io,wearpants/osf.io,samchrisinger/osf.io,amyshi188/osf.io,jnayak1/osf.io,TomHeatwole/osf.io,petermalcolm/osf.io,TomBaxter/osf.io,amyshi188/osf.io,TomHeatwole/osf.io,doublebits/osf.io,billyhunt/osf.io,crcresearch/osf.io,Ghalko/osf.io,saradbowman/osf.io,monikagrabowska/osf.io,mluo613/osf.io,alexschiller/osf.io,ticklemepierce/osf.io,njantrania/osf.io,caseyrollins/osf.io,zachjanicki/osf.io,brianjgeiger/osf.io,laurenrevere/osf.io,HalcyonChimera/osf.io,laurenrevere/osf.io,doublebits/osf.io,njantrania/osf.io,acshi/osf.io,danielneis/osf.io,mfraezz/osf.io,cslzchen/osf.io,Johnetordoff/osf.io,alexschiller/osf.io,ZobairAlijan/osf.io,caseyrollins/osf.io,KAsante95/osf.io,brandonPurvis/osf.io,kwierman/osf.io,zamattiac/osf.io,mluke93/osf.io,adlius/osf.io,Ghalko/osf.io,cslzchen/osf.io,kwierman/osf.io,DanielSBrown/osf.io,kch8qx/osf.io,mfraezz/osf.io,kwierman/osf.io,SSJohns/osf.io,asanfilippo7/osf.io,icereval/osf.io,RomanZWang/osf.io,alexschiller/osf.io,chennan47/osf.io,GageGaskins/osf.io,caneruguz/osf.io,baylee-d/osf.io,GageGaskins/osf.io,chennan47/osf.io,samanehsan/osf.io,petermalcolm/osf.io,RomanZWang/osf.io,DanielSBrown/osf.io,brianjgeiger/osf.io,kch8qx/osf.io,laurenrevere/osf.io,caseyrygt/osf.io,billyhunt/osf.io,billyhunt/osf.io,jnayak1/osf.io,chrisseto/osf.io,adlius/osf.io,wearpants/osf.io,cwisecarver/osf.io,samchrisinger/osf.io,cwisecarver/osf.io,alexschiller/osf.io,Nesiehr/osf.io,Johnetordoff/osf.io,acshi/osf.io,abought/osf.io,TomBaxter/osf.io,RomanZWang/osf.io,monikagrabowska/osf.io,wearpants/osf.io,binoculars/osf.io,caseyrollins/osf.io,zamattiac/osf.io,KAsante95/osf.io,mattclark/osf.io,felliott/osf.io,ticklemepierce/osf.io,SSJohns/osf.io,brandonPurvis/osf.io,brianjgeiger/osf.io,baylee-d/osf.io,HalcyonChimera/osf.io,HalcyonChimera/osf.io,mluo613/osf.io,zamattiac/osf.io,GageGaskins/osf.io,rdhyee/osf.io,Johnetordoff/osf.io,CenterForOpenScience/osf.io,cslzchen/osf.io,mattclark/osf.io,zachjanicki/osf.io,kch8qx/osf.io,mluke93/osf.io,KAsante95/osf.io,GageGaskins/osf.io,felliott/osf.io,haoyuchen1992/osf.io,erinspace/osf.io,HalcyonChimera/osf.io,hmoco/osf.io,felliott/osf.io,brandonPurvis/osf.io,zachjanicki/osf.io,ZobairAlijan/osf.io,mfraezz/osf.io,DanielSBrown/osf.io,cslzchen/osf.io,amyshi188/osf.io,saradbowman/osf.io,emetsger/osf.io,pattisdr/osf.io,rdhyee/osf.io,monikagrabowska/osf.io,zamattiac/osf.io,rdhyee/osf.io,erinspace/osf.io,haoyuchen1992/osf.io,chrisseto/osf.io,chrisseto/osf.io,Johnetordoff/osf.io,leb2dg/osf.io,samanehsan/osf.io,mluke93/osf.io,abought/osf.io,abought/osf.io,adlius/osf.io,RomanZWang/osf.io,caneruguz/osf.io,njantrania/osf.io,billyhunt/osf.io,erinspace/osf.io,brianjgeiger/osf.io,monikagrabowska/osf.io,SSJohns/osf.io,Nesiehr/osf.io,petermalcolm/osf.io,KAsante95/osf.io,hmoco/osf.io,caneruguz/osf.io,caseyrygt/osf.io,emetsger/osf.io,doublebits/osf.io,jnayak1/osf.io,danielneis/osf.io,danielneis/osf.io,alexschiller/osf.io,pattisdr/osf.io,CenterForOpenScience/osf.io,mattclark/osf.io,kch8qx/osf.io,cosenal/osf.io,caseyrygt/osf.io,kch8qx/osf.io,icereval/osf.io,aaxelb/osf.io,aaxelb/osf.io,crcresearch/osf.io,felliott/osf.io,mluo613/osf.io,haoyuchen1992/osf.io,sloria/osf.io,Ghalko/osf.io,hmoco/osf.io,aaxelb/osf.io,billyhunt/osf.io,cwisecarver/osf.io,RomanZWang/osf.io,samchrisinger/osf.io,GageGaskins/osf.io,jnayak1/osf.io,cosenal/osf.io,rdhyee/osf.io,Nesiehr/osf.io,acshi/osf.io,KAsante95/osf.io,samanehsan/osf.io,emetsger/osf.io,haoyuchen1992/osf.io,danielneis/osf.io,asanfilippo7/osf.io,TomHeatwole/osf.io,leb2dg/osf.io,adlius/osf.io,hmoco/osf.io,CenterForOpenScience/osf.io,cosenal/osf.io,brandonPurvis/osf.io,crcresearch/osf.io,wearpants/osf.io,mluo613/osf.io,ticklemepierce/osf.io,ticklemepierce/osf.io,aaxelb/osf.io,njantrania/osf.io,acshi/osf.io,mfraezz/osf.io,asanfilippo7/osf.io,brandonPurvis/osf.io,ZobairAlijan/osf.io,sloria/osf.io,binoculars/osf.io,CenterForOpenScience/osf.io,acshi/osf.io,asanfilippo7/osf.io,chrisseto/osf.io,DanielSBrown/osf.io,abought/osf.io,leb2dg/osf.io,TomBaxter/osf.io,samchrisinger/osf.io,amyshi188/osf.io,SSJohns/osf.io,doublebits/osf.io,cosenal/osf.io,caneruguz/osf.io,zachjanicki/osf.io,TomHeatwole/osf.io,pattisdr/osf.io,Nesiehr/osf.io,icereval/osf.io,chennan47/osf.io,monikagrabowska/osf.io,baylee-d/osf.io,doublebits/osf.io,sloria/osf.io,Ghalko/osf.io,ZobairAlijan/osf.io
import hashlib import urllib # Adapted from https://github.com/zzzsochi/Flask-Gravatar/blob/master/flaskext/gravatar.py def gravatar(user, use_ssl=False, d=None, r=None, size=None): if use_ssl: base_url = 'https://secure.gravatar.com/avatar/' else: base_url = 'http://www.gravatar.com/avatar/' # user can be a User instance or a username string username = user.username if hasattr(user, 'username') else user hash_code = hashlib.md5(unicode(username).encode('utf-8')).hexdigest() url = base_url + '?' # Order of query params matters, due to a quirk with gravatar params = [ ('s', size), ('d', 'identicon'), ] if r: params.append(('r', r)) url = base_url + hash_code + '?' + urllib.urlencode(params) return url Fix ordering of query params 3rd time's a charm
import hashlib import urllib # Adapted from https://github.com/zzzsochi/Flask-Gravatar/blob/master/flaskext/gravatar.py def gravatar(user, use_ssl=False, d=None, r=None, size=None): if use_ssl: base_url = 'https://secure.gravatar.com/avatar/' else: base_url = 'http://www.gravatar.com/avatar/' # user can be a User instance or a username string username = user.username if hasattr(user, 'username') else user hash_code = hashlib.md5(unicode(username).encode('utf-8')).hexdigest() url = base_url + '?' # Order of query params matters, due to a quirk with gravatar params = [ ('d', 'identicon'), ('s', size), ] if r: params.append(('r', r)) url = base_url + hash_code + '?' + urllib.urlencode(params) return url
<commit_before>import hashlib import urllib # Adapted from https://github.com/zzzsochi/Flask-Gravatar/blob/master/flaskext/gravatar.py def gravatar(user, use_ssl=False, d=None, r=None, size=None): if use_ssl: base_url = 'https://secure.gravatar.com/avatar/' else: base_url = 'http://www.gravatar.com/avatar/' # user can be a User instance or a username string username = user.username if hasattr(user, 'username') else user hash_code = hashlib.md5(unicode(username).encode('utf-8')).hexdigest() url = base_url + '?' # Order of query params matters, due to a quirk with gravatar params = [ ('s', size), ('d', 'identicon'), ] if r: params.append(('r', r)) url = base_url + hash_code + '?' + urllib.urlencode(params) return url <commit_msg>Fix ordering of query params 3rd time's a charm<commit_after>
import hashlib import urllib # Adapted from https://github.com/zzzsochi/Flask-Gravatar/blob/master/flaskext/gravatar.py def gravatar(user, use_ssl=False, d=None, r=None, size=None): if use_ssl: base_url = 'https://secure.gravatar.com/avatar/' else: base_url = 'http://www.gravatar.com/avatar/' # user can be a User instance or a username string username = user.username if hasattr(user, 'username') else user hash_code = hashlib.md5(unicode(username).encode('utf-8')).hexdigest() url = base_url + '?' # Order of query params matters, due to a quirk with gravatar params = [ ('d', 'identicon'), ('s', size), ] if r: params.append(('r', r)) url = base_url + hash_code + '?' + urllib.urlencode(params) return url
import hashlib import urllib # Adapted from https://github.com/zzzsochi/Flask-Gravatar/blob/master/flaskext/gravatar.py def gravatar(user, use_ssl=False, d=None, r=None, size=None): if use_ssl: base_url = 'https://secure.gravatar.com/avatar/' else: base_url = 'http://www.gravatar.com/avatar/' # user can be a User instance or a username string username = user.username if hasattr(user, 'username') else user hash_code = hashlib.md5(unicode(username).encode('utf-8')).hexdigest() url = base_url + '?' # Order of query params matters, due to a quirk with gravatar params = [ ('s', size), ('d', 'identicon'), ] if r: params.append(('r', r)) url = base_url + hash_code + '?' + urllib.urlencode(params) return url Fix ordering of query params 3rd time's a charmimport hashlib import urllib # Adapted from https://github.com/zzzsochi/Flask-Gravatar/blob/master/flaskext/gravatar.py def gravatar(user, use_ssl=False, d=None, r=None, size=None): if use_ssl: base_url = 'https://secure.gravatar.com/avatar/' else: base_url = 'http://www.gravatar.com/avatar/' # user can be a User instance or a username string username = user.username if hasattr(user, 'username') else user hash_code = hashlib.md5(unicode(username).encode('utf-8')).hexdigest() url = base_url + '?' # Order of query params matters, due to a quirk with gravatar params = [ ('d', 'identicon'), ('s', size), ] if r: params.append(('r', r)) url = base_url + hash_code + '?' + urllib.urlencode(params) return url
<commit_before>import hashlib import urllib # Adapted from https://github.com/zzzsochi/Flask-Gravatar/blob/master/flaskext/gravatar.py def gravatar(user, use_ssl=False, d=None, r=None, size=None): if use_ssl: base_url = 'https://secure.gravatar.com/avatar/' else: base_url = 'http://www.gravatar.com/avatar/' # user can be a User instance or a username string username = user.username if hasattr(user, 'username') else user hash_code = hashlib.md5(unicode(username).encode('utf-8')).hexdigest() url = base_url + '?' # Order of query params matters, due to a quirk with gravatar params = [ ('s', size), ('d', 'identicon'), ] if r: params.append(('r', r)) url = base_url + hash_code + '?' + urllib.urlencode(params) return url <commit_msg>Fix ordering of query params 3rd time's a charm<commit_after>import hashlib import urllib # Adapted from https://github.com/zzzsochi/Flask-Gravatar/blob/master/flaskext/gravatar.py def gravatar(user, use_ssl=False, d=None, r=None, size=None): if use_ssl: base_url = 'https://secure.gravatar.com/avatar/' else: base_url = 'http://www.gravatar.com/avatar/' # user can be a User instance or a username string username = user.username if hasattr(user, 'username') else user hash_code = hashlib.md5(unicode(username).encode('utf-8')).hexdigest() url = base_url + '?' # Order of query params matters, due to a quirk with gravatar params = [ ('d', 'identicon'), ('s', size), ] if r: params.append(('r', r)) url = base_url + hash_code + '?' + urllib.urlencode(params) return url
774b64779b18ff0d8fba048ab4c4cae53662628a
ummeli/vlive/auth/middleware.py
ummeli/vlive/auth/middleware.py
from django.contrib.auth.middleware import RemoteUserMiddleware class VodafoneLiveUserMiddleware(RemoteUserMiddleware): header = 'HTTP_X_UP_CALLING_LINE_ID' class VodafoneLiveInfo(object): pass class VodafoneLiveInfoMiddleware(object): """ Friendlier access to device / request info that Vodafone Live makes available to us via HTTP Headers """ def process_request(self, request): vlive = VodafoneLiveInfo() vlive.msisdn = request.META.get('HTTP_X_UP_CALLING_LINE_ID', 'unknown') vlive.area = request.META.get('HTTP_X_VODAFONE_AREA', 'unknown') request.vlive = vlive print request.META
from django.contrib.auth.middleware import RemoteUserMiddleware class VodafoneLiveUserMiddleware(RemoteUserMiddleware): header = 'HTTP_X_UP_CALLING_LINE_ID' class VodafoneLiveInfo(object): pass class VodafoneLiveInfoMiddleware(object): """ Friendlier access to device / request info that Vodafone Live makes available to us via HTTP Headers """ def process_request(self, request): vlive = VodafoneLiveInfo() vlive.msisdn = request.META.get('HTTP_X_UP_CALLING_LINE_ID', 'unknown') vlive.area = request.META.get('HTTP_X_VODAFONE_AREA', 'unknown') request.vlive = vlive
Revert "printing META for troubleshooting"
Revert "printing META for troubleshooting" This reverts commit 42d15d528da14866f2f0479da6462c17a02d8c84.
Python
bsd-3-clause
praekelt/ummeli,praekelt/ummeli,praekelt/ummeli
from django.contrib.auth.middleware import RemoteUserMiddleware class VodafoneLiveUserMiddleware(RemoteUserMiddleware): header = 'HTTP_X_UP_CALLING_LINE_ID' class VodafoneLiveInfo(object): pass class VodafoneLiveInfoMiddleware(object): """ Friendlier access to device / request info that Vodafone Live makes available to us via HTTP Headers """ def process_request(self, request): vlive = VodafoneLiveInfo() vlive.msisdn = request.META.get('HTTP_X_UP_CALLING_LINE_ID', 'unknown') vlive.area = request.META.get('HTTP_X_VODAFONE_AREA', 'unknown') request.vlive = vlive print request.META Revert "printing META for troubleshooting" This reverts commit 42d15d528da14866f2f0479da6462c17a02d8c84.
from django.contrib.auth.middleware import RemoteUserMiddleware class VodafoneLiveUserMiddleware(RemoteUserMiddleware): header = 'HTTP_X_UP_CALLING_LINE_ID' class VodafoneLiveInfo(object): pass class VodafoneLiveInfoMiddleware(object): """ Friendlier access to device / request info that Vodafone Live makes available to us via HTTP Headers """ def process_request(self, request): vlive = VodafoneLiveInfo() vlive.msisdn = request.META.get('HTTP_X_UP_CALLING_LINE_ID', 'unknown') vlive.area = request.META.get('HTTP_X_VODAFONE_AREA', 'unknown') request.vlive = vlive
<commit_before>from django.contrib.auth.middleware import RemoteUserMiddleware class VodafoneLiveUserMiddleware(RemoteUserMiddleware): header = 'HTTP_X_UP_CALLING_LINE_ID' class VodafoneLiveInfo(object): pass class VodafoneLiveInfoMiddleware(object): """ Friendlier access to device / request info that Vodafone Live makes available to us via HTTP Headers """ def process_request(self, request): vlive = VodafoneLiveInfo() vlive.msisdn = request.META.get('HTTP_X_UP_CALLING_LINE_ID', 'unknown') vlive.area = request.META.get('HTTP_X_VODAFONE_AREA', 'unknown') request.vlive = vlive print request.META <commit_msg>Revert "printing META for troubleshooting" This reverts commit 42d15d528da14866f2f0479da6462c17a02d8c84.<commit_after>
from django.contrib.auth.middleware import RemoteUserMiddleware class VodafoneLiveUserMiddleware(RemoteUserMiddleware): header = 'HTTP_X_UP_CALLING_LINE_ID' class VodafoneLiveInfo(object): pass class VodafoneLiveInfoMiddleware(object): """ Friendlier access to device / request info that Vodafone Live makes available to us via HTTP Headers """ def process_request(self, request): vlive = VodafoneLiveInfo() vlive.msisdn = request.META.get('HTTP_X_UP_CALLING_LINE_ID', 'unknown') vlive.area = request.META.get('HTTP_X_VODAFONE_AREA', 'unknown') request.vlive = vlive
from django.contrib.auth.middleware import RemoteUserMiddleware class VodafoneLiveUserMiddleware(RemoteUserMiddleware): header = 'HTTP_X_UP_CALLING_LINE_ID' class VodafoneLiveInfo(object): pass class VodafoneLiveInfoMiddleware(object): """ Friendlier access to device / request info that Vodafone Live makes available to us via HTTP Headers """ def process_request(self, request): vlive = VodafoneLiveInfo() vlive.msisdn = request.META.get('HTTP_X_UP_CALLING_LINE_ID', 'unknown') vlive.area = request.META.get('HTTP_X_VODAFONE_AREA', 'unknown') request.vlive = vlive print request.META Revert "printing META for troubleshooting" This reverts commit 42d15d528da14866f2f0479da6462c17a02d8c84.from django.contrib.auth.middleware import RemoteUserMiddleware class VodafoneLiveUserMiddleware(RemoteUserMiddleware): header = 'HTTP_X_UP_CALLING_LINE_ID' class VodafoneLiveInfo(object): pass class VodafoneLiveInfoMiddleware(object): """ Friendlier access to device / request info that Vodafone Live makes available to us via HTTP Headers """ def process_request(self, request): vlive = VodafoneLiveInfo() vlive.msisdn = request.META.get('HTTP_X_UP_CALLING_LINE_ID', 'unknown') vlive.area = request.META.get('HTTP_X_VODAFONE_AREA', 'unknown') request.vlive = vlive
<commit_before>from django.contrib.auth.middleware import RemoteUserMiddleware class VodafoneLiveUserMiddleware(RemoteUserMiddleware): header = 'HTTP_X_UP_CALLING_LINE_ID' class VodafoneLiveInfo(object): pass class VodafoneLiveInfoMiddleware(object): """ Friendlier access to device / request info that Vodafone Live makes available to us via HTTP Headers """ def process_request(self, request): vlive = VodafoneLiveInfo() vlive.msisdn = request.META.get('HTTP_X_UP_CALLING_LINE_ID', 'unknown') vlive.area = request.META.get('HTTP_X_VODAFONE_AREA', 'unknown') request.vlive = vlive print request.META <commit_msg>Revert "printing META for troubleshooting" This reverts commit 42d15d528da14866f2f0479da6462c17a02d8c84.<commit_after>from django.contrib.auth.middleware import RemoteUserMiddleware class VodafoneLiveUserMiddleware(RemoteUserMiddleware): header = 'HTTP_X_UP_CALLING_LINE_ID' class VodafoneLiveInfo(object): pass class VodafoneLiveInfoMiddleware(object): """ Friendlier access to device / request info that Vodafone Live makes available to us via HTTP Headers """ def process_request(self, request): vlive = VodafoneLiveInfo() vlive.msisdn = request.META.get('HTTP_X_UP_CALLING_LINE_ID', 'unknown') vlive.area = request.META.get('HTTP_X_VODAFONE_AREA', 'unknown') request.vlive = vlive
de1baa49fc34f8ecf4f7df4c723456348281df69
splunk_handler/__init__.py
splunk_handler/__init__.py
import logging import socket import traceback from threading import Thread import requests class SplunkHandler(logging.Handler): """ A logging handler to send events to a Splunk Enterprise instance """ def __init__(self, host, port, username, password, index): logging.Handler.__init__(self) self.host = host self.port = port self.username = username self.password = password self.index = index def emit(self, record): thread = Thread(target=self._async_emit, args=(record, )) thread.start() def _async_emit(self, record): try: params = { 'host': socket.gethostname(), 'index': self.index, 'source': record.pathname, 'sourcetype': 'json' } url = 'https://%s:%s/services/receivers/simple' % (self.host, self.port) payload = self.format(record) auth = (self.username, self.password) r = requests.post( url, auth=auth, data=payload, params=params ) r.close() except Exception, e: print "Traceback:\n" + traceback.format_exc() print "Exception in Splunk logging handler: %s" % str(e)
import logging import socket import traceback from threading import Thread import requests class SplunkHandler(logging.Handler): """ A logging handler to send events to a Splunk Enterprise instance """ def __init__(self, host, port, username, password, index): logging.Handler.__init__(self) self.host = host self.port = port self.username = username self.password = password self.index = index requests_log = logging.getLogger('requests') requests_log.propagate = False def emit(self, record): thread = Thread(target=self._async_emit, args=(record, )) thread.start() def _async_emit(self, record): try: params = { 'host': socket.gethostname(), 'index': self.index, 'source': record.pathname, 'sourcetype': 'json' } url = 'https://%s:%s/services/receivers/simple' % (self.host, self.port) payload = self.format(record) auth = (self.username, self.password) r = requests.post( url, auth=auth, data=payload, params=params ) r.close() except Exception, e: print "Traceback:\n" + traceback.format_exc() print "Exception in Splunk logging handler: %s" % str(e)
Add code to silence requests logger in the handler
Add code to silence requests logger in the handler
Python
mit
zach-taylor/splunk_handler,sullivanmatt/splunk_handler
import logging import socket import traceback from threading import Thread import requests class SplunkHandler(logging.Handler): """ A logging handler to send events to a Splunk Enterprise instance """ def __init__(self, host, port, username, password, index): logging.Handler.__init__(self) self.host = host self.port = port self.username = username self.password = password self.index = index def emit(self, record): thread = Thread(target=self._async_emit, args=(record, )) thread.start() def _async_emit(self, record): try: params = { 'host': socket.gethostname(), 'index': self.index, 'source': record.pathname, 'sourcetype': 'json' } url = 'https://%s:%s/services/receivers/simple' % (self.host, self.port) payload = self.format(record) auth = (self.username, self.password) r = requests.post( url, auth=auth, data=payload, params=params ) r.close() except Exception, e: print "Traceback:\n" + traceback.format_exc() print "Exception in Splunk logging handler: %s" % str(e) Add code to silence requests logger in the handler
import logging import socket import traceback from threading import Thread import requests class SplunkHandler(logging.Handler): """ A logging handler to send events to a Splunk Enterprise instance """ def __init__(self, host, port, username, password, index): logging.Handler.__init__(self) self.host = host self.port = port self.username = username self.password = password self.index = index requests_log = logging.getLogger('requests') requests_log.propagate = False def emit(self, record): thread = Thread(target=self._async_emit, args=(record, )) thread.start() def _async_emit(self, record): try: params = { 'host': socket.gethostname(), 'index': self.index, 'source': record.pathname, 'sourcetype': 'json' } url = 'https://%s:%s/services/receivers/simple' % (self.host, self.port) payload = self.format(record) auth = (self.username, self.password) r = requests.post( url, auth=auth, data=payload, params=params ) r.close() except Exception, e: print "Traceback:\n" + traceback.format_exc() print "Exception in Splunk logging handler: %s" % str(e)
<commit_before>import logging import socket import traceback from threading import Thread import requests class SplunkHandler(logging.Handler): """ A logging handler to send events to a Splunk Enterprise instance """ def __init__(self, host, port, username, password, index): logging.Handler.__init__(self) self.host = host self.port = port self.username = username self.password = password self.index = index def emit(self, record): thread = Thread(target=self._async_emit, args=(record, )) thread.start() def _async_emit(self, record): try: params = { 'host': socket.gethostname(), 'index': self.index, 'source': record.pathname, 'sourcetype': 'json' } url = 'https://%s:%s/services/receivers/simple' % (self.host, self.port) payload = self.format(record) auth = (self.username, self.password) r = requests.post( url, auth=auth, data=payload, params=params ) r.close() except Exception, e: print "Traceback:\n" + traceback.format_exc() print "Exception in Splunk logging handler: %s" % str(e) <commit_msg>Add code to silence requests logger in the handler<commit_after>
import logging import socket import traceback from threading import Thread import requests class SplunkHandler(logging.Handler): """ A logging handler to send events to a Splunk Enterprise instance """ def __init__(self, host, port, username, password, index): logging.Handler.__init__(self) self.host = host self.port = port self.username = username self.password = password self.index = index requests_log = logging.getLogger('requests') requests_log.propagate = False def emit(self, record): thread = Thread(target=self._async_emit, args=(record, )) thread.start() def _async_emit(self, record): try: params = { 'host': socket.gethostname(), 'index': self.index, 'source': record.pathname, 'sourcetype': 'json' } url = 'https://%s:%s/services/receivers/simple' % (self.host, self.port) payload = self.format(record) auth = (self.username, self.password) r = requests.post( url, auth=auth, data=payload, params=params ) r.close() except Exception, e: print "Traceback:\n" + traceback.format_exc() print "Exception in Splunk logging handler: %s" % str(e)
import logging import socket import traceback from threading import Thread import requests class SplunkHandler(logging.Handler): """ A logging handler to send events to a Splunk Enterprise instance """ def __init__(self, host, port, username, password, index): logging.Handler.__init__(self) self.host = host self.port = port self.username = username self.password = password self.index = index def emit(self, record): thread = Thread(target=self._async_emit, args=(record, )) thread.start() def _async_emit(self, record): try: params = { 'host': socket.gethostname(), 'index': self.index, 'source': record.pathname, 'sourcetype': 'json' } url = 'https://%s:%s/services/receivers/simple' % (self.host, self.port) payload = self.format(record) auth = (self.username, self.password) r = requests.post( url, auth=auth, data=payload, params=params ) r.close() except Exception, e: print "Traceback:\n" + traceback.format_exc() print "Exception in Splunk logging handler: %s" % str(e) Add code to silence requests logger in the handlerimport logging import socket import traceback from threading import Thread import requests class SplunkHandler(logging.Handler): """ A logging handler to send events to a Splunk Enterprise instance """ def __init__(self, host, port, username, password, index): logging.Handler.__init__(self) self.host = host self.port = port self.username = username self.password = password self.index = index requests_log = logging.getLogger('requests') requests_log.propagate = False def emit(self, record): thread = Thread(target=self._async_emit, args=(record, )) thread.start() def _async_emit(self, record): try: params = { 'host': socket.gethostname(), 'index': self.index, 'source': record.pathname, 'sourcetype': 'json' } url = 'https://%s:%s/services/receivers/simple' % (self.host, self.port) payload = self.format(record) auth = (self.username, self.password) r = requests.post( url, auth=auth, data=payload, params=params ) r.close() except Exception, e: print "Traceback:\n" + traceback.format_exc() print "Exception in Splunk logging handler: %s" % str(e)
<commit_before>import logging import socket import traceback from threading import Thread import requests class SplunkHandler(logging.Handler): """ A logging handler to send events to a Splunk Enterprise instance """ def __init__(self, host, port, username, password, index): logging.Handler.__init__(self) self.host = host self.port = port self.username = username self.password = password self.index = index def emit(self, record): thread = Thread(target=self._async_emit, args=(record, )) thread.start() def _async_emit(self, record): try: params = { 'host': socket.gethostname(), 'index': self.index, 'source': record.pathname, 'sourcetype': 'json' } url = 'https://%s:%s/services/receivers/simple' % (self.host, self.port) payload = self.format(record) auth = (self.username, self.password) r = requests.post( url, auth=auth, data=payload, params=params ) r.close() except Exception, e: print "Traceback:\n" + traceback.format_exc() print "Exception in Splunk logging handler: %s" % str(e) <commit_msg>Add code to silence requests logger in the handler<commit_after>import logging import socket import traceback from threading import Thread import requests class SplunkHandler(logging.Handler): """ A logging handler to send events to a Splunk Enterprise instance """ def __init__(self, host, port, username, password, index): logging.Handler.__init__(self) self.host = host self.port = port self.username = username self.password = password self.index = index requests_log = logging.getLogger('requests') requests_log.propagate = False def emit(self, record): thread = Thread(target=self._async_emit, args=(record, )) thread.start() def _async_emit(self, record): try: params = { 'host': socket.gethostname(), 'index': self.index, 'source': record.pathname, 'sourcetype': 'json' } url = 'https://%s:%s/services/receivers/simple' % (self.host, self.port) payload = self.format(record) auth = (self.username, self.password) r = requests.post( url, auth=auth, data=payload, params=params ) r.close() except Exception, e: print "Traceback:\n" + traceback.format_exc() print "Exception in Splunk logging handler: %s" % str(e)
d9abb2f56720480169d394a2cadd3cb9a77ac4f6
app/main/views/frameworks.py
app/main/views/frameworks.py
from flask import jsonify from sqlalchemy.types import String from sqlalchemy import func import datetime from .. import main from ...models import db, Framework, DraftService, Service, User, Supplier, SelectionAnswers, AuditEvent @main.route('/frameworks', methods=['GET']) def list_frameworks(): frameworks = Framework.query.all() return jsonify( frameworks=[f.serialize() for f in frameworks] ) @main.route('/frameworks/g-cloud-7/stats', methods=['GET']) def get_framework_stats(): seven_days_ago = datetime.datetime.utcnow() + datetime.timedelta(-7) lot_column = DraftService.data['lot'].cast(String).label('lot') return str({ 'services_drafts': DraftService.query.filter( DraftService.status == "not-submitted" ).count(), 'services_complete': DraftService.query.filter( DraftService.status == "submitted" ).count(), 'services_by_lot': dict(db.session.query( lot_column, func.count(lot_column) ).group_by(lot_column).all()), 'users': User.query.count(), 'active_users': User.query.filter(User.logged_in_at > seven_days_ago).count(), 'suppliers': Supplier.query.count(), 'suppliers_interested': AuditEvent.query.filter(AuditEvent.type == 'register_framework_interest').count(), 'suppliers_with_complete_declaration': SelectionAnswers.find_by_framework('g-cloud-7').count() })
from flask import jsonify from sqlalchemy.types import String from sqlalchemy import func import datetime from .. import main from ...models import db, Framework, DraftService, Service, User, Supplier, SelectionAnswers, AuditEvent @main.route('/frameworks', methods=['GET']) def list_frameworks(): frameworks = Framework.query.all() return jsonify( frameworks=[f.serialize() for f in frameworks] ) @main.route('/frameworks/g-cloud-7/stats', methods=['GET']) def get_framework_stats(): seven_days_ago = datetime.datetime.utcnow() + datetime.timedelta(-7) lot_column = DraftService.data['lot'].cast(String).label('lot') return str({ 'services_by_status': dict(db.session.query( DraftService.status, func.count(DraftService.status) ).group_by(DraftService.status)), 'services_by_lot': dict(db.session.query( lot_column, func.count(lot_column) ).group_by(lot_column).all()), 'users': User.query.count(), 'active_users': User.query.filter(User.logged_in_at > seven_days_ago).count(), 'suppliers': Supplier.query.count(), 'suppliers_interested': AuditEvent.query.filter(AuditEvent.type == 'register_framework_interest').count(), 'suppliers_with_complete_declaration': SelectionAnswers.find_by_framework('g-cloud-7').count() })
Use one query with group_by for service status
Use one query with group_by for service status
Python
mit
alphagov/digitalmarketplace-api,alphagov/digitalmarketplace-api,alphagov/digitalmarketplace-api
from flask import jsonify from sqlalchemy.types import String from sqlalchemy import func import datetime from .. import main from ...models import db, Framework, DraftService, Service, User, Supplier, SelectionAnswers, AuditEvent @main.route('/frameworks', methods=['GET']) def list_frameworks(): frameworks = Framework.query.all() return jsonify( frameworks=[f.serialize() for f in frameworks] ) @main.route('/frameworks/g-cloud-7/stats', methods=['GET']) def get_framework_stats(): seven_days_ago = datetime.datetime.utcnow() + datetime.timedelta(-7) lot_column = DraftService.data['lot'].cast(String).label('lot') return str({ 'services_drafts': DraftService.query.filter( DraftService.status == "not-submitted" ).count(), 'services_complete': DraftService.query.filter( DraftService.status == "submitted" ).count(), 'services_by_lot': dict(db.session.query( lot_column, func.count(lot_column) ).group_by(lot_column).all()), 'users': User.query.count(), 'active_users': User.query.filter(User.logged_in_at > seven_days_ago).count(), 'suppliers': Supplier.query.count(), 'suppliers_interested': AuditEvent.query.filter(AuditEvent.type == 'register_framework_interest').count(), 'suppliers_with_complete_declaration': SelectionAnswers.find_by_framework('g-cloud-7').count() }) Use one query with group_by for service status
from flask import jsonify from sqlalchemy.types import String from sqlalchemy import func import datetime from .. import main from ...models import db, Framework, DraftService, Service, User, Supplier, SelectionAnswers, AuditEvent @main.route('/frameworks', methods=['GET']) def list_frameworks(): frameworks = Framework.query.all() return jsonify( frameworks=[f.serialize() for f in frameworks] ) @main.route('/frameworks/g-cloud-7/stats', methods=['GET']) def get_framework_stats(): seven_days_ago = datetime.datetime.utcnow() + datetime.timedelta(-7) lot_column = DraftService.data['lot'].cast(String).label('lot') return str({ 'services_by_status': dict(db.session.query( DraftService.status, func.count(DraftService.status) ).group_by(DraftService.status)), 'services_by_lot': dict(db.session.query( lot_column, func.count(lot_column) ).group_by(lot_column).all()), 'users': User.query.count(), 'active_users': User.query.filter(User.logged_in_at > seven_days_ago).count(), 'suppliers': Supplier.query.count(), 'suppliers_interested': AuditEvent.query.filter(AuditEvent.type == 'register_framework_interest').count(), 'suppliers_with_complete_declaration': SelectionAnswers.find_by_framework('g-cloud-7').count() })
<commit_before>from flask import jsonify from sqlalchemy.types import String from sqlalchemy import func import datetime from .. import main from ...models import db, Framework, DraftService, Service, User, Supplier, SelectionAnswers, AuditEvent @main.route('/frameworks', methods=['GET']) def list_frameworks(): frameworks = Framework.query.all() return jsonify( frameworks=[f.serialize() for f in frameworks] ) @main.route('/frameworks/g-cloud-7/stats', methods=['GET']) def get_framework_stats(): seven_days_ago = datetime.datetime.utcnow() + datetime.timedelta(-7) lot_column = DraftService.data['lot'].cast(String).label('lot') return str({ 'services_drafts': DraftService.query.filter( DraftService.status == "not-submitted" ).count(), 'services_complete': DraftService.query.filter( DraftService.status == "submitted" ).count(), 'services_by_lot': dict(db.session.query( lot_column, func.count(lot_column) ).group_by(lot_column).all()), 'users': User.query.count(), 'active_users': User.query.filter(User.logged_in_at > seven_days_ago).count(), 'suppliers': Supplier.query.count(), 'suppliers_interested': AuditEvent.query.filter(AuditEvent.type == 'register_framework_interest').count(), 'suppliers_with_complete_declaration': SelectionAnswers.find_by_framework('g-cloud-7').count() }) <commit_msg>Use one query with group_by for service status<commit_after>
from flask import jsonify from sqlalchemy.types import String from sqlalchemy import func import datetime from .. import main from ...models import db, Framework, DraftService, Service, User, Supplier, SelectionAnswers, AuditEvent @main.route('/frameworks', methods=['GET']) def list_frameworks(): frameworks = Framework.query.all() return jsonify( frameworks=[f.serialize() for f in frameworks] ) @main.route('/frameworks/g-cloud-7/stats', methods=['GET']) def get_framework_stats(): seven_days_ago = datetime.datetime.utcnow() + datetime.timedelta(-7) lot_column = DraftService.data['lot'].cast(String).label('lot') return str({ 'services_by_status': dict(db.session.query( DraftService.status, func.count(DraftService.status) ).group_by(DraftService.status)), 'services_by_lot': dict(db.session.query( lot_column, func.count(lot_column) ).group_by(lot_column).all()), 'users': User.query.count(), 'active_users': User.query.filter(User.logged_in_at > seven_days_ago).count(), 'suppliers': Supplier.query.count(), 'suppliers_interested': AuditEvent.query.filter(AuditEvent.type == 'register_framework_interest').count(), 'suppliers_with_complete_declaration': SelectionAnswers.find_by_framework('g-cloud-7').count() })
from flask import jsonify from sqlalchemy.types import String from sqlalchemy import func import datetime from .. import main from ...models import db, Framework, DraftService, Service, User, Supplier, SelectionAnswers, AuditEvent @main.route('/frameworks', methods=['GET']) def list_frameworks(): frameworks = Framework.query.all() return jsonify( frameworks=[f.serialize() for f in frameworks] ) @main.route('/frameworks/g-cloud-7/stats', methods=['GET']) def get_framework_stats(): seven_days_ago = datetime.datetime.utcnow() + datetime.timedelta(-7) lot_column = DraftService.data['lot'].cast(String).label('lot') return str({ 'services_drafts': DraftService.query.filter( DraftService.status == "not-submitted" ).count(), 'services_complete': DraftService.query.filter( DraftService.status == "submitted" ).count(), 'services_by_lot': dict(db.session.query( lot_column, func.count(lot_column) ).group_by(lot_column).all()), 'users': User.query.count(), 'active_users': User.query.filter(User.logged_in_at > seven_days_ago).count(), 'suppliers': Supplier.query.count(), 'suppliers_interested': AuditEvent.query.filter(AuditEvent.type == 'register_framework_interest').count(), 'suppliers_with_complete_declaration': SelectionAnswers.find_by_framework('g-cloud-7').count() }) Use one query with group_by for service statusfrom flask import jsonify from sqlalchemy.types import String from sqlalchemy import func import datetime from .. import main from ...models import db, Framework, DraftService, Service, User, Supplier, SelectionAnswers, AuditEvent @main.route('/frameworks', methods=['GET']) def list_frameworks(): frameworks = Framework.query.all() return jsonify( frameworks=[f.serialize() for f in frameworks] ) @main.route('/frameworks/g-cloud-7/stats', methods=['GET']) def get_framework_stats(): seven_days_ago = datetime.datetime.utcnow() + datetime.timedelta(-7) lot_column = DraftService.data['lot'].cast(String).label('lot') return str({ 'services_by_status': dict(db.session.query( DraftService.status, func.count(DraftService.status) ).group_by(DraftService.status)), 'services_by_lot': dict(db.session.query( lot_column, func.count(lot_column) ).group_by(lot_column).all()), 'users': User.query.count(), 'active_users': User.query.filter(User.logged_in_at > seven_days_ago).count(), 'suppliers': Supplier.query.count(), 'suppliers_interested': AuditEvent.query.filter(AuditEvent.type == 'register_framework_interest').count(), 'suppliers_with_complete_declaration': SelectionAnswers.find_by_framework('g-cloud-7').count() })
<commit_before>from flask import jsonify from sqlalchemy.types import String from sqlalchemy import func import datetime from .. import main from ...models import db, Framework, DraftService, Service, User, Supplier, SelectionAnswers, AuditEvent @main.route('/frameworks', methods=['GET']) def list_frameworks(): frameworks = Framework.query.all() return jsonify( frameworks=[f.serialize() for f in frameworks] ) @main.route('/frameworks/g-cloud-7/stats', methods=['GET']) def get_framework_stats(): seven_days_ago = datetime.datetime.utcnow() + datetime.timedelta(-7) lot_column = DraftService.data['lot'].cast(String).label('lot') return str({ 'services_drafts': DraftService.query.filter( DraftService.status == "not-submitted" ).count(), 'services_complete': DraftService.query.filter( DraftService.status == "submitted" ).count(), 'services_by_lot': dict(db.session.query( lot_column, func.count(lot_column) ).group_by(lot_column).all()), 'users': User.query.count(), 'active_users': User.query.filter(User.logged_in_at > seven_days_ago).count(), 'suppliers': Supplier.query.count(), 'suppliers_interested': AuditEvent.query.filter(AuditEvent.type == 'register_framework_interest').count(), 'suppliers_with_complete_declaration': SelectionAnswers.find_by_framework('g-cloud-7').count() }) <commit_msg>Use one query with group_by for service status<commit_after>from flask import jsonify from sqlalchemy.types import String from sqlalchemy import func import datetime from .. import main from ...models import db, Framework, DraftService, Service, User, Supplier, SelectionAnswers, AuditEvent @main.route('/frameworks', methods=['GET']) def list_frameworks(): frameworks = Framework.query.all() return jsonify( frameworks=[f.serialize() for f in frameworks] ) @main.route('/frameworks/g-cloud-7/stats', methods=['GET']) def get_framework_stats(): seven_days_ago = datetime.datetime.utcnow() + datetime.timedelta(-7) lot_column = DraftService.data['lot'].cast(String).label('lot') return str({ 'services_by_status': dict(db.session.query( DraftService.status, func.count(DraftService.status) ).group_by(DraftService.status)), 'services_by_lot': dict(db.session.query( lot_column, func.count(lot_column) ).group_by(lot_column).all()), 'users': User.query.count(), 'active_users': User.query.filter(User.logged_in_at > seven_days_ago).count(), 'suppliers': Supplier.query.count(), 'suppliers_interested': AuditEvent.query.filter(AuditEvent.type == 'register_framework_interest').count(), 'suppliers_with_complete_declaration': SelectionAnswers.find_by_framework('g-cloud-7').count() })
a0903bb9fd988662269e9f2ef7e38acd877a63d5
src/nodeconductor_saltstack/saltstack/handlers.py
src/nodeconductor_saltstack/saltstack/handlers.py
from __future__ import unicode_literals import logging from .log import event_logger logger = logging.getLogger(__name__) def log_saltstack_property_created(sender, instance, created=False, **kwargs): if created: event_logger.saltstack_property.info( '%s {property_name} has been created.' % instance.get_type_display_name(), event_type='saltstack_property_creation_succeeded', event_context={ 'property': instance, }) else: event_logger.saltstack_property.info( '%s {property_name} has been updated.' % instance.get_type_display_name(), event_type='saltstack_property_update_succeeded', event_context={ 'property': instance, }) def log_saltstack_property_deleted(sender, instance, **kwargs): event_logger.saltstack_property.info( '%s {property_name} has been deleted.' % instance.get_type_display_name(), event_type='saltstack_property_deletion_succeeded', event_context={ 'property': instance, })
from __future__ import unicode_literals import logging from .log import event_logger logger = logging.getLogger(__name__) def log_saltstack_property_created(sender, instance, created=False, **kwargs): if created: event_logger.saltstack_property.info( '%s {property_name} has been created in {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_creation_succeeded', event_context={ 'property': instance, }) else: event_logger.saltstack_property.info( '%s {property_name} has been updated in {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_update_succeeded', event_context={ 'property': instance, }) def log_saltstack_property_deleted(sender, instance, **kwargs): event_logger.saltstack_property.info( '%s {property_name} has been deleted from {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_deletion_succeeded', event_context={ 'property': instance, })
Add more details to event logs for property CRUD
Add more details to event logs for property CRUD
Python
mit
opennode/nodeconductor-saltstack
from __future__ import unicode_literals import logging from .log import event_logger logger = logging.getLogger(__name__) def log_saltstack_property_created(sender, instance, created=False, **kwargs): if created: event_logger.saltstack_property.info( '%s {property_name} has been created.' % instance.get_type_display_name(), event_type='saltstack_property_creation_succeeded', event_context={ 'property': instance, }) else: event_logger.saltstack_property.info( '%s {property_name} has been updated.' % instance.get_type_display_name(), event_type='saltstack_property_update_succeeded', event_context={ 'property': instance, }) def log_saltstack_property_deleted(sender, instance, **kwargs): event_logger.saltstack_property.info( '%s {property_name} has been deleted.' % instance.get_type_display_name(), event_type='saltstack_property_deletion_succeeded', event_context={ 'property': instance, }) Add more details to event logs for property CRUD
from __future__ import unicode_literals import logging from .log import event_logger logger = logging.getLogger(__name__) def log_saltstack_property_created(sender, instance, created=False, **kwargs): if created: event_logger.saltstack_property.info( '%s {property_name} has been created in {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_creation_succeeded', event_context={ 'property': instance, }) else: event_logger.saltstack_property.info( '%s {property_name} has been updated in {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_update_succeeded', event_context={ 'property': instance, }) def log_saltstack_property_deleted(sender, instance, **kwargs): event_logger.saltstack_property.info( '%s {property_name} has been deleted from {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_deletion_succeeded', event_context={ 'property': instance, })
<commit_before>from __future__ import unicode_literals import logging from .log import event_logger logger = logging.getLogger(__name__) def log_saltstack_property_created(sender, instance, created=False, **kwargs): if created: event_logger.saltstack_property.info( '%s {property_name} has been created.' % instance.get_type_display_name(), event_type='saltstack_property_creation_succeeded', event_context={ 'property': instance, }) else: event_logger.saltstack_property.info( '%s {property_name} has been updated.' % instance.get_type_display_name(), event_type='saltstack_property_update_succeeded', event_context={ 'property': instance, }) def log_saltstack_property_deleted(sender, instance, **kwargs): event_logger.saltstack_property.info( '%s {property_name} has been deleted.' % instance.get_type_display_name(), event_type='saltstack_property_deletion_succeeded', event_context={ 'property': instance, }) <commit_msg>Add more details to event logs for property CRUD<commit_after>
from __future__ import unicode_literals import logging from .log import event_logger logger = logging.getLogger(__name__) def log_saltstack_property_created(sender, instance, created=False, **kwargs): if created: event_logger.saltstack_property.info( '%s {property_name} has been created in {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_creation_succeeded', event_context={ 'property': instance, }) else: event_logger.saltstack_property.info( '%s {property_name} has been updated in {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_update_succeeded', event_context={ 'property': instance, }) def log_saltstack_property_deleted(sender, instance, **kwargs): event_logger.saltstack_property.info( '%s {property_name} has been deleted from {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_deletion_succeeded', event_context={ 'property': instance, })
from __future__ import unicode_literals import logging from .log import event_logger logger = logging.getLogger(__name__) def log_saltstack_property_created(sender, instance, created=False, **kwargs): if created: event_logger.saltstack_property.info( '%s {property_name} has been created.' % instance.get_type_display_name(), event_type='saltstack_property_creation_succeeded', event_context={ 'property': instance, }) else: event_logger.saltstack_property.info( '%s {property_name} has been updated.' % instance.get_type_display_name(), event_type='saltstack_property_update_succeeded', event_context={ 'property': instance, }) def log_saltstack_property_deleted(sender, instance, **kwargs): event_logger.saltstack_property.info( '%s {property_name} has been deleted.' % instance.get_type_display_name(), event_type='saltstack_property_deletion_succeeded', event_context={ 'property': instance, }) Add more details to event logs for property CRUDfrom __future__ import unicode_literals import logging from .log import event_logger logger = logging.getLogger(__name__) def log_saltstack_property_created(sender, instance, created=False, **kwargs): if created: event_logger.saltstack_property.info( '%s {property_name} has been created in {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_creation_succeeded', event_context={ 'property': instance, }) else: event_logger.saltstack_property.info( '%s {property_name} has been updated in {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_update_succeeded', event_context={ 'property': instance, }) def log_saltstack_property_deleted(sender, instance, **kwargs): event_logger.saltstack_property.info( '%s {property_name} has been deleted from {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_deletion_succeeded', event_context={ 'property': instance, })
<commit_before>from __future__ import unicode_literals import logging from .log import event_logger logger = logging.getLogger(__name__) def log_saltstack_property_created(sender, instance, created=False, **kwargs): if created: event_logger.saltstack_property.info( '%s {property_name} has been created.' % instance.get_type_display_name(), event_type='saltstack_property_creation_succeeded', event_context={ 'property': instance, }) else: event_logger.saltstack_property.info( '%s {property_name} has been updated.' % instance.get_type_display_name(), event_type='saltstack_property_update_succeeded', event_context={ 'property': instance, }) def log_saltstack_property_deleted(sender, instance, **kwargs): event_logger.saltstack_property.info( '%s {property_name} has been deleted.' % instance.get_type_display_name(), event_type='saltstack_property_deletion_succeeded', event_context={ 'property': instance, }) <commit_msg>Add more details to event logs for property CRUD<commit_after>from __future__ import unicode_literals import logging from .log import event_logger logger = logging.getLogger(__name__) def log_saltstack_property_created(sender, instance, created=False, **kwargs): if created: event_logger.saltstack_property.info( '%s {property_name} has been created in {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_creation_succeeded', event_context={ 'property': instance, }) else: event_logger.saltstack_property.info( '%s {property_name} has been updated in {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_update_succeeded', event_context={ 'property': instance, }) def log_saltstack_property_deleted(sender, instance, **kwargs): event_logger.saltstack_property.info( '%s {property_name} has been deleted from {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_deletion_succeeded', event_context={ 'property': instance, })
18318b3bb431c8a5ec9261d6dd190997613cf1ed
src/pytest_django_casperjs/tests/test_fixtures.py
src/pytest_django_casperjs/tests/test_fixtures.py
from __future__ import with_statement import django import pytest from django.conf import settings as real_settings from django.utils.encoding import force_text from django.test.client import Client, RequestFactory from .app.models import Item from pytest_django_casperjs.compat import urlopen django # Avoid pyflakes complaints class TestCasperJSLiveServer: # Partially based on the LiveServer test case from pytest_django' pytestmark = [ pytest.mark.django_db() ] def test_url(self, casper_js): assert casper_js.url == force_text(casper_js) def test_db_changes_visibility(self, casper_js): response_data = urlopen(casper_js + '/item_count/').read() assert force_text(response_data) == 'Item count: 0' Item.objects.create(name='foo') response_data = urlopen(casper_js + '/item_count/').read() assert force_text(response_data) == 'Item count: 1' def test_fixture_db(self, db, casper_js): Item.objects.create(name='foo') response_data = urlopen(casper_js + '/item_count/').read() assert force_text(response_data) == 'Item count: 1' @pytest.fixture def item(self): # This has not requested database access so should fail. # Unfortunately the _live_server_helper autouse fixture makes this # test work. with pytest.raises(pytest.fail.Exception): Item.objects.create(name='foo') @pytest.mark.xfail def test_item(self, item, casper_js): # test should fail/pass in setup pass @pytest.fixture def item_db(self, db): return Item.objects.create(name='foo') def test_item_db(self, item_db, casper_js): response_data = urlopen(casper_js + '/item_count/').read() assert force_text(response_data) == 'Item count: 1'
from __future__ import with_statement import django import pytest from django.conf import settings as real_settings from django.utils.encoding import force_text from django.test.client import Client, RequestFactory from .app.models import Item from pytest_django_casperjs.compat import urlopen django # Avoid pyflakes complaints @pytest.mark.django_db class TestCasperJSLiveServer: # Partially based on the LiveServer test case from pytest_django' def test_url(self, casper_js): assert casper_js.url == force_text(casper_js)
Remove more irrelevant tests, those will be replaced with proper casperjs tests
Remove more irrelevant tests, those will be replaced with proper casperjs tests
Python
bsd-3-clause
EnTeQuAk/pytest-django-casperjs
from __future__ import with_statement import django import pytest from django.conf import settings as real_settings from django.utils.encoding import force_text from django.test.client import Client, RequestFactory from .app.models import Item from pytest_django_casperjs.compat import urlopen django # Avoid pyflakes complaints class TestCasperJSLiveServer: # Partially based on the LiveServer test case from pytest_django' pytestmark = [ pytest.mark.django_db() ] def test_url(self, casper_js): assert casper_js.url == force_text(casper_js) def test_db_changes_visibility(self, casper_js): response_data = urlopen(casper_js + '/item_count/').read() assert force_text(response_data) == 'Item count: 0' Item.objects.create(name='foo') response_data = urlopen(casper_js + '/item_count/').read() assert force_text(response_data) == 'Item count: 1' def test_fixture_db(self, db, casper_js): Item.objects.create(name='foo') response_data = urlopen(casper_js + '/item_count/').read() assert force_text(response_data) == 'Item count: 1' @pytest.fixture def item(self): # This has not requested database access so should fail. # Unfortunately the _live_server_helper autouse fixture makes this # test work. with pytest.raises(pytest.fail.Exception): Item.objects.create(name='foo') @pytest.mark.xfail def test_item(self, item, casper_js): # test should fail/pass in setup pass @pytest.fixture def item_db(self, db): return Item.objects.create(name='foo') def test_item_db(self, item_db, casper_js): response_data = urlopen(casper_js + '/item_count/').read() assert force_text(response_data) == 'Item count: 1' Remove more irrelevant tests, those will be replaced with proper casperjs tests
from __future__ import with_statement import django import pytest from django.conf import settings as real_settings from django.utils.encoding import force_text from django.test.client import Client, RequestFactory from .app.models import Item from pytest_django_casperjs.compat import urlopen django # Avoid pyflakes complaints @pytest.mark.django_db class TestCasperJSLiveServer: # Partially based on the LiveServer test case from pytest_django' def test_url(self, casper_js): assert casper_js.url == force_text(casper_js)
<commit_before>from __future__ import with_statement import django import pytest from django.conf import settings as real_settings from django.utils.encoding import force_text from django.test.client import Client, RequestFactory from .app.models import Item from pytest_django_casperjs.compat import urlopen django # Avoid pyflakes complaints class TestCasperJSLiveServer: # Partially based on the LiveServer test case from pytest_django' pytestmark = [ pytest.mark.django_db() ] def test_url(self, casper_js): assert casper_js.url == force_text(casper_js) def test_db_changes_visibility(self, casper_js): response_data = urlopen(casper_js + '/item_count/').read() assert force_text(response_data) == 'Item count: 0' Item.objects.create(name='foo') response_data = urlopen(casper_js + '/item_count/').read() assert force_text(response_data) == 'Item count: 1' def test_fixture_db(self, db, casper_js): Item.objects.create(name='foo') response_data = urlopen(casper_js + '/item_count/').read() assert force_text(response_data) == 'Item count: 1' @pytest.fixture def item(self): # This has not requested database access so should fail. # Unfortunately the _live_server_helper autouse fixture makes this # test work. with pytest.raises(pytest.fail.Exception): Item.objects.create(name='foo') @pytest.mark.xfail def test_item(self, item, casper_js): # test should fail/pass in setup pass @pytest.fixture def item_db(self, db): return Item.objects.create(name='foo') def test_item_db(self, item_db, casper_js): response_data = urlopen(casper_js + '/item_count/').read() assert force_text(response_data) == 'Item count: 1' <commit_msg>Remove more irrelevant tests, those will be replaced with proper casperjs tests<commit_after>
from __future__ import with_statement import django import pytest from django.conf import settings as real_settings from django.utils.encoding import force_text from django.test.client import Client, RequestFactory from .app.models import Item from pytest_django_casperjs.compat import urlopen django # Avoid pyflakes complaints @pytest.mark.django_db class TestCasperJSLiveServer: # Partially based on the LiveServer test case from pytest_django' def test_url(self, casper_js): assert casper_js.url == force_text(casper_js)
from __future__ import with_statement import django import pytest from django.conf import settings as real_settings from django.utils.encoding import force_text from django.test.client import Client, RequestFactory from .app.models import Item from pytest_django_casperjs.compat import urlopen django # Avoid pyflakes complaints class TestCasperJSLiveServer: # Partially based on the LiveServer test case from pytest_django' pytestmark = [ pytest.mark.django_db() ] def test_url(self, casper_js): assert casper_js.url == force_text(casper_js) def test_db_changes_visibility(self, casper_js): response_data = urlopen(casper_js + '/item_count/').read() assert force_text(response_data) == 'Item count: 0' Item.objects.create(name='foo') response_data = urlopen(casper_js + '/item_count/').read() assert force_text(response_data) == 'Item count: 1' def test_fixture_db(self, db, casper_js): Item.objects.create(name='foo') response_data = urlopen(casper_js + '/item_count/').read() assert force_text(response_data) == 'Item count: 1' @pytest.fixture def item(self): # This has not requested database access so should fail. # Unfortunately the _live_server_helper autouse fixture makes this # test work. with pytest.raises(pytest.fail.Exception): Item.objects.create(name='foo') @pytest.mark.xfail def test_item(self, item, casper_js): # test should fail/pass in setup pass @pytest.fixture def item_db(self, db): return Item.objects.create(name='foo') def test_item_db(self, item_db, casper_js): response_data = urlopen(casper_js + '/item_count/').read() assert force_text(response_data) == 'Item count: 1' Remove more irrelevant tests, those will be replaced with proper casperjs testsfrom __future__ import with_statement import django import pytest from django.conf import settings as real_settings from django.utils.encoding import force_text from django.test.client import Client, RequestFactory from .app.models import Item from pytest_django_casperjs.compat import urlopen django # Avoid pyflakes complaints @pytest.mark.django_db class TestCasperJSLiveServer: # Partially based on the LiveServer test case from pytest_django' def test_url(self, casper_js): assert casper_js.url == force_text(casper_js)
<commit_before>from __future__ import with_statement import django import pytest from django.conf import settings as real_settings from django.utils.encoding import force_text from django.test.client import Client, RequestFactory from .app.models import Item from pytest_django_casperjs.compat import urlopen django # Avoid pyflakes complaints class TestCasperJSLiveServer: # Partially based on the LiveServer test case from pytest_django' pytestmark = [ pytest.mark.django_db() ] def test_url(self, casper_js): assert casper_js.url == force_text(casper_js) def test_db_changes_visibility(self, casper_js): response_data = urlopen(casper_js + '/item_count/').read() assert force_text(response_data) == 'Item count: 0' Item.objects.create(name='foo') response_data = urlopen(casper_js + '/item_count/').read() assert force_text(response_data) == 'Item count: 1' def test_fixture_db(self, db, casper_js): Item.objects.create(name='foo') response_data = urlopen(casper_js + '/item_count/').read() assert force_text(response_data) == 'Item count: 1' @pytest.fixture def item(self): # This has not requested database access so should fail. # Unfortunately the _live_server_helper autouse fixture makes this # test work. with pytest.raises(pytest.fail.Exception): Item.objects.create(name='foo') @pytest.mark.xfail def test_item(self, item, casper_js): # test should fail/pass in setup pass @pytest.fixture def item_db(self, db): return Item.objects.create(name='foo') def test_item_db(self, item_db, casper_js): response_data = urlopen(casper_js + '/item_count/').read() assert force_text(response_data) == 'Item count: 1' <commit_msg>Remove more irrelevant tests, those will be replaced with proper casperjs tests<commit_after>from __future__ import with_statement import django import pytest from django.conf import settings as real_settings from django.utils.encoding import force_text from django.test.client import Client, RequestFactory from .app.models import Item from pytest_django_casperjs.compat import urlopen django # Avoid pyflakes complaints @pytest.mark.django_db class TestCasperJSLiveServer: # Partially based on the LiveServer test case from pytest_django' def test_url(self, casper_js): assert casper_js.url == force_text(casper_js)
8696885e9f1535bdfb8dbc0e285c67d1e6d41a95
datasets/admin.py
datasets/admin.py
from django.contrib import admin from datasets.models import Dataset, Sound, Annotation, Vote, Taxonomy, DatasetRelease, TaxonomyNode admin.site.register(Dataset) admin.site.register(Sound) admin.site.register(Annotation) admin.site.register(Vote) admin.site.register(Taxonomy) admin.site.register(DatasetRelease) admin.site.register(TaxonomyNode)
from django.contrib import admin from datasets.models import Dataset, Sound, Annotation, Vote, Taxonomy, DatasetRelease, TaxonomyNode class TaxonomyNodeAdmin(admin.ModelAdmin): fields = ('node_id', 'name', 'description', 'citation_uri', 'faq') admin.site.register(Dataset) admin.site.register(Sound) admin.site.register(Annotation) admin.site.register(Vote) admin.site.register(Taxonomy) admin.site.register(DatasetRelease) admin.site.register(TaxonomyNode, TaxonomyNodeAdmin)
Add custom Admin model TaxonomyNode, hide freesound ex
Add custom Admin model TaxonomyNode, hide freesound ex
Python
agpl-3.0
MTG/freesound-datasets,MTG/freesound-datasets,MTG/freesound-datasets,MTG/freesound-datasets
from django.contrib import admin from datasets.models import Dataset, Sound, Annotation, Vote, Taxonomy, DatasetRelease, TaxonomyNode admin.site.register(Dataset) admin.site.register(Sound) admin.site.register(Annotation) admin.site.register(Vote) admin.site.register(Taxonomy) admin.site.register(DatasetRelease) admin.site.register(TaxonomyNode) Add custom Admin model TaxonomyNode, hide freesound ex
from django.contrib import admin from datasets.models import Dataset, Sound, Annotation, Vote, Taxonomy, DatasetRelease, TaxonomyNode class TaxonomyNodeAdmin(admin.ModelAdmin): fields = ('node_id', 'name', 'description', 'citation_uri', 'faq') admin.site.register(Dataset) admin.site.register(Sound) admin.site.register(Annotation) admin.site.register(Vote) admin.site.register(Taxonomy) admin.site.register(DatasetRelease) admin.site.register(TaxonomyNode, TaxonomyNodeAdmin)
<commit_before>from django.contrib import admin from datasets.models import Dataset, Sound, Annotation, Vote, Taxonomy, DatasetRelease, TaxonomyNode admin.site.register(Dataset) admin.site.register(Sound) admin.site.register(Annotation) admin.site.register(Vote) admin.site.register(Taxonomy) admin.site.register(DatasetRelease) admin.site.register(TaxonomyNode) <commit_msg>Add custom Admin model TaxonomyNode, hide freesound ex<commit_after>
from django.contrib import admin from datasets.models import Dataset, Sound, Annotation, Vote, Taxonomy, DatasetRelease, TaxonomyNode class TaxonomyNodeAdmin(admin.ModelAdmin): fields = ('node_id', 'name', 'description', 'citation_uri', 'faq') admin.site.register(Dataset) admin.site.register(Sound) admin.site.register(Annotation) admin.site.register(Vote) admin.site.register(Taxonomy) admin.site.register(DatasetRelease) admin.site.register(TaxonomyNode, TaxonomyNodeAdmin)
from django.contrib import admin from datasets.models import Dataset, Sound, Annotation, Vote, Taxonomy, DatasetRelease, TaxonomyNode admin.site.register(Dataset) admin.site.register(Sound) admin.site.register(Annotation) admin.site.register(Vote) admin.site.register(Taxonomy) admin.site.register(DatasetRelease) admin.site.register(TaxonomyNode) Add custom Admin model TaxonomyNode, hide freesound exfrom django.contrib import admin from datasets.models import Dataset, Sound, Annotation, Vote, Taxonomy, DatasetRelease, TaxonomyNode class TaxonomyNodeAdmin(admin.ModelAdmin): fields = ('node_id', 'name', 'description', 'citation_uri', 'faq') admin.site.register(Dataset) admin.site.register(Sound) admin.site.register(Annotation) admin.site.register(Vote) admin.site.register(Taxonomy) admin.site.register(DatasetRelease) admin.site.register(TaxonomyNode, TaxonomyNodeAdmin)
<commit_before>from django.contrib import admin from datasets.models import Dataset, Sound, Annotation, Vote, Taxonomy, DatasetRelease, TaxonomyNode admin.site.register(Dataset) admin.site.register(Sound) admin.site.register(Annotation) admin.site.register(Vote) admin.site.register(Taxonomy) admin.site.register(DatasetRelease) admin.site.register(TaxonomyNode) <commit_msg>Add custom Admin model TaxonomyNode, hide freesound ex<commit_after>from django.contrib import admin from datasets.models import Dataset, Sound, Annotation, Vote, Taxonomy, DatasetRelease, TaxonomyNode class TaxonomyNodeAdmin(admin.ModelAdmin): fields = ('node_id', 'name', 'description', 'citation_uri', 'faq') admin.site.register(Dataset) admin.site.register(Sound) admin.site.register(Annotation) admin.site.register(Vote) admin.site.register(Taxonomy) admin.site.register(DatasetRelease) admin.site.register(TaxonomyNode, TaxonomyNodeAdmin)
a14256e715d51728ad4c2bde7ec52f13def6b2a6
director/views.py
director/views.py
from django.shortcuts import redirect from django.urls import reverse from django.views.generic import View class HomeView(View): def get(self, *args, **kwargs): if self.request.user.is_authenticated: return redirect(reverse('project_list')) else: return redirect(reverse('beta_token'))
from django.shortcuts import redirect from django.urls import reverse from accounts.views import BetaTokenView class HomeView(BetaTokenView): """ Home page view. Care needs to be taken that this view returns a 200 response (not a redirect) for unauthenticated users. This is because GCP load balancers ping the / path as a health check and will fail if anything other than a 200 is returned. """ def get(self, *args, **kwargs): if self.request.user.is_authenticated: return redirect(reverse('project_list')) else: return super().get(*args, **kwargs)
Fix home view so it returns 200 for unauthenticated health check
Fix home view so it returns 200 for unauthenticated health check
Python
apache-2.0
stencila/hub,stencila/hub,stencila/hub,stencila/hub,stencila/hub
from django.shortcuts import redirect from django.urls import reverse from django.views.generic import View class HomeView(View): def get(self, *args, **kwargs): if self.request.user.is_authenticated: return redirect(reverse('project_list')) else: return redirect(reverse('beta_token')) Fix home view so it returns 200 for unauthenticated health check
from django.shortcuts import redirect from django.urls import reverse from accounts.views import BetaTokenView class HomeView(BetaTokenView): """ Home page view. Care needs to be taken that this view returns a 200 response (not a redirect) for unauthenticated users. This is because GCP load balancers ping the / path as a health check and will fail if anything other than a 200 is returned. """ def get(self, *args, **kwargs): if self.request.user.is_authenticated: return redirect(reverse('project_list')) else: return super().get(*args, **kwargs)
<commit_before>from django.shortcuts import redirect from django.urls import reverse from django.views.generic import View class HomeView(View): def get(self, *args, **kwargs): if self.request.user.is_authenticated: return redirect(reverse('project_list')) else: return redirect(reverse('beta_token')) <commit_msg>Fix home view so it returns 200 for unauthenticated health check<commit_after>
from django.shortcuts import redirect from django.urls import reverse from accounts.views import BetaTokenView class HomeView(BetaTokenView): """ Home page view. Care needs to be taken that this view returns a 200 response (not a redirect) for unauthenticated users. This is because GCP load balancers ping the / path as a health check and will fail if anything other than a 200 is returned. """ def get(self, *args, **kwargs): if self.request.user.is_authenticated: return redirect(reverse('project_list')) else: return super().get(*args, **kwargs)
from django.shortcuts import redirect from django.urls import reverse from django.views.generic import View class HomeView(View): def get(self, *args, **kwargs): if self.request.user.is_authenticated: return redirect(reverse('project_list')) else: return redirect(reverse('beta_token')) Fix home view so it returns 200 for unauthenticated health checkfrom django.shortcuts import redirect from django.urls import reverse from accounts.views import BetaTokenView class HomeView(BetaTokenView): """ Home page view. Care needs to be taken that this view returns a 200 response (not a redirect) for unauthenticated users. This is because GCP load balancers ping the / path as a health check and will fail if anything other than a 200 is returned. """ def get(self, *args, **kwargs): if self.request.user.is_authenticated: return redirect(reverse('project_list')) else: return super().get(*args, **kwargs)
<commit_before>from django.shortcuts import redirect from django.urls import reverse from django.views.generic import View class HomeView(View): def get(self, *args, **kwargs): if self.request.user.is_authenticated: return redirect(reverse('project_list')) else: return redirect(reverse('beta_token')) <commit_msg>Fix home view so it returns 200 for unauthenticated health check<commit_after>from django.shortcuts import redirect from django.urls import reverse from accounts.views import BetaTokenView class HomeView(BetaTokenView): """ Home page view. Care needs to be taken that this view returns a 200 response (not a redirect) for unauthenticated users. This is because GCP load balancers ping the / path as a health check and will fail if anything other than a 200 is returned. """ def get(self, *args, **kwargs): if self.request.user.is_authenticated: return redirect(reverse('project_list')) else: return super().get(*args, **kwargs)
95ceea4ce45d531c277c00456639a42cfd18f129
djangae/patches/json.py
djangae/patches/json.py
from functools import wraps def additional_type_handler(func): @wraps(func) def _wrapper(self, o): if isinstance(o, set): # Return a string representing a set return "{" + ",".join([repr(x) for x in o]) + "}" else: return func(self, o) return _wrapper def patch(): """ This patches Djangos JSON encoder so it can deal with set(). This is necessary because otherwise we can't serialize SetFields """ from django.core.serializers.json import DjangoJSONEncoder DjangoJSONEncoder.default = additional_type_handler(DjangoJSONEncoder.default)
from functools import wraps def additional_type_handler(func): @wraps(func) def _wrapper(self, o): if isinstance(o, set): # Return a string representing a set return "{" + ",".join([str(x) for x in o]) + "}" else: return func(self, o) return _wrapper def patch(): """ This patches Djangos JSON encoder so it can deal with set(). This is necessary because otherwise we can't serialize SetFields """ from django.core.serializers.json import DjangoJSONEncoder DjangoJSONEncoder.default = additional_type_handler(DjangoJSONEncoder.default)
Use str() not repr() to avoid trailing L on longs
Use str() not repr() to avoid trailing L on longs
Python
bsd-3-clause
grzes/djangae,leekchan/djangae,SiPiggles/djangae,potatolondon/djangae,martinogden/djangae,potatolondon/djangae,jscissr/djangae,chargrizzle/djangae,trik/djangae,leekchan/djangae,kirberich/djangae,SiPiggles/djangae,chargrizzle/djangae,wangjun/djangae,jscissr/djangae,martinogden/djangae,pablorecio/djangae,jscissr/djangae,pablorecio/djangae,kirberich/djangae,asendecka/djangae,grzes/djangae,kirberich/djangae,armirusco/djangae,grzes/djangae,wangjun/djangae,armirusco/djangae,asendecka/djangae,trik/djangae,leekchan/djangae,asendecka/djangae,chargrizzle/djangae,martinogden/djangae,pablorecio/djangae,SiPiggles/djangae,trik/djangae,armirusco/djangae,wangjun/djangae
from functools import wraps def additional_type_handler(func): @wraps(func) def _wrapper(self, o): if isinstance(o, set): # Return a string representing a set return "{" + ",".join([repr(x) for x in o]) + "}" else: return func(self, o) return _wrapper def patch(): """ This patches Djangos JSON encoder so it can deal with set(). This is necessary because otherwise we can't serialize SetFields """ from django.core.serializers.json import DjangoJSONEncoder DjangoJSONEncoder.default = additional_type_handler(DjangoJSONEncoder.default) Use str() not repr() to avoid trailing L on longs
from functools import wraps def additional_type_handler(func): @wraps(func) def _wrapper(self, o): if isinstance(o, set): # Return a string representing a set return "{" + ",".join([str(x) for x in o]) + "}" else: return func(self, o) return _wrapper def patch(): """ This patches Djangos JSON encoder so it can deal with set(). This is necessary because otherwise we can't serialize SetFields """ from django.core.serializers.json import DjangoJSONEncoder DjangoJSONEncoder.default = additional_type_handler(DjangoJSONEncoder.default)
<commit_before>from functools import wraps def additional_type_handler(func): @wraps(func) def _wrapper(self, o): if isinstance(o, set): # Return a string representing a set return "{" + ",".join([repr(x) for x in o]) + "}" else: return func(self, o) return _wrapper def patch(): """ This patches Djangos JSON encoder so it can deal with set(). This is necessary because otherwise we can't serialize SetFields """ from django.core.serializers.json import DjangoJSONEncoder DjangoJSONEncoder.default = additional_type_handler(DjangoJSONEncoder.default) <commit_msg>Use str() not repr() to avoid trailing L on longs<commit_after>
from functools import wraps def additional_type_handler(func): @wraps(func) def _wrapper(self, o): if isinstance(o, set): # Return a string representing a set return "{" + ",".join([str(x) for x in o]) + "}" else: return func(self, o) return _wrapper def patch(): """ This patches Djangos JSON encoder so it can deal with set(). This is necessary because otherwise we can't serialize SetFields """ from django.core.serializers.json import DjangoJSONEncoder DjangoJSONEncoder.default = additional_type_handler(DjangoJSONEncoder.default)
from functools import wraps def additional_type_handler(func): @wraps(func) def _wrapper(self, o): if isinstance(o, set): # Return a string representing a set return "{" + ",".join([repr(x) for x in o]) + "}" else: return func(self, o) return _wrapper def patch(): """ This patches Djangos JSON encoder so it can deal with set(). This is necessary because otherwise we can't serialize SetFields """ from django.core.serializers.json import DjangoJSONEncoder DjangoJSONEncoder.default = additional_type_handler(DjangoJSONEncoder.default) Use str() not repr() to avoid trailing L on longsfrom functools import wraps def additional_type_handler(func): @wraps(func) def _wrapper(self, o): if isinstance(o, set): # Return a string representing a set return "{" + ",".join([str(x) for x in o]) + "}" else: return func(self, o) return _wrapper def patch(): """ This patches Djangos JSON encoder so it can deal with set(). This is necessary because otherwise we can't serialize SetFields """ from django.core.serializers.json import DjangoJSONEncoder DjangoJSONEncoder.default = additional_type_handler(DjangoJSONEncoder.default)
<commit_before>from functools import wraps def additional_type_handler(func): @wraps(func) def _wrapper(self, o): if isinstance(o, set): # Return a string representing a set return "{" + ",".join([repr(x) for x in o]) + "}" else: return func(self, o) return _wrapper def patch(): """ This patches Djangos JSON encoder so it can deal with set(). This is necessary because otherwise we can't serialize SetFields """ from django.core.serializers.json import DjangoJSONEncoder DjangoJSONEncoder.default = additional_type_handler(DjangoJSONEncoder.default) <commit_msg>Use str() not repr() to avoid trailing L on longs<commit_after>from functools import wraps def additional_type_handler(func): @wraps(func) def _wrapper(self, o): if isinstance(o, set): # Return a string representing a set return "{" + ",".join([str(x) for x in o]) + "}" else: return func(self, o) return _wrapper def patch(): """ This patches Djangos JSON encoder so it can deal with set(). This is necessary because otherwise we can't serialize SetFields """ from django.core.serializers.json import DjangoJSONEncoder DjangoJSONEncoder.default = additional_type_handler(DjangoJSONEncoder.default)
7bd19241e0502789bed482291554e8341034d377
bpmodule/testing/__init__.py
bpmodule/testing/__init__.py
from .modinfo import * # SO file from .testing import * # For output from bpmodule import output def PrintHeader(s): output.Output(output.Line("=")) output.Output("%1%\n", s) output.Output(output.Line("=")) def PrintResults(nfailed): output.Output("\n\n") if nfailed > 0: output.Output(output.Line("=")) output.Error("RESULTS: %1% failed\n", nfailed) output.Output(output.Line("=")) else: output.Output(output.Line("=")) output.Success("RESULTS: %1% failed\n", nfailed) output.Output(output.Line("=")) output.Output("\n\n")
from .modinfo import * # SO file from .testing import * # For output from bpmodule.output import Output, Error, Warning, Success, Debug, Line ################## # For testing on the python side ################## def PyTestFunc(itest, desc, expected, func, *args): fmt = "%|1$5| : %|2$-5| %|3$-5| %|4$-9| : %5%\n" try: func(*args) except Exception as e: if expected == False: Output(fmt, itest, "False", "False", "Success", desc) return 0 else: Debug(str(e) + "\n") Error(fmt, itest, "True", "False", "FAILED", desc) return 1 except: if expected == False: Output(fmt, itest, "False", "False", "Success", desc) return 0 else: Error(fmt, itest, "True", "False", "FAILED", desc) return 1 if expected == True: Output(fmt, itest, "True", "True", "Success", desc) return 0 else: Error(fmt, itest, "False", "True", "FAILED", desc) return 1 def PrintHeader(s): Output(Line("=")) Output("%1%\n", s) Output(Line("=")) def PrintResults(nfailed): Output("\n\n") if nfailed > 0: Output(Line("=")) Error("RESULTS: %1% failed\n", nfailed) Output(Line("=")) else: Output(Line("=")) Success("RESULTS: %1% failed\n", nfailed) Output(Line("=")) Output("\n\n")
Add testing function from python
Add testing function from python
Python
bsd-3-clause
pulsar-chem/Pulsar-Core,pulsar-chem/Pulsar-Core,pulsar-chem/Pulsar-Core,pulsar-chem/Pulsar-Core
from .modinfo import * # SO file from .testing import * # For output from bpmodule import output def PrintHeader(s): output.Output(output.Line("=")) output.Output("%1%\n", s) output.Output(output.Line("=")) def PrintResults(nfailed): output.Output("\n\n") if nfailed > 0: output.Output(output.Line("=")) output.Error("RESULTS: %1% failed\n", nfailed) output.Output(output.Line("=")) else: output.Output(output.Line("=")) output.Success("RESULTS: %1% failed\n", nfailed) output.Output(output.Line("=")) output.Output("\n\n") Add testing function from python
from .modinfo import * # SO file from .testing import * # For output from bpmodule.output import Output, Error, Warning, Success, Debug, Line ################## # For testing on the python side ################## def PyTestFunc(itest, desc, expected, func, *args): fmt = "%|1$5| : %|2$-5| %|3$-5| %|4$-9| : %5%\n" try: func(*args) except Exception as e: if expected == False: Output(fmt, itest, "False", "False", "Success", desc) return 0 else: Debug(str(e) + "\n") Error(fmt, itest, "True", "False", "FAILED", desc) return 1 except: if expected == False: Output(fmt, itest, "False", "False", "Success", desc) return 0 else: Error(fmt, itest, "True", "False", "FAILED", desc) return 1 if expected == True: Output(fmt, itest, "True", "True", "Success", desc) return 0 else: Error(fmt, itest, "False", "True", "FAILED", desc) return 1 def PrintHeader(s): Output(Line("=")) Output("%1%\n", s) Output(Line("=")) def PrintResults(nfailed): Output("\n\n") if nfailed > 0: Output(Line("=")) Error("RESULTS: %1% failed\n", nfailed) Output(Line("=")) else: Output(Line("=")) Success("RESULTS: %1% failed\n", nfailed) Output(Line("=")) Output("\n\n")
<commit_before>from .modinfo import * # SO file from .testing import * # For output from bpmodule import output def PrintHeader(s): output.Output(output.Line("=")) output.Output("%1%\n", s) output.Output(output.Line("=")) def PrintResults(nfailed): output.Output("\n\n") if nfailed > 0: output.Output(output.Line("=")) output.Error("RESULTS: %1% failed\n", nfailed) output.Output(output.Line("=")) else: output.Output(output.Line("=")) output.Success("RESULTS: %1% failed\n", nfailed) output.Output(output.Line("=")) output.Output("\n\n") <commit_msg>Add testing function from python<commit_after>
from .modinfo import * # SO file from .testing import * # For output from bpmodule.output import Output, Error, Warning, Success, Debug, Line ################## # For testing on the python side ################## def PyTestFunc(itest, desc, expected, func, *args): fmt = "%|1$5| : %|2$-5| %|3$-5| %|4$-9| : %5%\n" try: func(*args) except Exception as e: if expected == False: Output(fmt, itest, "False", "False", "Success", desc) return 0 else: Debug(str(e) + "\n") Error(fmt, itest, "True", "False", "FAILED", desc) return 1 except: if expected == False: Output(fmt, itest, "False", "False", "Success", desc) return 0 else: Error(fmt, itest, "True", "False", "FAILED", desc) return 1 if expected == True: Output(fmt, itest, "True", "True", "Success", desc) return 0 else: Error(fmt, itest, "False", "True", "FAILED", desc) return 1 def PrintHeader(s): Output(Line("=")) Output("%1%\n", s) Output(Line("=")) def PrintResults(nfailed): Output("\n\n") if nfailed > 0: Output(Line("=")) Error("RESULTS: %1% failed\n", nfailed) Output(Line("=")) else: Output(Line("=")) Success("RESULTS: %1% failed\n", nfailed) Output(Line("=")) Output("\n\n")
from .modinfo import * # SO file from .testing import * # For output from bpmodule import output def PrintHeader(s): output.Output(output.Line("=")) output.Output("%1%\n", s) output.Output(output.Line("=")) def PrintResults(nfailed): output.Output("\n\n") if nfailed > 0: output.Output(output.Line("=")) output.Error("RESULTS: %1% failed\n", nfailed) output.Output(output.Line("=")) else: output.Output(output.Line("=")) output.Success("RESULTS: %1% failed\n", nfailed) output.Output(output.Line("=")) output.Output("\n\n") Add testing function from pythonfrom .modinfo import * # SO file from .testing import * # For output from bpmodule.output import Output, Error, Warning, Success, Debug, Line ################## # For testing on the python side ################## def PyTestFunc(itest, desc, expected, func, *args): fmt = "%|1$5| : %|2$-5| %|3$-5| %|4$-9| : %5%\n" try: func(*args) except Exception as e: if expected == False: Output(fmt, itest, "False", "False", "Success", desc) return 0 else: Debug(str(e) + "\n") Error(fmt, itest, "True", "False", "FAILED", desc) return 1 except: if expected == False: Output(fmt, itest, "False", "False", "Success", desc) return 0 else: Error(fmt, itest, "True", "False", "FAILED", desc) return 1 if expected == True: Output(fmt, itest, "True", "True", "Success", desc) return 0 else: Error(fmt, itest, "False", "True", "FAILED", desc) return 1 def PrintHeader(s): Output(Line("=")) Output("%1%\n", s) Output(Line("=")) def PrintResults(nfailed): Output("\n\n") if nfailed > 0: Output(Line("=")) Error("RESULTS: %1% failed\n", nfailed) Output(Line("=")) else: Output(Line("=")) Success("RESULTS: %1% failed\n", nfailed) Output(Line("=")) Output("\n\n")
<commit_before>from .modinfo import * # SO file from .testing import * # For output from bpmodule import output def PrintHeader(s): output.Output(output.Line("=")) output.Output("%1%\n", s) output.Output(output.Line("=")) def PrintResults(nfailed): output.Output("\n\n") if nfailed > 0: output.Output(output.Line("=")) output.Error("RESULTS: %1% failed\n", nfailed) output.Output(output.Line("=")) else: output.Output(output.Line("=")) output.Success("RESULTS: %1% failed\n", nfailed) output.Output(output.Line("=")) output.Output("\n\n") <commit_msg>Add testing function from python<commit_after>from .modinfo import * # SO file from .testing import * # For output from bpmodule.output import Output, Error, Warning, Success, Debug, Line ################## # For testing on the python side ################## def PyTestFunc(itest, desc, expected, func, *args): fmt = "%|1$5| : %|2$-5| %|3$-5| %|4$-9| : %5%\n" try: func(*args) except Exception as e: if expected == False: Output(fmt, itest, "False", "False", "Success", desc) return 0 else: Debug(str(e) + "\n") Error(fmt, itest, "True", "False", "FAILED", desc) return 1 except: if expected == False: Output(fmt, itest, "False", "False", "Success", desc) return 0 else: Error(fmt, itest, "True", "False", "FAILED", desc) return 1 if expected == True: Output(fmt, itest, "True", "True", "Success", desc) return 0 else: Error(fmt, itest, "False", "True", "FAILED", desc) return 1 def PrintHeader(s): Output(Line("=")) Output("%1%\n", s) Output(Line("=")) def PrintResults(nfailed): Output("\n\n") if nfailed > 0: Output(Line("=")) Error("RESULTS: %1% failed\n", nfailed) Output(Line("=")) else: Output(Line("=")) Success("RESULTS: %1% failed\n", nfailed) Output(Line("=")) Output("\n\n")
a4b475120fd58f135695e071424a3fa1024ae649
lib/__init__.py
lib/__init__.py
"""Package providing an implementation of the TOPKAPI model and some utilities. The interface isn't stable yet so be prepared to update your code on a regular basis... """ __author__ = 'Theo Vischel' __version__ = '0.1' import model
"""Package providing an implementation of the TOPKAPI model and some utilities. The interface isn't stable yet so be prepared to update your code on a regular basis... """ __author__ = 'Theo Vischel' __version__ = '0.2' import model
Update version number to 0.2.
Update version number to 0.2.
Python
bsd-3-clause
sahg/PyTOPKAPI,scottza/PyTOPKAPI
"""Package providing an implementation of the TOPKAPI model and some utilities. The interface isn't stable yet so be prepared to update your code on a regular basis... """ __author__ = 'Theo Vischel' __version__ = '0.1' import model Update version number to 0.2.
"""Package providing an implementation of the TOPKAPI model and some utilities. The interface isn't stable yet so be prepared to update your code on a regular basis... """ __author__ = 'Theo Vischel' __version__ = '0.2' import model
<commit_before>"""Package providing an implementation of the TOPKAPI model and some utilities. The interface isn't stable yet so be prepared to update your code on a regular basis... """ __author__ = 'Theo Vischel' __version__ = '0.1' import model <commit_msg>Update version number to 0.2.<commit_after>
"""Package providing an implementation of the TOPKAPI model and some utilities. The interface isn't stable yet so be prepared to update your code on a regular basis... """ __author__ = 'Theo Vischel' __version__ = '0.2' import model
"""Package providing an implementation of the TOPKAPI model and some utilities. The interface isn't stable yet so be prepared to update your code on a regular basis... """ __author__ = 'Theo Vischel' __version__ = '0.1' import model Update version number to 0.2."""Package providing an implementation of the TOPKAPI model and some utilities. The interface isn't stable yet so be prepared to update your code on a regular basis... """ __author__ = 'Theo Vischel' __version__ = '0.2' import model
<commit_before>"""Package providing an implementation of the TOPKAPI model and some utilities. The interface isn't stable yet so be prepared to update your code on a regular basis... """ __author__ = 'Theo Vischel' __version__ = '0.1' import model <commit_msg>Update version number to 0.2.<commit_after>"""Package providing an implementation of the TOPKAPI model and some utilities. The interface isn't stable yet so be prepared to update your code on a regular basis... """ __author__ = 'Theo Vischel' __version__ = '0.2' import model
bb88b1d2e2c4d3eb482c3cf32d1a53c9e89f94cf
conftest.py
conftest.py
# -*- coding:utf-8 -*- from __future__ import unicode_literals from django.db import connection def pytest_report_header(config): with connection.cursor() as cursor: cursor.execute("SELECT VERSION()") version = cursor.fetchone()[0] return "MySQL version: {}".format(version)
# -*- coding:utf-8 -*- from __future__ import unicode_literals import django from django.db import connection def pytest_report_header(config): dot_version = '.'.join(str(x) for x in django.VERSION) header = "Django version: " + dot_version if hasattr(connection, '_nodb_connection'): with connection._nodb_connection.cursor() as cursor: cursor.execute("SELECT VERSION()") version = cursor.fetchone()[0] header += "\nMySQL version: {}".format(version) return header
Fix pytest version report when database does not exist, add Django version header
Fix pytest version report when database does not exist, add Django version header
Python
mit
nickmeharry/django-mysql,arnau126/django-mysql,arnau126/django-mysql,nickmeharry/django-mysql,adamchainz/django-mysql
# -*- coding:utf-8 -*- from __future__ import unicode_literals from django.db import connection def pytest_report_header(config): with connection.cursor() as cursor: cursor.execute("SELECT VERSION()") version = cursor.fetchone()[0] return "MySQL version: {}".format(version) Fix pytest version report when database does not exist, add Django version header
# -*- coding:utf-8 -*- from __future__ import unicode_literals import django from django.db import connection def pytest_report_header(config): dot_version = '.'.join(str(x) for x in django.VERSION) header = "Django version: " + dot_version if hasattr(connection, '_nodb_connection'): with connection._nodb_connection.cursor() as cursor: cursor.execute("SELECT VERSION()") version = cursor.fetchone()[0] header += "\nMySQL version: {}".format(version) return header
<commit_before># -*- coding:utf-8 -*- from __future__ import unicode_literals from django.db import connection def pytest_report_header(config): with connection.cursor() as cursor: cursor.execute("SELECT VERSION()") version = cursor.fetchone()[0] return "MySQL version: {}".format(version) <commit_msg>Fix pytest version report when database does not exist, add Django version header<commit_after>
# -*- coding:utf-8 -*- from __future__ import unicode_literals import django from django.db import connection def pytest_report_header(config): dot_version = '.'.join(str(x) for x in django.VERSION) header = "Django version: " + dot_version if hasattr(connection, '_nodb_connection'): with connection._nodb_connection.cursor() as cursor: cursor.execute("SELECT VERSION()") version = cursor.fetchone()[0] header += "\nMySQL version: {}".format(version) return header
# -*- coding:utf-8 -*- from __future__ import unicode_literals from django.db import connection def pytest_report_header(config): with connection.cursor() as cursor: cursor.execute("SELECT VERSION()") version = cursor.fetchone()[0] return "MySQL version: {}".format(version) Fix pytest version report when database does not exist, add Django version header# -*- coding:utf-8 -*- from __future__ import unicode_literals import django from django.db import connection def pytest_report_header(config): dot_version = '.'.join(str(x) for x in django.VERSION) header = "Django version: " + dot_version if hasattr(connection, '_nodb_connection'): with connection._nodb_connection.cursor() as cursor: cursor.execute("SELECT VERSION()") version = cursor.fetchone()[0] header += "\nMySQL version: {}".format(version) return header
<commit_before># -*- coding:utf-8 -*- from __future__ import unicode_literals from django.db import connection def pytest_report_header(config): with connection.cursor() as cursor: cursor.execute("SELECT VERSION()") version = cursor.fetchone()[0] return "MySQL version: {}".format(version) <commit_msg>Fix pytest version report when database does not exist, add Django version header<commit_after># -*- coding:utf-8 -*- from __future__ import unicode_literals import django from django.db import connection def pytest_report_header(config): dot_version = '.'.join(str(x) for x in django.VERSION) header = "Django version: " + dot_version if hasattr(connection, '_nodb_connection'): with connection._nodb_connection.cursor() as cursor: cursor.execute("SELECT VERSION()") version = cursor.fetchone()[0] header += "\nMySQL version: {}".format(version) return header
22a852a9ad0521496e8b0be52b37d111c3402bb4
conftest.py
conftest.py
import pytest from models.broadcasted_models import two_comp_model from utilities.phoenix_utils import load_starfish_spectrum @pytest.fixture def host(): """Host spectrum fixture.""" mod_spec = load_starfish_spectrum([5200, 4.50, 0.0], limits=[2110, 2170], normalize=True) return mod_spec @pytest.fixture def comp(): """Noramlized Companion spectrum fixture.""" mod_spec = load_starfish_spectrum([2600, 4.50, 0.0], limits=[2110, 2170], normalize=True) return mod_spec @pytest.fixture(params=["scalar", "linear", "quadratic", "exponential"]) def norm_method(request): return request.param @pytest.fixture() def tcm_model(host, comp): return two_comp_model(host.xaxis, host.flux, comp.xaxis, alphas=[0.1, 0.2, 0.3], rvs=[-0.25, 0.25], gammas=[1, 2, 3, 4])
import pytest from models.broadcasted_models import two_comp_model from utilities.phoenix_utils import load_starfish_spectrum @pytest.fixture def host(): """Host spectrum fixture.""" mod_spec = load_starfish_spectrum([5200, 4.50, 0.0], limits=[2110, 2165], normalize=True) return mod_spec @pytest.fixture def comp(): """Noramlized Companion spectrum fixture.""" mod_spec = load_starfish_spectrum([2600, 4.50, 0.0], limits=[2110, 2165], normalize=True) return mod_spec @pytest.fixture(params=["scalar", "linear", "quadratic", "exponential"]) def norm_method(request): return request.param @pytest.fixture() def tcm_model(host, comp): return two_comp_model(host.xaxis, host.flux, comp.xaxis, alphas=[0.1, 0.2, 0.3], rvs=[-0.25, 0.25], gammas=[1, 2, 3, 4])
Revert "tweak host fixture limits"
Revert "tweak host fixture limits" This reverts commit 06e9a964dec8392007e3af87d1a41bbe119158ca.
Python
mit
jason-neal/companion_simulations,jason-neal/companion_simulations
import pytest from models.broadcasted_models import two_comp_model from utilities.phoenix_utils import load_starfish_spectrum @pytest.fixture def host(): """Host spectrum fixture.""" mod_spec = load_starfish_spectrum([5200, 4.50, 0.0], limits=[2110, 2170], normalize=True) return mod_spec @pytest.fixture def comp(): """Noramlized Companion spectrum fixture.""" mod_spec = load_starfish_spectrum([2600, 4.50, 0.0], limits=[2110, 2170], normalize=True) return mod_spec @pytest.fixture(params=["scalar", "linear", "quadratic", "exponential"]) def norm_method(request): return request.param @pytest.fixture() def tcm_model(host, comp): return two_comp_model(host.xaxis, host.flux, comp.xaxis, alphas=[0.1, 0.2, 0.3], rvs=[-0.25, 0.25], gammas=[1, 2, 3, 4]) Revert "tweak host fixture limits" This reverts commit 06e9a964dec8392007e3af87d1a41bbe119158ca.
import pytest from models.broadcasted_models import two_comp_model from utilities.phoenix_utils import load_starfish_spectrum @pytest.fixture def host(): """Host spectrum fixture.""" mod_spec = load_starfish_spectrum([5200, 4.50, 0.0], limits=[2110, 2165], normalize=True) return mod_spec @pytest.fixture def comp(): """Noramlized Companion spectrum fixture.""" mod_spec = load_starfish_spectrum([2600, 4.50, 0.0], limits=[2110, 2165], normalize=True) return mod_spec @pytest.fixture(params=["scalar", "linear", "quadratic", "exponential"]) def norm_method(request): return request.param @pytest.fixture() def tcm_model(host, comp): return two_comp_model(host.xaxis, host.flux, comp.xaxis, alphas=[0.1, 0.2, 0.3], rvs=[-0.25, 0.25], gammas=[1, 2, 3, 4])
<commit_before>import pytest from models.broadcasted_models import two_comp_model from utilities.phoenix_utils import load_starfish_spectrum @pytest.fixture def host(): """Host spectrum fixture.""" mod_spec = load_starfish_spectrum([5200, 4.50, 0.0], limits=[2110, 2170], normalize=True) return mod_spec @pytest.fixture def comp(): """Noramlized Companion spectrum fixture.""" mod_spec = load_starfish_spectrum([2600, 4.50, 0.0], limits=[2110, 2170], normalize=True) return mod_spec @pytest.fixture(params=["scalar", "linear", "quadratic", "exponential"]) def norm_method(request): return request.param @pytest.fixture() def tcm_model(host, comp): return two_comp_model(host.xaxis, host.flux, comp.xaxis, alphas=[0.1, 0.2, 0.3], rvs=[-0.25, 0.25], gammas=[1, 2, 3, 4]) <commit_msg>Revert "tweak host fixture limits" This reverts commit 06e9a964dec8392007e3af87d1a41bbe119158ca.<commit_after>
import pytest from models.broadcasted_models import two_comp_model from utilities.phoenix_utils import load_starfish_spectrum @pytest.fixture def host(): """Host spectrum fixture.""" mod_spec = load_starfish_spectrum([5200, 4.50, 0.0], limits=[2110, 2165], normalize=True) return mod_spec @pytest.fixture def comp(): """Noramlized Companion spectrum fixture.""" mod_spec = load_starfish_spectrum([2600, 4.50, 0.0], limits=[2110, 2165], normalize=True) return mod_spec @pytest.fixture(params=["scalar", "linear", "quadratic", "exponential"]) def norm_method(request): return request.param @pytest.fixture() def tcm_model(host, comp): return two_comp_model(host.xaxis, host.flux, comp.xaxis, alphas=[0.1, 0.2, 0.3], rvs=[-0.25, 0.25], gammas=[1, 2, 3, 4])
import pytest from models.broadcasted_models import two_comp_model from utilities.phoenix_utils import load_starfish_spectrum @pytest.fixture def host(): """Host spectrum fixture.""" mod_spec = load_starfish_spectrum([5200, 4.50, 0.0], limits=[2110, 2170], normalize=True) return mod_spec @pytest.fixture def comp(): """Noramlized Companion spectrum fixture.""" mod_spec = load_starfish_spectrum([2600, 4.50, 0.0], limits=[2110, 2170], normalize=True) return mod_spec @pytest.fixture(params=["scalar", "linear", "quadratic", "exponential"]) def norm_method(request): return request.param @pytest.fixture() def tcm_model(host, comp): return two_comp_model(host.xaxis, host.flux, comp.xaxis, alphas=[0.1, 0.2, 0.3], rvs=[-0.25, 0.25], gammas=[1, 2, 3, 4]) Revert "tweak host fixture limits" This reverts commit 06e9a964dec8392007e3af87d1a41bbe119158ca.import pytest from models.broadcasted_models import two_comp_model from utilities.phoenix_utils import load_starfish_spectrum @pytest.fixture def host(): """Host spectrum fixture.""" mod_spec = load_starfish_spectrum([5200, 4.50, 0.0], limits=[2110, 2165], normalize=True) return mod_spec @pytest.fixture def comp(): """Noramlized Companion spectrum fixture.""" mod_spec = load_starfish_spectrum([2600, 4.50, 0.0], limits=[2110, 2165], normalize=True) return mod_spec @pytest.fixture(params=["scalar", "linear", "quadratic", "exponential"]) def norm_method(request): return request.param @pytest.fixture() def tcm_model(host, comp): return two_comp_model(host.xaxis, host.flux, comp.xaxis, alphas=[0.1, 0.2, 0.3], rvs=[-0.25, 0.25], gammas=[1, 2, 3, 4])
<commit_before>import pytest from models.broadcasted_models import two_comp_model from utilities.phoenix_utils import load_starfish_spectrum @pytest.fixture def host(): """Host spectrum fixture.""" mod_spec = load_starfish_spectrum([5200, 4.50, 0.0], limits=[2110, 2170], normalize=True) return mod_spec @pytest.fixture def comp(): """Noramlized Companion spectrum fixture.""" mod_spec = load_starfish_spectrum([2600, 4.50, 0.0], limits=[2110, 2170], normalize=True) return mod_spec @pytest.fixture(params=["scalar", "linear", "quadratic", "exponential"]) def norm_method(request): return request.param @pytest.fixture() def tcm_model(host, comp): return two_comp_model(host.xaxis, host.flux, comp.xaxis, alphas=[0.1, 0.2, 0.3], rvs=[-0.25, 0.25], gammas=[1, 2, 3, 4]) <commit_msg>Revert "tweak host fixture limits" This reverts commit 06e9a964dec8392007e3af87d1a41bbe119158ca.<commit_after>import pytest from models.broadcasted_models import two_comp_model from utilities.phoenix_utils import load_starfish_spectrum @pytest.fixture def host(): """Host spectrum fixture.""" mod_spec = load_starfish_spectrum([5200, 4.50, 0.0], limits=[2110, 2165], normalize=True) return mod_spec @pytest.fixture def comp(): """Noramlized Companion spectrum fixture.""" mod_spec = load_starfish_spectrum([2600, 4.50, 0.0], limits=[2110, 2165], normalize=True) return mod_spec @pytest.fixture(params=["scalar", "linear", "quadratic", "exponential"]) def norm_method(request): return request.param @pytest.fixture() def tcm_model(host, comp): return two_comp_model(host.xaxis, host.flux, comp.xaxis, alphas=[0.1, 0.2, 0.3], rvs=[-0.25, 0.25], gammas=[1, 2, 3, 4])
aaf7cb7ecc1a74fb2b222fd21aea7116dac2ca98
contribs.py
contribs.py
# Get Contribution Count import urllib import datetime import HTMLParser class ContribParser(HTMLParser.HTMLParser): today = datetime.date.today().isoformat() def handle_starttag(self, tag, attrs): if tag == 'rect' and self.is_today(attrs): self.count = self.get_count(attrs) def is_today(self, attrs): for name, value in attrs: if name == 'data-date' and value == self.today: return True return False def get_count(self, attrs): for name, value in attrs: if name == 'data-count': return value return None def getContribs(username): url = 'https://github.com/users/:user/contributions' req = urllib.urlopen(url.replace(':user', username)) parser = ContribParser() parser.feed(req.read()) return parser.count
# Get Contribution Count import urllib import datetime import HTMLParser class ContribParser(HTMLParser.HTMLParser): def __init__(self): self.today = datetime.date.today().isoformat() HTMLParser.HTMLParser.__init__(self) def handle_starttag(self, tag, attrs): if tag == 'rect' and self.is_today(attrs): self.count = self.get_count(attrs) def is_today(self, attrs): for name, value in attrs: if name == 'data-date' and value == self.today: return True return False def get_count(self, attrs): for name, value in attrs: if name == 'data-count': return value return None def getContribs(username): url = 'https://github.com/users/:user/contributions' req = urllib.urlopen(url.replace(':user', username)) parser = ContribParser() parser.feed(req.read()) return parser.count
Fix date issue (I think)
Fix date issue (I think)
Python
mit
chrisfosterelli/commitwatch
# Get Contribution Count import urllib import datetime import HTMLParser class ContribParser(HTMLParser.HTMLParser): today = datetime.date.today().isoformat() def handle_starttag(self, tag, attrs): if tag == 'rect' and self.is_today(attrs): self.count = self.get_count(attrs) def is_today(self, attrs): for name, value in attrs: if name == 'data-date' and value == self.today: return True return False def get_count(self, attrs): for name, value in attrs: if name == 'data-count': return value return None def getContribs(username): url = 'https://github.com/users/:user/contributions' req = urllib.urlopen(url.replace(':user', username)) parser = ContribParser() parser.feed(req.read()) return parser.count Fix date issue (I think)
# Get Contribution Count import urllib import datetime import HTMLParser class ContribParser(HTMLParser.HTMLParser): def __init__(self): self.today = datetime.date.today().isoformat() HTMLParser.HTMLParser.__init__(self) def handle_starttag(self, tag, attrs): if tag == 'rect' and self.is_today(attrs): self.count = self.get_count(attrs) def is_today(self, attrs): for name, value in attrs: if name == 'data-date' and value == self.today: return True return False def get_count(self, attrs): for name, value in attrs: if name == 'data-count': return value return None def getContribs(username): url = 'https://github.com/users/:user/contributions' req = urllib.urlopen(url.replace(':user', username)) parser = ContribParser() parser.feed(req.read()) return parser.count
<commit_before> # Get Contribution Count import urllib import datetime import HTMLParser class ContribParser(HTMLParser.HTMLParser): today = datetime.date.today().isoformat() def handle_starttag(self, tag, attrs): if tag == 'rect' and self.is_today(attrs): self.count = self.get_count(attrs) def is_today(self, attrs): for name, value in attrs: if name == 'data-date' and value == self.today: return True return False def get_count(self, attrs): for name, value in attrs: if name == 'data-count': return value return None def getContribs(username): url = 'https://github.com/users/:user/contributions' req = urllib.urlopen(url.replace(':user', username)) parser = ContribParser() parser.feed(req.read()) return parser.count <commit_msg>Fix date issue (I think)<commit_after>
# Get Contribution Count import urllib import datetime import HTMLParser class ContribParser(HTMLParser.HTMLParser): def __init__(self): self.today = datetime.date.today().isoformat() HTMLParser.HTMLParser.__init__(self) def handle_starttag(self, tag, attrs): if tag == 'rect' and self.is_today(attrs): self.count = self.get_count(attrs) def is_today(self, attrs): for name, value in attrs: if name == 'data-date' and value == self.today: return True return False def get_count(self, attrs): for name, value in attrs: if name == 'data-count': return value return None def getContribs(username): url = 'https://github.com/users/:user/contributions' req = urllib.urlopen(url.replace(':user', username)) parser = ContribParser() parser.feed(req.read()) return parser.count
# Get Contribution Count import urllib import datetime import HTMLParser class ContribParser(HTMLParser.HTMLParser): today = datetime.date.today().isoformat() def handle_starttag(self, tag, attrs): if tag == 'rect' and self.is_today(attrs): self.count = self.get_count(attrs) def is_today(self, attrs): for name, value in attrs: if name == 'data-date' and value == self.today: return True return False def get_count(self, attrs): for name, value in attrs: if name == 'data-count': return value return None def getContribs(username): url = 'https://github.com/users/:user/contributions' req = urllib.urlopen(url.replace(':user', username)) parser = ContribParser() parser.feed(req.read()) return parser.count Fix date issue (I think) # Get Contribution Count import urllib import datetime import HTMLParser class ContribParser(HTMLParser.HTMLParser): def __init__(self): self.today = datetime.date.today().isoformat() HTMLParser.HTMLParser.__init__(self) def handle_starttag(self, tag, attrs): if tag == 'rect' and self.is_today(attrs): self.count = self.get_count(attrs) def is_today(self, attrs): for name, value in attrs: if name == 'data-date' and value == self.today: return True return False def get_count(self, attrs): for name, value in attrs: if name == 'data-count': return value return None def getContribs(username): url = 'https://github.com/users/:user/contributions' req = urllib.urlopen(url.replace(':user', username)) parser = ContribParser() parser.feed(req.read()) return parser.count
<commit_before> # Get Contribution Count import urllib import datetime import HTMLParser class ContribParser(HTMLParser.HTMLParser): today = datetime.date.today().isoformat() def handle_starttag(self, tag, attrs): if tag == 'rect' and self.is_today(attrs): self.count = self.get_count(attrs) def is_today(self, attrs): for name, value in attrs: if name == 'data-date' and value == self.today: return True return False def get_count(self, attrs): for name, value in attrs: if name == 'data-count': return value return None def getContribs(username): url = 'https://github.com/users/:user/contributions' req = urllib.urlopen(url.replace(':user', username)) parser = ContribParser() parser.feed(req.read()) return parser.count <commit_msg>Fix date issue (I think)<commit_after> # Get Contribution Count import urllib import datetime import HTMLParser class ContribParser(HTMLParser.HTMLParser): def __init__(self): self.today = datetime.date.today().isoformat() HTMLParser.HTMLParser.__init__(self) def handle_starttag(self, tag, attrs): if tag == 'rect' and self.is_today(attrs): self.count = self.get_count(attrs) def is_today(self, attrs): for name, value in attrs: if name == 'data-date' and value == self.today: return True return False def get_count(self, attrs): for name, value in attrs: if name == 'data-count': return value return None def getContribs(username): url = 'https://github.com/users/:user/contributions' req = urllib.urlopen(url.replace(':user', username)) parser = ContribParser() parser.feed(req.read()) return parser.count
22be6bb3593f948893ab3f797d34e20e66fff841
example.py
example.py
import discord import asyncio client = discord.Client() @client.event async def on_ready(): print('Logged in as') print(client.user.name) print(client.user.id) print('------') @client.event async def on_message(message): if message.content.startswith('!test'): counter = 0 tmp = await client.send_message(message.channel, 'Calculating messages...') async for log in client.logs_from(message.channel, limit=100): if log.author == message.author: counter += 1 await client.edit_message(tmp, 'You have {} messages.'.format(counter)) elif message.content.startswith('!sleep'): await asyncio.sleep(5) await client.send_message(message.channel, 'Done sleeping') client.run('token')
import discord import asyncio import os #Set up Client State CLIENT_TOKEN=os.environ['TOKEN'] client = discord.Client() @client.event async def on_ready(): print('Logged in as') print(client.user.name) print(client.user.id) print('------') @client.event async def on_message(message): if message.content.startswith('!test'): counter = 0 tmp = await client.send_message(message.channel, 'Calculating messages...') async for log in client.logs_from(message.channel, limit=100): if log.author == message.author: counter += 1 await client.edit_message(tmp, 'You have {} messages.'.format(counter)) elif message.content.startswith('!sleep'): await asyncio.sleep(5) await client.send_message(message.channel, 'Done sleeping') client.run(CLIENT_TOKEN)
Use env value for client token
Use env value for client token
Python
mit
gryffon/SusumuTakuan,gryffon/SusumuTakuan
import discord import asyncio client = discord.Client() @client.event async def on_ready(): print('Logged in as') print(client.user.name) print(client.user.id) print('------') @client.event async def on_message(message): if message.content.startswith('!test'): counter = 0 tmp = await client.send_message(message.channel, 'Calculating messages...') async for log in client.logs_from(message.channel, limit=100): if log.author == message.author: counter += 1 await client.edit_message(tmp, 'You have {} messages.'.format(counter)) elif message.content.startswith('!sleep'): await asyncio.sleep(5) await client.send_message(message.channel, 'Done sleeping') client.run('token')Use env value for client token
import discord import asyncio import os #Set up Client State CLIENT_TOKEN=os.environ['TOKEN'] client = discord.Client() @client.event async def on_ready(): print('Logged in as') print(client.user.name) print(client.user.id) print('------') @client.event async def on_message(message): if message.content.startswith('!test'): counter = 0 tmp = await client.send_message(message.channel, 'Calculating messages...') async for log in client.logs_from(message.channel, limit=100): if log.author == message.author: counter += 1 await client.edit_message(tmp, 'You have {} messages.'.format(counter)) elif message.content.startswith('!sleep'): await asyncio.sleep(5) await client.send_message(message.channel, 'Done sleeping') client.run(CLIENT_TOKEN)
<commit_before>import discord import asyncio client = discord.Client() @client.event async def on_ready(): print('Logged in as') print(client.user.name) print(client.user.id) print('------') @client.event async def on_message(message): if message.content.startswith('!test'): counter = 0 tmp = await client.send_message(message.channel, 'Calculating messages...') async for log in client.logs_from(message.channel, limit=100): if log.author == message.author: counter += 1 await client.edit_message(tmp, 'You have {} messages.'.format(counter)) elif message.content.startswith('!sleep'): await asyncio.sleep(5) await client.send_message(message.channel, 'Done sleeping') client.run('token')<commit_msg>Use env value for client token<commit_after>
import discord import asyncio import os #Set up Client State CLIENT_TOKEN=os.environ['TOKEN'] client = discord.Client() @client.event async def on_ready(): print('Logged in as') print(client.user.name) print(client.user.id) print('------') @client.event async def on_message(message): if message.content.startswith('!test'): counter = 0 tmp = await client.send_message(message.channel, 'Calculating messages...') async for log in client.logs_from(message.channel, limit=100): if log.author == message.author: counter += 1 await client.edit_message(tmp, 'You have {} messages.'.format(counter)) elif message.content.startswith('!sleep'): await asyncio.sleep(5) await client.send_message(message.channel, 'Done sleeping') client.run(CLIENT_TOKEN)
import discord import asyncio client = discord.Client() @client.event async def on_ready(): print('Logged in as') print(client.user.name) print(client.user.id) print('------') @client.event async def on_message(message): if message.content.startswith('!test'): counter = 0 tmp = await client.send_message(message.channel, 'Calculating messages...') async for log in client.logs_from(message.channel, limit=100): if log.author == message.author: counter += 1 await client.edit_message(tmp, 'You have {} messages.'.format(counter)) elif message.content.startswith('!sleep'): await asyncio.sleep(5) await client.send_message(message.channel, 'Done sleeping') client.run('token')Use env value for client tokenimport discord import asyncio import os #Set up Client State CLIENT_TOKEN=os.environ['TOKEN'] client = discord.Client() @client.event async def on_ready(): print('Logged in as') print(client.user.name) print(client.user.id) print('------') @client.event async def on_message(message): if message.content.startswith('!test'): counter = 0 tmp = await client.send_message(message.channel, 'Calculating messages...') async for log in client.logs_from(message.channel, limit=100): if log.author == message.author: counter += 1 await client.edit_message(tmp, 'You have {} messages.'.format(counter)) elif message.content.startswith('!sleep'): await asyncio.sleep(5) await client.send_message(message.channel, 'Done sleeping') client.run(CLIENT_TOKEN)
<commit_before>import discord import asyncio client = discord.Client() @client.event async def on_ready(): print('Logged in as') print(client.user.name) print(client.user.id) print('------') @client.event async def on_message(message): if message.content.startswith('!test'): counter = 0 tmp = await client.send_message(message.channel, 'Calculating messages...') async for log in client.logs_from(message.channel, limit=100): if log.author == message.author: counter += 1 await client.edit_message(tmp, 'You have {} messages.'.format(counter)) elif message.content.startswith('!sleep'): await asyncio.sleep(5) await client.send_message(message.channel, 'Done sleeping') client.run('token')<commit_msg>Use env value for client token<commit_after>import discord import asyncio import os #Set up Client State CLIENT_TOKEN=os.environ['TOKEN'] client = discord.Client() @client.event async def on_ready(): print('Logged in as') print(client.user.name) print(client.user.id) print('------') @client.event async def on_message(message): if message.content.startswith('!test'): counter = 0 tmp = await client.send_message(message.channel, 'Calculating messages...') async for log in client.logs_from(message.channel, limit=100): if log.author == message.author: counter += 1 await client.edit_message(tmp, 'You have {} messages.'.format(counter)) elif message.content.startswith('!sleep'): await asyncio.sleep(5) await client.send_message(message.channel, 'Done sleeping') client.run(CLIENT_TOKEN)
e3d8b836681a0cb4795d317c7a23defd6004c967
pytest_run.py
pytest_run.py
# coding=utf-8 """This is a script for running pytest from the command line. This script exists so that the project directory gets added to sys.path, which prevents us from accidentally testing the globally installed willie version. pytest_run.py Copyright 2013, Ari Koivula, <ari@koivu.la> Licensed under the Eiffel Forum License 2. http://willie.dfbta.net """ if __name__ == "__main__": import sys import pytest returncode = pytest.main() sys.exit(returncode)
#!/usr/bin/env python # coding=utf-8 """This is a script for running pytest from the command line. This script exists so that the project directory gets added to sys.path, which prevents us from accidentally testing the globally installed willie version. pytest_run.py Copyright 2013, Ari Koivula, <ari@koivu.la> Licensed under the Eiffel Forum License 2. http://willie.dfbta.net """ if __name__ == "__main__": import sys import pytest returncode = pytest.main() sys.exit(returncode)
Add shebang to testing script
Add shebang to testing script
Python
mit
Uname-a/knife_scraper,Uname-a/knife_scraper,Uname-a/knife_scraper
# coding=utf-8 """This is a script for running pytest from the command line. This script exists so that the project directory gets added to sys.path, which prevents us from accidentally testing the globally installed willie version. pytest_run.py Copyright 2013, Ari Koivula, <ari@koivu.la> Licensed under the Eiffel Forum License 2. http://willie.dfbta.net """ if __name__ == "__main__": import sys import pytest returncode = pytest.main() sys.exit(returncode) Add shebang to testing script
#!/usr/bin/env python # coding=utf-8 """This is a script for running pytest from the command line. This script exists so that the project directory gets added to sys.path, which prevents us from accidentally testing the globally installed willie version. pytest_run.py Copyright 2013, Ari Koivula, <ari@koivu.la> Licensed under the Eiffel Forum License 2. http://willie.dfbta.net """ if __name__ == "__main__": import sys import pytest returncode = pytest.main() sys.exit(returncode)
<commit_before># coding=utf-8 """This is a script for running pytest from the command line. This script exists so that the project directory gets added to sys.path, which prevents us from accidentally testing the globally installed willie version. pytest_run.py Copyright 2013, Ari Koivula, <ari@koivu.la> Licensed under the Eiffel Forum License 2. http://willie.dfbta.net """ if __name__ == "__main__": import sys import pytest returncode = pytest.main() sys.exit(returncode) <commit_msg>Add shebang to testing script<commit_after>
#!/usr/bin/env python # coding=utf-8 """This is a script for running pytest from the command line. This script exists so that the project directory gets added to sys.path, which prevents us from accidentally testing the globally installed willie version. pytest_run.py Copyright 2013, Ari Koivula, <ari@koivu.la> Licensed under the Eiffel Forum License 2. http://willie.dfbta.net """ if __name__ == "__main__": import sys import pytest returncode = pytest.main() sys.exit(returncode)
# coding=utf-8 """This is a script for running pytest from the command line. This script exists so that the project directory gets added to sys.path, which prevents us from accidentally testing the globally installed willie version. pytest_run.py Copyright 2013, Ari Koivula, <ari@koivu.la> Licensed under the Eiffel Forum License 2. http://willie.dfbta.net """ if __name__ == "__main__": import sys import pytest returncode = pytest.main() sys.exit(returncode) Add shebang to testing script#!/usr/bin/env python # coding=utf-8 """This is a script for running pytest from the command line. This script exists so that the project directory gets added to sys.path, which prevents us from accidentally testing the globally installed willie version. pytest_run.py Copyright 2013, Ari Koivula, <ari@koivu.la> Licensed under the Eiffel Forum License 2. http://willie.dfbta.net """ if __name__ == "__main__": import sys import pytest returncode = pytest.main() sys.exit(returncode)
<commit_before># coding=utf-8 """This is a script for running pytest from the command line. This script exists so that the project directory gets added to sys.path, which prevents us from accidentally testing the globally installed willie version. pytest_run.py Copyright 2013, Ari Koivula, <ari@koivu.la> Licensed under the Eiffel Forum License 2. http://willie.dfbta.net """ if __name__ == "__main__": import sys import pytest returncode = pytest.main() sys.exit(returncode) <commit_msg>Add shebang to testing script<commit_after>#!/usr/bin/env python # coding=utf-8 """This is a script for running pytest from the command line. This script exists so that the project directory gets added to sys.path, which prevents us from accidentally testing the globally installed willie version. pytest_run.py Copyright 2013, Ari Koivula, <ari@koivu.la> Licensed under the Eiffel Forum License 2. http://willie.dfbta.net """ if __name__ == "__main__": import sys import pytest returncode = pytest.main() sys.exit(returncode)
d5a5e46b2fbc9284213aef3ec45f0605b002b7b1
axes/management/commands/axes_reset.py
axes/management/commands/axes_reset.py
from django.core.management.base import BaseCommand from axes.utils import reset class Command(BaseCommand): help = ("resets any lockouts or failed login records. If called with an " "IP, resets only for that IP") def add_arguments(self, parser): parser.add_argument('ip', nargs='*') def handle(self, *args, **kwargs): count = 0 if kwargs: for ip in kwargs['ip']: count += reset(ip=ip) else: count = reset() if count: print('{0} attempts removed.'.format(count)) else: print('No attempts found.')
from django.core.management.base import BaseCommand from axes.utils import reset class Command(BaseCommand): help = ("resets any lockouts or failed login records. If called with an " "IP, resets only for that IP") def add_arguments(self, parser): parser.add_argument('ip', nargs='*') def handle(self, *args, **kwargs): count = 0 if kwargs and kwargs.get('ip'): for ip in kwargs['ip']: count += reset(ip=ip) else: count = reset() if count: print('{0} attempts removed.'.format(count)) else: print('No attempts found.')
Reset all attempts when ip not specified
Reset all attempts when ip not specified When no ip address positional arguments are specified, reset all attempts, as with reset() and per documentation.
Python
mit
svenhertle/django-axes,django-pci/django-axes,jazzband/django-axes
from django.core.management.base import BaseCommand from axes.utils import reset class Command(BaseCommand): help = ("resets any lockouts or failed login records. If called with an " "IP, resets only for that IP") def add_arguments(self, parser): parser.add_argument('ip', nargs='*') def handle(self, *args, **kwargs): count = 0 if kwargs: for ip in kwargs['ip']: count += reset(ip=ip) else: count = reset() if count: print('{0} attempts removed.'.format(count)) else: print('No attempts found.') Reset all attempts when ip not specified When no ip address positional arguments are specified, reset all attempts, as with reset() and per documentation.
from django.core.management.base import BaseCommand from axes.utils import reset class Command(BaseCommand): help = ("resets any lockouts or failed login records. If called with an " "IP, resets only for that IP") def add_arguments(self, parser): parser.add_argument('ip', nargs='*') def handle(self, *args, **kwargs): count = 0 if kwargs and kwargs.get('ip'): for ip in kwargs['ip']: count += reset(ip=ip) else: count = reset() if count: print('{0} attempts removed.'.format(count)) else: print('No attempts found.')
<commit_before>from django.core.management.base import BaseCommand from axes.utils import reset class Command(BaseCommand): help = ("resets any lockouts or failed login records. If called with an " "IP, resets only for that IP") def add_arguments(self, parser): parser.add_argument('ip', nargs='*') def handle(self, *args, **kwargs): count = 0 if kwargs: for ip in kwargs['ip']: count += reset(ip=ip) else: count = reset() if count: print('{0} attempts removed.'.format(count)) else: print('No attempts found.') <commit_msg>Reset all attempts when ip not specified When no ip address positional arguments are specified, reset all attempts, as with reset() and per documentation.<commit_after>
from django.core.management.base import BaseCommand from axes.utils import reset class Command(BaseCommand): help = ("resets any lockouts or failed login records. If called with an " "IP, resets only for that IP") def add_arguments(self, parser): parser.add_argument('ip', nargs='*') def handle(self, *args, **kwargs): count = 0 if kwargs and kwargs.get('ip'): for ip in kwargs['ip']: count += reset(ip=ip) else: count = reset() if count: print('{0} attempts removed.'.format(count)) else: print('No attempts found.')
from django.core.management.base import BaseCommand from axes.utils import reset class Command(BaseCommand): help = ("resets any lockouts or failed login records. If called with an " "IP, resets only for that IP") def add_arguments(self, parser): parser.add_argument('ip', nargs='*') def handle(self, *args, **kwargs): count = 0 if kwargs: for ip in kwargs['ip']: count += reset(ip=ip) else: count = reset() if count: print('{0} attempts removed.'.format(count)) else: print('No attempts found.') Reset all attempts when ip not specified When no ip address positional arguments are specified, reset all attempts, as with reset() and per documentation.from django.core.management.base import BaseCommand from axes.utils import reset class Command(BaseCommand): help = ("resets any lockouts or failed login records. If called with an " "IP, resets only for that IP") def add_arguments(self, parser): parser.add_argument('ip', nargs='*') def handle(self, *args, **kwargs): count = 0 if kwargs and kwargs.get('ip'): for ip in kwargs['ip']: count += reset(ip=ip) else: count = reset() if count: print('{0} attempts removed.'.format(count)) else: print('No attempts found.')
<commit_before>from django.core.management.base import BaseCommand from axes.utils import reset class Command(BaseCommand): help = ("resets any lockouts or failed login records. If called with an " "IP, resets only for that IP") def add_arguments(self, parser): parser.add_argument('ip', nargs='*') def handle(self, *args, **kwargs): count = 0 if kwargs: for ip in kwargs['ip']: count += reset(ip=ip) else: count = reset() if count: print('{0} attempts removed.'.format(count)) else: print('No attempts found.') <commit_msg>Reset all attempts when ip not specified When no ip address positional arguments are specified, reset all attempts, as with reset() and per documentation.<commit_after>from django.core.management.base import BaseCommand from axes.utils import reset class Command(BaseCommand): help = ("resets any lockouts or failed login records. If called with an " "IP, resets only for that IP") def add_arguments(self, parser): parser.add_argument('ip', nargs='*') def handle(self, *args, **kwargs): count = 0 if kwargs and kwargs.get('ip'): for ip in kwargs['ip']: count += reset(ip=ip) else: count = reset() if count: print('{0} attempts removed.'.format(count)) else: print('No attempts found.')
132932747a1f7da67413b9c0cf7916707c1e3d19
src/python/services/CVMFSAppVersions.py
src/python/services/CVMFSAppVersions.py
"""CVMFS Servcice.""" import os import re import cherrypy import html from natsort import natsorted version_re = re.compile(r"^release-(\d{1,3}\.\d{1,3}\.\d{1,3})$") @cherrypy.popargs('appid') class CVMFSAppVersions(object): """ CVMFS App Version checking service. CVMFS Service to get the list of versions available on CVMFS for a given app. """ def __init__(self, cvmfs_root, valid_apps): """Initialise.""" self.cvmfs_root = cvmfs_root self.valid_apps = valid_apps @cherrypy.expose def index(self, appid=None): """Return the index page.""" print "IN CVMFSAppVersion: appid=(%s)" % appid if appid not in self.valid_apps: print "Invalid app type %s" % appid return '' html_ = html.HTML() _, dirs, _ = os.walk(os.path.join(self.cvmfs_root, appid)).next() for dir_ in natsorted(dirs): for version in version_re.findall(dir_): html_.option(version) return str(html_)
"""CVMFS Servcice.""" import os import re import cherrypy import html from natsort import natsorted VERSION_RE = re.compile(r"^release-(\d{1,3}\.\d{1,3}\.\d{1,3})$") @cherrypy.popargs('appid') class CVMFSAppVersions(object): """ CVMFS App Version checking service. CVMFS Service to get the list of versions available on CVMFS for a given app. """ def __init__(self, cvmfs_root, valid_apps): """Initialise.""" self.cvmfs_root = cvmfs_root self.valid_apps = valid_apps @cherrypy.expose def index(self, appid=None): """Return the index page.""" print "IN CVMFSAppVersion: appid=(%s)" % appid if appid not in self.valid_apps: print "Invalid app type %s" % appid return '' html_ = html.HTML() _, dirs, _ = os.walk(os.path.join(self.cvmfs_root, appid)).next() for dir_ in natsorted(dirs): for version in VERSION_RE.findall(dir_): html_.option(version) return str(html_)
Change the re const name to uppercase.
Change the re const name to uppercase.
Python
mit
alexanderrichards/LZProduction,alexanderrichards/LZProduction,alexanderrichards/LZProduction,alexanderrichards/LZProduction
"""CVMFS Servcice.""" import os import re import cherrypy import html from natsort import natsorted version_re = re.compile(r"^release-(\d{1,3}\.\d{1,3}\.\d{1,3})$") @cherrypy.popargs('appid') class CVMFSAppVersions(object): """ CVMFS App Version checking service. CVMFS Service to get the list of versions available on CVMFS for a given app. """ def __init__(self, cvmfs_root, valid_apps): """Initialise.""" self.cvmfs_root = cvmfs_root self.valid_apps = valid_apps @cherrypy.expose def index(self, appid=None): """Return the index page.""" print "IN CVMFSAppVersion: appid=(%s)" % appid if appid not in self.valid_apps: print "Invalid app type %s" % appid return '' html_ = html.HTML() _, dirs, _ = os.walk(os.path.join(self.cvmfs_root, appid)).next() for dir_ in natsorted(dirs): for version in version_re.findall(dir_): html_.option(version) return str(html_) Change the re const name to uppercase.
"""CVMFS Servcice.""" import os import re import cherrypy import html from natsort import natsorted VERSION_RE = re.compile(r"^release-(\d{1,3}\.\d{1,3}\.\d{1,3})$") @cherrypy.popargs('appid') class CVMFSAppVersions(object): """ CVMFS App Version checking service. CVMFS Service to get the list of versions available on CVMFS for a given app. """ def __init__(self, cvmfs_root, valid_apps): """Initialise.""" self.cvmfs_root = cvmfs_root self.valid_apps = valid_apps @cherrypy.expose def index(self, appid=None): """Return the index page.""" print "IN CVMFSAppVersion: appid=(%s)" % appid if appid not in self.valid_apps: print "Invalid app type %s" % appid return '' html_ = html.HTML() _, dirs, _ = os.walk(os.path.join(self.cvmfs_root, appid)).next() for dir_ in natsorted(dirs): for version in VERSION_RE.findall(dir_): html_.option(version) return str(html_)
<commit_before>"""CVMFS Servcice.""" import os import re import cherrypy import html from natsort import natsorted version_re = re.compile(r"^release-(\d{1,3}\.\d{1,3}\.\d{1,3})$") @cherrypy.popargs('appid') class CVMFSAppVersions(object): """ CVMFS App Version checking service. CVMFS Service to get the list of versions available on CVMFS for a given app. """ def __init__(self, cvmfs_root, valid_apps): """Initialise.""" self.cvmfs_root = cvmfs_root self.valid_apps = valid_apps @cherrypy.expose def index(self, appid=None): """Return the index page.""" print "IN CVMFSAppVersion: appid=(%s)" % appid if appid not in self.valid_apps: print "Invalid app type %s" % appid return '' html_ = html.HTML() _, dirs, _ = os.walk(os.path.join(self.cvmfs_root, appid)).next() for dir_ in natsorted(dirs): for version in version_re.findall(dir_): html_.option(version) return str(html_) <commit_msg>Change the re const name to uppercase.<commit_after>
"""CVMFS Servcice.""" import os import re import cherrypy import html from natsort import natsorted VERSION_RE = re.compile(r"^release-(\d{1,3}\.\d{1,3}\.\d{1,3})$") @cherrypy.popargs('appid') class CVMFSAppVersions(object): """ CVMFS App Version checking service. CVMFS Service to get the list of versions available on CVMFS for a given app. """ def __init__(self, cvmfs_root, valid_apps): """Initialise.""" self.cvmfs_root = cvmfs_root self.valid_apps = valid_apps @cherrypy.expose def index(self, appid=None): """Return the index page.""" print "IN CVMFSAppVersion: appid=(%s)" % appid if appid not in self.valid_apps: print "Invalid app type %s" % appid return '' html_ = html.HTML() _, dirs, _ = os.walk(os.path.join(self.cvmfs_root, appid)).next() for dir_ in natsorted(dirs): for version in VERSION_RE.findall(dir_): html_.option(version) return str(html_)
"""CVMFS Servcice.""" import os import re import cherrypy import html from natsort import natsorted version_re = re.compile(r"^release-(\d{1,3}\.\d{1,3}\.\d{1,3})$") @cherrypy.popargs('appid') class CVMFSAppVersions(object): """ CVMFS App Version checking service. CVMFS Service to get the list of versions available on CVMFS for a given app. """ def __init__(self, cvmfs_root, valid_apps): """Initialise.""" self.cvmfs_root = cvmfs_root self.valid_apps = valid_apps @cherrypy.expose def index(self, appid=None): """Return the index page.""" print "IN CVMFSAppVersion: appid=(%s)" % appid if appid not in self.valid_apps: print "Invalid app type %s" % appid return '' html_ = html.HTML() _, dirs, _ = os.walk(os.path.join(self.cvmfs_root, appid)).next() for dir_ in natsorted(dirs): for version in version_re.findall(dir_): html_.option(version) return str(html_) Change the re const name to uppercase."""CVMFS Servcice.""" import os import re import cherrypy import html from natsort import natsorted VERSION_RE = re.compile(r"^release-(\d{1,3}\.\d{1,3}\.\d{1,3})$") @cherrypy.popargs('appid') class CVMFSAppVersions(object): """ CVMFS App Version checking service. CVMFS Service to get the list of versions available on CVMFS for a given app. """ def __init__(self, cvmfs_root, valid_apps): """Initialise.""" self.cvmfs_root = cvmfs_root self.valid_apps = valid_apps @cherrypy.expose def index(self, appid=None): """Return the index page.""" print "IN CVMFSAppVersion: appid=(%s)" % appid if appid not in self.valid_apps: print "Invalid app type %s" % appid return '' html_ = html.HTML() _, dirs, _ = os.walk(os.path.join(self.cvmfs_root, appid)).next() for dir_ in natsorted(dirs): for version in VERSION_RE.findall(dir_): html_.option(version) return str(html_)
<commit_before>"""CVMFS Servcice.""" import os import re import cherrypy import html from natsort import natsorted version_re = re.compile(r"^release-(\d{1,3}\.\d{1,3}\.\d{1,3})$") @cherrypy.popargs('appid') class CVMFSAppVersions(object): """ CVMFS App Version checking service. CVMFS Service to get the list of versions available on CVMFS for a given app. """ def __init__(self, cvmfs_root, valid_apps): """Initialise.""" self.cvmfs_root = cvmfs_root self.valid_apps = valid_apps @cherrypy.expose def index(self, appid=None): """Return the index page.""" print "IN CVMFSAppVersion: appid=(%s)" % appid if appid not in self.valid_apps: print "Invalid app type %s" % appid return '' html_ = html.HTML() _, dirs, _ = os.walk(os.path.join(self.cvmfs_root, appid)).next() for dir_ in natsorted(dirs): for version in version_re.findall(dir_): html_.option(version) return str(html_) <commit_msg>Change the re const name to uppercase.<commit_after>"""CVMFS Servcice.""" import os import re import cherrypy import html from natsort import natsorted VERSION_RE = re.compile(r"^release-(\d{1,3}\.\d{1,3}\.\d{1,3})$") @cherrypy.popargs('appid') class CVMFSAppVersions(object): """ CVMFS App Version checking service. CVMFS Service to get the list of versions available on CVMFS for a given app. """ def __init__(self, cvmfs_root, valid_apps): """Initialise.""" self.cvmfs_root = cvmfs_root self.valid_apps = valid_apps @cherrypy.expose def index(self, appid=None): """Return the index page.""" print "IN CVMFSAppVersion: appid=(%s)" % appid if appid not in self.valid_apps: print "Invalid app type %s" % appid return '' html_ = html.HTML() _, dirs, _ = os.walk(os.path.join(self.cvmfs_root, appid)).next() for dir_ in natsorted(dirs): for version in VERSION_RE.findall(dir_): html_.option(version) return str(html_)
b26047600202a9776c99323813cf17b0aa951dcd
app/routes.py
app/routes.py
from flask import jsonify from . import app import mapper import utils from predict import predictor @app.route("/", methods=["GET"]) def index(): firebase_dump = mapper.get_dump_firebase() response = firebase_dump.get_all() response = response or {} return jsonify(response) @app.route("/build", methods=["POST"]) def build_model(): predictor.preprocess_airports() if not predictor.model: predictor.build_model() return jsonify({"message:" : "OK"}) @app.route("/predict", methods=["GET"]) def predict_all_delays(): results = None try: results = predictor.predict_all() except Exception as e: return jsonify({"message" : e.message}) return jsonify(results) @app.route("/predict/<airport_code>", methods=["GET"]) def predict_delay(airport_code): firebase_source = mapper.get_source_firebase() airport_status = firebase_source.get_airport(airport_code) cleaned_data = utils.get_clean_data(airport_status) res = predictor.predict(airport_code) cleaned_data["prediction"] = bool(res[0]) return jsonify(cleaned_data) @app.route("/status", methods=["GET"]) def get_airport_statuses(): firebase_source = mapper.get_source_firebase() airports = firebase_source.get_all() for airport_code, status in airports.items(): if "status" in status: del status["status"] return jsonify(airports)
from flask import jsonify from . import app import mapper import utils from predict import predictor @app.route("/", methods=["GET"]) def index(): return app.send_static_file("index.html") @app.route("/build", methods=["POST"]) def build_model(): predictor.preprocess_airports() if not predictor.model: predictor.build_model() return jsonify({"message:" : "OK"}) @app.route("/predict", methods=["GET"]) def predict_all_delays(): results = None try: results = predictor.predict_all() except Exception as e: print "ERROR", e.message return jsonify({"message" : e.message}) return jsonify(results) @app.route("/predict/<airport_code>", methods=["GET"]) def predict_delay(airport_code): firebase_source = mapper.get_source_firebase() airport_status = firebase_source.get_airport(airport_code) cleaned_data = utils.get_clean_data(airport_status) res = predictor.predict(airport_code) cleaned_data["prediction"] = bool(res[0]) return jsonify(cleaned_data) @app.route("/status", methods=["GET"]) def get_airport_statuses(): firebase_source = mapper.get_source_firebase() airports = firebase_source.get_all() results = [] for airport_code, status in airports.items(): try: results.append(utils.get_clean_data(status)) except: pass results = {"items":results} return jsonify(results)
Return index.html in root and transform /status results
Return index.html in root and transform /status results
Python
mit
MachineLearningProject/flight-delay-prediction,MachineLearningProject/flight-delay-prediction,MachineLearningProject/flight-delay-prediction
from flask import jsonify from . import app import mapper import utils from predict import predictor @app.route("/", methods=["GET"]) def index(): firebase_dump = mapper.get_dump_firebase() response = firebase_dump.get_all() response = response or {} return jsonify(response) @app.route("/build", methods=["POST"]) def build_model(): predictor.preprocess_airports() if not predictor.model: predictor.build_model() return jsonify({"message:" : "OK"}) @app.route("/predict", methods=["GET"]) def predict_all_delays(): results = None try: results = predictor.predict_all() except Exception as e: return jsonify({"message" : e.message}) return jsonify(results) @app.route("/predict/<airport_code>", methods=["GET"]) def predict_delay(airport_code): firebase_source = mapper.get_source_firebase() airport_status = firebase_source.get_airport(airport_code) cleaned_data = utils.get_clean_data(airport_status) res = predictor.predict(airport_code) cleaned_data["prediction"] = bool(res[0]) return jsonify(cleaned_data) @app.route("/status", methods=["GET"]) def get_airport_statuses(): firebase_source = mapper.get_source_firebase() airports = firebase_source.get_all() for airport_code, status in airports.items(): if "status" in status: del status["status"] return jsonify(airports) Return index.html in root and transform /status results
from flask import jsonify from . import app import mapper import utils from predict import predictor @app.route("/", methods=["GET"]) def index(): return app.send_static_file("index.html") @app.route("/build", methods=["POST"]) def build_model(): predictor.preprocess_airports() if not predictor.model: predictor.build_model() return jsonify({"message:" : "OK"}) @app.route("/predict", methods=["GET"]) def predict_all_delays(): results = None try: results = predictor.predict_all() except Exception as e: print "ERROR", e.message return jsonify({"message" : e.message}) return jsonify(results) @app.route("/predict/<airport_code>", methods=["GET"]) def predict_delay(airport_code): firebase_source = mapper.get_source_firebase() airport_status = firebase_source.get_airport(airport_code) cleaned_data = utils.get_clean_data(airport_status) res = predictor.predict(airport_code) cleaned_data["prediction"] = bool(res[0]) return jsonify(cleaned_data) @app.route("/status", methods=["GET"]) def get_airport_statuses(): firebase_source = mapper.get_source_firebase() airports = firebase_source.get_all() results = [] for airport_code, status in airports.items(): try: results.append(utils.get_clean_data(status)) except: pass results = {"items":results} return jsonify(results)
<commit_before>from flask import jsonify from . import app import mapper import utils from predict import predictor @app.route("/", methods=["GET"]) def index(): firebase_dump = mapper.get_dump_firebase() response = firebase_dump.get_all() response = response or {} return jsonify(response) @app.route("/build", methods=["POST"]) def build_model(): predictor.preprocess_airports() if not predictor.model: predictor.build_model() return jsonify({"message:" : "OK"}) @app.route("/predict", methods=["GET"]) def predict_all_delays(): results = None try: results = predictor.predict_all() except Exception as e: return jsonify({"message" : e.message}) return jsonify(results) @app.route("/predict/<airport_code>", methods=["GET"]) def predict_delay(airport_code): firebase_source = mapper.get_source_firebase() airport_status = firebase_source.get_airport(airport_code) cleaned_data = utils.get_clean_data(airport_status) res = predictor.predict(airport_code) cleaned_data["prediction"] = bool(res[0]) return jsonify(cleaned_data) @app.route("/status", methods=["GET"]) def get_airport_statuses(): firebase_source = mapper.get_source_firebase() airports = firebase_source.get_all() for airport_code, status in airports.items(): if "status" in status: del status["status"] return jsonify(airports) <commit_msg>Return index.html in root and transform /status results<commit_after>
from flask import jsonify from . import app import mapper import utils from predict import predictor @app.route("/", methods=["GET"]) def index(): return app.send_static_file("index.html") @app.route("/build", methods=["POST"]) def build_model(): predictor.preprocess_airports() if not predictor.model: predictor.build_model() return jsonify({"message:" : "OK"}) @app.route("/predict", methods=["GET"]) def predict_all_delays(): results = None try: results = predictor.predict_all() except Exception as e: print "ERROR", e.message return jsonify({"message" : e.message}) return jsonify(results) @app.route("/predict/<airport_code>", methods=["GET"]) def predict_delay(airport_code): firebase_source = mapper.get_source_firebase() airport_status = firebase_source.get_airport(airport_code) cleaned_data = utils.get_clean_data(airport_status) res = predictor.predict(airport_code) cleaned_data["prediction"] = bool(res[0]) return jsonify(cleaned_data) @app.route("/status", methods=["GET"]) def get_airport_statuses(): firebase_source = mapper.get_source_firebase() airports = firebase_source.get_all() results = [] for airport_code, status in airports.items(): try: results.append(utils.get_clean_data(status)) except: pass results = {"items":results} return jsonify(results)
from flask import jsonify from . import app import mapper import utils from predict import predictor @app.route("/", methods=["GET"]) def index(): firebase_dump = mapper.get_dump_firebase() response = firebase_dump.get_all() response = response or {} return jsonify(response) @app.route("/build", methods=["POST"]) def build_model(): predictor.preprocess_airports() if not predictor.model: predictor.build_model() return jsonify({"message:" : "OK"}) @app.route("/predict", methods=["GET"]) def predict_all_delays(): results = None try: results = predictor.predict_all() except Exception as e: return jsonify({"message" : e.message}) return jsonify(results) @app.route("/predict/<airport_code>", methods=["GET"]) def predict_delay(airport_code): firebase_source = mapper.get_source_firebase() airport_status = firebase_source.get_airport(airport_code) cleaned_data = utils.get_clean_data(airport_status) res = predictor.predict(airport_code) cleaned_data["prediction"] = bool(res[0]) return jsonify(cleaned_data) @app.route("/status", methods=["GET"]) def get_airport_statuses(): firebase_source = mapper.get_source_firebase() airports = firebase_source.get_all() for airport_code, status in airports.items(): if "status" in status: del status["status"] return jsonify(airports) Return index.html in root and transform /status resultsfrom flask import jsonify from . import app import mapper import utils from predict import predictor @app.route("/", methods=["GET"]) def index(): return app.send_static_file("index.html") @app.route("/build", methods=["POST"]) def build_model(): predictor.preprocess_airports() if not predictor.model: predictor.build_model() return jsonify({"message:" : "OK"}) @app.route("/predict", methods=["GET"]) def predict_all_delays(): results = None try: results = predictor.predict_all() except Exception as e: print "ERROR", e.message return jsonify({"message" : e.message}) return jsonify(results) @app.route("/predict/<airport_code>", methods=["GET"]) def predict_delay(airport_code): firebase_source = mapper.get_source_firebase() airport_status = firebase_source.get_airport(airport_code) cleaned_data = utils.get_clean_data(airport_status) res = predictor.predict(airport_code) cleaned_data["prediction"] = bool(res[0]) return jsonify(cleaned_data) @app.route("/status", methods=["GET"]) def get_airport_statuses(): firebase_source = mapper.get_source_firebase() airports = firebase_source.get_all() results = [] for airport_code, status in airports.items(): try: results.append(utils.get_clean_data(status)) except: pass results = {"items":results} return jsonify(results)
<commit_before>from flask import jsonify from . import app import mapper import utils from predict import predictor @app.route("/", methods=["GET"]) def index(): firebase_dump = mapper.get_dump_firebase() response = firebase_dump.get_all() response = response or {} return jsonify(response) @app.route("/build", methods=["POST"]) def build_model(): predictor.preprocess_airports() if not predictor.model: predictor.build_model() return jsonify({"message:" : "OK"}) @app.route("/predict", methods=["GET"]) def predict_all_delays(): results = None try: results = predictor.predict_all() except Exception as e: return jsonify({"message" : e.message}) return jsonify(results) @app.route("/predict/<airport_code>", methods=["GET"]) def predict_delay(airport_code): firebase_source = mapper.get_source_firebase() airport_status = firebase_source.get_airport(airport_code) cleaned_data = utils.get_clean_data(airport_status) res = predictor.predict(airport_code) cleaned_data["prediction"] = bool(res[0]) return jsonify(cleaned_data) @app.route("/status", methods=["GET"]) def get_airport_statuses(): firebase_source = mapper.get_source_firebase() airports = firebase_source.get_all() for airport_code, status in airports.items(): if "status" in status: del status["status"] return jsonify(airports) <commit_msg>Return index.html in root and transform /status results<commit_after>from flask import jsonify from . import app import mapper import utils from predict import predictor @app.route("/", methods=["GET"]) def index(): return app.send_static_file("index.html") @app.route("/build", methods=["POST"]) def build_model(): predictor.preprocess_airports() if not predictor.model: predictor.build_model() return jsonify({"message:" : "OK"}) @app.route("/predict", methods=["GET"]) def predict_all_delays(): results = None try: results = predictor.predict_all() except Exception as e: print "ERROR", e.message return jsonify({"message" : e.message}) return jsonify(results) @app.route("/predict/<airport_code>", methods=["GET"]) def predict_delay(airport_code): firebase_source = mapper.get_source_firebase() airport_status = firebase_source.get_airport(airport_code) cleaned_data = utils.get_clean_data(airport_status) res = predictor.predict(airport_code) cleaned_data["prediction"] = bool(res[0]) return jsonify(cleaned_data) @app.route("/status", methods=["GET"]) def get_airport_statuses(): firebase_source = mapper.get_source_firebase() airports = firebase_source.get_all() results = [] for airport_code, status in airports.items(): try: results.append(utils.get_clean_data(status)) except: pass results = {"items":results} return jsonify(results)
9bb14514a523484af6313008baef3b7cfd987951
tests/__init__.py
tests/__init__.py
import sys import doctest def fix_doctests(suite): if sys.version_info.major >= 3: return for case in suite._tests: # Add some more flags. case._dt_optionflags = ( (case._dt_optionflags or 0) | doctest.IGNORE_EXCEPTION_DETAIL | doctest.ELLIPSIS | doctest.NORMALIZE_WHITESPACE ) test = case._dt_test for example in test.examples: # Remove b prefix from strings. if example.want.startswith("b'"): example.want = example.want[1:] def get_doctests(mod): suite = doctest.DocTestSuite(mod) fix_doctests(suite) return suite
import sys import doctest def fix_doctests(suite): if sys.version_info[0] >= 3: return for case in suite._tests: # Add some more flags. case._dt_optionflags = ( (case._dt_optionflags or 0) | doctest.IGNORE_EXCEPTION_DETAIL | doctest.ELLIPSIS | doctest.NORMALIZE_WHITESPACE ) test = case._dt_test for example in test.examples: # Remove b prefix from strings. if example.want.startswith("b'"): example.want = example.want[1:] def get_doctests(mod): suite = doctest.DocTestSuite(mod) fix_doctests(suite) return suite
Fix version_info check for Python2.6
Tests: Fix version_info check for Python2.6
Python
bsd-3-clause
mikeboers/PyTomCrypt,mikeboers/PyTomCrypt,mikeboers/PyTomCrypt
import sys import doctest def fix_doctests(suite): if sys.version_info.major >= 3: return for case in suite._tests: # Add some more flags. case._dt_optionflags = ( (case._dt_optionflags or 0) | doctest.IGNORE_EXCEPTION_DETAIL | doctest.ELLIPSIS | doctest.NORMALIZE_WHITESPACE ) test = case._dt_test for example in test.examples: # Remove b prefix from strings. if example.want.startswith("b'"): example.want = example.want[1:] def get_doctests(mod): suite = doctest.DocTestSuite(mod) fix_doctests(suite) return suite Tests: Fix version_info check for Python2.6
import sys import doctest def fix_doctests(suite): if sys.version_info[0] >= 3: return for case in suite._tests: # Add some more flags. case._dt_optionflags = ( (case._dt_optionflags or 0) | doctest.IGNORE_EXCEPTION_DETAIL | doctest.ELLIPSIS | doctest.NORMALIZE_WHITESPACE ) test = case._dt_test for example in test.examples: # Remove b prefix from strings. if example.want.startswith("b'"): example.want = example.want[1:] def get_doctests(mod): suite = doctest.DocTestSuite(mod) fix_doctests(suite) return suite
<commit_before>import sys import doctest def fix_doctests(suite): if sys.version_info.major >= 3: return for case in suite._tests: # Add some more flags. case._dt_optionflags = ( (case._dt_optionflags or 0) | doctest.IGNORE_EXCEPTION_DETAIL | doctest.ELLIPSIS | doctest.NORMALIZE_WHITESPACE ) test = case._dt_test for example in test.examples: # Remove b prefix from strings. if example.want.startswith("b'"): example.want = example.want[1:] def get_doctests(mod): suite = doctest.DocTestSuite(mod) fix_doctests(suite) return suite <commit_msg>Tests: Fix version_info check for Python2.6<commit_after>
import sys import doctest def fix_doctests(suite): if sys.version_info[0] >= 3: return for case in suite._tests: # Add some more flags. case._dt_optionflags = ( (case._dt_optionflags or 0) | doctest.IGNORE_EXCEPTION_DETAIL | doctest.ELLIPSIS | doctest.NORMALIZE_WHITESPACE ) test = case._dt_test for example in test.examples: # Remove b prefix from strings. if example.want.startswith("b'"): example.want = example.want[1:] def get_doctests(mod): suite = doctest.DocTestSuite(mod) fix_doctests(suite) return suite
import sys import doctest def fix_doctests(suite): if sys.version_info.major >= 3: return for case in suite._tests: # Add some more flags. case._dt_optionflags = ( (case._dt_optionflags or 0) | doctest.IGNORE_EXCEPTION_DETAIL | doctest.ELLIPSIS | doctest.NORMALIZE_WHITESPACE ) test = case._dt_test for example in test.examples: # Remove b prefix from strings. if example.want.startswith("b'"): example.want = example.want[1:] def get_doctests(mod): suite = doctest.DocTestSuite(mod) fix_doctests(suite) return suite Tests: Fix version_info check for Python2.6import sys import doctest def fix_doctests(suite): if sys.version_info[0] >= 3: return for case in suite._tests: # Add some more flags. case._dt_optionflags = ( (case._dt_optionflags or 0) | doctest.IGNORE_EXCEPTION_DETAIL | doctest.ELLIPSIS | doctest.NORMALIZE_WHITESPACE ) test = case._dt_test for example in test.examples: # Remove b prefix from strings. if example.want.startswith("b'"): example.want = example.want[1:] def get_doctests(mod): suite = doctest.DocTestSuite(mod) fix_doctests(suite) return suite
<commit_before>import sys import doctest def fix_doctests(suite): if sys.version_info.major >= 3: return for case in suite._tests: # Add some more flags. case._dt_optionflags = ( (case._dt_optionflags or 0) | doctest.IGNORE_EXCEPTION_DETAIL | doctest.ELLIPSIS | doctest.NORMALIZE_WHITESPACE ) test = case._dt_test for example in test.examples: # Remove b prefix from strings. if example.want.startswith("b'"): example.want = example.want[1:] def get_doctests(mod): suite = doctest.DocTestSuite(mod) fix_doctests(suite) return suite <commit_msg>Tests: Fix version_info check for Python2.6<commit_after>import sys import doctest def fix_doctests(suite): if sys.version_info[0] >= 3: return for case in suite._tests: # Add some more flags. case._dt_optionflags = ( (case._dt_optionflags or 0) | doctest.IGNORE_EXCEPTION_DETAIL | doctest.ELLIPSIS | doctest.NORMALIZE_WHITESPACE ) test = case._dt_test for example in test.examples: # Remove b prefix from strings. if example.want.startswith("b'"): example.want = example.want[1:] def get_doctests(mod): suite = doctest.DocTestSuite(mod) fix_doctests(suite) return suite
600839e3c51d2091a6c434ac31ea11dc9ed2db85
foialist/forms.py
foialist/forms.py
from django import forms from foialist.models import * class FileForm(forms.ModelForm): class Meta: model = File exclude = ('entry', 'size') class EntryForm(forms.ModelForm): govt_entity = forms.CharField(label="Gov't. entity") class Meta: model = Entry # exclude = ('slug', 'poster_slug', 'show', 'date_posted', 'entity') fields = ('title', 'narrative', 'government_entity', 'date_requested', 'date_filed', 'poster', 'email') class CommentForm(forms.ModelForm): poster = forms.CharField() class Meta: model = Comment
from django import forms from foialist.models import * class FileForm(forms.ModelForm): class Meta: model = File exclude = ('entry', 'size') class EntryForm(forms.ModelForm): govt_entity = forms.CharField(label="Gov't. entity") class Meta: model = Entry fields = ('title', 'narrative', 'govt_entity', 'date_requested', 'date_filed', 'poster', 'email') class CommentForm(forms.ModelForm): poster = forms.CharField() class Meta: model = Comment
Correct mismatched field names in EntryForm.
Correct mismatched field names in EntryForm.
Python
bsd-3-clause
a2civictech/a2docs-sources,a2civictech/a2docs-sources,a2civictech/a2docs-sources
from django import forms from foialist.models import * class FileForm(forms.ModelForm): class Meta: model = File exclude = ('entry', 'size') class EntryForm(forms.ModelForm): govt_entity = forms.CharField(label="Gov't. entity") class Meta: model = Entry # exclude = ('slug', 'poster_slug', 'show', 'date_posted', 'entity') fields = ('title', 'narrative', 'government_entity', 'date_requested', 'date_filed', 'poster', 'email') class CommentForm(forms.ModelForm): poster = forms.CharField() class Meta: model = CommentCorrect mismatched field names in EntryForm.
from django import forms from foialist.models import * class FileForm(forms.ModelForm): class Meta: model = File exclude = ('entry', 'size') class EntryForm(forms.ModelForm): govt_entity = forms.CharField(label="Gov't. entity") class Meta: model = Entry fields = ('title', 'narrative', 'govt_entity', 'date_requested', 'date_filed', 'poster', 'email') class CommentForm(forms.ModelForm): poster = forms.CharField() class Meta: model = Comment
<commit_before>from django import forms from foialist.models import * class FileForm(forms.ModelForm): class Meta: model = File exclude = ('entry', 'size') class EntryForm(forms.ModelForm): govt_entity = forms.CharField(label="Gov't. entity") class Meta: model = Entry # exclude = ('slug', 'poster_slug', 'show', 'date_posted', 'entity') fields = ('title', 'narrative', 'government_entity', 'date_requested', 'date_filed', 'poster', 'email') class CommentForm(forms.ModelForm): poster = forms.CharField() class Meta: model = Comment<commit_msg>Correct mismatched field names in EntryForm.<commit_after>
from django import forms from foialist.models import * class FileForm(forms.ModelForm): class Meta: model = File exclude = ('entry', 'size') class EntryForm(forms.ModelForm): govt_entity = forms.CharField(label="Gov't. entity") class Meta: model = Entry fields = ('title', 'narrative', 'govt_entity', 'date_requested', 'date_filed', 'poster', 'email') class CommentForm(forms.ModelForm): poster = forms.CharField() class Meta: model = Comment
from django import forms from foialist.models import * class FileForm(forms.ModelForm): class Meta: model = File exclude = ('entry', 'size') class EntryForm(forms.ModelForm): govt_entity = forms.CharField(label="Gov't. entity") class Meta: model = Entry # exclude = ('slug', 'poster_slug', 'show', 'date_posted', 'entity') fields = ('title', 'narrative', 'government_entity', 'date_requested', 'date_filed', 'poster', 'email') class CommentForm(forms.ModelForm): poster = forms.CharField() class Meta: model = CommentCorrect mismatched field names in EntryForm.from django import forms from foialist.models import * class FileForm(forms.ModelForm): class Meta: model = File exclude = ('entry', 'size') class EntryForm(forms.ModelForm): govt_entity = forms.CharField(label="Gov't. entity") class Meta: model = Entry fields = ('title', 'narrative', 'govt_entity', 'date_requested', 'date_filed', 'poster', 'email') class CommentForm(forms.ModelForm): poster = forms.CharField() class Meta: model = Comment
<commit_before>from django import forms from foialist.models import * class FileForm(forms.ModelForm): class Meta: model = File exclude = ('entry', 'size') class EntryForm(forms.ModelForm): govt_entity = forms.CharField(label="Gov't. entity") class Meta: model = Entry # exclude = ('slug', 'poster_slug', 'show', 'date_posted', 'entity') fields = ('title', 'narrative', 'government_entity', 'date_requested', 'date_filed', 'poster', 'email') class CommentForm(forms.ModelForm): poster = forms.CharField() class Meta: model = Comment<commit_msg>Correct mismatched field names in EntryForm.<commit_after>from django import forms from foialist.models import * class FileForm(forms.ModelForm): class Meta: model = File exclude = ('entry', 'size') class EntryForm(forms.ModelForm): govt_entity = forms.CharField(label="Gov't. entity") class Meta: model = Entry fields = ('title', 'narrative', 'govt_entity', 'date_requested', 'date_filed', 'poster', 'email') class CommentForm(forms.ModelForm): poster = forms.CharField() class Meta: model = Comment
5d7a179e99632e2b8ca30bfa444497636492ca5a
catsnap/web/middleware/exception_logger.py
catsnap/web/middleware/exception_logger.py
import sys import traceback class ExceptionLogger(object): def __init__(self, app): self.app = app def __call__(self, environ, start_response): try: return self.app(environ, start_response) except Exception: (exc_type, exc_value, trace) = sys.exc_info() traceback.print_exception(exc_type, exc_value, trace, sys.stdout) raise
import sys import traceback class ExceptionLogger(object): def __init__(self, app): self.app = app def __call__(self, environ, start_response): try: return self.app(environ, start_response) except Exception: (exc_type, exc_value, trace) = sys.exc_info() traceback.print_exception(exc_type, exc_value, trace, file=sys.stdout) raise
Fix the exception logger to actually log to stdout
Fix the exception logger to actually log to stdout
Python
mit
ErinCall/catsnap,ErinCall/catsnap,ErinCall/catsnap
import sys import traceback class ExceptionLogger(object): def __init__(self, app): self.app = app def __call__(self, environ, start_response): try: return self.app(environ, start_response) except Exception: (exc_type, exc_value, trace) = sys.exc_info() traceback.print_exception(exc_type, exc_value, trace, sys.stdout) raise Fix the exception logger to actually log to stdout
import sys import traceback class ExceptionLogger(object): def __init__(self, app): self.app = app def __call__(self, environ, start_response): try: return self.app(environ, start_response) except Exception: (exc_type, exc_value, trace) = sys.exc_info() traceback.print_exception(exc_type, exc_value, trace, file=sys.stdout) raise
<commit_before>import sys import traceback class ExceptionLogger(object): def __init__(self, app): self.app = app def __call__(self, environ, start_response): try: return self.app(environ, start_response) except Exception: (exc_type, exc_value, trace) = sys.exc_info() traceback.print_exception(exc_type, exc_value, trace, sys.stdout) raise <commit_msg>Fix the exception logger to actually log to stdout<commit_after>
import sys import traceback class ExceptionLogger(object): def __init__(self, app): self.app = app def __call__(self, environ, start_response): try: return self.app(environ, start_response) except Exception: (exc_type, exc_value, trace) = sys.exc_info() traceback.print_exception(exc_type, exc_value, trace, file=sys.stdout) raise
import sys import traceback class ExceptionLogger(object): def __init__(self, app): self.app = app def __call__(self, environ, start_response): try: return self.app(environ, start_response) except Exception: (exc_type, exc_value, trace) = sys.exc_info() traceback.print_exception(exc_type, exc_value, trace, sys.stdout) raise Fix the exception logger to actually log to stdoutimport sys import traceback class ExceptionLogger(object): def __init__(self, app): self.app = app def __call__(self, environ, start_response): try: return self.app(environ, start_response) except Exception: (exc_type, exc_value, trace) = sys.exc_info() traceback.print_exception(exc_type, exc_value, trace, file=sys.stdout) raise
<commit_before>import sys import traceback class ExceptionLogger(object): def __init__(self, app): self.app = app def __call__(self, environ, start_response): try: return self.app(environ, start_response) except Exception: (exc_type, exc_value, trace) = sys.exc_info() traceback.print_exception(exc_type, exc_value, trace, sys.stdout) raise <commit_msg>Fix the exception logger to actually log to stdout<commit_after>import sys import traceback class ExceptionLogger(object): def __init__(self, app): self.app = app def __call__(self, environ, start_response): try: return self.app(environ, start_response) except Exception: (exc_type, exc_value, trace) = sys.exc_info() traceback.print_exception(exc_type, exc_value, trace, file=sys.stdout) raise
6a7302bed399aba98b01490f78728d3daa57e092
opps/images/generate.py
opps/images/generate.py
#!/usr/bin/env python # -*- coding: utf-8 -*- from libthumbor import CryptoURL from django.conf import settings crypto = CryptoURL(key=settings.THUMBOR_SECURITY_KEY) def _remove_prefix(url, prefix): if url.startswith(prefix): return url[len(prefix):] return url def _remove_schema(url): return _remove_prefix(url, 'http://') def _prepend_media_url(url): if url.startswith(settings.MEDIA_URL): url = _remove_prefix(url, settings.MEDIA_URL) url.lstrip('/') return u'{}/{}'.format(settings.THUMBOR_MEDIA_URL, url) return url def image_url(image_url, **kwargs): if not settings.THUMBOR_ENABLED: # TODO: alternative ways for generating the thumbnails return image_url image_url = _prepend_media_url(image_url) image_url = _remove_schema(image_url) encrypted_url = crypto.generate(image_url=image_url, **dict(settings.THUMBOR_ARGUMENTS, **kwargs)).strip('/') return u'{}/{}'.format(settings.THUMBOR_SERVER, encrypted_url)
#!/usr/bin/env python # -*- coding: utf-8 -*- from libthumbor import CryptoURL from django.conf import settings crypto = CryptoURL(key=settings.THUMBOR_SECURITY_KEY) def _remove_prefix(url, prefix): if url.startswith(prefix): return url[len(prefix):] return url def _remove_schema(url): return _remove_prefix(url, 'http://') def _prepend_media_url(url): if url.startswith(settings.MEDIA_URL): url = _remove_prefix(url, settings.MEDIA_URL) url.lstrip('/') return u'{}/{}'.format(settings.THUMBOR_MEDIA_URL, url) return url def image_url(image_url, **kwargs): if not settings.THUMBOR_ENABLED or image_url.split('.')[-1] == 'git': # TODO: alternative ways for generating the thumbnails return image_url image_url = _prepend_media_url(image_url) image_url = _remove_schema(image_url) encrypted_url = crypto.generate(image_url=image_url, **dict(settings.THUMBOR_ARGUMENTS, **kwargs)).strip('/') return u'{}/{}'.format(settings.THUMBOR_SERVER, encrypted_url)
Fix render git, if gif file not render via thumbor
Fix render git, if gif file not render via thumbor
Python
mit
opps/opps,YACOWS/opps,williamroot/opps,opps/opps,williamroot/opps,jeanmask/opps,jeanmask/opps,williamroot/opps,opps/opps,YACOWS/opps,jeanmask/opps,YACOWS/opps,williamroot/opps,jeanmask/opps,opps/opps,YACOWS/opps
#!/usr/bin/env python # -*- coding: utf-8 -*- from libthumbor import CryptoURL from django.conf import settings crypto = CryptoURL(key=settings.THUMBOR_SECURITY_KEY) def _remove_prefix(url, prefix): if url.startswith(prefix): return url[len(prefix):] return url def _remove_schema(url): return _remove_prefix(url, 'http://') def _prepend_media_url(url): if url.startswith(settings.MEDIA_URL): url = _remove_prefix(url, settings.MEDIA_URL) url.lstrip('/') return u'{}/{}'.format(settings.THUMBOR_MEDIA_URL, url) return url def image_url(image_url, **kwargs): if not settings.THUMBOR_ENABLED: # TODO: alternative ways for generating the thumbnails return image_url image_url = _prepend_media_url(image_url) image_url = _remove_schema(image_url) encrypted_url = crypto.generate(image_url=image_url, **dict(settings.THUMBOR_ARGUMENTS, **kwargs)).strip('/') return u'{}/{}'.format(settings.THUMBOR_SERVER, encrypted_url) Fix render git, if gif file not render via thumbor
#!/usr/bin/env python # -*- coding: utf-8 -*- from libthumbor import CryptoURL from django.conf import settings crypto = CryptoURL(key=settings.THUMBOR_SECURITY_KEY) def _remove_prefix(url, prefix): if url.startswith(prefix): return url[len(prefix):] return url def _remove_schema(url): return _remove_prefix(url, 'http://') def _prepend_media_url(url): if url.startswith(settings.MEDIA_URL): url = _remove_prefix(url, settings.MEDIA_URL) url.lstrip('/') return u'{}/{}'.format(settings.THUMBOR_MEDIA_URL, url) return url def image_url(image_url, **kwargs): if not settings.THUMBOR_ENABLED or image_url.split('.')[-1] == 'git': # TODO: alternative ways for generating the thumbnails return image_url image_url = _prepend_media_url(image_url) image_url = _remove_schema(image_url) encrypted_url = crypto.generate(image_url=image_url, **dict(settings.THUMBOR_ARGUMENTS, **kwargs)).strip('/') return u'{}/{}'.format(settings.THUMBOR_SERVER, encrypted_url)
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- from libthumbor import CryptoURL from django.conf import settings crypto = CryptoURL(key=settings.THUMBOR_SECURITY_KEY) def _remove_prefix(url, prefix): if url.startswith(prefix): return url[len(prefix):] return url def _remove_schema(url): return _remove_prefix(url, 'http://') def _prepend_media_url(url): if url.startswith(settings.MEDIA_URL): url = _remove_prefix(url, settings.MEDIA_URL) url.lstrip('/') return u'{}/{}'.format(settings.THUMBOR_MEDIA_URL, url) return url def image_url(image_url, **kwargs): if not settings.THUMBOR_ENABLED: # TODO: alternative ways for generating the thumbnails return image_url image_url = _prepend_media_url(image_url) image_url = _remove_schema(image_url) encrypted_url = crypto.generate(image_url=image_url, **dict(settings.THUMBOR_ARGUMENTS, **kwargs)).strip('/') return u'{}/{}'.format(settings.THUMBOR_SERVER, encrypted_url) <commit_msg>Fix render git, if gif file not render via thumbor<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- from libthumbor import CryptoURL from django.conf import settings crypto = CryptoURL(key=settings.THUMBOR_SECURITY_KEY) def _remove_prefix(url, prefix): if url.startswith(prefix): return url[len(prefix):] return url def _remove_schema(url): return _remove_prefix(url, 'http://') def _prepend_media_url(url): if url.startswith(settings.MEDIA_URL): url = _remove_prefix(url, settings.MEDIA_URL) url.lstrip('/') return u'{}/{}'.format(settings.THUMBOR_MEDIA_URL, url) return url def image_url(image_url, **kwargs): if not settings.THUMBOR_ENABLED or image_url.split('.')[-1] == 'git': # TODO: alternative ways for generating the thumbnails return image_url image_url = _prepend_media_url(image_url) image_url = _remove_schema(image_url) encrypted_url = crypto.generate(image_url=image_url, **dict(settings.THUMBOR_ARGUMENTS, **kwargs)).strip('/') return u'{}/{}'.format(settings.THUMBOR_SERVER, encrypted_url)
#!/usr/bin/env python # -*- coding: utf-8 -*- from libthumbor import CryptoURL from django.conf import settings crypto = CryptoURL(key=settings.THUMBOR_SECURITY_KEY) def _remove_prefix(url, prefix): if url.startswith(prefix): return url[len(prefix):] return url def _remove_schema(url): return _remove_prefix(url, 'http://') def _prepend_media_url(url): if url.startswith(settings.MEDIA_URL): url = _remove_prefix(url, settings.MEDIA_URL) url.lstrip('/') return u'{}/{}'.format(settings.THUMBOR_MEDIA_URL, url) return url def image_url(image_url, **kwargs): if not settings.THUMBOR_ENABLED: # TODO: alternative ways for generating the thumbnails return image_url image_url = _prepend_media_url(image_url) image_url = _remove_schema(image_url) encrypted_url = crypto.generate(image_url=image_url, **dict(settings.THUMBOR_ARGUMENTS, **kwargs)).strip('/') return u'{}/{}'.format(settings.THUMBOR_SERVER, encrypted_url) Fix render git, if gif file not render via thumbor#!/usr/bin/env python # -*- coding: utf-8 -*- from libthumbor import CryptoURL from django.conf import settings crypto = CryptoURL(key=settings.THUMBOR_SECURITY_KEY) def _remove_prefix(url, prefix): if url.startswith(prefix): return url[len(prefix):] return url def _remove_schema(url): return _remove_prefix(url, 'http://') def _prepend_media_url(url): if url.startswith(settings.MEDIA_URL): url = _remove_prefix(url, settings.MEDIA_URL) url.lstrip('/') return u'{}/{}'.format(settings.THUMBOR_MEDIA_URL, url) return url def image_url(image_url, **kwargs): if not settings.THUMBOR_ENABLED or image_url.split('.')[-1] == 'git': # TODO: alternative ways for generating the thumbnails return image_url image_url = _prepend_media_url(image_url) image_url = _remove_schema(image_url) encrypted_url = crypto.generate(image_url=image_url, **dict(settings.THUMBOR_ARGUMENTS, **kwargs)).strip('/') return u'{}/{}'.format(settings.THUMBOR_SERVER, encrypted_url)
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- from libthumbor import CryptoURL from django.conf import settings crypto = CryptoURL(key=settings.THUMBOR_SECURITY_KEY) def _remove_prefix(url, prefix): if url.startswith(prefix): return url[len(prefix):] return url def _remove_schema(url): return _remove_prefix(url, 'http://') def _prepend_media_url(url): if url.startswith(settings.MEDIA_URL): url = _remove_prefix(url, settings.MEDIA_URL) url.lstrip('/') return u'{}/{}'.format(settings.THUMBOR_MEDIA_URL, url) return url def image_url(image_url, **kwargs): if not settings.THUMBOR_ENABLED: # TODO: alternative ways for generating the thumbnails return image_url image_url = _prepend_media_url(image_url) image_url = _remove_schema(image_url) encrypted_url = crypto.generate(image_url=image_url, **dict(settings.THUMBOR_ARGUMENTS, **kwargs)).strip('/') return u'{}/{}'.format(settings.THUMBOR_SERVER, encrypted_url) <commit_msg>Fix render git, if gif file not render via thumbor<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- from libthumbor import CryptoURL from django.conf import settings crypto = CryptoURL(key=settings.THUMBOR_SECURITY_KEY) def _remove_prefix(url, prefix): if url.startswith(prefix): return url[len(prefix):] return url def _remove_schema(url): return _remove_prefix(url, 'http://') def _prepend_media_url(url): if url.startswith(settings.MEDIA_URL): url = _remove_prefix(url, settings.MEDIA_URL) url.lstrip('/') return u'{}/{}'.format(settings.THUMBOR_MEDIA_URL, url) return url def image_url(image_url, **kwargs): if not settings.THUMBOR_ENABLED or image_url.split('.')[-1] == 'git': # TODO: alternative ways for generating the thumbnails return image_url image_url = _prepend_media_url(image_url) image_url = _remove_schema(image_url) encrypted_url = crypto.generate(image_url=image_url, **dict(settings.THUMBOR_ARGUMENTS, **kwargs)).strip('/') return u'{}/{}'.format(settings.THUMBOR_SERVER, encrypted_url)
059230327fcebb35c881f8a6bc2ee12fed29d442
mcp/config.py
mcp/config.py
# # Copyright (c) 2005-2006 rPath, Inc. # # All rights reserved # import os from conary import conarycfg from conary.lib import cfgtypes class MCPConfig(conarycfg.ConfigFile): basePath = os.path.join(os.path.sep, 'srv', 'rbuilder', 'mcp') logPath = os.path.join(basePath, 'logs') queueHost = '127.0.0.1' queuePort = (cfgtypes.CfgInt, 61613) namespace = 'mcp' slaveTroveName = 'group-jobslave' slaveTroveLabel = 'products.rpath.com@rpath:js'
# # Copyright (c) 2005-2006 rPath, Inc. # # All rights reserved # import os from conary import conarycfg from conary.lib import cfgtypes class MCPConfig(conarycfg.ConfigFile): basePath = os.path.join(os.path.sep, 'srv', 'rbuilder', 'mcp') logPath = os.path.join(os.path.sep, 'var', 'log', 'rbuilder') queueHost = '127.0.0.1' queuePort = (cfgtypes.CfgInt, 61613) namespace = 'mcp' slaveTroveName = 'group-jobslave' slaveTroveLabel = 'products.rpath.com@rpath:js'
Move default location for MCP logs into /var/log/rbuilder/
Move default location for MCP logs into /var/log/rbuilder/
Python
apache-2.0
sassoftware/mcp,sassoftware/mcp
# # Copyright (c) 2005-2006 rPath, Inc. # # All rights reserved # import os from conary import conarycfg from conary.lib import cfgtypes class MCPConfig(conarycfg.ConfigFile): basePath = os.path.join(os.path.sep, 'srv', 'rbuilder', 'mcp') logPath = os.path.join(basePath, 'logs') queueHost = '127.0.0.1' queuePort = (cfgtypes.CfgInt, 61613) namespace = 'mcp' slaveTroveName = 'group-jobslave' slaveTroveLabel = 'products.rpath.com@rpath:js' Move default location for MCP logs into /var/log/rbuilder/
# # Copyright (c) 2005-2006 rPath, Inc. # # All rights reserved # import os from conary import conarycfg from conary.lib import cfgtypes class MCPConfig(conarycfg.ConfigFile): basePath = os.path.join(os.path.sep, 'srv', 'rbuilder', 'mcp') logPath = os.path.join(os.path.sep, 'var', 'log', 'rbuilder') queueHost = '127.0.0.1' queuePort = (cfgtypes.CfgInt, 61613) namespace = 'mcp' slaveTroveName = 'group-jobslave' slaveTroveLabel = 'products.rpath.com@rpath:js'
<commit_before># # Copyright (c) 2005-2006 rPath, Inc. # # All rights reserved # import os from conary import conarycfg from conary.lib import cfgtypes class MCPConfig(conarycfg.ConfigFile): basePath = os.path.join(os.path.sep, 'srv', 'rbuilder', 'mcp') logPath = os.path.join(basePath, 'logs') queueHost = '127.0.0.1' queuePort = (cfgtypes.CfgInt, 61613) namespace = 'mcp' slaveTroveName = 'group-jobslave' slaveTroveLabel = 'products.rpath.com@rpath:js' <commit_msg>Move default location for MCP logs into /var/log/rbuilder/<commit_after>
# # Copyright (c) 2005-2006 rPath, Inc. # # All rights reserved # import os from conary import conarycfg from conary.lib import cfgtypes class MCPConfig(conarycfg.ConfigFile): basePath = os.path.join(os.path.sep, 'srv', 'rbuilder', 'mcp') logPath = os.path.join(os.path.sep, 'var', 'log', 'rbuilder') queueHost = '127.0.0.1' queuePort = (cfgtypes.CfgInt, 61613) namespace = 'mcp' slaveTroveName = 'group-jobslave' slaveTroveLabel = 'products.rpath.com@rpath:js'
# # Copyright (c) 2005-2006 rPath, Inc. # # All rights reserved # import os from conary import conarycfg from conary.lib import cfgtypes class MCPConfig(conarycfg.ConfigFile): basePath = os.path.join(os.path.sep, 'srv', 'rbuilder', 'mcp') logPath = os.path.join(basePath, 'logs') queueHost = '127.0.0.1' queuePort = (cfgtypes.CfgInt, 61613) namespace = 'mcp' slaveTroveName = 'group-jobslave' slaveTroveLabel = 'products.rpath.com@rpath:js' Move default location for MCP logs into /var/log/rbuilder/# # Copyright (c) 2005-2006 rPath, Inc. # # All rights reserved # import os from conary import conarycfg from conary.lib import cfgtypes class MCPConfig(conarycfg.ConfigFile): basePath = os.path.join(os.path.sep, 'srv', 'rbuilder', 'mcp') logPath = os.path.join(os.path.sep, 'var', 'log', 'rbuilder') queueHost = '127.0.0.1' queuePort = (cfgtypes.CfgInt, 61613) namespace = 'mcp' slaveTroveName = 'group-jobslave' slaveTroveLabel = 'products.rpath.com@rpath:js'
<commit_before># # Copyright (c) 2005-2006 rPath, Inc. # # All rights reserved # import os from conary import conarycfg from conary.lib import cfgtypes class MCPConfig(conarycfg.ConfigFile): basePath = os.path.join(os.path.sep, 'srv', 'rbuilder', 'mcp') logPath = os.path.join(basePath, 'logs') queueHost = '127.0.0.1' queuePort = (cfgtypes.CfgInt, 61613) namespace = 'mcp' slaveTroveName = 'group-jobslave' slaveTroveLabel = 'products.rpath.com@rpath:js' <commit_msg>Move default location for MCP logs into /var/log/rbuilder/<commit_after># # Copyright (c) 2005-2006 rPath, Inc. # # All rights reserved # import os from conary import conarycfg from conary.lib import cfgtypes class MCPConfig(conarycfg.ConfigFile): basePath = os.path.join(os.path.sep, 'srv', 'rbuilder', 'mcp') logPath = os.path.join(os.path.sep, 'var', 'log', 'rbuilder') queueHost = '127.0.0.1' queuePort = (cfgtypes.CfgInt, 61613) namespace = 'mcp' slaveTroveName = 'group-jobslave' slaveTroveLabel = 'products.rpath.com@rpath:js'
520ad6a456cbd94e176bb54373669baf5e8cfbd9
sprockets/mixins/correlation/__init__.py
sprockets/mixins/correlation/__init__.py
from .mixins import HandlerMixin version_info = (1, 0, 2) __version__ = '.'.join(str(v) for v in version_info[:3])
try: from .mixins import HandlerMixin except ImportError as error: class HandlerMixin(object): def __init__(self, *args, **kwargs): raise error version_info = (1, 0, 2) __version__ = '.'.join(str(v) for v in version_info[:3])
Fix retrieving __version__ without Tornado installed.
Fix retrieving __version__ without Tornado installed.
Python
bsd-3-clause
sprockets/sprockets.mixins.correlation
from .mixins import HandlerMixin version_info = (1, 0, 2) __version__ = '.'.join(str(v) for v in version_info[:3]) Fix retrieving __version__ without Tornado installed.
try: from .mixins import HandlerMixin except ImportError as error: class HandlerMixin(object): def __init__(self, *args, **kwargs): raise error version_info = (1, 0, 2) __version__ = '.'.join(str(v) for v in version_info[:3])
<commit_before>from .mixins import HandlerMixin version_info = (1, 0, 2) __version__ = '.'.join(str(v) for v in version_info[:3]) <commit_msg>Fix retrieving __version__ without Tornado installed.<commit_after>
try: from .mixins import HandlerMixin except ImportError as error: class HandlerMixin(object): def __init__(self, *args, **kwargs): raise error version_info = (1, 0, 2) __version__ = '.'.join(str(v) for v in version_info[:3])
from .mixins import HandlerMixin version_info = (1, 0, 2) __version__ = '.'.join(str(v) for v in version_info[:3]) Fix retrieving __version__ without Tornado installed.try: from .mixins import HandlerMixin except ImportError as error: class HandlerMixin(object): def __init__(self, *args, **kwargs): raise error version_info = (1, 0, 2) __version__ = '.'.join(str(v) for v in version_info[:3])
<commit_before>from .mixins import HandlerMixin version_info = (1, 0, 2) __version__ = '.'.join(str(v) for v in version_info[:3]) <commit_msg>Fix retrieving __version__ without Tornado installed.<commit_after>try: from .mixins import HandlerMixin except ImportError as error: class HandlerMixin(object): def __init__(self, *args, **kwargs): raise error version_info = (1, 0, 2) __version__ = '.'.join(str(v) for v in version_info[:3])
99eafe1fb8ed3edce0d8d025b74ffdffa3bf8ae6
fabfile.py
fabfile.py
import sys import sh from fabric import api as fab sed = sh.sed.bake('-i bak -e') TRAVIS_YAML = '.travis.yml' REPLACE_LANGUAGE = 's/language: .*/language: {}/' def is_dirty(): return "" != sh.git.status(porcelain=True).strip() def release(language, message): if is_dirty(): sys.exit("Repo must be in clean state before deploying. Please commit changes.") sed(REPLACE_LANGUAGE.format(language), TRAVIS_YAML) if is_dirty(): sh.git.add(TRAVIS_YAML) sh.git.commit(m=message, allow_empty=True) sh.git.pull(rebase=True) sh.git.push() @fab.task def update(): if is_dirty(): sys.exit("Repo must be in clean state before deploying. Please commit changes.") sh.git.submodule.update(remote=True, rebase=True) if is_dirty(): sh.git.add(all=True) sh.git.commit(m="Update submodules to origin") @fab.task def release_osx(): release('objective-c', "Release OS X") @fab.task def release_linux(): release('python', "Release Linux")
import sys import sh from fabric import api as fab sed = sh.sed.bake('-i bak -e') TRAVIS_YAML = '.travis.yml' REPLACE_LANGUAGE = 's/language: .*/language: {}/' def is_dirty(): return "" != sh.git.status(porcelain=True).strip() def release(language, message): if is_dirty(): sys.exit("Repo must be in clean state before deploying. Please commit changes.") sed(REPLACE_LANGUAGE.format(language), TRAVIS_YAML) if is_dirty(): sh.git.add(TRAVIS_YAML) sh.git.commit(m=message, allow_empty=True) sh.git.pull(rebase=True) sh.git.push() @fab.task def update(): if is_dirty(): sys.exit("Repo must be in clean state before deploying. Please commit changes.") sh.git.submodule.update(remote=True, rebase=True) if is_dirty(): sh.git.add(all=True) sh.git.commit(m="Update submodules to origin") else: sys.exit('Nothing to update.') @fab.task def release_osx(): release('objective-c', "Release OS X") @fab.task def release_linux(): release('python', "Release Linux")
Print if nothing to update
Print if nothing to update
Python
bsd-3-clause
datamicroscopes/release,jzf2101/release,jzf2101/release,datamicroscopes/release
import sys import sh from fabric import api as fab sed = sh.sed.bake('-i bak -e') TRAVIS_YAML = '.travis.yml' REPLACE_LANGUAGE = 's/language: .*/language: {}/' def is_dirty(): return "" != sh.git.status(porcelain=True).strip() def release(language, message): if is_dirty(): sys.exit("Repo must be in clean state before deploying. Please commit changes.") sed(REPLACE_LANGUAGE.format(language), TRAVIS_YAML) if is_dirty(): sh.git.add(TRAVIS_YAML) sh.git.commit(m=message, allow_empty=True) sh.git.pull(rebase=True) sh.git.push() @fab.task def update(): if is_dirty(): sys.exit("Repo must be in clean state before deploying. Please commit changes.") sh.git.submodule.update(remote=True, rebase=True) if is_dirty(): sh.git.add(all=True) sh.git.commit(m="Update submodules to origin") @fab.task def release_osx(): release('objective-c', "Release OS X") @fab.task def release_linux(): release('python', "Release Linux") Print if nothing to update
import sys import sh from fabric import api as fab sed = sh.sed.bake('-i bak -e') TRAVIS_YAML = '.travis.yml' REPLACE_LANGUAGE = 's/language: .*/language: {}/' def is_dirty(): return "" != sh.git.status(porcelain=True).strip() def release(language, message): if is_dirty(): sys.exit("Repo must be in clean state before deploying. Please commit changes.") sed(REPLACE_LANGUAGE.format(language), TRAVIS_YAML) if is_dirty(): sh.git.add(TRAVIS_YAML) sh.git.commit(m=message, allow_empty=True) sh.git.pull(rebase=True) sh.git.push() @fab.task def update(): if is_dirty(): sys.exit("Repo must be in clean state before deploying. Please commit changes.") sh.git.submodule.update(remote=True, rebase=True) if is_dirty(): sh.git.add(all=True) sh.git.commit(m="Update submodules to origin") else: sys.exit('Nothing to update.') @fab.task def release_osx(): release('objective-c', "Release OS X") @fab.task def release_linux(): release('python', "Release Linux")
<commit_before>import sys import sh from fabric import api as fab sed = sh.sed.bake('-i bak -e') TRAVIS_YAML = '.travis.yml' REPLACE_LANGUAGE = 's/language: .*/language: {}/' def is_dirty(): return "" != sh.git.status(porcelain=True).strip() def release(language, message): if is_dirty(): sys.exit("Repo must be in clean state before deploying. Please commit changes.") sed(REPLACE_LANGUAGE.format(language), TRAVIS_YAML) if is_dirty(): sh.git.add(TRAVIS_YAML) sh.git.commit(m=message, allow_empty=True) sh.git.pull(rebase=True) sh.git.push() @fab.task def update(): if is_dirty(): sys.exit("Repo must be in clean state before deploying. Please commit changes.") sh.git.submodule.update(remote=True, rebase=True) if is_dirty(): sh.git.add(all=True) sh.git.commit(m="Update submodules to origin") @fab.task def release_osx(): release('objective-c', "Release OS X") @fab.task def release_linux(): release('python', "Release Linux") <commit_msg>Print if nothing to update<commit_after>
import sys import sh from fabric import api as fab sed = sh.sed.bake('-i bak -e') TRAVIS_YAML = '.travis.yml' REPLACE_LANGUAGE = 's/language: .*/language: {}/' def is_dirty(): return "" != sh.git.status(porcelain=True).strip() def release(language, message): if is_dirty(): sys.exit("Repo must be in clean state before deploying. Please commit changes.") sed(REPLACE_LANGUAGE.format(language), TRAVIS_YAML) if is_dirty(): sh.git.add(TRAVIS_YAML) sh.git.commit(m=message, allow_empty=True) sh.git.pull(rebase=True) sh.git.push() @fab.task def update(): if is_dirty(): sys.exit("Repo must be in clean state before deploying. Please commit changes.") sh.git.submodule.update(remote=True, rebase=True) if is_dirty(): sh.git.add(all=True) sh.git.commit(m="Update submodules to origin") else: sys.exit('Nothing to update.') @fab.task def release_osx(): release('objective-c', "Release OS X") @fab.task def release_linux(): release('python', "Release Linux")
import sys import sh from fabric import api as fab sed = sh.sed.bake('-i bak -e') TRAVIS_YAML = '.travis.yml' REPLACE_LANGUAGE = 's/language: .*/language: {}/' def is_dirty(): return "" != sh.git.status(porcelain=True).strip() def release(language, message): if is_dirty(): sys.exit("Repo must be in clean state before deploying. Please commit changes.") sed(REPLACE_LANGUAGE.format(language), TRAVIS_YAML) if is_dirty(): sh.git.add(TRAVIS_YAML) sh.git.commit(m=message, allow_empty=True) sh.git.pull(rebase=True) sh.git.push() @fab.task def update(): if is_dirty(): sys.exit("Repo must be in clean state before deploying. Please commit changes.") sh.git.submodule.update(remote=True, rebase=True) if is_dirty(): sh.git.add(all=True) sh.git.commit(m="Update submodules to origin") @fab.task def release_osx(): release('objective-c', "Release OS X") @fab.task def release_linux(): release('python', "Release Linux") Print if nothing to updateimport sys import sh from fabric import api as fab sed = sh.sed.bake('-i bak -e') TRAVIS_YAML = '.travis.yml' REPLACE_LANGUAGE = 's/language: .*/language: {}/' def is_dirty(): return "" != sh.git.status(porcelain=True).strip() def release(language, message): if is_dirty(): sys.exit("Repo must be in clean state before deploying. Please commit changes.") sed(REPLACE_LANGUAGE.format(language), TRAVIS_YAML) if is_dirty(): sh.git.add(TRAVIS_YAML) sh.git.commit(m=message, allow_empty=True) sh.git.pull(rebase=True) sh.git.push() @fab.task def update(): if is_dirty(): sys.exit("Repo must be in clean state before deploying. Please commit changes.") sh.git.submodule.update(remote=True, rebase=True) if is_dirty(): sh.git.add(all=True) sh.git.commit(m="Update submodules to origin") else: sys.exit('Nothing to update.') @fab.task def release_osx(): release('objective-c', "Release OS X") @fab.task def release_linux(): release('python', "Release Linux")
<commit_before>import sys import sh from fabric import api as fab sed = sh.sed.bake('-i bak -e') TRAVIS_YAML = '.travis.yml' REPLACE_LANGUAGE = 's/language: .*/language: {}/' def is_dirty(): return "" != sh.git.status(porcelain=True).strip() def release(language, message): if is_dirty(): sys.exit("Repo must be in clean state before deploying. Please commit changes.") sed(REPLACE_LANGUAGE.format(language), TRAVIS_YAML) if is_dirty(): sh.git.add(TRAVIS_YAML) sh.git.commit(m=message, allow_empty=True) sh.git.pull(rebase=True) sh.git.push() @fab.task def update(): if is_dirty(): sys.exit("Repo must be in clean state before deploying. Please commit changes.") sh.git.submodule.update(remote=True, rebase=True) if is_dirty(): sh.git.add(all=True) sh.git.commit(m="Update submodules to origin") @fab.task def release_osx(): release('objective-c', "Release OS X") @fab.task def release_linux(): release('python', "Release Linux") <commit_msg>Print if nothing to update<commit_after>import sys import sh from fabric import api as fab sed = sh.sed.bake('-i bak -e') TRAVIS_YAML = '.travis.yml' REPLACE_LANGUAGE = 's/language: .*/language: {}/' def is_dirty(): return "" != sh.git.status(porcelain=True).strip() def release(language, message): if is_dirty(): sys.exit("Repo must be in clean state before deploying. Please commit changes.") sed(REPLACE_LANGUAGE.format(language), TRAVIS_YAML) if is_dirty(): sh.git.add(TRAVIS_YAML) sh.git.commit(m=message, allow_empty=True) sh.git.pull(rebase=True) sh.git.push() @fab.task def update(): if is_dirty(): sys.exit("Repo must be in clean state before deploying. Please commit changes.") sh.git.submodule.update(remote=True, rebase=True) if is_dirty(): sh.git.add(all=True) sh.git.commit(m="Update submodules to origin") else: sys.exit('Nothing to update.') @fab.task def release_osx(): release('objective-c', "Release OS X") @fab.task def release_linux(): release('python', "Release Linux")
70f0be172801ee5fd205a90c78e2bf66f8e4ae07
playserver/webserver.py
playserver/webserver.py
import flask from . import track app = flask.Flask(__name__) @app.route("/") def root(): song = track.getCurrentSong() artist = track.getCurrentArtist() album = track.getCurrentAlbum() return "{} by {} - {}".format(song, artist, album)
import flask import json from . import track app = flask.Flask(__name__) @app.route("/") def root(): song = track.getCurrentSong() artist = track.getCurrentArtist() album = track.getCurrentAlbum() return "{} by {} - {}".format(song, artist, album) @app.route("/get_song_info") def getSongInfo(): return json.dumps({ "name": track.getCurrentSong(), "artist": track.getCurrentAritst(), "album": track.getCrrentAlbum() }) @app.route("/play_pause", methods = ["POST"]) def playPause(): track.playPause() return "" @app.route("/next", methods = ["POST"]) def next(): track.next() return "" @app.route("/previous", methods = ["POST"]) def previous(): track.previous() return ""
Add basic routes for controls and song info
Add basic routes for controls and song info
Python
mit
ollien/playserver,ollien/playserver,ollien/playserver
import flask from . import track app = flask.Flask(__name__) @app.route("/") def root(): song = track.getCurrentSong() artist = track.getCurrentArtist() album = track.getCurrentAlbum() return "{} by {} - {}".format(song, artist, album) Add basic routes for controls and song info
import flask import json from . import track app = flask.Flask(__name__) @app.route("/") def root(): song = track.getCurrentSong() artist = track.getCurrentArtist() album = track.getCurrentAlbum() return "{} by {} - {}".format(song, artist, album) @app.route("/get_song_info") def getSongInfo(): return json.dumps({ "name": track.getCurrentSong(), "artist": track.getCurrentAritst(), "album": track.getCrrentAlbum() }) @app.route("/play_pause", methods = ["POST"]) def playPause(): track.playPause() return "" @app.route("/next", methods = ["POST"]) def next(): track.next() return "" @app.route("/previous", methods = ["POST"]) def previous(): track.previous() return ""
<commit_before>import flask from . import track app = flask.Flask(__name__) @app.route("/") def root(): song = track.getCurrentSong() artist = track.getCurrentArtist() album = track.getCurrentAlbum() return "{} by {} - {}".format(song, artist, album) <commit_msg>Add basic routes for controls and song info<commit_after>
import flask import json from . import track app = flask.Flask(__name__) @app.route("/") def root(): song = track.getCurrentSong() artist = track.getCurrentArtist() album = track.getCurrentAlbum() return "{} by {} - {}".format(song, artist, album) @app.route("/get_song_info") def getSongInfo(): return json.dumps({ "name": track.getCurrentSong(), "artist": track.getCurrentAritst(), "album": track.getCrrentAlbum() }) @app.route("/play_pause", methods = ["POST"]) def playPause(): track.playPause() return "" @app.route("/next", methods = ["POST"]) def next(): track.next() return "" @app.route("/previous", methods = ["POST"]) def previous(): track.previous() return ""
import flask from . import track app = flask.Flask(__name__) @app.route("/") def root(): song = track.getCurrentSong() artist = track.getCurrentArtist() album = track.getCurrentAlbum() return "{} by {} - {}".format(song, artist, album) Add basic routes for controls and song infoimport flask import json from . import track app = flask.Flask(__name__) @app.route("/") def root(): song = track.getCurrentSong() artist = track.getCurrentArtist() album = track.getCurrentAlbum() return "{} by {} - {}".format(song, artist, album) @app.route("/get_song_info") def getSongInfo(): return json.dumps({ "name": track.getCurrentSong(), "artist": track.getCurrentAritst(), "album": track.getCrrentAlbum() }) @app.route("/play_pause", methods = ["POST"]) def playPause(): track.playPause() return "" @app.route("/next", methods = ["POST"]) def next(): track.next() return "" @app.route("/previous", methods = ["POST"]) def previous(): track.previous() return ""
<commit_before>import flask from . import track app = flask.Flask(__name__) @app.route("/") def root(): song = track.getCurrentSong() artist = track.getCurrentArtist() album = track.getCurrentAlbum() return "{} by {} - {}".format(song, artist, album) <commit_msg>Add basic routes for controls and song info<commit_after>import flask import json from . import track app = flask.Flask(__name__) @app.route("/") def root(): song = track.getCurrentSong() artist = track.getCurrentArtist() album = track.getCurrentAlbum() return "{} by {} - {}".format(song, artist, album) @app.route("/get_song_info") def getSongInfo(): return json.dumps({ "name": track.getCurrentSong(), "artist": track.getCurrentAritst(), "album": track.getCrrentAlbum() }) @app.route("/play_pause", methods = ["POST"]) def playPause(): track.playPause() return "" @app.route("/next", methods = ["POST"]) def next(): track.next() return "" @app.route("/previous", methods = ["POST"]) def previous(): track.previous() return ""
76d60adabc44fd3bbd432ee2cdad011b542a2fee
nel/features/mapping.py
nel/features/mapping.py
import math import numpy class FeatureMapper(object): def __init__(self, features): self.features = features def map(self, fv): raise NotImplementedError def __call__(self, doc): for chain in doc.chains: for c in chain.candidates: c.fv = self.map(numpy.array([c.features[f] for f in self.features])) return doc class ZeroMeanUnitVarianceMapper(FeatureMapper): def __init__(self, features, means, stds): super(ZeroMeanUnitVarianceMapper,self).__init__(features) self.mean = means self.std = stds def map(self, fv): return (fv - self.mean) / self.std class PolynomialMapper(ZeroMeanUnitVarianceMapper): def __init__(self, features, means, stds): super(PolynomialMapper,self).__init__(features, means, stds) def map(self, fv): fv = list(super(PolynomialMapper, self).map(fv)) sz = len(fv) for i in xrange(0, sz): for j in xrange(i, sz): weight = 1.0 if i != j else math.sqrt(2.0) fv.append(weight * fv[i]*fv[j]) return numpy.array(fv) FEATURE_MAPPERS = {cls.__name__:cls for cls in [ZeroMeanUnitVarianceMapper,PolynomialMapper]}
import math import numpy class FeatureMapper(object): def __init__(self, features): self.features = features def map(self, fv): raise NotImplementedError def __call__(self, doc): for chain in doc.chains: for c in chain.candidates: c.fv = self.map(numpy.array([c.features[f] for f in self.features])) return doc def feature_vector_length(self): raise NotImplementedError class ZeroMeanUnitVarianceMapper(FeatureMapper): def __init__(self, features, means, stds): super(ZeroMeanUnitVarianceMapper,self).__init__(features) self.mean = means self.std = stds def map(self, fv): return (fv - self.mean) / self.std def feature_vector_length(self): return len(self.features) class PolynomialMapper(ZeroMeanUnitVarianceMapper): def __init__(self, features, means, stds): super(PolynomialMapper,self).__init__(features, means, stds) def map(self, fv): fv = list(super(PolynomialMapper, self).map(fv)) sz = len(fv) for i in xrange(0, sz): for j in xrange(i, sz): weight = 1.0 if i != j else math.sqrt(2.0) fv.append(weight * fv[i]*fv[j]) return numpy.array(fv) def feature_vector_length(self): n = len(self.features) return n + n*(n+1)/2 FEATURE_MAPPERS = {cls.__name__:cls for cls in [ZeroMeanUnitVarianceMapper,PolynomialMapper]}
Add feature vector size calculation method to mapper interface
Add feature vector size calculation method to mapper interface
Python
mit
wikilinks/nel,wikilinks/nel
import math import numpy class FeatureMapper(object): def __init__(self, features): self.features = features def map(self, fv): raise NotImplementedError def __call__(self, doc): for chain in doc.chains: for c in chain.candidates: c.fv = self.map(numpy.array([c.features[f] for f in self.features])) return doc class ZeroMeanUnitVarianceMapper(FeatureMapper): def __init__(self, features, means, stds): super(ZeroMeanUnitVarianceMapper,self).__init__(features) self.mean = means self.std = stds def map(self, fv): return (fv - self.mean) / self.std class PolynomialMapper(ZeroMeanUnitVarianceMapper): def __init__(self, features, means, stds): super(PolynomialMapper,self).__init__(features, means, stds) def map(self, fv): fv = list(super(PolynomialMapper, self).map(fv)) sz = len(fv) for i in xrange(0, sz): for j in xrange(i, sz): weight = 1.0 if i != j else math.sqrt(2.0) fv.append(weight * fv[i]*fv[j]) return numpy.array(fv) FEATURE_MAPPERS = {cls.__name__:cls for cls in [ZeroMeanUnitVarianceMapper,PolynomialMapper]} Add feature vector size calculation method to mapper interface
import math import numpy class FeatureMapper(object): def __init__(self, features): self.features = features def map(self, fv): raise NotImplementedError def __call__(self, doc): for chain in doc.chains: for c in chain.candidates: c.fv = self.map(numpy.array([c.features[f] for f in self.features])) return doc def feature_vector_length(self): raise NotImplementedError class ZeroMeanUnitVarianceMapper(FeatureMapper): def __init__(self, features, means, stds): super(ZeroMeanUnitVarianceMapper,self).__init__(features) self.mean = means self.std = stds def map(self, fv): return (fv - self.mean) / self.std def feature_vector_length(self): return len(self.features) class PolynomialMapper(ZeroMeanUnitVarianceMapper): def __init__(self, features, means, stds): super(PolynomialMapper,self).__init__(features, means, stds) def map(self, fv): fv = list(super(PolynomialMapper, self).map(fv)) sz = len(fv) for i in xrange(0, sz): for j in xrange(i, sz): weight = 1.0 if i != j else math.sqrt(2.0) fv.append(weight * fv[i]*fv[j]) return numpy.array(fv) def feature_vector_length(self): n = len(self.features) return n + n*(n+1)/2 FEATURE_MAPPERS = {cls.__name__:cls for cls in [ZeroMeanUnitVarianceMapper,PolynomialMapper]}
<commit_before>import math import numpy class FeatureMapper(object): def __init__(self, features): self.features = features def map(self, fv): raise NotImplementedError def __call__(self, doc): for chain in doc.chains: for c in chain.candidates: c.fv = self.map(numpy.array([c.features[f] for f in self.features])) return doc class ZeroMeanUnitVarianceMapper(FeatureMapper): def __init__(self, features, means, stds): super(ZeroMeanUnitVarianceMapper,self).__init__(features) self.mean = means self.std = stds def map(self, fv): return (fv - self.mean) / self.std class PolynomialMapper(ZeroMeanUnitVarianceMapper): def __init__(self, features, means, stds): super(PolynomialMapper,self).__init__(features, means, stds) def map(self, fv): fv = list(super(PolynomialMapper, self).map(fv)) sz = len(fv) for i in xrange(0, sz): for j in xrange(i, sz): weight = 1.0 if i != j else math.sqrt(2.0) fv.append(weight * fv[i]*fv[j]) return numpy.array(fv) FEATURE_MAPPERS = {cls.__name__:cls for cls in [ZeroMeanUnitVarianceMapper,PolynomialMapper]} <commit_msg>Add feature vector size calculation method to mapper interface<commit_after>
import math import numpy class FeatureMapper(object): def __init__(self, features): self.features = features def map(self, fv): raise NotImplementedError def __call__(self, doc): for chain in doc.chains: for c in chain.candidates: c.fv = self.map(numpy.array([c.features[f] for f in self.features])) return doc def feature_vector_length(self): raise NotImplementedError class ZeroMeanUnitVarianceMapper(FeatureMapper): def __init__(self, features, means, stds): super(ZeroMeanUnitVarianceMapper,self).__init__(features) self.mean = means self.std = stds def map(self, fv): return (fv - self.mean) / self.std def feature_vector_length(self): return len(self.features) class PolynomialMapper(ZeroMeanUnitVarianceMapper): def __init__(self, features, means, stds): super(PolynomialMapper,self).__init__(features, means, stds) def map(self, fv): fv = list(super(PolynomialMapper, self).map(fv)) sz = len(fv) for i in xrange(0, sz): for j in xrange(i, sz): weight = 1.0 if i != j else math.sqrt(2.0) fv.append(weight * fv[i]*fv[j]) return numpy.array(fv) def feature_vector_length(self): n = len(self.features) return n + n*(n+1)/2 FEATURE_MAPPERS = {cls.__name__:cls for cls in [ZeroMeanUnitVarianceMapper,PolynomialMapper]}
import math import numpy class FeatureMapper(object): def __init__(self, features): self.features = features def map(self, fv): raise NotImplementedError def __call__(self, doc): for chain in doc.chains: for c in chain.candidates: c.fv = self.map(numpy.array([c.features[f] for f in self.features])) return doc class ZeroMeanUnitVarianceMapper(FeatureMapper): def __init__(self, features, means, stds): super(ZeroMeanUnitVarianceMapper,self).__init__(features) self.mean = means self.std = stds def map(self, fv): return (fv - self.mean) / self.std class PolynomialMapper(ZeroMeanUnitVarianceMapper): def __init__(self, features, means, stds): super(PolynomialMapper,self).__init__(features, means, stds) def map(self, fv): fv = list(super(PolynomialMapper, self).map(fv)) sz = len(fv) for i in xrange(0, sz): for j in xrange(i, sz): weight = 1.0 if i != j else math.sqrt(2.0) fv.append(weight * fv[i]*fv[j]) return numpy.array(fv) FEATURE_MAPPERS = {cls.__name__:cls for cls in [ZeroMeanUnitVarianceMapper,PolynomialMapper]} Add feature vector size calculation method to mapper interfaceimport math import numpy class FeatureMapper(object): def __init__(self, features): self.features = features def map(self, fv): raise NotImplementedError def __call__(self, doc): for chain in doc.chains: for c in chain.candidates: c.fv = self.map(numpy.array([c.features[f] for f in self.features])) return doc def feature_vector_length(self): raise NotImplementedError class ZeroMeanUnitVarianceMapper(FeatureMapper): def __init__(self, features, means, stds): super(ZeroMeanUnitVarianceMapper,self).__init__(features) self.mean = means self.std = stds def map(self, fv): return (fv - self.mean) / self.std def feature_vector_length(self): return len(self.features) class PolynomialMapper(ZeroMeanUnitVarianceMapper): def __init__(self, features, means, stds): super(PolynomialMapper,self).__init__(features, means, stds) def map(self, fv): fv = list(super(PolynomialMapper, self).map(fv)) sz = len(fv) for i in xrange(0, sz): for j in xrange(i, sz): weight = 1.0 if i != j else math.sqrt(2.0) fv.append(weight * fv[i]*fv[j]) return numpy.array(fv) def feature_vector_length(self): n = len(self.features) return n + n*(n+1)/2 FEATURE_MAPPERS = {cls.__name__:cls for cls in [ZeroMeanUnitVarianceMapper,PolynomialMapper]}
<commit_before>import math import numpy class FeatureMapper(object): def __init__(self, features): self.features = features def map(self, fv): raise NotImplementedError def __call__(self, doc): for chain in doc.chains: for c in chain.candidates: c.fv = self.map(numpy.array([c.features[f] for f in self.features])) return doc class ZeroMeanUnitVarianceMapper(FeatureMapper): def __init__(self, features, means, stds): super(ZeroMeanUnitVarianceMapper,self).__init__(features) self.mean = means self.std = stds def map(self, fv): return (fv - self.mean) / self.std class PolynomialMapper(ZeroMeanUnitVarianceMapper): def __init__(self, features, means, stds): super(PolynomialMapper,self).__init__(features, means, stds) def map(self, fv): fv = list(super(PolynomialMapper, self).map(fv)) sz = len(fv) for i in xrange(0, sz): for j in xrange(i, sz): weight = 1.0 if i != j else math.sqrt(2.0) fv.append(weight * fv[i]*fv[j]) return numpy.array(fv) FEATURE_MAPPERS = {cls.__name__:cls for cls in [ZeroMeanUnitVarianceMapper,PolynomialMapper]} <commit_msg>Add feature vector size calculation method to mapper interface<commit_after>import math import numpy class FeatureMapper(object): def __init__(self, features): self.features = features def map(self, fv): raise NotImplementedError def __call__(self, doc): for chain in doc.chains: for c in chain.candidates: c.fv = self.map(numpy.array([c.features[f] for f in self.features])) return doc def feature_vector_length(self): raise NotImplementedError class ZeroMeanUnitVarianceMapper(FeatureMapper): def __init__(self, features, means, stds): super(ZeroMeanUnitVarianceMapper,self).__init__(features) self.mean = means self.std = stds def map(self, fv): return (fv - self.mean) / self.std def feature_vector_length(self): return len(self.features) class PolynomialMapper(ZeroMeanUnitVarianceMapper): def __init__(self, features, means, stds): super(PolynomialMapper,self).__init__(features, means, stds) def map(self, fv): fv = list(super(PolynomialMapper, self).map(fv)) sz = len(fv) for i in xrange(0, sz): for j in xrange(i, sz): weight = 1.0 if i != j else math.sqrt(2.0) fv.append(weight * fv[i]*fv[j]) return numpy.array(fv) def feature_vector_length(self): n = len(self.features) return n + n*(n+1)/2 FEATURE_MAPPERS = {cls.__name__:cls for cls in [ZeroMeanUnitVarianceMapper,PolynomialMapper]}
141ad63b28eff5c7a034d479b98c83334ff1f0a3
provokator/site/util.py
provokator/site/util.py
#!/usr/bin/python3 -tt # -*- coding: utf-8 -*- __all__ = ['internal_origin_only'] from urllib.parse import urlparse from functools import wraps from werkzeug.exceptions import Forbidden import flask import re def internal_origin_only(fn): @wraps(fn) def wrapper(*args, **kwargs): h = urlparse('http://' + flask.request.headers.get('Host', '')) host = '%s:%i' % (h.hostname, h.port or 80) if 'Origin' in flask.request.headers: o = urlparse(flask.request.headers.get('Origin')) origin = '%s:%i' % (o.hostname, o.port or 80) elif 'Referer' in flask.request.headers: r = urlparse(flask.request.headers.get('Referer')) origin = '%s:%i' % (r.hostname, r.port or 80) else: origin = host if host != origin: raise Forbidden('Cross-Site Request Forbidden') return fn(*args, **kwargs) return wrapper # vim:set sw=4 ts=4 et:
#!/usr/bin/python3 -tt # -*- coding: utf-8 -*- __all__ = ['internal_origin_only'] from urllib.parse import urlparse from functools import wraps from werkzeug.exceptions import Forbidden import flask import re def internal_origin_only(fn): @wraps(fn) def wrapper(*args, **kwargs): host = flask.request.headers.get('X-Forwarded-Host') or \ flask.request.headers.get('Host', '') h = urlparse('http://' + host) origin = flask.request.headers.get('Origin') or \ flask.request.headers.get('Referer') or \ host o = urlparse(origin) if h.hostname != o.hostname: raise Forbidden('Cross-Site Request Forbidden') return fn(*args, **kwargs) return wrapper # vim:set sw=4 ts=4 et:
Change cross-origin check to work behind proxies
Change cross-origin check to work behind proxies Signed-off-by: Jan Dvořák <86df5a4870880bf501c926309e3bcfbe57789f3f@anilinux.org>
Python
mit
techlib/provokator,techlib/provokator
#!/usr/bin/python3 -tt # -*- coding: utf-8 -*- __all__ = ['internal_origin_only'] from urllib.parse import urlparse from functools import wraps from werkzeug.exceptions import Forbidden import flask import re def internal_origin_only(fn): @wraps(fn) def wrapper(*args, **kwargs): h = urlparse('http://' + flask.request.headers.get('Host', '')) host = '%s:%i' % (h.hostname, h.port or 80) if 'Origin' in flask.request.headers: o = urlparse(flask.request.headers.get('Origin')) origin = '%s:%i' % (o.hostname, o.port or 80) elif 'Referer' in flask.request.headers: r = urlparse(flask.request.headers.get('Referer')) origin = '%s:%i' % (r.hostname, r.port or 80) else: origin = host if host != origin: raise Forbidden('Cross-Site Request Forbidden') return fn(*args, **kwargs) return wrapper # vim:set sw=4 ts=4 et: Change cross-origin check to work behind proxies Signed-off-by: Jan Dvořák <86df5a4870880bf501c926309e3bcfbe57789f3f@anilinux.org>
#!/usr/bin/python3 -tt # -*- coding: utf-8 -*- __all__ = ['internal_origin_only'] from urllib.parse import urlparse from functools import wraps from werkzeug.exceptions import Forbidden import flask import re def internal_origin_only(fn): @wraps(fn) def wrapper(*args, **kwargs): host = flask.request.headers.get('X-Forwarded-Host') or \ flask.request.headers.get('Host', '') h = urlparse('http://' + host) origin = flask.request.headers.get('Origin') or \ flask.request.headers.get('Referer') or \ host o = urlparse(origin) if h.hostname != o.hostname: raise Forbidden('Cross-Site Request Forbidden') return fn(*args, **kwargs) return wrapper # vim:set sw=4 ts=4 et:
<commit_before>#!/usr/bin/python3 -tt # -*- coding: utf-8 -*- __all__ = ['internal_origin_only'] from urllib.parse import urlparse from functools import wraps from werkzeug.exceptions import Forbidden import flask import re def internal_origin_only(fn): @wraps(fn) def wrapper(*args, **kwargs): h = urlparse('http://' + flask.request.headers.get('Host', '')) host = '%s:%i' % (h.hostname, h.port or 80) if 'Origin' in flask.request.headers: o = urlparse(flask.request.headers.get('Origin')) origin = '%s:%i' % (o.hostname, o.port or 80) elif 'Referer' in flask.request.headers: r = urlparse(flask.request.headers.get('Referer')) origin = '%s:%i' % (r.hostname, r.port or 80) else: origin = host if host != origin: raise Forbidden('Cross-Site Request Forbidden') return fn(*args, **kwargs) return wrapper # vim:set sw=4 ts=4 et: <commit_msg>Change cross-origin check to work behind proxies Signed-off-by: Jan Dvořák <86df5a4870880bf501c926309e3bcfbe57789f3f@anilinux.org><commit_after>
#!/usr/bin/python3 -tt # -*- coding: utf-8 -*- __all__ = ['internal_origin_only'] from urllib.parse import urlparse from functools import wraps from werkzeug.exceptions import Forbidden import flask import re def internal_origin_only(fn): @wraps(fn) def wrapper(*args, **kwargs): host = flask.request.headers.get('X-Forwarded-Host') or \ flask.request.headers.get('Host', '') h = urlparse('http://' + host) origin = flask.request.headers.get('Origin') or \ flask.request.headers.get('Referer') or \ host o = urlparse(origin) if h.hostname != o.hostname: raise Forbidden('Cross-Site Request Forbidden') return fn(*args, **kwargs) return wrapper # vim:set sw=4 ts=4 et:
#!/usr/bin/python3 -tt # -*- coding: utf-8 -*- __all__ = ['internal_origin_only'] from urllib.parse import urlparse from functools import wraps from werkzeug.exceptions import Forbidden import flask import re def internal_origin_only(fn): @wraps(fn) def wrapper(*args, **kwargs): h = urlparse('http://' + flask.request.headers.get('Host', '')) host = '%s:%i' % (h.hostname, h.port or 80) if 'Origin' in flask.request.headers: o = urlparse(flask.request.headers.get('Origin')) origin = '%s:%i' % (o.hostname, o.port or 80) elif 'Referer' in flask.request.headers: r = urlparse(flask.request.headers.get('Referer')) origin = '%s:%i' % (r.hostname, r.port or 80) else: origin = host if host != origin: raise Forbidden('Cross-Site Request Forbidden') return fn(*args, **kwargs) return wrapper # vim:set sw=4 ts=4 et: Change cross-origin check to work behind proxies Signed-off-by: Jan Dvořák <86df5a4870880bf501c926309e3bcfbe57789f3f@anilinux.org>#!/usr/bin/python3 -tt # -*- coding: utf-8 -*- __all__ = ['internal_origin_only'] from urllib.parse import urlparse from functools import wraps from werkzeug.exceptions import Forbidden import flask import re def internal_origin_only(fn): @wraps(fn) def wrapper(*args, **kwargs): host = flask.request.headers.get('X-Forwarded-Host') or \ flask.request.headers.get('Host', '') h = urlparse('http://' + host) origin = flask.request.headers.get('Origin') or \ flask.request.headers.get('Referer') or \ host o = urlparse(origin) if h.hostname != o.hostname: raise Forbidden('Cross-Site Request Forbidden') return fn(*args, **kwargs) return wrapper # vim:set sw=4 ts=4 et:
<commit_before>#!/usr/bin/python3 -tt # -*- coding: utf-8 -*- __all__ = ['internal_origin_only'] from urllib.parse import urlparse from functools import wraps from werkzeug.exceptions import Forbidden import flask import re def internal_origin_only(fn): @wraps(fn) def wrapper(*args, **kwargs): h = urlparse('http://' + flask.request.headers.get('Host', '')) host = '%s:%i' % (h.hostname, h.port or 80) if 'Origin' in flask.request.headers: o = urlparse(flask.request.headers.get('Origin')) origin = '%s:%i' % (o.hostname, o.port or 80) elif 'Referer' in flask.request.headers: r = urlparse(flask.request.headers.get('Referer')) origin = '%s:%i' % (r.hostname, r.port or 80) else: origin = host if host != origin: raise Forbidden('Cross-Site Request Forbidden') return fn(*args, **kwargs) return wrapper # vim:set sw=4 ts=4 et: <commit_msg>Change cross-origin check to work behind proxies Signed-off-by: Jan Dvořák <86df5a4870880bf501c926309e3bcfbe57789f3f@anilinux.org><commit_after>#!/usr/bin/python3 -tt # -*- coding: utf-8 -*- __all__ = ['internal_origin_only'] from urllib.parse import urlparse from functools import wraps from werkzeug.exceptions import Forbidden import flask import re def internal_origin_only(fn): @wraps(fn) def wrapper(*args, **kwargs): host = flask.request.headers.get('X-Forwarded-Host') or \ flask.request.headers.get('Host', '') h = urlparse('http://' + host) origin = flask.request.headers.get('Origin') or \ flask.request.headers.get('Referer') or \ host o = urlparse(origin) if h.hostname != o.hostname: raise Forbidden('Cross-Site Request Forbidden') return fn(*args, **kwargs) return wrapper # vim:set sw=4 ts=4 et: