commit
stringlengths 40
40
| old_file
stringlengths 4
118
| new_file
stringlengths 4
118
| old_contents
stringlengths 0
2.94k
| new_contents
stringlengths 1
4.43k
| subject
stringlengths 15
444
| message
stringlengths 16
3.45k
| lang
stringclasses 1
value | license
stringclasses 13
values | repos
stringlengths 5
43.2k
| prompt
stringlengths 17
4.58k
| response
stringlengths 1
4.43k
| prompt_tagged
stringlengths 58
4.62k
| response_tagged
stringlengths 1
4.43k
| text
stringlengths 132
7.29k
| text_tagged
stringlengths 173
7.33k
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1cfdf9b1c11da15adb1e1603c815b76a4a286b1a
|
searchlogger/searchlogger/settings/production.py
|
searchlogger/searchlogger/settings/production.py
|
#! /usr/bin/env python
# -*- coding: utf-8 -*-
from defaults import * # noqa
import json
DEBUG = False
TEMPLATE_DEBUG = False
ALLOWED_HOSTS = ['.searchlogger.tutorons.com']
# Read in the Postgres database configuration from a file
DATABASE_CONFIG_FILENAME = os.path.join(
os.path.abspath(os.sep), # root directory
'etc', 'django', 'searchlogger', 'database_config.json'
)
with open(DATABASE_CONFIG_FILENAME) as database_config_file:
database_config = json.load(databse_config_file)
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.postgresql',
'NAME': database_config['name'],
'USER': database_config['user'],
'PASSWORD': database_config['password'],
'HOST': database_config['host'],
'PORT': database_config['port'],
}
}
|
#! /usr/bin/env python
# -*- coding: utf-8 -*-
from defaults import * # noqa
import json
DEBUG = False
TEMPLATE_DEBUG = False
ALLOWED_HOSTS = ['.searchlogger.tutorons.com']
# Read in the Postgres database configuration from a file
DATABASE_CONFIG_FILENAME = os.path.join(BASE_DIR, 'database_config.json')
with open(DATABASE_CONFIG_FILENAME) as database_config_file:
database_config = json.load(databse_config_file)
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.postgresql',
'NAME': database_config['name'],
'USER': database_config['user'],
'PASSWORD': database_config['password'],
'HOST': database_config['host'],
'PORT': database_config['port'],
}
}
|
Read database configuration from base directory
|
Settings: Read database configuration from base directory
|
Python
|
mit
|
andrewhead/Search-Task-Logger,andrewhead/Search-Task-Logger,andrewhead/Search-Task-Logger
|
#! /usr/bin/env python
# -*- coding: utf-8 -*-
from defaults import * # noqa
import json
DEBUG = False
TEMPLATE_DEBUG = False
ALLOWED_HOSTS = ['.searchlogger.tutorons.com']
# Read in the Postgres database configuration from a file
DATABASE_CONFIG_FILENAME = os.path.join(
os.path.abspath(os.sep), # root directory
'etc', 'django', 'searchlogger', 'database_config.json'
)
with open(DATABASE_CONFIG_FILENAME) as database_config_file:
database_config = json.load(databse_config_file)
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.postgresql',
'NAME': database_config['name'],
'USER': database_config['user'],
'PASSWORD': database_config['password'],
'HOST': database_config['host'],
'PORT': database_config['port'],
}
}
Settings: Read database configuration from base directory
|
#! /usr/bin/env python
# -*- coding: utf-8 -*-
from defaults import * # noqa
import json
DEBUG = False
TEMPLATE_DEBUG = False
ALLOWED_HOSTS = ['.searchlogger.tutorons.com']
# Read in the Postgres database configuration from a file
DATABASE_CONFIG_FILENAME = os.path.join(BASE_DIR, 'database_config.json')
with open(DATABASE_CONFIG_FILENAME) as database_config_file:
database_config = json.load(databse_config_file)
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.postgresql',
'NAME': database_config['name'],
'USER': database_config['user'],
'PASSWORD': database_config['password'],
'HOST': database_config['host'],
'PORT': database_config['port'],
}
}
|
<commit_before>#! /usr/bin/env python
# -*- coding: utf-8 -*-
from defaults import * # noqa
import json
DEBUG = False
TEMPLATE_DEBUG = False
ALLOWED_HOSTS = ['.searchlogger.tutorons.com']
# Read in the Postgres database configuration from a file
DATABASE_CONFIG_FILENAME = os.path.join(
os.path.abspath(os.sep), # root directory
'etc', 'django', 'searchlogger', 'database_config.json'
)
with open(DATABASE_CONFIG_FILENAME) as database_config_file:
database_config = json.load(databse_config_file)
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.postgresql',
'NAME': database_config['name'],
'USER': database_config['user'],
'PASSWORD': database_config['password'],
'HOST': database_config['host'],
'PORT': database_config['port'],
}
}
<commit_msg>Settings: Read database configuration from base directory<commit_after>
|
#! /usr/bin/env python
# -*- coding: utf-8 -*-
from defaults import * # noqa
import json
DEBUG = False
TEMPLATE_DEBUG = False
ALLOWED_HOSTS = ['.searchlogger.tutorons.com']
# Read in the Postgres database configuration from a file
DATABASE_CONFIG_FILENAME = os.path.join(BASE_DIR, 'database_config.json')
with open(DATABASE_CONFIG_FILENAME) as database_config_file:
database_config = json.load(databse_config_file)
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.postgresql',
'NAME': database_config['name'],
'USER': database_config['user'],
'PASSWORD': database_config['password'],
'HOST': database_config['host'],
'PORT': database_config['port'],
}
}
|
#! /usr/bin/env python
# -*- coding: utf-8 -*-
from defaults import * # noqa
import json
DEBUG = False
TEMPLATE_DEBUG = False
ALLOWED_HOSTS = ['.searchlogger.tutorons.com']
# Read in the Postgres database configuration from a file
DATABASE_CONFIG_FILENAME = os.path.join(
os.path.abspath(os.sep), # root directory
'etc', 'django', 'searchlogger', 'database_config.json'
)
with open(DATABASE_CONFIG_FILENAME) as database_config_file:
database_config = json.load(databse_config_file)
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.postgresql',
'NAME': database_config['name'],
'USER': database_config['user'],
'PASSWORD': database_config['password'],
'HOST': database_config['host'],
'PORT': database_config['port'],
}
}
Settings: Read database configuration from base directory#! /usr/bin/env python
# -*- coding: utf-8 -*-
from defaults import * # noqa
import json
DEBUG = False
TEMPLATE_DEBUG = False
ALLOWED_HOSTS = ['.searchlogger.tutorons.com']
# Read in the Postgres database configuration from a file
DATABASE_CONFIG_FILENAME = os.path.join(BASE_DIR, 'database_config.json')
with open(DATABASE_CONFIG_FILENAME) as database_config_file:
database_config = json.load(databse_config_file)
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.postgresql',
'NAME': database_config['name'],
'USER': database_config['user'],
'PASSWORD': database_config['password'],
'HOST': database_config['host'],
'PORT': database_config['port'],
}
}
|
<commit_before>#! /usr/bin/env python
# -*- coding: utf-8 -*-
from defaults import * # noqa
import json
DEBUG = False
TEMPLATE_DEBUG = False
ALLOWED_HOSTS = ['.searchlogger.tutorons.com']
# Read in the Postgres database configuration from a file
DATABASE_CONFIG_FILENAME = os.path.join(
os.path.abspath(os.sep), # root directory
'etc', 'django', 'searchlogger', 'database_config.json'
)
with open(DATABASE_CONFIG_FILENAME) as database_config_file:
database_config = json.load(databse_config_file)
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.postgresql',
'NAME': database_config['name'],
'USER': database_config['user'],
'PASSWORD': database_config['password'],
'HOST': database_config['host'],
'PORT': database_config['port'],
}
}
<commit_msg>Settings: Read database configuration from base directory<commit_after>#! /usr/bin/env python
# -*- coding: utf-8 -*-
from defaults import * # noqa
import json
DEBUG = False
TEMPLATE_DEBUG = False
ALLOWED_HOSTS = ['.searchlogger.tutorons.com']
# Read in the Postgres database configuration from a file
DATABASE_CONFIG_FILENAME = os.path.join(BASE_DIR, 'database_config.json')
with open(DATABASE_CONFIG_FILENAME) as database_config_file:
database_config = json.load(databse_config_file)
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.postgresql',
'NAME': database_config['name'],
'USER': database_config['user'],
'PASSWORD': database_config['password'],
'HOST': database_config['host'],
'PORT': database_config['port'],
}
}
|
14ff06097a72dc65a351bb6a8bf59963412d2f41
|
semillas_backend/users/serializers.py
|
semillas_backend/users/serializers.py
|
#from phonenumber_field.serializerfields import PhoneNumberField
from rest_framework import serializers
from drf_extra_fields.geo_fields import PointField
from .models import User
class UserSerializer(serializers.ModelSerializer):
""" Usage:
from rest_framework.renderers import JSONRenderer
from semillas_backend.users.serializers import UserSerializer
JSONRenderer().render(UserSerializer(user_instance).data)
"""
location = PointField()
class Meta:
model = User
fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login')
class UpdateUserSerializer(serializers.ModelSerializer):
name = serializers.CharField(required=False)
#phone = PhoneNumberField(required=False)
email = serializers.CharField(required=False)
picture = serializers.ImageField(required=False)
class Meta:
model = User
fields = ('name', 'picture', 'phone', 'email')
from wallet.serializers import WalletSerializer
class FullUserSerializer(UserSerializer):
wallet = WalletSerializer()
class Meta:
model = User
fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'wallet', 'email')
|
#from phonenumber_field.serializerfields import PhoneNumberField
from rest_framework import serializers
from drf_extra_fields.geo_fields import PointField
from .models import User
class UserSerializer(serializers.ModelSerializer):
""" Usage:
from rest_framework.renderers import JSONRenderer
from semillas_backend.users.serializers import UserSerializer
JSONRenderer().render(UserSerializer(user_instance).data)
"""
location = PointField()
class Meta:
model = User
fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'phone')
class UpdateUserSerializer(serializers.ModelSerializer):
name = serializers.CharField(required=False)
#phone = PhoneNumberField(required=False)
email = serializers.CharField(required=False)
picture = serializers.ImageField(required=False)
class Meta:
model = User
fields = ('name', 'picture', 'phone', 'email')
from wallet.serializers import WalletSerializer
class FullUserSerializer(UserSerializer):
wallet = WalletSerializer()
class Meta:
model = User
fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'wallet', 'email')
|
Add phone to user serializer
|
Add phone to user serializer
|
Python
|
mit
|
Semillas/semillas_platform,Semillas/semillas_backend,Semillas/semillas_backend,Semillas/semillas_backend,Semillas/semillas_platform,Semillas/semillas_platform,Semillas/semillas_platform,Semillas/semillas_backend
|
#from phonenumber_field.serializerfields import PhoneNumberField
from rest_framework import serializers
from drf_extra_fields.geo_fields import PointField
from .models import User
class UserSerializer(serializers.ModelSerializer):
""" Usage:
from rest_framework.renderers import JSONRenderer
from semillas_backend.users.serializers import UserSerializer
JSONRenderer().render(UserSerializer(user_instance).data)
"""
location = PointField()
class Meta:
model = User
fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login')
class UpdateUserSerializer(serializers.ModelSerializer):
name = serializers.CharField(required=False)
#phone = PhoneNumberField(required=False)
email = serializers.CharField(required=False)
picture = serializers.ImageField(required=False)
class Meta:
model = User
fields = ('name', 'picture', 'phone', 'email')
from wallet.serializers import WalletSerializer
class FullUserSerializer(UserSerializer):
wallet = WalletSerializer()
class Meta:
model = User
fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'wallet', 'email')
Add phone to user serializer
|
#from phonenumber_field.serializerfields import PhoneNumberField
from rest_framework import serializers
from drf_extra_fields.geo_fields import PointField
from .models import User
class UserSerializer(serializers.ModelSerializer):
""" Usage:
from rest_framework.renderers import JSONRenderer
from semillas_backend.users.serializers import UserSerializer
JSONRenderer().render(UserSerializer(user_instance).data)
"""
location = PointField()
class Meta:
model = User
fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'phone')
class UpdateUserSerializer(serializers.ModelSerializer):
name = serializers.CharField(required=False)
#phone = PhoneNumberField(required=False)
email = serializers.CharField(required=False)
picture = serializers.ImageField(required=False)
class Meta:
model = User
fields = ('name', 'picture', 'phone', 'email')
from wallet.serializers import WalletSerializer
class FullUserSerializer(UserSerializer):
wallet = WalletSerializer()
class Meta:
model = User
fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'wallet', 'email')
|
<commit_before>#from phonenumber_field.serializerfields import PhoneNumberField
from rest_framework import serializers
from drf_extra_fields.geo_fields import PointField
from .models import User
class UserSerializer(serializers.ModelSerializer):
""" Usage:
from rest_framework.renderers import JSONRenderer
from semillas_backend.users.serializers import UserSerializer
JSONRenderer().render(UserSerializer(user_instance).data)
"""
location = PointField()
class Meta:
model = User
fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login')
class UpdateUserSerializer(serializers.ModelSerializer):
name = serializers.CharField(required=False)
#phone = PhoneNumberField(required=False)
email = serializers.CharField(required=False)
picture = serializers.ImageField(required=False)
class Meta:
model = User
fields = ('name', 'picture', 'phone', 'email')
from wallet.serializers import WalletSerializer
class FullUserSerializer(UserSerializer):
wallet = WalletSerializer()
class Meta:
model = User
fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'wallet', 'email')
<commit_msg>Add phone to user serializer<commit_after>
|
#from phonenumber_field.serializerfields import PhoneNumberField
from rest_framework import serializers
from drf_extra_fields.geo_fields import PointField
from .models import User
class UserSerializer(serializers.ModelSerializer):
""" Usage:
from rest_framework.renderers import JSONRenderer
from semillas_backend.users.serializers import UserSerializer
JSONRenderer().render(UserSerializer(user_instance).data)
"""
location = PointField()
class Meta:
model = User
fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'phone')
class UpdateUserSerializer(serializers.ModelSerializer):
name = serializers.CharField(required=False)
#phone = PhoneNumberField(required=False)
email = serializers.CharField(required=False)
picture = serializers.ImageField(required=False)
class Meta:
model = User
fields = ('name', 'picture', 'phone', 'email')
from wallet.serializers import WalletSerializer
class FullUserSerializer(UserSerializer):
wallet = WalletSerializer()
class Meta:
model = User
fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'wallet', 'email')
|
#from phonenumber_field.serializerfields import PhoneNumberField
from rest_framework import serializers
from drf_extra_fields.geo_fields import PointField
from .models import User
class UserSerializer(serializers.ModelSerializer):
""" Usage:
from rest_framework.renderers import JSONRenderer
from semillas_backend.users.serializers import UserSerializer
JSONRenderer().render(UserSerializer(user_instance).data)
"""
location = PointField()
class Meta:
model = User
fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login')
class UpdateUserSerializer(serializers.ModelSerializer):
name = serializers.CharField(required=False)
#phone = PhoneNumberField(required=False)
email = serializers.CharField(required=False)
picture = serializers.ImageField(required=False)
class Meta:
model = User
fields = ('name', 'picture', 'phone', 'email')
from wallet.serializers import WalletSerializer
class FullUserSerializer(UserSerializer):
wallet = WalletSerializer()
class Meta:
model = User
fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'wallet', 'email')
Add phone to user serializer#from phonenumber_field.serializerfields import PhoneNumberField
from rest_framework import serializers
from drf_extra_fields.geo_fields import PointField
from .models import User
class UserSerializer(serializers.ModelSerializer):
""" Usage:
from rest_framework.renderers import JSONRenderer
from semillas_backend.users.serializers import UserSerializer
JSONRenderer().render(UserSerializer(user_instance).data)
"""
location = PointField()
class Meta:
model = User
fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'phone')
class UpdateUserSerializer(serializers.ModelSerializer):
name = serializers.CharField(required=False)
#phone = PhoneNumberField(required=False)
email = serializers.CharField(required=False)
picture = serializers.ImageField(required=False)
class Meta:
model = User
fields = ('name', 'picture', 'phone', 'email')
from wallet.serializers import WalletSerializer
class FullUserSerializer(UserSerializer):
wallet = WalletSerializer()
class Meta:
model = User
fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'wallet', 'email')
|
<commit_before>#from phonenumber_field.serializerfields import PhoneNumberField
from rest_framework import serializers
from drf_extra_fields.geo_fields import PointField
from .models import User
class UserSerializer(serializers.ModelSerializer):
""" Usage:
from rest_framework.renderers import JSONRenderer
from semillas_backend.users.serializers import UserSerializer
JSONRenderer().render(UserSerializer(user_instance).data)
"""
location = PointField()
class Meta:
model = User
fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login')
class UpdateUserSerializer(serializers.ModelSerializer):
name = serializers.CharField(required=False)
#phone = PhoneNumberField(required=False)
email = serializers.CharField(required=False)
picture = serializers.ImageField(required=False)
class Meta:
model = User
fields = ('name', 'picture', 'phone', 'email')
from wallet.serializers import WalletSerializer
class FullUserSerializer(UserSerializer):
wallet = WalletSerializer()
class Meta:
model = User
fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'wallet', 'email')
<commit_msg>Add phone to user serializer<commit_after>#from phonenumber_field.serializerfields import PhoneNumberField
from rest_framework import serializers
from drf_extra_fields.geo_fields import PointField
from .models import User
class UserSerializer(serializers.ModelSerializer):
""" Usage:
from rest_framework.renderers import JSONRenderer
from semillas_backend.users.serializers import UserSerializer
JSONRenderer().render(UserSerializer(user_instance).data)
"""
location = PointField()
class Meta:
model = User
fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'phone')
class UpdateUserSerializer(serializers.ModelSerializer):
name = serializers.CharField(required=False)
#phone = PhoneNumberField(required=False)
email = serializers.CharField(required=False)
picture = serializers.ImageField(required=False)
class Meta:
model = User
fields = ('name', 'picture', 'phone', 'email')
from wallet.serializers import WalletSerializer
class FullUserSerializer(UserSerializer):
wallet = WalletSerializer()
class Meta:
model = User
fields = ('uuid', 'name', 'picture', 'location', 'username', 'last_login', 'wallet', 'email')
|
23d12b1c4b755c7d35406bf2428eefbd682ef68f
|
examples/xor-classifier.py
|
examples/xor-classifier.py
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
'''Example using the theanets package for learning the XOR relation.'''
import climate
import logging
import numpy as np
import theanets
climate.enable_default_logging()
X = np.array([[0.0, 0.0], [0.0, 1.0], [1.0, 0.0], [1.0, 1.0]])
Y = np.array([0, 1, 1, 0, ])
Xi = np.random.randint(0, 2, size=(256, 2))
train = [
(Xi + 0.1 * np.random.randn(*Xi.shape)).astype('f'),
(Xi[:, 0] ^ Xi[:, 1]).astype('f')[:, None],
]
e = theanets.Experiment(theanets.Regressor,
layers=(2, 2, 1),
learning_rate=0.1,
momentum=0.5,
patience=300)
e.run(train, train)
logging.info("Input:\n%s", X)
logging.info("XOR output:\n%s", Y)
logging.info("NN XOR predictions:\n%s", e.network(X.astype('f')))
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
'''Example using the theanets package for learning the XOR relation.'''
import climate
import logging
import numpy as np
import theanets
climate.enable_default_logging()
X = np.array([[0, 0], [0, 1], [1, 0], [1, 1]], dtype='f')
Y = np.array([[0], [1], [1], [0]], dtype='f')
e = theanets.Experiment(theanets.Regressor, layers=(2, 2, 1))
e.train([X, Y], optimize='rprop', min_improvement=0.2, patience=500)
logging.info("Input:\n%s", X)
logging.info("XOR output:\n%s", Y)
logging.info("NN XOR predictions:\n%s", e.network(X.astype('f')).round(2))
|
Use rprop for xor example.
|
Use rprop for xor example.
|
Python
|
mit
|
lmjohns3/theanets,devdoer/theanets,chrinide/theanets
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
'''Example using the theanets package for learning the XOR relation.'''
import climate
import logging
import numpy as np
import theanets
climate.enable_default_logging()
X = np.array([[0.0, 0.0], [0.0, 1.0], [1.0, 0.0], [1.0, 1.0]])
Y = np.array([0, 1, 1, 0, ])
Xi = np.random.randint(0, 2, size=(256, 2))
train = [
(Xi + 0.1 * np.random.randn(*Xi.shape)).astype('f'),
(Xi[:, 0] ^ Xi[:, 1]).astype('f')[:, None],
]
e = theanets.Experiment(theanets.Regressor,
layers=(2, 2, 1),
learning_rate=0.1,
momentum=0.5,
patience=300)
e.run(train, train)
logging.info("Input:\n%s", X)
logging.info("XOR output:\n%s", Y)
logging.info("NN XOR predictions:\n%s", e.network(X.astype('f')))
Use rprop for xor example.
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
'''Example using the theanets package for learning the XOR relation.'''
import climate
import logging
import numpy as np
import theanets
climate.enable_default_logging()
X = np.array([[0, 0], [0, 1], [1, 0], [1, 1]], dtype='f')
Y = np.array([[0], [1], [1], [0]], dtype='f')
e = theanets.Experiment(theanets.Regressor, layers=(2, 2, 1))
e.train([X, Y], optimize='rprop', min_improvement=0.2, patience=500)
logging.info("Input:\n%s", X)
logging.info("XOR output:\n%s", Y)
logging.info("NN XOR predictions:\n%s", e.network(X.astype('f')).round(2))
|
<commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
'''Example using the theanets package for learning the XOR relation.'''
import climate
import logging
import numpy as np
import theanets
climate.enable_default_logging()
X = np.array([[0.0, 0.0], [0.0, 1.0], [1.0, 0.0], [1.0, 1.0]])
Y = np.array([0, 1, 1, 0, ])
Xi = np.random.randint(0, 2, size=(256, 2))
train = [
(Xi + 0.1 * np.random.randn(*Xi.shape)).astype('f'),
(Xi[:, 0] ^ Xi[:, 1]).astype('f')[:, None],
]
e = theanets.Experiment(theanets.Regressor,
layers=(2, 2, 1),
learning_rate=0.1,
momentum=0.5,
patience=300)
e.run(train, train)
logging.info("Input:\n%s", X)
logging.info("XOR output:\n%s", Y)
logging.info("NN XOR predictions:\n%s", e.network(X.astype('f')))
<commit_msg>Use rprop for xor example.<commit_after>
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
'''Example using the theanets package for learning the XOR relation.'''
import climate
import logging
import numpy as np
import theanets
climate.enable_default_logging()
X = np.array([[0, 0], [0, 1], [1, 0], [1, 1]], dtype='f')
Y = np.array([[0], [1], [1], [0]], dtype='f')
e = theanets.Experiment(theanets.Regressor, layers=(2, 2, 1))
e.train([X, Y], optimize='rprop', min_improvement=0.2, patience=500)
logging.info("Input:\n%s", X)
logging.info("XOR output:\n%s", Y)
logging.info("NN XOR predictions:\n%s", e.network(X.astype('f')).round(2))
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
'''Example using the theanets package for learning the XOR relation.'''
import climate
import logging
import numpy as np
import theanets
climate.enable_default_logging()
X = np.array([[0.0, 0.0], [0.0, 1.0], [1.0, 0.0], [1.0, 1.0]])
Y = np.array([0, 1, 1, 0, ])
Xi = np.random.randint(0, 2, size=(256, 2))
train = [
(Xi + 0.1 * np.random.randn(*Xi.shape)).astype('f'),
(Xi[:, 0] ^ Xi[:, 1]).astype('f')[:, None],
]
e = theanets.Experiment(theanets.Regressor,
layers=(2, 2, 1),
learning_rate=0.1,
momentum=0.5,
patience=300)
e.run(train, train)
logging.info("Input:\n%s", X)
logging.info("XOR output:\n%s", Y)
logging.info("NN XOR predictions:\n%s", e.network(X.astype('f')))
Use rprop for xor example.#!/usr/bin/env python
# -*- coding: utf-8 -*-
'''Example using the theanets package for learning the XOR relation.'''
import climate
import logging
import numpy as np
import theanets
climate.enable_default_logging()
X = np.array([[0, 0], [0, 1], [1, 0], [1, 1]], dtype='f')
Y = np.array([[0], [1], [1], [0]], dtype='f')
e = theanets.Experiment(theanets.Regressor, layers=(2, 2, 1))
e.train([X, Y], optimize='rprop', min_improvement=0.2, patience=500)
logging.info("Input:\n%s", X)
logging.info("XOR output:\n%s", Y)
logging.info("NN XOR predictions:\n%s", e.network(X.astype('f')).round(2))
|
<commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
'''Example using the theanets package for learning the XOR relation.'''
import climate
import logging
import numpy as np
import theanets
climate.enable_default_logging()
X = np.array([[0.0, 0.0], [0.0, 1.0], [1.0, 0.0], [1.0, 1.0]])
Y = np.array([0, 1, 1, 0, ])
Xi = np.random.randint(0, 2, size=(256, 2))
train = [
(Xi + 0.1 * np.random.randn(*Xi.shape)).astype('f'),
(Xi[:, 0] ^ Xi[:, 1]).astype('f')[:, None],
]
e = theanets.Experiment(theanets.Regressor,
layers=(2, 2, 1),
learning_rate=0.1,
momentum=0.5,
patience=300)
e.run(train, train)
logging.info("Input:\n%s", X)
logging.info("XOR output:\n%s", Y)
logging.info("NN XOR predictions:\n%s", e.network(X.astype('f')))
<commit_msg>Use rprop for xor example.<commit_after>#!/usr/bin/env python
# -*- coding: utf-8 -*-
'''Example using the theanets package for learning the XOR relation.'''
import climate
import logging
import numpy as np
import theanets
climate.enable_default_logging()
X = np.array([[0, 0], [0, 1], [1, 0], [1, 1]], dtype='f')
Y = np.array([[0], [1], [1], [0]], dtype='f')
e = theanets.Experiment(theanets.Regressor, layers=(2, 2, 1))
e.train([X, Y], optimize='rprop', min_improvement=0.2, patience=500)
logging.info("Input:\n%s", X)
logging.info("XOR output:\n%s", Y)
logging.info("NN XOR predictions:\n%s", e.network(X.astype('f')).round(2))
|
94796ca0107e6c676e3905675290bbe147169717
|
hoppy/deploy.py
|
hoppy/deploy.py
|
from restkit import Resource
from hoppy import api_key
class Deploy(Resource):
def __init__(self, use_ssl=False):
self.api_key = api_key
super(Deploy, self).__init__(self.host, follow_redirect=True)
def check_configuration(self):
if not self.api_key:
raise HoptoadError('API Key cannot be blank')
def request(self, *args, **kwargs):
response = super(Deploy, self).request(
api_key=self.api_key, *args, **kwargs)
return response.body_string()
def base_uri(self, use_ssl=False):
base = 'http://hoptoadapp.com/deploys.txt'
base = base.replace('http://', 'https://') if use_ssl else base
return base
def deploy(self, env, **kwargs):
""" Optional parameters accepted by Hoptoad are:
scm_revision
scm_repository
local_username
"""
params = {}
params['deploy[rails_env]'] = env
for key, value in kwargs:
params['deploy[%s]' % key] = value
return self.post(**params)
|
from hoppy.api import HoptoadResource
class Deploy(HoptoadResource):
def __init__(self, use_ssl=False):
from hoppy import api_key
self.api_key = api_key
super(Deploy, self).__init__(use_ssl)
def check_configuration(self):
if not self.api_key:
raise HoptoadError('API Key cannot be blank')
def request(self, *args, **kwargs):
response = super(Deploy, self).request(
api_key=self.api_key, *args, **kwargs)
return response
def base_uri(self, use_ssl=False):
base = 'http://hoptoadapp.com/deploys.txt'
base = base.replace('http://', 'https://') if use_ssl else base
return base
def deploy(self, env, **kwargs):
""" Optional parameters accepted by Hoptoad are:
scm_revision
scm_repository
local_username
"""
params = {}
params['deploy[rails_env]'] = env
for key, value in kwargs.iteritems():
params['deploy[%s]' % key] = value
return self.post(**params)
|
Test Deploy resource after reworking.
|
Test Deploy resource after reworking.
|
Python
|
mit
|
peplin/hoppy
|
from restkit import Resource
from hoppy import api_key
class Deploy(Resource):
def __init__(self, use_ssl=False):
self.api_key = api_key
super(Deploy, self).__init__(self.host, follow_redirect=True)
def check_configuration(self):
if not self.api_key:
raise HoptoadError('API Key cannot be blank')
def request(self, *args, **kwargs):
response = super(Deploy, self).request(
api_key=self.api_key, *args, **kwargs)
return response.body_string()
def base_uri(self, use_ssl=False):
base = 'http://hoptoadapp.com/deploys.txt'
base = base.replace('http://', 'https://') if use_ssl else base
return base
def deploy(self, env, **kwargs):
""" Optional parameters accepted by Hoptoad are:
scm_revision
scm_repository
local_username
"""
params = {}
params['deploy[rails_env]'] = env
for key, value in kwargs:
params['deploy[%s]' % key] = value
return self.post(**params)
Test Deploy resource after reworking.
|
from hoppy.api import HoptoadResource
class Deploy(HoptoadResource):
def __init__(self, use_ssl=False):
from hoppy import api_key
self.api_key = api_key
super(Deploy, self).__init__(use_ssl)
def check_configuration(self):
if not self.api_key:
raise HoptoadError('API Key cannot be blank')
def request(self, *args, **kwargs):
response = super(Deploy, self).request(
api_key=self.api_key, *args, **kwargs)
return response
def base_uri(self, use_ssl=False):
base = 'http://hoptoadapp.com/deploys.txt'
base = base.replace('http://', 'https://') if use_ssl else base
return base
def deploy(self, env, **kwargs):
""" Optional parameters accepted by Hoptoad are:
scm_revision
scm_repository
local_username
"""
params = {}
params['deploy[rails_env]'] = env
for key, value in kwargs.iteritems():
params['deploy[%s]' % key] = value
return self.post(**params)
|
<commit_before>from restkit import Resource
from hoppy import api_key
class Deploy(Resource):
def __init__(self, use_ssl=False):
self.api_key = api_key
super(Deploy, self).__init__(self.host, follow_redirect=True)
def check_configuration(self):
if not self.api_key:
raise HoptoadError('API Key cannot be blank')
def request(self, *args, **kwargs):
response = super(Deploy, self).request(
api_key=self.api_key, *args, **kwargs)
return response.body_string()
def base_uri(self, use_ssl=False):
base = 'http://hoptoadapp.com/deploys.txt'
base = base.replace('http://', 'https://') if use_ssl else base
return base
def deploy(self, env, **kwargs):
""" Optional parameters accepted by Hoptoad are:
scm_revision
scm_repository
local_username
"""
params = {}
params['deploy[rails_env]'] = env
for key, value in kwargs:
params['deploy[%s]' % key] = value
return self.post(**params)
<commit_msg>Test Deploy resource after reworking.<commit_after>
|
from hoppy.api import HoptoadResource
class Deploy(HoptoadResource):
def __init__(self, use_ssl=False):
from hoppy import api_key
self.api_key = api_key
super(Deploy, self).__init__(use_ssl)
def check_configuration(self):
if not self.api_key:
raise HoptoadError('API Key cannot be blank')
def request(self, *args, **kwargs):
response = super(Deploy, self).request(
api_key=self.api_key, *args, **kwargs)
return response
def base_uri(self, use_ssl=False):
base = 'http://hoptoadapp.com/deploys.txt'
base = base.replace('http://', 'https://') if use_ssl else base
return base
def deploy(self, env, **kwargs):
""" Optional parameters accepted by Hoptoad are:
scm_revision
scm_repository
local_username
"""
params = {}
params['deploy[rails_env]'] = env
for key, value in kwargs.iteritems():
params['deploy[%s]' % key] = value
return self.post(**params)
|
from restkit import Resource
from hoppy import api_key
class Deploy(Resource):
def __init__(self, use_ssl=False):
self.api_key = api_key
super(Deploy, self).__init__(self.host, follow_redirect=True)
def check_configuration(self):
if not self.api_key:
raise HoptoadError('API Key cannot be blank')
def request(self, *args, **kwargs):
response = super(Deploy, self).request(
api_key=self.api_key, *args, **kwargs)
return response.body_string()
def base_uri(self, use_ssl=False):
base = 'http://hoptoadapp.com/deploys.txt'
base = base.replace('http://', 'https://') if use_ssl else base
return base
def deploy(self, env, **kwargs):
""" Optional parameters accepted by Hoptoad are:
scm_revision
scm_repository
local_username
"""
params = {}
params['deploy[rails_env]'] = env
for key, value in kwargs:
params['deploy[%s]' % key] = value
return self.post(**params)
Test Deploy resource after reworking.from hoppy.api import HoptoadResource
class Deploy(HoptoadResource):
def __init__(self, use_ssl=False):
from hoppy import api_key
self.api_key = api_key
super(Deploy, self).__init__(use_ssl)
def check_configuration(self):
if not self.api_key:
raise HoptoadError('API Key cannot be blank')
def request(self, *args, **kwargs):
response = super(Deploy, self).request(
api_key=self.api_key, *args, **kwargs)
return response
def base_uri(self, use_ssl=False):
base = 'http://hoptoadapp.com/deploys.txt'
base = base.replace('http://', 'https://') if use_ssl else base
return base
def deploy(self, env, **kwargs):
""" Optional parameters accepted by Hoptoad are:
scm_revision
scm_repository
local_username
"""
params = {}
params['deploy[rails_env]'] = env
for key, value in kwargs.iteritems():
params['deploy[%s]' % key] = value
return self.post(**params)
|
<commit_before>from restkit import Resource
from hoppy import api_key
class Deploy(Resource):
def __init__(self, use_ssl=False):
self.api_key = api_key
super(Deploy, self).__init__(self.host, follow_redirect=True)
def check_configuration(self):
if not self.api_key:
raise HoptoadError('API Key cannot be blank')
def request(self, *args, **kwargs):
response = super(Deploy, self).request(
api_key=self.api_key, *args, **kwargs)
return response.body_string()
def base_uri(self, use_ssl=False):
base = 'http://hoptoadapp.com/deploys.txt'
base = base.replace('http://', 'https://') if use_ssl else base
return base
def deploy(self, env, **kwargs):
""" Optional parameters accepted by Hoptoad are:
scm_revision
scm_repository
local_username
"""
params = {}
params['deploy[rails_env]'] = env
for key, value in kwargs:
params['deploy[%s]' % key] = value
return self.post(**params)
<commit_msg>Test Deploy resource after reworking.<commit_after>from hoppy.api import HoptoadResource
class Deploy(HoptoadResource):
def __init__(self, use_ssl=False):
from hoppy import api_key
self.api_key = api_key
super(Deploy, self).__init__(use_ssl)
def check_configuration(self):
if not self.api_key:
raise HoptoadError('API Key cannot be blank')
def request(self, *args, **kwargs):
response = super(Deploy, self).request(
api_key=self.api_key, *args, **kwargs)
return response
def base_uri(self, use_ssl=False):
base = 'http://hoptoadapp.com/deploys.txt'
base = base.replace('http://', 'https://') if use_ssl else base
return base
def deploy(self, env, **kwargs):
""" Optional parameters accepted by Hoptoad are:
scm_revision
scm_repository
local_username
"""
params = {}
params['deploy[rails_env]'] = env
for key, value in kwargs.iteritems():
params['deploy[%s]' % key] = value
return self.post(**params)
|
9066d3e5bdbc95fb347b1a081d9b7db33ab68ea4
|
src/autobot/src/stopsign.py
|
src/autobot/src/stopsign.py
|
#!/usr/bin/env python
import rospy
class StopStates(object):
NORMAL = 0
FULL_STOP = 1
IGNORE_STOP_SIGNS = 2
class StopSign(object):
def __init__(self):
self.state = StopStates.NORMAL
self.stopDuration = 2
self.ignoreDuration = 2
def stopSignDetected(self):
self.state = StopStates.FULL_STOP
timer = rospy.Timer(rospy.Duration(self.stopDuration),
self.stepStateMachine,
oneshot=True)
def stepStateMachine(self):
if self.state is StopStates.NORMAL:
self.action = StopStates.FULL_STOP
elif self.state is StopStates.FULL_STOP:
self.action = StopStates.IGNORE_STOP_SIGNS
timer = rospy.Timer(rospy.Duration(self.ignoreDuration),
self.stepStateMachine,
oneshot=True)
elif self.state is StopStates.IGNORE_STOP_SIGNS:
self.action = StopStates.NORMAL
|
#!/usr/bin/env python
import rospy
class StopStates(object):
NORMAL = 0
FULL_STOP = 1
IGNORE_STOP_SIGNS = 2
class StopSign(object):
def __init__(self):
self.state = StopStates.NORMAL
self.stopDuration = 2
self.ignoreDuration = 4
def stopSignDetected(self):
self.state = StopStates.FULL_STOP
timer = rospy.Timer(rospy.Duration(self.stopDuration),
self.stepStateMachine,
oneshot=True)
def stepStateMachine(self, event):
if self.state is StopStates.NORMAL:
self.state = StopStates.FULL_STOP
elif self.state is StopStates.FULL_STOP:
self.state = StopStates.IGNORE_STOP_SIGNS
timer = rospy.Timer(rospy.Duration(self.ignoreDuration),
self.stepStateMachine,
oneshot=True)
elif self.state is StopStates.IGNORE_STOP_SIGNS:
self.state = StopStates.NORMAL
|
Fix state machine using wrong variable
|
Fix state machine using wrong variable
|
Python
|
mit
|
atkvo/masters-bot,atkvo/masters-bot,atkvo/masters-bot,atkvo/masters-bot,atkvo/masters-bot
|
#!/usr/bin/env python
import rospy
class StopStates(object):
NORMAL = 0
FULL_STOP = 1
IGNORE_STOP_SIGNS = 2
class StopSign(object):
def __init__(self):
self.state = StopStates.NORMAL
self.stopDuration = 2
self.ignoreDuration = 2
def stopSignDetected(self):
self.state = StopStates.FULL_STOP
timer = rospy.Timer(rospy.Duration(self.stopDuration),
self.stepStateMachine,
oneshot=True)
def stepStateMachine(self):
if self.state is StopStates.NORMAL:
self.action = StopStates.FULL_STOP
elif self.state is StopStates.FULL_STOP:
self.action = StopStates.IGNORE_STOP_SIGNS
timer = rospy.Timer(rospy.Duration(self.ignoreDuration),
self.stepStateMachine,
oneshot=True)
elif self.state is StopStates.IGNORE_STOP_SIGNS:
self.action = StopStates.NORMAL
Fix state machine using wrong variable
|
#!/usr/bin/env python
import rospy
class StopStates(object):
NORMAL = 0
FULL_STOP = 1
IGNORE_STOP_SIGNS = 2
class StopSign(object):
def __init__(self):
self.state = StopStates.NORMAL
self.stopDuration = 2
self.ignoreDuration = 4
def stopSignDetected(self):
self.state = StopStates.FULL_STOP
timer = rospy.Timer(rospy.Duration(self.stopDuration),
self.stepStateMachine,
oneshot=True)
def stepStateMachine(self, event):
if self.state is StopStates.NORMAL:
self.state = StopStates.FULL_STOP
elif self.state is StopStates.FULL_STOP:
self.state = StopStates.IGNORE_STOP_SIGNS
timer = rospy.Timer(rospy.Duration(self.ignoreDuration),
self.stepStateMachine,
oneshot=True)
elif self.state is StopStates.IGNORE_STOP_SIGNS:
self.state = StopStates.NORMAL
|
<commit_before>#!/usr/bin/env python
import rospy
class StopStates(object):
NORMAL = 0
FULL_STOP = 1
IGNORE_STOP_SIGNS = 2
class StopSign(object):
def __init__(self):
self.state = StopStates.NORMAL
self.stopDuration = 2
self.ignoreDuration = 2
def stopSignDetected(self):
self.state = StopStates.FULL_STOP
timer = rospy.Timer(rospy.Duration(self.stopDuration),
self.stepStateMachine,
oneshot=True)
def stepStateMachine(self):
if self.state is StopStates.NORMAL:
self.action = StopStates.FULL_STOP
elif self.state is StopStates.FULL_STOP:
self.action = StopStates.IGNORE_STOP_SIGNS
timer = rospy.Timer(rospy.Duration(self.ignoreDuration),
self.stepStateMachine,
oneshot=True)
elif self.state is StopStates.IGNORE_STOP_SIGNS:
self.action = StopStates.NORMAL
<commit_msg>Fix state machine using wrong variable<commit_after>
|
#!/usr/bin/env python
import rospy
class StopStates(object):
NORMAL = 0
FULL_STOP = 1
IGNORE_STOP_SIGNS = 2
class StopSign(object):
def __init__(self):
self.state = StopStates.NORMAL
self.stopDuration = 2
self.ignoreDuration = 4
def stopSignDetected(self):
self.state = StopStates.FULL_STOP
timer = rospy.Timer(rospy.Duration(self.stopDuration),
self.stepStateMachine,
oneshot=True)
def stepStateMachine(self, event):
if self.state is StopStates.NORMAL:
self.state = StopStates.FULL_STOP
elif self.state is StopStates.FULL_STOP:
self.state = StopStates.IGNORE_STOP_SIGNS
timer = rospy.Timer(rospy.Duration(self.ignoreDuration),
self.stepStateMachine,
oneshot=True)
elif self.state is StopStates.IGNORE_STOP_SIGNS:
self.state = StopStates.NORMAL
|
#!/usr/bin/env python
import rospy
class StopStates(object):
NORMAL = 0
FULL_STOP = 1
IGNORE_STOP_SIGNS = 2
class StopSign(object):
def __init__(self):
self.state = StopStates.NORMAL
self.stopDuration = 2
self.ignoreDuration = 2
def stopSignDetected(self):
self.state = StopStates.FULL_STOP
timer = rospy.Timer(rospy.Duration(self.stopDuration),
self.stepStateMachine,
oneshot=True)
def stepStateMachine(self):
if self.state is StopStates.NORMAL:
self.action = StopStates.FULL_STOP
elif self.state is StopStates.FULL_STOP:
self.action = StopStates.IGNORE_STOP_SIGNS
timer = rospy.Timer(rospy.Duration(self.ignoreDuration),
self.stepStateMachine,
oneshot=True)
elif self.state is StopStates.IGNORE_STOP_SIGNS:
self.action = StopStates.NORMAL
Fix state machine using wrong variable#!/usr/bin/env python
import rospy
class StopStates(object):
NORMAL = 0
FULL_STOP = 1
IGNORE_STOP_SIGNS = 2
class StopSign(object):
def __init__(self):
self.state = StopStates.NORMAL
self.stopDuration = 2
self.ignoreDuration = 4
def stopSignDetected(self):
self.state = StopStates.FULL_STOP
timer = rospy.Timer(rospy.Duration(self.stopDuration),
self.stepStateMachine,
oneshot=True)
def stepStateMachine(self, event):
if self.state is StopStates.NORMAL:
self.state = StopStates.FULL_STOP
elif self.state is StopStates.FULL_STOP:
self.state = StopStates.IGNORE_STOP_SIGNS
timer = rospy.Timer(rospy.Duration(self.ignoreDuration),
self.stepStateMachine,
oneshot=True)
elif self.state is StopStates.IGNORE_STOP_SIGNS:
self.state = StopStates.NORMAL
|
<commit_before>#!/usr/bin/env python
import rospy
class StopStates(object):
NORMAL = 0
FULL_STOP = 1
IGNORE_STOP_SIGNS = 2
class StopSign(object):
def __init__(self):
self.state = StopStates.NORMAL
self.stopDuration = 2
self.ignoreDuration = 2
def stopSignDetected(self):
self.state = StopStates.FULL_STOP
timer = rospy.Timer(rospy.Duration(self.stopDuration),
self.stepStateMachine,
oneshot=True)
def stepStateMachine(self):
if self.state is StopStates.NORMAL:
self.action = StopStates.FULL_STOP
elif self.state is StopStates.FULL_STOP:
self.action = StopStates.IGNORE_STOP_SIGNS
timer = rospy.Timer(rospy.Duration(self.ignoreDuration),
self.stepStateMachine,
oneshot=True)
elif self.state is StopStates.IGNORE_STOP_SIGNS:
self.action = StopStates.NORMAL
<commit_msg>Fix state machine using wrong variable<commit_after>#!/usr/bin/env python
import rospy
class StopStates(object):
NORMAL = 0
FULL_STOP = 1
IGNORE_STOP_SIGNS = 2
class StopSign(object):
def __init__(self):
self.state = StopStates.NORMAL
self.stopDuration = 2
self.ignoreDuration = 4
def stopSignDetected(self):
self.state = StopStates.FULL_STOP
timer = rospy.Timer(rospy.Duration(self.stopDuration),
self.stepStateMachine,
oneshot=True)
def stepStateMachine(self, event):
if self.state is StopStates.NORMAL:
self.state = StopStates.FULL_STOP
elif self.state is StopStates.FULL_STOP:
self.state = StopStates.IGNORE_STOP_SIGNS
timer = rospy.Timer(rospy.Duration(self.ignoreDuration),
self.stepStateMachine,
oneshot=True)
elif self.state is StopStates.IGNORE_STOP_SIGNS:
self.state = StopStates.NORMAL
|
2c5a1bebf805c9bf5208fc75c32d8998b865eb32
|
designate/objects/zone_transfer_request.py
|
designate/objects/zone_transfer_request.py
|
# Copyright 2014 Hewlett-Packard Development Company, L.P.
#
# Author: Graham Hayes <graham.hayes@hp.com>
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from designate.objects import base
class ZoneTransferRequest(base.DictObjectMixin, base.DesignateObject,
base.PersistentObjectMixin):
FIELDS = {
'domain_id': {},
'key': {},
'description': {},
'tenant_id': {},
'target_tenant_id': {},
'status': {},
'id': {},
'created_at': {},
'domain_name': {},
'updated_at': {},
'version': {},
}
class ZoneTransferRequestList(base.ListObjectMixin, base.DesignateObject):
LIST_ITEM_TYPE = ZoneTransferRequest
|
# Copyright 2014 Hewlett-Packard Development Company, L.P.
#
# Author: Graham Hayes <graham.hayes@hp.com>
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from designate.objects import base
class ZoneTransferRequest(base.DictObjectMixin, base.PersistentObjectMixin,
base.DesignateObject,):
FIELDS = {
'domain_id': {},
'key': {},
'description': {},
'tenant_id': {},
'target_tenant_id': {},
'status': {},
'domain_name': {},
}
class ZoneTransferRequestList(base.ListObjectMixin, base.DesignateObject):
LIST_ITEM_TYPE = ZoneTransferRequest
|
Remove duplicate fields from ZoneTransferRequest object
|
Remove duplicate fields from ZoneTransferRequest object
The fields id, version, created_at, updated_at are defined in the
PersistentObjectMixin which ZoneTransferRequest extends, so this
patch removes them from ZoneTransferRequest.
Change-Id: Iff20a31b4a208bff0bc879677a9901fedc43226b
Closes-Bug: #1403274
|
Python
|
apache-2.0
|
kiall/designate-py3,muraliselva10/designate,openstack/designate,kiall/designate-py3,ramsateesh/designate,kiall/designate-py3,openstack/designate,muraliselva10/designate,cneill/designate,tonyli71/designate,cneill/designate,cneill/designate-testing,cneill/designate-testing,cneill/designate,tonyli71/designate,muraliselva10/designate,ionrock/designate,grahamhayes/designate,cneill/designate,ramsateesh/designate,tonyli71/designate,ionrock/designate,openstack/designate,cneill/designate-testing,kiall/designate-py3,cneill/designate,ramsateesh/designate,kiall/designate-py3,ionrock/designate,grahamhayes/designate,grahamhayes/designate
|
# Copyright 2014 Hewlett-Packard Development Company, L.P.
#
# Author: Graham Hayes <graham.hayes@hp.com>
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from designate.objects import base
class ZoneTransferRequest(base.DictObjectMixin, base.DesignateObject,
base.PersistentObjectMixin):
FIELDS = {
'domain_id': {},
'key': {},
'description': {},
'tenant_id': {},
'target_tenant_id': {},
'status': {},
'id': {},
'created_at': {},
'domain_name': {},
'updated_at': {},
'version': {},
}
class ZoneTransferRequestList(base.ListObjectMixin, base.DesignateObject):
LIST_ITEM_TYPE = ZoneTransferRequest
Remove duplicate fields from ZoneTransferRequest object
The fields id, version, created_at, updated_at are defined in the
PersistentObjectMixin which ZoneTransferRequest extends, so this
patch removes them from ZoneTransferRequest.
Change-Id: Iff20a31b4a208bff0bc879677a9901fedc43226b
Closes-Bug: #1403274
|
# Copyright 2014 Hewlett-Packard Development Company, L.P.
#
# Author: Graham Hayes <graham.hayes@hp.com>
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from designate.objects import base
class ZoneTransferRequest(base.DictObjectMixin, base.PersistentObjectMixin,
base.DesignateObject,):
FIELDS = {
'domain_id': {},
'key': {},
'description': {},
'tenant_id': {},
'target_tenant_id': {},
'status': {},
'domain_name': {},
}
class ZoneTransferRequestList(base.ListObjectMixin, base.DesignateObject):
LIST_ITEM_TYPE = ZoneTransferRequest
|
<commit_before># Copyright 2014 Hewlett-Packard Development Company, L.P.
#
# Author: Graham Hayes <graham.hayes@hp.com>
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from designate.objects import base
class ZoneTransferRequest(base.DictObjectMixin, base.DesignateObject,
base.PersistentObjectMixin):
FIELDS = {
'domain_id': {},
'key': {},
'description': {},
'tenant_id': {},
'target_tenant_id': {},
'status': {},
'id': {},
'created_at': {},
'domain_name': {},
'updated_at': {},
'version': {},
}
class ZoneTransferRequestList(base.ListObjectMixin, base.DesignateObject):
LIST_ITEM_TYPE = ZoneTransferRequest
<commit_msg>Remove duplicate fields from ZoneTransferRequest object
The fields id, version, created_at, updated_at are defined in the
PersistentObjectMixin which ZoneTransferRequest extends, so this
patch removes them from ZoneTransferRequest.
Change-Id: Iff20a31b4a208bff0bc879677a9901fedc43226b
Closes-Bug: #1403274<commit_after>
|
# Copyright 2014 Hewlett-Packard Development Company, L.P.
#
# Author: Graham Hayes <graham.hayes@hp.com>
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from designate.objects import base
class ZoneTransferRequest(base.DictObjectMixin, base.PersistentObjectMixin,
base.DesignateObject,):
FIELDS = {
'domain_id': {},
'key': {},
'description': {},
'tenant_id': {},
'target_tenant_id': {},
'status': {},
'domain_name': {},
}
class ZoneTransferRequestList(base.ListObjectMixin, base.DesignateObject):
LIST_ITEM_TYPE = ZoneTransferRequest
|
# Copyright 2014 Hewlett-Packard Development Company, L.P.
#
# Author: Graham Hayes <graham.hayes@hp.com>
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from designate.objects import base
class ZoneTransferRequest(base.DictObjectMixin, base.DesignateObject,
base.PersistentObjectMixin):
FIELDS = {
'domain_id': {},
'key': {},
'description': {},
'tenant_id': {},
'target_tenant_id': {},
'status': {},
'id': {},
'created_at': {},
'domain_name': {},
'updated_at': {},
'version': {},
}
class ZoneTransferRequestList(base.ListObjectMixin, base.DesignateObject):
LIST_ITEM_TYPE = ZoneTransferRequest
Remove duplicate fields from ZoneTransferRequest object
The fields id, version, created_at, updated_at are defined in the
PersistentObjectMixin which ZoneTransferRequest extends, so this
patch removes them from ZoneTransferRequest.
Change-Id: Iff20a31b4a208bff0bc879677a9901fedc43226b
Closes-Bug: #1403274# Copyright 2014 Hewlett-Packard Development Company, L.P.
#
# Author: Graham Hayes <graham.hayes@hp.com>
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from designate.objects import base
class ZoneTransferRequest(base.DictObjectMixin, base.PersistentObjectMixin,
base.DesignateObject,):
FIELDS = {
'domain_id': {},
'key': {},
'description': {},
'tenant_id': {},
'target_tenant_id': {},
'status': {},
'domain_name': {},
}
class ZoneTransferRequestList(base.ListObjectMixin, base.DesignateObject):
LIST_ITEM_TYPE = ZoneTransferRequest
|
<commit_before># Copyright 2014 Hewlett-Packard Development Company, L.P.
#
# Author: Graham Hayes <graham.hayes@hp.com>
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from designate.objects import base
class ZoneTransferRequest(base.DictObjectMixin, base.DesignateObject,
base.PersistentObjectMixin):
FIELDS = {
'domain_id': {},
'key': {},
'description': {},
'tenant_id': {},
'target_tenant_id': {},
'status': {},
'id': {},
'created_at': {},
'domain_name': {},
'updated_at': {},
'version': {},
}
class ZoneTransferRequestList(base.ListObjectMixin, base.DesignateObject):
LIST_ITEM_TYPE = ZoneTransferRequest
<commit_msg>Remove duplicate fields from ZoneTransferRequest object
The fields id, version, created_at, updated_at are defined in the
PersistentObjectMixin which ZoneTransferRequest extends, so this
patch removes them from ZoneTransferRequest.
Change-Id: Iff20a31b4a208bff0bc879677a9901fedc43226b
Closes-Bug: #1403274<commit_after># Copyright 2014 Hewlett-Packard Development Company, L.P.
#
# Author: Graham Hayes <graham.hayes@hp.com>
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from designate.objects import base
class ZoneTransferRequest(base.DictObjectMixin, base.PersistentObjectMixin,
base.DesignateObject,):
FIELDS = {
'domain_id': {},
'key': {},
'description': {},
'tenant_id': {},
'target_tenant_id': {},
'status': {},
'domain_name': {},
}
class ZoneTransferRequestList(base.ListObjectMixin, base.DesignateObject):
LIST_ITEM_TYPE = ZoneTransferRequest
|
8441acfd5071e8b63fde816f67e167997045d510
|
Lib/misc/setup.py
|
Lib/misc/setup.py
|
import os
from numpy.distutils.misc_util import Configuration
def configuration(parent_package='',top_path=None):
config = Configuration('misc',parent_package, top_path)
config.add_data_files('lena.dat')
print "########", config
return config
if __name__ == '__main__':
from numpy.distutils.core import setup
setup(**configuration())
|
import os
from numpy.distutils.misc_util import Configuration
def configuration(parent_package='',top_path=None):
config = Configuration('misc',parent_package, top_path)
config.add_data_files('lena.dat')
return config
if __name__ == '__main__':
from numpy.distutils.core import setup
setup(**configuration())
|
Remove extra noise on install.
|
Remove extra noise on install.
|
Python
|
bsd-3-clause
|
jseabold/scipy,richardotis/scipy,anntzer/scipy,fredrikw/scipy,behzadnouri/scipy,aman-iitj/scipy,mortada/scipy,njwilson23/scipy,trankmichael/scipy,trankmichael/scipy,apbard/scipy,niknow/scipy,aman-iitj/scipy,behzadnouri/scipy,FRidh/scipy,vanpact/scipy,Eric89GXL/scipy,rmcgibbo/scipy,larsmans/scipy,Shaswat27/scipy,ogrisel/scipy,niknow/scipy,chatcannon/scipy,mhogg/scipy,petebachant/scipy,jor-/scipy,nonhermitian/scipy,mgaitan/scipy,chatcannon/scipy,Eric89GXL/scipy,gdooper/scipy,jonycgn/scipy,FRidh/scipy,fernand/scipy,juliantaylor/scipy,apbard/scipy,vigna/scipy,sargas/scipy,pschella/scipy,mhogg/scipy,chatcannon/scipy,Kamp9/scipy,juliantaylor/scipy,behzadnouri/scipy,surhudm/scipy,ales-erjavec/scipy,kalvdans/scipy,gertingold/scipy,piyush0609/scipy,surhudm/scipy,nonhermitian/scipy,mgaitan/scipy,pbrod/scipy,mortonjt/scipy,rmcgibbo/scipy,matthew-brett/scipy,aeklant/scipy,fredrikw/scipy,lukauskas/scipy,sonnyhu/scipy,raoulbq/scipy,sonnyhu/scipy,njwilson23/scipy,haudren/scipy,pyramania/scipy,zerothi/scipy,arokem/scipy,mikebenfield/scipy,minhlongdo/scipy,grlee77/scipy,gdooper/scipy,bkendzior/scipy,anntzer/scipy,dch312/scipy,grlee77/scipy,fredrikw/scipy,haudren/scipy,zerothi/scipy,witcxc/scipy,mtrbean/scipy,sriki18/scipy,jakevdp/scipy,sonnyhu/scipy,pbrod/scipy,giorgiop/scipy,vhaasteren/scipy,ogrisel/scipy,gertingold/scipy,nvoron23/scipy,andyfaff/scipy,andim/scipy,jamestwebber/scipy,maniteja123/scipy,FRidh/scipy,Kamp9/scipy,jjhelmus/scipy,WillieMaddox/scipy,nmayorov/scipy,Gillu13/scipy,jjhelmus/scipy,Eric89GXL/scipy,woodscn/scipy,sriki18/scipy,pizzathief/scipy,newemailjdm/scipy,ChanderG/scipy,felipebetancur/scipy,Stefan-Endres/scipy,gfyoung/scipy,endolith/scipy,ales-erjavec/scipy,giorgiop/scipy,WarrenWeckesser/scipy,surhudm/scipy,sriki18/scipy,Eric89GXL/scipy,anntzer/scipy,kleskjr/scipy,vberaudi/scipy,rgommers/scipy,nvoron23/scipy,pyramania/scipy,kalvdans/scipy,rgommers/scipy,felipebetancur/scipy,Newman101/scipy,ortylp/scipy,dominicelse/scipy,aarchiba/scipy,vhaasteren/scipy,endolith/scipy,nonhermitian/scipy,pizzathief/scipy,surhudm/scipy,Dapid/scipy,jseabold/scipy,kleskjr/scipy,Newman101/scipy,andyfaff/scipy,aman-iitj/scipy,tylerjereddy/scipy,mikebenfield/scipy,futurulus/scipy,kalvdans/scipy,teoliphant/scipy,maciejkula/scipy,mingwpy/scipy,jor-/scipy,ChanderG/scipy,mtrbean/scipy,endolith/scipy,cpaulik/scipy,surhudm/scipy,rgommers/scipy,mortonjt/scipy,perimosocordiae/scipy,rmcgibbo/scipy,mdhaber/scipy,Shaswat27/scipy,woodscn/scipy,jakevdp/scipy,maciejkula/scipy,andim/scipy,WarrenWeckesser/scipy,rgommers/scipy,Shaswat27/scipy,rmcgibbo/scipy,argriffing/scipy,jonycgn/scipy,gef756/scipy,befelix/scipy,andim/scipy,Srisai85/scipy,minhlongdo/scipy,ChanderG/scipy,sauliusl/scipy,sauliusl/scipy,trankmichael/scipy,ilayn/scipy,kleskjr/scipy,newemailjdm/scipy,gdooper/scipy,ortylp/scipy,vhaasteren/scipy,juliantaylor/scipy,josephcslater/scipy,lukauskas/scipy,teoliphant/scipy,Gillu13/scipy,vigna/scipy,zxsted/scipy,jamestwebber/scipy,apbard/scipy,pnedunuri/scipy,dominicelse/scipy,grlee77/scipy,witcxc/scipy,raoulbq/scipy,zaxliu/scipy,petebachant/scipy,perimosocordiae/scipy,Gillu13/scipy,mdhaber/scipy,jseabold/scipy,jamestwebber/scipy,hainm/scipy,zxsted/scipy,FRidh/scipy,cpaulik/scipy,gef756/scipy,FRidh/scipy,grlee77/scipy,trankmichael/scipy,e-q/scipy,maciejkula/scipy,richardotis/scipy,dominicelse/scipy,haudren/scipy,sriki18/scipy,dch312/scipy,efiring/scipy,piyush0609/scipy,jsilter/scipy,sargas/scipy,matthewalbani/scipy,Gillu13/scipy,maniteja123/scipy,gef756/scipy,lhilt/scipy,andim/scipy,Stefan-Endres/scipy,ilayn/scipy,niknow/scipy,mhogg/scipy,petebachant/scipy,giorgiop/scipy,anntzer/scipy,larsmans/scipy,minhlongdo/scipy,aman-iitj/scipy,kalvdans/scipy,zerothi/scipy,efiring/scipy,pnedunuri/scipy,perimosocordiae/scipy,sauliusl/scipy,jonycgn/scipy,Newman101/scipy,Gillu13/scipy,vhaasteren/scipy,andyfaff/scipy,apbard/scipy,jor-/scipy,sargas/scipy,ales-erjavec/scipy,raoulbq/scipy,jonycgn/scipy,nvoron23/scipy,richardotis/scipy,lukauskas/scipy,ChanderG/scipy,minhlongdo/scipy,Shaswat27/scipy,woodscn/scipy,niknow/scipy,aeklant/scipy,mingwpy/scipy,vhaasteren/scipy,anielsen001/scipy,njwilson23/scipy,ogrisel/scipy,fernand/scipy,anntzer/scipy,jakevdp/scipy,person142/scipy,pbrod/scipy,pbrod/scipy,aarchiba/scipy,vberaudi/scipy,tylerjereddy/scipy,Kamp9/scipy,dch312/scipy,njwilson23/scipy,hainm/scipy,endolith/scipy,zerothi/scipy,nvoron23/scipy,minhlongdo/scipy,sriki18/scipy,scipy/scipy,gertingold/scipy,petebachant/scipy,pizzathief/scipy,nmayorov/scipy,mhogg/scipy,mtrbean/scipy,kalvdans/scipy,mgaitan/scipy,mhogg/scipy,matthew-brett/scipy,scipy/scipy,jamestwebber/scipy,aarchiba/scipy,gfyoung/scipy,ales-erjavec/scipy,zaxliu/scipy,vberaudi/scipy,pschella/scipy,jseabold/scipy,perimosocordiae/scipy,zxsted/scipy,behzadnouri/scipy,gfyoung/scipy,woodscn/scipy,mingwpy/scipy,maniteja123/scipy,mdhaber/scipy,josephcslater/scipy,Kamp9/scipy,ndchorley/scipy,sriki18/scipy,trankmichael/scipy,Stefan-Endres/scipy,niknow/scipy,witcxc/scipy,argriffing/scipy,vanpact/scipy,person142/scipy,lhilt/scipy,mtrbean/scipy,haudren/scipy,larsmans/scipy,efiring/scipy,mtrbean/scipy,hainm/scipy,zxsted/scipy,ChanderG/scipy,mortada/scipy,giorgiop/scipy,aman-iitj/scipy,josephcslater/scipy,vanpact/scipy,matthew-brett/scipy,teoliphant/scipy,matthewalbani/scipy,newemailjdm/scipy,woodscn/scipy,sonnyhu/scipy,aeklant/scipy,anielsen001/scipy,andyfaff/scipy,cpaulik/scipy,perimosocordiae/scipy,pizzathief/scipy,njwilson23/scipy,e-q/scipy,trankmichael/scipy,lhilt/scipy,tylerjereddy/scipy,teoliphant/scipy,vigna/scipy,mdhaber/scipy,ogrisel/scipy,petebachant/scipy,raoulbq/scipy,felipebetancur/scipy,nmayorov/scipy,njwilson23/scipy,sargas/scipy,larsmans/scipy,jamestwebber/scipy,newemailjdm/scipy,rgommers/scipy,Stefan-Endres/scipy,rmcgibbo/scipy,mgaitan/scipy,anielsen001/scipy,arokem/scipy,jjhelmus/scipy,zaxliu/scipy,nvoron23/scipy,chatcannon/scipy,mikebenfield/scipy,richardotis/scipy,mortonjt/scipy,nonhermitian/scipy,befelix/scipy,mikebenfield/scipy,gertingold/scipy,gdooper/scipy,pnedunuri/scipy,befelix/scipy,scipy/scipy,nonhermitian/scipy,WillieMaddox/scipy,WillieMaddox/scipy,behzadnouri/scipy,haudren/scipy,piyush0609/scipy,dominicelse/scipy,Shaswat27/scipy,richardotis/scipy,WarrenWeckesser/scipy,gfyoung/scipy,mortada/scipy,efiring/scipy,ales-erjavec/scipy,WarrenWeckesser/scipy,mingwpy/scipy,aeklant/scipy,piyush0609/scipy,felipebetancur/scipy,mhogg/scipy,pbrod/scipy,mdhaber/scipy,maniteja123/scipy,anielsen001/scipy,Srisai85/scipy,mdhaber/scipy,pschella/scipy,pnedunuri/scipy,mortada/scipy,mingwpy/scipy,piyush0609/scipy,endolith/scipy,ortylp/scipy,mgaitan/scipy,surhudm/scipy,WillieMaddox/scipy,giorgiop/scipy,Srisai85/scipy,Newman101/scipy,maciejkula/scipy,perimosocordiae/scipy,person142/scipy,josephcslater/scipy,anntzer/scipy,gfyoung/scipy,pyramania/scipy,fredrikw/scipy,grlee77/scipy,kleskjr/scipy,Stefan-Endres/scipy,sonnyhu/scipy,cpaulik/scipy,Kamp9/scipy,ndchorley/scipy,nmayorov/scipy,gef756/scipy,vanpact/scipy,arokem/scipy,ndchorley/scipy,jonycgn/scipy,e-q/scipy,Dapid/scipy,ales-erjavec/scipy,jsilter/scipy,argriffing/scipy,zxsted/scipy,pyramania/scipy,jseabold/scipy,witcxc/scipy,fernand/scipy,sauliusl/scipy,zaxliu/scipy,lhilt/scipy,scipy/scipy,lhilt/scipy,juliantaylor/scipy,bkendzior/scipy,aman-iitj/scipy,Eric89GXL/scipy,cpaulik/scipy,aarchiba/scipy,fernand/scipy,fredrikw/scipy,Kamp9/scipy,zerothi/scipy,pizzathief/scipy,fernand/scipy,jonycgn/scipy,zxsted/scipy,andyfaff/scipy,kleskjr/scipy,mortonjt/scipy,jseabold/scipy,vigna/scipy,e-q/scipy,jakevdp/scipy,vanpact/scipy,mingwpy/scipy,futurulus/scipy,andim/scipy,pnedunuri/scipy,sonnyhu/scipy,matthew-brett/scipy,lukauskas/scipy,vanpact/scipy,jjhelmus/scipy,Srisai85/scipy,jsilter/scipy,argriffing/scipy,bkendzior/scipy,felipebetancur/scipy,hainm/scipy,pyramania/scipy,vberaudi/scipy,apbard/scipy,befelix/scipy,giorgiop/scipy,hainm/scipy,WarrenWeckesser/scipy,fredrikw/scipy,jor-/scipy,larsmans/scipy,ilayn/scipy,josephcslater/scipy,vigna/scipy,matthew-brett/scipy,person142/scipy,futurulus/scipy,zerothi/scipy,vberaudi/scipy,gertingold/scipy,ndchorley/scipy,arokem/scipy,arokem/scipy,raoulbq/scipy,matthewalbani/scipy,ndchorley/scipy,anielsen001/scipy,matthewalbani/scipy,pschella/scipy,cpaulik/scipy,bkendzior/scipy,gdooper/scipy,Shaswat27/scipy,sargas/scipy,Eric89GXL/scipy,aarchiba/scipy,ChanderG/scipy,ortylp/scipy,efiring/scipy,teoliphant/scipy,richardotis/scipy,andyfaff/scipy,jor-/scipy,ilayn/scipy,WarrenWeckesser/scipy,argriffing/scipy,maciejkula/scipy,mtrbean/scipy,futurulus/scipy,scipy/scipy,futurulus/scipy,person142/scipy,ilayn/scipy,aeklant/scipy,Newman101/scipy,nvoron23/scipy,felipebetancur/scipy,dch312/scipy,futurulus/scipy,raoulbq/scipy,lukauskas/scipy,anielsen001/scipy,zaxliu/scipy,efiring/scipy,lukauskas/scipy,minhlongdo/scipy,gef756/scipy,argriffing/scipy,ilayn/scipy,tylerjereddy/scipy,Newman101/scipy,witcxc/scipy,rmcgibbo/scipy,scipy/scipy,haudren/scipy,behzadnouri/scipy,chatcannon/scipy,larsmans/scipy,endolith/scipy,maniteja123/scipy,piyush0609/scipy,mortada/scipy,ortylp/scipy,jakevdp/scipy,Dapid/scipy,woodscn/scipy,newemailjdm/scipy,sauliusl/scipy,vhaasteren/scipy,vberaudi/scipy,FRidh/scipy,tylerjereddy/scipy,pschella/scipy,juliantaylor/scipy,newemailjdm/scipy,petebachant/scipy,chatcannon/scipy,Srisai85/scipy,sauliusl/scipy,mortonjt/scipy,kleskjr/scipy,dch312/scipy,Stefan-Endres/scipy,befelix/scipy,hainm/scipy,ndchorley/scipy,Gillu13/scipy,fernand/scipy,andim/scipy,gef756/scipy,pnedunuri/scipy,Dapid/scipy,jsilter/scipy,dominicelse/scipy,matthewalbani/scipy,mikebenfield/scipy,bkendzior/scipy,WillieMaddox/scipy,Dapid/scipy,jsilter/scipy,niknow/scipy,ortylp/scipy,mortonjt/scipy,pbrod/scipy,jjhelmus/scipy,mortada/scipy,WillieMaddox/scipy,e-q/scipy,Dapid/scipy,mgaitan/scipy,zaxliu/scipy,nmayorov/scipy,maniteja123/scipy,Srisai85/scipy,ogrisel/scipy
|
import os
from numpy.distutils.misc_util import Configuration
def configuration(parent_package='',top_path=None):
config = Configuration('misc',parent_package, top_path)
config.add_data_files('lena.dat')
print "########", config
return config
if __name__ == '__main__':
from numpy.distutils.core import setup
setup(**configuration())
Remove extra noise on install.
|
import os
from numpy.distutils.misc_util import Configuration
def configuration(parent_package='',top_path=None):
config = Configuration('misc',parent_package, top_path)
config.add_data_files('lena.dat')
return config
if __name__ == '__main__':
from numpy.distutils.core import setup
setup(**configuration())
|
<commit_before>
import os
from numpy.distutils.misc_util import Configuration
def configuration(parent_package='',top_path=None):
config = Configuration('misc',parent_package, top_path)
config.add_data_files('lena.dat')
print "########", config
return config
if __name__ == '__main__':
from numpy.distutils.core import setup
setup(**configuration())
<commit_msg>Remove extra noise on install.<commit_after>
|
import os
from numpy.distutils.misc_util import Configuration
def configuration(parent_package='',top_path=None):
config = Configuration('misc',parent_package, top_path)
config.add_data_files('lena.dat')
return config
if __name__ == '__main__':
from numpy.distutils.core import setup
setup(**configuration())
|
import os
from numpy.distutils.misc_util import Configuration
def configuration(parent_package='',top_path=None):
config = Configuration('misc',parent_package, top_path)
config.add_data_files('lena.dat')
print "########", config
return config
if __name__ == '__main__':
from numpy.distutils.core import setup
setup(**configuration())
Remove extra noise on install.
import os
from numpy.distutils.misc_util import Configuration
def configuration(parent_package='',top_path=None):
config = Configuration('misc',parent_package, top_path)
config.add_data_files('lena.dat')
return config
if __name__ == '__main__':
from numpy.distutils.core import setup
setup(**configuration())
|
<commit_before>
import os
from numpy.distutils.misc_util import Configuration
def configuration(parent_package='',top_path=None):
config = Configuration('misc',parent_package, top_path)
config.add_data_files('lena.dat')
print "########", config
return config
if __name__ == '__main__':
from numpy.distutils.core import setup
setup(**configuration())
<commit_msg>Remove extra noise on install.<commit_after>
import os
from numpy.distutils.misc_util import Configuration
def configuration(parent_package='',top_path=None):
config = Configuration('misc',parent_package, top_path)
config.add_data_files('lena.dat')
return config
if __name__ == '__main__':
from numpy.distutils.core import setup
setup(**configuration())
|
758f73e1ecc34f52929595dfcf5db4a3a24fcbc6
|
Python/views.py
|
Python/views.py
|
import requests
from django.shortcuts import render
from django.conf import settings
def oauthtest(request):
return render(request, 'oauthtest.html', {
'link': '{}o/authorize/?response_type=code&client_id={}&redirect_uri={}{}/oauthdone/'.format(
settings.API_URL,
settings.OAUTH_CLIENT_ID,
'https://' if request.is_secure() else 'http://',
request.META['HTTP_HOST'],
)
})
def oauthdone(request):
# Get token
r = requests.post(settings.API_URL + 'o/token/', data={
'grant_type': 'authorization_code',
'code': request.GET['code'],
'redirect_uri': 'http://localhost:8000/oauthdone/',
'client_id': settings.OAUTH_CLIENT_ID,
'client_secret': settings.OAUTH_CLIENT_SECRET,
})
oauth = r.json()
# Get user object + accounts
headers = {
'Authorization': oauth['token_type'] + ' ' + oauth['access_token'],
}
r = requests.get(settings.API_URL + 'api/users/me/?expand_accounts=True', headers=headers)
return render(request, 'accounts.html', {
'user': r.json(),
})
|
import requests
from django.shortcuts import render
from django.conf import settings
def oauthtest(request):
return render(request, 'oauthtest.html', {
'link': '{}o/authorize/?response_type=code&client_id={}&redirect_uri={}{}/oauthdone/'.format(
settings.API_URL,
settings.OAUTH_CLIENT_ID,
'https://' if request.is_secure() else 'http://',
request.META['HTTP_HOST'],
)
})
def oauthdone(request):
# Get token
r = requests.post(settings.API_URL + 'o/token/', data={
'grant_type': 'authorization_code',
'code': request.GET['code'],
'redirect_uri': '{}{}/oauthdone/'.format('https://' if request.is_secure() else 'http://', request.META['HTTP_HOST']),
'client_id': settings.OAUTH_CLIENT_ID,
'client_secret': settings.OAUTH_CLIENT_SECRET,
})
oauth = r.json()
# Get user object + accounts
headers = {
'Authorization': oauth['token_type'] + ' ' + oauth['access_token'],
}
r = requests.get(settings.API_URL + 'api/users/me/?expand_accounts=True', headers=headers)
return render(request, 'accounts.html', {
'user': r.json(),
})
|
Fix redirect URI in oauthdone
|
Fix redirect URI in oauthdone
|
Python
|
apache-2.0
|
SchoolIdolTomodachi/SchoolIdolAPIOAuthExample,SchoolIdolTomodachi/SchoolIdolAPIOAuthExample,SchoolIdolTomodachi/SchoolIdolAPIOAuthExample
|
import requests
from django.shortcuts import render
from django.conf import settings
def oauthtest(request):
return render(request, 'oauthtest.html', {
'link': '{}o/authorize/?response_type=code&client_id={}&redirect_uri={}{}/oauthdone/'.format(
settings.API_URL,
settings.OAUTH_CLIENT_ID,
'https://' if request.is_secure() else 'http://',
request.META['HTTP_HOST'],
)
})
def oauthdone(request):
# Get token
r = requests.post(settings.API_URL + 'o/token/', data={
'grant_type': 'authorization_code',
'code': request.GET['code'],
'redirect_uri': 'http://localhost:8000/oauthdone/',
'client_id': settings.OAUTH_CLIENT_ID,
'client_secret': settings.OAUTH_CLIENT_SECRET,
})
oauth = r.json()
# Get user object + accounts
headers = {
'Authorization': oauth['token_type'] + ' ' + oauth['access_token'],
}
r = requests.get(settings.API_URL + 'api/users/me/?expand_accounts=True', headers=headers)
return render(request, 'accounts.html', {
'user': r.json(),
})
Fix redirect URI in oauthdone
|
import requests
from django.shortcuts import render
from django.conf import settings
def oauthtest(request):
return render(request, 'oauthtest.html', {
'link': '{}o/authorize/?response_type=code&client_id={}&redirect_uri={}{}/oauthdone/'.format(
settings.API_URL,
settings.OAUTH_CLIENT_ID,
'https://' if request.is_secure() else 'http://',
request.META['HTTP_HOST'],
)
})
def oauthdone(request):
# Get token
r = requests.post(settings.API_URL + 'o/token/', data={
'grant_type': 'authorization_code',
'code': request.GET['code'],
'redirect_uri': '{}{}/oauthdone/'.format('https://' if request.is_secure() else 'http://', request.META['HTTP_HOST']),
'client_id': settings.OAUTH_CLIENT_ID,
'client_secret': settings.OAUTH_CLIENT_SECRET,
})
oauth = r.json()
# Get user object + accounts
headers = {
'Authorization': oauth['token_type'] + ' ' + oauth['access_token'],
}
r = requests.get(settings.API_URL + 'api/users/me/?expand_accounts=True', headers=headers)
return render(request, 'accounts.html', {
'user': r.json(),
})
|
<commit_before>import requests
from django.shortcuts import render
from django.conf import settings
def oauthtest(request):
return render(request, 'oauthtest.html', {
'link': '{}o/authorize/?response_type=code&client_id={}&redirect_uri={}{}/oauthdone/'.format(
settings.API_URL,
settings.OAUTH_CLIENT_ID,
'https://' if request.is_secure() else 'http://',
request.META['HTTP_HOST'],
)
})
def oauthdone(request):
# Get token
r = requests.post(settings.API_URL + 'o/token/', data={
'grant_type': 'authorization_code',
'code': request.GET['code'],
'redirect_uri': 'http://localhost:8000/oauthdone/',
'client_id': settings.OAUTH_CLIENT_ID,
'client_secret': settings.OAUTH_CLIENT_SECRET,
})
oauth = r.json()
# Get user object + accounts
headers = {
'Authorization': oauth['token_type'] + ' ' + oauth['access_token'],
}
r = requests.get(settings.API_URL + 'api/users/me/?expand_accounts=True', headers=headers)
return render(request, 'accounts.html', {
'user': r.json(),
})
<commit_msg>Fix redirect URI in oauthdone<commit_after>
|
import requests
from django.shortcuts import render
from django.conf import settings
def oauthtest(request):
return render(request, 'oauthtest.html', {
'link': '{}o/authorize/?response_type=code&client_id={}&redirect_uri={}{}/oauthdone/'.format(
settings.API_URL,
settings.OAUTH_CLIENT_ID,
'https://' if request.is_secure() else 'http://',
request.META['HTTP_HOST'],
)
})
def oauthdone(request):
# Get token
r = requests.post(settings.API_URL + 'o/token/', data={
'grant_type': 'authorization_code',
'code': request.GET['code'],
'redirect_uri': '{}{}/oauthdone/'.format('https://' if request.is_secure() else 'http://', request.META['HTTP_HOST']),
'client_id': settings.OAUTH_CLIENT_ID,
'client_secret': settings.OAUTH_CLIENT_SECRET,
})
oauth = r.json()
# Get user object + accounts
headers = {
'Authorization': oauth['token_type'] + ' ' + oauth['access_token'],
}
r = requests.get(settings.API_URL + 'api/users/me/?expand_accounts=True', headers=headers)
return render(request, 'accounts.html', {
'user': r.json(),
})
|
import requests
from django.shortcuts import render
from django.conf import settings
def oauthtest(request):
return render(request, 'oauthtest.html', {
'link': '{}o/authorize/?response_type=code&client_id={}&redirect_uri={}{}/oauthdone/'.format(
settings.API_URL,
settings.OAUTH_CLIENT_ID,
'https://' if request.is_secure() else 'http://',
request.META['HTTP_HOST'],
)
})
def oauthdone(request):
# Get token
r = requests.post(settings.API_URL + 'o/token/', data={
'grant_type': 'authorization_code',
'code': request.GET['code'],
'redirect_uri': 'http://localhost:8000/oauthdone/',
'client_id': settings.OAUTH_CLIENT_ID,
'client_secret': settings.OAUTH_CLIENT_SECRET,
})
oauth = r.json()
# Get user object + accounts
headers = {
'Authorization': oauth['token_type'] + ' ' + oauth['access_token'],
}
r = requests.get(settings.API_URL + 'api/users/me/?expand_accounts=True', headers=headers)
return render(request, 'accounts.html', {
'user': r.json(),
})
Fix redirect URI in oauthdoneimport requests
from django.shortcuts import render
from django.conf import settings
def oauthtest(request):
return render(request, 'oauthtest.html', {
'link': '{}o/authorize/?response_type=code&client_id={}&redirect_uri={}{}/oauthdone/'.format(
settings.API_URL,
settings.OAUTH_CLIENT_ID,
'https://' if request.is_secure() else 'http://',
request.META['HTTP_HOST'],
)
})
def oauthdone(request):
# Get token
r = requests.post(settings.API_URL + 'o/token/', data={
'grant_type': 'authorization_code',
'code': request.GET['code'],
'redirect_uri': '{}{}/oauthdone/'.format('https://' if request.is_secure() else 'http://', request.META['HTTP_HOST']),
'client_id': settings.OAUTH_CLIENT_ID,
'client_secret': settings.OAUTH_CLIENT_SECRET,
})
oauth = r.json()
# Get user object + accounts
headers = {
'Authorization': oauth['token_type'] + ' ' + oauth['access_token'],
}
r = requests.get(settings.API_URL + 'api/users/me/?expand_accounts=True', headers=headers)
return render(request, 'accounts.html', {
'user': r.json(),
})
|
<commit_before>import requests
from django.shortcuts import render
from django.conf import settings
def oauthtest(request):
return render(request, 'oauthtest.html', {
'link': '{}o/authorize/?response_type=code&client_id={}&redirect_uri={}{}/oauthdone/'.format(
settings.API_URL,
settings.OAUTH_CLIENT_ID,
'https://' if request.is_secure() else 'http://',
request.META['HTTP_HOST'],
)
})
def oauthdone(request):
# Get token
r = requests.post(settings.API_URL + 'o/token/', data={
'grant_type': 'authorization_code',
'code': request.GET['code'],
'redirect_uri': 'http://localhost:8000/oauthdone/',
'client_id': settings.OAUTH_CLIENT_ID,
'client_secret': settings.OAUTH_CLIENT_SECRET,
})
oauth = r.json()
# Get user object + accounts
headers = {
'Authorization': oauth['token_type'] + ' ' + oauth['access_token'],
}
r = requests.get(settings.API_URL + 'api/users/me/?expand_accounts=True', headers=headers)
return render(request, 'accounts.html', {
'user': r.json(),
})
<commit_msg>Fix redirect URI in oauthdone<commit_after>import requests
from django.shortcuts import render
from django.conf import settings
def oauthtest(request):
return render(request, 'oauthtest.html', {
'link': '{}o/authorize/?response_type=code&client_id={}&redirect_uri={}{}/oauthdone/'.format(
settings.API_URL,
settings.OAUTH_CLIENT_ID,
'https://' if request.is_secure() else 'http://',
request.META['HTTP_HOST'],
)
})
def oauthdone(request):
# Get token
r = requests.post(settings.API_URL + 'o/token/', data={
'grant_type': 'authorization_code',
'code': request.GET['code'],
'redirect_uri': '{}{}/oauthdone/'.format('https://' if request.is_secure() else 'http://', request.META['HTTP_HOST']),
'client_id': settings.OAUTH_CLIENT_ID,
'client_secret': settings.OAUTH_CLIENT_SECRET,
})
oauth = r.json()
# Get user object + accounts
headers = {
'Authorization': oauth['token_type'] + ' ' + oauth['access_token'],
}
r = requests.get(settings.API_URL + 'api/users/me/?expand_accounts=True', headers=headers)
return render(request, 'accounts.html', {
'user': r.json(),
})
|
68a1877bcd4511008aeff977cb45fa9edb5e9a8b
|
fusekiutils/__init__.py
|
fusekiutils/__init__.py
|
__author__ = 'adam'
import time
from subprocess import Popen
import shlex
import os
import urllib
def LaunchFuseki():
fuseki_url = "http://localhost:3030"
fuseki_dir = os.getcwd() + "/jena-fuseki"
fuseki_executable = fuseki_dir + "/fuseki-server"
f_log = open("fuseki.log","w")
fuseki = Popen( args=shlex.split("-q --update --loc=../fuseki-data /qudt4dt"),
executable=fuseki_executable,
cwd=fuseki_dir,
stdout=f_log)
f_log.close()
PollFusekiLaunch("http://localhost:3030")
return fuseki
def PollFusekiLaunch(fuseki_url):
while True:
if IsFusekiRunning(fuseki_url):
return
else:
print "polling..."
time.sleep(1)
def IsFusekiRunning(fuseki_url):
try:
urllib.urlopen(fuseki_url)
return True
except IOError:
return False
|
__author__ = 'adam'
import time
from subprocess import Popen
import shlex
import os
import urllib
import sys
def LaunchFuseki():
fuseki_dir = os.path.join(os.path.abspath(os.getcwd()), 'jena-fuseki')
if sys.platform == 'win32':
fuseki_executable = os.path.join(fuseki_dir, 'fuseki-server.bat')
else:
fuseki_executable = os.path.join(fuseki_dir, 'fuseki-server')
fuseki_data = os.path.join('..', 'fuseki-data')
f_log = open("fuseki.log", "w")
args = '%(exec)s -q --update --loc=%(data)s /qudt4dt' % {'exec': fuseki_executable, 'data': fuseki_data}
fuseki = Popen(args=args,
cwd=fuseki_dir,
stdout=f_log)
f_log.close()
PollFusekiLaunch("http://localhost:3030")
return fuseki
def PollFusekiLaunch(fuseki_url):
while True:
if IsFusekiRunning(fuseki_url):
return
else:
print "polling..."
time.sleep(1)
def IsFusekiRunning(fuseki_url):
try:
urllib.urlopen(fuseki_url)
return True
except IOError:
return False
|
Support both windows and shell environments when launching fuseki
|
Support both windows and shell environments when launching fuseki
|
Python
|
lgpl-2.1
|
adamnagel/qudt-for-domain-tools,adamnagel/qudt-for-domain-tools,adamnagel/qudt-for-domain-tools
|
__author__ = 'adam'
import time
from subprocess import Popen
import shlex
import os
import urllib
def LaunchFuseki():
fuseki_url = "http://localhost:3030"
fuseki_dir = os.getcwd() + "/jena-fuseki"
fuseki_executable = fuseki_dir + "/fuseki-server"
f_log = open("fuseki.log","w")
fuseki = Popen( args=shlex.split("-q --update --loc=../fuseki-data /qudt4dt"),
executable=fuseki_executable,
cwd=fuseki_dir,
stdout=f_log)
f_log.close()
PollFusekiLaunch("http://localhost:3030")
return fuseki
def PollFusekiLaunch(fuseki_url):
while True:
if IsFusekiRunning(fuseki_url):
return
else:
print "polling..."
time.sleep(1)
def IsFusekiRunning(fuseki_url):
try:
urllib.urlopen(fuseki_url)
return True
except IOError:
return FalseSupport both windows and shell environments when launching fuseki
|
__author__ = 'adam'
import time
from subprocess import Popen
import shlex
import os
import urllib
import sys
def LaunchFuseki():
fuseki_dir = os.path.join(os.path.abspath(os.getcwd()), 'jena-fuseki')
if sys.platform == 'win32':
fuseki_executable = os.path.join(fuseki_dir, 'fuseki-server.bat')
else:
fuseki_executable = os.path.join(fuseki_dir, 'fuseki-server')
fuseki_data = os.path.join('..', 'fuseki-data')
f_log = open("fuseki.log", "w")
args = '%(exec)s -q --update --loc=%(data)s /qudt4dt' % {'exec': fuseki_executable, 'data': fuseki_data}
fuseki = Popen(args=args,
cwd=fuseki_dir,
stdout=f_log)
f_log.close()
PollFusekiLaunch("http://localhost:3030")
return fuseki
def PollFusekiLaunch(fuseki_url):
while True:
if IsFusekiRunning(fuseki_url):
return
else:
print "polling..."
time.sleep(1)
def IsFusekiRunning(fuseki_url):
try:
urllib.urlopen(fuseki_url)
return True
except IOError:
return False
|
<commit_before>__author__ = 'adam'
import time
from subprocess import Popen
import shlex
import os
import urllib
def LaunchFuseki():
fuseki_url = "http://localhost:3030"
fuseki_dir = os.getcwd() + "/jena-fuseki"
fuseki_executable = fuseki_dir + "/fuseki-server"
f_log = open("fuseki.log","w")
fuseki = Popen( args=shlex.split("-q --update --loc=../fuseki-data /qudt4dt"),
executable=fuseki_executable,
cwd=fuseki_dir,
stdout=f_log)
f_log.close()
PollFusekiLaunch("http://localhost:3030")
return fuseki
def PollFusekiLaunch(fuseki_url):
while True:
if IsFusekiRunning(fuseki_url):
return
else:
print "polling..."
time.sleep(1)
def IsFusekiRunning(fuseki_url):
try:
urllib.urlopen(fuseki_url)
return True
except IOError:
return False<commit_msg>Support both windows and shell environments when launching fuseki<commit_after>
|
__author__ = 'adam'
import time
from subprocess import Popen
import shlex
import os
import urllib
import sys
def LaunchFuseki():
fuseki_dir = os.path.join(os.path.abspath(os.getcwd()), 'jena-fuseki')
if sys.platform == 'win32':
fuseki_executable = os.path.join(fuseki_dir, 'fuseki-server.bat')
else:
fuseki_executable = os.path.join(fuseki_dir, 'fuseki-server')
fuseki_data = os.path.join('..', 'fuseki-data')
f_log = open("fuseki.log", "w")
args = '%(exec)s -q --update --loc=%(data)s /qudt4dt' % {'exec': fuseki_executable, 'data': fuseki_data}
fuseki = Popen(args=args,
cwd=fuseki_dir,
stdout=f_log)
f_log.close()
PollFusekiLaunch("http://localhost:3030")
return fuseki
def PollFusekiLaunch(fuseki_url):
while True:
if IsFusekiRunning(fuseki_url):
return
else:
print "polling..."
time.sleep(1)
def IsFusekiRunning(fuseki_url):
try:
urllib.urlopen(fuseki_url)
return True
except IOError:
return False
|
__author__ = 'adam'
import time
from subprocess import Popen
import shlex
import os
import urllib
def LaunchFuseki():
fuseki_url = "http://localhost:3030"
fuseki_dir = os.getcwd() + "/jena-fuseki"
fuseki_executable = fuseki_dir + "/fuseki-server"
f_log = open("fuseki.log","w")
fuseki = Popen( args=shlex.split("-q --update --loc=../fuseki-data /qudt4dt"),
executable=fuseki_executable,
cwd=fuseki_dir,
stdout=f_log)
f_log.close()
PollFusekiLaunch("http://localhost:3030")
return fuseki
def PollFusekiLaunch(fuseki_url):
while True:
if IsFusekiRunning(fuseki_url):
return
else:
print "polling..."
time.sleep(1)
def IsFusekiRunning(fuseki_url):
try:
urllib.urlopen(fuseki_url)
return True
except IOError:
return FalseSupport both windows and shell environments when launching fuseki__author__ = 'adam'
import time
from subprocess import Popen
import shlex
import os
import urllib
import sys
def LaunchFuseki():
fuseki_dir = os.path.join(os.path.abspath(os.getcwd()), 'jena-fuseki')
if sys.platform == 'win32':
fuseki_executable = os.path.join(fuseki_dir, 'fuseki-server.bat')
else:
fuseki_executable = os.path.join(fuseki_dir, 'fuseki-server')
fuseki_data = os.path.join('..', 'fuseki-data')
f_log = open("fuseki.log", "w")
args = '%(exec)s -q --update --loc=%(data)s /qudt4dt' % {'exec': fuseki_executable, 'data': fuseki_data}
fuseki = Popen(args=args,
cwd=fuseki_dir,
stdout=f_log)
f_log.close()
PollFusekiLaunch("http://localhost:3030")
return fuseki
def PollFusekiLaunch(fuseki_url):
while True:
if IsFusekiRunning(fuseki_url):
return
else:
print "polling..."
time.sleep(1)
def IsFusekiRunning(fuseki_url):
try:
urllib.urlopen(fuseki_url)
return True
except IOError:
return False
|
<commit_before>__author__ = 'adam'
import time
from subprocess import Popen
import shlex
import os
import urllib
def LaunchFuseki():
fuseki_url = "http://localhost:3030"
fuseki_dir = os.getcwd() + "/jena-fuseki"
fuseki_executable = fuseki_dir + "/fuseki-server"
f_log = open("fuseki.log","w")
fuseki = Popen( args=shlex.split("-q --update --loc=../fuseki-data /qudt4dt"),
executable=fuseki_executable,
cwd=fuseki_dir,
stdout=f_log)
f_log.close()
PollFusekiLaunch("http://localhost:3030")
return fuseki
def PollFusekiLaunch(fuseki_url):
while True:
if IsFusekiRunning(fuseki_url):
return
else:
print "polling..."
time.sleep(1)
def IsFusekiRunning(fuseki_url):
try:
urllib.urlopen(fuseki_url)
return True
except IOError:
return False<commit_msg>Support both windows and shell environments when launching fuseki<commit_after>__author__ = 'adam'
import time
from subprocess import Popen
import shlex
import os
import urllib
import sys
def LaunchFuseki():
fuseki_dir = os.path.join(os.path.abspath(os.getcwd()), 'jena-fuseki')
if sys.platform == 'win32':
fuseki_executable = os.path.join(fuseki_dir, 'fuseki-server.bat')
else:
fuseki_executable = os.path.join(fuseki_dir, 'fuseki-server')
fuseki_data = os.path.join('..', 'fuseki-data')
f_log = open("fuseki.log", "w")
args = '%(exec)s -q --update --loc=%(data)s /qudt4dt' % {'exec': fuseki_executable, 'data': fuseki_data}
fuseki = Popen(args=args,
cwd=fuseki_dir,
stdout=f_log)
f_log.close()
PollFusekiLaunch("http://localhost:3030")
return fuseki
def PollFusekiLaunch(fuseki_url):
while True:
if IsFusekiRunning(fuseki_url):
return
else:
print "polling..."
time.sleep(1)
def IsFusekiRunning(fuseki_url):
try:
urllib.urlopen(fuseki_url)
return True
except IOError:
return False
|
94ff1527fb16c7a3557112f6e30cded4de99dda8
|
fabtastic/fabric/commands/c_supervisord.py
|
fabtastic/fabric/commands/c_supervisord.py
|
from fabric.api import *
from fabtastic.fabric.util import _current_host_has_role
def supervisord_restart_all(roles='webapp_servers'):
"""
Restarts all of supervisord's managed programs.
"""
if _current_host_has_role(roles):
print("=== RESTARTING SUPERVISORD PROGRAMS ===")
with cd(env.REMOTE_CODEBASE_PATH):
run("workon %s && ./manage.py ft_supervisord_restart_prog" % env.REMOTE_VIRTUALENV_NAME)
def supervisord_restart_prog(program, roles='webapp_servers'):
"""
Restarts all of supervisord's managed programs.
:arg str program: The name of the program to restart (as per supervisor's
conf.d/ contents).
"""
if _current_host_has_role(roles):
print("=== RESTARTING SUPERVISORD PROGRAMS ===")
with cd(env.REMOTE_CODEBASE_PATH):
run("workon %s && ./manage.py ft_supervisord_restart_prog %s" % (
program, env.REMOTE_VIRTUALENV_NAME))
|
from fabric.api import *
from fabtastic.fabric.util import _current_host_has_role
def supervisord_restart_all(roles='webapp_servers'):
"""
Restarts all of supervisord's managed programs.
"""
if _current_host_has_role(roles):
print("=== RESTARTING SUPERVISORD PROGRAMS ===")
with cd(env.REMOTE_CODEBASE_PATH):
run("workon %s && ./manage.py ft_supervisord_restart_prog" % env.REMOTE_VIRTUALENV_NAME)
def supervisord_restart_prog(program, roles='webapp_servers'):
"""
Restarts all of supervisord's managed programs.
:arg str program: The name of the program to restart (as per supervisor's
conf.d/ contents).
"""
if _current_host_has_role(roles):
print("=== RESTARTING SUPERVISORD PROGRAMS ===")
with cd(env.REMOTE_CODEBASE_PATH):
run("workon %s && ./manage.py ft_supervisord_restart_prog %s" % (
env.REMOTE_VIRTUALENV_NAME, program))
|
Fix arg order for supervisord_restart_prog
|
Fix arg order for supervisord_restart_prog
|
Python
|
bsd-3-clause
|
duointeractive/django-fabtastic
|
from fabric.api import *
from fabtastic.fabric.util import _current_host_has_role
def supervisord_restart_all(roles='webapp_servers'):
"""
Restarts all of supervisord's managed programs.
"""
if _current_host_has_role(roles):
print("=== RESTARTING SUPERVISORD PROGRAMS ===")
with cd(env.REMOTE_CODEBASE_PATH):
run("workon %s && ./manage.py ft_supervisord_restart_prog" % env.REMOTE_VIRTUALENV_NAME)
def supervisord_restart_prog(program, roles='webapp_servers'):
"""
Restarts all of supervisord's managed programs.
:arg str program: The name of the program to restart (as per supervisor's
conf.d/ contents).
"""
if _current_host_has_role(roles):
print("=== RESTARTING SUPERVISORD PROGRAMS ===")
with cd(env.REMOTE_CODEBASE_PATH):
run("workon %s && ./manage.py ft_supervisord_restart_prog %s" % (
program, env.REMOTE_VIRTUALENV_NAME))
Fix arg order for supervisord_restart_prog
|
from fabric.api import *
from fabtastic.fabric.util import _current_host_has_role
def supervisord_restart_all(roles='webapp_servers'):
"""
Restarts all of supervisord's managed programs.
"""
if _current_host_has_role(roles):
print("=== RESTARTING SUPERVISORD PROGRAMS ===")
with cd(env.REMOTE_CODEBASE_PATH):
run("workon %s && ./manage.py ft_supervisord_restart_prog" % env.REMOTE_VIRTUALENV_NAME)
def supervisord_restart_prog(program, roles='webapp_servers'):
"""
Restarts all of supervisord's managed programs.
:arg str program: The name of the program to restart (as per supervisor's
conf.d/ contents).
"""
if _current_host_has_role(roles):
print("=== RESTARTING SUPERVISORD PROGRAMS ===")
with cd(env.REMOTE_CODEBASE_PATH):
run("workon %s && ./manage.py ft_supervisord_restart_prog %s" % (
env.REMOTE_VIRTUALENV_NAME, program))
|
<commit_before>from fabric.api import *
from fabtastic.fabric.util import _current_host_has_role
def supervisord_restart_all(roles='webapp_servers'):
"""
Restarts all of supervisord's managed programs.
"""
if _current_host_has_role(roles):
print("=== RESTARTING SUPERVISORD PROGRAMS ===")
with cd(env.REMOTE_CODEBASE_PATH):
run("workon %s && ./manage.py ft_supervisord_restart_prog" % env.REMOTE_VIRTUALENV_NAME)
def supervisord_restart_prog(program, roles='webapp_servers'):
"""
Restarts all of supervisord's managed programs.
:arg str program: The name of the program to restart (as per supervisor's
conf.d/ contents).
"""
if _current_host_has_role(roles):
print("=== RESTARTING SUPERVISORD PROGRAMS ===")
with cd(env.REMOTE_CODEBASE_PATH):
run("workon %s && ./manage.py ft_supervisord_restart_prog %s" % (
program, env.REMOTE_VIRTUALENV_NAME))
<commit_msg>Fix arg order for supervisord_restart_prog<commit_after>
|
from fabric.api import *
from fabtastic.fabric.util import _current_host_has_role
def supervisord_restart_all(roles='webapp_servers'):
"""
Restarts all of supervisord's managed programs.
"""
if _current_host_has_role(roles):
print("=== RESTARTING SUPERVISORD PROGRAMS ===")
with cd(env.REMOTE_CODEBASE_PATH):
run("workon %s && ./manage.py ft_supervisord_restart_prog" % env.REMOTE_VIRTUALENV_NAME)
def supervisord_restart_prog(program, roles='webapp_servers'):
"""
Restarts all of supervisord's managed programs.
:arg str program: The name of the program to restart (as per supervisor's
conf.d/ contents).
"""
if _current_host_has_role(roles):
print("=== RESTARTING SUPERVISORD PROGRAMS ===")
with cd(env.REMOTE_CODEBASE_PATH):
run("workon %s && ./manage.py ft_supervisord_restart_prog %s" % (
env.REMOTE_VIRTUALENV_NAME, program))
|
from fabric.api import *
from fabtastic.fabric.util import _current_host_has_role
def supervisord_restart_all(roles='webapp_servers'):
"""
Restarts all of supervisord's managed programs.
"""
if _current_host_has_role(roles):
print("=== RESTARTING SUPERVISORD PROGRAMS ===")
with cd(env.REMOTE_CODEBASE_PATH):
run("workon %s && ./manage.py ft_supervisord_restart_prog" % env.REMOTE_VIRTUALENV_NAME)
def supervisord_restart_prog(program, roles='webapp_servers'):
"""
Restarts all of supervisord's managed programs.
:arg str program: The name of the program to restart (as per supervisor's
conf.d/ contents).
"""
if _current_host_has_role(roles):
print("=== RESTARTING SUPERVISORD PROGRAMS ===")
with cd(env.REMOTE_CODEBASE_PATH):
run("workon %s && ./manage.py ft_supervisord_restart_prog %s" % (
program, env.REMOTE_VIRTUALENV_NAME))
Fix arg order for supervisord_restart_progfrom fabric.api import *
from fabtastic.fabric.util import _current_host_has_role
def supervisord_restart_all(roles='webapp_servers'):
"""
Restarts all of supervisord's managed programs.
"""
if _current_host_has_role(roles):
print("=== RESTARTING SUPERVISORD PROGRAMS ===")
with cd(env.REMOTE_CODEBASE_PATH):
run("workon %s && ./manage.py ft_supervisord_restart_prog" % env.REMOTE_VIRTUALENV_NAME)
def supervisord_restart_prog(program, roles='webapp_servers'):
"""
Restarts all of supervisord's managed programs.
:arg str program: The name of the program to restart (as per supervisor's
conf.d/ contents).
"""
if _current_host_has_role(roles):
print("=== RESTARTING SUPERVISORD PROGRAMS ===")
with cd(env.REMOTE_CODEBASE_PATH):
run("workon %s && ./manage.py ft_supervisord_restart_prog %s" % (
env.REMOTE_VIRTUALENV_NAME, program))
|
<commit_before>from fabric.api import *
from fabtastic.fabric.util import _current_host_has_role
def supervisord_restart_all(roles='webapp_servers'):
"""
Restarts all of supervisord's managed programs.
"""
if _current_host_has_role(roles):
print("=== RESTARTING SUPERVISORD PROGRAMS ===")
with cd(env.REMOTE_CODEBASE_PATH):
run("workon %s && ./manage.py ft_supervisord_restart_prog" % env.REMOTE_VIRTUALENV_NAME)
def supervisord_restart_prog(program, roles='webapp_servers'):
"""
Restarts all of supervisord's managed programs.
:arg str program: The name of the program to restart (as per supervisor's
conf.d/ contents).
"""
if _current_host_has_role(roles):
print("=== RESTARTING SUPERVISORD PROGRAMS ===")
with cd(env.REMOTE_CODEBASE_PATH):
run("workon %s && ./manage.py ft_supervisord_restart_prog %s" % (
program, env.REMOTE_VIRTUALENV_NAME))
<commit_msg>Fix arg order for supervisord_restart_prog<commit_after>from fabric.api import *
from fabtastic.fabric.util import _current_host_has_role
def supervisord_restart_all(roles='webapp_servers'):
"""
Restarts all of supervisord's managed programs.
"""
if _current_host_has_role(roles):
print("=== RESTARTING SUPERVISORD PROGRAMS ===")
with cd(env.REMOTE_CODEBASE_PATH):
run("workon %s && ./manage.py ft_supervisord_restart_prog" % env.REMOTE_VIRTUALENV_NAME)
def supervisord_restart_prog(program, roles='webapp_servers'):
"""
Restarts all of supervisord's managed programs.
:arg str program: The name of the program to restart (as per supervisor's
conf.d/ contents).
"""
if _current_host_has_role(roles):
print("=== RESTARTING SUPERVISORD PROGRAMS ===")
with cd(env.REMOTE_CODEBASE_PATH):
run("workon %s && ./manage.py ft_supervisord_restart_prog %s" % (
env.REMOTE_VIRTUALENV_NAME, program))
|
2422e0eb14bc9ae0b79b88f9b02b7e9c7f6ee4fd
|
tests/window/window_util.py
|
tests/window/window_util.py
|
#!/usr/bin/python
# $Id:$
from pyglet.gl import *
def draw_client_border(window):
glClearColor(0, 0, 0, 1)
glClear(GL_COLOR_BUFFER_BIT)
glMatrixMode(GL_PROJECTION)
glLoadIdentity()
glOrtho(0, window.width, 0, window.height, -1, 1)
glMatrixMode(GL_MODELVIEW)
glLoadIdentity()
def rect(x1, y1, x2, y2):
glBegin(GL_LINE_LOOP)
glVertex2f(x1, y1)
glVertex2f(x2, y1)
glVertex2f(x2, y2)
glVertex2f(x1, y2)
glEnd()
glColor3f(1, 0, 0)
rect(-1, -1, window.width + 1, window.height - 1)
glColor3f(0, 1, 0)
rect(0, 0, window.width - 1, window.height - 1)
|
#!/usr/bin/python
# $Id:$
from pyglet.gl import *
def draw_client_border(window):
glClearColor(0, 0, 0, 1)
glClear(GL_COLOR_BUFFER_BIT)
glMatrixMode(GL_PROJECTION)
glLoadIdentity()
glOrtho(0, window.width, 0, window.height, -1, 1)
glMatrixMode(GL_MODELVIEW)
glLoadIdentity()
def rect(x1, y1, x2, y2):
glBegin(GL_LINE_LOOP)
glVertex2f(x1, y1)
glVertex2f(x2, y1)
glVertex2f(x2, y2)
glVertex2f(x1, y2)
glEnd()
glColor3f(1, 0, 0)
rect(-1, -1, window.width, window.height)
glColor3f(0, 1, 0)
rect(0, 0, window.width - 1, window.height - 1)
|
Fix window test border _again_ (more fixed).
|
Fix window test border _again_ (more fixed).
git-svn-id: d4fdfcd4de20a449196f78acc655f735742cd30d@1383 14d46d22-621c-0410-bb3d-6f67920f7d95
|
Python
|
bsd-3-clause
|
regular/pyglet-avbin-optimizations,regular/pyglet-avbin-optimizations,regular/pyglet-avbin-optimizations,regular/pyglet-avbin-optimizations
|
#!/usr/bin/python
# $Id:$
from pyglet.gl import *
def draw_client_border(window):
glClearColor(0, 0, 0, 1)
glClear(GL_COLOR_BUFFER_BIT)
glMatrixMode(GL_PROJECTION)
glLoadIdentity()
glOrtho(0, window.width, 0, window.height, -1, 1)
glMatrixMode(GL_MODELVIEW)
glLoadIdentity()
def rect(x1, y1, x2, y2):
glBegin(GL_LINE_LOOP)
glVertex2f(x1, y1)
glVertex2f(x2, y1)
glVertex2f(x2, y2)
glVertex2f(x1, y2)
glEnd()
glColor3f(1, 0, 0)
rect(-1, -1, window.width + 1, window.height - 1)
glColor3f(0, 1, 0)
rect(0, 0, window.width - 1, window.height - 1)
Fix window test border _again_ (more fixed).
git-svn-id: d4fdfcd4de20a449196f78acc655f735742cd30d@1383 14d46d22-621c-0410-bb3d-6f67920f7d95
|
#!/usr/bin/python
# $Id:$
from pyglet.gl import *
def draw_client_border(window):
glClearColor(0, 0, 0, 1)
glClear(GL_COLOR_BUFFER_BIT)
glMatrixMode(GL_PROJECTION)
glLoadIdentity()
glOrtho(0, window.width, 0, window.height, -1, 1)
glMatrixMode(GL_MODELVIEW)
glLoadIdentity()
def rect(x1, y1, x2, y2):
glBegin(GL_LINE_LOOP)
glVertex2f(x1, y1)
glVertex2f(x2, y1)
glVertex2f(x2, y2)
glVertex2f(x1, y2)
glEnd()
glColor3f(1, 0, 0)
rect(-1, -1, window.width, window.height)
glColor3f(0, 1, 0)
rect(0, 0, window.width - 1, window.height - 1)
|
<commit_before>#!/usr/bin/python
# $Id:$
from pyglet.gl import *
def draw_client_border(window):
glClearColor(0, 0, 0, 1)
glClear(GL_COLOR_BUFFER_BIT)
glMatrixMode(GL_PROJECTION)
glLoadIdentity()
glOrtho(0, window.width, 0, window.height, -1, 1)
glMatrixMode(GL_MODELVIEW)
glLoadIdentity()
def rect(x1, y1, x2, y2):
glBegin(GL_LINE_LOOP)
glVertex2f(x1, y1)
glVertex2f(x2, y1)
glVertex2f(x2, y2)
glVertex2f(x1, y2)
glEnd()
glColor3f(1, 0, 0)
rect(-1, -1, window.width + 1, window.height - 1)
glColor3f(0, 1, 0)
rect(0, 0, window.width - 1, window.height - 1)
<commit_msg>Fix window test border _again_ (more fixed).
git-svn-id: d4fdfcd4de20a449196f78acc655f735742cd30d@1383 14d46d22-621c-0410-bb3d-6f67920f7d95<commit_after>
|
#!/usr/bin/python
# $Id:$
from pyglet.gl import *
def draw_client_border(window):
glClearColor(0, 0, 0, 1)
glClear(GL_COLOR_BUFFER_BIT)
glMatrixMode(GL_PROJECTION)
glLoadIdentity()
glOrtho(0, window.width, 0, window.height, -1, 1)
glMatrixMode(GL_MODELVIEW)
glLoadIdentity()
def rect(x1, y1, x2, y2):
glBegin(GL_LINE_LOOP)
glVertex2f(x1, y1)
glVertex2f(x2, y1)
glVertex2f(x2, y2)
glVertex2f(x1, y2)
glEnd()
glColor3f(1, 0, 0)
rect(-1, -1, window.width, window.height)
glColor3f(0, 1, 0)
rect(0, 0, window.width - 1, window.height - 1)
|
#!/usr/bin/python
# $Id:$
from pyglet.gl import *
def draw_client_border(window):
glClearColor(0, 0, 0, 1)
glClear(GL_COLOR_BUFFER_BIT)
glMatrixMode(GL_PROJECTION)
glLoadIdentity()
glOrtho(0, window.width, 0, window.height, -1, 1)
glMatrixMode(GL_MODELVIEW)
glLoadIdentity()
def rect(x1, y1, x2, y2):
glBegin(GL_LINE_LOOP)
glVertex2f(x1, y1)
glVertex2f(x2, y1)
glVertex2f(x2, y2)
glVertex2f(x1, y2)
glEnd()
glColor3f(1, 0, 0)
rect(-1, -1, window.width + 1, window.height - 1)
glColor3f(0, 1, 0)
rect(0, 0, window.width - 1, window.height - 1)
Fix window test border _again_ (more fixed).
git-svn-id: d4fdfcd4de20a449196f78acc655f735742cd30d@1383 14d46d22-621c-0410-bb3d-6f67920f7d95#!/usr/bin/python
# $Id:$
from pyglet.gl import *
def draw_client_border(window):
glClearColor(0, 0, 0, 1)
glClear(GL_COLOR_BUFFER_BIT)
glMatrixMode(GL_PROJECTION)
glLoadIdentity()
glOrtho(0, window.width, 0, window.height, -1, 1)
glMatrixMode(GL_MODELVIEW)
glLoadIdentity()
def rect(x1, y1, x2, y2):
glBegin(GL_LINE_LOOP)
glVertex2f(x1, y1)
glVertex2f(x2, y1)
glVertex2f(x2, y2)
glVertex2f(x1, y2)
glEnd()
glColor3f(1, 0, 0)
rect(-1, -1, window.width, window.height)
glColor3f(0, 1, 0)
rect(0, 0, window.width - 1, window.height - 1)
|
<commit_before>#!/usr/bin/python
# $Id:$
from pyglet.gl import *
def draw_client_border(window):
glClearColor(0, 0, 0, 1)
glClear(GL_COLOR_BUFFER_BIT)
glMatrixMode(GL_PROJECTION)
glLoadIdentity()
glOrtho(0, window.width, 0, window.height, -1, 1)
glMatrixMode(GL_MODELVIEW)
glLoadIdentity()
def rect(x1, y1, x2, y2):
glBegin(GL_LINE_LOOP)
glVertex2f(x1, y1)
glVertex2f(x2, y1)
glVertex2f(x2, y2)
glVertex2f(x1, y2)
glEnd()
glColor3f(1, 0, 0)
rect(-1, -1, window.width + 1, window.height - 1)
glColor3f(0, 1, 0)
rect(0, 0, window.width - 1, window.height - 1)
<commit_msg>Fix window test border _again_ (more fixed).
git-svn-id: d4fdfcd4de20a449196f78acc655f735742cd30d@1383 14d46d22-621c-0410-bb3d-6f67920f7d95<commit_after>#!/usr/bin/python
# $Id:$
from pyglet.gl import *
def draw_client_border(window):
glClearColor(0, 0, 0, 1)
glClear(GL_COLOR_BUFFER_BIT)
glMatrixMode(GL_PROJECTION)
glLoadIdentity()
glOrtho(0, window.width, 0, window.height, -1, 1)
glMatrixMode(GL_MODELVIEW)
glLoadIdentity()
def rect(x1, y1, x2, y2):
glBegin(GL_LINE_LOOP)
glVertex2f(x1, y1)
glVertex2f(x2, y1)
glVertex2f(x2, y2)
glVertex2f(x1, y2)
glEnd()
glColor3f(1, 0, 0)
rect(-1, -1, window.width, window.height)
glColor3f(0, 1, 0)
rect(0, 0, window.width - 1, window.height - 1)
|
86008628f7bff187c956273fbf6f15376ab861d1
|
src/sgeparse/query.py
|
src/sgeparse/query.py
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import subprocess as sp
from .parser import JobsParser
def get_jobs():
xml_text = fetch_xml()
parser = JobsParser(xml_text)
return parser.jobs
def fetch_xml(user=None):
cmd = ['qstat', '-xml']
if user is not None:
cmd.extend(['-u', user])
return sp.check_output(list(map(str, cmd)))
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import subprocess as sp
from .parser import JobsParser
def get_jobs(user=None):
xml_text = fetch_xml(user=user)
parser = JobsParser(xml_text)
return parser.jobs
def fetch_xml(user=None):
cmd = ['qstat', '-xml']
if user is not None:
cmd.extend(['-u', user])
return sp.check_output(list(map(str, cmd)))
|
Add user argument to get_jobs
|
Add user argument to get_jobs
|
Python
|
mit
|
mindriot101/sgeparse
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import subprocess as sp
from .parser import JobsParser
def get_jobs():
xml_text = fetch_xml()
parser = JobsParser(xml_text)
return parser.jobs
def fetch_xml(user=None):
cmd = ['qstat', '-xml']
if user is not None:
cmd.extend(['-u', user])
return sp.check_output(list(map(str, cmd)))
Add user argument to get_jobs
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import subprocess as sp
from .parser import JobsParser
def get_jobs(user=None):
xml_text = fetch_xml(user=user)
parser = JobsParser(xml_text)
return parser.jobs
def fetch_xml(user=None):
cmd = ['qstat', '-xml']
if user is not None:
cmd.extend(['-u', user])
return sp.check_output(list(map(str, cmd)))
|
<commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
import subprocess as sp
from .parser import JobsParser
def get_jobs():
xml_text = fetch_xml()
parser = JobsParser(xml_text)
return parser.jobs
def fetch_xml(user=None):
cmd = ['qstat', '-xml']
if user is not None:
cmd.extend(['-u', user])
return sp.check_output(list(map(str, cmd)))
<commit_msg>Add user argument to get_jobs<commit_after>
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import subprocess as sp
from .parser import JobsParser
def get_jobs(user=None):
xml_text = fetch_xml(user=user)
parser = JobsParser(xml_text)
return parser.jobs
def fetch_xml(user=None):
cmd = ['qstat', '-xml']
if user is not None:
cmd.extend(['-u', user])
return sp.check_output(list(map(str, cmd)))
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import subprocess as sp
from .parser import JobsParser
def get_jobs():
xml_text = fetch_xml()
parser = JobsParser(xml_text)
return parser.jobs
def fetch_xml(user=None):
cmd = ['qstat', '-xml']
if user is not None:
cmd.extend(['-u', user])
return sp.check_output(list(map(str, cmd)))
Add user argument to get_jobs#!/usr/bin/env python
# -*- coding: utf-8 -*-
import subprocess as sp
from .parser import JobsParser
def get_jobs(user=None):
xml_text = fetch_xml(user=user)
parser = JobsParser(xml_text)
return parser.jobs
def fetch_xml(user=None):
cmd = ['qstat', '-xml']
if user is not None:
cmd.extend(['-u', user])
return sp.check_output(list(map(str, cmd)))
|
<commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
import subprocess as sp
from .parser import JobsParser
def get_jobs():
xml_text = fetch_xml()
parser = JobsParser(xml_text)
return parser.jobs
def fetch_xml(user=None):
cmd = ['qstat', '-xml']
if user is not None:
cmd.extend(['-u', user])
return sp.check_output(list(map(str, cmd)))
<commit_msg>Add user argument to get_jobs<commit_after>#!/usr/bin/env python
# -*- coding: utf-8 -*-
import subprocess as sp
from .parser import JobsParser
def get_jobs(user=None):
xml_text = fetch_xml(user=user)
parser = JobsParser(xml_text)
return parser.jobs
def fetch_xml(user=None):
cmd = ['qstat', '-xml']
if user is not None:
cmd.extend(['-u', user])
return sp.check_output(list(map(str, cmd)))
|
43a209bd122329d5a70e5f0bdc2066e952676c6a
|
tests/unit/output/yaml_out_test.py
|
tests/unit/output/yaml_out_test.py
|
# -*- coding: utf-8 -*-
'''
unittests for yaml outputter
'''
# Import Python Libs
from __future__ import absolute_import
from StringIO import StringIO
import sys
# Import Salt Testing Libs
from salttesting import TestCase
from salttesting.helpers import ensure_in_syspath
ensure_in_syspath('../../')
# Import Salt Libs
from salt.output import yaml_out as yaml
class YamlTestCase(TestCase):
'''
Test cases for salt.output.json_out
'''
def setUp(self):
# reset to default behavior
yaml.__opts__ = {}
self.data = {'test': 'two', 'example': 'one'}
def test_default_output(self):
ret = yaml.output(self.data)
expect = 'example: one\ntest: two\n'
self.assertEqual(expect, ret)
def test_negative_int_output(self):
yaml.__opts__['output_indent'] = -1
ret = yaml.output(self.data)
expect = '{example: one, test: two}\n'
self.assertEqual(expect, ret)
if __name__ == '__main__':
from integration import run_tests
run_tests(YamlTestCase, needs_daemon=False)
|
# -*- coding: utf-8 -*-
'''
unittests for yaml outputter
'''
# Import Python Libs
from __future__ import absolute_import
# Import Salt Testing Libs
from salttesting import TestCase
from salttesting.helpers import ensure_in_syspath
ensure_in_syspath('../../')
# Import Salt Libs
from salt.output import yaml_out as yaml
class YamlTestCase(TestCase):
'''
Test cases for salt.output.json_out
'''
def setUp(self):
# reset to default behavior
yaml.__opts__ = {}
self.data = {'test': 'two', 'example': 'one'}
def test_default_output(self):
ret = yaml.output(self.data)
expect = 'example: one\ntest: two\n'
self.assertEqual(expect, ret)
def test_negative_int_output(self):
yaml.__opts__['output_indent'] = -1
ret = yaml.output(self.data)
expect = '{example: one, test: two}\n'
self.assertEqual(expect, ret)
if __name__ == '__main__':
from integration import run_tests
run_tests(YamlTestCase, needs_daemon=False)
|
Remove unused imports for lint
|
Remove unused imports for lint
|
Python
|
apache-2.0
|
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
|
# -*- coding: utf-8 -*-
'''
unittests for yaml outputter
'''
# Import Python Libs
from __future__ import absolute_import
from StringIO import StringIO
import sys
# Import Salt Testing Libs
from salttesting import TestCase
from salttesting.helpers import ensure_in_syspath
ensure_in_syspath('../../')
# Import Salt Libs
from salt.output import yaml_out as yaml
class YamlTestCase(TestCase):
'''
Test cases for salt.output.json_out
'''
def setUp(self):
# reset to default behavior
yaml.__opts__ = {}
self.data = {'test': 'two', 'example': 'one'}
def test_default_output(self):
ret = yaml.output(self.data)
expect = 'example: one\ntest: two\n'
self.assertEqual(expect, ret)
def test_negative_int_output(self):
yaml.__opts__['output_indent'] = -1
ret = yaml.output(self.data)
expect = '{example: one, test: two}\n'
self.assertEqual(expect, ret)
if __name__ == '__main__':
from integration import run_tests
run_tests(YamlTestCase, needs_daemon=False)
Remove unused imports for lint
|
# -*- coding: utf-8 -*-
'''
unittests for yaml outputter
'''
# Import Python Libs
from __future__ import absolute_import
# Import Salt Testing Libs
from salttesting import TestCase
from salttesting.helpers import ensure_in_syspath
ensure_in_syspath('../../')
# Import Salt Libs
from salt.output import yaml_out as yaml
class YamlTestCase(TestCase):
'''
Test cases for salt.output.json_out
'''
def setUp(self):
# reset to default behavior
yaml.__opts__ = {}
self.data = {'test': 'two', 'example': 'one'}
def test_default_output(self):
ret = yaml.output(self.data)
expect = 'example: one\ntest: two\n'
self.assertEqual(expect, ret)
def test_negative_int_output(self):
yaml.__opts__['output_indent'] = -1
ret = yaml.output(self.data)
expect = '{example: one, test: two}\n'
self.assertEqual(expect, ret)
if __name__ == '__main__':
from integration import run_tests
run_tests(YamlTestCase, needs_daemon=False)
|
<commit_before># -*- coding: utf-8 -*-
'''
unittests for yaml outputter
'''
# Import Python Libs
from __future__ import absolute_import
from StringIO import StringIO
import sys
# Import Salt Testing Libs
from salttesting import TestCase
from salttesting.helpers import ensure_in_syspath
ensure_in_syspath('../../')
# Import Salt Libs
from salt.output import yaml_out as yaml
class YamlTestCase(TestCase):
'''
Test cases for salt.output.json_out
'''
def setUp(self):
# reset to default behavior
yaml.__opts__ = {}
self.data = {'test': 'two', 'example': 'one'}
def test_default_output(self):
ret = yaml.output(self.data)
expect = 'example: one\ntest: two\n'
self.assertEqual(expect, ret)
def test_negative_int_output(self):
yaml.__opts__['output_indent'] = -1
ret = yaml.output(self.data)
expect = '{example: one, test: two}\n'
self.assertEqual(expect, ret)
if __name__ == '__main__':
from integration import run_tests
run_tests(YamlTestCase, needs_daemon=False)
<commit_msg>Remove unused imports for lint<commit_after>
|
# -*- coding: utf-8 -*-
'''
unittests for yaml outputter
'''
# Import Python Libs
from __future__ import absolute_import
# Import Salt Testing Libs
from salttesting import TestCase
from salttesting.helpers import ensure_in_syspath
ensure_in_syspath('../../')
# Import Salt Libs
from salt.output import yaml_out as yaml
class YamlTestCase(TestCase):
'''
Test cases for salt.output.json_out
'''
def setUp(self):
# reset to default behavior
yaml.__opts__ = {}
self.data = {'test': 'two', 'example': 'one'}
def test_default_output(self):
ret = yaml.output(self.data)
expect = 'example: one\ntest: two\n'
self.assertEqual(expect, ret)
def test_negative_int_output(self):
yaml.__opts__['output_indent'] = -1
ret = yaml.output(self.data)
expect = '{example: one, test: two}\n'
self.assertEqual(expect, ret)
if __name__ == '__main__':
from integration import run_tests
run_tests(YamlTestCase, needs_daemon=False)
|
# -*- coding: utf-8 -*-
'''
unittests for yaml outputter
'''
# Import Python Libs
from __future__ import absolute_import
from StringIO import StringIO
import sys
# Import Salt Testing Libs
from salttesting import TestCase
from salttesting.helpers import ensure_in_syspath
ensure_in_syspath('../../')
# Import Salt Libs
from salt.output import yaml_out as yaml
class YamlTestCase(TestCase):
'''
Test cases for salt.output.json_out
'''
def setUp(self):
# reset to default behavior
yaml.__opts__ = {}
self.data = {'test': 'two', 'example': 'one'}
def test_default_output(self):
ret = yaml.output(self.data)
expect = 'example: one\ntest: two\n'
self.assertEqual(expect, ret)
def test_negative_int_output(self):
yaml.__opts__['output_indent'] = -1
ret = yaml.output(self.data)
expect = '{example: one, test: two}\n'
self.assertEqual(expect, ret)
if __name__ == '__main__':
from integration import run_tests
run_tests(YamlTestCase, needs_daemon=False)
Remove unused imports for lint# -*- coding: utf-8 -*-
'''
unittests for yaml outputter
'''
# Import Python Libs
from __future__ import absolute_import
# Import Salt Testing Libs
from salttesting import TestCase
from salttesting.helpers import ensure_in_syspath
ensure_in_syspath('../../')
# Import Salt Libs
from salt.output import yaml_out as yaml
class YamlTestCase(TestCase):
'''
Test cases for salt.output.json_out
'''
def setUp(self):
# reset to default behavior
yaml.__opts__ = {}
self.data = {'test': 'two', 'example': 'one'}
def test_default_output(self):
ret = yaml.output(self.data)
expect = 'example: one\ntest: two\n'
self.assertEqual(expect, ret)
def test_negative_int_output(self):
yaml.__opts__['output_indent'] = -1
ret = yaml.output(self.data)
expect = '{example: one, test: two}\n'
self.assertEqual(expect, ret)
if __name__ == '__main__':
from integration import run_tests
run_tests(YamlTestCase, needs_daemon=False)
|
<commit_before># -*- coding: utf-8 -*-
'''
unittests for yaml outputter
'''
# Import Python Libs
from __future__ import absolute_import
from StringIO import StringIO
import sys
# Import Salt Testing Libs
from salttesting import TestCase
from salttesting.helpers import ensure_in_syspath
ensure_in_syspath('../../')
# Import Salt Libs
from salt.output import yaml_out as yaml
class YamlTestCase(TestCase):
'''
Test cases for salt.output.json_out
'''
def setUp(self):
# reset to default behavior
yaml.__opts__ = {}
self.data = {'test': 'two', 'example': 'one'}
def test_default_output(self):
ret = yaml.output(self.data)
expect = 'example: one\ntest: two\n'
self.assertEqual(expect, ret)
def test_negative_int_output(self):
yaml.__opts__['output_indent'] = -1
ret = yaml.output(self.data)
expect = '{example: one, test: two}\n'
self.assertEqual(expect, ret)
if __name__ == '__main__':
from integration import run_tests
run_tests(YamlTestCase, needs_daemon=False)
<commit_msg>Remove unused imports for lint<commit_after># -*- coding: utf-8 -*-
'''
unittests for yaml outputter
'''
# Import Python Libs
from __future__ import absolute_import
# Import Salt Testing Libs
from salttesting import TestCase
from salttesting.helpers import ensure_in_syspath
ensure_in_syspath('../../')
# Import Salt Libs
from salt.output import yaml_out as yaml
class YamlTestCase(TestCase):
'''
Test cases for salt.output.json_out
'''
def setUp(self):
# reset to default behavior
yaml.__opts__ = {}
self.data = {'test': 'two', 'example': 'one'}
def test_default_output(self):
ret = yaml.output(self.data)
expect = 'example: one\ntest: two\n'
self.assertEqual(expect, ret)
def test_negative_int_output(self):
yaml.__opts__['output_indent'] = -1
ret = yaml.output(self.data)
expect = '{example: one, test: two}\n'
self.assertEqual(expect, ret)
if __name__ == '__main__':
from integration import run_tests
run_tests(YamlTestCase, needs_daemon=False)
|
b970f230864b40eaddb8e5faa76538c9f8e5c59c
|
txircd/modules/rfc/cmd_userhost.py
|
txircd/modules/rfc/cmd_userhost.py
|
from twisted.plugin import IPlugin
from twisted.words.protocols import irc
from txircd.module_interface import Command, ICommand, IModuleData, ModuleData
from zope.interface import implements
class UserhostCommand(ModuleData, Command):
implements(IPlugin, IModuleData, ICommand)
name = "UserhostCommand"
core = True
def userCommands(self):
return [ ("USERHOST", 1, self) ]
def parseParams(self, user, params, prefix, tags):
if not params:
user.sendSingleError("UserhostParams", irc.ERR_NEEDMOREPARAMS, "USERHOST", "Not enough parameters")
return None
return {
"nicks": params[:5]
}
def execute(self, user, data):
userHosts = []
for nick in data["nicks"]:
if nick not in self.ircd.userNicks:
continue
targetUser = self.ircd.users[self.ircd.userNicks[nick]]
output = targetUser.nick
if self.ircd.runActionUntilValue("userhasoperpermission", targetUser, ""):
output += "*"
output += "="
if user.metadataKeyExists("away"):
output += "-"
else:
output += "+"
output += "{}@{}".format(targetUser.ident, targetUser.host())
userHosts.append(output)
user.sendMessage(irc.RPL_USERHOST, " ".join(userHosts))
return True
userhostCmd = UserhostCommand()
|
from twisted.plugin import IPlugin
from twisted.words.protocols import irc
from txircd.module_interface import Command, ICommand, IModuleData, ModuleData
from zope.interface import implements
class UserhostCommand(ModuleData, Command):
implements(IPlugin, IModuleData, ICommand)
name = "UserhostCommand"
core = True
def userCommands(self):
return [ ("USERHOST", 1, self) ]
def parseParams(self, user, params, prefix, tags):
if not params:
user.sendSingleError("UserhostParams", irc.ERR_NEEDMOREPARAMS, "USERHOST", "Not enough parameters")
return None
return {
"nicks": params[:5]
}
def execute(self, user, data):
userHosts = []
for nick in data["nicks"]:
if nick not in self.ircd.userNicks:
continue
targetUser = self.ircd.users[self.ircd.userNicks[nick]]
output = targetUser.nick
if self.ircd.runActionUntilValue("userhasoperpermission", targetUser, "", users=[targetUser]):
output += "*"
output += "="
if user.metadataKeyExists("away"):
output += "-"
else:
output += "+"
output += "{}@{}".format(targetUser.ident, targetUser.host())
userHosts.append(output)
user.sendMessage(irc.RPL_USERHOST, " ".join(userHosts))
return True
userhostCmd = UserhostCommand()
|
Add affected users to userhasoperpermission call in USERHOST
|
Add affected users to userhasoperpermission call in USERHOST
|
Python
|
bsd-3-clause
|
Heufneutje/txircd,ElementalAlchemist/txircd
|
from twisted.plugin import IPlugin
from twisted.words.protocols import irc
from txircd.module_interface import Command, ICommand, IModuleData, ModuleData
from zope.interface import implements
class UserhostCommand(ModuleData, Command):
implements(IPlugin, IModuleData, ICommand)
name = "UserhostCommand"
core = True
def userCommands(self):
return [ ("USERHOST", 1, self) ]
def parseParams(self, user, params, prefix, tags):
if not params:
user.sendSingleError("UserhostParams", irc.ERR_NEEDMOREPARAMS, "USERHOST", "Not enough parameters")
return None
return {
"nicks": params[:5]
}
def execute(self, user, data):
userHosts = []
for nick in data["nicks"]:
if nick not in self.ircd.userNicks:
continue
targetUser = self.ircd.users[self.ircd.userNicks[nick]]
output = targetUser.nick
if self.ircd.runActionUntilValue("userhasoperpermission", targetUser, ""):
output += "*"
output += "="
if user.metadataKeyExists("away"):
output += "-"
else:
output += "+"
output += "{}@{}".format(targetUser.ident, targetUser.host())
userHosts.append(output)
user.sendMessage(irc.RPL_USERHOST, " ".join(userHosts))
return True
userhostCmd = UserhostCommand()Add affected users to userhasoperpermission call in USERHOST
|
from twisted.plugin import IPlugin
from twisted.words.protocols import irc
from txircd.module_interface import Command, ICommand, IModuleData, ModuleData
from zope.interface import implements
class UserhostCommand(ModuleData, Command):
implements(IPlugin, IModuleData, ICommand)
name = "UserhostCommand"
core = True
def userCommands(self):
return [ ("USERHOST", 1, self) ]
def parseParams(self, user, params, prefix, tags):
if not params:
user.sendSingleError("UserhostParams", irc.ERR_NEEDMOREPARAMS, "USERHOST", "Not enough parameters")
return None
return {
"nicks": params[:5]
}
def execute(self, user, data):
userHosts = []
for nick in data["nicks"]:
if nick not in self.ircd.userNicks:
continue
targetUser = self.ircd.users[self.ircd.userNicks[nick]]
output = targetUser.nick
if self.ircd.runActionUntilValue("userhasoperpermission", targetUser, "", users=[targetUser]):
output += "*"
output += "="
if user.metadataKeyExists("away"):
output += "-"
else:
output += "+"
output += "{}@{}".format(targetUser.ident, targetUser.host())
userHosts.append(output)
user.sendMessage(irc.RPL_USERHOST, " ".join(userHosts))
return True
userhostCmd = UserhostCommand()
|
<commit_before>from twisted.plugin import IPlugin
from twisted.words.protocols import irc
from txircd.module_interface import Command, ICommand, IModuleData, ModuleData
from zope.interface import implements
class UserhostCommand(ModuleData, Command):
implements(IPlugin, IModuleData, ICommand)
name = "UserhostCommand"
core = True
def userCommands(self):
return [ ("USERHOST", 1, self) ]
def parseParams(self, user, params, prefix, tags):
if not params:
user.sendSingleError("UserhostParams", irc.ERR_NEEDMOREPARAMS, "USERHOST", "Not enough parameters")
return None
return {
"nicks": params[:5]
}
def execute(self, user, data):
userHosts = []
for nick in data["nicks"]:
if nick not in self.ircd.userNicks:
continue
targetUser = self.ircd.users[self.ircd.userNicks[nick]]
output = targetUser.nick
if self.ircd.runActionUntilValue("userhasoperpermission", targetUser, ""):
output += "*"
output += "="
if user.metadataKeyExists("away"):
output += "-"
else:
output += "+"
output += "{}@{}".format(targetUser.ident, targetUser.host())
userHosts.append(output)
user.sendMessage(irc.RPL_USERHOST, " ".join(userHosts))
return True
userhostCmd = UserhostCommand()<commit_msg>Add affected users to userhasoperpermission call in USERHOST<commit_after>
|
from twisted.plugin import IPlugin
from twisted.words.protocols import irc
from txircd.module_interface import Command, ICommand, IModuleData, ModuleData
from zope.interface import implements
class UserhostCommand(ModuleData, Command):
implements(IPlugin, IModuleData, ICommand)
name = "UserhostCommand"
core = True
def userCommands(self):
return [ ("USERHOST", 1, self) ]
def parseParams(self, user, params, prefix, tags):
if not params:
user.sendSingleError("UserhostParams", irc.ERR_NEEDMOREPARAMS, "USERHOST", "Not enough parameters")
return None
return {
"nicks": params[:5]
}
def execute(self, user, data):
userHosts = []
for nick in data["nicks"]:
if nick not in self.ircd.userNicks:
continue
targetUser = self.ircd.users[self.ircd.userNicks[nick]]
output = targetUser.nick
if self.ircd.runActionUntilValue("userhasoperpermission", targetUser, "", users=[targetUser]):
output += "*"
output += "="
if user.metadataKeyExists("away"):
output += "-"
else:
output += "+"
output += "{}@{}".format(targetUser.ident, targetUser.host())
userHosts.append(output)
user.sendMessage(irc.RPL_USERHOST, " ".join(userHosts))
return True
userhostCmd = UserhostCommand()
|
from twisted.plugin import IPlugin
from twisted.words.protocols import irc
from txircd.module_interface import Command, ICommand, IModuleData, ModuleData
from zope.interface import implements
class UserhostCommand(ModuleData, Command):
implements(IPlugin, IModuleData, ICommand)
name = "UserhostCommand"
core = True
def userCommands(self):
return [ ("USERHOST", 1, self) ]
def parseParams(self, user, params, prefix, tags):
if not params:
user.sendSingleError("UserhostParams", irc.ERR_NEEDMOREPARAMS, "USERHOST", "Not enough parameters")
return None
return {
"nicks": params[:5]
}
def execute(self, user, data):
userHosts = []
for nick in data["nicks"]:
if nick not in self.ircd.userNicks:
continue
targetUser = self.ircd.users[self.ircd.userNicks[nick]]
output = targetUser.nick
if self.ircd.runActionUntilValue("userhasoperpermission", targetUser, ""):
output += "*"
output += "="
if user.metadataKeyExists("away"):
output += "-"
else:
output += "+"
output += "{}@{}".format(targetUser.ident, targetUser.host())
userHosts.append(output)
user.sendMessage(irc.RPL_USERHOST, " ".join(userHosts))
return True
userhostCmd = UserhostCommand()Add affected users to userhasoperpermission call in USERHOSTfrom twisted.plugin import IPlugin
from twisted.words.protocols import irc
from txircd.module_interface import Command, ICommand, IModuleData, ModuleData
from zope.interface import implements
class UserhostCommand(ModuleData, Command):
implements(IPlugin, IModuleData, ICommand)
name = "UserhostCommand"
core = True
def userCommands(self):
return [ ("USERHOST", 1, self) ]
def parseParams(self, user, params, prefix, tags):
if not params:
user.sendSingleError("UserhostParams", irc.ERR_NEEDMOREPARAMS, "USERHOST", "Not enough parameters")
return None
return {
"nicks": params[:5]
}
def execute(self, user, data):
userHosts = []
for nick in data["nicks"]:
if nick not in self.ircd.userNicks:
continue
targetUser = self.ircd.users[self.ircd.userNicks[nick]]
output = targetUser.nick
if self.ircd.runActionUntilValue("userhasoperpermission", targetUser, "", users=[targetUser]):
output += "*"
output += "="
if user.metadataKeyExists("away"):
output += "-"
else:
output += "+"
output += "{}@{}".format(targetUser.ident, targetUser.host())
userHosts.append(output)
user.sendMessage(irc.RPL_USERHOST, " ".join(userHosts))
return True
userhostCmd = UserhostCommand()
|
<commit_before>from twisted.plugin import IPlugin
from twisted.words.protocols import irc
from txircd.module_interface import Command, ICommand, IModuleData, ModuleData
from zope.interface import implements
class UserhostCommand(ModuleData, Command):
implements(IPlugin, IModuleData, ICommand)
name = "UserhostCommand"
core = True
def userCommands(self):
return [ ("USERHOST", 1, self) ]
def parseParams(self, user, params, prefix, tags):
if not params:
user.sendSingleError("UserhostParams", irc.ERR_NEEDMOREPARAMS, "USERHOST", "Not enough parameters")
return None
return {
"nicks": params[:5]
}
def execute(self, user, data):
userHosts = []
for nick in data["nicks"]:
if nick not in self.ircd.userNicks:
continue
targetUser = self.ircd.users[self.ircd.userNicks[nick]]
output = targetUser.nick
if self.ircd.runActionUntilValue("userhasoperpermission", targetUser, ""):
output += "*"
output += "="
if user.metadataKeyExists("away"):
output += "-"
else:
output += "+"
output += "{}@{}".format(targetUser.ident, targetUser.host())
userHosts.append(output)
user.sendMessage(irc.RPL_USERHOST, " ".join(userHosts))
return True
userhostCmd = UserhostCommand()<commit_msg>Add affected users to userhasoperpermission call in USERHOST<commit_after>from twisted.plugin import IPlugin
from twisted.words.protocols import irc
from txircd.module_interface import Command, ICommand, IModuleData, ModuleData
from zope.interface import implements
class UserhostCommand(ModuleData, Command):
implements(IPlugin, IModuleData, ICommand)
name = "UserhostCommand"
core = True
def userCommands(self):
return [ ("USERHOST", 1, self) ]
def parseParams(self, user, params, prefix, tags):
if not params:
user.sendSingleError("UserhostParams", irc.ERR_NEEDMOREPARAMS, "USERHOST", "Not enough parameters")
return None
return {
"nicks": params[:5]
}
def execute(self, user, data):
userHosts = []
for nick in data["nicks"]:
if nick not in self.ircd.userNicks:
continue
targetUser = self.ircd.users[self.ircd.userNicks[nick]]
output = targetUser.nick
if self.ircd.runActionUntilValue("userhasoperpermission", targetUser, "", users=[targetUser]):
output += "*"
output += "="
if user.metadataKeyExists("away"):
output += "-"
else:
output += "+"
output += "{}@{}".format(targetUser.ident, targetUser.host())
userHosts.append(output)
user.sendMessage(irc.RPL_USERHOST, " ".join(userHosts))
return True
userhostCmd = UserhostCommand()
|
4c58426a88ba056841b1d1b44536f2f85de120cc
|
pythonx/completers/javascript/__init__.py
|
pythonx/completers/javascript/__init__.py
|
# -*- coding: utf-8 -*-
import json
import os.path
import re
from completor import Completor
from completor.compat import to_unicode
dirname = os.path.dirname(__file__)
class Tern(Completor):
filetype = 'javascript'
daemon = True
ident = re.compile(r"""(\w+)|(('|").+)""", re.U)
trigger = r"""\w+$|[\w\)\]\}\'\"]+\.\w*$|('|").*$"""
def format_cmd(self):
binary = self.get_option('node_binary') or 'node'
tern_config = self.find_config_file('.tern-project')
cmd = [binary, os.path.join(dirname, 'tern_wrapper.js')]
if tern_config:
cmd.append(os.path.dirname(tern_config))
return cmd
def parse(self, data):
try:
data = to_unicode(data[0], 'utf-8')
return [i for i in json.loads(data)
if not self.input_data.endswith(i['word'])]
except Exception:
return []
|
# -*- coding: utf-8 -*-
import json
import os.path
import re
from completor import Completor
from completor.compat import to_unicode
dirname = os.path.dirname(__file__)
class Tern(Completor):
filetype = 'javascript'
daemon = True
ident = re.compile(r"""(\w+)|(["'][^"']*)""", re.U)
trigger = r"""\w+$|[\w\)\]\}\'\"]+\.\w*$|["']\w*$"""
def format_cmd(self):
binary = self.get_option('node_binary') or 'node'
tern_config = self.find_config_file('.tern-project')
cmd = [binary, os.path.join(dirname, 'tern_wrapper.js')]
if tern_config:
cmd.append(os.path.dirname(tern_config))
return cmd
def parse(self, data):
try:
data = to_unicode(data[0], 'utf-8')
return [i for i in json.loads(data)
if not self.input_data.endswith(i['word'])]
except Exception:
return []
|
Fix regex for tern complete_strings plugin
|
Fix regex for tern complete_strings plugin
|
Python
|
mit
|
maralla/completor.vim,maralla/completor.vim
|
# -*- coding: utf-8 -*-
import json
import os.path
import re
from completor import Completor
from completor.compat import to_unicode
dirname = os.path.dirname(__file__)
class Tern(Completor):
filetype = 'javascript'
daemon = True
ident = re.compile(r"""(\w+)|(('|").+)""", re.U)
trigger = r"""\w+$|[\w\)\]\}\'\"]+\.\w*$|('|").*$"""
def format_cmd(self):
binary = self.get_option('node_binary') or 'node'
tern_config = self.find_config_file('.tern-project')
cmd = [binary, os.path.join(dirname, 'tern_wrapper.js')]
if tern_config:
cmd.append(os.path.dirname(tern_config))
return cmd
def parse(self, data):
try:
data = to_unicode(data[0], 'utf-8')
return [i for i in json.loads(data)
if not self.input_data.endswith(i['word'])]
except Exception:
return []
Fix regex for tern complete_strings plugin
|
# -*- coding: utf-8 -*-
import json
import os.path
import re
from completor import Completor
from completor.compat import to_unicode
dirname = os.path.dirname(__file__)
class Tern(Completor):
filetype = 'javascript'
daemon = True
ident = re.compile(r"""(\w+)|(["'][^"']*)""", re.U)
trigger = r"""\w+$|[\w\)\]\}\'\"]+\.\w*$|["']\w*$"""
def format_cmd(self):
binary = self.get_option('node_binary') or 'node'
tern_config = self.find_config_file('.tern-project')
cmd = [binary, os.path.join(dirname, 'tern_wrapper.js')]
if tern_config:
cmd.append(os.path.dirname(tern_config))
return cmd
def parse(self, data):
try:
data = to_unicode(data[0], 'utf-8')
return [i for i in json.loads(data)
if not self.input_data.endswith(i['word'])]
except Exception:
return []
|
<commit_before># -*- coding: utf-8 -*-
import json
import os.path
import re
from completor import Completor
from completor.compat import to_unicode
dirname = os.path.dirname(__file__)
class Tern(Completor):
filetype = 'javascript'
daemon = True
ident = re.compile(r"""(\w+)|(('|").+)""", re.U)
trigger = r"""\w+$|[\w\)\]\}\'\"]+\.\w*$|('|").*$"""
def format_cmd(self):
binary = self.get_option('node_binary') or 'node'
tern_config = self.find_config_file('.tern-project')
cmd = [binary, os.path.join(dirname, 'tern_wrapper.js')]
if tern_config:
cmd.append(os.path.dirname(tern_config))
return cmd
def parse(self, data):
try:
data = to_unicode(data[0], 'utf-8')
return [i for i in json.loads(data)
if not self.input_data.endswith(i['word'])]
except Exception:
return []
<commit_msg>Fix regex for tern complete_strings plugin<commit_after>
|
# -*- coding: utf-8 -*-
import json
import os.path
import re
from completor import Completor
from completor.compat import to_unicode
dirname = os.path.dirname(__file__)
class Tern(Completor):
filetype = 'javascript'
daemon = True
ident = re.compile(r"""(\w+)|(["'][^"']*)""", re.U)
trigger = r"""\w+$|[\w\)\]\}\'\"]+\.\w*$|["']\w*$"""
def format_cmd(self):
binary = self.get_option('node_binary') or 'node'
tern_config = self.find_config_file('.tern-project')
cmd = [binary, os.path.join(dirname, 'tern_wrapper.js')]
if tern_config:
cmd.append(os.path.dirname(tern_config))
return cmd
def parse(self, data):
try:
data = to_unicode(data[0], 'utf-8')
return [i for i in json.loads(data)
if not self.input_data.endswith(i['word'])]
except Exception:
return []
|
# -*- coding: utf-8 -*-
import json
import os.path
import re
from completor import Completor
from completor.compat import to_unicode
dirname = os.path.dirname(__file__)
class Tern(Completor):
filetype = 'javascript'
daemon = True
ident = re.compile(r"""(\w+)|(('|").+)""", re.U)
trigger = r"""\w+$|[\w\)\]\}\'\"]+\.\w*$|('|").*$"""
def format_cmd(self):
binary = self.get_option('node_binary') or 'node'
tern_config = self.find_config_file('.tern-project')
cmd = [binary, os.path.join(dirname, 'tern_wrapper.js')]
if tern_config:
cmd.append(os.path.dirname(tern_config))
return cmd
def parse(self, data):
try:
data = to_unicode(data[0], 'utf-8')
return [i for i in json.loads(data)
if not self.input_data.endswith(i['word'])]
except Exception:
return []
Fix regex for tern complete_strings plugin# -*- coding: utf-8 -*-
import json
import os.path
import re
from completor import Completor
from completor.compat import to_unicode
dirname = os.path.dirname(__file__)
class Tern(Completor):
filetype = 'javascript'
daemon = True
ident = re.compile(r"""(\w+)|(["'][^"']*)""", re.U)
trigger = r"""\w+$|[\w\)\]\}\'\"]+\.\w*$|["']\w*$"""
def format_cmd(self):
binary = self.get_option('node_binary') or 'node'
tern_config = self.find_config_file('.tern-project')
cmd = [binary, os.path.join(dirname, 'tern_wrapper.js')]
if tern_config:
cmd.append(os.path.dirname(tern_config))
return cmd
def parse(self, data):
try:
data = to_unicode(data[0], 'utf-8')
return [i for i in json.loads(data)
if not self.input_data.endswith(i['word'])]
except Exception:
return []
|
<commit_before># -*- coding: utf-8 -*-
import json
import os.path
import re
from completor import Completor
from completor.compat import to_unicode
dirname = os.path.dirname(__file__)
class Tern(Completor):
filetype = 'javascript'
daemon = True
ident = re.compile(r"""(\w+)|(('|").+)""", re.U)
trigger = r"""\w+$|[\w\)\]\}\'\"]+\.\w*$|('|").*$"""
def format_cmd(self):
binary = self.get_option('node_binary') or 'node'
tern_config = self.find_config_file('.tern-project')
cmd = [binary, os.path.join(dirname, 'tern_wrapper.js')]
if tern_config:
cmd.append(os.path.dirname(tern_config))
return cmd
def parse(self, data):
try:
data = to_unicode(data[0], 'utf-8')
return [i for i in json.loads(data)
if not self.input_data.endswith(i['word'])]
except Exception:
return []
<commit_msg>Fix regex for tern complete_strings plugin<commit_after># -*- coding: utf-8 -*-
import json
import os.path
import re
from completor import Completor
from completor.compat import to_unicode
dirname = os.path.dirname(__file__)
class Tern(Completor):
filetype = 'javascript'
daemon = True
ident = re.compile(r"""(\w+)|(["'][^"']*)""", re.U)
trigger = r"""\w+$|[\w\)\]\}\'\"]+\.\w*$|["']\w*$"""
def format_cmd(self):
binary = self.get_option('node_binary') or 'node'
tern_config = self.find_config_file('.tern-project')
cmd = [binary, os.path.join(dirname, 'tern_wrapper.js')]
if tern_config:
cmd.append(os.path.dirname(tern_config))
return cmd
def parse(self, data):
try:
data = to_unicode(data[0], 'utf-8')
return [i for i in json.loads(data)
if not self.input_data.endswith(i['word'])]
except Exception:
return []
|
2890660ee3e87eb9af2c81caac0dc3131a264310
|
app.py
|
app.py
|
from flask import Flask
from flask import jsonify
from red_api import get_suspect_mentions, get_tweet_count, get_suspect_tweets
from jsonp_flask import support_jsonp
app = Flask(__name__)
@app.route('/api/tweets/suspects/count', methods=['GET'])
@support_jsonp
def get_mentions():
return jsonify(results=get_suspect_mentions())
@app.route('/api/tweets/count', methods=['GET'])
@support_jsonp
def get_total_tweet_count():
return jsonify(results=get_tweet_count())
@app.route('/api/tweets/<suspect>/<limit>')
@support_jsonp
def get_tweets(suspect, limit):
return jsonify(results=get_suspect_tweets(suspect, limit))
if __name__ == '__main__':
app.run(debug=True, host='0.0.0.0')
|
from flask import Flask
from flask import jsonify
from red_api import get_suspect_mentions, get_tweet_count, get_suspect_tweets
from jsonp_flask import support_jsonp
app = Flask(__name__)
@app.route('/api/tweets/suspects/count', methods=['GET'])
@support_jsonp
def get_mentions():
return jsonify(results=get_suspect_mentions())
@app.route('/api/tweets/count', methods=['GET'])
@support_jsonp
def get_total_tweet_count():
return jsonify(results=get_tweet_count())
@app.route('/api/tweets/<suspect>/<limit>')
@support_jsonp
def get_tweets(suspect, limit):
return jsonify(results=get_suspect_tweets(suspect, int(limit)))
if __name__ == '__main__':
app.run(debug=True, host='0.0.0.0')
|
Make sure the limit is an int
|
Make sure the limit is an int
|
Python
|
mit
|
AnSavvides/redjohn,AnSavvides/redjohn
|
from flask import Flask
from flask import jsonify
from red_api import get_suspect_mentions, get_tweet_count, get_suspect_tweets
from jsonp_flask import support_jsonp
app = Flask(__name__)
@app.route('/api/tweets/suspects/count', methods=['GET'])
@support_jsonp
def get_mentions():
return jsonify(results=get_suspect_mentions())
@app.route('/api/tweets/count', methods=['GET'])
@support_jsonp
def get_total_tweet_count():
return jsonify(results=get_tweet_count())
@app.route('/api/tweets/<suspect>/<limit>')
@support_jsonp
def get_tweets(suspect, limit):
return jsonify(results=get_suspect_tweets(suspect, limit))
if __name__ == '__main__':
app.run(debug=True, host='0.0.0.0')Make sure the limit is an int
|
from flask import Flask
from flask import jsonify
from red_api import get_suspect_mentions, get_tweet_count, get_suspect_tweets
from jsonp_flask import support_jsonp
app = Flask(__name__)
@app.route('/api/tweets/suspects/count', methods=['GET'])
@support_jsonp
def get_mentions():
return jsonify(results=get_suspect_mentions())
@app.route('/api/tweets/count', methods=['GET'])
@support_jsonp
def get_total_tweet_count():
return jsonify(results=get_tweet_count())
@app.route('/api/tweets/<suspect>/<limit>')
@support_jsonp
def get_tweets(suspect, limit):
return jsonify(results=get_suspect_tweets(suspect, int(limit)))
if __name__ == '__main__':
app.run(debug=True, host='0.0.0.0')
|
<commit_before>from flask import Flask
from flask import jsonify
from red_api import get_suspect_mentions, get_tweet_count, get_suspect_tweets
from jsonp_flask import support_jsonp
app = Flask(__name__)
@app.route('/api/tweets/suspects/count', methods=['GET'])
@support_jsonp
def get_mentions():
return jsonify(results=get_suspect_mentions())
@app.route('/api/tweets/count', methods=['GET'])
@support_jsonp
def get_total_tweet_count():
return jsonify(results=get_tweet_count())
@app.route('/api/tweets/<suspect>/<limit>')
@support_jsonp
def get_tweets(suspect, limit):
return jsonify(results=get_suspect_tweets(suspect, limit))
if __name__ == '__main__':
app.run(debug=True, host='0.0.0.0')<commit_msg>Make sure the limit is an int<commit_after>
|
from flask import Flask
from flask import jsonify
from red_api import get_suspect_mentions, get_tweet_count, get_suspect_tweets
from jsonp_flask import support_jsonp
app = Flask(__name__)
@app.route('/api/tweets/suspects/count', methods=['GET'])
@support_jsonp
def get_mentions():
return jsonify(results=get_suspect_mentions())
@app.route('/api/tweets/count', methods=['GET'])
@support_jsonp
def get_total_tweet_count():
return jsonify(results=get_tweet_count())
@app.route('/api/tweets/<suspect>/<limit>')
@support_jsonp
def get_tweets(suspect, limit):
return jsonify(results=get_suspect_tweets(suspect, int(limit)))
if __name__ == '__main__':
app.run(debug=True, host='0.0.0.0')
|
from flask import Flask
from flask import jsonify
from red_api import get_suspect_mentions, get_tweet_count, get_suspect_tweets
from jsonp_flask import support_jsonp
app = Flask(__name__)
@app.route('/api/tweets/suspects/count', methods=['GET'])
@support_jsonp
def get_mentions():
return jsonify(results=get_suspect_mentions())
@app.route('/api/tweets/count', methods=['GET'])
@support_jsonp
def get_total_tweet_count():
return jsonify(results=get_tweet_count())
@app.route('/api/tweets/<suspect>/<limit>')
@support_jsonp
def get_tweets(suspect, limit):
return jsonify(results=get_suspect_tweets(suspect, limit))
if __name__ == '__main__':
app.run(debug=True, host='0.0.0.0')Make sure the limit is an intfrom flask import Flask
from flask import jsonify
from red_api import get_suspect_mentions, get_tweet_count, get_suspect_tweets
from jsonp_flask import support_jsonp
app = Flask(__name__)
@app.route('/api/tweets/suspects/count', methods=['GET'])
@support_jsonp
def get_mentions():
return jsonify(results=get_suspect_mentions())
@app.route('/api/tweets/count', methods=['GET'])
@support_jsonp
def get_total_tweet_count():
return jsonify(results=get_tweet_count())
@app.route('/api/tweets/<suspect>/<limit>')
@support_jsonp
def get_tweets(suspect, limit):
return jsonify(results=get_suspect_tweets(suspect, int(limit)))
if __name__ == '__main__':
app.run(debug=True, host='0.0.0.0')
|
<commit_before>from flask import Flask
from flask import jsonify
from red_api import get_suspect_mentions, get_tweet_count, get_suspect_tweets
from jsonp_flask import support_jsonp
app = Flask(__name__)
@app.route('/api/tweets/suspects/count', methods=['GET'])
@support_jsonp
def get_mentions():
return jsonify(results=get_suspect_mentions())
@app.route('/api/tweets/count', methods=['GET'])
@support_jsonp
def get_total_tweet_count():
return jsonify(results=get_tweet_count())
@app.route('/api/tweets/<suspect>/<limit>')
@support_jsonp
def get_tweets(suspect, limit):
return jsonify(results=get_suspect_tweets(suspect, limit))
if __name__ == '__main__':
app.run(debug=True, host='0.0.0.0')<commit_msg>Make sure the limit is an int<commit_after>from flask import Flask
from flask import jsonify
from red_api import get_suspect_mentions, get_tweet_count, get_suspect_tweets
from jsonp_flask import support_jsonp
app = Flask(__name__)
@app.route('/api/tweets/suspects/count', methods=['GET'])
@support_jsonp
def get_mentions():
return jsonify(results=get_suspect_mentions())
@app.route('/api/tweets/count', methods=['GET'])
@support_jsonp
def get_total_tweet_count():
return jsonify(results=get_tweet_count())
@app.route('/api/tweets/<suspect>/<limit>')
@support_jsonp
def get_tweets(suspect, limit):
return jsonify(results=get_suspect_tweets(suspect, int(limit)))
if __name__ == '__main__':
app.run(debug=True, host='0.0.0.0')
|
7ac8ae993a30ce6ea221e2474df4a8eb7eada1ef
|
scrapy/trunk/scrapy/conf/core_settings.py
|
scrapy/trunk/scrapy/conf/core_settings.py
|
import scrapy
# Scrapy core settings
BOT_NAME = 'scrapy'
BOT_VERSION = scrapy.__version__
ENGINE_DEBUG = False
# Download configuration options
USER_AGENT = '%s/%s' % (BOT_NAME, BOT_VERSION)
DOWNLOAD_TIMEOUT = 180 # 3mins
CONCURRENT_DOMAINS = 8 # number of domains to scrape in parallel
REQUESTS_PER_DOMAIN = 8 # max simultaneous requests per domain
CACHE2_EXPIRATION_SECS = 48 * 60 * 60 # seconds while cached response is still valid (a negative value means "never expires")
LOG_ENABLED = True #
LOGLEVEL = 'DEBUG' # default loglevel
LOGFILE = None # None means sys.stderr by default
LOG_STDOUT = False #
DEFAULT_ITEM_CLASS = 'scrapy.item.ScrapedItem'
SCHEDULER = 'scrapy.core.scheduler.Scheduler'
MEMORYSTORE = 'scrapy.core.scheduler.MemoryStore'
PRIORITIZER = 'scrapy.core.prioritizers.RandomPrioritizer'
EXTENSIONS = []
# contrib.middleware.retry.RetryMiddleware default settings
RETRY_TIMES = 3
RETRY_HTTP_CODES = ['500', '503', '504', '400', '408', '200']
# amazon s3 images backend sometimes fails to authenticate or validate http protocol
RETRY_HTTP_CODES = RETRY_HTTP_CODES + ['505', '403']
|
import scrapy
# Scrapy core settings
BOT_NAME = 'scrapy'
BOT_VERSION = scrapy.__version__
ENGINE_DEBUG = False
# Download configuration options
USER_AGENT = '%s/%s' % (BOT_NAME, BOT_VERSION)
DOWNLOAD_TIMEOUT = 180 # 3mins
CONCURRENT_DOMAINS = 8 # number of domains to scrape in parallel
REQUESTS_PER_DOMAIN = 8 # max simultaneous requests per domain
CACHE2_EXPIRATION_SECS = 48 * 60 * 60 # seconds while cached response is still valid (a negative value means "never expires")
LOG_ENABLED = True #
LOGLEVEL = 'DEBUG' # default loglevel
LOGFILE = None # None means sys.stderr by default
LOG_STDOUT = False #
DEFAULT_ITEM_CLASS = 'scrapy.item.ScrapedItem'
SCHEDULER = 'scrapy.core.scheduler.Scheduler'
MEMORYSTORE = 'scrapy.core.scheduler.MemoryStore'
PRIORITIZER = 'scrapy.core.prioritizers.RandomPrioritizer'
EXTENSIONS = []
# contrib.middleware.retry.RetryMiddleware default settings
RETRY_TIMES = 3
RETRY_HTTP_CODES = ['500', '503', '504', '400', '408', '200']
|
Revert "add 505 and 403 to retry status codes due to amazon s3 random fails while uploading images"
|
Revert "add 505 and 403 to retry status codes due to amazon s3 random fails while uploading images"
This reverts changeset r457
--HG--
extra : convert_revision : svn%3Ab85faa78-f9eb-468e-a121-7cced6da292c%40461
|
Python
|
bsd-3-clause
|
AaronTao1990/scrapy,yidongliu/scrapy,ArturGaspar/scrapy,tagatac/scrapy,URXtech/scrapy,olafdietsche/scrapy,w495/scrapy,agusc/scrapy,Preetwinder/scrapy,foromer4/scrapy,zorojean/scrapy,xiao26/scrapy,kmike/scrapy,w495/scrapy,ENjOyAbLE1991/scrapy,Digenis/scrapy,haiiiiiyun/scrapy,eLRuLL/scrapy,wujuguang/scrapy,scrapy/scrapy,mlyundin/scrapy,jiezhu2007/scrapy,rolando-contrib/scrapy,taito/scrapy,umrashrf/scrapy,heamon7/scrapy,fpy171/scrapy,irwinlove/scrapy,kmike/scrapy,fontenele/scrapy,carlosp420/scrapy,redapple/scrapy,github-account-because-they-want-it/scrapy,webmakin/scrapy,nfunato/scrapy,rahulsharma1991/scrapy,cursesun/scrapy,foromer4/scrapy,rahul-c1/scrapy,tntC4stl3/scrapy,jc0n/scrapy,johnardavies/scrapy,ylcolala/scrapy,curita/scrapy,CodeJuan/scrapy,ashishnerkar1/scrapy,pombredanne/scrapy,agreen/scrapy,aivarsk/scrapy,zackslash/scrapy,finfish/scrapy,nett55/scrapy,elacuesta/scrapy,Cnfc19932/scrapy,scorphus/scrapy,elijah513/scrapy,beni55/scrapy,amboxer21/scrapy,liyy7/scrapy,taito/scrapy,cursesun/scrapy,Chenmxs/scrapy,Zephor5/scrapy,lacrazyboy/scrapy,redapple/scrapy,scrapy/scrapy,agusc/scrapy,AaronTao1990/scrapy,Parlin-Galanodel/scrapy,jeffreyjinfeng/scrapy,stenskjaer/scrapy,olafdietsche/scrapy,fafaman/scrapy,Geeglee/scrapy,Preetwinder/scrapy,rolando/scrapy,hwsyy/scrapy,famorted/scrapy,zhangtao11/scrapy,dracony/scrapy,rklabs/scrapy,Bourneer/scrapy,wzyuliyang/scrapy,livepy/scrapy,codebhendi/scrapy,fafaman/scrapy,rolando/scrapy,sigma-random/scrapy,Geeglee/scrapy,aivarsk/scrapy,ArturGaspar/scrapy,elacuesta/scrapy,Djlavoy/scrapy,github-account-because-they-want-it/scrapy,smaty1/scrapy,wujuguang/scrapy,Timeship/scrapy,mouadino/scrapy,curita/scrapy,codebhendi/scrapy,Timeship/scrapy,AaronTao1990/scrapy,ramiro/scrapy,CodeJuan/scrapy,wzyuliyang/scrapy,nguyenhongson03/scrapy,hyrole/scrapy,kalessin/scrapy,JacobStevenR/scrapy,famorted/scrapy,dgillis/scrapy,hansenDise/scrapy,wangjun/scrapy,TarasRudnyk/scrapy,dracony/scrapy,Cnfc19932/scrapy,rahulsharma1991/scrapy,nguyenhongson03/scrapy,CodeJuan/scrapy,Chenmxs/scrapy,z-fork/scrapy,ENjOyAbLE1991/scrapy,umrashrf/scrapy,finfish/scrapy,farhan0581/scrapy,yusofm/scrapy,rootAvish/scrapy,mgedmin/scrapy,KublaikhanGeek/scrapy,devGregA/scrapy,crasker/scrapy,xiao26/scrapy,pranjalpatil/scrapy,IvanGavran/scrapy,liyy7/scrapy,elacuesta/scrapy,kimimj/scrapy,emschorsch/scrapy,jiezhu2007/scrapy,barraponto/scrapy,eliasdorneles/scrapy,pfctdayelise/scrapy,KublaikhanGeek/scrapy,rdowinton/scrapy,godfreyy/scrapy,nfunato/scrapy,beni55/scrapy,z-fork/scrapy,heamon7/scrapy,smaty1/scrapy,livepy/scrapy,jorik041/scrapy,fontenele/scrapy,jc0n/scrapy,Djlavoy/scrapy,songfj/scrapy,Geeglee/scrapy,tntC4stl3/scrapy,famorted/scrapy,fpy171/scrapy,irwinlove/scrapy,z-fork/scrapy,amboxer21/scrapy,kimimj/scrapy,fqul/scrapy,xiao26/scrapy,coderabhishek/scrapy,johnardavies/scrapy,Partoo/scrapy,ssteo/scrapy,yarikoptic/scrapy,wujuguang/scrapy,hyrole/scrapy,cyrixhero/scrapy,jorik041/scrapy,cleydson/scrapy,taito/scrapy,huoxudong125/scrapy,IvanGavran/scrapy,crasker/scrapy,nfunato/scrapy,Cnfc19932/scrapy,olorz/scrapy,Slater-Victoroff/scrapy,avtoritet/scrapy,hyrole/scrapy,CENDARI/scrapy,OpenWhere/scrapy,ssh-odoo/scrapy,rolando-contrib/scrapy,ssh-odoo/scrapy,csalazar/scrapy,Lucifer-Kim/scrapy,YeelerG/scrapy,gbirke/scrapy,CENDARI/scrapy,profjrr/scrapy,farhan0581/scrapy,cyrixhero/scrapy,devGregA/scrapy,ylcolala/scrapy,arush0311/scrapy,fontenele/scrapy,hbwzhsh/scrapy,kalessin/scrapy,JacobStevenR/scrapy,pombredanne/scrapy,Slater-Victoroff/scrapy,dacjames/scrapy,ashishnerkar1/scrapy,zackslash/scrapy,dangra/scrapy,Digenis/scrapy,rdowinton/scrapy,emschorsch/scrapy,fafaman/scrapy,csalazar/scrapy,rootAvish/scrapy,beni55/scrapy,hansenDise/scrapy,darkrho/scrapy-scrapy,GregoryVigoTorres/scrapy,rolando-contrib/scrapy,wenyu1001/scrapy,darkrho/scrapy-scrapy,nowopen/scrapy,agusc/scrapy,Slater-Victoroff/scrapy,huoxudong125/scrapy,barraponto/scrapy,yusofm/scrapy,curita/scrapy,webmakin/scrapy,devGregA/scrapy,joshlk/scrapy,rdowinton/scrapy,zjuwangg/scrapy,shaform/scrapy,joshlk/scrapy,olorz/scrapy,zhangtao11/scrapy,eLRuLL/scrapy,sigma-random/scrapy,haiiiiiyun/scrapy,irwinlove/scrapy,chekunkov/scrapy,jeffreyjinfeng/scrapy,cursesun/scrapy,jdemaeyer/scrapy,sardok/scrapy,TarasRudnyk/scrapy,songfj/scrapy,zjuwangg/scrapy,dhenyjarasandy/scrapy,ramiro/scrapy,eliasdorneles/scrapy,huoxudong125/scrapy,yusofm/scrapy,heamon7/scrapy,pranjalpatil/scrapy,elijah513/scrapy,rahul-c1/scrapy,dgillis/scrapy,tliber/scrapy,YeelerG/scrapy,zjuwangg/scrapy,nett55/scrapy,WilliamKinaan/scrapy,raphaelfruneaux/scrapy,haiiiiiyun/scrapy,dhenyjarasandy/scrapy,github-account-because-they-want-it/scrapy,wangjun/scrapy,WilliamKinaan/scrapy,cleydson/scrapy,pawelmhm/scrapy,farhan0581/scrapy,stenskjaer/scrapy,Ryezhang/scrapy,godfreyy/scrapy,Bourneer/scrapy,kimimj/scrapy,eliasdorneles/scrapy,tntC4stl3/scrapy,bmess/scrapy,1yvT0s/scrapy,moraesnicol/scrapy,snowdream1314/scrapy,Partoo/scrapy,nowopen/scrapy,jc0n/scrapy,eLRuLL/scrapy,dgillis/scrapy,kmike/scrapy,dacjames/scrapy,rolando/scrapy,moraesnicol/scrapy,rklabs/scrapy,JacobStevenR/scrapy,Allianzcortex/scrapy,Djlavoy/scrapy,Digenis/scrapy,liyy7/scrapy,starrify/scrapy,olorz/scrapy,finfish/scrapy,hectoruelo/scrapy,chekunkov/scrapy,yidongliu/scrapy,sardok/scrapy,mlyundin/scrapy,ssteo/scrapy,WilliamKinaan/scrapy,legendtkl/scrapy,snowdream1314/scrapy,nguyenhongson03/scrapy,jamesblunt/scrapy,legendtkl/scrapy,wenyu1001/scrapy,kashyap32/scrapy,moraesnicol/scrapy,jiezhu2007/scrapy,livepy/scrapy,tagatac/scrapy,wangjun/scrapy,codebhendi/scrapy,avtoritet/scrapy,raphaelfruneaux/scrapy,pombredanne/scrapy,csalazar/scrapy,tagatac/scrapy,kazitanvirahsan/scrapy,smaty1/scrapy,fqul/scrapy,Parlin-Galanodel/scrapy,lacrazyboy/scrapy,kashyap32/scrapy,shaform/scrapy,arush0311/scrapy,bmess/scrapy,Partoo/scrapy,YeelerG/scrapy,cyberplant/scrapy,wenyu1001/scrapy,cyberplant/scrapy,hwsyy/scrapy,gbirke/scrapy,johnardavies/scrapy,darkrho/scrapy-scrapy,mgedmin/scrapy,GregoryVigoTorres/scrapy,fpy171/scrapy,olafdietsche/scrapy,starrify/scrapy,mgedmin/scrapy,agreen/scrapy,yidongliu/scrapy,URXtech/scrapy,Lucifer-Kim/scrapy,kazitanvirahsan/scrapy,webmakin/scrapy,kashyap32/scrapy,ylcolala/scrapy,kalessin/scrapy,dangra/scrapy,zhangtao11/scrapy,KublaikhanGeek/scrapy,ENjOyAbLE1991/scrapy,jorik041/scrapy,nowopen/scrapy,scorphus/scrapy,yarikoptic/scrapy,Chenmxs/scrapy,mlyundin/scrapy,rklabs/scrapy,hectoruelo/scrapy,nett55/scrapy,chekunkov/scrapy,1yvT0s/scrapy,zorojean/scrapy,Allianzcortex/scrapy,stenskjaer/scrapy,cleydson/scrapy,dracony/scrapy,TarasRudnyk/scrapy,hbwzhsh/scrapy,Ryezhang/scrapy,ArturGaspar/scrapy,avtoritet/scrapy,starrify/scrapy,Adai0808/scrapy-1,Lucifer-Kim/scrapy,rootAvish/scrapy,nikgr95/scrapy,pablohoffman/scrapy,ssteo/scrapy,ndemir/scrapy,profjrr/scrapy,pablohoffman/scrapy,gnemoug/scrapy,scrapy/scrapy,profjrr/scrapy,scorphus/scrapy,joshlk/scrapy,rahul-c1/scrapy,jeffreyjinfeng/scrapy,cyrixhero/scrapy,shaform/scrapy,raphaelfruneaux/scrapy,arush0311/scrapy,1yvT0s/scrapy,tliber/scrapy,yarikoptic/scrapy,pablohoffman/scrapy,godfreyy/scrapy,jamesblunt/scrapy,Zephor5/scrapy,kazitanvirahsan/scrapy,Parlin-Galanodel/scrapy,Timeship/scrapy,agreen/scrapy,cyberplant/scrapy,coderabhishek/scrapy,carlosp420/scrapy,zackslash/scrapy,hwsyy/scrapy,OpenWhere/scrapy,redapple/scrapy,Allianzcortex/scrapy,fqul/scrapy,Adai0808/scrapy-1,dhenyjarasandy/scrapy,wzyuliyang/scrapy,w495/scrapy,snowdream1314/scrapy,ramiro/scrapy,bmess/scrapy,dangra/scrapy,IvanGavran/scrapy,pawelmhm/scrapy,URXtech/scrapy,elijah513/scrapy,gnemoug/scrapy,dacjames/scrapy,jdemaeyer/scrapy,rahulsharma1991/scrapy,Zephor5/scrapy,nikgr95/scrapy,CENDARI/scrapy,songfj/scrapy,nikgr95/scrapy,ndemir/scrapy,lacrazyboy/scrapy,ssh-odoo/scrapy,hectoruelo/scrapy,barraponto/scrapy,Preetwinder/scrapy,GregoryVigoTorres/scrapy,Adai0808/scrapy-1,OpenWhere/scrapy,umrashrf/scrapy,zorojean/scrapy,legendtkl/scrapy,amboxer21/scrapy,hansenDise/scrapy,carlosp420/scrapy,hbwzhsh/scrapy,pfctdayelise/scrapy,crasker/scrapy,jdemaeyer/scrapy,pfctdayelise/scrapy,Bourneer/scrapy,aivarsk/scrapy,mouadino/scrapy,tliber/scrapy,foromer4/scrapy,pranjalpatil/scrapy,pawelmhm/scrapy,emschorsch/scrapy,Ryezhang/scrapy,coderabhishek/scrapy
|
import scrapy
# Scrapy core settings
BOT_NAME = 'scrapy'
BOT_VERSION = scrapy.__version__
ENGINE_DEBUG = False
# Download configuration options
USER_AGENT = '%s/%s' % (BOT_NAME, BOT_VERSION)
DOWNLOAD_TIMEOUT = 180 # 3mins
CONCURRENT_DOMAINS = 8 # number of domains to scrape in parallel
REQUESTS_PER_DOMAIN = 8 # max simultaneous requests per domain
CACHE2_EXPIRATION_SECS = 48 * 60 * 60 # seconds while cached response is still valid (a negative value means "never expires")
LOG_ENABLED = True #
LOGLEVEL = 'DEBUG' # default loglevel
LOGFILE = None # None means sys.stderr by default
LOG_STDOUT = False #
DEFAULT_ITEM_CLASS = 'scrapy.item.ScrapedItem'
SCHEDULER = 'scrapy.core.scheduler.Scheduler'
MEMORYSTORE = 'scrapy.core.scheduler.MemoryStore'
PRIORITIZER = 'scrapy.core.prioritizers.RandomPrioritizer'
EXTENSIONS = []
# contrib.middleware.retry.RetryMiddleware default settings
RETRY_TIMES = 3
RETRY_HTTP_CODES = ['500', '503', '504', '400', '408', '200']
# amazon s3 images backend sometimes fails to authenticate or validate http protocol
RETRY_HTTP_CODES = RETRY_HTTP_CODES + ['505', '403']
Revert "add 505 and 403 to retry status codes due to amazon s3 random fails while uploading images"
This reverts changeset r457
--HG--
extra : convert_revision : svn%3Ab85faa78-f9eb-468e-a121-7cced6da292c%40461
|
import scrapy
# Scrapy core settings
BOT_NAME = 'scrapy'
BOT_VERSION = scrapy.__version__
ENGINE_DEBUG = False
# Download configuration options
USER_AGENT = '%s/%s' % (BOT_NAME, BOT_VERSION)
DOWNLOAD_TIMEOUT = 180 # 3mins
CONCURRENT_DOMAINS = 8 # number of domains to scrape in parallel
REQUESTS_PER_DOMAIN = 8 # max simultaneous requests per domain
CACHE2_EXPIRATION_SECS = 48 * 60 * 60 # seconds while cached response is still valid (a negative value means "never expires")
LOG_ENABLED = True #
LOGLEVEL = 'DEBUG' # default loglevel
LOGFILE = None # None means sys.stderr by default
LOG_STDOUT = False #
DEFAULT_ITEM_CLASS = 'scrapy.item.ScrapedItem'
SCHEDULER = 'scrapy.core.scheduler.Scheduler'
MEMORYSTORE = 'scrapy.core.scheduler.MemoryStore'
PRIORITIZER = 'scrapy.core.prioritizers.RandomPrioritizer'
EXTENSIONS = []
# contrib.middleware.retry.RetryMiddleware default settings
RETRY_TIMES = 3
RETRY_HTTP_CODES = ['500', '503', '504', '400', '408', '200']
|
<commit_before>import scrapy
# Scrapy core settings
BOT_NAME = 'scrapy'
BOT_VERSION = scrapy.__version__
ENGINE_DEBUG = False
# Download configuration options
USER_AGENT = '%s/%s' % (BOT_NAME, BOT_VERSION)
DOWNLOAD_TIMEOUT = 180 # 3mins
CONCURRENT_DOMAINS = 8 # number of domains to scrape in parallel
REQUESTS_PER_DOMAIN = 8 # max simultaneous requests per domain
CACHE2_EXPIRATION_SECS = 48 * 60 * 60 # seconds while cached response is still valid (a negative value means "never expires")
LOG_ENABLED = True #
LOGLEVEL = 'DEBUG' # default loglevel
LOGFILE = None # None means sys.stderr by default
LOG_STDOUT = False #
DEFAULT_ITEM_CLASS = 'scrapy.item.ScrapedItem'
SCHEDULER = 'scrapy.core.scheduler.Scheduler'
MEMORYSTORE = 'scrapy.core.scheduler.MemoryStore'
PRIORITIZER = 'scrapy.core.prioritizers.RandomPrioritizer'
EXTENSIONS = []
# contrib.middleware.retry.RetryMiddleware default settings
RETRY_TIMES = 3
RETRY_HTTP_CODES = ['500', '503', '504', '400', '408', '200']
# amazon s3 images backend sometimes fails to authenticate or validate http protocol
RETRY_HTTP_CODES = RETRY_HTTP_CODES + ['505', '403']
<commit_msg>Revert "add 505 and 403 to retry status codes due to amazon s3 random fails while uploading images"
This reverts changeset r457
--HG--
extra : convert_revision : svn%3Ab85faa78-f9eb-468e-a121-7cced6da292c%40461<commit_after>
|
import scrapy
# Scrapy core settings
BOT_NAME = 'scrapy'
BOT_VERSION = scrapy.__version__
ENGINE_DEBUG = False
# Download configuration options
USER_AGENT = '%s/%s' % (BOT_NAME, BOT_VERSION)
DOWNLOAD_TIMEOUT = 180 # 3mins
CONCURRENT_DOMAINS = 8 # number of domains to scrape in parallel
REQUESTS_PER_DOMAIN = 8 # max simultaneous requests per domain
CACHE2_EXPIRATION_SECS = 48 * 60 * 60 # seconds while cached response is still valid (a negative value means "never expires")
LOG_ENABLED = True #
LOGLEVEL = 'DEBUG' # default loglevel
LOGFILE = None # None means sys.stderr by default
LOG_STDOUT = False #
DEFAULT_ITEM_CLASS = 'scrapy.item.ScrapedItem'
SCHEDULER = 'scrapy.core.scheduler.Scheduler'
MEMORYSTORE = 'scrapy.core.scheduler.MemoryStore'
PRIORITIZER = 'scrapy.core.prioritizers.RandomPrioritizer'
EXTENSIONS = []
# contrib.middleware.retry.RetryMiddleware default settings
RETRY_TIMES = 3
RETRY_HTTP_CODES = ['500', '503', '504', '400', '408', '200']
|
import scrapy
# Scrapy core settings
BOT_NAME = 'scrapy'
BOT_VERSION = scrapy.__version__
ENGINE_DEBUG = False
# Download configuration options
USER_AGENT = '%s/%s' % (BOT_NAME, BOT_VERSION)
DOWNLOAD_TIMEOUT = 180 # 3mins
CONCURRENT_DOMAINS = 8 # number of domains to scrape in parallel
REQUESTS_PER_DOMAIN = 8 # max simultaneous requests per domain
CACHE2_EXPIRATION_SECS = 48 * 60 * 60 # seconds while cached response is still valid (a negative value means "never expires")
LOG_ENABLED = True #
LOGLEVEL = 'DEBUG' # default loglevel
LOGFILE = None # None means sys.stderr by default
LOG_STDOUT = False #
DEFAULT_ITEM_CLASS = 'scrapy.item.ScrapedItem'
SCHEDULER = 'scrapy.core.scheduler.Scheduler'
MEMORYSTORE = 'scrapy.core.scheduler.MemoryStore'
PRIORITIZER = 'scrapy.core.prioritizers.RandomPrioritizer'
EXTENSIONS = []
# contrib.middleware.retry.RetryMiddleware default settings
RETRY_TIMES = 3
RETRY_HTTP_CODES = ['500', '503', '504', '400', '408', '200']
# amazon s3 images backend sometimes fails to authenticate or validate http protocol
RETRY_HTTP_CODES = RETRY_HTTP_CODES + ['505', '403']
Revert "add 505 and 403 to retry status codes due to amazon s3 random fails while uploading images"
This reverts changeset r457
--HG--
extra : convert_revision : svn%3Ab85faa78-f9eb-468e-a121-7cced6da292c%40461import scrapy
# Scrapy core settings
BOT_NAME = 'scrapy'
BOT_VERSION = scrapy.__version__
ENGINE_DEBUG = False
# Download configuration options
USER_AGENT = '%s/%s' % (BOT_NAME, BOT_VERSION)
DOWNLOAD_TIMEOUT = 180 # 3mins
CONCURRENT_DOMAINS = 8 # number of domains to scrape in parallel
REQUESTS_PER_DOMAIN = 8 # max simultaneous requests per domain
CACHE2_EXPIRATION_SECS = 48 * 60 * 60 # seconds while cached response is still valid (a negative value means "never expires")
LOG_ENABLED = True #
LOGLEVEL = 'DEBUG' # default loglevel
LOGFILE = None # None means sys.stderr by default
LOG_STDOUT = False #
DEFAULT_ITEM_CLASS = 'scrapy.item.ScrapedItem'
SCHEDULER = 'scrapy.core.scheduler.Scheduler'
MEMORYSTORE = 'scrapy.core.scheduler.MemoryStore'
PRIORITIZER = 'scrapy.core.prioritizers.RandomPrioritizer'
EXTENSIONS = []
# contrib.middleware.retry.RetryMiddleware default settings
RETRY_TIMES = 3
RETRY_HTTP_CODES = ['500', '503', '504', '400', '408', '200']
|
<commit_before>import scrapy
# Scrapy core settings
BOT_NAME = 'scrapy'
BOT_VERSION = scrapy.__version__
ENGINE_DEBUG = False
# Download configuration options
USER_AGENT = '%s/%s' % (BOT_NAME, BOT_VERSION)
DOWNLOAD_TIMEOUT = 180 # 3mins
CONCURRENT_DOMAINS = 8 # number of domains to scrape in parallel
REQUESTS_PER_DOMAIN = 8 # max simultaneous requests per domain
CACHE2_EXPIRATION_SECS = 48 * 60 * 60 # seconds while cached response is still valid (a negative value means "never expires")
LOG_ENABLED = True #
LOGLEVEL = 'DEBUG' # default loglevel
LOGFILE = None # None means sys.stderr by default
LOG_STDOUT = False #
DEFAULT_ITEM_CLASS = 'scrapy.item.ScrapedItem'
SCHEDULER = 'scrapy.core.scheduler.Scheduler'
MEMORYSTORE = 'scrapy.core.scheduler.MemoryStore'
PRIORITIZER = 'scrapy.core.prioritizers.RandomPrioritizer'
EXTENSIONS = []
# contrib.middleware.retry.RetryMiddleware default settings
RETRY_TIMES = 3
RETRY_HTTP_CODES = ['500', '503', '504', '400', '408', '200']
# amazon s3 images backend sometimes fails to authenticate or validate http protocol
RETRY_HTTP_CODES = RETRY_HTTP_CODES + ['505', '403']
<commit_msg>Revert "add 505 and 403 to retry status codes due to amazon s3 random fails while uploading images"
This reverts changeset r457
--HG--
extra : convert_revision : svn%3Ab85faa78-f9eb-468e-a121-7cced6da292c%40461<commit_after>import scrapy
# Scrapy core settings
BOT_NAME = 'scrapy'
BOT_VERSION = scrapy.__version__
ENGINE_DEBUG = False
# Download configuration options
USER_AGENT = '%s/%s' % (BOT_NAME, BOT_VERSION)
DOWNLOAD_TIMEOUT = 180 # 3mins
CONCURRENT_DOMAINS = 8 # number of domains to scrape in parallel
REQUESTS_PER_DOMAIN = 8 # max simultaneous requests per domain
CACHE2_EXPIRATION_SECS = 48 * 60 * 60 # seconds while cached response is still valid (a negative value means "never expires")
LOG_ENABLED = True #
LOGLEVEL = 'DEBUG' # default loglevel
LOGFILE = None # None means sys.stderr by default
LOG_STDOUT = False #
DEFAULT_ITEM_CLASS = 'scrapy.item.ScrapedItem'
SCHEDULER = 'scrapy.core.scheduler.Scheduler'
MEMORYSTORE = 'scrapy.core.scheduler.MemoryStore'
PRIORITIZER = 'scrapy.core.prioritizers.RandomPrioritizer'
EXTENSIONS = []
# contrib.middleware.retry.RetryMiddleware default settings
RETRY_TIMES = 3
RETRY_HTTP_CODES = ['500', '503', '504', '400', '408', '200']
|
bad65df528da18293d38b0f50dbbb16390af465e
|
sphinx/source/docs/user_guide/source_examples/plotting_label.py
|
sphinx/source/docs/user_guide/source_examples/plotting_label.py
|
from bokeh.plotting import figure, show, output_file
from bokeh.models import ColumnDataSource, Range1d, Label
output_file("label.html", title="label.py example")
source = ColumnDataSource(data=dict(height=[66, 71, 72, 68, 58, 62],
weight=[165, 189, 220, 141, 260, 174],
names=['Mark', 'Amir', 'Matt', 'Greg',
'Owen', 'Juan']))
p = figure(title='Dist. of 10th Grade Students at Lee High',
x_range=Range1d(140, 275))
p.scatter(x='weight', y='height', size=8, source=source)
p.xaxis[0].axis_label = 'Weight (lbs)'
p.yaxis[0].axis_label = 'Height (in)'
label = Label(x='weight', y='height', text='names', level='glyph',
x_offset=5, y_offset=-5, source=source)
p.add_annotation(label)
show(p)
|
from bokeh.plotting import figure, show, output_file
from bokeh.models import ColumnDataSource, Range1d, Label
output_file("label.html", title="label.py example")
source = ColumnDataSource(data=dict(height=[66, 71, 72, 68, 58, 62],
weight=[165, 189, 220, 141, 260, 174],
names=['Mark', 'Amir', 'Matt', 'Greg',
'Owen', 'Juan']))
p = figure(title='Dist. of 10th Grade Students at Lee High',
x_range=Range1d(140, 275))
p.scatter(x='weight', y='height', size=8, source=source)
p.xaxis[0].axis_label = 'Weight (lbs)'
p.yaxis[0].axis_label = 'Height (in)'
labels = Label(x='weight', y='height', text='names', level='glyph',
x_offset=5, y_offset=-5, source=source, render_mode='canvas')
citation = Label(x=70, y=70, x_units='screen', y_units='screen',
text=['Collected by Luke C. 2016-04-01'], render_mode='css',
border_line_color='black', border_line_alpha=1.0,
background_fill_color='white', background_fill_alpha=1.0)
p.add_annotation(labels)
p.add_annotation(citation)
show(p)
|
Include example of css render_mode
|
Include example of css render_mode
|
Python
|
bsd-3-clause
|
clairetang6/bokeh,Karel-van-de-Plassche/bokeh,mindriot101/bokeh,aiguofer/bokeh,rs2/bokeh,Karel-van-de-Plassche/bokeh,KasperPRasmussen/bokeh,dennisobrien/bokeh,draperjames/bokeh,bokeh/bokeh,quasiben/bokeh,KasperPRasmussen/bokeh,philippjfr/bokeh,stonebig/bokeh,justacec/bokeh,KasperPRasmussen/bokeh,phobson/bokeh,phobson/bokeh,jakirkham/bokeh,rs2/bokeh,ericmjl/bokeh,jakirkham/bokeh,bokeh/bokeh,phobson/bokeh,bokeh/bokeh,aavanian/bokeh,dennisobrien/bokeh,rs2/bokeh,timsnyder/bokeh,ptitjano/bokeh,schoolie/bokeh,DuCorey/bokeh,dennisobrien/bokeh,quasiben/bokeh,mindriot101/bokeh,draperjames/bokeh,stonebig/bokeh,dennisobrien/bokeh,philippjfr/bokeh,rs2/bokeh,Karel-van-de-Plassche/bokeh,rs2/bokeh,azjps/bokeh,ptitjano/bokeh,jakirkham/bokeh,aavanian/bokeh,mindriot101/bokeh,KasperPRasmussen/bokeh,percyfal/bokeh,clairetang6/bokeh,ericmjl/bokeh,philippjfr/bokeh,clairetang6/bokeh,ericmjl/bokeh,mindriot101/bokeh,aiguofer/bokeh,draperjames/bokeh,bokeh/bokeh,bokeh/bokeh,Karel-van-de-Plassche/bokeh,philippjfr/bokeh,ericmjl/bokeh,phobson/bokeh,aavanian/bokeh,jakirkham/bokeh,timsnyder/bokeh,azjps/bokeh,schoolie/bokeh,DuCorey/bokeh,azjps/bokeh,quasiben/bokeh,DuCorey/bokeh,justacec/bokeh,draperjames/bokeh,justacec/bokeh,jakirkham/bokeh,schoolie/bokeh,justacec/bokeh,stonebig/bokeh,philippjfr/bokeh,Karel-van-de-Plassche/bokeh,aiguofer/bokeh,aavanian/bokeh,ptitjano/bokeh,ptitjano/bokeh,DuCorey/bokeh,azjps/bokeh,percyfal/bokeh,timsnyder/bokeh,aiguofer/bokeh,KasperPRasmussen/bokeh,timsnyder/bokeh,schoolie/bokeh,ptitjano/bokeh,stonebig/bokeh,phobson/bokeh,timsnyder/bokeh,dennisobrien/bokeh,draperjames/bokeh,schoolie/bokeh,DuCorey/bokeh,azjps/bokeh,percyfal/bokeh,percyfal/bokeh,clairetang6/bokeh,percyfal/bokeh,ericmjl/bokeh,aiguofer/bokeh,aavanian/bokeh
|
from bokeh.plotting import figure, show, output_file
from bokeh.models import ColumnDataSource, Range1d, Label
output_file("label.html", title="label.py example")
source = ColumnDataSource(data=dict(height=[66, 71, 72, 68, 58, 62],
weight=[165, 189, 220, 141, 260, 174],
names=['Mark', 'Amir', 'Matt', 'Greg',
'Owen', 'Juan']))
p = figure(title='Dist. of 10th Grade Students at Lee High',
x_range=Range1d(140, 275))
p.scatter(x='weight', y='height', size=8, source=source)
p.xaxis[0].axis_label = 'Weight (lbs)'
p.yaxis[0].axis_label = 'Height (in)'
label = Label(x='weight', y='height', text='names', level='glyph',
x_offset=5, y_offset=-5, source=source)
p.add_annotation(label)
show(p)
Include example of css render_mode
|
from bokeh.plotting import figure, show, output_file
from bokeh.models import ColumnDataSource, Range1d, Label
output_file("label.html", title="label.py example")
source = ColumnDataSource(data=dict(height=[66, 71, 72, 68, 58, 62],
weight=[165, 189, 220, 141, 260, 174],
names=['Mark', 'Amir', 'Matt', 'Greg',
'Owen', 'Juan']))
p = figure(title='Dist. of 10th Grade Students at Lee High',
x_range=Range1d(140, 275))
p.scatter(x='weight', y='height', size=8, source=source)
p.xaxis[0].axis_label = 'Weight (lbs)'
p.yaxis[0].axis_label = 'Height (in)'
labels = Label(x='weight', y='height', text='names', level='glyph',
x_offset=5, y_offset=-5, source=source, render_mode='canvas')
citation = Label(x=70, y=70, x_units='screen', y_units='screen',
text=['Collected by Luke C. 2016-04-01'], render_mode='css',
border_line_color='black', border_line_alpha=1.0,
background_fill_color='white', background_fill_alpha=1.0)
p.add_annotation(labels)
p.add_annotation(citation)
show(p)
|
<commit_before>from bokeh.plotting import figure, show, output_file
from bokeh.models import ColumnDataSource, Range1d, Label
output_file("label.html", title="label.py example")
source = ColumnDataSource(data=dict(height=[66, 71, 72, 68, 58, 62],
weight=[165, 189, 220, 141, 260, 174],
names=['Mark', 'Amir', 'Matt', 'Greg',
'Owen', 'Juan']))
p = figure(title='Dist. of 10th Grade Students at Lee High',
x_range=Range1d(140, 275))
p.scatter(x='weight', y='height', size=8, source=source)
p.xaxis[0].axis_label = 'Weight (lbs)'
p.yaxis[0].axis_label = 'Height (in)'
label = Label(x='weight', y='height', text='names', level='glyph',
x_offset=5, y_offset=-5, source=source)
p.add_annotation(label)
show(p)
<commit_msg>Include example of css render_mode<commit_after>
|
from bokeh.plotting import figure, show, output_file
from bokeh.models import ColumnDataSource, Range1d, Label
output_file("label.html", title="label.py example")
source = ColumnDataSource(data=dict(height=[66, 71, 72, 68, 58, 62],
weight=[165, 189, 220, 141, 260, 174],
names=['Mark', 'Amir', 'Matt', 'Greg',
'Owen', 'Juan']))
p = figure(title='Dist. of 10th Grade Students at Lee High',
x_range=Range1d(140, 275))
p.scatter(x='weight', y='height', size=8, source=source)
p.xaxis[0].axis_label = 'Weight (lbs)'
p.yaxis[0].axis_label = 'Height (in)'
labels = Label(x='weight', y='height', text='names', level='glyph',
x_offset=5, y_offset=-5, source=source, render_mode='canvas')
citation = Label(x=70, y=70, x_units='screen', y_units='screen',
text=['Collected by Luke C. 2016-04-01'], render_mode='css',
border_line_color='black', border_line_alpha=1.0,
background_fill_color='white', background_fill_alpha=1.0)
p.add_annotation(labels)
p.add_annotation(citation)
show(p)
|
from bokeh.plotting import figure, show, output_file
from bokeh.models import ColumnDataSource, Range1d, Label
output_file("label.html", title="label.py example")
source = ColumnDataSource(data=dict(height=[66, 71, 72, 68, 58, 62],
weight=[165, 189, 220, 141, 260, 174],
names=['Mark', 'Amir', 'Matt', 'Greg',
'Owen', 'Juan']))
p = figure(title='Dist. of 10th Grade Students at Lee High',
x_range=Range1d(140, 275))
p.scatter(x='weight', y='height', size=8, source=source)
p.xaxis[0].axis_label = 'Weight (lbs)'
p.yaxis[0].axis_label = 'Height (in)'
label = Label(x='weight', y='height', text='names', level='glyph',
x_offset=5, y_offset=-5, source=source)
p.add_annotation(label)
show(p)
Include example of css render_modefrom bokeh.plotting import figure, show, output_file
from bokeh.models import ColumnDataSource, Range1d, Label
output_file("label.html", title="label.py example")
source = ColumnDataSource(data=dict(height=[66, 71, 72, 68, 58, 62],
weight=[165, 189, 220, 141, 260, 174],
names=['Mark', 'Amir', 'Matt', 'Greg',
'Owen', 'Juan']))
p = figure(title='Dist. of 10th Grade Students at Lee High',
x_range=Range1d(140, 275))
p.scatter(x='weight', y='height', size=8, source=source)
p.xaxis[0].axis_label = 'Weight (lbs)'
p.yaxis[0].axis_label = 'Height (in)'
labels = Label(x='weight', y='height', text='names', level='glyph',
x_offset=5, y_offset=-5, source=source, render_mode='canvas')
citation = Label(x=70, y=70, x_units='screen', y_units='screen',
text=['Collected by Luke C. 2016-04-01'], render_mode='css',
border_line_color='black', border_line_alpha=1.0,
background_fill_color='white', background_fill_alpha=1.0)
p.add_annotation(labels)
p.add_annotation(citation)
show(p)
|
<commit_before>from bokeh.plotting import figure, show, output_file
from bokeh.models import ColumnDataSource, Range1d, Label
output_file("label.html", title="label.py example")
source = ColumnDataSource(data=dict(height=[66, 71, 72, 68, 58, 62],
weight=[165, 189, 220, 141, 260, 174],
names=['Mark', 'Amir', 'Matt', 'Greg',
'Owen', 'Juan']))
p = figure(title='Dist. of 10th Grade Students at Lee High',
x_range=Range1d(140, 275))
p.scatter(x='weight', y='height', size=8, source=source)
p.xaxis[0].axis_label = 'Weight (lbs)'
p.yaxis[0].axis_label = 'Height (in)'
label = Label(x='weight', y='height', text='names', level='glyph',
x_offset=5, y_offset=-5, source=source)
p.add_annotation(label)
show(p)
<commit_msg>Include example of css render_mode<commit_after>from bokeh.plotting import figure, show, output_file
from bokeh.models import ColumnDataSource, Range1d, Label
output_file("label.html", title="label.py example")
source = ColumnDataSource(data=dict(height=[66, 71, 72, 68, 58, 62],
weight=[165, 189, 220, 141, 260, 174],
names=['Mark', 'Amir', 'Matt', 'Greg',
'Owen', 'Juan']))
p = figure(title='Dist. of 10th Grade Students at Lee High',
x_range=Range1d(140, 275))
p.scatter(x='weight', y='height', size=8, source=source)
p.xaxis[0].axis_label = 'Weight (lbs)'
p.yaxis[0].axis_label = 'Height (in)'
labels = Label(x='weight', y='height', text='names', level='glyph',
x_offset=5, y_offset=-5, source=source, render_mode='canvas')
citation = Label(x=70, y=70, x_units='screen', y_units='screen',
text=['Collected by Luke C. 2016-04-01'], render_mode='css',
border_line_color='black', border_line_alpha=1.0,
background_fill_color='white', background_fill_alpha=1.0)
p.add_annotation(labels)
p.add_annotation(citation)
show(p)
|
c6cdf543f6bfd0049594eeb530551371bf21bae4
|
test/test_scraping.py
|
test/test_scraping.py
|
from datetime import datetime
import sys
import unittest
import btceapi
class TestScraping(unittest.TestCase):
def test_scrape_main_page(self):
mainPage = btceapi.scrapeMainPage()
for message in mainPage.messages:
msgId, user, time, text = message
self.assertIs(type(time), datetime)
if sys.version_info[0] == 2:
# python2.x
assert type(msgId) in (str, unicode)
assert type(user) in (str, unicode)
assert type(text) in (str, unicode)
else:
# python3.x
self.assertIs(type(msgId), str)
self.assertIs(type(user), str)
self.assertIs(type(text), str)
if __name__ == '__main__':
unittest.main()
|
from datetime import datetime
import sys
import unittest
import btceapi
class TestScraping(unittest.TestCase):
def test_scrape_main_page(self):
mainPage = btceapi.scrapeMainPage()
for message in mainPage.messages:
msgId, user, time, text = message
assert type(time) is datetime
if sys.version_info[0] == 2:
# python2.x
assert type(msgId) in (str, unicode)
assert type(user) in (str, unicode)
assert type(text) in (str, unicode)
else:
# python3.x
self.assertIs(type(msgId), str)
self.assertIs(type(user), str)
self.assertIs(type(text), str)
if __name__ == '__main__':
unittest.main()
|
Fix for assertIs method not being present in Python 2.6.
|
Fix for assertIs method not being present in Python 2.6.
|
Python
|
mit
|
lromanov/tidex-api,CodeReclaimers/btce-api,alanmcintyre/btce-api
|
from datetime import datetime
import sys
import unittest
import btceapi
class TestScraping(unittest.TestCase):
def test_scrape_main_page(self):
mainPage = btceapi.scrapeMainPage()
for message in mainPage.messages:
msgId, user, time, text = message
self.assertIs(type(time), datetime)
if sys.version_info[0] == 2:
# python2.x
assert type(msgId) in (str, unicode)
assert type(user) in (str, unicode)
assert type(text) in (str, unicode)
else:
# python3.x
self.assertIs(type(msgId), str)
self.assertIs(type(user), str)
self.assertIs(type(text), str)
if __name__ == '__main__':
unittest.main()
Fix for assertIs method not being present in Python 2.6.
|
from datetime import datetime
import sys
import unittest
import btceapi
class TestScraping(unittest.TestCase):
def test_scrape_main_page(self):
mainPage = btceapi.scrapeMainPage()
for message in mainPage.messages:
msgId, user, time, text = message
assert type(time) is datetime
if sys.version_info[0] == 2:
# python2.x
assert type(msgId) in (str, unicode)
assert type(user) in (str, unicode)
assert type(text) in (str, unicode)
else:
# python3.x
self.assertIs(type(msgId), str)
self.assertIs(type(user), str)
self.assertIs(type(text), str)
if __name__ == '__main__':
unittest.main()
|
<commit_before>from datetime import datetime
import sys
import unittest
import btceapi
class TestScraping(unittest.TestCase):
def test_scrape_main_page(self):
mainPage = btceapi.scrapeMainPage()
for message in mainPage.messages:
msgId, user, time, text = message
self.assertIs(type(time), datetime)
if sys.version_info[0] == 2:
# python2.x
assert type(msgId) in (str, unicode)
assert type(user) in (str, unicode)
assert type(text) in (str, unicode)
else:
# python3.x
self.assertIs(type(msgId), str)
self.assertIs(type(user), str)
self.assertIs(type(text), str)
if __name__ == '__main__':
unittest.main()
<commit_msg>Fix for assertIs method not being present in Python 2.6.<commit_after>
|
from datetime import datetime
import sys
import unittest
import btceapi
class TestScraping(unittest.TestCase):
def test_scrape_main_page(self):
mainPage = btceapi.scrapeMainPage()
for message in mainPage.messages:
msgId, user, time, text = message
assert type(time) is datetime
if sys.version_info[0] == 2:
# python2.x
assert type(msgId) in (str, unicode)
assert type(user) in (str, unicode)
assert type(text) in (str, unicode)
else:
# python3.x
self.assertIs(type(msgId), str)
self.assertIs(type(user), str)
self.assertIs(type(text), str)
if __name__ == '__main__':
unittest.main()
|
from datetime import datetime
import sys
import unittest
import btceapi
class TestScraping(unittest.TestCase):
def test_scrape_main_page(self):
mainPage = btceapi.scrapeMainPage()
for message in mainPage.messages:
msgId, user, time, text = message
self.assertIs(type(time), datetime)
if sys.version_info[0] == 2:
# python2.x
assert type(msgId) in (str, unicode)
assert type(user) in (str, unicode)
assert type(text) in (str, unicode)
else:
# python3.x
self.assertIs(type(msgId), str)
self.assertIs(type(user), str)
self.assertIs(type(text), str)
if __name__ == '__main__':
unittest.main()
Fix for assertIs method not being present in Python 2.6.from datetime import datetime
import sys
import unittest
import btceapi
class TestScraping(unittest.TestCase):
def test_scrape_main_page(self):
mainPage = btceapi.scrapeMainPage()
for message in mainPage.messages:
msgId, user, time, text = message
assert type(time) is datetime
if sys.version_info[0] == 2:
# python2.x
assert type(msgId) in (str, unicode)
assert type(user) in (str, unicode)
assert type(text) in (str, unicode)
else:
# python3.x
self.assertIs(type(msgId), str)
self.assertIs(type(user), str)
self.assertIs(type(text), str)
if __name__ == '__main__':
unittest.main()
|
<commit_before>from datetime import datetime
import sys
import unittest
import btceapi
class TestScraping(unittest.TestCase):
def test_scrape_main_page(self):
mainPage = btceapi.scrapeMainPage()
for message in mainPage.messages:
msgId, user, time, text = message
self.assertIs(type(time), datetime)
if sys.version_info[0] == 2:
# python2.x
assert type(msgId) in (str, unicode)
assert type(user) in (str, unicode)
assert type(text) in (str, unicode)
else:
# python3.x
self.assertIs(type(msgId), str)
self.assertIs(type(user), str)
self.assertIs(type(text), str)
if __name__ == '__main__':
unittest.main()
<commit_msg>Fix for assertIs method not being present in Python 2.6.<commit_after>from datetime import datetime
import sys
import unittest
import btceapi
class TestScraping(unittest.TestCase):
def test_scrape_main_page(self):
mainPage = btceapi.scrapeMainPage()
for message in mainPage.messages:
msgId, user, time, text = message
assert type(time) is datetime
if sys.version_info[0] == 2:
# python2.x
assert type(msgId) in (str, unicode)
assert type(user) in (str, unicode)
assert type(text) in (str, unicode)
else:
# python3.x
self.assertIs(type(msgId), str)
self.assertIs(type(user), str)
self.assertIs(type(text), str)
if __name__ == '__main__':
unittest.main()
|
96076567bac3329cba55b61c59781c7670c7a02b
|
anybox/recipe/odoo/runtime/patch_odoo.py
|
anybox/recipe/odoo/runtime/patch_odoo.py
|
"""Necessary monkey patches to make Odoo work in the buildout context.
"""
import subprocess
def do_patch(gevent_script_path):
"""
Patch odoo prefork so that --workers execute the correct gevent script.
This monkey patch could be safer, if the script path determination could be
isolated from the actual process management logic in the original.
"""
from odoo.service.server import PreforkServer, stripped_sys_argv
def long_polling_spawn(server):
nargs = stripped_sys_argv()
nargs[0] = gevent_script_path
popen = subprocess.Popen(nargs)
server.long_polling_pid = popen.pid
PreforkServer.long_polling_spawn = long_polling_spawn
|
"""Necessary monkey patches to make Odoo work in the buildout context.
"""
import subprocess
def do_patch(gevent_script_path):
"""
Patch odoo prefork so that --workers execute the correct gevent script.
This monkey patch could be safer, if the script path determination could be
isolated from the actual process management logic in the original.
"""
try:
from odoo.service.server import PreforkServer, stripped_sys_argv
except ImportError:
from openerp.service.server import PreforkServer, stripped_sys_argv
def long_polling_spawn(server):
nargs = stripped_sys_argv()
nargs[0] = gevent_script_path
popen = subprocess.Popen(nargs)
server.long_polling_pid = popen.pid
PreforkServer.long_polling_spawn = long_polling_spawn
|
Maintain compatilbility with <10 version
|
Maintain compatilbility with <10 version
|
Python
|
agpl-3.0
|
anybox/anybox.recipe.odoo
|
"""Necessary monkey patches to make Odoo work in the buildout context.
"""
import subprocess
def do_patch(gevent_script_path):
"""
Patch odoo prefork so that --workers execute the correct gevent script.
This monkey patch could be safer, if the script path determination could be
isolated from the actual process management logic in the original.
"""
from odoo.service.server import PreforkServer, stripped_sys_argv
def long_polling_spawn(server):
nargs = stripped_sys_argv()
nargs[0] = gevent_script_path
popen = subprocess.Popen(nargs)
server.long_polling_pid = popen.pid
PreforkServer.long_polling_spawn = long_polling_spawn
Maintain compatilbility with <10 version
|
"""Necessary monkey patches to make Odoo work in the buildout context.
"""
import subprocess
def do_patch(gevent_script_path):
"""
Patch odoo prefork so that --workers execute the correct gevent script.
This monkey patch could be safer, if the script path determination could be
isolated from the actual process management logic in the original.
"""
try:
from odoo.service.server import PreforkServer, stripped_sys_argv
except ImportError:
from openerp.service.server import PreforkServer, stripped_sys_argv
def long_polling_spawn(server):
nargs = stripped_sys_argv()
nargs[0] = gevent_script_path
popen = subprocess.Popen(nargs)
server.long_polling_pid = popen.pid
PreforkServer.long_polling_spawn = long_polling_spawn
|
<commit_before>"""Necessary monkey patches to make Odoo work in the buildout context.
"""
import subprocess
def do_patch(gevent_script_path):
"""
Patch odoo prefork so that --workers execute the correct gevent script.
This monkey patch could be safer, if the script path determination could be
isolated from the actual process management logic in the original.
"""
from odoo.service.server import PreforkServer, stripped_sys_argv
def long_polling_spawn(server):
nargs = stripped_sys_argv()
nargs[0] = gevent_script_path
popen = subprocess.Popen(nargs)
server.long_polling_pid = popen.pid
PreforkServer.long_polling_spawn = long_polling_spawn
<commit_msg>Maintain compatilbility with <10 version<commit_after>
|
"""Necessary monkey patches to make Odoo work in the buildout context.
"""
import subprocess
def do_patch(gevent_script_path):
"""
Patch odoo prefork so that --workers execute the correct gevent script.
This monkey patch could be safer, if the script path determination could be
isolated from the actual process management logic in the original.
"""
try:
from odoo.service.server import PreforkServer, stripped_sys_argv
except ImportError:
from openerp.service.server import PreforkServer, stripped_sys_argv
def long_polling_spawn(server):
nargs = stripped_sys_argv()
nargs[0] = gevent_script_path
popen = subprocess.Popen(nargs)
server.long_polling_pid = popen.pid
PreforkServer.long_polling_spawn = long_polling_spawn
|
"""Necessary monkey patches to make Odoo work in the buildout context.
"""
import subprocess
def do_patch(gevent_script_path):
"""
Patch odoo prefork so that --workers execute the correct gevent script.
This monkey patch could be safer, if the script path determination could be
isolated from the actual process management logic in the original.
"""
from odoo.service.server import PreforkServer, stripped_sys_argv
def long_polling_spawn(server):
nargs = stripped_sys_argv()
nargs[0] = gevent_script_path
popen = subprocess.Popen(nargs)
server.long_polling_pid = popen.pid
PreforkServer.long_polling_spawn = long_polling_spawn
Maintain compatilbility with <10 version"""Necessary monkey patches to make Odoo work in the buildout context.
"""
import subprocess
def do_patch(gevent_script_path):
"""
Patch odoo prefork so that --workers execute the correct gevent script.
This monkey patch could be safer, if the script path determination could be
isolated from the actual process management logic in the original.
"""
try:
from odoo.service.server import PreforkServer, stripped_sys_argv
except ImportError:
from openerp.service.server import PreforkServer, stripped_sys_argv
def long_polling_spawn(server):
nargs = stripped_sys_argv()
nargs[0] = gevent_script_path
popen = subprocess.Popen(nargs)
server.long_polling_pid = popen.pid
PreforkServer.long_polling_spawn = long_polling_spawn
|
<commit_before>"""Necessary monkey patches to make Odoo work in the buildout context.
"""
import subprocess
def do_patch(gevent_script_path):
"""
Patch odoo prefork so that --workers execute the correct gevent script.
This monkey patch could be safer, if the script path determination could be
isolated from the actual process management logic in the original.
"""
from odoo.service.server import PreforkServer, stripped_sys_argv
def long_polling_spawn(server):
nargs = stripped_sys_argv()
nargs[0] = gevent_script_path
popen = subprocess.Popen(nargs)
server.long_polling_pid = popen.pid
PreforkServer.long_polling_spawn = long_polling_spawn
<commit_msg>Maintain compatilbility with <10 version<commit_after>"""Necessary monkey patches to make Odoo work in the buildout context.
"""
import subprocess
def do_patch(gevent_script_path):
"""
Patch odoo prefork so that --workers execute the correct gevent script.
This monkey patch could be safer, if the script path determination could be
isolated from the actual process management logic in the original.
"""
try:
from odoo.service.server import PreforkServer, stripped_sys_argv
except ImportError:
from openerp.service.server import PreforkServer, stripped_sys_argv
def long_polling_spawn(server):
nargs = stripped_sys_argv()
nargs[0] = gevent_script_path
popen = subprocess.Popen(nargs)
server.long_polling_pid = popen.pid
PreforkServer.long_polling_spawn = long_polling_spawn
|
3cc3c0b90714bbf7a2638b16faec69aba82a4050
|
op_robot_tests/tests_files/brokers/openprocurement_client_helper.py
|
op_robot_tests/tests_files/brokers/openprocurement_client_helper.py
|
from openprocurement_client.client import Client
import sys
def prepare_api_wrapper(key='', host_url="https://api-sandbox.openprocurement.org", api_version='0.8' ):
return Client(key, host_url, api_version )
def get_internal_id(get_tenders_function, date):
result = get_tenders_function({"offset": date, "opt_fields": 'tenderID', })
#import pdb; pdb.Pdb(stdout=sys.__stdout__).set_trace()
return result
|
from openprocurement_client.client import Client
import sys
def prepare_api_wrapper(key='', host_url="https://api-sandbox.openprocurement.org", api_version='0.8'):
return Client(key, host_url, api_version)
def get_internal_id(get_tenders_function, date):
result = get_tenders_function({"offset": date, "opt_fields": 'tenderID'})
return result
|
Improve PEP8 compliance in op_client_helper.py
|
Improve PEP8 compliance in op_client_helper.py
|
Python
|
apache-2.0
|
SlaOne/robot_tests,kosaniak/robot_tests,selurvedu/robot_tests,Leits/robot_tests,cleardevice/robot_tests,VadimShurhal/robot_tests.broker.aps,mykhaly/robot_tests,Rzaporozhets/robot_tests,bubanoid/robot_tests,openprocurement/robot_tests
|
from openprocurement_client.client import Client
import sys
def prepare_api_wrapper(key='', host_url="https://api-sandbox.openprocurement.org", api_version='0.8' ):
return Client(key, host_url, api_version )
def get_internal_id(get_tenders_function, date):
result = get_tenders_function({"offset": date, "opt_fields": 'tenderID', })
#import pdb; pdb.Pdb(stdout=sys.__stdout__).set_trace()
return result
Improve PEP8 compliance in op_client_helper.py
|
from openprocurement_client.client import Client
import sys
def prepare_api_wrapper(key='', host_url="https://api-sandbox.openprocurement.org", api_version='0.8'):
return Client(key, host_url, api_version)
def get_internal_id(get_tenders_function, date):
result = get_tenders_function({"offset": date, "opt_fields": 'tenderID'})
return result
|
<commit_before>from openprocurement_client.client import Client
import sys
def prepare_api_wrapper(key='', host_url="https://api-sandbox.openprocurement.org", api_version='0.8' ):
return Client(key, host_url, api_version )
def get_internal_id(get_tenders_function, date):
result = get_tenders_function({"offset": date, "opt_fields": 'tenderID', })
#import pdb; pdb.Pdb(stdout=sys.__stdout__).set_trace()
return result
<commit_msg>Improve PEP8 compliance in op_client_helper.py<commit_after>
|
from openprocurement_client.client import Client
import sys
def prepare_api_wrapper(key='', host_url="https://api-sandbox.openprocurement.org", api_version='0.8'):
return Client(key, host_url, api_version)
def get_internal_id(get_tenders_function, date):
result = get_tenders_function({"offset": date, "opt_fields": 'tenderID'})
return result
|
from openprocurement_client.client import Client
import sys
def prepare_api_wrapper(key='', host_url="https://api-sandbox.openprocurement.org", api_version='0.8' ):
return Client(key, host_url, api_version )
def get_internal_id(get_tenders_function, date):
result = get_tenders_function({"offset": date, "opt_fields": 'tenderID', })
#import pdb; pdb.Pdb(stdout=sys.__stdout__).set_trace()
return result
Improve PEP8 compliance in op_client_helper.pyfrom openprocurement_client.client import Client
import sys
def prepare_api_wrapper(key='', host_url="https://api-sandbox.openprocurement.org", api_version='0.8'):
return Client(key, host_url, api_version)
def get_internal_id(get_tenders_function, date):
result = get_tenders_function({"offset": date, "opt_fields": 'tenderID'})
return result
|
<commit_before>from openprocurement_client.client import Client
import sys
def prepare_api_wrapper(key='', host_url="https://api-sandbox.openprocurement.org", api_version='0.8' ):
return Client(key, host_url, api_version )
def get_internal_id(get_tenders_function, date):
result = get_tenders_function({"offset": date, "opt_fields": 'tenderID', })
#import pdb; pdb.Pdb(stdout=sys.__stdout__).set_trace()
return result
<commit_msg>Improve PEP8 compliance in op_client_helper.py<commit_after>from openprocurement_client.client import Client
import sys
def prepare_api_wrapper(key='', host_url="https://api-sandbox.openprocurement.org", api_version='0.8'):
return Client(key, host_url, api_version)
def get_internal_id(get_tenders_function, date):
result = get_tenders_function({"offset": date, "opt_fields": 'tenderID'})
return result
|
27c54cfd5eaf180595e671c80bd7c39406c8a24c
|
databroker/__init__.py
|
databroker/__init__.py
|
# Import intake to run driver discovery first and avoid circular import issues.
import intake
del intake
import warnings
import logging
logger = logging.getLogger(__name__)
from ._core import (Broker, BrokerES, Header, ALL,
lookup_config, list_configs, describe_configs, temp_config,
wrap_in_doct,
DeprecatedDoct, wrap_in_deprecated_doct)
from .discovery import MergedCatalog, EntrypointsCatalog, V0Catalog
# A catalog created from discovered entrypoints and v0 catalogs.
catalog = MergedCatalog([EntrypointsCatalog(), V0Catalog()])
# set version string using versioneer
from ._version import get_versions
__version__ = get_versions()['version']
del get_versions
### Legacy imports ###
try:
from .databroker import DataBroker
except ImportError:
pass
else:
from .databroker import (DataBroker, DataBroker as db,
get_events, get_table, stream, get_fields,
restream, process)
from .pims_readers import get_images
|
# Import intake to run driver discovery first and avoid circular import issues.
import intake
del intake
import warnings
import logging
logger = logging.getLogger(__name__)
from .v1 import Broker, Header, ALL, temp, temp_config
from .utils import (lookup_config, list_configs, describe_configs,
wrap_in_doct, DeprecatedDoct, wrap_in_deprecated_doct)
from .discovery import MergedCatalog, EntrypointsCatalog, V0Catalog
# A catalog created from discovered entrypoints and v0 catalogs.
catalog = MergedCatalog([EntrypointsCatalog(), V0Catalog()])
# set version string using versioneer
from ._version import get_versions
__version__ = get_versions()['version']
del get_versions
### Legacy imports ###
try:
from .databroker import DataBroker
except ImportError:
pass
else:
from .databroker import (DataBroker, DataBroker as db,
get_events, get_table, stream, get_fields,
restream, process)
from .pims_readers import get_images
|
Move top-level imports from v0 to v1.
|
Move top-level imports from v0 to v1.
|
Python
|
bsd-3-clause
|
ericdill/databroker,ericdill/databroker
|
# Import intake to run driver discovery first and avoid circular import issues.
import intake
del intake
import warnings
import logging
logger = logging.getLogger(__name__)
from ._core import (Broker, BrokerES, Header, ALL,
lookup_config, list_configs, describe_configs, temp_config,
wrap_in_doct,
DeprecatedDoct, wrap_in_deprecated_doct)
from .discovery import MergedCatalog, EntrypointsCatalog, V0Catalog
# A catalog created from discovered entrypoints and v0 catalogs.
catalog = MergedCatalog([EntrypointsCatalog(), V0Catalog()])
# set version string using versioneer
from ._version import get_versions
__version__ = get_versions()['version']
del get_versions
### Legacy imports ###
try:
from .databroker import DataBroker
except ImportError:
pass
else:
from .databroker import (DataBroker, DataBroker as db,
get_events, get_table, stream, get_fields,
restream, process)
from .pims_readers import get_images
Move top-level imports from v0 to v1.
|
# Import intake to run driver discovery first and avoid circular import issues.
import intake
del intake
import warnings
import logging
logger = logging.getLogger(__name__)
from .v1 import Broker, Header, ALL, temp, temp_config
from .utils import (lookup_config, list_configs, describe_configs,
wrap_in_doct, DeprecatedDoct, wrap_in_deprecated_doct)
from .discovery import MergedCatalog, EntrypointsCatalog, V0Catalog
# A catalog created from discovered entrypoints and v0 catalogs.
catalog = MergedCatalog([EntrypointsCatalog(), V0Catalog()])
# set version string using versioneer
from ._version import get_versions
__version__ = get_versions()['version']
del get_versions
### Legacy imports ###
try:
from .databroker import DataBroker
except ImportError:
pass
else:
from .databroker import (DataBroker, DataBroker as db,
get_events, get_table, stream, get_fields,
restream, process)
from .pims_readers import get_images
|
<commit_before># Import intake to run driver discovery first and avoid circular import issues.
import intake
del intake
import warnings
import logging
logger = logging.getLogger(__name__)
from ._core import (Broker, BrokerES, Header, ALL,
lookup_config, list_configs, describe_configs, temp_config,
wrap_in_doct,
DeprecatedDoct, wrap_in_deprecated_doct)
from .discovery import MergedCatalog, EntrypointsCatalog, V0Catalog
# A catalog created from discovered entrypoints and v0 catalogs.
catalog = MergedCatalog([EntrypointsCatalog(), V0Catalog()])
# set version string using versioneer
from ._version import get_versions
__version__ = get_versions()['version']
del get_versions
### Legacy imports ###
try:
from .databroker import DataBroker
except ImportError:
pass
else:
from .databroker import (DataBroker, DataBroker as db,
get_events, get_table, stream, get_fields,
restream, process)
from .pims_readers import get_images
<commit_msg>Move top-level imports from v0 to v1.<commit_after>
|
# Import intake to run driver discovery first and avoid circular import issues.
import intake
del intake
import warnings
import logging
logger = logging.getLogger(__name__)
from .v1 import Broker, Header, ALL, temp, temp_config
from .utils import (lookup_config, list_configs, describe_configs,
wrap_in_doct, DeprecatedDoct, wrap_in_deprecated_doct)
from .discovery import MergedCatalog, EntrypointsCatalog, V0Catalog
# A catalog created from discovered entrypoints and v0 catalogs.
catalog = MergedCatalog([EntrypointsCatalog(), V0Catalog()])
# set version string using versioneer
from ._version import get_versions
__version__ = get_versions()['version']
del get_versions
### Legacy imports ###
try:
from .databroker import DataBroker
except ImportError:
pass
else:
from .databroker import (DataBroker, DataBroker as db,
get_events, get_table, stream, get_fields,
restream, process)
from .pims_readers import get_images
|
# Import intake to run driver discovery first and avoid circular import issues.
import intake
del intake
import warnings
import logging
logger = logging.getLogger(__name__)
from ._core import (Broker, BrokerES, Header, ALL,
lookup_config, list_configs, describe_configs, temp_config,
wrap_in_doct,
DeprecatedDoct, wrap_in_deprecated_doct)
from .discovery import MergedCatalog, EntrypointsCatalog, V0Catalog
# A catalog created from discovered entrypoints and v0 catalogs.
catalog = MergedCatalog([EntrypointsCatalog(), V0Catalog()])
# set version string using versioneer
from ._version import get_versions
__version__ = get_versions()['version']
del get_versions
### Legacy imports ###
try:
from .databroker import DataBroker
except ImportError:
pass
else:
from .databroker import (DataBroker, DataBroker as db,
get_events, get_table, stream, get_fields,
restream, process)
from .pims_readers import get_images
Move top-level imports from v0 to v1.# Import intake to run driver discovery first and avoid circular import issues.
import intake
del intake
import warnings
import logging
logger = logging.getLogger(__name__)
from .v1 import Broker, Header, ALL, temp, temp_config
from .utils import (lookup_config, list_configs, describe_configs,
wrap_in_doct, DeprecatedDoct, wrap_in_deprecated_doct)
from .discovery import MergedCatalog, EntrypointsCatalog, V0Catalog
# A catalog created from discovered entrypoints and v0 catalogs.
catalog = MergedCatalog([EntrypointsCatalog(), V0Catalog()])
# set version string using versioneer
from ._version import get_versions
__version__ = get_versions()['version']
del get_versions
### Legacy imports ###
try:
from .databroker import DataBroker
except ImportError:
pass
else:
from .databroker import (DataBroker, DataBroker as db,
get_events, get_table, stream, get_fields,
restream, process)
from .pims_readers import get_images
|
<commit_before># Import intake to run driver discovery first and avoid circular import issues.
import intake
del intake
import warnings
import logging
logger = logging.getLogger(__name__)
from ._core import (Broker, BrokerES, Header, ALL,
lookup_config, list_configs, describe_configs, temp_config,
wrap_in_doct,
DeprecatedDoct, wrap_in_deprecated_doct)
from .discovery import MergedCatalog, EntrypointsCatalog, V0Catalog
# A catalog created from discovered entrypoints and v0 catalogs.
catalog = MergedCatalog([EntrypointsCatalog(), V0Catalog()])
# set version string using versioneer
from ._version import get_versions
__version__ = get_versions()['version']
del get_versions
### Legacy imports ###
try:
from .databroker import DataBroker
except ImportError:
pass
else:
from .databroker import (DataBroker, DataBroker as db,
get_events, get_table, stream, get_fields,
restream, process)
from .pims_readers import get_images
<commit_msg>Move top-level imports from v0 to v1.<commit_after># Import intake to run driver discovery first and avoid circular import issues.
import intake
del intake
import warnings
import logging
logger = logging.getLogger(__name__)
from .v1 import Broker, Header, ALL, temp, temp_config
from .utils import (lookup_config, list_configs, describe_configs,
wrap_in_doct, DeprecatedDoct, wrap_in_deprecated_doct)
from .discovery import MergedCatalog, EntrypointsCatalog, V0Catalog
# A catalog created from discovered entrypoints and v0 catalogs.
catalog = MergedCatalog([EntrypointsCatalog(), V0Catalog()])
# set version string using versioneer
from ._version import get_versions
__version__ = get_versions()['version']
del get_versions
### Legacy imports ###
try:
from .databroker import DataBroker
except ImportError:
pass
else:
from .databroker import (DataBroker, DataBroker as db,
get_events, get_table, stream, get_fields,
restream, process)
from .pims_readers import get_images
|
09f86488096880870bbd3363e0a4c018f11e935d
|
lingcod/layers/urls.py
|
lingcod/layers/urls.py
|
from django.conf.urls.defaults import *
urlpatterns = patterns('lingcod.layers.views',
url(r'^public/',
'get_public_layers',
name='public-data-layers'),
url(r'^kml_file/(?P<session_key>\w+)/(?P<uid>[\w_]+).kml',
'get_kml_file',
name='kml-file'),
url(r'^privatekml/(?P<session_key>\w+)/$',
'get_privatekml_list',
name='layers-privatekml-list'),
url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/$',
'get_privatekml',
name='layers-privatekml'),
url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/(?P<path>[^\z]+)$',
'get_relative_to_privatekml',
name='layers-privatekml-relative'),
)
|
from django.conf.urls.defaults import *
import time
urlpatterns = patterns('lingcod.layers.views',
url(r'^public/$',
'get_public_layers',
name='public-data-layers'),
# Useful for debugging, avoids GE caching interference
url(r'^public/cachebuster/%s' % str(time.time()),
'get_public_layers',
name='public-data-layers-cachebuster'),
url(r'^kml_file/(?P<session_key>\w+)/(?P<uid>[\w_]+).kml',
'get_kml_file',
name='kml-file'),
url(r'^privatekml/(?P<session_key>\w+)/$',
'get_privatekml_list',
name='layers-privatekml-list'),
url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/$',
'get_privatekml',
name='layers-privatekml'),
url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/(?P<path>[^\z]+)$',
'get_relative_to_privatekml',
name='layers-privatekml-relative'),
)
|
Add another url pattern for debugging public layers
|
Add another url pattern for debugging public layers
|
Python
|
bsd-3-clause
|
Alwnikrotikz/marinemap,google-code-export/marinemap,Alwnikrotikz/marinemap,google-code-export/marinemap,Alwnikrotikz/marinemap,google-code-export/marinemap,google-code-export/marinemap,Alwnikrotikz/marinemap
|
from django.conf.urls.defaults import *
urlpatterns = patterns('lingcod.layers.views',
url(r'^public/',
'get_public_layers',
name='public-data-layers'),
url(r'^kml_file/(?P<session_key>\w+)/(?P<uid>[\w_]+).kml',
'get_kml_file',
name='kml-file'),
url(r'^privatekml/(?P<session_key>\w+)/$',
'get_privatekml_list',
name='layers-privatekml-list'),
url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/$',
'get_privatekml',
name='layers-privatekml'),
url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/(?P<path>[^\z]+)$',
'get_relative_to_privatekml',
name='layers-privatekml-relative'),
)
Add another url pattern for debugging public layers
|
from django.conf.urls.defaults import *
import time
urlpatterns = patterns('lingcod.layers.views',
url(r'^public/$',
'get_public_layers',
name='public-data-layers'),
# Useful for debugging, avoids GE caching interference
url(r'^public/cachebuster/%s' % str(time.time()),
'get_public_layers',
name='public-data-layers-cachebuster'),
url(r'^kml_file/(?P<session_key>\w+)/(?P<uid>[\w_]+).kml',
'get_kml_file',
name='kml-file'),
url(r'^privatekml/(?P<session_key>\w+)/$',
'get_privatekml_list',
name='layers-privatekml-list'),
url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/$',
'get_privatekml',
name='layers-privatekml'),
url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/(?P<path>[^\z]+)$',
'get_relative_to_privatekml',
name='layers-privatekml-relative'),
)
|
<commit_before>from django.conf.urls.defaults import *
urlpatterns = patterns('lingcod.layers.views',
url(r'^public/',
'get_public_layers',
name='public-data-layers'),
url(r'^kml_file/(?P<session_key>\w+)/(?P<uid>[\w_]+).kml',
'get_kml_file',
name='kml-file'),
url(r'^privatekml/(?P<session_key>\w+)/$',
'get_privatekml_list',
name='layers-privatekml-list'),
url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/$',
'get_privatekml',
name='layers-privatekml'),
url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/(?P<path>[^\z]+)$',
'get_relative_to_privatekml',
name='layers-privatekml-relative'),
)
<commit_msg>Add another url pattern for debugging public layers<commit_after>
|
from django.conf.urls.defaults import *
import time
urlpatterns = patterns('lingcod.layers.views',
url(r'^public/$',
'get_public_layers',
name='public-data-layers'),
# Useful for debugging, avoids GE caching interference
url(r'^public/cachebuster/%s' % str(time.time()),
'get_public_layers',
name='public-data-layers-cachebuster'),
url(r'^kml_file/(?P<session_key>\w+)/(?P<uid>[\w_]+).kml',
'get_kml_file',
name='kml-file'),
url(r'^privatekml/(?P<session_key>\w+)/$',
'get_privatekml_list',
name='layers-privatekml-list'),
url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/$',
'get_privatekml',
name='layers-privatekml'),
url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/(?P<path>[^\z]+)$',
'get_relative_to_privatekml',
name='layers-privatekml-relative'),
)
|
from django.conf.urls.defaults import *
urlpatterns = patterns('lingcod.layers.views',
url(r'^public/',
'get_public_layers',
name='public-data-layers'),
url(r'^kml_file/(?P<session_key>\w+)/(?P<uid>[\w_]+).kml',
'get_kml_file',
name='kml-file'),
url(r'^privatekml/(?P<session_key>\w+)/$',
'get_privatekml_list',
name='layers-privatekml-list'),
url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/$',
'get_privatekml',
name='layers-privatekml'),
url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/(?P<path>[^\z]+)$',
'get_relative_to_privatekml',
name='layers-privatekml-relative'),
)
Add another url pattern for debugging public layersfrom django.conf.urls.defaults import *
import time
urlpatterns = patterns('lingcod.layers.views',
url(r'^public/$',
'get_public_layers',
name='public-data-layers'),
# Useful for debugging, avoids GE caching interference
url(r'^public/cachebuster/%s' % str(time.time()),
'get_public_layers',
name='public-data-layers-cachebuster'),
url(r'^kml_file/(?P<session_key>\w+)/(?P<uid>[\w_]+).kml',
'get_kml_file',
name='kml-file'),
url(r'^privatekml/(?P<session_key>\w+)/$',
'get_privatekml_list',
name='layers-privatekml-list'),
url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/$',
'get_privatekml',
name='layers-privatekml'),
url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/(?P<path>[^\z]+)$',
'get_relative_to_privatekml',
name='layers-privatekml-relative'),
)
|
<commit_before>from django.conf.urls.defaults import *
urlpatterns = patterns('lingcod.layers.views',
url(r'^public/',
'get_public_layers',
name='public-data-layers'),
url(r'^kml_file/(?P<session_key>\w+)/(?P<uid>[\w_]+).kml',
'get_kml_file',
name='kml-file'),
url(r'^privatekml/(?P<session_key>\w+)/$',
'get_privatekml_list',
name='layers-privatekml-list'),
url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/$',
'get_privatekml',
name='layers-privatekml'),
url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/(?P<path>[^\z]+)$',
'get_relative_to_privatekml',
name='layers-privatekml-relative'),
)
<commit_msg>Add another url pattern for debugging public layers<commit_after>from django.conf.urls.defaults import *
import time
urlpatterns = patterns('lingcod.layers.views',
url(r'^public/$',
'get_public_layers',
name='public-data-layers'),
# Useful for debugging, avoids GE caching interference
url(r'^public/cachebuster/%s' % str(time.time()),
'get_public_layers',
name='public-data-layers-cachebuster'),
url(r'^kml_file/(?P<session_key>\w+)/(?P<uid>[\w_]+).kml',
'get_kml_file',
name='kml-file'),
url(r'^privatekml/(?P<session_key>\w+)/$',
'get_privatekml_list',
name='layers-privatekml-list'),
url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/$',
'get_privatekml',
name='layers-privatekml'),
url(r'^privatekml/(?P<session_key>\w+)/(?P<pk>\d+)/(?P<path>[^\z]+)$',
'get_relative_to_privatekml',
name='layers-privatekml-relative'),
)
|
a2e7642034bf89bf1d7d513ef155da3375482373
|
virtool/user_permissions.py
|
virtool/user_permissions.py
|
#: A list of the permission strings used by Virtool.
PERMISSIONS = [
"add_virus",
"modify_virus",
"remove_virus",
"modify_hmm",
"add_sample",
"add_host",
"remove_host",
"cancel_job",
"remove_job",
"archive_job",
"rebuild_index",
"modify_options",
"manage_users"
]
|
#: A list of the permission strings used by Virtool.
PERMISSIONS = [
"add_virus",
"modify_virus",
"remove_virus",
"modify_hmm",
"create_sample",
"add_host",
"remove_host",
"cancel_job",
"remove_job",
"archive_job",
"rebuild_index",
"modify_options",
"manage_users"
]
|
Change 'add_sample' permission to 'create_sample'
|
Change 'add_sample' permission to 'create_sample'
|
Python
|
mit
|
igboyes/virtool,igboyes/virtool,virtool/virtool,virtool/virtool
|
#: A list of the permission strings used by Virtool.
PERMISSIONS = [
"add_virus",
"modify_virus",
"remove_virus",
"modify_hmm",
"add_sample",
"add_host",
"remove_host",
"cancel_job",
"remove_job",
"archive_job",
"rebuild_index",
"modify_options",
"manage_users"
]
Change 'add_sample' permission to 'create_sample'
|
#: A list of the permission strings used by Virtool.
PERMISSIONS = [
"add_virus",
"modify_virus",
"remove_virus",
"modify_hmm",
"create_sample",
"add_host",
"remove_host",
"cancel_job",
"remove_job",
"archive_job",
"rebuild_index",
"modify_options",
"manage_users"
]
|
<commit_before>#: A list of the permission strings used by Virtool.
PERMISSIONS = [
"add_virus",
"modify_virus",
"remove_virus",
"modify_hmm",
"add_sample",
"add_host",
"remove_host",
"cancel_job",
"remove_job",
"archive_job",
"rebuild_index",
"modify_options",
"manage_users"
]
<commit_msg>Change 'add_sample' permission to 'create_sample'<commit_after>
|
#: A list of the permission strings used by Virtool.
PERMISSIONS = [
"add_virus",
"modify_virus",
"remove_virus",
"modify_hmm",
"create_sample",
"add_host",
"remove_host",
"cancel_job",
"remove_job",
"archive_job",
"rebuild_index",
"modify_options",
"manage_users"
]
|
#: A list of the permission strings used by Virtool.
PERMISSIONS = [
"add_virus",
"modify_virus",
"remove_virus",
"modify_hmm",
"add_sample",
"add_host",
"remove_host",
"cancel_job",
"remove_job",
"archive_job",
"rebuild_index",
"modify_options",
"manage_users"
]
Change 'add_sample' permission to 'create_sample'#: A list of the permission strings used by Virtool.
PERMISSIONS = [
"add_virus",
"modify_virus",
"remove_virus",
"modify_hmm",
"create_sample",
"add_host",
"remove_host",
"cancel_job",
"remove_job",
"archive_job",
"rebuild_index",
"modify_options",
"manage_users"
]
|
<commit_before>#: A list of the permission strings used by Virtool.
PERMISSIONS = [
"add_virus",
"modify_virus",
"remove_virus",
"modify_hmm",
"add_sample",
"add_host",
"remove_host",
"cancel_job",
"remove_job",
"archive_job",
"rebuild_index",
"modify_options",
"manage_users"
]
<commit_msg>Change 'add_sample' permission to 'create_sample'<commit_after>#: A list of the permission strings used by Virtool.
PERMISSIONS = [
"add_virus",
"modify_virus",
"remove_virus",
"modify_hmm",
"create_sample",
"add_host",
"remove_host",
"cancel_job",
"remove_job",
"archive_job",
"rebuild_index",
"modify_options",
"manage_users"
]
|
685116d1a2799399819ed780679403e7576e67b5
|
keystone/tests/unit/common/test_manager.py
|
keystone/tests/unit/common/test_manager.py
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
import mock
from keystone import catalog
from keystone.common import manager
from keystone.tests import unit
class TestCreateLegacyDriver(unit.BaseTestCase):
@mock.patch('oslo_log.versionutils.report_deprecated_feature')
def test_class_is_properly_deprecated(self, mock_reporter):
Driver = manager.create_legacy_driver(catalog.CatalogDriverV8)
# NOTE(dstanek): I want to subvert the requirement for this
# class to implement all of the abstract methods.
Driver.__abstractmethods__ = set()
impl = Driver()
details = {
'as_of': 'Liberty',
'what': 'keystone.catalog.core.Driver',
'in_favor_of': 'keystone.catalog.core.CatalogDriverV8',
'remove_in': 'N',
}
mock_reporter.assert_called_with(mock.ANY, mock.ANY, details)
self.assertIsInstance(impl, catalog.CatalogDriverV8)
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
import mock
from keystone import catalog
from keystone.common import manager
from keystone.tests import unit
class TestCreateLegacyDriver(unit.BaseTestCase):
@mock.patch('oslo_log.versionutils.report_deprecated_feature')
def test_class_is_properly_deprecated(self, mock_reporter):
Driver = manager.create_legacy_driver(catalog.CatalogDriverV8)
# NOTE(dstanek): I want to subvert the requirement for this
# class to implement all of the abstract methods.
Driver.__abstractmethods__ = set()
impl = Driver()
details = {
'as_of': 'Liberty',
'what': 'keystone.catalog.core.Driver',
'in_favor_of': 'keystone.catalog.core.CatalogDriverV8',
'remove_in': mock.ANY,
}
mock_reporter.assert_called_with(mock.ANY, mock.ANY, details)
self.assertEqual('N', mock_reporter.call_args[0][2]['remove_in'][0])
self.assertIsInstance(impl, catalog.CatalogDriverV8)
|
Correct test to support changing N release name
|
Correct test to support changing N release name
oslo.log is going to change to use Newton rather than N so this test
should not make an assumption about the way that
versionutils.deprecated is calling report_deprecated_feature.
Change-Id: I06aa6d085232376811f73597b2d84b5174bc7a8d
Closes-Bug: 1561121
(cherry picked from commit 8556437ee02de028ec5de3b867abaab82533cb91)
|
Python
|
apache-2.0
|
openstack/keystone,openstack/keystone,cernops/keystone,klmitch/keystone,mahak/keystone,mahak/keystone,rajalokan/keystone,ilay09/keystone,openstack/keystone,ilay09/keystone,ilay09/keystone,cernops/keystone,klmitch/keystone,mahak/keystone,rajalokan/keystone,rajalokan/keystone
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
import mock
from keystone import catalog
from keystone.common import manager
from keystone.tests import unit
class TestCreateLegacyDriver(unit.BaseTestCase):
@mock.patch('oslo_log.versionutils.report_deprecated_feature')
def test_class_is_properly_deprecated(self, mock_reporter):
Driver = manager.create_legacy_driver(catalog.CatalogDriverV8)
# NOTE(dstanek): I want to subvert the requirement for this
# class to implement all of the abstract methods.
Driver.__abstractmethods__ = set()
impl = Driver()
details = {
'as_of': 'Liberty',
'what': 'keystone.catalog.core.Driver',
'in_favor_of': 'keystone.catalog.core.CatalogDriverV8',
'remove_in': 'N',
}
mock_reporter.assert_called_with(mock.ANY, mock.ANY, details)
self.assertIsInstance(impl, catalog.CatalogDriverV8)
Correct test to support changing N release name
oslo.log is going to change to use Newton rather than N so this test
should not make an assumption about the way that
versionutils.deprecated is calling report_deprecated_feature.
Change-Id: I06aa6d085232376811f73597b2d84b5174bc7a8d
Closes-Bug: 1561121
(cherry picked from commit 8556437ee02de028ec5de3b867abaab82533cb91)
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
import mock
from keystone import catalog
from keystone.common import manager
from keystone.tests import unit
class TestCreateLegacyDriver(unit.BaseTestCase):
@mock.patch('oslo_log.versionutils.report_deprecated_feature')
def test_class_is_properly_deprecated(self, mock_reporter):
Driver = manager.create_legacy_driver(catalog.CatalogDriverV8)
# NOTE(dstanek): I want to subvert the requirement for this
# class to implement all of the abstract methods.
Driver.__abstractmethods__ = set()
impl = Driver()
details = {
'as_of': 'Liberty',
'what': 'keystone.catalog.core.Driver',
'in_favor_of': 'keystone.catalog.core.CatalogDriverV8',
'remove_in': mock.ANY,
}
mock_reporter.assert_called_with(mock.ANY, mock.ANY, details)
self.assertEqual('N', mock_reporter.call_args[0][2]['remove_in'][0])
self.assertIsInstance(impl, catalog.CatalogDriverV8)
|
<commit_before># Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
import mock
from keystone import catalog
from keystone.common import manager
from keystone.tests import unit
class TestCreateLegacyDriver(unit.BaseTestCase):
@mock.patch('oslo_log.versionutils.report_deprecated_feature')
def test_class_is_properly_deprecated(self, mock_reporter):
Driver = manager.create_legacy_driver(catalog.CatalogDriverV8)
# NOTE(dstanek): I want to subvert the requirement for this
# class to implement all of the abstract methods.
Driver.__abstractmethods__ = set()
impl = Driver()
details = {
'as_of': 'Liberty',
'what': 'keystone.catalog.core.Driver',
'in_favor_of': 'keystone.catalog.core.CatalogDriverV8',
'remove_in': 'N',
}
mock_reporter.assert_called_with(mock.ANY, mock.ANY, details)
self.assertIsInstance(impl, catalog.CatalogDriverV8)
<commit_msg>Correct test to support changing N release name
oslo.log is going to change to use Newton rather than N so this test
should not make an assumption about the way that
versionutils.deprecated is calling report_deprecated_feature.
Change-Id: I06aa6d085232376811f73597b2d84b5174bc7a8d
Closes-Bug: 1561121
(cherry picked from commit 8556437ee02de028ec5de3b867abaab82533cb91)<commit_after>
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
import mock
from keystone import catalog
from keystone.common import manager
from keystone.tests import unit
class TestCreateLegacyDriver(unit.BaseTestCase):
@mock.patch('oslo_log.versionutils.report_deprecated_feature')
def test_class_is_properly_deprecated(self, mock_reporter):
Driver = manager.create_legacy_driver(catalog.CatalogDriverV8)
# NOTE(dstanek): I want to subvert the requirement for this
# class to implement all of the abstract methods.
Driver.__abstractmethods__ = set()
impl = Driver()
details = {
'as_of': 'Liberty',
'what': 'keystone.catalog.core.Driver',
'in_favor_of': 'keystone.catalog.core.CatalogDriverV8',
'remove_in': mock.ANY,
}
mock_reporter.assert_called_with(mock.ANY, mock.ANY, details)
self.assertEqual('N', mock_reporter.call_args[0][2]['remove_in'][0])
self.assertIsInstance(impl, catalog.CatalogDriverV8)
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
import mock
from keystone import catalog
from keystone.common import manager
from keystone.tests import unit
class TestCreateLegacyDriver(unit.BaseTestCase):
@mock.patch('oslo_log.versionutils.report_deprecated_feature')
def test_class_is_properly_deprecated(self, mock_reporter):
Driver = manager.create_legacy_driver(catalog.CatalogDriverV8)
# NOTE(dstanek): I want to subvert the requirement for this
# class to implement all of the abstract methods.
Driver.__abstractmethods__ = set()
impl = Driver()
details = {
'as_of': 'Liberty',
'what': 'keystone.catalog.core.Driver',
'in_favor_of': 'keystone.catalog.core.CatalogDriverV8',
'remove_in': 'N',
}
mock_reporter.assert_called_with(mock.ANY, mock.ANY, details)
self.assertIsInstance(impl, catalog.CatalogDriverV8)
Correct test to support changing N release name
oslo.log is going to change to use Newton rather than N so this test
should not make an assumption about the way that
versionutils.deprecated is calling report_deprecated_feature.
Change-Id: I06aa6d085232376811f73597b2d84b5174bc7a8d
Closes-Bug: 1561121
(cherry picked from commit 8556437ee02de028ec5de3b867abaab82533cb91)# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
import mock
from keystone import catalog
from keystone.common import manager
from keystone.tests import unit
class TestCreateLegacyDriver(unit.BaseTestCase):
@mock.patch('oslo_log.versionutils.report_deprecated_feature')
def test_class_is_properly_deprecated(self, mock_reporter):
Driver = manager.create_legacy_driver(catalog.CatalogDriverV8)
# NOTE(dstanek): I want to subvert the requirement for this
# class to implement all of the abstract methods.
Driver.__abstractmethods__ = set()
impl = Driver()
details = {
'as_of': 'Liberty',
'what': 'keystone.catalog.core.Driver',
'in_favor_of': 'keystone.catalog.core.CatalogDriverV8',
'remove_in': mock.ANY,
}
mock_reporter.assert_called_with(mock.ANY, mock.ANY, details)
self.assertEqual('N', mock_reporter.call_args[0][2]['remove_in'][0])
self.assertIsInstance(impl, catalog.CatalogDriverV8)
|
<commit_before># Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
import mock
from keystone import catalog
from keystone.common import manager
from keystone.tests import unit
class TestCreateLegacyDriver(unit.BaseTestCase):
@mock.patch('oslo_log.versionutils.report_deprecated_feature')
def test_class_is_properly_deprecated(self, mock_reporter):
Driver = manager.create_legacy_driver(catalog.CatalogDriverV8)
# NOTE(dstanek): I want to subvert the requirement for this
# class to implement all of the abstract methods.
Driver.__abstractmethods__ = set()
impl = Driver()
details = {
'as_of': 'Liberty',
'what': 'keystone.catalog.core.Driver',
'in_favor_of': 'keystone.catalog.core.CatalogDriverV8',
'remove_in': 'N',
}
mock_reporter.assert_called_with(mock.ANY, mock.ANY, details)
self.assertIsInstance(impl, catalog.CatalogDriverV8)
<commit_msg>Correct test to support changing N release name
oslo.log is going to change to use Newton rather than N so this test
should not make an assumption about the way that
versionutils.deprecated is calling report_deprecated_feature.
Change-Id: I06aa6d085232376811f73597b2d84b5174bc7a8d
Closes-Bug: 1561121
(cherry picked from commit 8556437ee02de028ec5de3b867abaab82533cb91)<commit_after># Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
import mock
from keystone import catalog
from keystone.common import manager
from keystone.tests import unit
class TestCreateLegacyDriver(unit.BaseTestCase):
@mock.patch('oslo_log.versionutils.report_deprecated_feature')
def test_class_is_properly_deprecated(self, mock_reporter):
Driver = manager.create_legacy_driver(catalog.CatalogDriverV8)
# NOTE(dstanek): I want to subvert the requirement for this
# class to implement all of the abstract methods.
Driver.__abstractmethods__ = set()
impl = Driver()
details = {
'as_of': 'Liberty',
'what': 'keystone.catalog.core.Driver',
'in_favor_of': 'keystone.catalog.core.CatalogDriverV8',
'remove_in': mock.ANY,
}
mock_reporter.assert_called_with(mock.ANY, mock.ANY, details)
self.assertEqual('N', mock_reporter.call_args[0][2]['remove_in'][0])
self.assertIsInstance(impl, catalog.CatalogDriverV8)
|
74bd9ffd412f22671232cb301b3762660a73d912
|
lot/landmapper/urls.py
|
lot/landmapper/urls.py
|
from django.urls import include, re_path, path
from landmapper.views import *
urlpatterns = [
# What is difference between re_path and path?
# re_path(r'',
# home, name='landmapper-home'),
path('', home, name="home"),
path('/identify/', identify, name="identify"),
path('/report/', report, name="report"),
path('/pages/', include('django.contrib.flatpages.urls')),
url(r'^get_taxlot_json', views.get_taxlot_json, name='get taxlot json'),
]
|
from django.urls import include, re_path, path
from landmapper.views import *
urlpatterns = [
# What is difference between re_path and path?
# re_path(r'',
# home, name='landmapper-home'),
path('', home, name="home"),
path('/identify', identify, name="identify"),
path('/report', report, name="report"),
path('/get_taxlot_json', get_taxlot_json, name='get taxlot json'),
]
|
Fix get taxlot url and remove trailing slashes
|
Fix get taxlot url and remove trailing slashes
|
Python
|
bsd-3-clause
|
Ecotrust/forestplanner,Ecotrust/forestplanner,Ecotrust/forestplanner,Ecotrust/forestplanner,Ecotrust/forestplanner,Ecotrust/forestplanner,Ecotrust/forestplanner,Ecotrust/forestplanner
|
from django.urls import include, re_path, path
from landmapper.views import *
urlpatterns = [
# What is difference between re_path and path?
# re_path(r'',
# home, name='landmapper-home'),
path('', home, name="home"),
path('/identify/', identify, name="identify"),
path('/report/', report, name="report"),
path('/pages/', include('django.contrib.flatpages.urls')),
url(r'^get_taxlot_json', views.get_taxlot_json, name='get taxlot json'),
]
Fix get taxlot url and remove trailing slashes
|
from django.urls import include, re_path, path
from landmapper.views import *
urlpatterns = [
# What is difference between re_path and path?
# re_path(r'',
# home, name='landmapper-home'),
path('', home, name="home"),
path('/identify', identify, name="identify"),
path('/report', report, name="report"),
path('/get_taxlot_json', get_taxlot_json, name='get taxlot json'),
]
|
<commit_before>from django.urls import include, re_path, path
from landmapper.views import *
urlpatterns = [
# What is difference between re_path and path?
# re_path(r'',
# home, name='landmapper-home'),
path('', home, name="home"),
path('/identify/', identify, name="identify"),
path('/report/', report, name="report"),
path('/pages/', include('django.contrib.flatpages.urls')),
url(r'^get_taxlot_json', views.get_taxlot_json, name='get taxlot json'),
]
<commit_msg>Fix get taxlot url and remove trailing slashes<commit_after>
|
from django.urls import include, re_path, path
from landmapper.views import *
urlpatterns = [
# What is difference between re_path and path?
# re_path(r'',
# home, name='landmapper-home'),
path('', home, name="home"),
path('/identify', identify, name="identify"),
path('/report', report, name="report"),
path('/get_taxlot_json', get_taxlot_json, name='get taxlot json'),
]
|
from django.urls import include, re_path, path
from landmapper.views import *
urlpatterns = [
# What is difference between re_path and path?
# re_path(r'',
# home, name='landmapper-home'),
path('', home, name="home"),
path('/identify/', identify, name="identify"),
path('/report/', report, name="report"),
path('/pages/', include('django.contrib.flatpages.urls')),
url(r'^get_taxlot_json', views.get_taxlot_json, name='get taxlot json'),
]
Fix get taxlot url and remove trailing slashesfrom django.urls import include, re_path, path
from landmapper.views import *
urlpatterns = [
# What is difference between re_path and path?
# re_path(r'',
# home, name='landmapper-home'),
path('', home, name="home"),
path('/identify', identify, name="identify"),
path('/report', report, name="report"),
path('/get_taxlot_json', get_taxlot_json, name='get taxlot json'),
]
|
<commit_before>from django.urls import include, re_path, path
from landmapper.views import *
urlpatterns = [
# What is difference between re_path and path?
# re_path(r'',
# home, name='landmapper-home'),
path('', home, name="home"),
path('/identify/', identify, name="identify"),
path('/report/', report, name="report"),
path('/pages/', include('django.contrib.flatpages.urls')),
url(r'^get_taxlot_json', views.get_taxlot_json, name='get taxlot json'),
]
<commit_msg>Fix get taxlot url and remove trailing slashes<commit_after>from django.urls import include, re_path, path
from landmapper.views import *
urlpatterns = [
# What is difference between re_path and path?
# re_path(r'',
# home, name='landmapper-home'),
path('', home, name="home"),
path('/identify', identify, name="identify"),
path('/report', report, name="report"),
path('/get_taxlot_json', get_taxlot_json, name='get taxlot json'),
]
|
789b33f8c6d4ddad4c46e7a3815d9f9543485caa
|
usb/blueprints/api.py
|
usb/blueprints/api.py
|
from flask import Blueprint, jsonify, request
from usb.models import db, Redirect, DeviceType
from usb.shortener import get_short_id, get_short_url
api = Blueprint('api', __name__)
@api.route('/links')
def get_links():
return jsonify({}), 200
@api.route('/links', methods=['POST'])
def shorten_url():
short_id = get_short_id()
long_url = request.json['url']
for device_type in DeviceType:
db.session.add(Redirect(short_id, device_type, long_url))
db.session.commit()
short_url = get_short_url(short_id)
return jsonify(url=short_url), 200
|
from flask import Blueprint, jsonify, request
from usb.models import db, Redirect, DeviceType
from usb.shortener import get_short_id, get_short_url
api = Blueprint('api', __name__)
@api.route('/links')
def get_links():
return jsonify({}), 200
@api.route('/links', methods=['POST'])
def shorten_url():
short_id = get_short_id()
long_url = request.json['url']
redirect = Redirect.query.filter_by(url=long_url).first()
if redirect:
short_url = get_short_url(redirect.short)
return jsonify(url=short_url), 409
for device_type in DeviceType:
db.session.add(Redirect(short_id, device_type, long_url))
db.session.commit()
short_url = get_short_url(short_id)
return jsonify(url=short_url), 200
|
Return short URL if it's already exists
|
Return short URL if it's already exists
|
Python
|
mit
|
dizpers/usb
|
from flask import Blueprint, jsonify, request
from usb.models import db, Redirect, DeviceType
from usb.shortener import get_short_id, get_short_url
api = Blueprint('api', __name__)
@api.route('/links')
def get_links():
return jsonify({}), 200
@api.route('/links', methods=['POST'])
def shorten_url():
short_id = get_short_id()
long_url = request.json['url']
for device_type in DeviceType:
db.session.add(Redirect(short_id, device_type, long_url))
db.session.commit()
short_url = get_short_url(short_id)
return jsonify(url=short_url), 200
Return short URL if it's already exists
|
from flask import Blueprint, jsonify, request
from usb.models import db, Redirect, DeviceType
from usb.shortener import get_short_id, get_short_url
api = Blueprint('api', __name__)
@api.route('/links')
def get_links():
return jsonify({}), 200
@api.route('/links', methods=['POST'])
def shorten_url():
short_id = get_short_id()
long_url = request.json['url']
redirect = Redirect.query.filter_by(url=long_url).first()
if redirect:
short_url = get_short_url(redirect.short)
return jsonify(url=short_url), 409
for device_type in DeviceType:
db.session.add(Redirect(short_id, device_type, long_url))
db.session.commit()
short_url = get_short_url(short_id)
return jsonify(url=short_url), 200
|
<commit_before>from flask import Blueprint, jsonify, request
from usb.models import db, Redirect, DeviceType
from usb.shortener import get_short_id, get_short_url
api = Blueprint('api', __name__)
@api.route('/links')
def get_links():
return jsonify({}), 200
@api.route('/links', methods=['POST'])
def shorten_url():
short_id = get_short_id()
long_url = request.json['url']
for device_type in DeviceType:
db.session.add(Redirect(short_id, device_type, long_url))
db.session.commit()
short_url = get_short_url(short_id)
return jsonify(url=short_url), 200
<commit_msg>Return short URL if it's already exists<commit_after>
|
from flask import Blueprint, jsonify, request
from usb.models import db, Redirect, DeviceType
from usb.shortener import get_short_id, get_short_url
api = Blueprint('api', __name__)
@api.route('/links')
def get_links():
return jsonify({}), 200
@api.route('/links', methods=['POST'])
def shorten_url():
short_id = get_short_id()
long_url = request.json['url']
redirect = Redirect.query.filter_by(url=long_url).first()
if redirect:
short_url = get_short_url(redirect.short)
return jsonify(url=short_url), 409
for device_type in DeviceType:
db.session.add(Redirect(short_id, device_type, long_url))
db.session.commit()
short_url = get_short_url(short_id)
return jsonify(url=short_url), 200
|
from flask import Blueprint, jsonify, request
from usb.models import db, Redirect, DeviceType
from usb.shortener import get_short_id, get_short_url
api = Blueprint('api', __name__)
@api.route('/links')
def get_links():
return jsonify({}), 200
@api.route('/links', methods=['POST'])
def shorten_url():
short_id = get_short_id()
long_url = request.json['url']
for device_type in DeviceType:
db.session.add(Redirect(short_id, device_type, long_url))
db.session.commit()
short_url = get_short_url(short_id)
return jsonify(url=short_url), 200
Return short URL if it's already existsfrom flask import Blueprint, jsonify, request
from usb.models import db, Redirect, DeviceType
from usb.shortener import get_short_id, get_short_url
api = Blueprint('api', __name__)
@api.route('/links')
def get_links():
return jsonify({}), 200
@api.route('/links', methods=['POST'])
def shorten_url():
short_id = get_short_id()
long_url = request.json['url']
redirect = Redirect.query.filter_by(url=long_url).first()
if redirect:
short_url = get_short_url(redirect.short)
return jsonify(url=short_url), 409
for device_type in DeviceType:
db.session.add(Redirect(short_id, device_type, long_url))
db.session.commit()
short_url = get_short_url(short_id)
return jsonify(url=short_url), 200
|
<commit_before>from flask import Blueprint, jsonify, request
from usb.models import db, Redirect, DeviceType
from usb.shortener import get_short_id, get_short_url
api = Blueprint('api', __name__)
@api.route('/links')
def get_links():
return jsonify({}), 200
@api.route('/links', methods=['POST'])
def shorten_url():
short_id = get_short_id()
long_url = request.json['url']
for device_type in DeviceType:
db.session.add(Redirect(short_id, device_type, long_url))
db.session.commit()
short_url = get_short_url(short_id)
return jsonify(url=short_url), 200
<commit_msg>Return short URL if it's already exists<commit_after>from flask import Blueprint, jsonify, request
from usb.models import db, Redirect, DeviceType
from usb.shortener import get_short_id, get_short_url
api = Blueprint('api', __name__)
@api.route('/links')
def get_links():
return jsonify({}), 200
@api.route('/links', methods=['POST'])
def shorten_url():
short_id = get_short_id()
long_url = request.json['url']
redirect = Redirect.query.filter_by(url=long_url).first()
if redirect:
short_url = get_short_url(redirect.short)
return jsonify(url=short_url), 409
for device_type in DeviceType:
db.session.add(Redirect(short_id, device_type, long_url))
db.session.commit()
short_url = get_short_url(short_id)
return jsonify(url=short_url), 200
|
9b4e7a06932d6ed6a5a9032619fa433629187d69
|
utilkit/stringutil.py
|
utilkit/stringutil.py
|
"""
String/unicode helper functions
"""
def safe_unicode(obj, *args):
""" return the unicode representation of obj """
try:
return unicode(obj, *args) # pylint:disable=undefined-variable
except UnicodeDecodeError:
# obj is byte string
ascii_text = str(obj).encode('string_escape')
return unicode(ascii_text) # pylint:disable=undefined-variable
def safe_str(obj):
""" return the byte string representation of obj """
try:
return str(obj)
except UnicodeEncodeError:
# obj is unicode
return unicode(obj).encode('unicode_escape') # pylint:disable=undefined-variable
|
"""
String/unicode helper functions
"""
def safe_unicode(obj, *args):
""" return the unicode representation of obj """
try:
return unicode(obj, *args) # noqa for undefined-variable
except UnicodeDecodeError:
# obj is byte string
ascii_text = str(obj).encode('string_escape')
return unicode(ascii_text) # noqa for undefined-variable
def safe_str(obj):
""" return the byte string representation of obj """
try:
return str(obj)
except UnicodeEncodeError:
# obj is unicode
return unicode(obj).encode('unicode_escape') # noqa for undefined-variable
|
Disable error-checking that assumes Python 3 for these Python 2 helpers, landscape.io style
|
Disable error-checking that assumes Python 3 for these Python 2 helpers,
landscape.io style
|
Python
|
mit
|
aquatix/python-utilkit
|
"""
String/unicode helper functions
"""
def safe_unicode(obj, *args):
""" return the unicode representation of obj """
try:
return unicode(obj, *args) # pylint:disable=undefined-variable
except UnicodeDecodeError:
# obj is byte string
ascii_text = str(obj).encode('string_escape')
return unicode(ascii_text) # pylint:disable=undefined-variable
def safe_str(obj):
""" return the byte string representation of obj """
try:
return str(obj)
except UnicodeEncodeError:
# obj is unicode
return unicode(obj).encode('unicode_escape') # pylint:disable=undefined-variable
Disable error-checking that assumes Python 3 for these Python 2 helpers,
landscape.io style
|
"""
String/unicode helper functions
"""
def safe_unicode(obj, *args):
""" return the unicode representation of obj """
try:
return unicode(obj, *args) # noqa for undefined-variable
except UnicodeDecodeError:
# obj is byte string
ascii_text = str(obj).encode('string_escape')
return unicode(ascii_text) # noqa for undefined-variable
def safe_str(obj):
""" return the byte string representation of obj """
try:
return str(obj)
except UnicodeEncodeError:
# obj is unicode
return unicode(obj).encode('unicode_escape') # noqa for undefined-variable
|
<commit_before>"""
String/unicode helper functions
"""
def safe_unicode(obj, *args):
""" return the unicode representation of obj """
try:
return unicode(obj, *args) # pylint:disable=undefined-variable
except UnicodeDecodeError:
# obj is byte string
ascii_text = str(obj).encode('string_escape')
return unicode(ascii_text) # pylint:disable=undefined-variable
def safe_str(obj):
""" return the byte string representation of obj """
try:
return str(obj)
except UnicodeEncodeError:
# obj is unicode
return unicode(obj).encode('unicode_escape') # pylint:disable=undefined-variable
<commit_msg>Disable error-checking that assumes Python 3 for these Python 2 helpers,
landscape.io style<commit_after>
|
"""
String/unicode helper functions
"""
def safe_unicode(obj, *args):
""" return the unicode representation of obj """
try:
return unicode(obj, *args) # noqa for undefined-variable
except UnicodeDecodeError:
# obj is byte string
ascii_text = str(obj).encode('string_escape')
return unicode(ascii_text) # noqa for undefined-variable
def safe_str(obj):
""" return the byte string representation of obj """
try:
return str(obj)
except UnicodeEncodeError:
# obj is unicode
return unicode(obj).encode('unicode_escape') # noqa for undefined-variable
|
"""
String/unicode helper functions
"""
def safe_unicode(obj, *args):
""" return the unicode representation of obj """
try:
return unicode(obj, *args) # pylint:disable=undefined-variable
except UnicodeDecodeError:
# obj is byte string
ascii_text = str(obj).encode('string_escape')
return unicode(ascii_text) # pylint:disable=undefined-variable
def safe_str(obj):
""" return the byte string representation of obj """
try:
return str(obj)
except UnicodeEncodeError:
# obj is unicode
return unicode(obj).encode('unicode_escape') # pylint:disable=undefined-variable
Disable error-checking that assumes Python 3 for these Python 2 helpers,
landscape.io style"""
String/unicode helper functions
"""
def safe_unicode(obj, *args):
""" return the unicode representation of obj """
try:
return unicode(obj, *args) # noqa for undefined-variable
except UnicodeDecodeError:
# obj is byte string
ascii_text = str(obj).encode('string_escape')
return unicode(ascii_text) # noqa for undefined-variable
def safe_str(obj):
""" return the byte string representation of obj """
try:
return str(obj)
except UnicodeEncodeError:
# obj is unicode
return unicode(obj).encode('unicode_escape') # noqa for undefined-variable
|
<commit_before>"""
String/unicode helper functions
"""
def safe_unicode(obj, *args):
""" return the unicode representation of obj """
try:
return unicode(obj, *args) # pylint:disable=undefined-variable
except UnicodeDecodeError:
# obj is byte string
ascii_text = str(obj).encode('string_escape')
return unicode(ascii_text) # pylint:disable=undefined-variable
def safe_str(obj):
""" return the byte string representation of obj """
try:
return str(obj)
except UnicodeEncodeError:
# obj is unicode
return unicode(obj).encode('unicode_escape') # pylint:disable=undefined-variable
<commit_msg>Disable error-checking that assumes Python 3 for these Python 2 helpers,
landscape.io style<commit_after>"""
String/unicode helper functions
"""
def safe_unicode(obj, *args):
""" return the unicode representation of obj """
try:
return unicode(obj, *args) # noqa for undefined-variable
except UnicodeDecodeError:
# obj is byte string
ascii_text = str(obj).encode('string_escape')
return unicode(ascii_text) # noqa for undefined-variable
def safe_str(obj):
""" return the byte string representation of obj """
try:
return str(obj)
except UnicodeEncodeError:
# obj is unicode
return unicode(obj).encode('unicode_escape') # noqa for undefined-variable
|
8995cbf71454e3424e15913661ee659c48f7b8fa
|
volunteer_planner/settings/local_mysql.py
|
volunteer_planner/settings/local_mysql.py
|
# coding: utf-8
from volunteer_planner.settings.local import *
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.mysql',
'NAME': 'volunteer_planner',
'PASSWORD': os.environ.get('DATABASE_PW', 'volunteer_planner'),
'USER': os.environ.get('DB_USER', 'vp')
}
}
|
# coding: utf-8
from volunteer_planner.settings.local import *
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.mysql',
'NAME': os.environ.get('DATABASE_NAME', 'volunteer_planner'),
'PASSWORD': os.environ.get('DATABASE_PW', 'volunteer_planner'),
'USER': os.environ.get('DB_USER', 'vp')
}
}
|
Make local mysql db name overridable with DATABASE_NAME environment variable
|
Make local mysql db name overridable with DATABASE_NAME environment variable
|
Python
|
agpl-3.0
|
christophmeissner/volunteer_planner,christophmeissner/volunteer_planner,coders4help/volunteer_planner,klinger/volunteer_planner,klinger/volunteer_planner,pitpalme/volunteer_planner,pitpalme/volunteer_planner,pitpalme/volunteer_planner,coders4help/volunteer_planner,pitpalme/volunteer_planner,coders4help/volunteer_planner,christophmeissner/volunteer_planner,alper/volunteer_planner,klinger/volunteer_planner,coders4help/volunteer_planner,alper/volunteer_planner,volunteer-planner/volunteer_planner,christophmeissner/volunteer_planner,volunteer-planner/volunteer_planner,klinger/volunteer_planner,volunteer-planner/volunteer_planner,volunteer-planner/volunteer_planner,alper/volunteer_planner
|
# coding: utf-8
from volunteer_planner.settings.local import *
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.mysql',
'NAME': 'volunteer_planner',
'PASSWORD': os.environ.get('DATABASE_PW', 'volunteer_planner'),
'USER': os.environ.get('DB_USER', 'vp')
}
}
Make local mysql db name overridable with DATABASE_NAME environment variable
|
# coding: utf-8
from volunteer_planner.settings.local import *
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.mysql',
'NAME': os.environ.get('DATABASE_NAME', 'volunteer_planner'),
'PASSWORD': os.environ.get('DATABASE_PW', 'volunteer_planner'),
'USER': os.environ.get('DB_USER', 'vp')
}
}
|
<commit_before># coding: utf-8
from volunteer_planner.settings.local import *
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.mysql',
'NAME': 'volunteer_planner',
'PASSWORD': os.environ.get('DATABASE_PW', 'volunteer_planner'),
'USER': os.environ.get('DB_USER', 'vp')
}
}
<commit_msg>Make local mysql db name overridable with DATABASE_NAME environment variable<commit_after>
|
# coding: utf-8
from volunteer_planner.settings.local import *
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.mysql',
'NAME': os.environ.get('DATABASE_NAME', 'volunteer_planner'),
'PASSWORD': os.environ.get('DATABASE_PW', 'volunteer_planner'),
'USER': os.environ.get('DB_USER', 'vp')
}
}
|
# coding: utf-8
from volunteer_planner.settings.local import *
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.mysql',
'NAME': 'volunteer_planner',
'PASSWORD': os.environ.get('DATABASE_PW', 'volunteer_planner'),
'USER': os.environ.get('DB_USER', 'vp')
}
}
Make local mysql db name overridable with DATABASE_NAME environment variable# coding: utf-8
from volunteer_planner.settings.local import *
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.mysql',
'NAME': os.environ.get('DATABASE_NAME', 'volunteer_planner'),
'PASSWORD': os.environ.get('DATABASE_PW', 'volunteer_planner'),
'USER': os.environ.get('DB_USER', 'vp')
}
}
|
<commit_before># coding: utf-8
from volunteer_planner.settings.local import *
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.mysql',
'NAME': 'volunteer_planner',
'PASSWORD': os.environ.get('DATABASE_PW', 'volunteer_planner'),
'USER': os.environ.get('DB_USER', 'vp')
}
}
<commit_msg>Make local mysql db name overridable with DATABASE_NAME environment variable<commit_after># coding: utf-8
from volunteer_planner.settings.local import *
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.mysql',
'NAME': os.environ.get('DATABASE_NAME', 'volunteer_planner'),
'PASSWORD': os.environ.get('DATABASE_PW', 'volunteer_planner'),
'USER': os.environ.get('DB_USER', 'vp')
}
}
|
7fb1b95205de32ec27b4e5428928b1bba417c9c8
|
build/fbcode_builder/specs/fbthrift.py
|
build/fbcode_builder/specs/fbthrift.py
|
#!/usr/bin/env python
# Copyright (c) Facebook, Inc. and its affiliates.
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import specs.folly as folly
import specs.fizz as fizz
import specs.rsocket as rsocket
import specs.sodium as sodium
import specs.wangle as wangle
import specs.zstd as zstd
from shell_quoting import ShellQuoted
def fbcode_builder_spec(builder):
# This API should change rarely, so build the latest tag instead of master.
builder.add_option(
'no1msd/mstch:git_hash',
ShellQuoted('$(git describe --abbrev=0 --tags)')
)
builder.add_option('krb5/krb5:git_hash', 'krb5-1.16.1-final')
return {
'depends_on': [folly, fizz, sodium, rsocket, wangle, zstd],
'steps': [
# This isn't a separete spec, since only fbthrift uses mstch.
builder.github_project_workdir('no1msd/mstch', 'build'),
builder.cmake_install('no1msd/mstch'),
builder.github_project_workdir('krb5/krb5', 'src'),
builder.autoconf_install('krb5/krb5'),
builder.fb_github_cmake_install('fbthrift/thrift'),
],
}
|
#!/usr/bin/env python
# Copyright (c) Facebook, Inc. and its affiliates.
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import specs.folly as folly
import specs.fizz as fizz
import specs.rsocket as rsocket
import specs.sodium as sodium
import specs.wangle as wangle
import specs.zstd as zstd
from shell_quoting import ShellQuoted
def fbcode_builder_spec(builder):
# This API should change rarely, so build the latest tag instead of master.
builder.add_option(
'no1msd/mstch:git_hash',
ShellQuoted('$(git describe --abbrev=0 --tags)')
)
return {
'depends_on': [folly, fizz, sodium, rsocket, wangle, zstd],
'steps': [
# This isn't a separete spec, since only fbthrift uses mstch.
builder.github_project_workdir('no1msd/mstch', 'build'),
builder.cmake_install('no1msd/mstch'),
builder.fb_github_cmake_install('fbthrift/thrift'),
],
}
|
Cut fbcode_builder dep for thrift on krb5
|
Cut fbcode_builder dep for thrift on krb5
Summary: [Thrift] Cut `fbcode_builder` dep for `thrift` on `krb5`. In the past, Thrift depended on Kerberos and the `krb5` implementation for its transport-layer security. However, Thrift has since migrated fully to Transport Layer Security for its transport-layer security and no longer has any build-time dependency on `krb5`. Clean this up.
Reviewed By: stevegury, vitaut
Differential Revision: D14814205
fbshipit-source-id: dca469d22098e34573674194facaaac6c4c6aa32
|
Python
|
unknown
|
ReactiveSocket/reactivesocket-cpp,ReactiveSocket/reactivesocket-cpp,phoad/rsocket-cpp,phoad/rsocket-cpp,rsocket/rsocket-cpp,phoad/rsocket-cpp,rsocket/rsocket-cpp,rsocket/rsocket-cpp,ReactiveSocket/reactivesocket-cpp,rsocket/rsocket-cpp,phoad/rsocket-cpp,phoad/rsocket-cpp
|
#!/usr/bin/env python
# Copyright (c) Facebook, Inc. and its affiliates.
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import specs.folly as folly
import specs.fizz as fizz
import specs.rsocket as rsocket
import specs.sodium as sodium
import specs.wangle as wangle
import specs.zstd as zstd
from shell_quoting import ShellQuoted
def fbcode_builder_spec(builder):
# This API should change rarely, so build the latest tag instead of master.
builder.add_option(
'no1msd/mstch:git_hash',
ShellQuoted('$(git describe --abbrev=0 --tags)')
)
builder.add_option('krb5/krb5:git_hash', 'krb5-1.16.1-final')
return {
'depends_on': [folly, fizz, sodium, rsocket, wangle, zstd],
'steps': [
# This isn't a separete spec, since only fbthrift uses mstch.
builder.github_project_workdir('no1msd/mstch', 'build'),
builder.cmake_install('no1msd/mstch'),
builder.github_project_workdir('krb5/krb5', 'src'),
builder.autoconf_install('krb5/krb5'),
builder.fb_github_cmake_install('fbthrift/thrift'),
],
}
Cut fbcode_builder dep for thrift on krb5
Summary: [Thrift] Cut `fbcode_builder` dep for `thrift` on `krb5`. In the past, Thrift depended on Kerberos and the `krb5` implementation for its transport-layer security. However, Thrift has since migrated fully to Transport Layer Security for its transport-layer security and no longer has any build-time dependency on `krb5`. Clean this up.
Reviewed By: stevegury, vitaut
Differential Revision: D14814205
fbshipit-source-id: dca469d22098e34573674194facaaac6c4c6aa32
|
#!/usr/bin/env python
# Copyright (c) Facebook, Inc. and its affiliates.
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import specs.folly as folly
import specs.fizz as fizz
import specs.rsocket as rsocket
import specs.sodium as sodium
import specs.wangle as wangle
import specs.zstd as zstd
from shell_quoting import ShellQuoted
def fbcode_builder_spec(builder):
# This API should change rarely, so build the latest tag instead of master.
builder.add_option(
'no1msd/mstch:git_hash',
ShellQuoted('$(git describe --abbrev=0 --tags)')
)
return {
'depends_on': [folly, fizz, sodium, rsocket, wangle, zstd],
'steps': [
# This isn't a separete spec, since only fbthrift uses mstch.
builder.github_project_workdir('no1msd/mstch', 'build'),
builder.cmake_install('no1msd/mstch'),
builder.fb_github_cmake_install('fbthrift/thrift'),
],
}
|
<commit_before>#!/usr/bin/env python
# Copyright (c) Facebook, Inc. and its affiliates.
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import specs.folly as folly
import specs.fizz as fizz
import specs.rsocket as rsocket
import specs.sodium as sodium
import specs.wangle as wangle
import specs.zstd as zstd
from shell_quoting import ShellQuoted
def fbcode_builder_spec(builder):
# This API should change rarely, so build the latest tag instead of master.
builder.add_option(
'no1msd/mstch:git_hash',
ShellQuoted('$(git describe --abbrev=0 --tags)')
)
builder.add_option('krb5/krb5:git_hash', 'krb5-1.16.1-final')
return {
'depends_on': [folly, fizz, sodium, rsocket, wangle, zstd],
'steps': [
# This isn't a separete spec, since only fbthrift uses mstch.
builder.github_project_workdir('no1msd/mstch', 'build'),
builder.cmake_install('no1msd/mstch'),
builder.github_project_workdir('krb5/krb5', 'src'),
builder.autoconf_install('krb5/krb5'),
builder.fb_github_cmake_install('fbthrift/thrift'),
],
}
<commit_msg>Cut fbcode_builder dep for thrift on krb5
Summary: [Thrift] Cut `fbcode_builder` dep for `thrift` on `krb5`. In the past, Thrift depended on Kerberos and the `krb5` implementation for its transport-layer security. However, Thrift has since migrated fully to Transport Layer Security for its transport-layer security and no longer has any build-time dependency on `krb5`. Clean this up.
Reviewed By: stevegury, vitaut
Differential Revision: D14814205
fbshipit-source-id: dca469d22098e34573674194facaaac6c4c6aa32<commit_after>
|
#!/usr/bin/env python
# Copyright (c) Facebook, Inc. and its affiliates.
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import specs.folly as folly
import specs.fizz as fizz
import specs.rsocket as rsocket
import specs.sodium as sodium
import specs.wangle as wangle
import specs.zstd as zstd
from shell_quoting import ShellQuoted
def fbcode_builder_spec(builder):
# This API should change rarely, so build the latest tag instead of master.
builder.add_option(
'no1msd/mstch:git_hash',
ShellQuoted('$(git describe --abbrev=0 --tags)')
)
return {
'depends_on': [folly, fizz, sodium, rsocket, wangle, zstd],
'steps': [
# This isn't a separete spec, since only fbthrift uses mstch.
builder.github_project_workdir('no1msd/mstch', 'build'),
builder.cmake_install('no1msd/mstch'),
builder.fb_github_cmake_install('fbthrift/thrift'),
],
}
|
#!/usr/bin/env python
# Copyright (c) Facebook, Inc. and its affiliates.
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import specs.folly as folly
import specs.fizz as fizz
import specs.rsocket as rsocket
import specs.sodium as sodium
import specs.wangle as wangle
import specs.zstd as zstd
from shell_quoting import ShellQuoted
def fbcode_builder_spec(builder):
# This API should change rarely, so build the latest tag instead of master.
builder.add_option(
'no1msd/mstch:git_hash',
ShellQuoted('$(git describe --abbrev=0 --tags)')
)
builder.add_option('krb5/krb5:git_hash', 'krb5-1.16.1-final')
return {
'depends_on': [folly, fizz, sodium, rsocket, wangle, zstd],
'steps': [
# This isn't a separete spec, since only fbthrift uses mstch.
builder.github_project_workdir('no1msd/mstch', 'build'),
builder.cmake_install('no1msd/mstch'),
builder.github_project_workdir('krb5/krb5', 'src'),
builder.autoconf_install('krb5/krb5'),
builder.fb_github_cmake_install('fbthrift/thrift'),
],
}
Cut fbcode_builder dep for thrift on krb5
Summary: [Thrift] Cut `fbcode_builder` dep for `thrift` on `krb5`. In the past, Thrift depended on Kerberos and the `krb5` implementation for its transport-layer security. However, Thrift has since migrated fully to Transport Layer Security for its transport-layer security and no longer has any build-time dependency on `krb5`. Clean this up.
Reviewed By: stevegury, vitaut
Differential Revision: D14814205
fbshipit-source-id: dca469d22098e34573674194facaaac6c4c6aa32#!/usr/bin/env python
# Copyright (c) Facebook, Inc. and its affiliates.
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import specs.folly as folly
import specs.fizz as fizz
import specs.rsocket as rsocket
import specs.sodium as sodium
import specs.wangle as wangle
import specs.zstd as zstd
from shell_quoting import ShellQuoted
def fbcode_builder_spec(builder):
# This API should change rarely, so build the latest tag instead of master.
builder.add_option(
'no1msd/mstch:git_hash',
ShellQuoted('$(git describe --abbrev=0 --tags)')
)
return {
'depends_on': [folly, fizz, sodium, rsocket, wangle, zstd],
'steps': [
# This isn't a separete spec, since only fbthrift uses mstch.
builder.github_project_workdir('no1msd/mstch', 'build'),
builder.cmake_install('no1msd/mstch'),
builder.fb_github_cmake_install('fbthrift/thrift'),
],
}
|
<commit_before>#!/usr/bin/env python
# Copyright (c) Facebook, Inc. and its affiliates.
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import specs.folly as folly
import specs.fizz as fizz
import specs.rsocket as rsocket
import specs.sodium as sodium
import specs.wangle as wangle
import specs.zstd as zstd
from shell_quoting import ShellQuoted
def fbcode_builder_spec(builder):
# This API should change rarely, so build the latest tag instead of master.
builder.add_option(
'no1msd/mstch:git_hash',
ShellQuoted('$(git describe --abbrev=0 --tags)')
)
builder.add_option('krb5/krb5:git_hash', 'krb5-1.16.1-final')
return {
'depends_on': [folly, fizz, sodium, rsocket, wangle, zstd],
'steps': [
# This isn't a separete spec, since only fbthrift uses mstch.
builder.github_project_workdir('no1msd/mstch', 'build'),
builder.cmake_install('no1msd/mstch'),
builder.github_project_workdir('krb5/krb5', 'src'),
builder.autoconf_install('krb5/krb5'),
builder.fb_github_cmake_install('fbthrift/thrift'),
],
}
<commit_msg>Cut fbcode_builder dep for thrift on krb5
Summary: [Thrift] Cut `fbcode_builder` dep for `thrift` on `krb5`. In the past, Thrift depended on Kerberos and the `krb5` implementation for its transport-layer security. However, Thrift has since migrated fully to Transport Layer Security for its transport-layer security and no longer has any build-time dependency on `krb5`. Clean this up.
Reviewed By: stevegury, vitaut
Differential Revision: D14814205
fbshipit-source-id: dca469d22098e34573674194facaaac6c4c6aa32<commit_after>#!/usr/bin/env python
# Copyright (c) Facebook, Inc. and its affiliates.
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import specs.folly as folly
import specs.fizz as fizz
import specs.rsocket as rsocket
import specs.sodium as sodium
import specs.wangle as wangle
import specs.zstd as zstd
from shell_quoting import ShellQuoted
def fbcode_builder_spec(builder):
# This API should change rarely, so build the latest tag instead of master.
builder.add_option(
'no1msd/mstch:git_hash',
ShellQuoted('$(git describe --abbrev=0 --tags)')
)
return {
'depends_on': [folly, fizz, sodium, rsocket, wangle, zstd],
'steps': [
# This isn't a separete spec, since only fbthrift uses mstch.
builder.github_project_workdir('no1msd/mstch', 'build'),
builder.cmake_install('no1msd/mstch'),
builder.fb_github_cmake_install('fbthrift/thrift'),
],
}
|
6ec13485a475aeabf8a7fc461b160bbc4a453a00
|
windmill/server/__init__.py
|
windmill/server/__init__.py
|
# Copyright (c) 2006-2007 Open Source Applications Foundation
# Copyright (c) 2008-2009 Mikeal Rogers <mikeal.rogers@gmail.com>
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import wsgi, convergence
forwarding_conditions = [
lambda e : 'google.com/safebrowsing/downloads' not in e['reconstructed_url'],
lambda e : 'mozilla.org/en-US/firefox/livebookmarks.html' not in e['reconstructed_url'],
]
def add_forward_condition(condition):
forwarding_conditions.append(condition)
def remove_forward_condition(condition):
while condition in forwarding_conditions:
forwarding_conditions.remove(condition)
|
# Copyright (c) 2006-2007 Open Source Applications Foundation
# Copyright (c) 2008-2009 Mikeal Rogers <mikeal.rogers@gmail.com>
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import wsgi, convergence
forwarding_conditions = [
lambda e : 'google.com/safebrowsing/downloads' not in e['reconstructed_url'],
lambda e : 'mozilla.org/en-US/firefox/livebookmarks.html' not in e['reconstructed_url'],
lambda e : e.get('CONTENT_TYPE') != 'application/x-shockwave-flash',
]
def add_forward_condition(condition):
forwarding_conditions.append(condition)
def remove_forward_condition(condition):
while condition in forwarding_conditions:
forwarding_conditions.remove(condition)
|
Stop forwarding flash by default, it breaks more than it doesn't.
|
Stop forwarding flash by default, it breaks more than it doesn't.
git-svn-id: 87d19257dd11500985d055ec4730e446075a5f07@1279 78c7df6f-8922-0410-bcd3-9426b1ad491b
|
Python
|
apache-2.0
|
ept/windmill,ept/windmill,ept/windmill
|
# Copyright (c) 2006-2007 Open Source Applications Foundation
# Copyright (c) 2008-2009 Mikeal Rogers <mikeal.rogers@gmail.com>
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import wsgi, convergence
forwarding_conditions = [
lambda e : 'google.com/safebrowsing/downloads' not in e['reconstructed_url'],
lambda e : 'mozilla.org/en-US/firefox/livebookmarks.html' not in e['reconstructed_url'],
]
def add_forward_condition(condition):
forwarding_conditions.append(condition)
def remove_forward_condition(condition):
while condition in forwarding_conditions:
forwarding_conditions.remove(condition)
Stop forwarding flash by default, it breaks more than it doesn't.
git-svn-id: 87d19257dd11500985d055ec4730e446075a5f07@1279 78c7df6f-8922-0410-bcd3-9426b1ad491b
|
# Copyright (c) 2006-2007 Open Source Applications Foundation
# Copyright (c) 2008-2009 Mikeal Rogers <mikeal.rogers@gmail.com>
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import wsgi, convergence
forwarding_conditions = [
lambda e : 'google.com/safebrowsing/downloads' not in e['reconstructed_url'],
lambda e : 'mozilla.org/en-US/firefox/livebookmarks.html' not in e['reconstructed_url'],
lambda e : e.get('CONTENT_TYPE') != 'application/x-shockwave-flash',
]
def add_forward_condition(condition):
forwarding_conditions.append(condition)
def remove_forward_condition(condition):
while condition in forwarding_conditions:
forwarding_conditions.remove(condition)
|
<commit_before># Copyright (c) 2006-2007 Open Source Applications Foundation
# Copyright (c) 2008-2009 Mikeal Rogers <mikeal.rogers@gmail.com>
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import wsgi, convergence
forwarding_conditions = [
lambda e : 'google.com/safebrowsing/downloads' not in e['reconstructed_url'],
lambda e : 'mozilla.org/en-US/firefox/livebookmarks.html' not in e['reconstructed_url'],
]
def add_forward_condition(condition):
forwarding_conditions.append(condition)
def remove_forward_condition(condition):
while condition in forwarding_conditions:
forwarding_conditions.remove(condition)
<commit_msg>Stop forwarding flash by default, it breaks more than it doesn't.
git-svn-id: 87d19257dd11500985d055ec4730e446075a5f07@1279 78c7df6f-8922-0410-bcd3-9426b1ad491b<commit_after>
|
# Copyright (c) 2006-2007 Open Source Applications Foundation
# Copyright (c) 2008-2009 Mikeal Rogers <mikeal.rogers@gmail.com>
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import wsgi, convergence
forwarding_conditions = [
lambda e : 'google.com/safebrowsing/downloads' not in e['reconstructed_url'],
lambda e : 'mozilla.org/en-US/firefox/livebookmarks.html' not in e['reconstructed_url'],
lambda e : e.get('CONTENT_TYPE') != 'application/x-shockwave-flash',
]
def add_forward_condition(condition):
forwarding_conditions.append(condition)
def remove_forward_condition(condition):
while condition in forwarding_conditions:
forwarding_conditions.remove(condition)
|
# Copyright (c) 2006-2007 Open Source Applications Foundation
# Copyright (c) 2008-2009 Mikeal Rogers <mikeal.rogers@gmail.com>
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import wsgi, convergence
forwarding_conditions = [
lambda e : 'google.com/safebrowsing/downloads' not in e['reconstructed_url'],
lambda e : 'mozilla.org/en-US/firefox/livebookmarks.html' not in e['reconstructed_url'],
]
def add_forward_condition(condition):
forwarding_conditions.append(condition)
def remove_forward_condition(condition):
while condition in forwarding_conditions:
forwarding_conditions.remove(condition)
Stop forwarding flash by default, it breaks more than it doesn't.
git-svn-id: 87d19257dd11500985d055ec4730e446075a5f07@1279 78c7df6f-8922-0410-bcd3-9426b1ad491b# Copyright (c) 2006-2007 Open Source Applications Foundation
# Copyright (c) 2008-2009 Mikeal Rogers <mikeal.rogers@gmail.com>
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import wsgi, convergence
forwarding_conditions = [
lambda e : 'google.com/safebrowsing/downloads' not in e['reconstructed_url'],
lambda e : 'mozilla.org/en-US/firefox/livebookmarks.html' not in e['reconstructed_url'],
lambda e : e.get('CONTENT_TYPE') != 'application/x-shockwave-flash',
]
def add_forward_condition(condition):
forwarding_conditions.append(condition)
def remove_forward_condition(condition):
while condition in forwarding_conditions:
forwarding_conditions.remove(condition)
|
<commit_before># Copyright (c) 2006-2007 Open Source Applications Foundation
# Copyright (c) 2008-2009 Mikeal Rogers <mikeal.rogers@gmail.com>
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import wsgi, convergence
forwarding_conditions = [
lambda e : 'google.com/safebrowsing/downloads' not in e['reconstructed_url'],
lambda e : 'mozilla.org/en-US/firefox/livebookmarks.html' not in e['reconstructed_url'],
]
def add_forward_condition(condition):
forwarding_conditions.append(condition)
def remove_forward_condition(condition):
while condition in forwarding_conditions:
forwarding_conditions.remove(condition)
<commit_msg>Stop forwarding flash by default, it breaks more than it doesn't.
git-svn-id: 87d19257dd11500985d055ec4730e446075a5f07@1279 78c7df6f-8922-0410-bcd3-9426b1ad491b<commit_after># Copyright (c) 2006-2007 Open Source Applications Foundation
# Copyright (c) 2008-2009 Mikeal Rogers <mikeal.rogers@gmail.com>
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import wsgi, convergence
forwarding_conditions = [
lambda e : 'google.com/safebrowsing/downloads' not in e['reconstructed_url'],
lambda e : 'mozilla.org/en-US/firefox/livebookmarks.html' not in e['reconstructed_url'],
lambda e : e.get('CONTENT_TYPE') != 'application/x-shockwave-flash',
]
def add_forward_condition(condition):
forwarding_conditions.append(condition)
def remove_forward_condition(condition):
while condition in forwarding_conditions:
forwarding_conditions.remove(condition)
|
9b676c6a4945540a6b23333b43e75c3f539862ae
|
propertyfrontend/__init__.py
|
propertyfrontend/__init__.py
|
import os, logging
from flask import Flask
from flask.ext.basicauth import BasicAuth
from raven.contrib.flask import Sentry
app = Flask(__name__)
app.config.from_object(os.environ.get('SETTINGS'))
if app.config.get('BASIC_AUTH_USERNAME'):
app.config['BASIC_AUTH_FORCE'] = True
basic_auth = BasicAuth(app)
# Sentry exception reporting
if 'SENTRY_DSN' in os.environ:
sentry = Sentry(app, dsn=os.environ['SENTRY_DSN'])
if not app.debug:
app.logger.addHandler(logging.StreamHandler())
app.logger.setLevel(logging.INFO)
app.logger.info("\nConfiguration\n%s\n" % app.config)
@app.context_processor
def asset_path_context_processor():
return {
'asset_path': '/static/build/',
'landregistry_asset_path': '/static/build/'
}
@app.context_processor
def address_processor():
from lrutils import build_address
def process_address_json(address_json):
return build_address(address_json)
return dict(formatted=process_address_json)
|
import os, logging
from flask import Flask
from flask.ext.basicauth import BasicAuth
from raven.contrib.flask import Sentry
app = Flask(__name__)
app.config.from_object(os.environ.get('SETTINGS'))
if app.config.get('BASIC_AUTH_USERNAME'):
app.config['BASIC_AUTH_FORCE'] = True
basic_auth = BasicAuth(app)
# Sentry exception reporting
if 'SENTRY_DSN' in os.environ:
sentry = Sentry(app, dsn=os.environ['SENTRY_DSN'])
if not app.debug:
app.logger.addHandler(logging.StreamHandler())
app.logger.setLevel(logging.INFO)
app.logger.debug("\nConfiguration\n%s\n" % app.config)
@app.context_processor
def asset_path_context_processor():
return {
'asset_path': '/static/build/',
'landregistry_asset_path': '/static/build/'
}
@app.context_processor
def address_processor():
from lrutils import build_address
def process_address_json(address_json):
return build_address(address_json)
return dict(formatted=process_address_json)
|
Set config logging in init to debug
|
Set config logging in init to debug
|
Python
|
mit
|
LandRegistry/property-frontend-alpha,LandRegistry/property-frontend-alpha,LandRegistry/property-frontend-alpha,LandRegistry/property-frontend-alpha
|
import os, logging
from flask import Flask
from flask.ext.basicauth import BasicAuth
from raven.contrib.flask import Sentry
app = Flask(__name__)
app.config.from_object(os.environ.get('SETTINGS'))
if app.config.get('BASIC_AUTH_USERNAME'):
app.config['BASIC_AUTH_FORCE'] = True
basic_auth = BasicAuth(app)
# Sentry exception reporting
if 'SENTRY_DSN' in os.environ:
sentry = Sentry(app, dsn=os.environ['SENTRY_DSN'])
if not app.debug:
app.logger.addHandler(logging.StreamHandler())
app.logger.setLevel(logging.INFO)
app.logger.info("\nConfiguration\n%s\n" % app.config)
@app.context_processor
def asset_path_context_processor():
return {
'asset_path': '/static/build/',
'landregistry_asset_path': '/static/build/'
}
@app.context_processor
def address_processor():
from lrutils import build_address
def process_address_json(address_json):
return build_address(address_json)
return dict(formatted=process_address_json)
Set config logging in init to debug
|
import os, logging
from flask import Flask
from flask.ext.basicauth import BasicAuth
from raven.contrib.flask import Sentry
app = Flask(__name__)
app.config.from_object(os.environ.get('SETTINGS'))
if app.config.get('BASIC_AUTH_USERNAME'):
app.config['BASIC_AUTH_FORCE'] = True
basic_auth = BasicAuth(app)
# Sentry exception reporting
if 'SENTRY_DSN' in os.environ:
sentry = Sentry(app, dsn=os.environ['SENTRY_DSN'])
if not app.debug:
app.logger.addHandler(logging.StreamHandler())
app.logger.setLevel(logging.INFO)
app.logger.debug("\nConfiguration\n%s\n" % app.config)
@app.context_processor
def asset_path_context_processor():
return {
'asset_path': '/static/build/',
'landregistry_asset_path': '/static/build/'
}
@app.context_processor
def address_processor():
from lrutils import build_address
def process_address_json(address_json):
return build_address(address_json)
return dict(formatted=process_address_json)
|
<commit_before>import os, logging
from flask import Flask
from flask.ext.basicauth import BasicAuth
from raven.contrib.flask import Sentry
app = Flask(__name__)
app.config.from_object(os.environ.get('SETTINGS'))
if app.config.get('BASIC_AUTH_USERNAME'):
app.config['BASIC_AUTH_FORCE'] = True
basic_auth = BasicAuth(app)
# Sentry exception reporting
if 'SENTRY_DSN' in os.environ:
sentry = Sentry(app, dsn=os.environ['SENTRY_DSN'])
if not app.debug:
app.logger.addHandler(logging.StreamHandler())
app.logger.setLevel(logging.INFO)
app.logger.info("\nConfiguration\n%s\n" % app.config)
@app.context_processor
def asset_path_context_processor():
return {
'asset_path': '/static/build/',
'landregistry_asset_path': '/static/build/'
}
@app.context_processor
def address_processor():
from lrutils import build_address
def process_address_json(address_json):
return build_address(address_json)
return dict(formatted=process_address_json)
<commit_msg>Set config logging in init to debug<commit_after>
|
import os, logging
from flask import Flask
from flask.ext.basicauth import BasicAuth
from raven.contrib.flask import Sentry
app = Flask(__name__)
app.config.from_object(os.environ.get('SETTINGS'))
if app.config.get('BASIC_AUTH_USERNAME'):
app.config['BASIC_AUTH_FORCE'] = True
basic_auth = BasicAuth(app)
# Sentry exception reporting
if 'SENTRY_DSN' in os.environ:
sentry = Sentry(app, dsn=os.environ['SENTRY_DSN'])
if not app.debug:
app.logger.addHandler(logging.StreamHandler())
app.logger.setLevel(logging.INFO)
app.logger.debug("\nConfiguration\n%s\n" % app.config)
@app.context_processor
def asset_path_context_processor():
return {
'asset_path': '/static/build/',
'landregistry_asset_path': '/static/build/'
}
@app.context_processor
def address_processor():
from lrutils import build_address
def process_address_json(address_json):
return build_address(address_json)
return dict(formatted=process_address_json)
|
import os, logging
from flask import Flask
from flask.ext.basicauth import BasicAuth
from raven.contrib.flask import Sentry
app = Flask(__name__)
app.config.from_object(os.environ.get('SETTINGS'))
if app.config.get('BASIC_AUTH_USERNAME'):
app.config['BASIC_AUTH_FORCE'] = True
basic_auth = BasicAuth(app)
# Sentry exception reporting
if 'SENTRY_DSN' in os.environ:
sentry = Sentry(app, dsn=os.environ['SENTRY_DSN'])
if not app.debug:
app.logger.addHandler(logging.StreamHandler())
app.logger.setLevel(logging.INFO)
app.logger.info("\nConfiguration\n%s\n" % app.config)
@app.context_processor
def asset_path_context_processor():
return {
'asset_path': '/static/build/',
'landregistry_asset_path': '/static/build/'
}
@app.context_processor
def address_processor():
from lrutils import build_address
def process_address_json(address_json):
return build_address(address_json)
return dict(formatted=process_address_json)
Set config logging in init to debugimport os, logging
from flask import Flask
from flask.ext.basicauth import BasicAuth
from raven.contrib.flask import Sentry
app = Flask(__name__)
app.config.from_object(os.environ.get('SETTINGS'))
if app.config.get('BASIC_AUTH_USERNAME'):
app.config['BASIC_AUTH_FORCE'] = True
basic_auth = BasicAuth(app)
# Sentry exception reporting
if 'SENTRY_DSN' in os.environ:
sentry = Sentry(app, dsn=os.environ['SENTRY_DSN'])
if not app.debug:
app.logger.addHandler(logging.StreamHandler())
app.logger.setLevel(logging.INFO)
app.logger.debug("\nConfiguration\n%s\n" % app.config)
@app.context_processor
def asset_path_context_processor():
return {
'asset_path': '/static/build/',
'landregistry_asset_path': '/static/build/'
}
@app.context_processor
def address_processor():
from lrutils import build_address
def process_address_json(address_json):
return build_address(address_json)
return dict(formatted=process_address_json)
|
<commit_before>import os, logging
from flask import Flask
from flask.ext.basicauth import BasicAuth
from raven.contrib.flask import Sentry
app = Flask(__name__)
app.config.from_object(os.environ.get('SETTINGS'))
if app.config.get('BASIC_AUTH_USERNAME'):
app.config['BASIC_AUTH_FORCE'] = True
basic_auth = BasicAuth(app)
# Sentry exception reporting
if 'SENTRY_DSN' in os.environ:
sentry = Sentry(app, dsn=os.environ['SENTRY_DSN'])
if not app.debug:
app.logger.addHandler(logging.StreamHandler())
app.logger.setLevel(logging.INFO)
app.logger.info("\nConfiguration\n%s\n" % app.config)
@app.context_processor
def asset_path_context_processor():
return {
'asset_path': '/static/build/',
'landregistry_asset_path': '/static/build/'
}
@app.context_processor
def address_processor():
from lrutils import build_address
def process_address_json(address_json):
return build_address(address_json)
return dict(formatted=process_address_json)
<commit_msg>Set config logging in init to debug<commit_after>import os, logging
from flask import Flask
from flask.ext.basicauth import BasicAuth
from raven.contrib.flask import Sentry
app = Flask(__name__)
app.config.from_object(os.environ.get('SETTINGS'))
if app.config.get('BASIC_AUTH_USERNAME'):
app.config['BASIC_AUTH_FORCE'] = True
basic_auth = BasicAuth(app)
# Sentry exception reporting
if 'SENTRY_DSN' in os.environ:
sentry = Sentry(app, dsn=os.environ['SENTRY_DSN'])
if not app.debug:
app.logger.addHandler(logging.StreamHandler())
app.logger.setLevel(logging.INFO)
app.logger.debug("\nConfiguration\n%s\n" % app.config)
@app.context_processor
def asset_path_context_processor():
return {
'asset_path': '/static/build/',
'landregistry_asset_path': '/static/build/'
}
@app.context_processor
def address_processor():
from lrutils import build_address
def process_address_json(address_json):
return build_address(address_json)
return dict(formatted=process_address_json)
|
c57910adc6e907881a99e092837fc35e5f45518b
|
survey_creation/config/de_17.py
|
survey_creation/config/de_17.py
|
"""
Config file specific to uk to create automated survey
"""
class config:
# To modify, just add the keys of the dictionary
header_to_modify = [{'class': 'S', 'name': 'sid', 'text': '421498'},
{'class': 'S', 'name': 'admin_email', 'text': 'olivier.philippe@soton.ac.uk'}]
# Same as header_to_modify
description_to_modify = []
# Add header and description. Tuple of a dictionary + the position where it is supposed
# to be inserted
header_to_add = []
# The index positions starts at 0
description_to_add = [({'class': 'SL', 'name': 'additional_languages', 'text': 'nl'}, 12)]
|
"""
Config file specific to uk to create automated survey
"""
class config:
# To modify, just add the keys of the dictionary
header_to_modify = [{'class': 'S', 'name': 'sid', 'text': '421498'},
{'class': 'S', 'name': 'admin_email', 'text': 'olivier.philippe@soton.ac.uk'}]
# Same as header_to_modify
description_to_modify = []
# Add header and description. Tuple of a dictionary + the position where it is supposed
# to be inserted
header_to_add = [({'class': 'S', 'name': 'additional_languages', 'text': 'nl'}, 12)]
# The index positions starts at 0
description_to_add = []
|
Fix issue with headers about additional language in description rather than header
|
Fix issue with headers about additional language in description rather than header
|
Python
|
bsd-3-clause
|
softwaresaved/international-survey
|
"""
Config file specific to uk to create automated survey
"""
class config:
# To modify, just add the keys of the dictionary
header_to_modify = [{'class': 'S', 'name': 'sid', 'text': '421498'},
{'class': 'S', 'name': 'admin_email', 'text': 'olivier.philippe@soton.ac.uk'}]
# Same as header_to_modify
description_to_modify = []
# Add header and description. Tuple of a dictionary + the position where it is supposed
# to be inserted
header_to_add = []
# The index positions starts at 0
description_to_add = [({'class': 'SL', 'name': 'additional_languages', 'text': 'nl'}, 12)]
Fix issue with headers about additional language in description rather than header
|
"""
Config file specific to uk to create automated survey
"""
class config:
# To modify, just add the keys of the dictionary
header_to_modify = [{'class': 'S', 'name': 'sid', 'text': '421498'},
{'class': 'S', 'name': 'admin_email', 'text': 'olivier.philippe@soton.ac.uk'}]
# Same as header_to_modify
description_to_modify = []
# Add header and description. Tuple of a dictionary + the position where it is supposed
# to be inserted
header_to_add = [({'class': 'S', 'name': 'additional_languages', 'text': 'nl'}, 12)]
# The index positions starts at 0
description_to_add = []
|
<commit_before>"""
Config file specific to uk to create automated survey
"""
class config:
# To modify, just add the keys of the dictionary
header_to_modify = [{'class': 'S', 'name': 'sid', 'text': '421498'},
{'class': 'S', 'name': 'admin_email', 'text': 'olivier.philippe@soton.ac.uk'}]
# Same as header_to_modify
description_to_modify = []
# Add header and description. Tuple of a dictionary + the position where it is supposed
# to be inserted
header_to_add = []
# The index positions starts at 0
description_to_add = [({'class': 'SL', 'name': 'additional_languages', 'text': 'nl'}, 12)]
<commit_msg>Fix issue with headers about additional language in description rather than header<commit_after>
|
"""
Config file specific to uk to create automated survey
"""
class config:
# To modify, just add the keys of the dictionary
header_to_modify = [{'class': 'S', 'name': 'sid', 'text': '421498'},
{'class': 'S', 'name': 'admin_email', 'text': 'olivier.philippe@soton.ac.uk'}]
# Same as header_to_modify
description_to_modify = []
# Add header and description. Tuple of a dictionary + the position where it is supposed
# to be inserted
header_to_add = [({'class': 'S', 'name': 'additional_languages', 'text': 'nl'}, 12)]
# The index positions starts at 0
description_to_add = []
|
"""
Config file specific to uk to create automated survey
"""
class config:
# To modify, just add the keys of the dictionary
header_to_modify = [{'class': 'S', 'name': 'sid', 'text': '421498'},
{'class': 'S', 'name': 'admin_email', 'text': 'olivier.philippe@soton.ac.uk'}]
# Same as header_to_modify
description_to_modify = []
# Add header and description. Tuple of a dictionary + the position where it is supposed
# to be inserted
header_to_add = []
# The index positions starts at 0
description_to_add = [({'class': 'SL', 'name': 'additional_languages', 'text': 'nl'}, 12)]
Fix issue with headers about additional language in description rather than header"""
Config file specific to uk to create automated survey
"""
class config:
# To modify, just add the keys of the dictionary
header_to_modify = [{'class': 'S', 'name': 'sid', 'text': '421498'},
{'class': 'S', 'name': 'admin_email', 'text': 'olivier.philippe@soton.ac.uk'}]
# Same as header_to_modify
description_to_modify = []
# Add header and description. Tuple of a dictionary + the position where it is supposed
# to be inserted
header_to_add = [({'class': 'S', 'name': 'additional_languages', 'text': 'nl'}, 12)]
# The index positions starts at 0
description_to_add = []
|
<commit_before>"""
Config file specific to uk to create automated survey
"""
class config:
# To modify, just add the keys of the dictionary
header_to_modify = [{'class': 'S', 'name': 'sid', 'text': '421498'},
{'class': 'S', 'name': 'admin_email', 'text': 'olivier.philippe@soton.ac.uk'}]
# Same as header_to_modify
description_to_modify = []
# Add header and description. Tuple of a dictionary + the position where it is supposed
# to be inserted
header_to_add = []
# The index positions starts at 0
description_to_add = [({'class': 'SL', 'name': 'additional_languages', 'text': 'nl'}, 12)]
<commit_msg>Fix issue with headers about additional language in description rather than header<commit_after>"""
Config file specific to uk to create automated survey
"""
class config:
# To modify, just add the keys of the dictionary
header_to_modify = [{'class': 'S', 'name': 'sid', 'text': '421498'},
{'class': 'S', 'name': 'admin_email', 'text': 'olivier.philippe@soton.ac.uk'}]
# Same as header_to_modify
description_to_modify = []
# Add header and description. Tuple of a dictionary + the position where it is supposed
# to be inserted
header_to_add = [({'class': 'S', 'name': 'additional_languages', 'text': 'nl'}, 12)]
# The index positions starts at 0
description_to_add = []
|
4de89e1d1cf258e903b469deff9d2a7df34a1db9
|
dotfiles/.ipython/profile_default/startup/bytes.py
|
dotfiles/.ipython/profile_default/startup/bytes.py
|
def bytes(b):
""" Print bytes in a humanized way """
def humanize(b, base, suffices=[]):
bb = int(b)
for suffix in suffices:
if bb < base:
break
bb /= float(base)
return "%.2f %s" % (bb, suffix)
print "Base 1024: ", humanize(
b, 1024, ['B', 'KiB', 'MiB', 'GiB', 'TiB', 'PiB', 'EiB'])
print "Base 1000: ", humanize(
b, 1000, ['B', 'KB', 'MB', 'GB', 'TB', 'PB', 'EB'])
|
def bytes(b):
""" Print bytes in a humanized way """
def humanize(b, base, suffices=[]):
bb = int(b)
for suffix in suffices:
if bb < base:
break
bb /= float(base)
return "%.2f %s" % (bb, suffix)
print("Base 1024: ", humanize(
b, 1024, ['B', 'KiB', 'MiB', 'GiB', 'TiB', 'PiB', 'EiB']))
print("Base 1000: ", humanize(
b, 1000, ['B', 'KB', 'MB', 'GB', 'TB', 'PB', 'EB']))
|
Make ipython profile python3 compliant
|
Make ipython profile python3 compliant
|
Python
|
mit
|
izidormatusov/dotfiles,izidormatusov/dotfiles
|
def bytes(b):
""" Print bytes in a humanized way """
def humanize(b, base, suffices=[]):
bb = int(b)
for suffix in suffices:
if bb < base:
break
bb /= float(base)
return "%.2f %s" % (bb, suffix)
print "Base 1024: ", humanize(
b, 1024, ['B', 'KiB', 'MiB', 'GiB', 'TiB', 'PiB', 'EiB'])
print "Base 1000: ", humanize(
b, 1000, ['B', 'KB', 'MB', 'GB', 'TB', 'PB', 'EB'])
Make ipython profile python3 compliant
|
def bytes(b):
""" Print bytes in a humanized way """
def humanize(b, base, suffices=[]):
bb = int(b)
for suffix in suffices:
if bb < base:
break
bb /= float(base)
return "%.2f %s" % (bb, suffix)
print("Base 1024: ", humanize(
b, 1024, ['B', 'KiB', 'MiB', 'GiB', 'TiB', 'PiB', 'EiB']))
print("Base 1000: ", humanize(
b, 1000, ['B', 'KB', 'MB', 'GB', 'TB', 'PB', 'EB']))
|
<commit_before>def bytes(b):
""" Print bytes in a humanized way """
def humanize(b, base, suffices=[]):
bb = int(b)
for suffix in suffices:
if bb < base:
break
bb /= float(base)
return "%.2f %s" % (bb, suffix)
print "Base 1024: ", humanize(
b, 1024, ['B', 'KiB', 'MiB', 'GiB', 'TiB', 'PiB', 'EiB'])
print "Base 1000: ", humanize(
b, 1000, ['B', 'KB', 'MB', 'GB', 'TB', 'PB', 'EB'])
<commit_msg>Make ipython profile python3 compliant<commit_after>
|
def bytes(b):
""" Print bytes in a humanized way """
def humanize(b, base, suffices=[]):
bb = int(b)
for suffix in suffices:
if bb < base:
break
bb /= float(base)
return "%.2f %s" % (bb, suffix)
print("Base 1024: ", humanize(
b, 1024, ['B', 'KiB', 'MiB', 'GiB', 'TiB', 'PiB', 'EiB']))
print("Base 1000: ", humanize(
b, 1000, ['B', 'KB', 'MB', 'GB', 'TB', 'PB', 'EB']))
|
def bytes(b):
""" Print bytes in a humanized way """
def humanize(b, base, suffices=[]):
bb = int(b)
for suffix in suffices:
if bb < base:
break
bb /= float(base)
return "%.2f %s" % (bb, suffix)
print "Base 1024: ", humanize(
b, 1024, ['B', 'KiB', 'MiB', 'GiB', 'TiB', 'PiB', 'EiB'])
print "Base 1000: ", humanize(
b, 1000, ['B', 'KB', 'MB', 'GB', 'TB', 'PB', 'EB'])
Make ipython profile python3 compliantdef bytes(b):
""" Print bytes in a humanized way """
def humanize(b, base, suffices=[]):
bb = int(b)
for suffix in suffices:
if bb < base:
break
bb /= float(base)
return "%.2f %s" % (bb, suffix)
print("Base 1024: ", humanize(
b, 1024, ['B', 'KiB', 'MiB', 'GiB', 'TiB', 'PiB', 'EiB']))
print("Base 1000: ", humanize(
b, 1000, ['B', 'KB', 'MB', 'GB', 'TB', 'PB', 'EB']))
|
<commit_before>def bytes(b):
""" Print bytes in a humanized way """
def humanize(b, base, suffices=[]):
bb = int(b)
for suffix in suffices:
if bb < base:
break
bb /= float(base)
return "%.2f %s" % (bb, suffix)
print "Base 1024: ", humanize(
b, 1024, ['B', 'KiB', 'MiB', 'GiB', 'TiB', 'PiB', 'EiB'])
print "Base 1000: ", humanize(
b, 1000, ['B', 'KB', 'MB', 'GB', 'TB', 'PB', 'EB'])
<commit_msg>Make ipython profile python3 compliant<commit_after>def bytes(b):
""" Print bytes in a humanized way """
def humanize(b, base, suffices=[]):
bb = int(b)
for suffix in suffices:
if bb < base:
break
bb /= float(base)
return "%.2f %s" % (bb, suffix)
print("Base 1024: ", humanize(
b, 1024, ['B', 'KiB', 'MiB', 'GiB', 'TiB', 'PiB', 'EiB']))
print("Base 1000: ", humanize(
b, 1000, ['B', 'KB', 'MB', 'GB', 'TB', 'PB', 'EB']))
|
bed671bdd7dc221e55b5f60c4f9daca3c338a737
|
artists/views.py
|
artists/views.py
|
from django.shortcuts import get_object_or_404
from rest_framework import permissions, viewsets
from similarities.utils import get_similar
from .models import Artist
from similarities.models import UserSimilarity
from .serializers import ArtistSerializer, SimilaritySerializer
class ArtistViewSet(viewsets.ModelViewSet):
"""API endpoint that allows artists to be viewed or edited"""
queryset = Artist.objects.all()
serializer_class = ArtistSerializer
permission_classes = (permissions.IsAuthenticatedOrReadOnly,)
def get_queryset(self):
name = self.request.GET.get('name', "")
if name:
qs = get_similar(name)
else:
qs = super().get_queryset()
return qs[:100]
class SimilarViewSet(viewsets.ModelViewSet):
queryset = UserSimilarity.objects.all()
serializer_class = SimilaritySerializer
permission_classes = (permissions.IsAuthenticated,)
http_method_names = ['get', 'post', 'put', 'delete']
filter_fields = ['cc_artist']
def get_queryset(self):
return super().get_queryset().filter(user=self.request.user)
def pre_save(self, obj):
obj.user = self.request.user
|
from django.shortcuts import get_object_or_404
from rest_framework import permissions, viewsets
from similarities.utils import get_similar
from .models import Artist
from similarities.models import UserSimilarity, Similarity, update_similarities
from .serializers import ArtistSerializer, SimilaritySerializer
class ArtistViewSet(viewsets.ModelViewSet):
"""API endpoint that allows artists to be viewed or edited"""
queryset = Artist.objects.all()
serializer_class = ArtistSerializer
permission_classes = (permissions.IsAuthenticatedOrReadOnly,)
def get_queryset(self):
name = self.request.GET.get('name', "")
if name:
qs = get_similar(name)
else:
qs = super().get_queryset()
return qs[:100]
class SimilarViewSet(viewsets.ModelViewSet):
queryset = UserSimilarity.objects.all()
serializer_class = SimilaritySerializer
permission_classes = (permissions.IsAuthenticated,)
http_method_names = ['get', 'post', 'put', 'delete']
filter_fields = ['cc_artist']
def get_queryset(self):
return super().get_queryset().filter(user=self.request.user)
def pre_save(self, obj):
obj.user = self.request.user
def post_save(self, obj, created=False):
cumulative_similarity, _ = Similarity.objects.get_or_create(
other_artist=obj.other_artist,
cc_artist=obj.cc_artist,
)
update_similarities([cumulative_similarity])
|
Update cumulative similarities on save
|
Update cumulative similarities on save
|
Python
|
bsd-3-clause
|
FreeMusicNinja/api.freemusic.ninja
|
from django.shortcuts import get_object_or_404
from rest_framework import permissions, viewsets
from similarities.utils import get_similar
from .models import Artist
from similarities.models import UserSimilarity
from .serializers import ArtistSerializer, SimilaritySerializer
class ArtistViewSet(viewsets.ModelViewSet):
"""API endpoint that allows artists to be viewed or edited"""
queryset = Artist.objects.all()
serializer_class = ArtistSerializer
permission_classes = (permissions.IsAuthenticatedOrReadOnly,)
def get_queryset(self):
name = self.request.GET.get('name', "")
if name:
qs = get_similar(name)
else:
qs = super().get_queryset()
return qs[:100]
class SimilarViewSet(viewsets.ModelViewSet):
queryset = UserSimilarity.objects.all()
serializer_class = SimilaritySerializer
permission_classes = (permissions.IsAuthenticated,)
http_method_names = ['get', 'post', 'put', 'delete']
filter_fields = ['cc_artist']
def get_queryset(self):
return super().get_queryset().filter(user=self.request.user)
def pre_save(self, obj):
obj.user = self.request.user
Update cumulative similarities on save
|
from django.shortcuts import get_object_or_404
from rest_framework import permissions, viewsets
from similarities.utils import get_similar
from .models import Artist
from similarities.models import UserSimilarity, Similarity, update_similarities
from .serializers import ArtistSerializer, SimilaritySerializer
class ArtistViewSet(viewsets.ModelViewSet):
"""API endpoint that allows artists to be viewed or edited"""
queryset = Artist.objects.all()
serializer_class = ArtistSerializer
permission_classes = (permissions.IsAuthenticatedOrReadOnly,)
def get_queryset(self):
name = self.request.GET.get('name', "")
if name:
qs = get_similar(name)
else:
qs = super().get_queryset()
return qs[:100]
class SimilarViewSet(viewsets.ModelViewSet):
queryset = UserSimilarity.objects.all()
serializer_class = SimilaritySerializer
permission_classes = (permissions.IsAuthenticated,)
http_method_names = ['get', 'post', 'put', 'delete']
filter_fields = ['cc_artist']
def get_queryset(self):
return super().get_queryset().filter(user=self.request.user)
def pre_save(self, obj):
obj.user = self.request.user
def post_save(self, obj, created=False):
cumulative_similarity, _ = Similarity.objects.get_or_create(
other_artist=obj.other_artist,
cc_artist=obj.cc_artist,
)
update_similarities([cumulative_similarity])
|
<commit_before>from django.shortcuts import get_object_or_404
from rest_framework import permissions, viewsets
from similarities.utils import get_similar
from .models import Artist
from similarities.models import UserSimilarity
from .serializers import ArtistSerializer, SimilaritySerializer
class ArtistViewSet(viewsets.ModelViewSet):
"""API endpoint that allows artists to be viewed or edited"""
queryset = Artist.objects.all()
serializer_class = ArtistSerializer
permission_classes = (permissions.IsAuthenticatedOrReadOnly,)
def get_queryset(self):
name = self.request.GET.get('name', "")
if name:
qs = get_similar(name)
else:
qs = super().get_queryset()
return qs[:100]
class SimilarViewSet(viewsets.ModelViewSet):
queryset = UserSimilarity.objects.all()
serializer_class = SimilaritySerializer
permission_classes = (permissions.IsAuthenticated,)
http_method_names = ['get', 'post', 'put', 'delete']
filter_fields = ['cc_artist']
def get_queryset(self):
return super().get_queryset().filter(user=self.request.user)
def pre_save(self, obj):
obj.user = self.request.user
<commit_msg>Update cumulative similarities on save<commit_after>
|
from django.shortcuts import get_object_or_404
from rest_framework import permissions, viewsets
from similarities.utils import get_similar
from .models import Artist
from similarities.models import UserSimilarity, Similarity, update_similarities
from .serializers import ArtistSerializer, SimilaritySerializer
class ArtistViewSet(viewsets.ModelViewSet):
"""API endpoint that allows artists to be viewed or edited"""
queryset = Artist.objects.all()
serializer_class = ArtistSerializer
permission_classes = (permissions.IsAuthenticatedOrReadOnly,)
def get_queryset(self):
name = self.request.GET.get('name', "")
if name:
qs = get_similar(name)
else:
qs = super().get_queryset()
return qs[:100]
class SimilarViewSet(viewsets.ModelViewSet):
queryset = UserSimilarity.objects.all()
serializer_class = SimilaritySerializer
permission_classes = (permissions.IsAuthenticated,)
http_method_names = ['get', 'post', 'put', 'delete']
filter_fields = ['cc_artist']
def get_queryset(self):
return super().get_queryset().filter(user=self.request.user)
def pre_save(self, obj):
obj.user = self.request.user
def post_save(self, obj, created=False):
cumulative_similarity, _ = Similarity.objects.get_or_create(
other_artist=obj.other_artist,
cc_artist=obj.cc_artist,
)
update_similarities([cumulative_similarity])
|
from django.shortcuts import get_object_or_404
from rest_framework import permissions, viewsets
from similarities.utils import get_similar
from .models import Artist
from similarities.models import UserSimilarity
from .serializers import ArtistSerializer, SimilaritySerializer
class ArtistViewSet(viewsets.ModelViewSet):
"""API endpoint that allows artists to be viewed or edited"""
queryset = Artist.objects.all()
serializer_class = ArtistSerializer
permission_classes = (permissions.IsAuthenticatedOrReadOnly,)
def get_queryset(self):
name = self.request.GET.get('name', "")
if name:
qs = get_similar(name)
else:
qs = super().get_queryset()
return qs[:100]
class SimilarViewSet(viewsets.ModelViewSet):
queryset = UserSimilarity.objects.all()
serializer_class = SimilaritySerializer
permission_classes = (permissions.IsAuthenticated,)
http_method_names = ['get', 'post', 'put', 'delete']
filter_fields = ['cc_artist']
def get_queryset(self):
return super().get_queryset().filter(user=self.request.user)
def pre_save(self, obj):
obj.user = self.request.user
Update cumulative similarities on savefrom django.shortcuts import get_object_or_404
from rest_framework import permissions, viewsets
from similarities.utils import get_similar
from .models import Artist
from similarities.models import UserSimilarity, Similarity, update_similarities
from .serializers import ArtistSerializer, SimilaritySerializer
class ArtistViewSet(viewsets.ModelViewSet):
"""API endpoint that allows artists to be viewed or edited"""
queryset = Artist.objects.all()
serializer_class = ArtistSerializer
permission_classes = (permissions.IsAuthenticatedOrReadOnly,)
def get_queryset(self):
name = self.request.GET.get('name', "")
if name:
qs = get_similar(name)
else:
qs = super().get_queryset()
return qs[:100]
class SimilarViewSet(viewsets.ModelViewSet):
queryset = UserSimilarity.objects.all()
serializer_class = SimilaritySerializer
permission_classes = (permissions.IsAuthenticated,)
http_method_names = ['get', 'post', 'put', 'delete']
filter_fields = ['cc_artist']
def get_queryset(self):
return super().get_queryset().filter(user=self.request.user)
def pre_save(self, obj):
obj.user = self.request.user
def post_save(self, obj, created=False):
cumulative_similarity, _ = Similarity.objects.get_or_create(
other_artist=obj.other_artist,
cc_artist=obj.cc_artist,
)
update_similarities([cumulative_similarity])
|
<commit_before>from django.shortcuts import get_object_or_404
from rest_framework import permissions, viewsets
from similarities.utils import get_similar
from .models import Artist
from similarities.models import UserSimilarity
from .serializers import ArtistSerializer, SimilaritySerializer
class ArtistViewSet(viewsets.ModelViewSet):
"""API endpoint that allows artists to be viewed or edited"""
queryset = Artist.objects.all()
serializer_class = ArtistSerializer
permission_classes = (permissions.IsAuthenticatedOrReadOnly,)
def get_queryset(self):
name = self.request.GET.get('name', "")
if name:
qs = get_similar(name)
else:
qs = super().get_queryset()
return qs[:100]
class SimilarViewSet(viewsets.ModelViewSet):
queryset = UserSimilarity.objects.all()
serializer_class = SimilaritySerializer
permission_classes = (permissions.IsAuthenticated,)
http_method_names = ['get', 'post', 'put', 'delete']
filter_fields = ['cc_artist']
def get_queryset(self):
return super().get_queryset().filter(user=self.request.user)
def pre_save(self, obj):
obj.user = self.request.user
<commit_msg>Update cumulative similarities on save<commit_after>from django.shortcuts import get_object_or_404
from rest_framework import permissions, viewsets
from similarities.utils import get_similar
from .models import Artist
from similarities.models import UserSimilarity, Similarity, update_similarities
from .serializers import ArtistSerializer, SimilaritySerializer
class ArtistViewSet(viewsets.ModelViewSet):
"""API endpoint that allows artists to be viewed or edited"""
queryset = Artist.objects.all()
serializer_class = ArtistSerializer
permission_classes = (permissions.IsAuthenticatedOrReadOnly,)
def get_queryset(self):
name = self.request.GET.get('name', "")
if name:
qs = get_similar(name)
else:
qs = super().get_queryset()
return qs[:100]
class SimilarViewSet(viewsets.ModelViewSet):
queryset = UserSimilarity.objects.all()
serializer_class = SimilaritySerializer
permission_classes = (permissions.IsAuthenticated,)
http_method_names = ['get', 'post', 'put', 'delete']
filter_fields = ['cc_artist']
def get_queryset(self):
return super().get_queryset().filter(user=self.request.user)
def pre_save(self, obj):
obj.user = self.request.user
def post_save(self, obj, created=False):
cumulative_similarity, _ = Similarity.objects.get_or_create(
other_artist=obj.other_artist,
cc_artist=obj.cc_artist,
)
update_similarities([cumulative_similarity])
|
42560625d8f83a60320e111503521a9a17d8ae09
|
mollie/api/objects/list.py
|
mollie/api/objects/list.py
|
from .base import Base
class List(Base):
def __init__(self, result, object_type):
Base.__init__(self, result)
self.object_type = object_type
def get_resource_name(self):
return self.object_type.__name__.lower() + 's'
def __iter__(self):
for item in self['_embedded'][self.get_resource_name()]:
yield self.object_type(item)
@property
def count(self):
if 'count' not in self:
return None
return int(self['count'])
def get_offset(self):
if 'offset' not in self:
return None
return self['offset']
|
from .base import Base
class List(Base):
def __init__(self, result, object_type):
Base.__init__(self, result)
self.object_type = object_type
def get_object_name(self):
return self.object_type.__name__.lower() + 's'
def __iter__(self):
for item in self['_embedded'][self.get_object_name()]:
yield self.object_type(item)
@property
def count(self):
if 'count' not in self:
return None
return int(self['count'])
def get_offset(self):
if 'offset' not in self:
return None
return self['offset']
|
Rename method to be more logical
|
Rename method to be more logical
|
Python
|
bsd-2-clause
|
mollie/mollie-api-python
|
from .base import Base
class List(Base):
def __init__(self, result, object_type):
Base.__init__(self, result)
self.object_type = object_type
def get_resource_name(self):
return self.object_type.__name__.lower() + 's'
def __iter__(self):
for item in self['_embedded'][self.get_resource_name()]:
yield self.object_type(item)
@property
def count(self):
if 'count' not in self:
return None
return int(self['count'])
def get_offset(self):
if 'offset' not in self:
return None
return self['offset']
Rename method to be more logical
|
from .base import Base
class List(Base):
def __init__(self, result, object_type):
Base.__init__(self, result)
self.object_type = object_type
def get_object_name(self):
return self.object_type.__name__.lower() + 's'
def __iter__(self):
for item in self['_embedded'][self.get_object_name()]:
yield self.object_type(item)
@property
def count(self):
if 'count' not in self:
return None
return int(self['count'])
def get_offset(self):
if 'offset' not in self:
return None
return self['offset']
|
<commit_before>from .base import Base
class List(Base):
def __init__(self, result, object_type):
Base.__init__(self, result)
self.object_type = object_type
def get_resource_name(self):
return self.object_type.__name__.lower() + 's'
def __iter__(self):
for item in self['_embedded'][self.get_resource_name()]:
yield self.object_type(item)
@property
def count(self):
if 'count' not in self:
return None
return int(self['count'])
def get_offset(self):
if 'offset' not in self:
return None
return self['offset']
<commit_msg>Rename method to be more logical<commit_after>
|
from .base import Base
class List(Base):
def __init__(self, result, object_type):
Base.__init__(self, result)
self.object_type = object_type
def get_object_name(self):
return self.object_type.__name__.lower() + 's'
def __iter__(self):
for item in self['_embedded'][self.get_object_name()]:
yield self.object_type(item)
@property
def count(self):
if 'count' not in self:
return None
return int(self['count'])
def get_offset(self):
if 'offset' not in self:
return None
return self['offset']
|
from .base import Base
class List(Base):
def __init__(self, result, object_type):
Base.__init__(self, result)
self.object_type = object_type
def get_resource_name(self):
return self.object_type.__name__.lower() + 's'
def __iter__(self):
for item in self['_embedded'][self.get_resource_name()]:
yield self.object_type(item)
@property
def count(self):
if 'count' not in self:
return None
return int(self['count'])
def get_offset(self):
if 'offset' not in self:
return None
return self['offset']
Rename method to be more logicalfrom .base import Base
class List(Base):
def __init__(self, result, object_type):
Base.__init__(self, result)
self.object_type = object_type
def get_object_name(self):
return self.object_type.__name__.lower() + 's'
def __iter__(self):
for item in self['_embedded'][self.get_object_name()]:
yield self.object_type(item)
@property
def count(self):
if 'count' not in self:
return None
return int(self['count'])
def get_offset(self):
if 'offset' not in self:
return None
return self['offset']
|
<commit_before>from .base import Base
class List(Base):
def __init__(self, result, object_type):
Base.__init__(self, result)
self.object_type = object_type
def get_resource_name(self):
return self.object_type.__name__.lower() + 's'
def __iter__(self):
for item in self['_embedded'][self.get_resource_name()]:
yield self.object_type(item)
@property
def count(self):
if 'count' not in self:
return None
return int(self['count'])
def get_offset(self):
if 'offset' not in self:
return None
return self['offset']
<commit_msg>Rename method to be more logical<commit_after>from .base import Base
class List(Base):
def __init__(self, result, object_type):
Base.__init__(self, result)
self.object_type = object_type
def get_object_name(self):
return self.object_type.__name__.lower() + 's'
def __iter__(self):
for item in self['_embedded'][self.get_object_name()]:
yield self.object_type(item)
@property
def count(self):
if 'count' not in self:
return None
return int(self['count'])
def get_offset(self):
if 'offset' not in self:
return None
return self['offset']
|
616d92fed79bbfe6ea70ed7e053622819d99088d
|
python/getmonotime.py
|
python/getmonotime.py
|
import getopt, sys
if __name__ == '__main__':
sippy_path = None
try:
opts, args = getopt.getopt(sys.argv[1:], 's:S:i:o:b')
except getopt.GetoptError:
usage()
for o, a in opts:
if o == '-S':
sippy_path = a.strip()
continue
if sippy_path != None:
sys.path.insert(0, sippy_path)
from sippy.Time.clock_dtime import clock_getdtime, CLOCK_MONOTONIC
print clock_getdtime(CLOCK_MONOTONIC)
|
import getopt, sys
if __name__ == '__main__':
sippy_path = None
try:
opts, args = getopt.getopt(sys.argv[1:], 'rS:')
except getopt.GetoptError:
usage()
out_realtime = False
for o, a in opts:
if o == '-S':
sippy_path = a.strip()
continue
if o == '-r':
out_realtime = True
if sippy_path != None:
sys.path.insert(0, sippy_path)
from sippy.Time.clock_dtime import clock_getdtime, CLOCK_MONOTONIC
if not out_realtime:
print(clock_getdtime(CLOCK_MONOTONIC))
else:
from sippy.Time.clock_dtime import CLOCK_REALTIME
print("%f %f" % (clock_getdtime(CLOCK_MONOTONIC), clock_getdtime(CLOCK_REALTIME)))
|
Add an option to also output realtime along with monotime.
|
Add an option to also output realtime along with monotime.
|
Python
|
bsd-2-clause
|
sippy/rtp_cluster,sippy/rtp_cluster
|
import getopt, sys
if __name__ == '__main__':
sippy_path = None
try:
opts, args = getopt.getopt(sys.argv[1:], 's:S:i:o:b')
except getopt.GetoptError:
usage()
for o, a in opts:
if o == '-S':
sippy_path = a.strip()
continue
if sippy_path != None:
sys.path.insert(0, sippy_path)
from sippy.Time.clock_dtime import clock_getdtime, CLOCK_MONOTONIC
print clock_getdtime(CLOCK_MONOTONIC)
Add an option to also output realtime along with monotime.
|
import getopt, sys
if __name__ == '__main__':
sippy_path = None
try:
opts, args = getopt.getopt(sys.argv[1:], 'rS:')
except getopt.GetoptError:
usage()
out_realtime = False
for o, a in opts:
if o == '-S':
sippy_path = a.strip()
continue
if o == '-r':
out_realtime = True
if sippy_path != None:
sys.path.insert(0, sippy_path)
from sippy.Time.clock_dtime import clock_getdtime, CLOCK_MONOTONIC
if not out_realtime:
print(clock_getdtime(CLOCK_MONOTONIC))
else:
from sippy.Time.clock_dtime import CLOCK_REALTIME
print("%f %f" % (clock_getdtime(CLOCK_MONOTONIC), clock_getdtime(CLOCK_REALTIME)))
|
<commit_before>import getopt, sys
if __name__ == '__main__':
sippy_path = None
try:
opts, args = getopt.getopt(sys.argv[1:], 's:S:i:o:b')
except getopt.GetoptError:
usage()
for o, a in opts:
if o == '-S':
sippy_path = a.strip()
continue
if sippy_path != None:
sys.path.insert(0, sippy_path)
from sippy.Time.clock_dtime import clock_getdtime, CLOCK_MONOTONIC
print clock_getdtime(CLOCK_MONOTONIC)
<commit_msg>Add an option to also output realtime along with monotime.<commit_after>
|
import getopt, sys
if __name__ == '__main__':
sippy_path = None
try:
opts, args = getopt.getopt(sys.argv[1:], 'rS:')
except getopt.GetoptError:
usage()
out_realtime = False
for o, a in opts:
if o == '-S':
sippy_path = a.strip()
continue
if o == '-r':
out_realtime = True
if sippy_path != None:
sys.path.insert(0, sippy_path)
from sippy.Time.clock_dtime import clock_getdtime, CLOCK_MONOTONIC
if not out_realtime:
print(clock_getdtime(CLOCK_MONOTONIC))
else:
from sippy.Time.clock_dtime import CLOCK_REALTIME
print("%f %f" % (clock_getdtime(CLOCK_MONOTONIC), clock_getdtime(CLOCK_REALTIME)))
|
import getopt, sys
if __name__ == '__main__':
sippy_path = None
try:
opts, args = getopt.getopt(sys.argv[1:], 's:S:i:o:b')
except getopt.GetoptError:
usage()
for o, a in opts:
if o == '-S':
sippy_path = a.strip()
continue
if sippy_path != None:
sys.path.insert(0, sippy_path)
from sippy.Time.clock_dtime import clock_getdtime, CLOCK_MONOTONIC
print clock_getdtime(CLOCK_MONOTONIC)
Add an option to also output realtime along with monotime.import getopt, sys
if __name__ == '__main__':
sippy_path = None
try:
opts, args = getopt.getopt(sys.argv[1:], 'rS:')
except getopt.GetoptError:
usage()
out_realtime = False
for o, a in opts:
if o == '-S':
sippy_path = a.strip()
continue
if o == '-r':
out_realtime = True
if sippy_path != None:
sys.path.insert(0, sippy_path)
from sippy.Time.clock_dtime import clock_getdtime, CLOCK_MONOTONIC
if not out_realtime:
print(clock_getdtime(CLOCK_MONOTONIC))
else:
from sippy.Time.clock_dtime import CLOCK_REALTIME
print("%f %f" % (clock_getdtime(CLOCK_MONOTONIC), clock_getdtime(CLOCK_REALTIME)))
|
<commit_before>import getopt, sys
if __name__ == '__main__':
sippy_path = None
try:
opts, args = getopt.getopt(sys.argv[1:], 's:S:i:o:b')
except getopt.GetoptError:
usage()
for o, a in opts:
if o == '-S':
sippy_path = a.strip()
continue
if sippy_path != None:
sys.path.insert(0, sippy_path)
from sippy.Time.clock_dtime import clock_getdtime, CLOCK_MONOTONIC
print clock_getdtime(CLOCK_MONOTONIC)
<commit_msg>Add an option to also output realtime along with monotime.<commit_after>import getopt, sys
if __name__ == '__main__':
sippy_path = None
try:
opts, args = getopt.getopt(sys.argv[1:], 'rS:')
except getopt.GetoptError:
usage()
out_realtime = False
for o, a in opts:
if o == '-S':
sippy_path = a.strip()
continue
if o == '-r':
out_realtime = True
if sippy_path != None:
sys.path.insert(0, sippy_path)
from sippy.Time.clock_dtime import clock_getdtime, CLOCK_MONOTONIC
if not out_realtime:
print(clock_getdtime(CLOCK_MONOTONIC))
else:
from sippy.Time.clock_dtime import CLOCK_REALTIME
print("%f %f" % (clock_getdtime(CLOCK_MONOTONIC), clock_getdtime(CLOCK_REALTIME)))
|
ff45b8c21f99b20ed044e8b194bc84f21f4f15d7
|
httpserver_with_post.py
|
httpserver_with_post.py
|
# Adapted from http://stackoverflow.com/questions/10017859/how-to-build-a-simple-http-post-server
# Thank you!
import sys
import BaseHTTPServer
import cgi
class MyHandler(BaseHTTPServer.BaseHTTPRequestHandler):
def do_POST(self):
ctype, pdict = cgi.parse_header(self.headers.getheader('content-type'))
postvars = {}
try:
if ctype == 'application/x-www-form-urlencoded':
length = int(self.headers.getheader('content-length'))
postvars = cgi.parse_qs(self.rfile.read(length), keep_blank_values=1)
self.send_response(200)
self.send_header("Content-type", "text")
self.send_header("Content-length", str(len(body)))
self.end_headers()
self.wfile.write(body)
except:
print "Error"
def httpd(handler_class=MyHandler, server_address = ('127.0.0.1', 8000)):
try:
print "Server started"
srvr = BaseHTTPServer.HTTPServer(server_address, handler_class)
srvr.serve_forever() # serve_forever
except KeyboardInterrupt:
srvr.socket.close()
if __name__ == "__main__":
httpd(server_address = (sys.argv[1], int(sys.argv[2])))
|
# Adapted from http://stackoverflow.com/questions/10017859/how-to-build-a-simple-http-post-server
# Thank you!
import sys
import BaseHTTPServer
import cgi
class MyHandler(BaseHTTPServer.BaseHTTPRequestHandler):
def do_POST(self):
ctype, pdict = cgi.parse_header(self.headers.getheader('content-type'))
postvars = {}
try:
if ctype == 'application/x-www-form-urlencoded':
length = int(self.headers.getheader('content-length'))
postvars = cgi.parse_qs(self.rfile.read(length), keep_blank_values=1)
print "Client posted", postvars
self.send_response(200)
"""
self.send_header("Content-type", "text")
self.send_header("Content-length", str(len(body))) """
self.end_headers()
# self.wfile.write(body)
except Exception, e:
print "Error", repr(e)
def httpd(handler_class=MyHandler, server_address = ('127.0.0.1', 8000)):
try:
print "Server started"
srvr = BaseHTTPServer.HTTPServer(server_address, handler_class)
srvr.serve_forever() # serve_forever
except KeyboardInterrupt:
srvr.socket.close()
if __name__ == "__main__":
httpd(server_address = (sys.argv[1], int(sys.argv[2])))
|
Print client-POSTed data, more verbose error handling
|
Print client-POSTed data, more verbose error handling
And less fiddling with the returned header. For the time being,
I don't care about correcting the bugs in that part of the code.
|
Python
|
unlicense
|
aaaaalbert/repy-doodles
|
# Adapted from http://stackoverflow.com/questions/10017859/how-to-build-a-simple-http-post-server
# Thank you!
import sys
import BaseHTTPServer
import cgi
class MyHandler(BaseHTTPServer.BaseHTTPRequestHandler):
def do_POST(self):
ctype, pdict = cgi.parse_header(self.headers.getheader('content-type'))
postvars = {}
try:
if ctype == 'application/x-www-form-urlencoded':
length = int(self.headers.getheader('content-length'))
postvars = cgi.parse_qs(self.rfile.read(length), keep_blank_values=1)
self.send_response(200)
self.send_header("Content-type", "text")
self.send_header("Content-length", str(len(body)))
self.end_headers()
self.wfile.write(body)
except:
print "Error"
def httpd(handler_class=MyHandler, server_address = ('127.0.0.1', 8000)):
try:
print "Server started"
srvr = BaseHTTPServer.HTTPServer(server_address, handler_class)
srvr.serve_forever() # serve_forever
except KeyboardInterrupt:
srvr.socket.close()
if __name__ == "__main__":
httpd(server_address = (sys.argv[1], int(sys.argv[2])))
Print client-POSTed data, more verbose error handling
And less fiddling with the returned header. For the time being,
I don't care about correcting the bugs in that part of the code.
|
# Adapted from http://stackoverflow.com/questions/10017859/how-to-build-a-simple-http-post-server
# Thank you!
import sys
import BaseHTTPServer
import cgi
class MyHandler(BaseHTTPServer.BaseHTTPRequestHandler):
def do_POST(self):
ctype, pdict = cgi.parse_header(self.headers.getheader('content-type'))
postvars = {}
try:
if ctype == 'application/x-www-form-urlencoded':
length = int(self.headers.getheader('content-length'))
postvars = cgi.parse_qs(self.rfile.read(length), keep_blank_values=1)
print "Client posted", postvars
self.send_response(200)
"""
self.send_header("Content-type", "text")
self.send_header("Content-length", str(len(body))) """
self.end_headers()
# self.wfile.write(body)
except Exception, e:
print "Error", repr(e)
def httpd(handler_class=MyHandler, server_address = ('127.0.0.1', 8000)):
try:
print "Server started"
srvr = BaseHTTPServer.HTTPServer(server_address, handler_class)
srvr.serve_forever() # serve_forever
except KeyboardInterrupt:
srvr.socket.close()
if __name__ == "__main__":
httpd(server_address = (sys.argv[1], int(sys.argv[2])))
|
<commit_before># Adapted from http://stackoverflow.com/questions/10017859/how-to-build-a-simple-http-post-server
# Thank you!
import sys
import BaseHTTPServer
import cgi
class MyHandler(BaseHTTPServer.BaseHTTPRequestHandler):
def do_POST(self):
ctype, pdict = cgi.parse_header(self.headers.getheader('content-type'))
postvars = {}
try:
if ctype == 'application/x-www-form-urlencoded':
length = int(self.headers.getheader('content-length'))
postvars = cgi.parse_qs(self.rfile.read(length), keep_blank_values=1)
self.send_response(200)
self.send_header("Content-type", "text")
self.send_header("Content-length", str(len(body)))
self.end_headers()
self.wfile.write(body)
except:
print "Error"
def httpd(handler_class=MyHandler, server_address = ('127.0.0.1', 8000)):
try:
print "Server started"
srvr = BaseHTTPServer.HTTPServer(server_address, handler_class)
srvr.serve_forever() # serve_forever
except KeyboardInterrupt:
srvr.socket.close()
if __name__ == "__main__":
httpd(server_address = (sys.argv[1], int(sys.argv[2])))
<commit_msg>Print client-POSTed data, more verbose error handling
And less fiddling with the returned header. For the time being,
I don't care about correcting the bugs in that part of the code.<commit_after>
|
# Adapted from http://stackoverflow.com/questions/10017859/how-to-build-a-simple-http-post-server
# Thank you!
import sys
import BaseHTTPServer
import cgi
class MyHandler(BaseHTTPServer.BaseHTTPRequestHandler):
def do_POST(self):
ctype, pdict = cgi.parse_header(self.headers.getheader('content-type'))
postvars = {}
try:
if ctype == 'application/x-www-form-urlencoded':
length = int(self.headers.getheader('content-length'))
postvars = cgi.parse_qs(self.rfile.read(length), keep_blank_values=1)
print "Client posted", postvars
self.send_response(200)
"""
self.send_header("Content-type", "text")
self.send_header("Content-length", str(len(body))) """
self.end_headers()
# self.wfile.write(body)
except Exception, e:
print "Error", repr(e)
def httpd(handler_class=MyHandler, server_address = ('127.0.0.1', 8000)):
try:
print "Server started"
srvr = BaseHTTPServer.HTTPServer(server_address, handler_class)
srvr.serve_forever() # serve_forever
except KeyboardInterrupt:
srvr.socket.close()
if __name__ == "__main__":
httpd(server_address = (sys.argv[1], int(sys.argv[2])))
|
# Adapted from http://stackoverflow.com/questions/10017859/how-to-build-a-simple-http-post-server
# Thank you!
import sys
import BaseHTTPServer
import cgi
class MyHandler(BaseHTTPServer.BaseHTTPRequestHandler):
def do_POST(self):
ctype, pdict = cgi.parse_header(self.headers.getheader('content-type'))
postvars = {}
try:
if ctype == 'application/x-www-form-urlencoded':
length = int(self.headers.getheader('content-length'))
postvars = cgi.parse_qs(self.rfile.read(length), keep_blank_values=1)
self.send_response(200)
self.send_header("Content-type", "text")
self.send_header("Content-length", str(len(body)))
self.end_headers()
self.wfile.write(body)
except:
print "Error"
def httpd(handler_class=MyHandler, server_address = ('127.0.0.1', 8000)):
try:
print "Server started"
srvr = BaseHTTPServer.HTTPServer(server_address, handler_class)
srvr.serve_forever() # serve_forever
except KeyboardInterrupt:
srvr.socket.close()
if __name__ == "__main__":
httpd(server_address = (sys.argv[1], int(sys.argv[2])))
Print client-POSTed data, more verbose error handling
And less fiddling with the returned header. For the time being,
I don't care about correcting the bugs in that part of the code.# Adapted from http://stackoverflow.com/questions/10017859/how-to-build-a-simple-http-post-server
# Thank you!
import sys
import BaseHTTPServer
import cgi
class MyHandler(BaseHTTPServer.BaseHTTPRequestHandler):
def do_POST(self):
ctype, pdict = cgi.parse_header(self.headers.getheader('content-type'))
postvars = {}
try:
if ctype == 'application/x-www-form-urlencoded':
length = int(self.headers.getheader('content-length'))
postvars = cgi.parse_qs(self.rfile.read(length), keep_blank_values=1)
print "Client posted", postvars
self.send_response(200)
"""
self.send_header("Content-type", "text")
self.send_header("Content-length", str(len(body))) """
self.end_headers()
# self.wfile.write(body)
except Exception, e:
print "Error", repr(e)
def httpd(handler_class=MyHandler, server_address = ('127.0.0.1', 8000)):
try:
print "Server started"
srvr = BaseHTTPServer.HTTPServer(server_address, handler_class)
srvr.serve_forever() # serve_forever
except KeyboardInterrupt:
srvr.socket.close()
if __name__ == "__main__":
httpd(server_address = (sys.argv[1], int(sys.argv[2])))
|
<commit_before># Adapted from http://stackoverflow.com/questions/10017859/how-to-build-a-simple-http-post-server
# Thank you!
import sys
import BaseHTTPServer
import cgi
class MyHandler(BaseHTTPServer.BaseHTTPRequestHandler):
def do_POST(self):
ctype, pdict = cgi.parse_header(self.headers.getheader('content-type'))
postvars = {}
try:
if ctype == 'application/x-www-form-urlencoded':
length = int(self.headers.getheader('content-length'))
postvars = cgi.parse_qs(self.rfile.read(length), keep_blank_values=1)
self.send_response(200)
self.send_header("Content-type", "text")
self.send_header("Content-length", str(len(body)))
self.end_headers()
self.wfile.write(body)
except:
print "Error"
def httpd(handler_class=MyHandler, server_address = ('127.0.0.1', 8000)):
try:
print "Server started"
srvr = BaseHTTPServer.HTTPServer(server_address, handler_class)
srvr.serve_forever() # serve_forever
except KeyboardInterrupt:
srvr.socket.close()
if __name__ == "__main__":
httpd(server_address = (sys.argv[1], int(sys.argv[2])))
<commit_msg>Print client-POSTed data, more verbose error handling
And less fiddling with the returned header. For the time being,
I don't care about correcting the bugs in that part of the code.<commit_after># Adapted from http://stackoverflow.com/questions/10017859/how-to-build-a-simple-http-post-server
# Thank you!
import sys
import BaseHTTPServer
import cgi
class MyHandler(BaseHTTPServer.BaseHTTPRequestHandler):
def do_POST(self):
ctype, pdict = cgi.parse_header(self.headers.getheader('content-type'))
postvars = {}
try:
if ctype == 'application/x-www-form-urlencoded':
length = int(self.headers.getheader('content-length'))
postvars = cgi.parse_qs(self.rfile.read(length), keep_blank_values=1)
print "Client posted", postvars
self.send_response(200)
"""
self.send_header("Content-type", "text")
self.send_header("Content-length", str(len(body))) """
self.end_headers()
# self.wfile.write(body)
except Exception, e:
print "Error", repr(e)
def httpd(handler_class=MyHandler, server_address = ('127.0.0.1', 8000)):
try:
print "Server started"
srvr = BaseHTTPServer.HTTPServer(server_address, handler_class)
srvr.serve_forever() # serve_forever
except KeyboardInterrupt:
srvr.socket.close()
if __name__ == "__main__":
httpd(server_address = (sys.argv[1], int(sys.argv[2])))
|
5b54df50752b3f661ad43f2086734f90a8d1a11e
|
src/ggrc/migrations/versions/20150205020509_5254f4f31427_system_editable_object_state.py
|
src/ggrc/migrations/versions/20150205020509_5254f4f31427_system_editable_object_state.py
|
"""System editable object state
Revision ID: 5254f4f31427
Revises: 512c71e4d93b
Create Date: 2015-02-05 02:05:09.351265
"""
# revision identifiers, used by Alembic.
revision = '5254f4f31427'
down_revision = '512c71e4d93b'
import sqlalchemy as sa
from sqlalchemy.sql import table, column
from alembic import op
from ggrc.models.track_object_state import ObjectStates, ObjectStateTables
def upgrade():
for table_name in ObjectStateTables.table_names:
op.add_column(table_name, sa.Column('os_state', sa.String(length=16), nullable=True))
# Set the value into all existing records
object_table = table(table_name,
column('os_state', sa.String(length=16)))
connection = op.get_bind()
connection.execute(
object_table.update().values(
{
'os_state': ObjectStates.DRAFT
}
)
)
# Make the field not-nullable
op.alter_column(table_name, 'os_state',existing_type=sa.String(length=16),nullable=False)
def downgrade():
for table_name in ObjectStateTables:
op.drop_column(table_name, 'os_state')
|
"""System editable object state
Revision ID: 5254f4f31427
Revises: 512c71e4d93b
Create Date: 2015-02-05 02:05:09.351265
"""
# revision identifiers, used by Alembic.
revision = '5254f4f31427'
down_revision = '512c71e4d93b'
import sqlalchemy as sa
from sqlalchemy.sql import table, column
from alembic import op
from ggrc.models.track_object_state import ObjectStates, ObjectStateTables
def upgrade():
for table_name in ObjectStateTables.table_names:
op.add_column(table_name, sa.Column('os_state', sa.String(length=16), nullable=True))
# Set the value into all existing records
object_table = table(table_name,
column('os_state', sa.String(length=16)))
connection = op.get_bind()
connection.execute(
object_table.update().values(
{
'os_state': ObjectStates.DRAFT
}
)
)
# Make the field not-nullable
op.alter_column(table_name, 'os_state',existing_type=sa.String(length=16),nullable=False)
def downgrade():
for table_name in ObjectStateTables.table_names:
op.drop_column(table_name, 'os_state')
|
Fix db_downgrade for "System editable object state"
|
Fix db_downgrade for "System editable object state"
|
Python
|
apache-2.0
|
jmakov/ggrc-core,edofic/ggrc-core,uskudnik/ggrc-core,plamut/ggrc-core,vladan-m/ggrc-core,j0gurt/ggrc-core,selahssea/ggrc-core,andrei-karalionak/ggrc-core,edofic/ggrc-core,AleksNeStu/ggrc-core,selahssea/ggrc-core,j0gurt/ggrc-core,NejcZupec/ggrc-core,VinnieJohns/ggrc-core,hasanalom/ggrc-core,uskudnik/ggrc-core,j0gurt/ggrc-core,kr41/ggrc-core,josthkko/ggrc-core,jmakov/ggrc-core,andrei-karalionak/ggrc-core,prasannav7/ggrc-core,plamut/ggrc-core,hyperNURb/ggrc-core,selahssea/ggrc-core,jmakov/ggrc-core,kr41/ggrc-core,hyperNURb/ggrc-core,andrei-karalionak/ggrc-core,hasanalom/ggrc-core,hasanalom/ggrc-core,uskudnik/ggrc-core,vladan-m/ggrc-core,AleksNeStu/ggrc-core,VinnieJohns/ggrc-core,AleksNeStu/ggrc-core,uskudnik/ggrc-core,kr41/ggrc-core,VinnieJohns/ggrc-core,hasanalom/ggrc-core,josthkko/ggrc-core,vladan-m/ggrc-core,edofic/ggrc-core,AleksNeStu/ggrc-core,NejcZupec/ggrc-core,vladan-m/ggrc-core,hyperNURb/ggrc-core,prasannav7/ggrc-core,NejcZupec/ggrc-core,VinnieJohns/ggrc-core,andrei-karalionak/ggrc-core,j0gurt/ggrc-core,uskudnik/ggrc-core,kr41/ggrc-core,plamut/ggrc-core,selahssea/ggrc-core,vladan-m/ggrc-core,hasanalom/ggrc-core,jmakov/ggrc-core,hyperNURb/ggrc-core,prasannav7/ggrc-core,josthkko/ggrc-core,prasannav7/ggrc-core,josthkko/ggrc-core,hyperNURb/ggrc-core,NejcZupec/ggrc-core,edofic/ggrc-core,jmakov/ggrc-core,plamut/ggrc-core
|
"""System editable object state
Revision ID: 5254f4f31427
Revises: 512c71e4d93b
Create Date: 2015-02-05 02:05:09.351265
"""
# revision identifiers, used by Alembic.
revision = '5254f4f31427'
down_revision = '512c71e4d93b'
import sqlalchemy as sa
from sqlalchemy.sql import table, column
from alembic import op
from ggrc.models.track_object_state import ObjectStates, ObjectStateTables
def upgrade():
for table_name in ObjectStateTables.table_names:
op.add_column(table_name, sa.Column('os_state', sa.String(length=16), nullable=True))
# Set the value into all existing records
object_table = table(table_name,
column('os_state', sa.String(length=16)))
connection = op.get_bind()
connection.execute(
object_table.update().values(
{
'os_state': ObjectStates.DRAFT
}
)
)
# Make the field not-nullable
op.alter_column(table_name, 'os_state',existing_type=sa.String(length=16),nullable=False)
def downgrade():
for table_name in ObjectStateTables:
op.drop_column(table_name, 'os_state')
Fix db_downgrade for "System editable object state"
|
"""System editable object state
Revision ID: 5254f4f31427
Revises: 512c71e4d93b
Create Date: 2015-02-05 02:05:09.351265
"""
# revision identifiers, used by Alembic.
revision = '5254f4f31427'
down_revision = '512c71e4d93b'
import sqlalchemy as sa
from sqlalchemy.sql import table, column
from alembic import op
from ggrc.models.track_object_state import ObjectStates, ObjectStateTables
def upgrade():
for table_name in ObjectStateTables.table_names:
op.add_column(table_name, sa.Column('os_state', sa.String(length=16), nullable=True))
# Set the value into all existing records
object_table = table(table_name,
column('os_state', sa.String(length=16)))
connection = op.get_bind()
connection.execute(
object_table.update().values(
{
'os_state': ObjectStates.DRAFT
}
)
)
# Make the field not-nullable
op.alter_column(table_name, 'os_state',existing_type=sa.String(length=16),nullable=False)
def downgrade():
for table_name in ObjectStateTables.table_names:
op.drop_column(table_name, 'os_state')
|
<commit_before>
"""System editable object state
Revision ID: 5254f4f31427
Revises: 512c71e4d93b
Create Date: 2015-02-05 02:05:09.351265
"""
# revision identifiers, used by Alembic.
revision = '5254f4f31427'
down_revision = '512c71e4d93b'
import sqlalchemy as sa
from sqlalchemy.sql import table, column
from alembic import op
from ggrc.models.track_object_state import ObjectStates, ObjectStateTables
def upgrade():
for table_name in ObjectStateTables.table_names:
op.add_column(table_name, sa.Column('os_state', sa.String(length=16), nullable=True))
# Set the value into all existing records
object_table = table(table_name,
column('os_state', sa.String(length=16)))
connection = op.get_bind()
connection.execute(
object_table.update().values(
{
'os_state': ObjectStates.DRAFT
}
)
)
# Make the field not-nullable
op.alter_column(table_name, 'os_state',existing_type=sa.String(length=16),nullable=False)
def downgrade():
for table_name in ObjectStateTables:
op.drop_column(table_name, 'os_state')
<commit_msg>Fix db_downgrade for "System editable object state"<commit_after>
|
"""System editable object state
Revision ID: 5254f4f31427
Revises: 512c71e4d93b
Create Date: 2015-02-05 02:05:09.351265
"""
# revision identifiers, used by Alembic.
revision = '5254f4f31427'
down_revision = '512c71e4d93b'
import sqlalchemy as sa
from sqlalchemy.sql import table, column
from alembic import op
from ggrc.models.track_object_state import ObjectStates, ObjectStateTables
def upgrade():
for table_name in ObjectStateTables.table_names:
op.add_column(table_name, sa.Column('os_state', sa.String(length=16), nullable=True))
# Set the value into all existing records
object_table = table(table_name,
column('os_state', sa.String(length=16)))
connection = op.get_bind()
connection.execute(
object_table.update().values(
{
'os_state': ObjectStates.DRAFT
}
)
)
# Make the field not-nullable
op.alter_column(table_name, 'os_state',existing_type=sa.String(length=16),nullable=False)
def downgrade():
for table_name in ObjectStateTables.table_names:
op.drop_column(table_name, 'os_state')
|
"""System editable object state
Revision ID: 5254f4f31427
Revises: 512c71e4d93b
Create Date: 2015-02-05 02:05:09.351265
"""
# revision identifiers, used by Alembic.
revision = '5254f4f31427'
down_revision = '512c71e4d93b'
import sqlalchemy as sa
from sqlalchemy.sql import table, column
from alembic import op
from ggrc.models.track_object_state import ObjectStates, ObjectStateTables
def upgrade():
for table_name in ObjectStateTables.table_names:
op.add_column(table_name, sa.Column('os_state', sa.String(length=16), nullable=True))
# Set the value into all existing records
object_table = table(table_name,
column('os_state', sa.String(length=16)))
connection = op.get_bind()
connection.execute(
object_table.update().values(
{
'os_state': ObjectStates.DRAFT
}
)
)
# Make the field not-nullable
op.alter_column(table_name, 'os_state',existing_type=sa.String(length=16),nullable=False)
def downgrade():
for table_name in ObjectStateTables:
op.drop_column(table_name, 'os_state')
Fix db_downgrade for "System editable object state"
"""System editable object state
Revision ID: 5254f4f31427
Revises: 512c71e4d93b
Create Date: 2015-02-05 02:05:09.351265
"""
# revision identifiers, used by Alembic.
revision = '5254f4f31427'
down_revision = '512c71e4d93b'
import sqlalchemy as sa
from sqlalchemy.sql import table, column
from alembic import op
from ggrc.models.track_object_state import ObjectStates, ObjectStateTables
def upgrade():
for table_name in ObjectStateTables.table_names:
op.add_column(table_name, sa.Column('os_state', sa.String(length=16), nullable=True))
# Set the value into all existing records
object_table = table(table_name,
column('os_state', sa.String(length=16)))
connection = op.get_bind()
connection.execute(
object_table.update().values(
{
'os_state': ObjectStates.DRAFT
}
)
)
# Make the field not-nullable
op.alter_column(table_name, 'os_state',existing_type=sa.String(length=16),nullable=False)
def downgrade():
for table_name in ObjectStateTables.table_names:
op.drop_column(table_name, 'os_state')
|
<commit_before>
"""System editable object state
Revision ID: 5254f4f31427
Revises: 512c71e4d93b
Create Date: 2015-02-05 02:05:09.351265
"""
# revision identifiers, used by Alembic.
revision = '5254f4f31427'
down_revision = '512c71e4d93b'
import sqlalchemy as sa
from sqlalchemy.sql import table, column
from alembic import op
from ggrc.models.track_object_state import ObjectStates, ObjectStateTables
def upgrade():
for table_name in ObjectStateTables.table_names:
op.add_column(table_name, sa.Column('os_state', sa.String(length=16), nullable=True))
# Set the value into all existing records
object_table = table(table_name,
column('os_state', sa.String(length=16)))
connection = op.get_bind()
connection.execute(
object_table.update().values(
{
'os_state': ObjectStates.DRAFT
}
)
)
# Make the field not-nullable
op.alter_column(table_name, 'os_state',existing_type=sa.String(length=16),nullable=False)
def downgrade():
for table_name in ObjectStateTables:
op.drop_column(table_name, 'os_state')
<commit_msg>Fix db_downgrade for "System editable object state"<commit_after>
"""System editable object state
Revision ID: 5254f4f31427
Revises: 512c71e4d93b
Create Date: 2015-02-05 02:05:09.351265
"""
# revision identifiers, used by Alembic.
revision = '5254f4f31427'
down_revision = '512c71e4d93b'
import sqlalchemy as sa
from sqlalchemy.sql import table, column
from alembic import op
from ggrc.models.track_object_state import ObjectStates, ObjectStateTables
def upgrade():
for table_name in ObjectStateTables.table_names:
op.add_column(table_name, sa.Column('os_state', sa.String(length=16), nullable=True))
# Set the value into all existing records
object_table = table(table_name,
column('os_state', sa.String(length=16)))
connection = op.get_bind()
connection.execute(
object_table.update().values(
{
'os_state': ObjectStates.DRAFT
}
)
)
# Make the field not-nullable
op.alter_column(table_name, 'os_state',existing_type=sa.String(length=16),nullable=False)
def downgrade():
for table_name in ObjectStateTables.table_names:
op.drop_column(table_name, 'os_state')
|
877a3470044c98d3a938633479d38df6df6d26bd
|
boltiot/urls.py
|
boltiot/urls.py
|
#Creating a key value store for all the urls
BASE_URL = 'http://cloud.boltiot.com/remote/'
url_list = {
'digitalWrite' : '{}/digitalWrite?pin={}&state={}&deviceName={}',
'digitalRead' : '{}/digitalRead?pin={}&deviceName={}',
'analogWrite' : '{}/analogWrite?pin=1&value={}&state={}&deviceName={}',
'analogRead' : '{}/analogRead?pin={}&deviceName={}',
'serialBegin' : '{}/serialBegin?baud={}&deviceName={}',
'serialWrite' : '{}/serialWrite?data={}&deviceName={}',
'serialRead' : '{}/serialRead?till={}&deviceName={}',
'version' : '{}/version?&deviceName={}',
'restart' : '{}/restart?&deviceName={}',
'isAlive' : '{}/isAlive?&deviceName={}',
}
def url(operation):
return BASE_URL+url_list[operation]
|
#Creating a key value store for all the urls
BASE_URL = 'http://cloud.boltiot.com/remote/'
url_list = {
'digitalWrite' : '{}/digitalWrite?pin={}&state={}&deviceName={}',
'digitalRead' : '{}/digitalRead?pin={}&deviceName={}',
'analogWrite' : '{}/analogWrite?pin={}&value={}&deviceName={}',
'analogRead' : '{}/analogRead?pin={}&deviceName={}',
'serialBegin' : '{}/serialBegin?baud={}&deviceName={}',
'serialWrite' : '{}/serialWrite?data={}&deviceName={}',
'serialRead' : '{}/serialRead?till={}&deviceName={}',
'version' : '{}/version?&deviceName={}',
'restart' : '{}/restart?&deviceName={}',
'isAlive' : '{}/isAlive?&deviceName={}',
}
def url(operation):
return BASE_URL+url_list[operation]
|
Remove the static pin fir analog read
|
Remove the static pin fir analog read
|
Python
|
mit
|
Inventrom/bolt-api-python
|
#Creating a key value store for all the urls
BASE_URL = 'http://cloud.boltiot.com/remote/'
url_list = {
'digitalWrite' : '{}/digitalWrite?pin={}&state={}&deviceName={}',
'digitalRead' : '{}/digitalRead?pin={}&deviceName={}',
'analogWrite' : '{}/analogWrite?pin=1&value={}&state={}&deviceName={}',
'analogRead' : '{}/analogRead?pin={}&deviceName={}',
'serialBegin' : '{}/serialBegin?baud={}&deviceName={}',
'serialWrite' : '{}/serialWrite?data={}&deviceName={}',
'serialRead' : '{}/serialRead?till={}&deviceName={}',
'version' : '{}/version?&deviceName={}',
'restart' : '{}/restart?&deviceName={}',
'isAlive' : '{}/isAlive?&deviceName={}',
}
def url(operation):
return BASE_URL+url_list[operation]
Remove the static pin fir analog read
|
#Creating a key value store for all the urls
BASE_URL = 'http://cloud.boltiot.com/remote/'
url_list = {
'digitalWrite' : '{}/digitalWrite?pin={}&state={}&deviceName={}',
'digitalRead' : '{}/digitalRead?pin={}&deviceName={}',
'analogWrite' : '{}/analogWrite?pin={}&value={}&deviceName={}',
'analogRead' : '{}/analogRead?pin={}&deviceName={}',
'serialBegin' : '{}/serialBegin?baud={}&deviceName={}',
'serialWrite' : '{}/serialWrite?data={}&deviceName={}',
'serialRead' : '{}/serialRead?till={}&deviceName={}',
'version' : '{}/version?&deviceName={}',
'restart' : '{}/restart?&deviceName={}',
'isAlive' : '{}/isAlive?&deviceName={}',
}
def url(operation):
return BASE_URL+url_list[operation]
|
<commit_before>#Creating a key value store for all the urls
BASE_URL = 'http://cloud.boltiot.com/remote/'
url_list = {
'digitalWrite' : '{}/digitalWrite?pin={}&state={}&deviceName={}',
'digitalRead' : '{}/digitalRead?pin={}&deviceName={}',
'analogWrite' : '{}/analogWrite?pin=1&value={}&state={}&deviceName={}',
'analogRead' : '{}/analogRead?pin={}&deviceName={}',
'serialBegin' : '{}/serialBegin?baud={}&deviceName={}',
'serialWrite' : '{}/serialWrite?data={}&deviceName={}',
'serialRead' : '{}/serialRead?till={}&deviceName={}',
'version' : '{}/version?&deviceName={}',
'restart' : '{}/restart?&deviceName={}',
'isAlive' : '{}/isAlive?&deviceName={}',
}
def url(operation):
return BASE_URL+url_list[operation]
<commit_msg>Remove the static pin fir analog read<commit_after>
|
#Creating a key value store for all the urls
BASE_URL = 'http://cloud.boltiot.com/remote/'
url_list = {
'digitalWrite' : '{}/digitalWrite?pin={}&state={}&deviceName={}',
'digitalRead' : '{}/digitalRead?pin={}&deviceName={}',
'analogWrite' : '{}/analogWrite?pin={}&value={}&deviceName={}',
'analogRead' : '{}/analogRead?pin={}&deviceName={}',
'serialBegin' : '{}/serialBegin?baud={}&deviceName={}',
'serialWrite' : '{}/serialWrite?data={}&deviceName={}',
'serialRead' : '{}/serialRead?till={}&deviceName={}',
'version' : '{}/version?&deviceName={}',
'restart' : '{}/restart?&deviceName={}',
'isAlive' : '{}/isAlive?&deviceName={}',
}
def url(operation):
return BASE_URL+url_list[operation]
|
#Creating a key value store for all the urls
BASE_URL = 'http://cloud.boltiot.com/remote/'
url_list = {
'digitalWrite' : '{}/digitalWrite?pin={}&state={}&deviceName={}',
'digitalRead' : '{}/digitalRead?pin={}&deviceName={}',
'analogWrite' : '{}/analogWrite?pin=1&value={}&state={}&deviceName={}',
'analogRead' : '{}/analogRead?pin={}&deviceName={}',
'serialBegin' : '{}/serialBegin?baud={}&deviceName={}',
'serialWrite' : '{}/serialWrite?data={}&deviceName={}',
'serialRead' : '{}/serialRead?till={}&deviceName={}',
'version' : '{}/version?&deviceName={}',
'restart' : '{}/restart?&deviceName={}',
'isAlive' : '{}/isAlive?&deviceName={}',
}
def url(operation):
return BASE_URL+url_list[operation]
Remove the static pin fir analog read#Creating a key value store for all the urls
BASE_URL = 'http://cloud.boltiot.com/remote/'
url_list = {
'digitalWrite' : '{}/digitalWrite?pin={}&state={}&deviceName={}',
'digitalRead' : '{}/digitalRead?pin={}&deviceName={}',
'analogWrite' : '{}/analogWrite?pin={}&value={}&deviceName={}',
'analogRead' : '{}/analogRead?pin={}&deviceName={}',
'serialBegin' : '{}/serialBegin?baud={}&deviceName={}',
'serialWrite' : '{}/serialWrite?data={}&deviceName={}',
'serialRead' : '{}/serialRead?till={}&deviceName={}',
'version' : '{}/version?&deviceName={}',
'restart' : '{}/restart?&deviceName={}',
'isAlive' : '{}/isAlive?&deviceName={}',
}
def url(operation):
return BASE_URL+url_list[operation]
|
<commit_before>#Creating a key value store for all the urls
BASE_URL = 'http://cloud.boltiot.com/remote/'
url_list = {
'digitalWrite' : '{}/digitalWrite?pin={}&state={}&deviceName={}',
'digitalRead' : '{}/digitalRead?pin={}&deviceName={}',
'analogWrite' : '{}/analogWrite?pin=1&value={}&state={}&deviceName={}',
'analogRead' : '{}/analogRead?pin={}&deviceName={}',
'serialBegin' : '{}/serialBegin?baud={}&deviceName={}',
'serialWrite' : '{}/serialWrite?data={}&deviceName={}',
'serialRead' : '{}/serialRead?till={}&deviceName={}',
'version' : '{}/version?&deviceName={}',
'restart' : '{}/restart?&deviceName={}',
'isAlive' : '{}/isAlive?&deviceName={}',
}
def url(operation):
return BASE_URL+url_list[operation]
<commit_msg>Remove the static pin fir analog read<commit_after>#Creating a key value store for all the urls
BASE_URL = 'http://cloud.boltiot.com/remote/'
url_list = {
'digitalWrite' : '{}/digitalWrite?pin={}&state={}&deviceName={}',
'digitalRead' : '{}/digitalRead?pin={}&deviceName={}',
'analogWrite' : '{}/analogWrite?pin={}&value={}&deviceName={}',
'analogRead' : '{}/analogRead?pin={}&deviceName={}',
'serialBegin' : '{}/serialBegin?baud={}&deviceName={}',
'serialWrite' : '{}/serialWrite?data={}&deviceName={}',
'serialRead' : '{}/serialRead?till={}&deviceName={}',
'version' : '{}/version?&deviceName={}',
'restart' : '{}/restart?&deviceName={}',
'isAlive' : '{}/isAlive?&deviceName={}',
}
def url(operation):
return BASE_URL+url_list[operation]
|
ab5d570b92aca2c598d12fcdb0b063782ad4c871
|
templates/root/appfiles/urls.py
|
templates/root/appfiles/urls.py
|
"""template URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url, patterns
from django.contrib import admin
from <%= appName %>.forms import AuthenticationForm
urlpatterns = patterns('',
url(r'^$', '<%= appName %>.views.index', name="index"),
url(r'^', include('<%= appName %>.urls')),
url(r'^admin/', include(admin.site.urls)),
)
<% if (includeLoginPage == true) { %>
urlpatterns += patterns(
'django.contrib.auth',
url(r'^accounts/login/$', 'views.login', {
'template_name': 'login.html',
'authentication_form': AuthenticationForm
}, name='login'),
url(r'^accounts/logout/$', 'views.logout', {
'next_page': '/'
}, name='logout'),
)
<% } %>
|
"""template URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url, patterns
from django.contrib import admin
<% if (includeLoginPage == true) { %>
from <%= appName %>.forms import AuthenticationForm
<% } %>
urlpatterns = patterns('',
url(r'^$', '<%= appName %>.views.index', name="index"),
url(r'^', include('<%= appName %>.urls')),
url(r'^admin/', include(admin.site.urls)),
)
<% if (includeLoginPage == true) { %>
urlpatterns += patterns(
'django.contrib.auth',
url(r'^accounts/login/$', 'views.login', {
'template_name': 'login.html',
'authentication_form': AuthenticationForm
}, name='login'),
url(r'^accounts/logout/$', 'views.logout', {
'next_page': '/'
}, name='logout'),
)
<% } %>
|
Fix Import error as a result of answering No to include Login
|
Fix Import error as a result of answering No to include Login
|
Python
|
mit
|
dfurtado/generator-djangospa,dfurtado/generator-djangospa,dfurtado/generator-djangospa
|
"""template URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url, patterns
from django.contrib import admin
from <%= appName %>.forms import AuthenticationForm
urlpatterns = patterns('',
url(r'^$', '<%= appName %>.views.index', name="index"),
url(r'^', include('<%= appName %>.urls')),
url(r'^admin/', include(admin.site.urls)),
)
<% if (includeLoginPage == true) { %>
urlpatterns += patterns(
'django.contrib.auth',
url(r'^accounts/login/$', 'views.login', {
'template_name': 'login.html',
'authentication_form': AuthenticationForm
}, name='login'),
url(r'^accounts/logout/$', 'views.logout', {
'next_page': '/'
}, name='logout'),
)
<% } %>
Fix Import error as a result of answering No to include Login
|
"""template URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url, patterns
from django.contrib import admin
<% if (includeLoginPage == true) { %>
from <%= appName %>.forms import AuthenticationForm
<% } %>
urlpatterns = patterns('',
url(r'^$', '<%= appName %>.views.index', name="index"),
url(r'^', include('<%= appName %>.urls')),
url(r'^admin/', include(admin.site.urls)),
)
<% if (includeLoginPage == true) { %>
urlpatterns += patterns(
'django.contrib.auth',
url(r'^accounts/login/$', 'views.login', {
'template_name': 'login.html',
'authentication_form': AuthenticationForm
}, name='login'),
url(r'^accounts/logout/$', 'views.logout', {
'next_page': '/'
}, name='logout'),
)
<% } %>
|
<commit_before>"""template URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url, patterns
from django.contrib import admin
from <%= appName %>.forms import AuthenticationForm
urlpatterns = patterns('',
url(r'^$', '<%= appName %>.views.index', name="index"),
url(r'^', include('<%= appName %>.urls')),
url(r'^admin/', include(admin.site.urls)),
)
<% if (includeLoginPage == true) { %>
urlpatterns += patterns(
'django.contrib.auth',
url(r'^accounts/login/$', 'views.login', {
'template_name': 'login.html',
'authentication_form': AuthenticationForm
}, name='login'),
url(r'^accounts/logout/$', 'views.logout', {
'next_page': '/'
}, name='logout'),
)
<% } %>
<commit_msg>Fix Import error as a result of answering No to include Login<commit_after>
|
"""template URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url, patterns
from django.contrib import admin
<% if (includeLoginPage == true) { %>
from <%= appName %>.forms import AuthenticationForm
<% } %>
urlpatterns = patterns('',
url(r'^$', '<%= appName %>.views.index', name="index"),
url(r'^', include('<%= appName %>.urls')),
url(r'^admin/', include(admin.site.urls)),
)
<% if (includeLoginPage == true) { %>
urlpatterns += patterns(
'django.contrib.auth',
url(r'^accounts/login/$', 'views.login', {
'template_name': 'login.html',
'authentication_form': AuthenticationForm
}, name='login'),
url(r'^accounts/logout/$', 'views.logout', {
'next_page': '/'
}, name='logout'),
)
<% } %>
|
"""template URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url, patterns
from django.contrib import admin
from <%= appName %>.forms import AuthenticationForm
urlpatterns = patterns('',
url(r'^$', '<%= appName %>.views.index', name="index"),
url(r'^', include('<%= appName %>.urls')),
url(r'^admin/', include(admin.site.urls)),
)
<% if (includeLoginPage == true) { %>
urlpatterns += patterns(
'django.contrib.auth',
url(r'^accounts/login/$', 'views.login', {
'template_name': 'login.html',
'authentication_form': AuthenticationForm
}, name='login'),
url(r'^accounts/logout/$', 'views.logout', {
'next_page': '/'
}, name='logout'),
)
<% } %>
Fix Import error as a result of answering No to include Login"""template URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url, patterns
from django.contrib import admin
<% if (includeLoginPage == true) { %>
from <%= appName %>.forms import AuthenticationForm
<% } %>
urlpatterns = patterns('',
url(r'^$', '<%= appName %>.views.index', name="index"),
url(r'^', include('<%= appName %>.urls')),
url(r'^admin/', include(admin.site.urls)),
)
<% if (includeLoginPage == true) { %>
urlpatterns += patterns(
'django.contrib.auth',
url(r'^accounts/login/$', 'views.login', {
'template_name': 'login.html',
'authentication_form': AuthenticationForm
}, name='login'),
url(r'^accounts/logout/$', 'views.logout', {
'next_page': '/'
}, name='logout'),
)
<% } %>
|
<commit_before>"""template URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url, patterns
from django.contrib import admin
from <%= appName %>.forms import AuthenticationForm
urlpatterns = patterns('',
url(r'^$', '<%= appName %>.views.index', name="index"),
url(r'^', include('<%= appName %>.urls')),
url(r'^admin/', include(admin.site.urls)),
)
<% if (includeLoginPage == true) { %>
urlpatterns += patterns(
'django.contrib.auth',
url(r'^accounts/login/$', 'views.login', {
'template_name': 'login.html',
'authentication_form': AuthenticationForm
}, name='login'),
url(r'^accounts/logout/$', 'views.logout', {
'next_page': '/'
}, name='logout'),
)
<% } %>
<commit_msg>Fix Import error as a result of answering No to include Login<commit_after>"""template URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url, patterns
from django.contrib import admin
<% if (includeLoginPage == true) { %>
from <%= appName %>.forms import AuthenticationForm
<% } %>
urlpatterns = patterns('',
url(r'^$', '<%= appName %>.views.index', name="index"),
url(r'^', include('<%= appName %>.urls')),
url(r'^admin/', include(admin.site.urls)),
)
<% if (includeLoginPage == true) { %>
urlpatterns += patterns(
'django.contrib.auth',
url(r'^accounts/login/$', 'views.login', {
'template_name': 'login.html',
'authentication_form': AuthenticationForm
}, name='login'),
url(r'^accounts/logout/$', 'views.logout', {
'next_page': '/'
}, name='logout'),
)
<% } %>
|
115615a2a183684eed4f11e98a7da12190059fb1
|
armstrong/core/arm_layout/utils.py
|
armstrong/core/arm_layout/utils.py
|
# Here for backwards compatibility (deprecated)
from django.utils.safestring import mark_safe
from django.template.loader import render_to_string
from armstrong.utils.backends import GenericBackend
render_model = (GenericBackend("ARMSTRONG_RENDER_MODEL_BACKEND",
defaults="armstrong.core.arm_layout.backends.BasicRenderModelBackend")
.get_backend())
# Here for backwards compatibility (deprecated)
get_layout_template_name = render_model.get_layout_template_name
|
import warnings
from armstrong.utils.backends import GenericBackend
render_model = (GenericBackend("ARMSTRONG_RENDER_MODEL_BACKEND",
defaults="armstrong.core.arm_layout.backends.BasicRenderModelBackend")
.get_backend())
# DEPRECATED: To be removed in ArmLayout 1.4. Here for backwards compatibility
from django.utils.safestring import mark_safe
from django.template.loader import render_to_string
def deprecate(func):
def wrapper(*args, **kwargs):
msg = "Importing `{}` from this module is deprecated and will be removed in ArmLayout 1.4"
warnings.warn(msg.format(func.__name__), DeprecationWarning)
return func(*args, **kwargs)
return wrapper
mark_safe = deprecate(mark_safe)
render_to_string = deprecate(render_to_string)
get_layout_template_name = deprecate(render_model.get_layout_template_name)
|
Throw deprecation warnings for these imports, which will be removed in the next version. They've been deprecated for two years so we can make it happen.
|
Throw deprecation warnings for these imports, which will be removed in the next version. They've been deprecated for two years so we can make it happen.
|
Python
|
apache-2.0
|
armstrong/armstrong.core.arm_layout,armstrong/armstrong.core.arm_layout
|
# Here for backwards compatibility (deprecated)
from django.utils.safestring import mark_safe
from django.template.loader import render_to_string
from armstrong.utils.backends import GenericBackend
render_model = (GenericBackend("ARMSTRONG_RENDER_MODEL_BACKEND",
defaults="armstrong.core.arm_layout.backends.BasicRenderModelBackend")
.get_backend())
# Here for backwards compatibility (deprecated)
get_layout_template_name = render_model.get_layout_template_name
Throw deprecation warnings for these imports, which will be removed in the next version. They've been deprecated for two years so we can make it happen.
|
import warnings
from armstrong.utils.backends import GenericBackend
render_model = (GenericBackend("ARMSTRONG_RENDER_MODEL_BACKEND",
defaults="armstrong.core.arm_layout.backends.BasicRenderModelBackend")
.get_backend())
# DEPRECATED: To be removed in ArmLayout 1.4. Here for backwards compatibility
from django.utils.safestring import mark_safe
from django.template.loader import render_to_string
def deprecate(func):
def wrapper(*args, **kwargs):
msg = "Importing `{}` from this module is deprecated and will be removed in ArmLayout 1.4"
warnings.warn(msg.format(func.__name__), DeprecationWarning)
return func(*args, **kwargs)
return wrapper
mark_safe = deprecate(mark_safe)
render_to_string = deprecate(render_to_string)
get_layout_template_name = deprecate(render_model.get_layout_template_name)
|
<commit_before># Here for backwards compatibility (deprecated)
from django.utils.safestring import mark_safe
from django.template.loader import render_to_string
from armstrong.utils.backends import GenericBackend
render_model = (GenericBackend("ARMSTRONG_RENDER_MODEL_BACKEND",
defaults="armstrong.core.arm_layout.backends.BasicRenderModelBackend")
.get_backend())
# Here for backwards compatibility (deprecated)
get_layout_template_name = render_model.get_layout_template_name
<commit_msg>Throw deprecation warnings for these imports, which will be removed in the next version. They've been deprecated for two years so we can make it happen.<commit_after>
|
import warnings
from armstrong.utils.backends import GenericBackend
render_model = (GenericBackend("ARMSTRONG_RENDER_MODEL_BACKEND",
defaults="armstrong.core.arm_layout.backends.BasicRenderModelBackend")
.get_backend())
# DEPRECATED: To be removed in ArmLayout 1.4. Here for backwards compatibility
from django.utils.safestring import mark_safe
from django.template.loader import render_to_string
def deprecate(func):
def wrapper(*args, **kwargs):
msg = "Importing `{}` from this module is deprecated and will be removed in ArmLayout 1.4"
warnings.warn(msg.format(func.__name__), DeprecationWarning)
return func(*args, **kwargs)
return wrapper
mark_safe = deprecate(mark_safe)
render_to_string = deprecate(render_to_string)
get_layout_template_name = deprecate(render_model.get_layout_template_name)
|
# Here for backwards compatibility (deprecated)
from django.utils.safestring import mark_safe
from django.template.loader import render_to_string
from armstrong.utils.backends import GenericBackend
render_model = (GenericBackend("ARMSTRONG_RENDER_MODEL_BACKEND",
defaults="armstrong.core.arm_layout.backends.BasicRenderModelBackend")
.get_backend())
# Here for backwards compatibility (deprecated)
get_layout_template_name = render_model.get_layout_template_name
Throw deprecation warnings for these imports, which will be removed in the next version. They've been deprecated for two years so we can make it happen.import warnings
from armstrong.utils.backends import GenericBackend
render_model = (GenericBackend("ARMSTRONG_RENDER_MODEL_BACKEND",
defaults="armstrong.core.arm_layout.backends.BasicRenderModelBackend")
.get_backend())
# DEPRECATED: To be removed in ArmLayout 1.4. Here for backwards compatibility
from django.utils.safestring import mark_safe
from django.template.loader import render_to_string
def deprecate(func):
def wrapper(*args, **kwargs):
msg = "Importing `{}` from this module is deprecated and will be removed in ArmLayout 1.4"
warnings.warn(msg.format(func.__name__), DeprecationWarning)
return func(*args, **kwargs)
return wrapper
mark_safe = deprecate(mark_safe)
render_to_string = deprecate(render_to_string)
get_layout_template_name = deprecate(render_model.get_layout_template_name)
|
<commit_before># Here for backwards compatibility (deprecated)
from django.utils.safestring import mark_safe
from django.template.loader import render_to_string
from armstrong.utils.backends import GenericBackend
render_model = (GenericBackend("ARMSTRONG_RENDER_MODEL_BACKEND",
defaults="armstrong.core.arm_layout.backends.BasicRenderModelBackend")
.get_backend())
# Here for backwards compatibility (deprecated)
get_layout_template_name = render_model.get_layout_template_name
<commit_msg>Throw deprecation warnings for these imports, which will be removed in the next version. They've been deprecated for two years so we can make it happen.<commit_after>import warnings
from armstrong.utils.backends import GenericBackend
render_model = (GenericBackend("ARMSTRONG_RENDER_MODEL_BACKEND",
defaults="armstrong.core.arm_layout.backends.BasicRenderModelBackend")
.get_backend())
# DEPRECATED: To be removed in ArmLayout 1.4. Here for backwards compatibility
from django.utils.safestring import mark_safe
from django.template.loader import render_to_string
def deprecate(func):
def wrapper(*args, **kwargs):
msg = "Importing `{}` from this module is deprecated and will be removed in ArmLayout 1.4"
warnings.warn(msg.format(func.__name__), DeprecationWarning)
return func(*args, **kwargs)
return wrapper
mark_safe = deprecate(mark_safe)
render_to_string = deprecate(render_to_string)
get_layout_template_name = deprecate(render_model.get_layout_template_name)
|
4753dffc6a1672dfa99a5a5da8f082d6554bbb8f
|
http_request_translator/templates/bash_template.py
|
http_request_translator/templates/bash_template.py
|
begin_code = """
#!/usr/bin/env bash
curl -s --request """
request_header = """ --header "{header} : {header_value}" """
code_search = " | egrep --color ' {search_string} |$' "
code_simple = "{method} {url} {headers} --include "
proxy_code = "-x {proxy}"
body_code = " --data '{body}' "
|
begin_code = """
#!/usr/bin/env bash
curl"""
request_header = """ --header "{header} : {header_value}" """
code_search = " | egrep --color ' {search_string} |$'"
code_simple = " -s --request {method} {url} {headers} --include"
proxy_code = " -x {proxy}"
body_code = " --data '{body}'"
|
Fix whitespace in bash script code template
|
Fix whitespace in bash script code template
Signed-off-by: Arun Sori <e3bf7af6e125f7de61de92cd66a64411bed42bee@gmail.com>
|
Python
|
bsd-3-clause
|
owtf/http-request-translator,dhruvagarwal/http-request-translator
|
begin_code = """
#!/usr/bin/env bash
curl -s --request """
request_header = """ --header "{header} : {header_value}" """
code_search = " | egrep --color ' {search_string} |$' "
code_simple = "{method} {url} {headers} --include "
proxy_code = "-x {proxy}"
body_code = " --data '{body}' "
Fix whitespace in bash script code template
Signed-off-by: Arun Sori <e3bf7af6e125f7de61de92cd66a64411bed42bee@gmail.com>
|
begin_code = """
#!/usr/bin/env bash
curl"""
request_header = """ --header "{header} : {header_value}" """
code_search = " | egrep --color ' {search_string} |$'"
code_simple = " -s --request {method} {url} {headers} --include"
proxy_code = " -x {proxy}"
body_code = " --data '{body}'"
|
<commit_before>begin_code = """
#!/usr/bin/env bash
curl -s --request """
request_header = """ --header "{header} : {header_value}" """
code_search = " | egrep --color ' {search_string} |$' "
code_simple = "{method} {url} {headers} --include "
proxy_code = "-x {proxy}"
body_code = " --data '{body}' "
<commit_msg>Fix whitespace in bash script code template
Signed-off-by: Arun Sori <e3bf7af6e125f7de61de92cd66a64411bed42bee@gmail.com><commit_after>
|
begin_code = """
#!/usr/bin/env bash
curl"""
request_header = """ --header "{header} : {header_value}" """
code_search = " | egrep --color ' {search_string} |$'"
code_simple = " -s --request {method} {url} {headers} --include"
proxy_code = " -x {proxy}"
body_code = " --data '{body}'"
|
begin_code = """
#!/usr/bin/env bash
curl -s --request """
request_header = """ --header "{header} : {header_value}" """
code_search = " | egrep --color ' {search_string} |$' "
code_simple = "{method} {url} {headers} --include "
proxy_code = "-x {proxy}"
body_code = " --data '{body}' "
Fix whitespace in bash script code template
Signed-off-by: Arun Sori <e3bf7af6e125f7de61de92cd66a64411bed42bee@gmail.com>begin_code = """
#!/usr/bin/env bash
curl"""
request_header = """ --header "{header} : {header_value}" """
code_search = " | egrep --color ' {search_string} |$'"
code_simple = " -s --request {method} {url} {headers} --include"
proxy_code = " -x {proxy}"
body_code = " --data '{body}'"
|
<commit_before>begin_code = """
#!/usr/bin/env bash
curl -s --request """
request_header = """ --header "{header} : {header_value}" """
code_search = " | egrep --color ' {search_string} |$' "
code_simple = "{method} {url} {headers} --include "
proxy_code = "-x {proxy}"
body_code = " --data '{body}' "
<commit_msg>Fix whitespace in bash script code template
Signed-off-by: Arun Sori <e3bf7af6e125f7de61de92cd66a64411bed42bee@gmail.com><commit_after>begin_code = """
#!/usr/bin/env bash
curl"""
request_header = """ --header "{header} : {header_value}" """
code_search = " | egrep --color ' {search_string} |$'"
code_simple = " -s --request {method} {url} {headers} --include"
proxy_code = " -x {proxy}"
body_code = " --data '{body}'"
|
f1760fe01ae82289d8de2bb9323271edb80d4c08
|
f8a_jobs/graph_sync.py
|
f8a_jobs/graph_sync.py
|
"""Functions to retrieve pending list and invoke Graph Sync."""
import f8a_jobs.defaults as configuration
import requests
import traceback
import logging
logger = logging.getLogger(__name__)
def _api_call(url, params={}):
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending")
try:
logger.info("API Call for url: %s, params: %s" % (url, params))
r = requests.get(url, params=params)
if r is None:
logger.error("Returned response is: %s" % r)
raise Exception("Empty response found")
result = {"data": r.json()}
except Exception:
logger.error(traceback.format_exc())
result = {"error": "Failed to retrieve data from Data Model Importer backend"}
return result
def fetch_pending(params={}):
"""Invoke Pending Graph Sync APIs for given parameters."""
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending")
return _api_call(url, params)
def invoke_sync(params={}):
"""Invoke Graph Sync APIs to sync for given parameters."""
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/sync_all")
return _api_call(url, params)
|
"""Functions to retrieve pending list and invoke Graph Sync."""
import f8a_jobs.defaults as configuration
import requests
import traceback
import logging
logger = logging.getLogger(__name__)
def _api_call(url, params={}):
try:
logger.info("API Call for url: %s, params: %s" % (url, params))
r = requests.get(url, params=params)
if r is None:
logger.error("Returned response is: %s" % r)
raise Exception("Empty response found")
result = {"data": r.json()}
except Exception:
logger.error(traceback.format_exc())
result = {"error": "Failed to retrieve data from Data Model Importer backend"}
return result
def fetch_pending(params={}):
"""Invoke Pending Graph Sync APIs for given parameters."""
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending")
return _api_call(url, params)
def invoke_sync(params={}):
"""Invoke Graph Sync APIs to sync for given parameters."""
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/sync_all")
return _api_call(url, params)
|
Use url from the parameters
|
Use url from the parameters
|
Python
|
apache-2.0
|
fabric8-analytics/fabric8-analytics-jobs,fabric8-analytics/fabric8-analytics-jobs
|
"""Functions to retrieve pending list and invoke Graph Sync."""
import f8a_jobs.defaults as configuration
import requests
import traceback
import logging
logger = logging.getLogger(__name__)
def _api_call(url, params={}):
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending")
try:
logger.info("API Call for url: %s, params: %s" % (url, params))
r = requests.get(url, params=params)
if r is None:
logger.error("Returned response is: %s" % r)
raise Exception("Empty response found")
result = {"data": r.json()}
except Exception:
logger.error(traceback.format_exc())
result = {"error": "Failed to retrieve data from Data Model Importer backend"}
return result
def fetch_pending(params={}):
"""Invoke Pending Graph Sync APIs for given parameters."""
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending")
return _api_call(url, params)
def invoke_sync(params={}):
"""Invoke Graph Sync APIs to sync for given parameters."""
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/sync_all")
return _api_call(url, params)
Use url from the parameters
|
"""Functions to retrieve pending list and invoke Graph Sync."""
import f8a_jobs.defaults as configuration
import requests
import traceback
import logging
logger = logging.getLogger(__name__)
def _api_call(url, params={}):
try:
logger.info("API Call for url: %s, params: %s" % (url, params))
r = requests.get(url, params=params)
if r is None:
logger.error("Returned response is: %s" % r)
raise Exception("Empty response found")
result = {"data": r.json()}
except Exception:
logger.error(traceback.format_exc())
result = {"error": "Failed to retrieve data from Data Model Importer backend"}
return result
def fetch_pending(params={}):
"""Invoke Pending Graph Sync APIs for given parameters."""
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending")
return _api_call(url, params)
def invoke_sync(params={}):
"""Invoke Graph Sync APIs to sync for given parameters."""
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/sync_all")
return _api_call(url, params)
|
<commit_before>"""Functions to retrieve pending list and invoke Graph Sync."""
import f8a_jobs.defaults as configuration
import requests
import traceback
import logging
logger = logging.getLogger(__name__)
def _api_call(url, params={}):
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending")
try:
logger.info("API Call for url: %s, params: %s" % (url, params))
r = requests.get(url, params=params)
if r is None:
logger.error("Returned response is: %s" % r)
raise Exception("Empty response found")
result = {"data": r.json()}
except Exception:
logger.error(traceback.format_exc())
result = {"error": "Failed to retrieve data from Data Model Importer backend"}
return result
def fetch_pending(params={}):
"""Invoke Pending Graph Sync APIs for given parameters."""
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending")
return _api_call(url, params)
def invoke_sync(params={}):
"""Invoke Graph Sync APIs to sync for given parameters."""
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/sync_all")
return _api_call(url, params)
<commit_msg>Use url from the parameters<commit_after>
|
"""Functions to retrieve pending list and invoke Graph Sync."""
import f8a_jobs.defaults as configuration
import requests
import traceback
import logging
logger = logging.getLogger(__name__)
def _api_call(url, params={}):
try:
logger.info("API Call for url: %s, params: %s" % (url, params))
r = requests.get(url, params=params)
if r is None:
logger.error("Returned response is: %s" % r)
raise Exception("Empty response found")
result = {"data": r.json()}
except Exception:
logger.error(traceback.format_exc())
result = {"error": "Failed to retrieve data from Data Model Importer backend"}
return result
def fetch_pending(params={}):
"""Invoke Pending Graph Sync APIs for given parameters."""
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending")
return _api_call(url, params)
def invoke_sync(params={}):
"""Invoke Graph Sync APIs to sync for given parameters."""
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/sync_all")
return _api_call(url, params)
|
"""Functions to retrieve pending list and invoke Graph Sync."""
import f8a_jobs.defaults as configuration
import requests
import traceback
import logging
logger = logging.getLogger(__name__)
def _api_call(url, params={}):
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending")
try:
logger.info("API Call for url: %s, params: %s" % (url, params))
r = requests.get(url, params=params)
if r is None:
logger.error("Returned response is: %s" % r)
raise Exception("Empty response found")
result = {"data": r.json()}
except Exception:
logger.error(traceback.format_exc())
result = {"error": "Failed to retrieve data from Data Model Importer backend"}
return result
def fetch_pending(params={}):
"""Invoke Pending Graph Sync APIs for given parameters."""
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending")
return _api_call(url, params)
def invoke_sync(params={}):
"""Invoke Graph Sync APIs to sync for given parameters."""
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/sync_all")
return _api_call(url, params)
Use url from the parameters"""Functions to retrieve pending list and invoke Graph Sync."""
import f8a_jobs.defaults as configuration
import requests
import traceback
import logging
logger = logging.getLogger(__name__)
def _api_call(url, params={}):
try:
logger.info("API Call for url: %s, params: %s" % (url, params))
r = requests.get(url, params=params)
if r is None:
logger.error("Returned response is: %s" % r)
raise Exception("Empty response found")
result = {"data": r.json()}
except Exception:
logger.error(traceback.format_exc())
result = {"error": "Failed to retrieve data from Data Model Importer backend"}
return result
def fetch_pending(params={}):
"""Invoke Pending Graph Sync APIs for given parameters."""
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending")
return _api_call(url, params)
def invoke_sync(params={}):
"""Invoke Graph Sync APIs to sync for given parameters."""
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/sync_all")
return _api_call(url, params)
|
<commit_before>"""Functions to retrieve pending list and invoke Graph Sync."""
import f8a_jobs.defaults as configuration
import requests
import traceback
import logging
logger = logging.getLogger(__name__)
def _api_call(url, params={}):
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending")
try:
logger.info("API Call for url: %s, params: %s" % (url, params))
r = requests.get(url, params=params)
if r is None:
logger.error("Returned response is: %s" % r)
raise Exception("Empty response found")
result = {"data": r.json()}
except Exception:
logger.error(traceback.format_exc())
result = {"error": "Failed to retrieve data from Data Model Importer backend"}
return result
def fetch_pending(params={}):
"""Invoke Pending Graph Sync APIs for given parameters."""
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending")
return _api_call(url, params)
def invoke_sync(params={}):
"""Invoke Graph Sync APIs to sync for given parameters."""
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/sync_all")
return _api_call(url, params)
<commit_msg>Use url from the parameters<commit_after>"""Functions to retrieve pending list and invoke Graph Sync."""
import f8a_jobs.defaults as configuration
import requests
import traceback
import logging
logger = logging.getLogger(__name__)
def _api_call(url, params={}):
try:
logger.info("API Call for url: %s, params: %s" % (url, params))
r = requests.get(url, params=params)
if r is None:
logger.error("Returned response is: %s" % r)
raise Exception("Empty response found")
result = {"data": r.json()}
except Exception:
logger.error(traceback.format_exc())
result = {"error": "Failed to retrieve data from Data Model Importer backend"}
return result
def fetch_pending(params={}):
"""Invoke Pending Graph Sync APIs for given parameters."""
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/pending")
return _api_call(url, params)
def invoke_sync(params={}):
"""Invoke Graph Sync APIs to sync for given parameters."""
url = "%s%s" % (configuration.DATA_IMPORTER_ENDPOINT, "/api/v1/sync_all")
return _api_call(url, params)
|
d3af229c5c692fdb52c211cd8785bcb7c869090b
|
reobject/query.py
|
reobject/query.py
|
from reobject.utils import signed_attrgetter
class QuerySet(list):
def __init__(self, *args, **kwargs):
super(QuerySet, self).__init__(*args, **kwargs)
def count(self):
return len(self)
def delete(self):
for item in self:
item.delete()
def exists(self):
return bool(self)
def order_by(self, *args):
return type(self)(
sorted(self, key=signed_attrgetter(*args))
)
|
from reobject.utils import signed_attrgetter
class QuerySet(list):
def __init__(self, *args, **kwargs):
super(QuerySet, self).__init__(*args, **kwargs)
def count(self):
return len(self)
def delete(self):
for item in self:
item.delete()
def exists(self):
return bool(self)
def order_by(self, *args):
return type(self)(
sorted(self, key=signed_attrgetter(*args))
)
def reverse(self):
return type(self)(
reversed(self)
)
|
Allow QuerySet objects to be reversed
|
Allow QuerySet objects to be reversed
|
Python
|
apache-2.0
|
onyb/reobject,onyb/reobject
|
from reobject.utils import signed_attrgetter
class QuerySet(list):
def __init__(self, *args, **kwargs):
super(QuerySet, self).__init__(*args, **kwargs)
def count(self):
return len(self)
def delete(self):
for item in self:
item.delete()
def exists(self):
return bool(self)
def order_by(self, *args):
return type(self)(
sorted(self, key=signed_attrgetter(*args))
)
Allow QuerySet objects to be reversed
|
from reobject.utils import signed_attrgetter
class QuerySet(list):
def __init__(self, *args, **kwargs):
super(QuerySet, self).__init__(*args, **kwargs)
def count(self):
return len(self)
def delete(self):
for item in self:
item.delete()
def exists(self):
return bool(self)
def order_by(self, *args):
return type(self)(
sorted(self, key=signed_attrgetter(*args))
)
def reverse(self):
return type(self)(
reversed(self)
)
|
<commit_before>from reobject.utils import signed_attrgetter
class QuerySet(list):
def __init__(self, *args, **kwargs):
super(QuerySet, self).__init__(*args, **kwargs)
def count(self):
return len(self)
def delete(self):
for item in self:
item.delete()
def exists(self):
return bool(self)
def order_by(self, *args):
return type(self)(
sorted(self, key=signed_attrgetter(*args))
)
<commit_msg>Allow QuerySet objects to be reversed<commit_after>
|
from reobject.utils import signed_attrgetter
class QuerySet(list):
def __init__(self, *args, **kwargs):
super(QuerySet, self).__init__(*args, **kwargs)
def count(self):
return len(self)
def delete(self):
for item in self:
item.delete()
def exists(self):
return bool(self)
def order_by(self, *args):
return type(self)(
sorted(self, key=signed_attrgetter(*args))
)
def reverse(self):
return type(self)(
reversed(self)
)
|
from reobject.utils import signed_attrgetter
class QuerySet(list):
def __init__(self, *args, **kwargs):
super(QuerySet, self).__init__(*args, **kwargs)
def count(self):
return len(self)
def delete(self):
for item in self:
item.delete()
def exists(self):
return bool(self)
def order_by(self, *args):
return type(self)(
sorted(self, key=signed_attrgetter(*args))
)
Allow QuerySet objects to be reversedfrom reobject.utils import signed_attrgetter
class QuerySet(list):
def __init__(self, *args, **kwargs):
super(QuerySet, self).__init__(*args, **kwargs)
def count(self):
return len(self)
def delete(self):
for item in self:
item.delete()
def exists(self):
return bool(self)
def order_by(self, *args):
return type(self)(
sorted(self, key=signed_attrgetter(*args))
)
def reverse(self):
return type(self)(
reversed(self)
)
|
<commit_before>from reobject.utils import signed_attrgetter
class QuerySet(list):
def __init__(self, *args, **kwargs):
super(QuerySet, self).__init__(*args, **kwargs)
def count(self):
return len(self)
def delete(self):
for item in self:
item.delete()
def exists(self):
return bool(self)
def order_by(self, *args):
return type(self)(
sorted(self, key=signed_attrgetter(*args))
)
<commit_msg>Allow QuerySet objects to be reversed<commit_after>from reobject.utils import signed_attrgetter
class QuerySet(list):
def __init__(self, *args, **kwargs):
super(QuerySet, self).__init__(*args, **kwargs)
def count(self):
return len(self)
def delete(self):
for item in self:
item.delete()
def exists(self):
return bool(self)
def order_by(self, *args):
return type(self)(
sorted(self, key=signed_attrgetter(*args))
)
def reverse(self):
return type(self)(
reversed(self)
)
|
06c5f27c04de9fa62f6ac4834e0a920349c27084
|
rules/binutils.py
|
rules/binutils.py
|
import xyz
import os
import shutil
class Binutils(xyz.BuildProtocol):
pkg_name = 'binutils'
supported_targets = ['arm-none-eabi']
def check(self, builder):
if builder.target not in self.supported_targets:
raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name))
def configure(self, builder, config):
builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib',
config=config)
def install(self, builder, config):
super().install(builder, config)
# For some reason binutils plonks libiberty.a in the output directory
libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config)
if os.path.exists(libdir):
shutil.rmtree(libdir)
rules = Binutils()
|
import xyz
import os
import shutil
class Binutils(xyz.BuildProtocol):
pkg_name = 'binutils'
supported_targets = ['arm-none-eabi']
def check(self, builder):
if builder.target not in self.supported_targets:
raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name))
def configure(self, builder, config):
builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib',
config=config)
def install(self, builder, config):
super().install(builder, config)
# For some reason binutils plonks libiberty.a in the output directory
libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config)
if os.path.exists(libdir):
shutil.rmtree(libdir)
# For now we strip the man pages.
# man pages created on different systems are (for no good reason) different!
man_dir = builder.j('{install_dir}', config['prefix'][1:], 'share', 'man', config=config)
shutil.rmtree(man_dir)
rules = Binutils()
|
Remove man pages post-install (for now)
|
Remove man pages post-install (for now)
|
Python
|
mit
|
BreakawayConsulting/xyz
|
import xyz
import os
import shutil
class Binutils(xyz.BuildProtocol):
pkg_name = 'binutils'
supported_targets = ['arm-none-eabi']
def check(self, builder):
if builder.target not in self.supported_targets:
raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name))
def configure(self, builder, config):
builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib',
config=config)
def install(self, builder, config):
super().install(builder, config)
# For some reason binutils plonks libiberty.a in the output directory
libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config)
if os.path.exists(libdir):
shutil.rmtree(libdir)
rules = Binutils()
Remove man pages post-install (for now)
|
import xyz
import os
import shutil
class Binutils(xyz.BuildProtocol):
pkg_name = 'binutils'
supported_targets = ['arm-none-eabi']
def check(self, builder):
if builder.target not in self.supported_targets:
raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name))
def configure(self, builder, config):
builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib',
config=config)
def install(self, builder, config):
super().install(builder, config)
# For some reason binutils plonks libiberty.a in the output directory
libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config)
if os.path.exists(libdir):
shutil.rmtree(libdir)
# For now we strip the man pages.
# man pages created on different systems are (for no good reason) different!
man_dir = builder.j('{install_dir}', config['prefix'][1:], 'share', 'man', config=config)
shutil.rmtree(man_dir)
rules = Binutils()
|
<commit_before>import xyz
import os
import shutil
class Binutils(xyz.BuildProtocol):
pkg_name = 'binutils'
supported_targets = ['arm-none-eabi']
def check(self, builder):
if builder.target not in self.supported_targets:
raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name))
def configure(self, builder, config):
builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib',
config=config)
def install(self, builder, config):
super().install(builder, config)
# For some reason binutils plonks libiberty.a in the output directory
libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config)
if os.path.exists(libdir):
shutil.rmtree(libdir)
rules = Binutils()
<commit_msg>Remove man pages post-install (for now)<commit_after>
|
import xyz
import os
import shutil
class Binutils(xyz.BuildProtocol):
pkg_name = 'binutils'
supported_targets = ['arm-none-eabi']
def check(self, builder):
if builder.target not in self.supported_targets:
raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name))
def configure(self, builder, config):
builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib',
config=config)
def install(self, builder, config):
super().install(builder, config)
# For some reason binutils plonks libiberty.a in the output directory
libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config)
if os.path.exists(libdir):
shutil.rmtree(libdir)
# For now we strip the man pages.
# man pages created on different systems are (for no good reason) different!
man_dir = builder.j('{install_dir}', config['prefix'][1:], 'share', 'man', config=config)
shutil.rmtree(man_dir)
rules = Binutils()
|
import xyz
import os
import shutil
class Binutils(xyz.BuildProtocol):
pkg_name = 'binutils'
supported_targets = ['arm-none-eabi']
def check(self, builder):
if builder.target not in self.supported_targets:
raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name))
def configure(self, builder, config):
builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib',
config=config)
def install(self, builder, config):
super().install(builder, config)
# For some reason binutils plonks libiberty.a in the output directory
libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config)
if os.path.exists(libdir):
shutil.rmtree(libdir)
rules = Binutils()
Remove man pages post-install (for now)import xyz
import os
import shutil
class Binutils(xyz.BuildProtocol):
pkg_name = 'binutils'
supported_targets = ['arm-none-eabi']
def check(self, builder):
if builder.target not in self.supported_targets:
raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name))
def configure(self, builder, config):
builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib',
config=config)
def install(self, builder, config):
super().install(builder, config)
# For some reason binutils plonks libiberty.a in the output directory
libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config)
if os.path.exists(libdir):
shutil.rmtree(libdir)
# For now we strip the man pages.
# man pages created on different systems are (for no good reason) different!
man_dir = builder.j('{install_dir}', config['prefix'][1:], 'share', 'man', config=config)
shutil.rmtree(man_dir)
rules = Binutils()
|
<commit_before>import xyz
import os
import shutil
class Binutils(xyz.BuildProtocol):
pkg_name = 'binutils'
supported_targets = ['arm-none-eabi']
def check(self, builder):
if builder.target not in self.supported_targets:
raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name))
def configure(self, builder, config):
builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib',
config=config)
def install(self, builder, config):
super().install(builder, config)
# For some reason binutils plonks libiberty.a in the output directory
libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config)
if os.path.exists(libdir):
shutil.rmtree(libdir)
rules = Binutils()
<commit_msg>Remove man pages post-install (for now)<commit_after>import xyz
import os
import shutil
class Binutils(xyz.BuildProtocol):
pkg_name = 'binutils'
supported_targets = ['arm-none-eabi']
def check(self, builder):
if builder.target not in self.supported_targets:
raise xyz.UsageError("Invalid target ({}) for {}".format(builder.target, self.pkg_name))
def configure(self, builder, config):
builder.cross_configure('--disable-nls', '--enable-lto', '--enable-ld=yes', '--without-zlib',
config=config)
def install(self, builder, config):
super().install(builder, config)
# For some reason binutils plonks libiberty.a in the output directory
libdir = builder.j('{install_dir_abs}', config['eprefix'][1:], 'lib', config=config)
if os.path.exists(libdir):
shutil.rmtree(libdir)
# For now we strip the man pages.
# man pages created on different systems are (for no good reason) different!
man_dir = builder.j('{install_dir}', config['prefix'][1:], 'share', 'man', config=config)
shutil.rmtree(man_dir)
rules = Binutils()
|
7a5f2d0397f8ecda1c1b0517e844eec9d0e3e9d4
|
geotrek/common/urls.py
|
geotrek/common/urls.py
|
from django.conf.urls import patterns, url
from .views import settings_json
urlpatterns += patterns('',
url(r'^api/settings.json', settings_json, name='settings_json'),
)
|
from django.conf.urls import patterns, url
from .views import settings_json
urlpatterns = patterns('',
url(r'^api/settings.json', settings_json, name='settings_json'),
)
|
Fix URL pattern after removing
|
Fix URL pattern after removing
|
Python
|
bsd-2-clause
|
makinacorpus/Geotrek,mabhub/Geotrek,GeotrekCE/Geotrek-admin,makinacorpus/Geotrek,Anaethelion/Geotrek,johan--/Geotrek,Anaethelion/Geotrek,Anaethelion/Geotrek,GeotrekCE/Geotrek-admin,mabhub/Geotrek,GeotrekCE/Geotrek-admin,camillemonchicourt/Geotrek,makinacorpus/Geotrek,makinacorpus/Geotrek,mabhub/Geotrek,johan--/Geotrek,Anaethelion/Geotrek,johan--/Geotrek,johan--/Geotrek,camillemonchicourt/Geotrek,mabhub/Geotrek,camillemonchicourt/Geotrek,GeotrekCE/Geotrek-admin
|
from django.conf.urls import patterns, url
from .views import settings_json
urlpatterns += patterns('',
url(r'^api/settings.json', settings_json, name='settings_json'),
)
Fix URL pattern after removing
|
from django.conf.urls import patterns, url
from .views import settings_json
urlpatterns = patterns('',
url(r'^api/settings.json', settings_json, name='settings_json'),
)
|
<commit_before>from django.conf.urls import patterns, url
from .views import settings_json
urlpatterns += patterns('',
url(r'^api/settings.json', settings_json, name='settings_json'),
)
<commit_msg>Fix URL pattern after removing<commit_after>
|
from django.conf.urls import patterns, url
from .views import settings_json
urlpatterns = patterns('',
url(r'^api/settings.json', settings_json, name='settings_json'),
)
|
from django.conf.urls import patterns, url
from .views import settings_json
urlpatterns += patterns('',
url(r'^api/settings.json', settings_json, name='settings_json'),
)
Fix URL pattern after removingfrom django.conf.urls import patterns, url
from .views import settings_json
urlpatterns = patterns('',
url(r'^api/settings.json', settings_json, name='settings_json'),
)
|
<commit_before>from django.conf.urls import patterns, url
from .views import settings_json
urlpatterns += patterns('',
url(r'^api/settings.json', settings_json, name='settings_json'),
)
<commit_msg>Fix URL pattern after removing<commit_after>from django.conf.urls import patterns, url
from .views import settings_json
urlpatterns = patterns('',
url(r'^api/settings.json', settings_json, name='settings_json'),
)
|
9c0d88ba1681949c02f2cd136efc0de1c23d170d
|
simuvex/procedures/libc___so___6/fileno.py
|
simuvex/procedures/libc___so___6/fileno.py
|
import simuvex
from simuvex.s_type import SimTypeFd
import logging
l = logging.getLogger("simuvex.procedures.fileno")
######################################
# memset
######################################
class fileno(simuvex.SimProcedure):
#pylint:disable=arguments-differ
def run(self, f):
self.argument_types = {0: SimTypeFd()}
self.return_type = SimTypeFd()
return f
|
import simuvex
from simuvex.s_type import SimTypeFd, SimTypeTop
from . import io_file_data_for_arch
import logging
l = logging.getLogger("simuvex.procedures.fileno")
######################################
# fileno
######################################
class fileno(simuvex.SimProcedure):
#pylint:disable=arguments-differ
def run(self, f):
self.argument_types = {0: self.ty_ptr(SimTypeTop())}
self.return_type = SimTypeFd()
# Get FILE struct
io_file_data = io_file_data_for_arch(self.state.arch)
# Get the file descriptor from FILE struct
fd = self.state.se.any_int(self.state.memory.load(f + io_file_data['fd'],
4 * 8, # int
endness=self.state.arch.memory_endness))
return fd
|
Add logic for grabbing file descriptor from FILE struct
|
Add logic for grabbing file descriptor from FILE struct
|
Python
|
bsd-2-clause
|
chubbymaggie/angr,angr/angr,schieb/angr,tyb0807/angr,axt/angr,f-prettyland/angr,axt/angr,iamahuman/angr,angr/angr,iamahuman/angr,axt/angr,iamahuman/angr,f-prettyland/angr,chubbymaggie/angr,schieb/angr,angr/angr,tyb0807/angr,tyb0807/angr,chubbymaggie/angr,f-prettyland/angr,angr/simuvex,schieb/angr
|
import simuvex
from simuvex.s_type import SimTypeFd
import logging
l = logging.getLogger("simuvex.procedures.fileno")
######################################
# memset
######################################
class fileno(simuvex.SimProcedure):
#pylint:disable=arguments-differ
def run(self, f):
self.argument_types = {0: SimTypeFd()}
self.return_type = SimTypeFd()
return f
Add logic for grabbing file descriptor from FILE struct
|
import simuvex
from simuvex.s_type import SimTypeFd, SimTypeTop
from . import io_file_data_for_arch
import logging
l = logging.getLogger("simuvex.procedures.fileno")
######################################
# fileno
######################################
class fileno(simuvex.SimProcedure):
#pylint:disable=arguments-differ
def run(self, f):
self.argument_types = {0: self.ty_ptr(SimTypeTop())}
self.return_type = SimTypeFd()
# Get FILE struct
io_file_data = io_file_data_for_arch(self.state.arch)
# Get the file descriptor from FILE struct
fd = self.state.se.any_int(self.state.memory.load(f + io_file_data['fd'],
4 * 8, # int
endness=self.state.arch.memory_endness))
return fd
|
<commit_before>import simuvex
from simuvex.s_type import SimTypeFd
import logging
l = logging.getLogger("simuvex.procedures.fileno")
######################################
# memset
######################################
class fileno(simuvex.SimProcedure):
#pylint:disable=arguments-differ
def run(self, f):
self.argument_types = {0: SimTypeFd()}
self.return_type = SimTypeFd()
return f
<commit_msg>Add logic for grabbing file descriptor from FILE struct<commit_after>
|
import simuvex
from simuvex.s_type import SimTypeFd, SimTypeTop
from . import io_file_data_for_arch
import logging
l = logging.getLogger("simuvex.procedures.fileno")
######################################
# fileno
######################################
class fileno(simuvex.SimProcedure):
#pylint:disable=arguments-differ
def run(self, f):
self.argument_types = {0: self.ty_ptr(SimTypeTop())}
self.return_type = SimTypeFd()
# Get FILE struct
io_file_data = io_file_data_for_arch(self.state.arch)
# Get the file descriptor from FILE struct
fd = self.state.se.any_int(self.state.memory.load(f + io_file_data['fd'],
4 * 8, # int
endness=self.state.arch.memory_endness))
return fd
|
import simuvex
from simuvex.s_type import SimTypeFd
import logging
l = logging.getLogger("simuvex.procedures.fileno")
######################################
# memset
######################################
class fileno(simuvex.SimProcedure):
#pylint:disable=arguments-differ
def run(self, f):
self.argument_types = {0: SimTypeFd()}
self.return_type = SimTypeFd()
return f
Add logic for grabbing file descriptor from FILE structimport simuvex
from simuvex.s_type import SimTypeFd, SimTypeTop
from . import io_file_data_for_arch
import logging
l = logging.getLogger("simuvex.procedures.fileno")
######################################
# fileno
######################################
class fileno(simuvex.SimProcedure):
#pylint:disable=arguments-differ
def run(self, f):
self.argument_types = {0: self.ty_ptr(SimTypeTop())}
self.return_type = SimTypeFd()
# Get FILE struct
io_file_data = io_file_data_for_arch(self.state.arch)
# Get the file descriptor from FILE struct
fd = self.state.se.any_int(self.state.memory.load(f + io_file_data['fd'],
4 * 8, # int
endness=self.state.arch.memory_endness))
return fd
|
<commit_before>import simuvex
from simuvex.s_type import SimTypeFd
import logging
l = logging.getLogger("simuvex.procedures.fileno")
######################################
# memset
######################################
class fileno(simuvex.SimProcedure):
#pylint:disable=arguments-differ
def run(self, f):
self.argument_types = {0: SimTypeFd()}
self.return_type = SimTypeFd()
return f
<commit_msg>Add logic for grabbing file descriptor from FILE struct<commit_after>import simuvex
from simuvex.s_type import SimTypeFd, SimTypeTop
from . import io_file_data_for_arch
import logging
l = logging.getLogger("simuvex.procedures.fileno")
######################################
# fileno
######################################
class fileno(simuvex.SimProcedure):
#pylint:disable=arguments-differ
def run(self, f):
self.argument_types = {0: self.ty_ptr(SimTypeTop())}
self.return_type = SimTypeFd()
# Get FILE struct
io_file_data = io_file_data_for_arch(self.state.arch)
# Get the file descriptor from FILE struct
fd = self.state.se.any_int(self.state.memory.load(f + io_file_data['fd'],
4 * 8, # int
endness=self.state.arch.memory_endness))
return fd
|
d387ab236634f91186805dd114ee85455d1244f8
|
pywikibot/echo.py
|
pywikibot/echo.py
|
# -*- coding: utf-8 -*-
"""Classes and functions for working with the Echo extension."""
from __future__ import absolute_import, unicode_literals
import pywikibot
class Notification(object):
"""A notification issued by the Echo extension."""
def __init__(self, site):
"""Construct an empty Notification object."""
self.site = site
@classmethod
def fromJSON(cls, site, data):
"""
Construct a Notification object from JSON data returned by the API.
@rtype: Notification
"""
notif = cls(site)
notif.id = data['id'] # TODO: use numeric id ?
notif.type = data['type']
notif.category = data['category']
notif.timestamp = pywikibot.Timestamp.fromtimestampformat(data['timestamp']['mw'])
# TODO: use 'namespace-key' + 'text' ?
notif.page = pywikibot.Page(site, data['title']['full'])
if 'agent' in data and 'name' in data['agent']:
notif.agent = pywikibot.User(site, data['agent']['name'])
else:
notif.agent = None
if 'read' in data:
notif.read = pywikibot.Timestamp.fromtimestampformat(data['read'])
else:
notif.read = False
notif.content = data.get('*', None)
return notif
def mark_as_read(self):
"""Mark the notification as read."""
return self.site.notifications_mark_read(list=self.id)
|
# -*- coding: utf-8 -*-
"""Classes and functions for working with the Echo extension."""
from __future__ import absolute_import, unicode_literals
import pywikibot
class Notification(object):
"""A notification issued by the Echo extension."""
def __init__(self, site):
"""Construct an empty Notification object."""
self.site = site
@classmethod
def fromJSON(cls, site, data):
"""
Construct a Notification object from JSON data returned by the API.
@rtype: Notification
"""
notif = cls(site)
notif.id = data['id'] # TODO: use numeric id ?
notif.type = data['type']
notif.category = data['category']
notif.timestamp = pywikibot.Timestamp.fromtimestampformat(data['timestamp']['mw'])
if 'title' in data and 'full' in data['title']:
notif.page = pywikibot.Page(site, data['title']['full'])
else:
notif.page = None
if 'agent' in data and 'name' in data['agent']:
notif.agent = pywikibot.User(site, data['agent']['name'])
else:
notif.agent = None
if 'read' in data:
notif.read = pywikibot.Timestamp.fromtimestampformat(data['read'])
else:
notif.read = False
notif.content = data.get('*', None)
return notif
def mark_as_read(self):
"""Mark the notification as read."""
return self.site.notifications_mark_read(list=self.id)
|
Fix notifications building from JSON
|
Fix notifications building from JSON
Sometimes (like in welcome messages), notifications don't have a
'title' property, so we shouldn't assume there is one.
Bug: T139015
Change-Id: I83e480d04e8e09aa9bcb5edef4f56b47d150e199
|
Python
|
mit
|
magul/pywikibot-core,hasteur/g13bot_tools_new,PersianWikipedia/pywikibot-core,npdoty/pywikibot,npdoty/pywikibot,wikimedia/pywikibot-core,happy5214/pywikibot-core,happy5214/pywikibot-core,wikimedia/pywikibot-core,Darkdadaah/pywikibot-core,jayvdb/pywikibot-core,hasteur/g13bot_tools_new,jayvdb/pywikibot-core,hasteur/g13bot_tools_new,magul/pywikibot-core,Darkdadaah/pywikibot-core
|
# -*- coding: utf-8 -*-
"""Classes and functions for working with the Echo extension."""
from __future__ import absolute_import, unicode_literals
import pywikibot
class Notification(object):
"""A notification issued by the Echo extension."""
def __init__(self, site):
"""Construct an empty Notification object."""
self.site = site
@classmethod
def fromJSON(cls, site, data):
"""
Construct a Notification object from JSON data returned by the API.
@rtype: Notification
"""
notif = cls(site)
notif.id = data['id'] # TODO: use numeric id ?
notif.type = data['type']
notif.category = data['category']
notif.timestamp = pywikibot.Timestamp.fromtimestampformat(data['timestamp']['mw'])
# TODO: use 'namespace-key' + 'text' ?
notif.page = pywikibot.Page(site, data['title']['full'])
if 'agent' in data and 'name' in data['agent']:
notif.agent = pywikibot.User(site, data['agent']['name'])
else:
notif.agent = None
if 'read' in data:
notif.read = pywikibot.Timestamp.fromtimestampformat(data['read'])
else:
notif.read = False
notif.content = data.get('*', None)
return notif
def mark_as_read(self):
"""Mark the notification as read."""
return self.site.notifications_mark_read(list=self.id)
Fix notifications building from JSON
Sometimes (like in welcome messages), notifications don't have a
'title' property, so we shouldn't assume there is one.
Bug: T139015
Change-Id: I83e480d04e8e09aa9bcb5edef4f56b47d150e199
|
# -*- coding: utf-8 -*-
"""Classes and functions for working with the Echo extension."""
from __future__ import absolute_import, unicode_literals
import pywikibot
class Notification(object):
"""A notification issued by the Echo extension."""
def __init__(self, site):
"""Construct an empty Notification object."""
self.site = site
@classmethod
def fromJSON(cls, site, data):
"""
Construct a Notification object from JSON data returned by the API.
@rtype: Notification
"""
notif = cls(site)
notif.id = data['id'] # TODO: use numeric id ?
notif.type = data['type']
notif.category = data['category']
notif.timestamp = pywikibot.Timestamp.fromtimestampformat(data['timestamp']['mw'])
if 'title' in data and 'full' in data['title']:
notif.page = pywikibot.Page(site, data['title']['full'])
else:
notif.page = None
if 'agent' in data and 'name' in data['agent']:
notif.agent = pywikibot.User(site, data['agent']['name'])
else:
notif.agent = None
if 'read' in data:
notif.read = pywikibot.Timestamp.fromtimestampformat(data['read'])
else:
notif.read = False
notif.content = data.get('*', None)
return notif
def mark_as_read(self):
"""Mark the notification as read."""
return self.site.notifications_mark_read(list=self.id)
|
<commit_before># -*- coding: utf-8 -*-
"""Classes and functions for working with the Echo extension."""
from __future__ import absolute_import, unicode_literals
import pywikibot
class Notification(object):
"""A notification issued by the Echo extension."""
def __init__(self, site):
"""Construct an empty Notification object."""
self.site = site
@classmethod
def fromJSON(cls, site, data):
"""
Construct a Notification object from JSON data returned by the API.
@rtype: Notification
"""
notif = cls(site)
notif.id = data['id'] # TODO: use numeric id ?
notif.type = data['type']
notif.category = data['category']
notif.timestamp = pywikibot.Timestamp.fromtimestampformat(data['timestamp']['mw'])
# TODO: use 'namespace-key' + 'text' ?
notif.page = pywikibot.Page(site, data['title']['full'])
if 'agent' in data and 'name' in data['agent']:
notif.agent = pywikibot.User(site, data['agent']['name'])
else:
notif.agent = None
if 'read' in data:
notif.read = pywikibot.Timestamp.fromtimestampformat(data['read'])
else:
notif.read = False
notif.content = data.get('*', None)
return notif
def mark_as_read(self):
"""Mark the notification as read."""
return self.site.notifications_mark_read(list=self.id)
<commit_msg>Fix notifications building from JSON
Sometimes (like in welcome messages), notifications don't have a
'title' property, so we shouldn't assume there is one.
Bug: T139015
Change-Id: I83e480d04e8e09aa9bcb5edef4f56b47d150e199<commit_after>
|
# -*- coding: utf-8 -*-
"""Classes and functions for working with the Echo extension."""
from __future__ import absolute_import, unicode_literals
import pywikibot
class Notification(object):
"""A notification issued by the Echo extension."""
def __init__(self, site):
"""Construct an empty Notification object."""
self.site = site
@classmethod
def fromJSON(cls, site, data):
"""
Construct a Notification object from JSON data returned by the API.
@rtype: Notification
"""
notif = cls(site)
notif.id = data['id'] # TODO: use numeric id ?
notif.type = data['type']
notif.category = data['category']
notif.timestamp = pywikibot.Timestamp.fromtimestampformat(data['timestamp']['mw'])
if 'title' in data and 'full' in data['title']:
notif.page = pywikibot.Page(site, data['title']['full'])
else:
notif.page = None
if 'agent' in data and 'name' in data['agent']:
notif.agent = pywikibot.User(site, data['agent']['name'])
else:
notif.agent = None
if 'read' in data:
notif.read = pywikibot.Timestamp.fromtimestampformat(data['read'])
else:
notif.read = False
notif.content = data.get('*', None)
return notif
def mark_as_read(self):
"""Mark the notification as read."""
return self.site.notifications_mark_read(list=self.id)
|
# -*- coding: utf-8 -*-
"""Classes and functions for working with the Echo extension."""
from __future__ import absolute_import, unicode_literals
import pywikibot
class Notification(object):
"""A notification issued by the Echo extension."""
def __init__(self, site):
"""Construct an empty Notification object."""
self.site = site
@classmethod
def fromJSON(cls, site, data):
"""
Construct a Notification object from JSON data returned by the API.
@rtype: Notification
"""
notif = cls(site)
notif.id = data['id'] # TODO: use numeric id ?
notif.type = data['type']
notif.category = data['category']
notif.timestamp = pywikibot.Timestamp.fromtimestampformat(data['timestamp']['mw'])
# TODO: use 'namespace-key' + 'text' ?
notif.page = pywikibot.Page(site, data['title']['full'])
if 'agent' in data and 'name' in data['agent']:
notif.agent = pywikibot.User(site, data['agent']['name'])
else:
notif.agent = None
if 'read' in data:
notif.read = pywikibot.Timestamp.fromtimestampformat(data['read'])
else:
notif.read = False
notif.content = data.get('*', None)
return notif
def mark_as_read(self):
"""Mark the notification as read."""
return self.site.notifications_mark_read(list=self.id)
Fix notifications building from JSON
Sometimes (like in welcome messages), notifications don't have a
'title' property, so we shouldn't assume there is one.
Bug: T139015
Change-Id: I83e480d04e8e09aa9bcb5edef4f56b47d150e199# -*- coding: utf-8 -*-
"""Classes and functions for working with the Echo extension."""
from __future__ import absolute_import, unicode_literals
import pywikibot
class Notification(object):
"""A notification issued by the Echo extension."""
def __init__(self, site):
"""Construct an empty Notification object."""
self.site = site
@classmethod
def fromJSON(cls, site, data):
"""
Construct a Notification object from JSON data returned by the API.
@rtype: Notification
"""
notif = cls(site)
notif.id = data['id'] # TODO: use numeric id ?
notif.type = data['type']
notif.category = data['category']
notif.timestamp = pywikibot.Timestamp.fromtimestampformat(data['timestamp']['mw'])
if 'title' in data and 'full' in data['title']:
notif.page = pywikibot.Page(site, data['title']['full'])
else:
notif.page = None
if 'agent' in data and 'name' in data['agent']:
notif.agent = pywikibot.User(site, data['agent']['name'])
else:
notif.agent = None
if 'read' in data:
notif.read = pywikibot.Timestamp.fromtimestampformat(data['read'])
else:
notif.read = False
notif.content = data.get('*', None)
return notif
def mark_as_read(self):
"""Mark the notification as read."""
return self.site.notifications_mark_read(list=self.id)
|
<commit_before># -*- coding: utf-8 -*-
"""Classes and functions for working with the Echo extension."""
from __future__ import absolute_import, unicode_literals
import pywikibot
class Notification(object):
"""A notification issued by the Echo extension."""
def __init__(self, site):
"""Construct an empty Notification object."""
self.site = site
@classmethod
def fromJSON(cls, site, data):
"""
Construct a Notification object from JSON data returned by the API.
@rtype: Notification
"""
notif = cls(site)
notif.id = data['id'] # TODO: use numeric id ?
notif.type = data['type']
notif.category = data['category']
notif.timestamp = pywikibot.Timestamp.fromtimestampformat(data['timestamp']['mw'])
# TODO: use 'namespace-key' + 'text' ?
notif.page = pywikibot.Page(site, data['title']['full'])
if 'agent' in data and 'name' in data['agent']:
notif.agent = pywikibot.User(site, data['agent']['name'])
else:
notif.agent = None
if 'read' in data:
notif.read = pywikibot.Timestamp.fromtimestampformat(data['read'])
else:
notif.read = False
notif.content = data.get('*', None)
return notif
def mark_as_read(self):
"""Mark the notification as read."""
return self.site.notifications_mark_read(list=self.id)
<commit_msg>Fix notifications building from JSON
Sometimes (like in welcome messages), notifications don't have a
'title' property, so we shouldn't assume there is one.
Bug: T139015
Change-Id: I83e480d04e8e09aa9bcb5edef4f56b47d150e199<commit_after># -*- coding: utf-8 -*-
"""Classes and functions for working with the Echo extension."""
from __future__ import absolute_import, unicode_literals
import pywikibot
class Notification(object):
"""A notification issued by the Echo extension."""
def __init__(self, site):
"""Construct an empty Notification object."""
self.site = site
@classmethod
def fromJSON(cls, site, data):
"""
Construct a Notification object from JSON data returned by the API.
@rtype: Notification
"""
notif = cls(site)
notif.id = data['id'] # TODO: use numeric id ?
notif.type = data['type']
notif.category = data['category']
notif.timestamp = pywikibot.Timestamp.fromtimestampformat(data['timestamp']['mw'])
if 'title' in data and 'full' in data['title']:
notif.page = pywikibot.Page(site, data['title']['full'])
else:
notif.page = None
if 'agent' in data and 'name' in data['agent']:
notif.agent = pywikibot.User(site, data['agent']['name'])
else:
notif.agent = None
if 'read' in data:
notif.read = pywikibot.Timestamp.fromtimestampformat(data['read'])
else:
notif.read = False
notif.content = data.get('*', None)
return notif
def mark_as_read(self):
"""Mark the notification as read."""
return self.site.notifications_mark_read(list=self.id)
|
6664d075b4037ae40a91267afaca5731aa73ed3c
|
bluebottle/utils/widgets.py
|
bluebottle/utils/widgets.py
|
from __future__ import unicode_literals
from urlparse import urlparse
from django.contrib.admin.widgets import AdminURLFieldWidget
from django.forms.widgets import CheckboxFieldRenderer, CheckboxSelectMultiple, CheckboxChoiceInput
from django.utils.html import format_html
class NiceCheckboxChoiceInput(CheckboxChoiceInput):
def render(self, name=None, value=None, attrs=None):
if self.id_for_label:
label_for = format_html(' for="{}"', self.id_for_label)
else:
label_for = ''
attrs = dict(self.attrs, **attrs) if attrs else self.attrs
return format_html(
'{} <label{}>{}</label>', self.tag(attrs), label_for, self.choice_label
)
def is_checked(self):
return self.choice_value in self.value
class MultiCheckboxRenderer(CheckboxFieldRenderer):
choice_input_class = NiceCheckboxChoiceInput
class CheckboxSelectMultipleWidget(CheckboxSelectMultiple):
renderer = MultiCheckboxRenderer
class SecureAdminURLFieldWidget(AdminURLFieldWidget):
def render(self, name, value, attrs=None):
if urlparse(value).scheme not in ('http', 'https', ):
return super(AdminURLFieldWidget, self).render(name, value, attrs)
else:
return super(SecureAdminURLFieldWidget, self).render(name, value, attrs)
|
from __future__ import unicode_literals
from urlparse import urlparse
from django.contrib.admin.widgets import AdminURLFieldWidget
from django.forms.widgets import CheckboxFieldRenderer, CheckboxSelectMultiple, CheckboxChoiceInput
from django.utils.html import format_html
class NiceCheckboxChoiceInput(CheckboxChoiceInput):
def render(self, name=None, value=None, attrs=None):
if self.id_for_label:
label_for = format_html(' for="{}"', self.id_for_label)
else:
label_for = ''
attrs = dict(self.attrs, **attrs) if attrs else self.attrs
return format_html(
'{} <label{}>{}</label>', self.tag(attrs), label_for, self.choice_label
)
def is_checked(self):
return self.choice_value in self.value
class MultiCheckboxRenderer(CheckboxFieldRenderer):
choice_input_class = NiceCheckboxChoiceInput
class CheckboxSelectMultipleWidget(CheckboxSelectMultiple):
renderer = MultiCheckboxRenderer
class SecureAdminURLFieldWidget(AdminURLFieldWidget):
def render(self, name, value, attrs=None):
if value and urlparse(value).scheme not in ('http', 'https', ):
return super(AdminURLFieldWidget, self).render(name, value, attrs)
else:
return super(SecureAdminURLFieldWidget, self).render(name, value, attrs)
|
Fix url fields when no value is set
|
Fix url fields when no value is set
|
Python
|
bsd-3-clause
|
onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle
|
from __future__ import unicode_literals
from urlparse import urlparse
from django.contrib.admin.widgets import AdminURLFieldWidget
from django.forms.widgets import CheckboxFieldRenderer, CheckboxSelectMultiple, CheckboxChoiceInput
from django.utils.html import format_html
class NiceCheckboxChoiceInput(CheckboxChoiceInput):
def render(self, name=None, value=None, attrs=None):
if self.id_for_label:
label_for = format_html(' for="{}"', self.id_for_label)
else:
label_for = ''
attrs = dict(self.attrs, **attrs) if attrs else self.attrs
return format_html(
'{} <label{}>{}</label>', self.tag(attrs), label_for, self.choice_label
)
def is_checked(self):
return self.choice_value in self.value
class MultiCheckboxRenderer(CheckboxFieldRenderer):
choice_input_class = NiceCheckboxChoiceInput
class CheckboxSelectMultipleWidget(CheckboxSelectMultiple):
renderer = MultiCheckboxRenderer
class SecureAdminURLFieldWidget(AdminURLFieldWidget):
def render(self, name, value, attrs=None):
if urlparse(value).scheme not in ('http', 'https', ):
return super(AdminURLFieldWidget, self).render(name, value, attrs)
else:
return super(SecureAdminURLFieldWidget, self).render(name, value, attrs)
Fix url fields when no value is set
|
from __future__ import unicode_literals
from urlparse import urlparse
from django.contrib.admin.widgets import AdminURLFieldWidget
from django.forms.widgets import CheckboxFieldRenderer, CheckboxSelectMultiple, CheckboxChoiceInput
from django.utils.html import format_html
class NiceCheckboxChoiceInput(CheckboxChoiceInput):
def render(self, name=None, value=None, attrs=None):
if self.id_for_label:
label_for = format_html(' for="{}"', self.id_for_label)
else:
label_for = ''
attrs = dict(self.attrs, **attrs) if attrs else self.attrs
return format_html(
'{} <label{}>{}</label>', self.tag(attrs), label_for, self.choice_label
)
def is_checked(self):
return self.choice_value in self.value
class MultiCheckboxRenderer(CheckboxFieldRenderer):
choice_input_class = NiceCheckboxChoiceInput
class CheckboxSelectMultipleWidget(CheckboxSelectMultiple):
renderer = MultiCheckboxRenderer
class SecureAdminURLFieldWidget(AdminURLFieldWidget):
def render(self, name, value, attrs=None):
if value and urlparse(value).scheme not in ('http', 'https', ):
return super(AdminURLFieldWidget, self).render(name, value, attrs)
else:
return super(SecureAdminURLFieldWidget, self).render(name, value, attrs)
|
<commit_before>from __future__ import unicode_literals
from urlparse import urlparse
from django.contrib.admin.widgets import AdminURLFieldWidget
from django.forms.widgets import CheckboxFieldRenderer, CheckboxSelectMultiple, CheckboxChoiceInput
from django.utils.html import format_html
class NiceCheckboxChoiceInput(CheckboxChoiceInput):
def render(self, name=None, value=None, attrs=None):
if self.id_for_label:
label_for = format_html(' for="{}"', self.id_for_label)
else:
label_for = ''
attrs = dict(self.attrs, **attrs) if attrs else self.attrs
return format_html(
'{} <label{}>{}</label>', self.tag(attrs), label_for, self.choice_label
)
def is_checked(self):
return self.choice_value in self.value
class MultiCheckboxRenderer(CheckboxFieldRenderer):
choice_input_class = NiceCheckboxChoiceInput
class CheckboxSelectMultipleWidget(CheckboxSelectMultiple):
renderer = MultiCheckboxRenderer
class SecureAdminURLFieldWidget(AdminURLFieldWidget):
def render(self, name, value, attrs=None):
if urlparse(value).scheme not in ('http', 'https', ):
return super(AdminURLFieldWidget, self).render(name, value, attrs)
else:
return super(SecureAdminURLFieldWidget, self).render(name, value, attrs)
<commit_msg>Fix url fields when no value is set<commit_after>
|
from __future__ import unicode_literals
from urlparse import urlparse
from django.contrib.admin.widgets import AdminURLFieldWidget
from django.forms.widgets import CheckboxFieldRenderer, CheckboxSelectMultiple, CheckboxChoiceInput
from django.utils.html import format_html
class NiceCheckboxChoiceInput(CheckboxChoiceInput):
def render(self, name=None, value=None, attrs=None):
if self.id_for_label:
label_for = format_html(' for="{}"', self.id_for_label)
else:
label_for = ''
attrs = dict(self.attrs, **attrs) if attrs else self.attrs
return format_html(
'{} <label{}>{}</label>', self.tag(attrs), label_for, self.choice_label
)
def is_checked(self):
return self.choice_value in self.value
class MultiCheckboxRenderer(CheckboxFieldRenderer):
choice_input_class = NiceCheckboxChoiceInput
class CheckboxSelectMultipleWidget(CheckboxSelectMultiple):
renderer = MultiCheckboxRenderer
class SecureAdminURLFieldWidget(AdminURLFieldWidget):
def render(self, name, value, attrs=None):
if value and urlparse(value).scheme not in ('http', 'https', ):
return super(AdminURLFieldWidget, self).render(name, value, attrs)
else:
return super(SecureAdminURLFieldWidget, self).render(name, value, attrs)
|
from __future__ import unicode_literals
from urlparse import urlparse
from django.contrib.admin.widgets import AdminURLFieldWidget
from django.forms.widgets import CheckboxFieldRenderer, CheckboxSelectMultiple, CheckboxChoiceInput
from django.utils.html import format_html
class NiceCheckboxChoiceInput(CheckboxChoiceInput):
def render(self, name=None, value=None, attrs=None):
if self.id_for_label:
label_for = format_html(' for="{}"', self.id_for_label)
else:
label_for = ''
attrs = dict(self.attrs, **attrs) if attrs else self.attrs
return format_html(
'{} <label{}>{}</label>', self.tag(attrs), label_for, self.choice_label
)
def is_checked(self):
return self.choice_value in self.value
class MultiCheckboxRenderer(CheckboxFieldRenderer):
choice_input_class = NiceCheckboxChoiceInput
class CheckboxSelectMultipleWidget(CheckboxSelectMultiple):
renderer = MultiCheckboxRenderer
class SecureAdminURLFieldWidget(AdminURLFieldWidget):
def render(self, name, value, attrs=None):
if urlparse(value).scheme not in ('http', 'https', ):
return super(AdminURLFieldWidget, self).render(name, value, attrs)
else:
return super(SecureAdminURLFieldWidget, self).render(name, value, attrs)
Fix url fields when no value is setfrom __future__ import unicode_literals
from urlparse import urlparse
from django.contrib.admin.widgets import AdminURLFieldWidget
from django.forms.widgets import CheckboxFieldRenderer, CheckboxSelectMultiple, CheckboxChoiceInput
from django.utils.html import format_html
class NiceCheckboxChoiceInput(CheckboxChoiceInput):
def render(self, name=None, value=None, attrs=None):
if self.id_for_label:
label_for = format_html(' for="{}"', self.id_for_label)
else:
label_for = ''
attrs = dict(self.attrs, **attrs) if attrs else self.attrs
return format_html(
'{} <label{}>{}</label>', self.tag(attrs), label_for, self.choice_label
)
def is_checked(self):
return self.choice_value in self.value
class MultiCheckboxRenderer(CheckboxFieldRenderer):
choice_input_class = NiceCheckboxChoiceInput
class CheckboxSelectMultipleWidget(CheckboxSelectMultiple):
renderer = MultiCheckboxRenderer
class SecureAdminURLFieldWidget(AdminURLFieldWidget):
def render(self, name, value, attrs=None):
if value and urlparse(value).scheme not in ('http', 'https', ):
return super(AdminURLFieldWidget, self).render(name, value, attrs)
else:
return super(SecureAdminURLFieldWidget, self).render(name, value, attrs)
|
<commit_before>from __future__ import unicode_literals
from urlparse import urlparse
from django.contrib.admin.widgets import AdminURLFieldWidget
from django.forms.widgets import CheckboxFieldRenderer, CheckboxSelectMultiple, CheckboxChoiceInput
from django.utils.html import format_html
class NiceCheckboxChoiceInput(CheckboxChoiceInput):
def render(self, name=None, value=None, attrs=None):
if self.id_for_label:
label_for = format_html(' for="{}"', self.id_for_label)
else:
label_for = ''
attrs = dict(self.attrs, **attrs) if attrs else self.attrs
return format_html(
'{} <label{}>{}</label>', self.tag(attrs), label_for, self.choice_label
)
def is_checked(self):
return self.choice_value in self.value
class MultiCheckboxRenderer(CheckboxFieldRenderer):
choice_input_class = NiceCheckboxChoiceInput
class CheckboxSelectMultipleWidget(CheckboxSelectMultiple):
renderer = MultiCheckboxRenderer
class SecureAdminURLFieldWidget(AdminURLFieldWidget):
def render(self, name, value, attrs=None):
if urlparse(value).scheme not in ('http', 'https', ):
return super(AdminURLFieldWidget, self).render(name, value, attrs)
else:
return super(SecureAdminURLFieldWidget, self).render(name, value, attrs)
<commit_msg>Fix url fields when no value is set<commit_after>from __future__ import unicode_literals
from urlparse import urlparse
from django.contrib.admin.widgets import AdminURLFieldWidget
from django.forms.widgets import CheckboxFieldRenderer, CheckboxSelectMultiple, CheckboxChoiceInput
from django.utils.html import format_html
class NiceCheckboxChoiceInput(CheckboxChoiceInput):
def render(self, name=None, value=None, attrs=None):
if self.id_for_label:
label_for = format_html(' for="{}"', self.id_for_label)
else:
label_for = ''
attrs = dict(self.attrs, **attrs) if attrs else self.attrs
return format_html(
'{} <label{}>{}</label>', self.tag(attrs), label_for, self.choice_label
)
def is_checked(self):
return self.choice_value in self.value
class MultiCheckboxRenderer(CheckboxFieldRenderer):
choice_input_class = NiceCheckboxChoiceInput
class CheckboxSelectMultipleWidget(CheckboxSelectMultiple):
renderer = MultiCheckboxRenderer
class SecureAdminURLFieldWidget(AdminURLFieldWidget):
def render(self, name, value, attrs=None):
if value and urlparse(value).scheme not in ('http', 'https', ):
return super(AdminURLFieldWidget, self).render(name, value, attrs)
else:
return super(SecureAdminURLFieldWidget, self).render(name, value, attrs)
|
bfe45a24800817e7445fa12e7cd859679e6452c3
|
porchlightapi/views.py
|
porchlightapi/views.py
|
# -*- coding: utf-8 -*-
from django.shortcuts import render
# Create your views here.
import django_filters
from rest_framework import viewsets
from rest_framework import filters
from porchlightapi.models import Repository, ValueDataPoint
from porchlightapi.serializers import RepositorySerializer, ValueDataPointSerializer
class RepositoryFilter(django_filters.FilterSet):
"""
Provide filtering of repository objects based on name or project.
This is 'icontains' filtering, so a repo with the name "Porchlight"
will match 'por', 'Por', etc.
"""
name = django_filters.CharFilter(name="name", lookup_type='icontains')
project = django_filters.CharFilter(name="project", lookup_type='icontains')
class Meta:
model = Repository
fields = ['name', 'project',]
class RepositoryViewSet(viewsets.ReadOnlyModelViewSet):
"""
A REST view of repositories. Provides GET listing of Repositories,
filtering on name and project, and individual repsositories on id.
"""
queryset = Repository.objects.all()
serializer_class = RepositorySerializer
filter_backends = (filters.DjangoFilterBackend,)
filter_class = RepositoryFilter
# filter_fields = ('name', 'project')
class ValueDataPointViewSet(viewsets.ReadOnlyModelViewSet):
"""
A REST view of data points.
"""
queryset = ValueDataPoint.objects.all()
serializer_class = ValueDataPointSerializer
|
# -*- coding: utf-8 -*-
from django.shortcuts import render
# Create your views here.
from rest_framework import viewsets
from rest_framework import filters
from porchlightapi.models import Repository, ValueDataPoint
from porchlightapi.serializers import RepositorySerializer, ValueDataPointSerializer
class RepositoryViewSet(viewsets.ReadOnlyModelViewSet):
"""
A REST view of repositories. Provides GET listing of Repositories,
filtering on name and project, and individual repsositories on id.
"""
queryset = Repository.objects.all()
serializer_class = RepositorySerializer
filter_backends = (filters.SearchFilter,)
search_fields = ('name', 'project', 'url')
class ValueDataPointViewSet(viewsets.ReadOnlyModelViewSet):
"""
A REST view of data points.
"""
queryset = ValueDataPoint.objects.all()
serializer_class = ValueDataPointSerializer
|
Use DRF's built-in search filter
|
Use DRF's built-in search filter
|
Python
|
cc0-1.0
|
cfpb/porchlight,cfpb/porchlight,cfpb/porchlight
|
# -*- coding: utf-8 -*-
from django.shortcuts import render
# Create your views here.
import django_filters
from rest_framework import viewsets
from rest_framework import filters
from porchlightapi.models import Repository, ValueDataPoint
from porchlightapi.serializers import RepositorySerializer, ValueDataPointSerializer
class RepositoryFilter(django_filters.FilterSet):
"""
Provide filtering of repository objects based on name or project.
This is 'icontains' filtering, so a repo with the name "Porchlight"
will match 'por', 'Por', etc.
"""
name = django_filters.CharFilter(name="name", lookup_type='icontains')
project = django_filters.CharFilter(name="project", lookup_type='icontains')
class Meta:
model = Repository
fields = ['name', 'project',]
class RepositoryViewSet(viewsets.ReadOnlyModelViewSet):
"""
A REST view of repositories. Provides GET listing of Repositories,
filtering on name and project, and individual repsositories on id.
"""
queryset = Repository.objects.all()
serializer_class = RepositorySerializer
filter_backends = (filters.DjangoFilterBackend,)
filter_class = RepositoryFilter
# filter_fields = ('name', 'project')
class ValueDataPointViewSet(viewsets.ReadOnlyModelViewSet):
"""
A REST view of data points.
"""
queryset = ValueDataPoint.objects.all()
serializer_class = ValueDataPointSerializer
Use DRF's built-in search filter
|
# -*- coding: utf-8 -*-
from django.shortcuts import render
# Create your views here.
from rest_framework import viewsets
from rest_framework import filters
from porchlightapi.models import Repository, ValueDataPoint
from porchlightapi.serializers import RepositorySerializer, ValueDataPointSerializer
class RepositoryViewSet(viewsets.ReadOnlyModelViewSet):
"""
A REST view of repositories. Provides GET listing of Repositories,
filtering on name and project, and individual repsositories on id.
"""
queryset = Repository.objects.all()
serializer_class = RepositorySerializer
filter_backends = (filters.SearchFilter,)
search_fields = ('name', 'project', 'url')
class ValueDataPointViewSet(viewsets.ReadOnlyModelViewSet):
"""
A REST view of data points.
"""
queryset = ValueDataPoint.objects.all()
serializer_class = ValueDataPointSerializer
|
<commit_before># -*- coding: utf-8 -*-
from django.shortcuts import render
# Create your views here.
import django_filters
from rest_framework import viewsets
from rest_framework import filters
from porchlightapi.models import Repository, ValueDataPoint
from porchlightapi.serializers import RepositorySerializer, ValueDataPointSerializer
class RepositoryFilter(django_filters.FilterSet):
"""
Provide filtering of repository objects based on name or project.
This is 'icontains' filtering, so a repo with the name "Porchlight"
will match 'por', 'Por', etc.
"""
name = django_filters.CharFilter(name="name", lookup_type='icontains')
project = django_filters.CharFilter(name="project", lookup_type='icontains')
class Meta:
model = Repository
fields = ['name', 'project',]
class RepositoryViewSet(viewsets.ReadOnlyModelViewSet):
"""
A REST view of repositories. Provides GET listing of Repositories,
filtering on name and project, and individual repsositories on id.
"""
queryset = Repository.objects.all()
serializer_class = RepositorySerializer
filter_backends = (filters.DjangoFilterBackend,)
filter_class = RepositoryFilter
# filter_fields = ('name', 'project')
class ValueDataPointViewSet(viewsets.ReadOnlyModelViewSet):
"""
A REST view of data points.
"""
queryset = ValueDataPoint.objects.all()
serializer_class = ValueDataPointSerializer
<commit_msg>Use DRF's built-in search filter<commit_after>
|
# -*- coding: utf-8 -*-
from django.shortcuts import render
# Create your views here.
from rest_framework import viewsets
from rest_framework import filters
from porchlightapi.models import Repository, ValueDataPoint
from porchlightapi.serializers import RepositorySerializer, ValueDataPointSerializer
class RepositoryViewSet(viewsets.ReadOnlyModelViewSet):
"""
A REST view of repositories. Provides GET listing of Repositories,
filtering on name and project, and individual repsositories on id.
"""
queryset = Repository.objects.all()
serializer_class = RepositorySerializer
filter_backends = (filters.SearchFilter,)
search_fields = ('name', 'project', 'url')
class ValueDataPointViewSet(viewsets.ReadOnlyModelViewSet):
"""
A REST view of data points.
"""
queryset = ValueDataPoint.objects.all()
serializer_class = ValueDataPointSerializer
|
# -*- coding: utf-8 -*-
from django.shortcuts import render
# Create your views here.
import django_filters
from rest_framework import viewsets
from rest_framework import filters
from porchlightapi.models import Repository, ValueDataPoint
from porchlightapi.serializers import RepositorySerializer, ValueDataPointSerializer
class RepositoryFilter(django_filters.FilterSet):
"""
Provide filtering of repository objects based on name or project.
This is 'icontains' filtering, so a repo with the name "Porchlight"
will match 'por', 'Por', etc.
"""
name = django_filters.CharFilter(name="name", lookup_type='icontains')
project = django_filters.CharFilter(name="project", lookup_type='icontains')
class Meta:
model = Repository
fields = ['name', 'project',]
class RepositoryViewSet(viewsets.ReadOnlyModelViewSet):
"""
A REST view of repositories. Provides GET listing of Repositories,
filtering on name and project, and individual repsositories on id.
"""
queryset = Repository.objects.all()
serializer_class = RepositorySerializer
filter_backends = (filters.DjangoFilterBackend,)
filter_class = RepositoryFilter
# filter_fields = ('name', 'project')
class ValueDataPointViewSet(viewsets.ReadOnlyModelViewSet):
"""
A REST view of data points.
"""
queryset = ValueDataPoint.objects.all()
serializer_class = ValueDataPointSerializer
Use DRF's built-in search filter# -*- coding: utf-8 -*-
from django.shortcuts import render
# Create your views here.
from rest_framework import viewsets
from rest_framework import filters
from porchlightapi.models import Repository, ValueDataPoint
from porchlightapi.serializers import RepositorySerializer, ValueDataPointSerializer
class RepositoryViewSet(viewsets.ReadOnlyModelViewSet):
"""
A REST view of repositories. Provides GET listing of Repositories,
filtering on name and project, and individual repsositories on id.
"""
queryset = Repository.objects.all()
serializer_class = RepositorySerializer
filter_backends = (filters.SearchFilter,)
search_fields = ('name', 'project', 'url')
class ValueDataPointViewSet(viewsets.ReadOnlyModelViewSet):
"""
A REST view of data points.
"""
queryset = ValueDataPoint.objects.all()
serializer_class = ValueDataPointSerializer
|
<commit_before># -*- coding: utf-8 -*-
from django.shortcuts import render
# Create your views here.
import django_filters
from rest_framework import viewsets
from rest_framework import filters
from porchlightapi.models import Repository, ValueDataPoint
from porchlightapi.serializers import RepositorySerializer, ValueDataPointSerializer
class RepositoryFilter(django_filters.FilterSet):
"""
Provide filtering of repository objects based on name or project.
This is 'icontains' filtering, so a repo with the name "Porchlight"
will match 'por', 'Por', etc.
"""
name = django_filters.CharFilter(name="name", lookup_type='icontains')
project = django_filters.CharFilter(name="project", lookup_type='icontains')
class Meta:
model = Repository
fields = ['name', 'project',]
class RepositoryViewSet(viewsets.ReadOnlyModelViewSet):
"""
A REST view of repositories. Provides GET listing of Repositories,
filtering on name and project, and individual repsositories on id.
"""
queryset = Repository.objects.all()
serializer_class = RepositorySerializer
filter_backends = (filters.DjangoFilterBackend,)
filter_class = RepositoryFilter
# filter_fields = ('name', 'project')
class ValueDataPointViewSet(viewsets.ReadOnlyModelViewSet):
"""
A REST view of data points.
"""
queryset = ValueDataPoint.objects.all()
serializer_class = ValueDataPointSerializer
<commit_msg>Use DRF's built-in search filter<commit_after># -*- coding: utf-8 -*-
from django.shortcuts import render
# Create your views here.
from rest_framework import viewsets
from rest_framework import filters
from porchlightapi.models import Repository, ValueDataPoint
from porchlightapi.serializers import RepositorySerializer, ValueDataPointSerializer
class RepositoryViewSet(viewsets.ReadOnlyModelViewSet):
"""
A REST view of repositories. Provides GET listing of Repositories,
filtering on name and project, and individual repsositories on id.
"""
queryset = Repository.objects.all()
serializer_class = RepositorySerializer
filter_backends = (filters.SearchFilter,)
search_fields = ('name', 'project', 'url')
class ValueDataPointViewSet(viewsets.ReadOnlyModelViewSet):
"""
A REST view of data points.
"""
queryset = ValueDataPoint.objects.all()
serializer_class = ValueDataPointSerializer
|
1ecbd06083ac65a9520bcf0f87c5f5f1b4a4e532
|
helloworld.py
|
helloworld.py
|
#This is my hello world program
str1='Hello'
str2='Tarun'
print str1 +' '+ str2
# this is my hello world program
print 'Hello World!'
#This is my Hello world program
str1='Hello'
str2='Akash'
print str1 + ' ' + str2 + '!'
#this is a comment
str1='Hello'
str2='Priyanka'
print str1+' '+str2
|
print "helloworld"
|
Add strings to print hello world
|
Add strings to print hello world
|
Python
|
apache-2.0
|
ctsit/J.O.B-Training-Repo-1
|
#This is my hello world program
str1='Hello'
str2='Tarun'
print str1 +' '+ str2
# this is my hello world program
print 'Hello World!'
#This is my Hello world program
str1='Hello'
str2='Akash'
print str1 + ' ' + str2 + '!'
#this is a comment
str1='Hello'
str2='Priyanka'
print str1+' '+str2Add strings to print hello world
|
print "helloworld"
|
<commit_before>
#This is my hello world program
str1='Hello'
str2='Tarun'
print str1 +' '+ str2
# this is my hello world program
print 'Hello World!'
#This is my Hello world program
str1='Hello'
str2='Akash'
print str1 + ' ' + str2 + '!'
#this is a comment
str1='Hello'
str2='Priyanka'
print str1+' '+str2<commit_msg>Add strings to print hello world<commit_after>
|
print "helloworld"
|
#This is my hello world program
str1='Hello'
str2='Tarun'
print str1 +' '+ str2
# this is my hello world program
print 'Hello World!'
#This is my Hello world program
str1='Hello'
str2='Akash'
print str1 + ' ' + str2 + '!'
#this is a comment
str1='Hello'
str2='Priyanka'
print str1+' '+str2Add strings to print hello world
print "helloworld"
|
<commit_before>
#This is my hello world program
str1='Hello'
str2='Tarun'
print str1 +' '+ str2
# this is my hello world program
print 'Hello World!'
#This is my Hello world program
str1='Hello'
str2='Akash'
print str1 + ' ' + str2 + '!'
#this is a comment
str1='Hello'
str2='Priyanka'
print str1+' '+str2<commit_msg>Add strings to print hello world<commit_after>
print "helloworld"
|
14c473b8bef44ee5b521ce365ad89249c7f6e39e
|
linter.py
|
linter.py
|
#
# linter.py
# Linter for SublimeLinter3, a code checking framework for Sublime Text 3
#
# Written by Jon Surrell
# Copyright (c) 2014 Jon Surrell
#
# License: MIT
#
"""This module exports the Gotype plugin class."""
from SublimeLinter.lint import Linter, util
class Gotype(Linter):
"""Provides an interface to gotype."""
syntax = 'go'
cmd = ('gotype', '-e')
regex = r'^.+:(?P<line>\d+):(?P<col>\d+):\s+(?P<message>.+)$'
tempfile_suffix = 'go'
error_stream = util.STREAM_STDERR
|
#
# linter.py
# Linter for SublimeLinter3, a code checking framework for Sublime Text 3
#
# Written by Jon Surrell
# Copyright (c) 2014 Jon Surrell
#
# License: MIT
#
"""This module exports the Gotype plugin class."""
from SublimeLinter.lint import Linter, util
class Gotype(Linter):
"""Provides an interface to gotype."""
syntax = 'go'
cmd = ('gotype', '-e')
regex = r'^.+:(?P<line>\d+):(?P<col>\d+):\s+(?P<message>.+)'
tempfile_suffix = 'go'
error_stream = util.STREAM_STDERR
|
Remove trailing $ from regex
|
Remove trailing $ from regex
|
Python
|
mit
|
sirreal/SublimeLinter-contrib-gotype
|
#
# linter.py
# Linter for SublimeLinter3, a code checking framework for Sublime Text 3
#
# Written by Jon Surrell
# Copyright (c) 2014 Jon Surrell
#
# License: MIT
#
"""This module exports the Gotype plugin class."""
from SublimeLinter.lint import Linter, util
class Gotype(Linter):
"""Provides an interface to gotype."""
syntax = 'go'
cmd = ('gotype', '-e')
regex = r'^.+:(?P<line>\d+):(?P<col>\d+):\s+(?P<message>.+)$'
tempfile_suffix = 'go'
error_stream = util.STREAM_STDERR
Remove trailing $ from regex
|
#
# linter.py
# Linter for SublimeLinter3, a code checking framework for Sublime Text 3
#
# Written by Jon Surrell
# Copyright (c) 2014 Jon Surrell
#
# License: MIT
#
"""This module exports the Gotype plugin class."""
from SublimeLinter.lint import Linter, util
class Gotype(Linter):
"""Provides an interface to gotype."""
syntax = 'go'
cmd = ('gotype', '-e')
regex = r'^.+:(?P<line>\d+):(?P<col>\d+):\s+(?P<message>.+)'
tempfile_suffix = 'go'
error_stream = util.STREAM_STDERR
|
<commit_before>#
# linter.py
# Linter for SublimeLinter3, a code checking framework for Sublime Text 3
#
# Written by Jon Surrell
# Copyright (c) 2014 Jon Surrell
#
# License: MIT
#
"""This module exports the Gotype plugin class."""
from SublimeLinter.lint import Linter, util
class Gotype(Linter):
"""Provides an interface to gotype."""
syntax = 'go'
cmd = ('gotype', '-e')
regex = r'^.+:(?P<line>\d+):(?P<col>\d+):\s+(?P<message>.+)$'
tempfile_suffix = 'go'
error_stream = util.STREAM_STDERR
<commit_msg>Remove trailing $ from regex<commit_after>
|
#
# linter.py
# Linter for SublimeLinter3, a code checking framework for Sublime Text 3
#
# Written by Jon Surrell
# Copyright (c) 2014 Jon Surrell
#
# License: MIT
#
"""This module exports the Gotype plugin class."""
from SublimeLinter.lint import Linter, util
class Gotype(Linter):
"""Provides an interface to gotype."""
syntax = 'go'
cmd = ('gotype', '-e')
regex = r'^.+:(?P<line>\d+):(?P<col>\d+):\s+(?P<message>.+)'
tempfile_suffix = 'go'
error_stream = util.STREAM_STDERR
|
#
# linter.py
# Linter for SublimeLinter3, a code checking framework for Sublime Text 3
#
# Written by Jon Surrell
# Copyright (c) 2014 Jon Surrell
#
# License: MIT
#
"""This module exports the Gotype plugin class."""
from SublimeLinter.lint import Linter, util
class Gotype(Linter):
"""Provides an interface to gotype."""
syntax = 'go'
cmd = ('gotype', '-e')
regex = r'^.+:(?P<line>\d+):(?P<col>\d+):\s+(?P<message>.+)$'
tempfile_suffix = 'go'
error_stream = util.STREAM_STDERR
Remove trailing $ from regex#
# linter.py
# Linter for SublimeLinter3, a code checking framework for Sublime Text 3
#
# Written by Jon Surrell
# Copyright (c) 2014 Jon Surrell
#
# License: MIT
#
"""This module exports the Gotype plugin class."""
from SublimeLinter.lint import Linter, util
class Gotype(Linter):
"""Provides an interface to gotype."""
syntax = 'go'
cmd = ('gotype', '-e')
regex = r'^.+:(?P<line>\d+):(?P<col>\d+):\s+(?P<message>.+)'
tempfile_suffix = 'go'
error_stream = util.STREAM_STDERR
|
<commit_before>#
# linter.py
# Linter for SublimeLinter3, a code checking framework for Sublime Text 3
#
# Written by Jon Surrell
# Copyright (c) 2014 Jon Surrell
#
# License: MIT
#
"""This module exports the Gotype plugin class."""
from SublimeLinter.lint import Linter, util
class Gotype(Linter):
"""Provides an interface to gotype."""
syntax = 'go'
cmd = ('gotype', '-e')
regex = r'^.+:(?P<line>\d+):(?P<col>\d+):\s+(?P<message>.+)$'
tempfile_suffix = 'go'
error_stream = util.STREAM_STDERR
<commit_msg>Remove trailing $ from regex<commit_after>#
# linter.py
# Linter for SublimeLinter3, a code checking framework for Sublime Text 3
#
# Written by Jon Surrell
# Copyright (c) 2014 Jon Surrell
#
# License: MIT
#
"""This module exports the Gotype plugin class."""
from SublimeLinter.lint import Linter, util
class Gotype(Linter):
"""Provides an interface to gotype."""
syntax = 'go'
cmd = ('gotype', '-e')
regex = r'^.+:(?P<line>\d+):(?P<col>\d+):\s+(?P<message>.+)'
tempfile_suffix = 'go'
error_stream = util.STREAM_STDERR
|
d1928f0b1c98093b977ae208613c2b7eeb9a3ce5
|
carepoint/tests/models/cph/test_address.py
|
carepoint/tests/models/cph/test_address.py
|
# -*- coding: utf-8 -*-
##############################################################################
#
# Author: Dave Lasley <dave@laslabs.com>
# Copyright: 2015 LasLabs, Inc [https://laslabs.com]
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
import os
import unittest
import mock
from carepoint.tests.db.db import DatabaseTest
from carepoint.models.cph.address import Address
class ModelCphAddressTest(DatabaseTest):
def test_primary_key(self, ):
print Address.__table__
if __name__ == '__main__':
unittest.main()
|
# -*- coding: utf-8 -*-
##############################################################################
#
# Author: Dave Lasley <dave@laslabs.com>
# Copyright: 2015 LasLabs, Inc [https://laslabs.com]
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
import os
import unittest
import mock
from sqlalchemy.schema import Table
from carepoint.tests.db.db import DatabaseTest
from carepoint.models.cph.address import Address
class ModelCphAddressTest(DatabaseTest):
def test_table_initialization(self, ):
self.assertIsInstance(Address.__table__, Table)
if __name__ == '__main__':
unittest.main()
|
Add instance assertion to table
|
Add instance assertion to table
|
Python
|
mit
|
laslabs/Python-Carepoint
|
# -*- coding: utf-8 -*-
##############################################################################
#
# Author: Dave Lasley <dave@laslabs.com>
# Copyright: 2015 LasLabs, Inc [https://laslabs.com]
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
import os
import unittest
import mock
from carepoint.tests.db.db import DatabaseTest
from carepoint.models.cph.address import Address
class ModelCphAddressTest(DatabaseTest):
def test_primary_key(self, ):
print Address.__table__
if __name__ == '__main__':
unittest.main()
Add instance assertion to table
|
# -*- coding: utf-8 -*-
##############################################################################
#
# Author: Dave Lasley <dave@laslabs.com>
# Copyright: 2015 LasLabs, Inc [https://laslabs.com]
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
import os
import unittest
import mock
from sqlalchemy.schema import Table
from carepoint.tests.db.db import DatabaseTest
from carepoint.models.cph.address import Address
class ModelCphAddressTest(DatabaseTest):
def test_table_initialization(self, ):
self.assertIsInstance(Address.__table__, Table)
if __name__ == '__main__':
unittest.main()
|
<commit_before># -*- coding: utf-8 -*-
##############################################################################
#
# Author: Dave Lasley <dave@laslabs.com>
# Copyright: 2015 LasLabs, Inc [https://laslabs.com]
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
import os
import unittest
import mock
from carepoint.tests.db.db import DatabaseTest
from carepoint.models.cph.address import Address
class ModelCphAddressTest(DatabaseTest):
def test_primary_key(self, ):
print Address.__table__
if __name__ == '__main__':
unittest.main()
<commit_msg>Add instance assertion to table<commit_after>
|
# -*- coding: utf-8 -*-
##############################################################################
#
# Author: Dave Lasley <dave@laslabs.com>
# Copyright: 2015 LasLabs, Inc [https://laslabs.com]
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
import os
import unittest
import mock
from sqlalchemy.schema import Table
from carepoint.tests.db.db import DatabaseTest
from carepoint.models.cph.address import Address
class ModelCphAddressTest(DatabaseTest):
def test_table_initialization(self, ):
self.assertIsInstance(Address.__table__, Table)
if __name__ == '__main__':
unittest.main()
|
# -*- coding: utf-8 -*-
##############################################################################
#
# Author: Dave Lasley <dave@laslabs.com>
# Copyright: 2015 LasLabs, Inc [https://laslabs.com]
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
import os
import unittest
import mock
from carepoint.tests.db.db import DatabaseTest
from carepoint.models.cph.address import Address
class ModelCphAddressTest(DatabaseTest):
def test_primary_key(self, ):
print Address.__table__
if __name__ == '__main__':
unittest.main()
Add instance assertion to table# -*- coding: utf-8 -*-
##############################################################################
#
# Author: Dave Lasley <dave@laslabs.com>
# Copyright: 2015 LasLabs, Inc [https://laslabs.com]
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
import os
import unittest
import mock
from sqlalchemy.schema import Table
from carepoint.tests.db.db import DatabaseTest
from carepoint.models.cph.address import Address
class ModelCphAddressTest(DatabaseTest):
def test_table_initialization(self, ):
self.assertIsInstance(Address.__table__, Table)
if __name__ == '__main__':
unittest.main()
|
<commit_before># -*- coding: utf-8 -*-
##############################################################################
#
# Author: Dave Lasley <dave@laslabs.com>
# Copyright: 2015 LasLabs, Inc [https://laslabs.com]
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
import os
import unittest
import mock
from carepoint.tests.db.db import DatabaseTest
from carepoint.models.cph.address import Address
class ModelCphAddressTest(DatabaseTest):
def test_primary_key(self, ):
print Address.__table__
if __name__ == '__main__':
unittest.main()
<commit_msg>Add instance assertion to table<commit_after># -*- coding: utf-8 -*-
##############################################################################
#
# Author: Dave Lasley <dave@laslabs.com>
# Copyright: 2015 LasLabs, Inc [https://laslabs.com]
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
#
##############################################################################
import os
import unittest
import mock
from sqlalchemy.schema import Table
from carepoint.tests.db.db import DatabaseTest
from carepoint.models.cph.address import Address
class ModelCphAddressTest(DatabaseTest):
def test_table_initialization(self, ):
self.assertIsInstance(Address.__table__, Table)
if __name__ == '__main__':
unittest.main()
|
60951f30d8b5e2a450c13aa2b146be14ceb53c4d
|
rolldembones.py
|
rolldembones.py
|
#!/usr/bin/python
import argparse
import dice
def main():
roller = dice.Roller(args)
for repeat in range(args.repeats):
roller.do_roll()
for result in roller:
if isinstance(result, list):
print(' '.join(map(str, result)))
else:
print(result)
if __name__ == '__main__':
parser = argparse.ArgumentParser(description="Roll some dice.")
parser.add_argument("-r, --repeat", dest="repeats", type=int, metavar="N", default=1, help="Repeat these rolls N times.")
parser.add_argument("-e, --explode", dest="explode", metavar="E", type=int, default=None, help="Any die whose roll matches or exceeds E is counted and rolled again. Set to 1 or lower to disable this behavior on special dice.")
parser.add_argument("dice", nargs='*', help="Dice to roll, given in pairs of the number of dice to roll, and the sides those dice have.")
args = parser.parse_args()
# some basic error checking
if len(args.dice)%2 != 0:
parser.error("Incorrect number of arguments: Rolls and faces must be paired")
main()
|
#!/usr/bin/python3
import argparse
import dice
def main():
roller = dice.Roller(args)
for repeat in range(args.repeats):
roller.do_roll()
for result in roller:
if isinstance(result, list):
print(' '.join(map(str, result)))
else:
print(result)
if __name__ == '__main__':
parser = argparse.ArgumentParser(description="Roll some dice.")
parser.add_argument("-r, --repeat", dest="repeats", type=int, metavar="N", default=1, help="Repeat these rolls N times.")
parser.add_argument("-e, --explode", dest="explode", metavar="E", type=int, default=None, help="Any die whose roll matches or exceeds E is counted and rolled again. Set to 1 or lower to disable this behavior on special dice.")
parser.add_argument("dice", nargs='*', help="Dice to roll, given in pairs of the number of dice to roll, and the sides those dice have.")
args = parser.parse_args()
# some basic error checking
if len(args.dice)%2 != 0:
parser.error("Incorrect number of arguments: Rolls and faces must be paired")
main()
|
Update shebang to request python 3
|
Update shebang to request python 3
|
Python
|
mit
|
aurule/rolldembones
|
#!/usr/bin/python
import argparse
import dice
def main():
roller = dice.Roller(args)
for repeat in range(args.repeats):
roller.do_roll()
for result in roller:
if isinstance(result, list):
print(' '.join(map(str, result)))
else:
print(result)
if __name__ == '__main__':
parser = argparse.ArgumentParser(description="Roll some dice.")
parser.add_argument("-r, --repeat", dest="repeats", type=int, metavar="N", default=1, help="Repeat these rolls N times.")
parser.add_argument("-e, --explode", dest="explode", metavar="E", type=int, default=None, help="Any die whose roll matches or exceeds E is counted and rolled again. Set to 1 or lower to disable this behavior on special dice.")
parser.add_argument("dice", nargs='*', help="Dice to roll, given in pairs of the number of dice to roll, and the sides those dice have.")
args = parser.parse_args()
# some basic error checking
if len(args.dice)%2 != 0:
parser.error("Incorrect number of arguments: Rolls and faces must be paired")
main()Update shebang to request python 3
|
#!/usr/bin/python3
import argparse
import dice
def main():
roller = dice.Roller(args)
for repeat in range(args.repeats):
roller.do_roll()
for result in roller:
if isinstance(result, list):
print(' '.join(map(str, result)))
else:
print(result)
if __name__ == '__main__':
parser = argparse.ArgumentParser(description="Roll some dice.")
parser.add_argument("-r, --repeat", dest="repeats", type=int, metavar="N", default=1, help="Repeat these rolls N times.")
parser.add_argument("-e, --explode", dest="explode", metavar="E", type=int, default=None, help="Any die whose roll matches or exceeds E is counted and rolled again. Set to 1 or lower to disable this behavior on special dice.")
parser.add_argument("dice", nargs='*', help="Dice to roll, given in pairs of the number of dice to roll, and the sides those dice have.")
args = parser.parse_args()
# some basic error checking
if len(args.dice)%2 != 0:
parser.error("Incorrect number of arguments: Rolls and faces must be paired")
main()
|
<commit_before>#!/usr/bin/python
import argparse
import dice
def main():
roller = dice.Roller(args)
for repeat in range(args.repeats):
roller.do_roll()
for result in roller:
if isinstance(result, list):
print(' '.join(map(str, result)))
else:
print(result)
if __name__ == '__main__':
parser = argparse.ArgumentParser(description="Roll some dice.")
parser.add_argument("-r, --repeat", dest="repeats", type=int, metavar="N", default=1, help="Repeat these rolls N times.")
parser.add_argument("-e, --explode", dest="explode", metavar="E", type=int, default=None, help="Any die whose roll matches or exceeds E is counted and rolled again. Set to 1 or lower to disable this behavior on special dice.")
parser.add_argument("dice", nargs='*', help="Dice to roll, given in pairs of the number of dice to roll, and the sides those dice have.")
args = parser.parse_args()
# some basic error checking
if len(args.dice)%2 != 0:
parser.error("Incorrect number of arguments: Rolls and faces must be paired")
main()<commit_msg>Update shebang to request python 3<commit_after>
|
#!/usr/bin/python3
import argparse
import dice
def main():
roller = dice.Roller(args)
for repeat in range(args.repeats):
roller.do_roll()
for result in roller:
if isinstance(result, list):
print(' '.join(map(str, result)))
else:
print(result)
if __name__ == '__main__':
parser = argparse.ArgumentParser(description="Roll some dice.")
parser.add_argument("-r, --repeat", dest="repeats", type=int, metavar="N", default=1, help="Repeat these rolls N times.")
parser.add_argument("-e, --explode", dest="explode", metavar="E", type=int, default=None, help="Any die whose roll matches or exceeds E is counted and rolled again. Set to 1 or lower to disable this behavior on special dice.")
parser.add_argument("dice", nargs='*', help="Dice to roll, given in pairs of the number of dice to roll, and the sides those dice have.")
args = parser.parse_args()
# some basic error checking
if len(args.dice)%2 != 0:
parser.error("Incorrect number of arguments: Rolls and faces must be paired")
main()
|
#!/usr/bin/python
import argparse
import dice
def main():
roller = dice.Roller(args)
for repeat in range(args.repeats):
roller.do_roll()
for result in roller:
if isinstance(result, list):
print(' '.join(map(str, result)))
else:
print(result)
if __name__ == '__main__':
parser = argparse.ArgumentParser(description="Roll some dice.")
parser.add_argument("-r, --repeat", dest="repeats", type=int, metavar="N", default=1, help="Repeat these rolls N times.")
parser.add_argument("-e, --explode", dest="explode", metavar="E", type=int, default=None, help="Any die whose roll matches or exceeds E is counted and rolled again. Set to 1 or lower to disable this behavior on special dice.")
parser.add_argument("dice", nargs='*', help="Dice to roll, given in pairs of the number of dice to roll, and the sides those dice have.")
args = parser.parse_args()
# some basic error checking
if len(args.dice)%2 != 0:
parser.error("Incorrect number of arguments: Rolls and faces must be paired")
main()Update shebang to request python 3#!/usr/bin/python3
import argparse
import dice
def main():
roller = dice.Roller(args)
for repeat in range(args.repeats):
roller.do_roll()
for result in roller:
if isinstance(result, list):
print(' '.join(map(str, result)))
else:
print(result)
if __name__ == '__main__':
parser = argparse.ArgumentParser(description="Roll some dice.")
parser.add_argument("-r, --repeat", dest="repeats", type=int, metavar="N", default=1, help="Repeat these rolls N times.")
parser.add_argument("-e, --explode", dest="explode", metavar="E", type=int, default=None, help="Any die whose roll matches or exceeds E is counted and rolled again. Set to 1 or lower to disable this behavior on special dice.")
parser.add_argument("dice", nargs='*', help="Dice to roll, given in pairs of the number of dice to roll, and the sides those dice have.")
args = parser.parse_args()
# some basic error checking
if len(args.dice)%2 != 0:
parser.error("Incorrect number of arguments: Rolls and faces must be paired")
main()
|
<commit_before>#!/usr/bin/python
import argparse
import dice
def main():
roller = dice.Roller(args)
for repeat in range(args.repeats):
roller.do_roll()
for result in roller:
if isinstance(result, list):
print(' '.join(map(str, result)))
else:
print(result)
if __name__ == '__main__':
parser = argparse.ArgumentParser(description="Roll some dice.")
parser.add_argument("-r, --repeat", dest="repeats", type=int, metavar="N", default=1, help="Repeat these rolls N times.")
parser.add_argument("-e, --explode", dest="explode", metavar="E", type=int, default=None, help="Any die whose roll matches or exceeds E is counted and rolled again. Set to 1 or lower to disable this behavior on special dice.")
parser.add_argument("dice", nargs='*', help="Dice to roll, given in pairs of the number of dice to roll, and the sides those dice have.")
args = parser.parse_args()
# some basic error checking
if len(args.dice)%2 != 0:
parser.error("Incorrect number of arguments: Rolls and faces must be paired")
main()<commit_msg>Update shebang to request python 3<commit_after>#!/usr/bin/python3
import argparse
import dice
def main():
roller = dice.Roller(args)
for repeat in range(args.repeats):
roller.do_roll()
for result in roller:
if isinstance(result, list):
print(' '.join(map(str, result)))
else:
print(result)
if __name__ == '__main__':
parser = argparse.ArgumentParser(description="Roll some dice.")
parser.add_argument("-r, --repeat", dest="repeats", type=int, metavar="N", default=1, help="Repeat these rolls N times.")
parser.add_argument("-e, --explode", dest="explode", metavar="E", type=int, default=None, help="Any die whose roll matches or exceeds E is counted and rolled again. Set to 1 or lower to disable this behavior on special dice.")
parser.add_argument("dice", nargs='*', help="Dice to roll, given in pairs of the number of dice to roll, and the sides those dice have.")
args = parser.parse_args()
# some basic error checking
if len(args.dice)%2 != 0:
parser.error("Incorrect number of arguments: Rolls and faces must be paired")
main()
|
0f67d19a2cc38d8781946e20f6cd17b5287848a4
|
common/djangoapps/track/backends/logger.py
|
common/djangoapps/track/backends/logger.py
|
"""Event tracker backend that saves events to a python logger."""
from __future__ import absolute_import
import logging
import json
from django.conf import settings
from track.backends import BaseBackend
from track.utils import DateTimeJSONEncoder
log = logging.getLogger('track.backends.logger')
class LoggerBackend(BaseBackend):
"""Event tracker backend that uses a python logger.
Events are logged to the INFO level as JSON strings.
"""
def __init__(self, name, **kwargs):
"""Event tracker backend that uses a python logger.
:Parameters:
- `name`: identifier of the logger, which should have
been configured using the default python mechanisms.
"""
super(LoggerBackend, self).__init__(**kwargs)
self.event_logger = logging.getLogger(name)
def send(self, event):
event_str = json.dumps(event, cls=DateTimeJSONEncoder)
# TODO: remove trucation of the serialized event, either at a
# higher level during the emittion of the event, or by
# providing warnings when the events exceed certain size.
event_str = event_str[:settings.TRACK_MAX_EVENT]
self.event_logger.info(event_str)
|
"""Event tracker backend that saves events to a python logger."""
from __future__ import absolute_import
import logging
import json
from django.conf import settings
from track.backends import BaseBackend
from track.utils import DateTimeJSONEncoder
log = logging.getLogger('track.backends.logger')
application_log = logging.getLogger('track.backends.application_log') # pylint: disable=invalid-name
class LoggerBackend(BaseBackend):
"""Event tracker backend that uses a python logger.
Events are logged to the INFO level as JSON strings.
"""
def __init__(self, name, **kwargs):
"""Event tracker backend that uses a python logger.
:Parameters:
- `name`: identifier of the logger, which should have
been configured using the default python mechanisms.
"""
super(LoggerBackend, self).__init__(**kwargs)
self.event_logger = logging.getLogger(name)
def send(self, event):
try:
event_str = json.dumps(event, cls=DateTimeJSONEncoder)
except UnicodeDecodeError:
application_log.exception(
"UnicodeDecodeError Event_type: %r, Event_source: %r, Page: %r, Referer: %r",
event.get('event_type'), event.get('event_source'), event.get('page'), event.get('referer')
)
# TODO: remove trucation of the serialized event, either at a
# higher level during the emittion of the event, or by
# providing warnings when the events exceed certain size.
event_str = event_str[:settings.TRACK_MAX_EVENT]
self.event_logger.info(event_str)
|
Add logging for UnicodeDecodeError excpetion LoggerBackend
|
Add logging for UnicodeDecodeError excpetion LoggerBackend
|
Python
|
agpl-3.0
|
cognitiveclass/edx-platform,jolyonb/edx-platform,miptliot/edx-platform,jjmiranda/edx-platform,ZLLab-Mooc/edx-platform,procangroup/edx-platform,edx-solutions/edx-platform,jjmiranda/edx-platform,synergeticsedx/deployment-wipro,edx/edx-platform,UOMx/edx-platform,defance/edx-platform,EDUlib/edx-platform,raccoongang/edx-platform,antoviaque/edx-platform,alu042/edx-platform,marcore/edx-platform,simbs/edx-platform,jzoldak/edx-platform,philanthropy-u/edx-platform,waheedahmed/edx-platform,appsembler/edx-platform,shurihell/testasia,edx/edx-platform,kmoocdev2/edx-platform,10clouds/edx-platform,pabloborrego93/edx-platform,lduarte1991/edx-platform,ZLLab-Mooc/edx-platform,longmen21/edx-platform,Livit/Livit.Learn.EdX,caesar2164/edx-platform,fintech-circle/edx-platform,tanmaykm/edx-platform,raccoongang/edx-platform,defance/edx-platform,wwj718/edx-platform,jolyonb/edx-platform,RPI-OPENEDX/edx-platform,cpennington/edx-platform,Ayub-Khan/edx-platform,Stanford-Online/edx-platform,Edraak/circleci-edx-platform,chrisndodge/edx-platform,msegado/edx-platform,raccoongang/edx-platform,halvertoluke/edx-platform,Stanford-Online/edx-platform,procangroup/edx-platform,prarthitm/edxplatform,IndonesiaX/edx-platform,cpennington/edx-platform,devs1991/test_edx_docmode,a-parhom/edx-platform,10clouds/edx-platform,deepsrijit1105/edx-platform,devs1991/test_edx_docmode,eduNEXT/edunext-platform,ESOedX/edx-platform,JioEducation/edx-platform,ampax/edx-platform,prarthitm/edxplatform,solashirai/edx-platform,nttks/edx-platform,ampax/edx-platform,EDUlib/edx-platform,lduarte1991/edx-platform,franosincic/edx-platform,pomegranited/edx-platform,Endika/edx-platform,Edraak/edx-platform,waheedahmed/edx-platform,jjmiranda/edx-platform,nttks/edx-platform,TeachAtTUM/edx-platform,Livit/Livit.Learn.EdX,ahmedaljazzar/edx-platform,doganov/edx-platform,edx/edx-platform,deepsrijit1105/edx-platform,RPI-OPENEDX/edx-platform,10clouds/edx-platform,zhenzhai/edx-platform,Edraak/edx-platform,zhenzhai/edx-platform,wwj718/edx-platform,JioEducation/edx-platform,amir-qayyum-khan/edx-platform,Lektorium-LLC/edx-platform,cognitiveclass/edx-platform,kursitet/edx-platform,pomegranited/edx-platform,MakeHer/edx-platform,doganov/edx-platform,mbareta/edx-platform-ft,bigdatauniversity/edx-platform,BehavioralInsightsTeam/edx-platform,ESOedX/edx-platform,cecep-edu/edx-platform,simbs/edx-platform,IndonesiaX/edx-platform,CredoReference/edx-platform,raccoongang/edx-platform,bigdatauniversity/edx-platform,romain-li/edx-platform,appsembler/edx-platform,simbs/edx-platform,devs1991/test_edx_docmode,kursitet/edx-platform,eduNEXT/edunext-platform,analyseuc3m/ANALYSE-v1,BehavioralInsightsTeam/edx-platform,shurihell/testasia,jzoldak/edx-platform,mitocw/edx-platform,ovnicraft/edx-platform,defance/edx-platform,eduNEXT/edx-platform,proversity-org/edx-platform,marcore/edx-platform,stvstnfrd/edx-platform,jjmiranda/edx-platform,Edraak/circleci-edx-platform,edx-solutions/edx-platform,IndonesiaX/edx-platform,MakeHer/edx-platform,halvertoluke/edx-platform,Edraak/edraak-platform,itsjeyd/edx-platform,edx/edx-platform,shabab12/edx-platform,cognitiveclass/edx-platform,waheedahmed/edx-platform,mbareta/edx-platform-ft,msegado/edx-platform,appsembler/edx-platform,proversity-org/edx-platform,angelapper/edx-platform,marcore/edx-platform,zhenzhai/edx-platform,stvstnfrd/edx-platform,gymnasium/edx-platform,shurihell/testasia,halvertoluke/edx-platform,fintech-circle/edx-platform,Ayub-Khan/edx-platform,gymnasium/edx-platform,pabloborrego93/edx-platform,hastexo/edx-platform,romain-li/edx-platform,caesar2164/edx-platform,pepeportela/edx-platform,hastexo/edx-platform,wwj718/edx-platform,Ayub-Khan/edx-platform,naresh21/synergetics-edx-platform,hastexo/edx-platform,franosincic/edx-platform,pepeportela/edx-platform,marcore/edx-platform,synergeticsedx/deployment-wipro,IndonesiaX/edx-platform,kmoocdev2/edx-platform,romain-li/edx-platform,jzoldak/edx-platform,naresh21/synergetics-edx-platform,cecep-edu/edx-platform,franosincic/edx-platform,Stanford-Online/edx-platform,analyseuc3m/ANALYSE-v1,caesar2164/edx-platform,angelapper/edx-platform,msegado/edx-platform,msegado/edx-platform,waheedahmed/edx-platform,MakeHer/edx-platform,ahmedaljazzar/edx-platform,itsjeyd/edx-platform,kmoocdev2/edx-platform,synergeticsedx/deployment-wipro,antoviaque/edx-platform,proversity-org/edx-platform,nttks/edx-platform,kursitet/edx-platform,antoviaque/edx-platform,RPI-OPENEDX/edx-platform,kursitet/edx-platform,Livit/Livit.Learn.EdX,procangroup/edx-platform,shurihell/testasia,edx-solutions/edx-platform,kmoocdev2/edx-platform,teltek/edx-platform,shabab12/edx-platform,shabab12/edx-platform,shurihell/testasia,Edraak/circleci-edx-platform,BehavioralInsightsTeam/edx-platform,halvertoluke/edx-platform,analyseuc3m/ANALYSE-v1,cecep-edu/edx-platform,CourseTalk/edx-platform,deepsrijit1105/edx-platform,pepeportela/edx-platform,eduNEXT/edunext-platform,pabloborrego93/edx-platform,appsembler/edx-platform,UOMx/edx-platform,hastexo/edx-platform,Endika/edx-platform,JioEducation/edx-platform,chrisndodge/edx-platform,ESOedX/edx-platform,JioEducation/edx-platform,synergeticsedx/deployment-wipro,chrisndodge/edx-platform,halvertoluke/edx-platform,romain-li/edx-platform,philanthropy-u/edx-platform,defance/edx-platform,CredoReference/edx-platform,EDUlib/edx-platform,Lektorium-LLC/edx-platform,cognitiveclass/edx-platform,philanthropy-u/edx-platform,naresh21/synergetics-edx-platform,MakeHer/edx-platform,teltek/edx-platform,angelapper/edx-platform,jbzdak/edx-platform,jzoldak/edx-platform,alu042/edx-platform,bigdatauniversity/edx-platform,solashirai/edx-platform,RPI-OPENEDX/edx-platform,alu042/edx-platform,zhenzhai/edx-platform,UOMx/edx-platform,fintech-circle/edx-platform,pepeportela/edx-platform,angelapper/edx-platform,longmen21/edx-platform,solashirai/edx-platform,doganov/edx-platform,TeachAtTUM/edx-platform,chrisndodge/edx-platform,simbs/edx-platform,simbs/edx-platform,pomegranited/edx-platform,Edraak/edx-platform,procangroup/edx-platform,itsjeyd/edx-platform,antoviaque/edx-platform,tanmaykm/edx-platform,UOMx/edx-platform,eduNEXT/edx-platform,BehavioralInsightsTeam/edx-platform,ZLLab-Mooc/edx-platform,louyihua/edx-platform,eduNEXT/edunext-platform,Stanford-Online/edx-platform,CredoReference/edx-platform,jolyonb/edx-platform,Edraak/edraak-platform,10clouds/edx-platform,Edraak/circleci-edx-platform,devs1991/test_edx_docmode,zhenzhai/edx-platform,Ayub-Khan/edx-platform,devs1991/test_edx_docmode,pomegranited/edx-platform,ampax/edx-platform,arbrandes/edx-platform,shabab12/edx-platform,proversity-org/edx-platform,kursitet/edx-platform,ZLLab-Mooc/edx-platform,mitocw/edx-platform,miptliot/edx-platform,stvstnfrd/edx-platform,cpennington/edx-platform,louyihua/edx-platform,louyihua/edx-platform,Endika/edx-platform,fintech-circle/edx-platform,analyseuc3m/ANALYSE-v1,solashirai/edx-platform,ovnicraft/edx-platform,waheedahmed/edx-platform,lduarte1991/edx-platform,msegado/edx-platform,Lektorium-LLC/edx-platform,RPI-OPENEDX/edx-platform,jbzdak/edx-platform,Edraak/edraak-platform,pomegranited/edx-platform,tanmaykm/edx-platform,eduNEXT/edx-platform,miptliot/edx-platform,doganov/edx-platform,amir-qayyum-khan/edx-platform,stvstnfrd/edx-platform,philanthropy-u/edx-platform,arbrandes/edx-platform,tanmaykm/edx-platform,nttks/edx-platform,EDUlib/edx-platform,amir-qayyum-khan/edx-platform,ovnicraft/edx-platform,cecep-edu/edx-platform,wwj718/edx-platform,ovnicraft/edx-platform,longmen21/edx-platform,longmen21/edx-platform,TeachAtTUM/edx-platform,deepsrijit1105/edx-platform,prarthitm/edxplatform,pabloborrego93/edx-platform,CourseTalk/edx-platform,naresh21/synergetics-edx-platform,gymnasium/edx-platform,gsehub/edx-platform,a-parhom/edx-platform,CourseTalk/edx-platform,louyihua/edx-platform,jbzdak/edx-platform,mitocw/edx-platform,CourseTalk/edx-platform,ESOedX/edx-platform,arbrandes/edx-platform,eduNEXT/edx-platform,devs1991/test_edx_docmode,alu042/edx-platform,jbzdak/edx-platform,ZLLab-Mooc/edx-platform,Endika/edx-platform,prarthitm/edxplatform,teltek/edx-platform,Edraak/edx-platform,doganov/edx-platform,Edraak/edx-platform,a-parhom/edx-platform,TeachAtTUM/edx-platform,gsehub/edx-platform,jolyonb/edx-platform,kmoocdev2/edx-platform,Edraak/circleci-edx-platform,gymnasium/edx-platform,amir-qayyum-khan/edx-platform,edx-solutions/edx-platform,teltek/edx-platform,caesar2164/edx-platform,mbareta/edx-platform-ft,Livit/Livit.Learn.EdX,ampax/edx-platform,Edraak/edraak-platform,longmen21/edx-platform,lduarte1991/edx-platform,nttks/edx-platform,bigdatauniversity/edx-platform,cognitiveclass/edx-platform,devs1991/test_edx_docmode,ahmedaljazzar/edx-platform,mbareta/edx-platform-ft,Lektorium-LLC/edx-platform,miptliot/edx-platform,gsehub/edx-platform,CredoReference/edx-platform,devs1991/test_edx_docmode,ahmedaljazzar/edx-platform,ovnicraft/edx-platform,wwj718/edx-platform,gsehub/edx-platform,cpennington/edx-platform,Ayub-Khan/edx-platform,franosincic/edx-platform,solashirai/edx-platform,jbzdak/edx-platform,a-parhom/edx-platform,IndonesiaX/edx-platform,MakeHer/edx-platform,itsjeyd/edx-platform,franosincic/edx-platform,mitocw/edx-platform,romain-li/edx-platform,arbrandes/edx-platform,cecep-edu/edx-platform,bigdatauniversity/edx-platform
|
"""Event tracker backend that saves events to a python logger."""
from __future__ import absolute_import
import logging
import json
from django.conf import settings
from track.backends import BaseBackend
from track.utils import DateTimeJSONEncoder
log = logging.getLogger('track.backends.logger')
class LoggerBackend(BaseBackend):
"""Event tracker backend that uses a python logger.
Events are logged to the INFO level as JSON strings.
"""
def __init__(self, name, **kwargs):
"""Event tracker backend that uses a python logger.
:Parameters:
- `name`: identifier of the logger, which should have
been configured using the default python mechanisms.
"""
super(LoggerBackend, self).__init__(**kwargs)
self.event_logger = logging.getLogger(name)
def send(self, event):
event_str = json.dumps(event, cls=DateTimeJSONEncoder)
# TODO: remove trucation of the serialized event, either at a
# higher level during the emittion of the event, or by
# providing warnings when the events exceed certain size.
event_str = event_str[:settings.TRACK_MAX_EVENT]
self.event_logger.info(event_str)
Add logging for UnicodeDecodeError excpetion LoggerBackend
|
"""Event tracker backend that saves events to a python logger."""
from __future__ import absolute_import
import logging
import json
from django.conf import settings
from track.backends import BaseBackend
from track.utils import DateTimeJSONEncoder
log = logging.getLogger('track.backends.logger')
application_log = logging.getLogger('track.backends.application_log') # pylint: disable=invalid-name
class LoggerBackend(BaseBackend):
"""Event tracker backend that uses a python logger.
Events are logged to the INFO level as JSON strings.
"""
def __init__(self, name, **kwargs):
"""Event tracker backend that uses a python logger.
:Parameters:
- `name`: identifier of the logger, which should have
been configured using the default python mechanisms.
"""
super(LoggerBackend, self).__init__(**kwargs)
self.event_logger = logging.getLogger(name)
def send(self, event):
try:
event_str = json.dumps(event, cls=DateTimeJSONEncoder)
except UnicodeDecodeError:
application_log.exception(
"UnicodeDecodeError Event_type: %r, Event_source: %r, Page: %r, Referer: %r",
event.get('event_type'), event.get('event_source'), event.get('page'), event.get('referer')
)
# TODO: remove trucation of the serialized event, either at a
# higher level during the emittion of the event, or by
# providing warnings when the events exceed certain size.
event_str = event_str[:settings.TRACK_MAX_EVENT]
self.event_logger.info(event_str)
|
<commit_before>"""Event tracker backend that saves events to a python logger."""
from __future__ import absolute_import
import logging
import json
from django.conf import settings
from track.backends import BaseBackend
from track.utils import DateTimeJSONEncoder
log = logging.getLogger('track.backends.logger')
class LoggerBackend(BaseBackend):
"""Event tracker backend that uses a python logger.
Events are logged to the INFO level as JSON strings.
"""
def __init__(self, name, **kwargs):
"""Event tracker backend that uses a python logger.
:Parameters:
- `name`: identifier of the logger, which should have
been configured using the default python mechanisms.
"""
super(LoggerBackend, self).__init__(**kwargs)
self.event_logger = logging.getLogger(name)
def send(self, event):
event_str = json.dumps(event, cls=DateTimeJSONEncoder)
# TODO: remove trucation of the serialized event, either at a
# higher level during the emittion of the event, or by
# providing warnings when the events exceed certain size.
event_str = event_str[:settings.TRACK_MAX_EVENT]
self.event_logger.info(event_str)
<commit_msg>Add logging for UnicodeDecodeError excpetion LoggerBackend<commit_after>
|
"""Event tracker backend that saves events to a python logger."""
from __future__ import absolute_import
import logging
import json
from django.conf import settings
from track.backends import BaseBackend
from track.utils import DateTimeJSONEncoder
log = logging.getLogger('track.backends.logger')
application_log = logging.getLogger('track.backends.application_log') # pylint: disable=invalid-name
class LoggerBackend(BaseBackend):
"""Event tracker backend that uses a python logger.
Events are logged to the INFO level as JSON strings.
"""
def __init__(self, name, **kwargs):
"""Event tracker backend that uses a python logger.
:Parameters:
- `name`: identifier of the logger, which should have
been configured using the default python mechanisms.
"""
super(LoggerBackend, self).__init__(**kwargs)
self.event_logger = logging.getLogger(name)
def send(self, event):
try:
event_str = json.dumps(event, cls=DateTimeJSONEncoder)
except UnicodeDecodeError:
application_log.exception(
"UnicodeDecodeError Event_type: %r, Event_source: %r, Page: %r, Referer: %r",
event.get('event_type'), event.get('event_source'), event.get('page'), event.get('referer')
)
# TODO: remove trucation of the serialized event, either at a
# higher level during the emittion of the event, or by
# providing warnings when the events exceed certain size.
event_str = event_str[:settings.TRACK_MAX_EVENT]
self.event_logger.info(event_str)
|
"""Event tracker backend that saves events to a python logger."""
from __future__ import absolute_import
import logging
import json
from django.conf import settings
from track.backends import BaseBackend
from track.utils import DateTimeJSONEncoder
log = logging.getLogger('track.backends.logger')
class LoggerBackend(BaseBackend):
"""Event tracker backend that uses a python logger.
Events are logged to the INFO level as JSON strings.
"""
def __init__(self, name, **kwargs):
"""Event tracker backend that uses a python logger.
:Parameters:
- `name`: identifier of the logger, which should have
been configured using the default python mechanisms.
"""
super(LoggerBackend, self).__init__(**kwargs)
self.event_logger = logging.getLogger(name)
def send(self, event):
event_str = json.dumps(event, cls=DateTimeJSONEncoder)
# TODO: remove trucation of the serialized event, either at a
# higher level during the emittion of the event, or by
# providing warnings when the events exceed certain size.
event_str = event_str[:settings.TRACK_MAX_EVENT]
self.event_logger.info(event_str)
Add logging for UnicodeDecodeError excpetion LoggerBackend"""Event tracker backend that saves events to a python logger."""
from __future__ import absolute_import
import logging
import json
from django.conf import settings
from track.backends import BaseBackend
from track.utils import DateTimeJSONEncoder
log = logging.getLogger('track.backends.logger')
application_log = logging.getLogger('track.backends.application_log') # pylint: disable=invalid-name
class LoggerBackend(BaseBackend):
"""Event tracker backend that uses a python logger.
Events are logged to the INFO level as JSON strings.
"""
def __init__(self, name, **kwargs):
"""Event tracker backend that uses a python logger.
:Parameters:
- `name`: identifier of the logger, which should have
been configured using the default python mechanisms.
"""
super(LoggerBackend, self).__init__(**kwargs)
self.event_logger = logging.getLogger(name)
def send(self, event):
try:
event_str = json.dumps(event, cls=DateTimeJSONEncoder)
except UnicodeDecodeError:
application_log.exception(
"UnicodeDecodeError Event_type: %r, Event_source: %r, Page: %r, Referer: %r",
event.get('event_type'), event.get('event_source'), event.get('page'), event.get('referer')
)
# TODO: remove trucation of the serialized event, either at a
# higher level during the emittion of the event, or by
# providing warnings when the events exceed certain size.
event_str = event_str[:settings.TRACK_MAX_EVENT]
self.event_logger.info(event_str)
|
<commit_before>"""Event tracker backend that saves events to a python logger."""
from __future__ import absolute_import
import logging
import json
from django.conf import settings
from track.backends import BaseBackend
from track.utils import DateTimeJSONEncoder
log = logging.getLogger('track.backends.logger')
class LoggerBackend(BaseBackend):
"""Event tracker backend that uses a python logger.
Events are logged to the INFO level as JSON strings.
"""
def __init__(self, name, **kwargs):
"""Event tracker backend that uses a python logger.
:Parameters:
- `name`: identifier of the logger, which should have
been configured using the default python mechanisms.
"""
super(LoggerBackend, self).__init__(**kwargs)
self.event_logger = logging.getLogger(name)
def send(self, event):
event_str = json.dumps(event, cls=DateTimeJSONEncoder)
# TODO: remove trucation of the serialized event, either at a
# higher level during the emittion of the event, or by
# providing warnings when the events exceed certain size.
event_str = event_str[:settings.TRACK_MAX_EVENT]
self.event_logger.info(event_str)
<commit_msg>Add logging for UnicodeDecodeError excpetion LoggerBackend<commit_after>"""Event tracker backend that saves events to a python logger."""
from __future__ import absolute_import
import logging
import json
from django.conf import settings
from track.backends import BaseBackend
from track.utils import DateTimeJSONEncoder
log = logging.getLogger('track.backends.logger')
application_log = logging.getLogger('track.backends.application_log') # pylint: disable=invalid-name
class LoggerBackend(BaseBackend):
"""Event tracker backend that uses a python logger.
Events are logged to the INFO level as JSON strings.
"""
def __init__(self, name, **kwargs):
"""Event tracker backend that uses a python logger.
:Parameters:
- `name`: identifier of the logger, which should have
been configured using the default python mechanisms.
"""
super(LoggerBackend, self).__init__(**kwargs)
self.event_logger = logging.getLogger(name)
def send(self, event):
try:
event_str = json.dumps(event, cls=DateTimeJSONEncoder)
except UnicodeDecodeError:
application_log.exception(
"UnicodeDecodeError Event_type: %r, Event_source: %r, Page: %r, Referer: %r",
event.get('event_type'), event.get('event_source'), event.get('page'), event.get('referer')
)
# TODO: remove trucation of the serialized event, either at a
# higher level during the emittion of the event, or by
# providing warnings when the events exceed certain size.
event_str = event_str[:settings.TRACK_MAX_EVENT]
self.event_logger.info(event_str)
|
8fc274021a8c0813f3fc3568d1d7984112952b9c
|
pytilemap/qtsupport.py
|
pytilemap/qtsupport.py
|
import sys
import sip
import qtpy
__all__ = [
'getQVariantValue',
'wheelAngleDelta',
]
try:
if qtpy.PYQT5:
QVARIANT_API = 2
else:
QVARIANT_API = sip.getapi('QVariant')
except ValueError:
QVARIANT_API = 1
if QVARIANT_API == 1:
def getQVariantValue(variant):
return variant.toPyObject()
else:
def getQVariantValue(variant):
return variant
if qtpy.PYQT5:
def wheelAngleDelta(wheelEvent):
return wheelEvent.angleDelta().y()
else:
def wheelAngleDelta(wheelEvent):
return wheelEvent.delta()
if qtpy.PYQT5:
from qtpy.QtCore import QStandardPaths
def getTemporaryFolder():
return QStandardPaths.writableLocation(QStandardPaths.TempLocation)
else:
from qtpy.QtGui import QDesktopServices
def getTemporaryFolder():
return QDesktopServices.storageLocation(QDesktopServices.TempLocation)
|
import sys
import sip
import qtpy
__all__ = [
'getQVariantValue',
'wheelAngleDelta',
]
try:
if qtpy.PYQT5:
QVARIANT_API = 2
else:
QVARIANT_API = sip.getapi('QVariant')
except ValueError:
QVARIANT_API = 1
if QVARIANT_API == 1:
def getQVariantValue(variant):
return variant.toPyObject()
else:
def getQVariantValue(variant):
return variant
if qtpy.PYQT5:
def wheelAngleDelta(wheelEvent):
return wheelEvent.angleDelta().y()
else:
def wheelAngleDelta(wheelEvent):
return wheelEvent.delta()
if qtpy.PYQT5:
from qtpy.QtCore import QStandardPaths
def getTemporaryFolder():
return QStandardPaths.writableLocation(QStandardPaths.CacheLocation)
else:
from qtpy.QtGui import QDesktopServices
def getTemporaryFolder():
return QDesktopServices.storageLocation(QDesktopServices.CacheLocation)
|
Use Cache location instead of temp folder
|
Use Cache location instead of temp folder
|
Python
|
mit
|
allebacco/PyTileMap
|
import sys
import sip
import qtpy
__all__ = [
'getQVariantValue',
'wheelAngleDelta',
]
try:
if qtpy.PYQT5:
QVARIANT_API = 2
else:
QVARIANT_API = sip.getapi('QVariant')
except ValueError:
QVARIANT_API = 1
if QVARIANT_API == 1:
def getQVariantValue(variant):
return variant.toPyObject()
else:
def getQVariantValue(variant):
return variant
if qtpy.PYQT5:
def wheelAngleDelta(wheelEvent):
return wheelEvent.angleDelta().y()
else:
def wheelAngleDelta(wheelEvent):
return wheelEvent.delta()
if qtpy.PYQT5:
from qtpy.QtCore import QStandardPaths
def getTemporaryFolder():
return QStandardPaths.writableLocation(QStandardPaths.TempLocation)
else:
from qtpy.QtGui import QDesktopServices
def getTemporaryFolder():
return QDesktopServices.storageLocation(QDesktopServices.TempLocation)
Use Cache location instead of temp folder
|
import sys
import sip
import qtpy
__all__ = [
'getQVariantValue',
'wheelAngleDelta',
]
try:
if qtpy.PYQT5:
QVARIANT_API = 2
else:
QVARIANT_API = sip.getapi('QVariant')
except ValueError:
QVARIANT_API = 1
if QVARIANT_API == 1:
def getQVariantValue(variant):
return variant.toPyObject()
else:
def getQVariantValue(variant):
return variant
if qtpy.PYQT5:
def wheelAngleDelta(wheelEvent):
return wheelEvent.angleDelta().y()
else:
def wheelAngleDelta(wheelEvent):
return wheelEvent.delta()
if qtpy.PYQT5:
from qtpy.QtCore import QStandardPaths
def getTemporaryFolder():
return QStandardPaths.writableLocation(QStandardPaths.CacheLocation)
else:
from qtpy.QtGui import QDesktopServices
def getTemporaryFolder():
return QDesktopServices.storageLocation(QDesktopServices.CacheLocation)
|
<commit_before>
import sys
import sip
import qtpy
__all__ = [
'getQVariantValue',
'wheelAngleDelta',
]
try:
if qtpy.PYQT5:
QVARIANT_API = 2
else:
QVARIANT_API = sip.getapi('QVariant')
except ValueError:
QVARIANT_API = 1
if QVARIANT_API == 1:
def getQVariantValue(variant):
return variant.toPyObject()
else:
def getQVariantValue(variant):
return variant
if qtpy.PYQT5:
def wheelAngleDelta(wheelEvent):
return wheelEvent.angleDelta().y()
else:
def wheelAngleDelta(wheelEvent):
return wheelEvent.delta()
if qtpy.PYQT5:
from qtpy.QtCore import QStandardPaths
def getTemporaryFolder():
return QStandardPaths.writableLocation(QStandardPaths.TempLocation)
else:
from qtpy.QtGui import QDesktopServices
def getTemporaryFolder():
return QDesktopServices.storageLocation(QDesktopServices.TempLocation)
<commit_msg>Use Cache location instead of temp folder<commit_after>
|
import sys
import sip
import qtpy
__all__ = [
'getQVariantValue',
'wheelAngleDelta',
]
try:
if qtpy.PYQT5:
QVARIANT_API = 2
else:
QVARIANT_API = sip.getapi('QVariant')
except ValueError:
QVARIANT_API = 1
if QVARIANT_API == 1:
def getQVariantValue(variant):
return variant.toPyObject()
else:
def getQVariantValue(variant):
return variant
if qtpy.PYQT5:
def wheelAngleDelta(wheelEvent):
return wheelEvent.angleDelta().y()
else:
def wheelAngleDelta(wheelEvent):
return wheelEvent.delta()
if qtpy.PYQT5:
from qtpy.QtCore import QStandardPaths
def getTemporaryFolder():
return QStandardPaths.writableLocation(QStandardPaths.CacheLocation)
else:
from qtpy.QtGui import QDesktopServices
def getTemporaryFolder():
return QDesktopServices.storageLocation(QDesktopServices.CacheLocation)
|
import sys
import sip
import qtpy
__all__ = [
'getQVariantValue',
'wheelAngleDelta',
]
try:
if qtpy.PYQT5:
QVARIANT_API = 2
else:
QVARIANT_API = sip.getapi('QVariant')
except ValueError:
QVARIANT_API = 1
if QVARIANT_API == 1:
def getQVariantValue(variant):
return variant.toPyObject()
else:
def getQVariantValue(variant):
return variant
if qtpy.PYQT5:
def wheelAngleDelta(wheelEvent):
return wheelEvent.angleDelta().y()
else:
def wheelAngleDelta(wheelEvent):
return wheelEvent.delta()
if qtpy.PYQT5:
from qtpy.QtCore import QStandardPaths
def getTemporaryFolder():
return QStandardPaths.writableLocation(QStandardPaths.TempLocation)
else:
from qtpy.QtGui import QDesktopServices
def getTemporaryFolder():
return QDesktopServices.storageLocation(QDesktopServices.TempLocation)
Use Cache location instead of temp folder
import sys
import sip
import qtpy
__all__ = [
'getQVariantValue',
'wheelAngleDelta',
]
try:
if qtpy.PYQT5:
QVARIANT_API = 2
else:
QVARIANT_API = sip.getapi('QVariant')
except ValueError:
QVARIANT_API = 1
if QVARIANT_API == 1:
def getQVariantValue(variant):
return variant.toPyObject()
else:
def getQVariantValue(variant):
return variant
if qtpy.PYQT5:
def wheelAngleDelta(wheelEvent):
return wheelEvent.angleDelta().y()
else:
def wheelAngleDelta(wheelEvent):
return wheelEvent.delta()
if qtpy.PYQT5:
from qtpy.QtCore import QStandardPaths
def getTemporaryFolder():
return QStandardPaths.writableLocation(QStandardPaths.CacheLocation)
else:
from qtpy.QtGui import QDesktopServices
def getTemporaryFolder():
return QDesktopServices.storageLocation(QDesktopServices.CacheLocation)
|
<commit_before>
import sys
import sip
import qtpy
__all__ = [
'getQVariantValue',
'wheelAngleDelta',
]
try:
if qtpy.PYQT5:
QVARIANT_API = 2
else:
QVARIANT_API = sip.getapi('QVariant')
except ValueError:
QVARIANT_API = 1
if QVARIANT_API == 1:
def getQVariantValue(variant):
return variant.toPyObject()
else:
def getQVariantValue(variant):
return variant
if qtpy.PYQT5:
def wheelAngleDelta(wheelEvent):
return wheelEvent.angleDelta().y()
else:
def wheelAngleDelta(wheelEvent):
return wheelEvent.delta()
if qtpy.PYQT5:
from qtpy.QtCore import QStandardPaths
def getTemporaryFolder():
return QStandardPaths.writableLocation(QStandardPaths.TempLocation)
else:
from qtpy.QtGui import QDesktopServices
def getTemporaryFolder():
return QDesktopServices.storageLocation(QDesktopServices.TempLocation)
<commit_msg>Use Cache location instead of temp folder<commit_after>
import sys
import sip
import qtpy
__all__ = [
'getQVariantValue',
'wheelAngleDelta',
]
try:
if qtpy.PYQT5:
QVARIANT_API = 2
else:
QVARIANT_API = sip.getapi('QVariant')
except ValueError:
QVARIANT_API = 1
if QVARIANT_API == 1:
def getQVariantValue(variant):
return variant.toPyObject()
else:
def getQVariantValue(variant):
return variant
if qtpy.PYQT5:
def wheelAngleDelta(wheelEvent):
return wheelEvent.angleDelta().y()
else:
def wheelAngleDelta(wheelEvent):
return wheelEvent.delta()
if qtpy.PYQT5:
from qtpy.QtCore import QStandardPaths
def getTemporaryFolder():
return QStandardPaths.writableLocation(QStandardPaths.CacheLocation)
else:
from qtpy.QtGui import QDesktopServices
def getTemporaryFolder():
return QDesktopServices.storageLocation(QDesktopServices.CacheLocation)
|
4fd6abddcc3457e53046f5a1c1bcc277083a8b15
|
entrypoint.py
|
entrypoint.py
|
#!/usr/bin/python3
#
# Define containerized environment for running Diosix on Qemu
#
# On Google Cloud Run: Creates HTTP server on port 8080
# or whatever was specified using the PORT system variable.
# Use this to signal the build was successful and the container\
# can be run via the command line.
#
# On all other environments: Log to stdout
#
# syntax: entrypoint.py <command>
#
# Author: Chris Williams <diodesign@tuta.io>
#
import os
import sys
global command_result
from flask import Flask
app = Flask(__name__)
# for Google Cloud Run
@app.route('/')
def ContainerService():
return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\n'
if __name__ == "__main__":
if (os.environ.get('K_SERVICE')) != '':
app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080)))
else:
stream = os.popen('. $HOME/.cargo/env && cd /build/diosix && {}'.format(' '.join(sys.argv[1:])))
output = stream.read()
output
|
#!/usr/bin/python3
#
# Define containerized environment for running Diosix on Qemu
#
# On Google Cloud Run: Creates HTTP server on port 8080
# or whatever was specified using the PORT system variable.
# Outputs via the HTTP port. This requires K_SERVICE to be set.
#
# On all other environments: Log to stdout
#
# syntax: entrypoint.py <command>
#
# Author: Chris Williams <diodesign@tuta.io>
#
import os
import sys
global command_result
from flask import Flask
app = Flask(__name__)
# for Google Cloud Run
@app.route('/')
def ContainerService():
return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\n'
if __name__ == "__main__":
if (os.environ.get('K_SERVICE')) != '':
print('Running HTTP service for Google Cloud')
app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080)))
else:
print('Running locally')
stream = os.popen('. $HOME/.cargo/env && cd /build/diosix && {}'.format(' '.join(sys.argv[1:])))
output = stream.read()
output
|
Debug Google Cloud Run support
|
Debug Google Cloud Run support
|
Python
|
mit
|
diodesign/diosix
|
#!/usr/bin/python3
#
# Define containerized environment for running Diosix on Qemu
#
# On Google Cloud Run: Creates HTTP server on port 8080
# or whatever was specified using the PORT system variable.
# Use this to signal the build was successful and the container\
# can be run via the command line.
#
# On all other environments: Log to stdout
#
# syntax: entrypoint.py <command>
#
# Author: Chris Williams <diodesign@tuta.io>
#
import os
import sys
global command_result
from flask import Flask
app = Flask(__name__)
# for Google Cloud Run
@app.route('/')
def ContainerService():
return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\n'
if __name__ == "__main__":
if (os.environ.get('K_SERVICE')) != '':
app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080)))
else:
stream = os.popen('. $HOME/.cargo/env && cd /build/diosix && {}'.format(' '.join(sys.argv[1:])))
output = stream.read()
output
Debug Google Cloud Run support
|
#!/usr/bin/python3
#
# Define containerized environment for running Diosix on Qemu
#
# On Google Cloud Run: Creates HTTP server on port 8080
# or whatever was specified using the PORT system variable.
# Outputs via the HTTP port. This requires K_SERVICE to be set.
#
# On all other environments: Log to stdout
#
# syntax: entrypoint.py <command>
#
# Author: Chris Williams <diodesign@tuta.io>
#
import os
import sys
global command_result
from flask import Flask
app = Flask(__name__)
# for Google Cloud Run
@app.route('/')
def ContainerService():
return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\n'
if __name__ == "__main__":
if (os.environ.get('K_SERVICE')) != '':
print('Running HTTP service for Google Cloud')
app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080)))
else:
print('Running locally')
stream = os.popen('. $HOME/.cargo/env && cd /build/diosix && {}'.format(' '.join(sys.argv[1:])))
output = stream.read()
output
|
<commit_before>#!/usr/bin/python3
#
# Define containerized environment for running Diosix on Qemu
#
# On Google Cloud Run: Creates HTTP server on port 8080
# or whatever was specified using the PORT system variable.
# Use this to signal the build was successful and the container\
# can be run via the command line.
#
# On all other environments: Log to stdout
#
# syntax: entrypoint.py <command>
#
# Author: Chris Williams <diodesign@tuta.io>
#
import os
import sys
global command_result
from flask import Flask
app = Flask(__name__)
# for Google Cloud Run
@app.route('/')
def ContainerService():
return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\n'
if __name__ == "__main__":
if (os.environ.get('K_SERVICE')) != '':
app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080)))
else:
stream = os.popen('. $HOME/.cargo/env && cd /build/diosix && {}'.format(' '.join(sys.argv[1:])))
output = stream.read()
output
<commit_msg>Debug Google Cloud Run support<commit_after>
|
#!/usr/bin/python3
#
# Define containerized environment for running Diosix on Qemu
#
# On Google Cloud Run: Creates HTTP server on port 8080
# or whatever was specified using the PORT system variable.
# Outputs via the HTTP port. This requires K_SERVICE to be set.
#
# On all other environments: Log to stdout
#
# syntax: entrypoint.py <command>
#
# Author: Chris Williams <diodesign@tuta.io>
#
import os
import sys
global command_result
from flask import Flask
app = Flask(__name__)
# for Google Cloud Run
@app.route('/')
def ContainerService():
return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\n'
if __name__ == "__main__":
if (os.environ.get('K_SERVICE')) != '':
print('Running HTTP service for Google Cloud')
app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080)))
else:
print('Running locally')
stream = os.popen('. $HOME/.cargo/env && cd /build/diosix && {}'.format(' '.join(sys.argv[1:])))
output = stream.read()
output
|
#!/usr/bin/python3
#
# Define containerized environment for running Diosix on Qemu
#
# On Google Cloud Run: Creates HTTP server on port 8080
# or whatever was specified using the PORT system variable.
# Use this to signal the build was successful and the container\
# can be run via the command line.
#
# On all other environments: Log to stdout
#
# syntax: entrypoint.py <command>
#
# Author: Chris Williams <diodesign@tuta.io>
#
import os
import sys
global command_result
from flask import Flask
app = Flask(__name__)
# for Google Cloud Run
@app.route('/')
def ContainerService():
return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\n'
if __name__ == "__main__":
if (os.environ.get('K_SERVICE')) != '':
app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080)))
else:
stream = os.popen('. $HOME/.cargo/env && cd /build/diosix && {}'.format(' '.join(sys.argv[1:])))
output = stream.read()
output
Debug Google Cloud Run support#!/usr/bin/python3
#
# Define containerized environment for running Diosix on Qemu
#
# On Google Cloud Run: Creates HTTP server on port 8080
# or whatever was specified using the PORT system variable.
# Outputs via the HTTP port. This requires K_SERVICE to be set.
#
# On all other environments: Log to stdout
#
# syntax: entrypoint.py <command>
#
# Author: Chris Williams <diodesign@tuta.io>
#
import os
import sys
global command_result
from flask import Flask
app = Flask(__name__)
# for Google Cloud Run
@app.route('/')
def ContainerService():
return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\n'
if __name__ == "__main__":
if (os.environ.get('K_SERVICE')) != '':
print('Running HTTP service for Google Cloud')
app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080)))
else:
print('Running locally')
stream = os.popen('. $HOME/.cargo/env && cd /build/diosix && {}'.format(' '.join(sys.argv[1:])))
output = stream.read()
output
|
<commit_before>#!/usr/bin/python3
#
# Define containerized environment for running Diosix on Qemu
#
# On Google Cloud Run: Creates HTTP server on port 8080
# or whatever was specified using the PORT system variable.
# Use this to signal the build was successful and the container\
# can be run via the command line.
#
# On all other environments: Log to stdout
#
# syntax: entrypoint.py <command>
#
# Author: Chris Williams <diodesign@tuta.io>
#
import os
import sys
global command_result
from flask import Flask
app = Flask(__name__)
# for Google Cloud Run
@app.route('/')
def ContainerService():
return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\n'
if __name__ == "__main__":
if (os.environ.get('K_SERVICE')) != '':
app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080)))
else:
stream = os.popen('. $HOME/.cargo/env && cd /build/diosix && {}'.format(' '.join(sys.argv[1:])))
output = stream.read()
output
<commit_msg>Debug Google Cloud Run support<commit_after>#!/usr/bin/python3
#
# Define containerized environment for running Diosix on Qemu
#
# On Google Cloud Run: Creates HTTP server on port 8080
# or whatever was specified using the PORT system variable.
# Outputs via the HTTP port. This requires K_SERVICE to be set.
#
# On all other environments: Log to stdout
#
# syntax: entrypoint.py <command>
#
# Author: Chris Williams <diodesign@tuta.io>
#
import os
import sys
global command_result
from flask import Flask
app = Flask(__name__)
# for Google Cloud Run
@app.route('/')
def ContainerService():
return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\n'
if __name__ == "__main__":
if (os.environ.get('K_SERVICE')) != '':
print('Running HTTP service for Google Cloud')
app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080)))
else:
print('Running locally')
stream = os.popen('. $HOME/.cargo/env && cd /build/diosix && {}'.format(' '.join(sys.argv[1:])))
output = stream.read()
output
|
739cf9a93afd9c742675e24cc637634e67d2c3b9
|
src/lavatory/utils/get_artifactory_info.py
|
src/lavatory/utils/get_artifactory_info.py
|
"""Helper method for getting artifactory information."""
import logging
from .artifactory import Artifactory
def get_artifactory_info(repo_names=None, repo_type='local'):
"""Get storage info from Artifactory.
Args:
repo_names (tuple, optional): Name of artifactory repo.
repo_type (str): Type of artifactory repo.
Returns:
keys (dict, optional): Dictionary of repo data.
storage_info (dict): Storage information api call.
"""
artifactory = Artifactory(repo_name=repo_names)
storage_info = artifactory.repos(repo_type=repo_type)
if repo_names:
keys = repo_names
else:
keys = storage_info.keys()
logging.debug('Storage info: %s', storage_info)
logging.debug('Keys: %s', keys)
return storage_info, keys
|
"""Helper method for getting artifactory information."""
import logging
import requests
from .artifactory import Artifactory
def _artifactory(artifactory=None, repo_names=None):
if not artifactory:
artifactory = Artifactory(repo_name=repo_names)
return artifactory
def get_storage(repo_names=None, repo_type=None):
artifactory = _artifactory(repo_names=repo_names)
storage_info = []
try:
storage_info = artifactory.repos(repo_type=repo_type)
except requests.exceptions.HTTPError:
logging.warning('Account is not an admin and may not be able to get storage details.')
logging.debug('Storage info: %s', storage_info)
return storage_info
def get_repos(repo_names=None, repo_type='local'):
repos = []
if repo_names:
repos = repo_names
else:
repos = get_storage(repo_names=repo_names, repo_type=repo_type)
return repos
def get_artifactory_info(repo_names=None, repo_type='local'):
"""Get storage info from Artifactory.
Args:
repo_names (tuple, optional): Name of artifactory repo.
repo_type (str): Type of artifactory repo.
Returns:
keys (dict, optional): Dictionary of repo data.
storage_info (dict): Storage information api call.
"""
artifactory = Artifactory(repo_name=repo_names)
storage_info = artifactory.repos(repo_type=repo_type)
if repo_names:
keys = repo_names
else:
keys = storage_info.keys()
logging.debug('Storage info: %s', storage_info)
logging.debug('Keys: %s', keys)
return storage_info, keys
|
Add storage and repo helper functions
|
feat: Add storage and repo helper functions
|
Python
|
apache-2.0
|
gogoair/lavatory
|
"""Helper method for getting artifactory information."""
import logging
from .artifactory import Artifactory
def get_artifactory_info(repo_names=None, repo_type='local'):
"""Get storage info from Artifactory.
Args:
repo_names (tuple, optional): Name of artifactory repo.
repo_type (str): Type of artifactory repo.
Returns:
keys (dict, optional): Dictionary of repo data.
storage_info (dict): Storage information api call.
"""
artifactory = Artifactory(repo_name=repo_names)
storage_info = artifactory.repos(repo_type=repo_type)
if repo_names:
keys = repo_names
else:
keys = storage_info.keys()
logging.debug('Storage info: %s', storage_info)
logging.debug('Keys: %s', keys)
return storage_info, keys
feat: Add storage and repo helper functions
|
"""Helper method for getting artifactory information."""
import logging
import requests
from .artifactory import Artifactory
def _artifactory(artifactory=None, repo_names=None):
if not artifactory:
artifactory = Artifactory(repo_name=repo_names)
return artifactory
def get_storage(repo_names=None, repo_type=None):
artifactory = _artifactory(repo_names=repo_names)
storage_info = []
try:
storage_info = artifactory.repos(repo_type=repo_type)
except requests.exceptions.HTTPError:
logging.warning('Account is not an admin and may not be able to get storage details.')
logging.debug('Storage info: %s', storage_info)
return storage_info
def get_repos(repo_names=None, repo_type='local'):
repos = []
if repo_names:
repos = repo_names
else:
repos = get_storage(repo_names=repo_names, repo_type=repo_type)
return repos
def get_artifactory_info(repo_names=None, repo_type='local'):
"""Get storage info from Artifactory.
Args:
repo_names (tuple, optional): Name of artifactory repo.
repo_type (str): Type of artifactory repo.
Returns:
keys (dict, optional): Dictionary of repo data.
storage_info (dict): Storage information api call.
"""
artifactory = Artifactory(repo_name=repo_names)
storage_info = artifactory.repos(repo_type=repo_type)
if repo_names:
keys = repo_names
else:
keys = storage_info.keys()
logging.debug('Storage info: %s', storage_info)
logging.debug('Keys: %s', keys)
return storage_info, keys
|
<commit_before>"""Helper method for getting artifactory information."""
import logging
from .artifactory import Artifactory
def get_artifactory_info(repo_names=None, repo_type='local'):
"""Get storage info from Artifactory.
Args:
repo_names (tuple, optional): Name of artifactory repo.
repo_type (str): Type of artifactory repo.
Returns:
keys (dict, optional): Dictionary of repo data.
storage_info (dict): Storage information api call.
"""
artifactory = Artifactory(repo_name=repo_names)
storage_info = artifactory.repos(repo_type=repo_type)
if repo_names:
keys = repo_names
else:
keys = storage_info.keys()
logging.debug('Storage info: %s', storage_info)
logging.debug('Keys: %s', keys)
return storage_info, keys
<commit_msg>feat: Add storage and repo helper functions<commit_after>
|
"""Helper method for getting artifactory information."""
import logging
import requests
from .artifactory import Artifactory
def _artifactory(artifactory=None, repo_names=None):
if not artifactory:
artifactory = Artifactory(repo_name=repo_names)
return artifactory
def get_storage(repo_names=None, repo_type=None):
artifactory = _artifactory(repo_names=repo_names)
storage_info = []
try:
storage_info = artifactory.repos(repo_type=repo_type)
except requests.exceptions.HTTPError:
logging.warning('Account is not an admin and may not be able to get storage details.')
logging.debug('Storage info: %s', storage_info)
return storage_info
def get_repos(repo_names=None, repo_type='local'):
repos = []
if repo_names:
repos = repo_names
else:
repos = get_storage(repo_names=repo_names, repo_type=repo_type)
return repos
def get_artifactory_info(repo_names=None, repo_type='local'):
"""Get storage info from Artifactory.
Args:
repo_names (tuple, optional): Name of artifactory repo.
repo_type (str): Type of artifactory repo.
Returns:
keys (dict, optional): Dictionary of repo data.
storage_info (dict): Storage information api call.
"""
artifactory = Artifactory(repo_name=repo_names)
storage_info = artifactory.repos(repo_type=repo_type)
if repo_names:
keys = repo_names
else:
keys = storage_info.keys()
logging.debug('Storage info: %s', storage_info)
logging.debug('Keys: %s', keys)
return storage_info, keys
|
"""Helper method for getting artifactory information."""
import logging
from .artifactory import Artifactory
def get_artifactory_info(repo_names=None, repo_type='local'):
"""Get storage info from Artifactory.
Args:
repo_names (tuple, optional): Name of artifactory repo.
repo_type (str): Type of artifactory repo.
Returns:
keys (dict, optional): Dictionary of repo data.
storage_info (dict): Storage information api call.
"""
artifactory = Artifactory(repo_name=repo_names)
storage_info = artifactory.repos(repo_type=repo_type)
if repo_names:
keys = repo_names
else:
keys = storage_info.keys()
logging.debug('Storage info: %s', storage_info)
logging.debug('Keys: %s', keys)
return storage_info, keys
feat: Add storage and repo helper functions"""Helper method for getting artifactory information."""
import logging
import requests
from .artifactory import Artifactory
def _artifactory(artifactory=None, repo_names=None):
if not artifactory:
artifactory = Artifactory(repo_name=repo_names)
return artifactory
def get_storage(repo_names=None, repo_type=None):
artifactory = _artifactory(repo_names=repo_names)
storage_info = []
try:
storage_info = artifactory.repos(repo_type=repo_type)
except requests.exceptions.HTTPError:
logging.warning('Account is not an admin and may not be able to get storage details.')
logging.debug('Storage info: %s', storage_info)
return storage_info
def get_repos(repo_names=None, repo_type='local'):
repos = []
if repo_names:
repos = repo_names
else:
repos = get_storage(repo_names=repo_names, repo_type=repo_type)
return repos
def get_artifactory_info(repo_names=None, repo_type='local'):
"""Get storage info from Artifactory.
Args:
repo_names (tuple, optional): Name of artifactory repo.
repo_type (str): Type of artifactory repo.
Returns:
keys (dict, optional): Dictionary of repo data.
storage_info (dict): Storage information api call.
"""
artifactory = Artifactory(repo_name=repo_names)
storage_info = artifactory.repos(repo_type=repo_type)
if repo_names:
keys = repo_names
else:
keys = storage_info.keys()
logging.debug('Storage info: %s', storage_info)
logging.debug('Keys: %s', keys)
return storage_info, keys
|
<commit_before>"""Helper method for getting artifactory information."""
import logging
from .artifactory import Artifactory
def get_artifactory_info(repo_names=None, repo_type='local'):
"""Get storage info from Artifactory.
Args:
repo_names (tuple, optional): Name of artifactory repo.
repo_type (str): Type of artifactory repo.
Returns:
keys (dict, optional): Dictionary of repo data.
storage_info (dict): Storage information api call.
"""
artifactory = Artifactory(repo_name=repo_names)
storage_info = artifactory.repos(repo_type=repo_type)
if repo_names:
keys = repo_names
else:
keys = storage_info.keys()
logging.debug('Storage info: %s', storage_info)
logging.debug('Keys: %s', keys)
return storage_info, keys
<commit_msg>feat: Add storage and repo helper functions<commit_after>"""Helper method for getting artifactory information."""
import logging
import requests
from .artifactory import Artifactory
def _artifactory(artifactory=None, repo_names=None):
if not artifactory:
artifactory = Artifactory(repo_name=repo_names)
return artifactory
def get_storage(repo_names=None, repo_type=None):
artifactory = _artifactory(repo_names=repo_names)
storage_info = []
try:
storage_info = artifactory.repos(repo_type=repo_type)
except requests.exceptions.HTTPError:
logging.warning('Account is not an admin and may not be able to get storage details.')
logging.debug('Storage info: %s', storage_info)
return storage_info
def get_repos(repo_names=None, repo_type='local'):
repos = []
if repo_names:
repos = repo_names
else:
repos = get_storage(repo_names=repo_names, repo_type=repo_type)
return repos
def get_artifactory_info(repo_names=None, repo_type='local'):
"""Get storage info from Artifactory.
Args:
repo_names (tuple, optional): Name of artifactory repo.
repo_type (str): Type of artifactory repo.
Returns:
keys (dict, optional): Dictionary of repo data.
storage_info (dict): Storage information api call.
"""
artifactory = Artifactory(repo_name=repo_names)
storage_info = artifactory.repos(repo_type=repo_type)
if repo_names:
keys = repo_names
else:
keys = storage_info.keys()
logging.debug('Storage info: %s', storage_info)
logging.debug('Keys: %s', keys)
return storage_info, keys
|
06b99c4415a6605cbd6123271d44af96585fbb9d
|
conda_env/exceptions.py
|
conda_env/exceptions.py
|
class CondaEnvException(Exception):
pass
class EnvironmentFileNotFound(CondaEnvException):
def __init__(self, filename, *args, **kwargs):
msg = '{} file not found'.format(filename)
self.filename = filename
super(EnvironmentFileNotFound, self).__init__(msg, *args, **kwargs)
|
class CondaEnvException(Exception):
pass
class EnvironmentFileNotFound(CondaEnvException):
def __init__(self, filename, *args, **kwargs):
msg = '{} file not found'.format(filename)
self.filename = filename
super(EnvironmentFileNotFound, self).__init__(msg, *args, **kwargs)
class EnvironmentFileNotDownloaded(CondaEnvException):
def __init__(self, filename, *args, **kwargs):
msg = '{} file not downloaded'.format(filename)
self.filename = filename
super(EnvironmentFileNotDownloaded, self).__init__(msg, *args, **kwargs)
|
Add environment not found exception
|
Add environment not found exception
|
Python
|
bsd-3-clause
|
nicoddemus/conda-env,mikecroucher/conda-env,dan-blanchard/conda-env,conda/conda-env,dan-blanchard/conda-env,phobson/conda-env,conda/conda-env,ESSS/conda-env,mikecroucher/conda-env,asmeurer/conda-env,nicoddemus/conda-env,ESSS/conda-env,phobson/conda-env,isaac-kit/conda-env,asmeurer/conda-env,isaac-kit/conda-env
|
class CondaEnvException(Exception):
pass
class EnvironmentFileNotFound(CondaEnvException):
def __init__(self, filename, *args, **kwargs):
msg = '{} file not found'.format(filename)
self.filename = filename
super(EnvironmentFileNotFound, self).__init__(msg, *args, **kwargs)
Add environment not found exception
|
class CondaEnvException(Exception):
pass
class EnvironmentFileNotFound(CondaEnvException):
def __init__(self, filename, *args, **kwargs):
msg = '{} file not found'.format(filename)
self.filename = filename
super(EnvironmentFileNotFound, self).__init__(msg, *args, **kwargs)
class EnvironmentFileNotDownloaded(CondaEnvException):
def __init__(self, filename, *args, **kwargs):
msg = '{} file not downloaded'.format(filename)
self.filename = filename
super(EnvironmentFileNotDownloaded, self).__init__(msg, *args, **kwargs)
|
<commit_before>class CondaEnvException(Exception):
pass
class EnvironmentFileNotFound(CondaEnvException):
def __init__(self, filename, *args, **kwargs):
msg = '{} file not found'.format(filename)
self.filename = filename
super(EnvironmentFileNotFound, self).__init__(msg, *args, **kwargs)
<commit_msg>Add environment not found exception<commit_after>
|
class CondaEnvException(Exception):
pass
class EnvironmentFileNotFound(CondaEnvException):
def __init__(self, filename, *args, **kwargs):
msg = '{} file not found'.format(filename)
self.filename = filename
super(EnvironmentFileNotFound, self).__init__(msg, *args, **kwargs)
class EnvironmentFileNotDownloaded(CondaEnvException):
def __init__(self, filename, *args, **kwargs):
msg = '{} file not downloaded'.format(filename)
self.filename = filename
super(EnvironmentFileNotDownloaded, self).__init__(msg, *args, **kwargs)
|
class CondaEnvException(Exception):
pass
class EnvironmentFileNotFound(CondaEnvException):
def __init__(self, filename, *args, **kwargs):
msg = '{} file not found'.format(filename)
self.filename = filename
super(EnvironmentFileNotFound, self).__init__(msg, *args, **kwargs)
Add environment not found exceptionclass CondaEnvException(Exception):
pass
class EnvironmentFileNotFound(CondaEnvException):
def __init__(self, filename, *args, **kwargs):
msg = '{} file not found'.format(filename)
self.filename = filename
super(EnvironmentFileNotFound, self).__init__(msg, *args, **kwargs)
class EnvironmentFileNotDownloaded(CondaEnvException):
def __init__(self, filename, *args, **kwargs):
msg = '{} file not downloaded'.format(filename)
self.filename = filename
super(EnvironmentFileNotDownloaded, self).__init__(msg, *args, **kwargs)
|
<commit_before>class CondaEnvException(Exception):
pass
class EnvironmentFileNotFound(CondaEnvException):
def __init__(self, filename, *args, **kwargs):
msg = '{} file not found'.format(filename)
self.filename = filename
super(EnvironmentFileNotFound, self).__init__(msg, *args, **kwargs)
<commit_msg>Add environment not found exception<commit_after>class CondaEnvException(Exception):
pass
class EnvironmentFileNotFound(CondaEnvException):
def __init__(self, filename, *args, **kwargs):
msg = '{} file not found'.format(filename)
self.filename = filename
super(EnvironmentFileNotFound, self).__init__(msg, *args, **kwargs)
class EnvironmentFileNotDownloaded(CondaEnvException):
def __init__(self, filename, *args, **kwargs):
msg = '{} file not downloaded'.format(filename)
self.filename = filename
super(EnvironmentFileNotDownloaded, self).__init__(msg, *args, **kwargs)
|
484f42b6fc1a8129a53480bc6e7913c5c7d58f46
|
froide/foirequest/search_indexes.py
|
froide/foirequest/search_indexes.py
|
from haystack import indexes
from celery_haystack.indexes import CelerySearchIndex
from .models import FoiRequest
class FoiRequestIndex(CelerySearchIndex, indexes.Indexable):
text = indexes.EdgeNgramField(document=True, use_template=True)
title = indexes.CharField(model_attr='title')
description = indexes.CharField(model_attr='description')
resolution = indexes.CharField(model_attr='resolution', default="")
status = indexes.CharField(model_attr='status')
readable_status = indexes.CharField(model_attr='readable_status')
first_message = indexes.DateTimeField(model_attr='first_message')
last_message = indexes.DateTimeField(model_attr='last_message')
url = indexes.CharField(model_attr='get_absolute_url')
public_body_name = indexes.CharField(model_attr='public_body__name', default="")
def get_model(self):
return FoiRequest
def index_queryset(self):
"""Used when the entire index for model is updated."""
return self.get_model().published.get_for_search_index()
def should_update(self, instance, **kwargs):
return instance.visibility > 1
|
from haystack import indexes
from celery_haystack.indexes import CelerySearchIndex
from .models import FoiRequest
class FoiRequestIndex(CelerySearchIndex, indexes.Indexable):
text = indexes.EdgeNgramField(document=True, use_template=True)
title = indexes.CharField(model_attr='title')
description = indexes.CharField(model_attr='description')
resolution = indexes.CharField(model_attr='resolution', default="")
status = indexes.CharField(model_attr='status')
readable_status = indexes.CharField(model_attr='readable_status')
first_message = indexes.DateTimeField(model_attr='first_message')
last_message = indexes.DateTimeField(model_attr='last_message')
url = indexes.CharField(model_attr='get_absolute_url')
public_body_name = indexes.CharField(model_attr='public_body__name', default="")
def get_model(self):
return FoiRequest
def index_queryset(self):
"""Used when the entire index for model is updated."""
return self.get_model().published.get_for_search_index()
def should_update(self, instance, **kwargs):
return instance.visibility > 1 and instance.is_foi
|
Index only FoiRequests marked is_foi
|
Index only FoiRequests marked is_foi
|
Python
|
mit
|
fin/froide,CodeforHawaii/froide,catcosmo/froide,okfse/froide,LilithWittmann/froide,stefanw/froide,catcosmo/froide,catcosmo/froide,okfse/froide,fin/froide,ryankanno/froide,stefanw/froide,CodeforHawaii/froide,okfse/froide,LilithWittmann/froide,CodeforHawaii/froide,ryankanno/froide,ryankanno/froide,stefanw/froide,LilithWittmann/froide,okfse/froide,ryankanno/froide,stefanw/froide,okfse/froide,LilithWittmann/froide,LilithWittmann/froide,stefanw/froide,fin/froide,catcosmo/froide,CodeforHawaii/froide,catcosmo/froide,CodeforHawaii/froide,fin/froide,ryankanno/froide
|
from haystack import indexes
from celery_haystack.indexes import CelerySearchIndex
from .models import FoiRequest
class FoiRequestIndex(CelerySearchIndex, indexes.Indexable):
text = indexes.EdgeNgramField(document=True, use_template=True)
title = indexes.CharField(model_attr='title')
description = indexes.CharField(model_attr='description')
resolution = indexes.CharField(model_attr='resolution', default="")
status = indexes.CharField(model_attr='status')
readable_status = indexes.CharField(model_attr='readable_status')
first_message = indexes.DateTimeField(model_attr='first_message')
last_message = indexes.DateTimeField(model_attr='last_message')
url = indexes.CharField(model_attr='get_absolute_url')
public_body_name = indexes.CharField(model_attr='public_body__name', default="")
def get_model(self):
return FoiRequest
def index_queryset(self):
"""Used when the entire index for model is updated."""
return self.get_model().published.get_for_search_index()
def should_update(self, instance, **kwargs):
return instance.visibility > 1
Index only FoiRequests marked is_foi
|
from haystack import indexes
from celery_haystack.indexes import CelerySearchIndex
from .models import FoiRequest
class FoiRequestIndex(CelerySearchIndex, indexes.Indexable):
text = indexes.EdgeNgramField(document=True, use_template=True)
title = indexes.CharField(model_attr='title')
description = indexes.CharField(model_attr='description')
resolution = indexes.CharField(model_attr='resolution', default="")
status = indexes.CharField(model_attr='status')
readable_status = indexes.CharField(model_attr='readable_status')
first_message = indexes.DateTimeField(model_attr='first_message')
last_message = indexes.DateTimeField(model_attr='last_message')
url = indexes.CharField(model_attr='get_absolute_url')
public_body_name = indexes.CharField(model_attr='public_body__name', default="")
def get_model(self):
return FoiRequest
def index_queryset(self):
"""Used when the entire index for model is updated."""
return self.get_model().published.get_for_search_index()
def should_update(self, instance, **kwargs):
return instance.visibility > 1 and instance.is_foi
|
<commit_before>from haystack import indexes
from celery_haystack.indexes import CelerySearchIndex
from .models import FoiRequest
class FoiRequestIndex(CelerySearchIndex, indexes.Indexable):
text = indexes.EdgeNgramField(document=True, use_template=True)
title = indexes.CharField(model_attr='title')
description = indexes.CharField(model_attr='description')
resolution = indexes.CharField(model_attr='resolution', default="")
status = indexes.CharField(model_attr='status')
readable_status = indexes.CharField(model_attr='readable_status')
first_message = indexes.DateTimeField(model_attr='first_message')
last_message = indexes.DateTimeField(model_attr='last_message')
url = indexes.CharField(model_attr='get_absolute_url')
public_body_name = indexes.CharField(model_attr='public_body__name', default="")
def get_model(self):
return FoiRequest
def index_queryset(self):
"""Used when the entire index for model is updated."""
return self.get_model().published.get_for_search_index()
def should_update(self, instance, **kwargs):
return instance.visibility > 1
<commit_msg>Index only FoiRequests marked is_foi<commit_after>
|
from haystack import indexes
from celery_haystack.indexes import CelerySearchIndex
from .models import FoiRequest
class FoiRequestIndex(CelerySearchIndex, indexes.Indexable):
text = indexes.EdgeNgramField(document=True, use_template=True)
title = indexes.CharField(model_attr='title')
description = indexes.CharField(model_attr='description')
resolution = indexes.CharField(model_attr='resolution', default="")
status = indexes.CharField(model_attr='status')
readable_status = indexes.CharField(model_attr='readable_status')
first_message = indexes.DateTimeField(model_attr='first_message')
last_message = indexes.DateTimeField(model_attr='last_message')
url = indexes.CharField(model_attr='get_absolute_url')
public_body_name = indexes.CharField(model_attr='public_body__name', default="")
def get_model(self):
return FoiRequest
def index_queryset(self):
"""Used when the entire index for model is updated."""
return self.get_model().published.get_for_search_index()
def should_update(self, instance, **kwargs):
return instance.visibility > 1 and instance.is_foi
|
from haystack import indexes
from celery_haystack.indexes import CelerySearchIndex
from .models import FoiRequest
class FoiRequestIndex(CelerySearchIndex, indexes.Indexable):
text = indexes.EdgeNgramField(document=True, use_template=True)
title = indexes.CharField(model_attr='title')
description = indexes.CharField(model_attr='description')
resolution = indexes.CharField(model_attr='resolution', default="")
status = indexes.CharField(model_attr='status')
readable_status = indexes.CharField(model_attr='readable_status')
first_message = indexes.DateTimeField(model_attr='first_message')
last_message = indexes.DateTimeField(model_attr='last_message')
url = indexes.CharField(model_attr='get_absolute_url')
public_body_name = indexes.CharField(model_attr='public_body__name', default="")
def get_model(self):
return FoiRequest
def index_queryset(self):
"""Used when the entire index for model is updated."""
return self.get_model().published.get_for_search_index()
def should_update(self, instance, **kwargs):
return instance.visibility > 1
Index only FoiRequests marked is_foifrom haystack import indexes
from celery_haystack.indexes import CelerySearchIndex
from .models import FoiRequest
class FoiRequestIndex(CelerySearchIndex, indexes.Indexable):
text = indexes.EdgeNgramField(document=True, use_template=True)
title = indexes.CharField(model_attr='title')
description = indexes.CharField(model_attr='description')
resolution = indexes.CharField(model_attr='resolution', default="")
status = indexes.CharField(model_attr='status')
readable_status = indexes.CharField(model_attr='readable_status')
first_message = indexes.DateTimeField(model_attr='first_message')
last_message = indexes.DateTimeField(model_attr='last_message')
url = indexes.CharField(model_attr='get_absolute_url')
public_body_name = indexes.CharField(model_attr='public_body__name', default="")
def get_model(self):
return FoiRequest
def index_queryset(self):
"""Used when the entire index for model is updated."""
return self.get_model().published.get_for_search_index()
def should_update(self, instance, **kwargs):
return instance.visibility > 1 and instance.is_foi
|
<commit_before>from haystack import indexes
from celery_haystack.indexes import CelerySearchIndex
from .models import FoiRequest
class FoiRequestIndex(CelerySearchIndex, indexes.Indexable):
text = indexes.EdgeNgramField(document=True, use_template=True)
title = indexes.CharField(model_attr='title')
description = indexes.CharField(model_attr='description')
resolution = indexes.CharField(model_attr='resolution', default="")
status = indexes.CharField(model_attr='status')
readable_status = indexes.CharField(model_attr='readable_status')
first_message = indexes.DateTimeField(model_attr='first_message')
last_message = indexes.DateTimeField(model_attr='last_message')
url = indexes.CharField(model_attr='get_absolute_url')
public_body_name = indexes.CharField(model_attr='public_body__name', default="")
def get_model(self):
return FoiRequest
def index_queryset(self):
"""Used when the entire index for model is updated."""
return self.get_model().published.get_for_search_index()
def should_update(self, instance, **kwargs):
return instance.visibility > 1
<commit_msg>Index only FoiRequests marked is_foi<commit_after>from haystack import indexes
from celery_haystack.indexes import CelerySearchIndex
from .models import FoiRequest
class FoiRequestIndex(CelerySearchIndex, indexes.Indexable):
text = indexes.EdgeNgramField(document=True, use_template=True)
title = indexes.CharField(model_attr='title')
description = indexes.CharField(model_attr='description')
resolution = indexes.CharField(model_attr='resolution', default="")
status = indexes.CharField(model_attr='status')
readable_status = indexes.CharField(model_attr='readable_status')
first_message = indexes.DateTimeField(model_attr='first_message')
last_message = indexes.DateTimeField(model_attr='last_message')
url = indexes.CharField(model_attr='get_absolute_url')
public_body_name = indexes.CharField(model_attr='public_body__name', default="")
def get_model(self):
return FoiRequest
def index_queryset(self):
"""Used when the entire index for model is updated."""
return self.get_model().published.get_for_search_index()
def should_update(self, instance, **kwargs):
return instance.visibility > 1 and instance.is_foi
|
359595413071ff706b484a875a23a4a7d1508f50
|
bindings/python/llvm/tests/base.py
|
bindings/python/llvm/tests/base.py
|
import os.path
import unittest
POSSIBLE_TEST_BINARIES = [
'libreadline.so.5',
'libreadline.so.6',
]
POSSIBLE_TEST_BINARY_PATHS = [
'/lib',
'/usr/lib',
'/usr/local/lib',
]
class TestBase(unittest.TestCase):
def get_test_binary(self):
"""Helper to obtain a test binary for object file testing.
FIXME Support additional, highly-likely targets or create one
ourselves.
"""
for d in POSSIBLE_TEST_BINARY_PATHS:
for lib in POSSIBLE_TEST_BINARIES:
path = os.path.join(d, lib)
if os.path.exists(path):
return path
raise Exception('No suitable test binaries available!')
|
import os.path
import unittest
POSSIBLE_TEST_BINARIES = [
'libreadline.so.5',
'libreadline.so.6',
]
POSSIBLE_TEST_BINARY_PATHS = [
'/lib',
'/usr/lib',
'/usr/local/lib',
]
class TestBase(unittest.TestCase):
def get_test_binary(self):
"""Helper to obtain a test binary for object file testing.
FIXME Support additional, highly-likely targets or create one
ourselves.
"""
for d in POSSIBLE_TEST_BINARY_PATHS:
for lib in POSSIBLE_TEST_BINARIES:
path = os.path.join(d, lib)
if os.path.exists(path):
return path
raise Exception('No suitable test binaries available!')
get_test_binary.__test__ = False
|
Mark get_test_binary as not being a test
|
[python] Mark get_test_binary as not being a test
get_test_binary is a helper method, not a test, make sure nosetests
doesn't pick it up as a test.
git-svn-id: 0ff597fd157e6f4fc38580e8d64ab130330d2411@153173 91177308-0d34-0410-b5e6-96231b3b80d8
|
Python
|
apache-2.0
|
llvm-mirror/llvm,GPUOpen-Drivers/llvm,GPUOpen-Drivers/llvm,llvm-mirror/llvm,apple/swift-llvm,llvm-mirror/llvm,apple/swift-llvm,GPUOpen-Drivers/llvm,GPUOpen-Drivers/llvm,chubbymaggie/asap,dslab-epfl/asap,llvm-mirror/llvm,llvm-mirror/llvm,apple/swift-llvm,llvm-mirror/llvm,apple/swift-llvm,GPUOpen-Drivers/llvm,apple/swift-llvm,chubbymaggie/asap,chubbymaggie/asap,dslab-epfl/asap,dslab-epfl/asap,llvm-mirror/llvm,chubbymaggie/asap,llvm-mirror/llvm,dslab-epfl/asap,llvm-mirror/llvm,apple/swift-llvm,dslab-epfl/asap,apple/swift-llvm,apple/swift-llvm,dslab-epfl/asap,dslab-epfl/asap,GPUOpen-Drivers/llvm,chubbymaggie/asap,chubbymaggie/asap,GPUOpen-Drivers/llvm,GPUOpen-Drivers/llvm
|
import os.path
import unittest
POSSIBLE_TEST_BINARIES = [
'libreadline.so.5',
'libreadline.so.6',
]
POSSIBLE_TEST_BINARY_PATHS = [
'/lib',
'/usr/lib',
'/usr/local/lib',
]
class TestBase(unittest.TestCase):
def get_test_binary(self):
"""Helper to obtain a test binary for object file testing.
FIXME Support additional, highly-likely targets or create one
ourselves.
"""
for d in POSSIBLE_TEST_BINARY_PATHS:
for lib in POSSIBLE_TEST_BINARIES:
path = os.path.join(d, lib)
if os.path.exists(path):
return path
raise Exception('No suitable test binaries available!')
[python] Mark get_test_binary as not being a test
get_test_binary is a helper method, not a test, make sure nosetests
doesn't pick it up as a test.
git-svn-id: 0ff597fd157e6f4fc38580e8d64ab130330d2411@153173 91177308-0d34-0410-b5e6-96231b3b80d8
|
import os.path
import unittest
POSSIBLE_TEST_BINARIES = [
'libreadline.so.5',
'libreadline.so.6',
]
POSSIBLE_TEST_BINARY_PATHS = [
'/lib',
'/usr/lib',
'/usr/local/lib',
]
class TestBase(unittest.TestCase):
def get_test_binary(self):
"""Helper to obtain a test binary for object file testing.
FIXME Support additional, highly-likely targets or create one
ourselves.
"""
for d in POSSIBLE_TEST_BINARY_PATHS:
for lib in POSSIBLE_TEST_BINARIES:
path = os.path.join(d, lib)
if os.path.exists(path):
return path
raise Exception('No suitable test binaries available!')
get_test_binary.__test__ = False
|
<commit_before>import os.path
import unittest
POSSIBLE_TEST_BINARIES = [
'libreadline.so.5',
'libreadline.so.6',
]
POSSIBLE_TEST_BINARY_PATHS = [
'/lib',
'/usr/lib',
'/usr/local/lib',
]
class TestBase(unittest.TestCase):
def get_test_binary(self):
"""Helper to obtain a test binary for object file testing.
FIXME Support additional, highly-likely targets or create one
ourselves.
"""
for d in POSSIBLE_TEST_BINARY_PATHS:
for lib in POSSIBLE_TEST_BINARIES:
path = os.path.join(d, lib)
if os.path.exists(path):
return path
raise Exception('No suitable test binaries available!')
<commit_msg>[python] Mark get_test_binary as not being a test
get_test_binary is a helper method, not a test, make sure nosetests
doesn't pick it up as a test.
git-svn-id: 0ff597fd157e6f4fc38580e8d64ab130330d2411@153173 91177308-0d34-0410-b5e6-96231b3b80d8<commit_after>
|
import os.path
import unittest
POSSIBLE_TEST_BINARIES = [
'libreadline.so.5',
'libreadline.so.6',
]
POSSIBLE_TEST_BINARY_PATHS = [
'/lib',
'/usr/lib',
'/usr/local/lib',
]
class TestBase(unittest.TestCase):
def get_test_binary(self):
"""Helper to obtain a test binary for object file testing.
FIXME Support additional, highly-likely targets or create one
ourselves.
"""
for d in POSSIBLE_TEST_BINARY_PATHS:
for lib in POSSIBLE_TEST_BINARIES:
path = os.path.join(d, lib)
if os.path.exists(path):
return path
raise Exception('No suitable test binaries available!')
get_test_binary.__test__ = False
|
import os.path
import unittest
POSSIBLE_TEST_BINARIES = [
'libreadline.so.5',
'libreadline.so.6',
]
POSSIBLE_TEST_BINARY_PATHS = [
'/lib',
'/usr/lib',
'/usr/local/lib',
]
class TestBase(unittest.TestCase):
def get_test_binary(self):
"""Helper to obtain a test binary for object file testing.
FIXME Support additional, highly-likely targets or create one
ourselves.
"""
for d in POSSIBLE_TEST_BINARY_PATHS:
for lib in POSSIBLE_TEST_BINARIES:
path = os.path.join(d, lib)
if os.path.exists(path):
return path
raise Exception('No suitable test binaries available!')
[python] Mark get_test_binary as not being a test
get_test_binary is a helper method, not a test, make sure nosetests
doesn't pick it up as a test.
git-svn-id: 0ff597fd157e6f4fc38580e8d64ab130330d2411@153173 91177308-0d34-0410-b5e6-96231b3b80d8import os.path
import unittest
POSSIBLE_TEST_BINARIES = [
'libreadline.so.5',
'libreadline.so.6',
]
POSSIBLE_TEST_BINARY_PATHS = [
'/lib',
'/usr/lib',
'/usr/local/lib',
]
class TestBase(unittest.TestCase):
def get_test_binary(self):
"""Helper to obtain a test binary for object file testing.
FIXME Support additional, highly-likely targets or create one
ourselves.
"""
for d in POSSIBLE_TEST_BINARY_PATHS:
for lib in POSSIBLE_TEST_BINARIES:
path = os.path.join(d, lib)
if os.path.exists(path):
return path
raise Exception('No suitable test binaries available!')
get_test_binary.__test__ = False
|
<commit_before>import os.path
import unittest
POSSIBLE_TEST_BINARIES = [
'libreadline.so.5',
'libreadline.so.6',
]
POSSIBLE_TEST_BINARY_PATHS = [
'/lib',
'/usr/lib',
'/usr/local/lib',
]
class TestBase(unittest.TestCase):
def get_test_binary(self):
"""Helper to obtain a test binary for object file testing.
FIXME Support additional, highly-likely targets or create one
ourselves.
"""
for d in POSSIBLE_TEST_BINARY_PATHS:
for lib in POSSIBLE_TEST_BINARIES:
path = os.path.join(d, lib)
if os.path.exists(path):
return path
raise Exception('No suitable test binaries available!')
<commit_msg>[python] Mark get_test_binary as not being a test
get_test_binary is a helper method, not a test, make sure nosetests
doesn't pick it up as a test.
git-svn-id: 0ff597fd157e6f4fc38580e8d64ab130330d2411@153173 91177308-0d34-0410-b5e6-96231b3b80d8<commit_after>import os.path
import unittest
POSSIBLE_TEST_BINARIES = [
'libreadline.so.5',
'libreadline.so.6',
]
POSSIBLE_TEST_BINARY_PATHS = [
'/lib',
'/usr/lib',
'/usr/local/lib',
]
class TestBase(unittest.TestCase):
def get_test_binary(self):
"""Helper to obtain a test binary for object file testing.
FIXME Support additional, highly-likely targets or create one
ourselves.
"""
for d in POSSIBLE_TEST_BINARY_PATHS:
for lib in POSSIBLE_TEST_BINARIES:
path = os.path.join(d, lib)
if os.path.exists(path):
return path
raise Exception('No suitable test binaries available!')
get_test_binary.__test__ = False
|
50442966938b532cc759089692ffb52e94c6e89b
|
config_example.py
|
config_example.py
|
"""Example config.py"""
webhook_urls = ["DISCORD WEBHOOK", "DISCORD WEBHOOK"] # Used to update webhooks on Discord
key_path = "/path/to/key/in/format/of/file.pem" # Private key to sign the file
file_path = "/path/to/folder" # Path to save the file to
lzss_path = "/path/to/lzss" # Path to lzss
production = None # Use production mode
cachet_url = "http://status.domain.tld/api/v1" # URL for Cachet
cachet_key = "api_key" # API Key for Cachet
sentry_url = "http://status.domain.tld/" # URL for Sentry
"""News Channel only"""
force_all = False # Force the script to replace all news files
google_maps_api_key = "api_key" # API Key for Google Maps geocoding API
geoparser_keys = ["GEOPARSER KEY", "GEOPARSER KEY"] # API Key for Geoparser
"""Forecast Channel only"""
import forecastlists
useVerbose = None # Print more verbose messages
useMultithreaded = None # Use multithreading
weathercities = ["CITY LIST", "CITY LIST"] # Lists of cities to use
cachet_elapsed_time = None # ID of the Cachet point to log elapsed time.
"""Everybody Votes Channel only"""
mysql_user = "user" # MySQL username
mysql_password = "password" # MySQL password
mysql_database = "database" # MySQL database
|
"""Example config.py"""
webhook_urls = ["DISCORD WEBHOOK", "DISCORD WEBHOOK"] # Used to update webhooks on Discord
key_path = "/path/to/key/in/format/of/file.pem" # Private key to sign the file
file_path = "/path/to/folder" # Path to save the file to
lzss_path = "/path/to/lzss" # Path to lzss
production = None # Use production mode
cachet_url = "http://status.domain.tld/api/v1" # URL for Cachet
cachet_key = "api_key" # API Key for Cachet
sentry_url = "http://status.domain.tld/" # URL for Sentry
"""News Channel only"""
force_all = False # Force the script to replace all news files
google_maps_api_key = "api_key" # API Key for Google Maps geocoding API
geoparser_keys = ["GEOPARSER KEY", "GEOPARSER KEY"] # API Key for Geoparser
"""Forecast Channel only"""
import forecastlists
useVerbose = None # Print more verbose messages
useMultithreaded = None # Use multithreading
weathercities = ["CITY LIST", "CITY LIST"] # Lists of cities to use
cachet_elapsed_time = None # ID of the Cachet point to log elapsed time.
"""Everybody Votes Channel only"""
mysql_user = "user" # MySQL username
mysql_password = "password" # MySQL password
mysql_database = "database" # MySQL database
|
Fix PEP 8 coding violations
|
Fix PEP 8 coding violations
|
Python
|
agpl-3.0
|
RiiConnect24/File-Maker,RiiConnect24/File-Maker
|
"""Example config.py"""
webhook_urls = ["DISCORD WEBHOOK", "DISCORD WEBHOOK"] # Used to update webhooks on Discord
key_path = "/path/to/key/in/format/of/file.pem" # Private key to sign the file
file_path = "/path/to/folder" # Path to save the file to
lzss_path = "/path/to/lzss" # Path to lzss
production = None # Use production mode
cachet_url = "http://status.domain.tld/api/v1" # URL for Cachet
cachet_key = "api_key" # API Key for Cachet
sentry_url = "http://status.domain.tld/" # URL for Sentry
"""News Channel only"""
force_all = False # Force the script to replace all news files
google_maps_api_key = "api_key" # API Key for Google Maps geocoding API
geoparser_keys = ["GEOPARSER KEY", "GEOPARSER KEY"] # API Key for Geoparser
"""Forecast Channel only"""
import forecastlists
useVerbose = None # Print more verbose messages
useMultithreaded = None # Use multithreading
weathercities = ["CITY LIST", "CITY LIST"] # Lists of cities to use
cachet_elapsed_time = None # ID of the Cachet point to log elapsed time.
"""Everybody Votes Channel only"""
mysql_user = "user" # MySQL username
mysql_password = "password" # MySQL password
mysql_database = "database" # MySQL database
Fix PEP 8 coding violations
|
"""Example config.py"""
webhook_urls = ["DISCORD WEBHOOK", "DISCORD WEBHOOK"] # Used to update webhooks on Discord
key_path = "/path/to/key/in/format/of/file.pem" # Private key to sign the file
file_path = "/path/to/folder" # Path to save the file to
lzss_path = "/path/to/lzss" # Path to lzss
production = None # Use production mode
cachet_url = "http://status.domain.tld/api/v1" # URL for Cachet
cachet_key = "api_key" # API Key for Cachet
sentry_url = "http://status.domain.tld/" # URL for Sentry
"""News Channel only"""
force_all = False # Force the script to replace all news files
google_maps_api_key = "api_key" # API Key for Google Maps geocoding API
geoparser_keys = ["GEOPARSER KEY", "GEOPARSER KEY"] # API Key for Geoparser
"""Forecast Channel only"""
import forecastlists
useVerbose = None # Print more verbose messages
useMultithreaded = None # Use multithreading
weathercities = ["CITY LIST", "CITY LIST"] # Lists of cities to use
cachet_elapsed_time = None # ID of the Cachet point to log elapsed time.
"""Everybody Votes Channel only"""
mysql_user = "user" # MySQL username
mysql_password = "password" # MySQL password
mysql_database = "database" # MySQL database
|
<commit_before>"""Example config.py"""
webhook_urls = ["DISCORD WEBHOOK", "DISCORD WEBHOOK"] # Used to update webhooks on Discord
key_path = "/path/to/key/in/format/of/file.pem" # Private key to sign the file
file_path = "/path/to/folder" # Path to save the file to
lzss_path = "/path/to/lzss" # Path to lzss
production = None # Use production mode
cachet_url = "http://status.domain.tld/api/v1" # URL for Cachet
cachet_key = "api_key" # API Key for Cachet
sentry_url = "http://status.domain.tld/" # URL for Sentry
"""News Channel only"""
force_all = False # Force the script to replace all news files
google_maps_api_key = "api_key" # API Key for Google Maps geocoding API
geoparser_keys = ["GEOPARSER KEY", "GEOPARSER KEY"] # API Key for Geoparser
"""Forecast Channel only"""
import forecastlists
useVerbose = None # Print more verbose messages
useMultithreaded = None # Use multithreading
weathercities = ["CITY LIST", "CITY LIST"] # Lists of cities to use
cachet_elapsed_time = None # ID of the Cachet point to log elapsed time.
"""Everybody Votes Channel only"""
mysql_user = "user" # MySQL username
mysql_password = "password" # MySQL password
mysql_database = "database" # MySQL database
<commit_msg>Fix PEP 8 coding violations<commit_after>
|
"""Example config.py"""
webhook_urls = ["DISCORD WEBHOOK", "DISCORD WEBHOOK"] # Used to update webhooks on Discord
key_path = "/path/to/key/in/format/of/file.pem" # Private key to sign the file
file_path = "/path/to/folder" # Path to save the file to
lzss_path = "/path/to/lzss" # Path to lzss
production = None # Use production mode
cachet_url = "http://status.domain.tld/api/v1" # URL for Cachet
cachet_key = "api_key" # API Key for Cachet
sentry_url = "http://status.domain.tld/" # URL for Sentry
"""News Channel only"""
force_all = False # Force the script to replace all news files
google_maps_api_key = "api_key" # API Key for Google Maps geocoding API
geoparser_keys = ["GEOPARSER KEY", "GEOPARSER KEY"] # API Key for Geoparser
"""Forecast Channel only"""
import forecastlists
useVerbose = None # Print more verbose messages
useMultithreaded = None # Use multithreading
weathercities = ["CITY LIST", "CITY LIST"] # Lists of cities to use
cachet_elapsed_time = None # ID of the Cachet point to log elapsed time.
"""Everybody Votes Channel only"""
mysql_user = "user" # MySQL username
mysql_password = "password" # MySQL password
mysql_database = "database" # MySQL database
|
"""Example config.py"""
webhook_urls = ["DISCORD WEBHOOK", "DISCORD WEBHOOK"] # Used to update webhooks on Discord
key_path = "/path/to/key/in/format/of/file.pem" # Private key to sign the file
file_path = "/path/to/folder" # Path to save the file to
lzss_path = "/path/to/lzss" # Path to lzss
production = None # Use production mode
cachet_url = "http://status.domain.tld/api/v1" # URL for Cachet
cachet_key = "api_key" # API Key for Cachet
sentry_url = "http://status.domain.tld/" # URL for Sentry
"""News Channel only"""
force_all = False # Force the script to replace all news files
google_maps_api_key = "api_key" # API Key for Google Maps geocoding API
geoparser_keys = ["GEOPARSER KEY", "GEOPARSER KEY"] # API Key for Geoparser
"""Forecast Channel only"""
import forecastlists
useVerbose = None # Print more verbose messages
useMultithreaded = None # Use multithreading
weathercities = ["CITY LIST", "CITY LIST"] # Lists of cities to use
cachet_elapsed_time = None # ID of the Cachet point to log elapsed time.
"""Everybody Votes Channel only"""
mysql_user = "user" # MySQL username
mysql_password = "password" # MySQL password
mysql_database = "database" # MySQL database
Fix PEP 8 coding violations"""Example config.py"""
webhook_urls = ["DISCORD WEBHOOK", "DISCORD WEBHOOK"] # Used to update webhooks on Discord
key_path = "/path/to/key/in/format/of/file.pem" # Private key to sign the file
file_path = "/path/to/folder" # Path to save the file to
lzss_path = "/path/to/lzss" # Path to lzss
production = None # Use production mode
cachet_url = "http://status.domain.tld/api/v1" # URL for Cachet
cachet_key = "api_key" # API Key for Cachet
sentry_url = "http://status.domain.tld/" # URL for Sentry
"""News Channel only"""
force_all = False # Force the script to replace all news files
google_maps_api_key = "api_key" # API Key for Google Maps geocoding API
geoparser_keys = ["GEOPARSER KEY", "GEOPARSER KEY"] # API Key for Geoparser
"""Forecast Channel only"""
import forecastlists
useVerbose = None # Print more verbose messages
useMultithreaded = None # Use multithreading
weathercities = ["CITY LIST", "CITY LIST"] # Lists of cities to use
cachet_elapsed_time = None # ID of the Cachet point to log elapsed time.
"""Everybody Votes Channel only"""
mysql_user = "user" # MySQL username
mysql_password = "password" # MySQL password
mysql_database = "database" # MySQL database
|
<commit_before>"""Example config.py"""
webhook_urls = ["DISCORD WEBHOOK", "DISCORD WEBHOOK"] # Used to update webhooks on Discord
key_path = "/path/to/key/in/format/of/file.pem" # Private key to sign the file
file_path = "/path/to/folder" # Path to save the file to
lzss_path = "/path/to/lzss" # Path to lzss
production = None # Use production mode
cachet_url = "http://status.domain.tld/api/v1" # URL for Cachet
cachet_key = "api_key" # API Key for Cachet
sentry_url = "http://status.domain.tld/" # URL for Sentry
"""News Channel only"""
force_all = False # Force the script to replace all news files
google_maps_api_key = "api_key" # API Key for Google Maps geocoding API
geoparser_keys = ["GEOPARSER KEY", "GEOPARSER KEY"] # API Key for Geoparser
"""Forecast Channel only"""
import forecastlists
useVerbose = None # Print more verbose messages
useMultithreaded = None # Use multithreading
weathercities = ["CITY LIST", "CITY LIST"] # Lists of cities to use
cachet_elapsed_time = None # ID of the Cachet point to log elapsed time.
"""Everybody Votes Channel only"""
mysql_user = "user" # MySQL username
mysql_password = "password" # MySQL password
mysql_database = "database" # MySQL database
<commit_msg>Fix PEP 8 coding violations<commit_after>"""Example config.py"""
webhook_urls = ["DISCORD WEBHOOK", "DISCORD WEBHOOK"] # Used to update webhooks on Discord
key_path = "/path/to/key/in/format/of/file.pem" # Private key to sign the file
file_path = "/path/to/folder" # Path to save the file to
lzss_path = "/path/to/lzss" # Path to lzss
production = None # Use production mode
cachet_url = "http://status.domain.tld/api/v1" # URL for Cachet
cachet_key = "api_key" # API Key for Cachet
sentry_url = "http://status.domain.tld/" # URL for Sentry
"""News Channel only"""
force_all = False # Force the script to replace all news files
google_maps_api_key = "api_key" # API Key for Google Maps geocoding API
geoparser_keys = ["GEOPARSER KEY", "GEOPARSER KEY"] # API Key for Geoparser
"""Forecast Channel only"""
import forecastlists
useVerbose = None # Print more verbose messages
useMultithreaded = None # Use multithreading
weathercities = ["CITY LIST", "CITY LIST"] # Lists of cities to use
cachet_elapsed_time = None # ID of the Cachet point to log elapsed time.
"""Everybody Votes Channel only"""
mysql_user = "user" # MySQL username
mysql_password = "password" # MySQL password
mysql_database = "database" # MySQL database
|
b085d519da9869be8c4bc4f56cb0e040a6b1525b
|
build/combine.py
|
build/combine.py
|
import os, sys, re
from simplejson import load as json
from simplejson import dumps as dump
from glob import glob
VERSION = 0.1
all = []
for p in glob("../plugins/*.json"):
fp = open(p, "r")
x = json(fp, "utf-8")
x['date'] = int(os.path.getmtime(p) * 1000)
fp.close()
all += x,
fp = open("../modules/plugins.json", "w")
fp.write(dump(all))
|
import os, sys, re
try:
from simplejson import load as json
from simplejson import dumps as dump
except:
from json import load as json
from json import dumps as dump
from glob import glob
VERSION = 0.1
all = []
for p in glob("../plugins/*.json"):
fp = open(p, "r")
x = json(fp, "utf-8")
x['date'] = int(os.path.getmtime(p) * 1000)
fp.close()
all += x,
fp = open("../modules/plugins.json", "w")
fp.write(dump(all))
|
Use json standard module if simplejson is not present
|
Use json standard module if simplejson is not present
|
Python
|
mpl-2.0
|
marianocarrazana/anticontainer,downthemall/anticontainer,downthemall/anticontainer,marianocarrazana/anticontainer,downthemall/anticontainer,marianocarrazana/anticontainer
|
import os, sys, re
from simplejson import load as json
from simplejson import dumps as dump
from glob import glob
VERSION = 0.1
all = []
for p in glob("../plugins/*.json"):
fp = open(p, "r")
x = json(fp, "utf-8")
x['date'] = int(os.path.getmtime(p) * 1000)
fp.close()
all += x,
fp = open("../modules/plugins.json", "w")
fp.write(dump(all))Use json standard module if simplejson is not present
|
import os, sys, re
try:
from simplejson import load as json
from simplejson import dumps as dump
except:
from json import load as json
from json import dumps as dump
from glob import glob
VERSION = 0.1
all = []
for p in glob("../plugins/*.json"):
fp = open(p, "r")
x = json(fp, "utf-8")
x['date'] = int(os.path.getmtime(p) * 1000)
fp.close()
all += x,
fp = open("../modules/plugins.json", "w")
fp.write(dump(all))
|
<commit_before>import os, sys, re
from simplejson import load as json
from simplejson import dumps as dump
from glob import glob
VERSION = 0.1
all = []
for p in glob("../plugins/*.json"):
fp = open(p, "r")
x = json(fp, "utf-8")
x['date'] = int(os.path.getmtime(p) * 1000)
fp.close()
all += x,
fp = open("../modules/plugins.json", "w")
fp.write(dump(all))<commit_msg>Use json standard module if simplejson is not present<commit_after>
|
import os, sys, re
try:
from simplejson import load as json
from simplejson import dumps as dump
except:
from json import load as json
from json import dumps as dump
from glob import glob
VERSION = 0.1
all = []
for p in glob("../plugins/*.json"):
fp = open(p, "r")
x = json(fp, "utf-8")
x['date'] = int(os.path.getmtime(p) * 1000)
fp.close()
all += x,
fp = open("../modules/plugins.json", "w")
fp.write(dump(all))
|
import os, sys, re
from simplejson import load as json
from simplejson import dumps as dump
from glob import glob
VERSION = 0.1
all = []
for p in glob("../plugins/*.json"):
fp = open(p, "r")
x = json(fp, "utf-8")
x['date'] = int(os.path.getmtime(p) * 1000)
fp.close()
all += x,
fp = open("../modules/plugins.json", "w")
fp.write(dump(all))Use json standard module if simplejson is not presentimport os, sys, re
try:
from simplejson import load as json
from simplejson import dumps as dump
except:
from json import load as json
from json import dumps as dump
from glob import glob
VERSION = 0.1
all = []
for p in glob("../plugins/*.json"):
fp = open(p, "r")
x = json(fp, "utf-8")
x['date'] = int(os.path.getmtime(p) * 1000)
fp.close()
all += x,
fp = open("../modules/plugins.json", "w")
fp.write(dump(all))
|
<commit_before>import os, sys, re
from simplejson import load as json
from simplejson import dumps as dump
from glob import glob
VERSION = 0.1
all = []
for p in glob("../plugins/*.json"):
fp = open(p, "r")
x = json(fp, "utf-8")
x['date'] = int(os.path.getmtime(p) * 1000)
fp.close()
all += x,
fp = open("../modules/plugins.json", "w")
fp.write(dump(all))<commit_msg>Use json standard module if simplejson is not present<commit_after>import os, sys, re
try:
from simplejson import load as json
from simplejson import dumps as dump
except:
from json import load as json
from json import dumps as dump
from glob import glob
VERSION = 0.1
all = []
for p in glob("../plugins/*.json"):
fp = open(p, "r")
x = json(fp, "utf-8")
x['date'] = int(os.path.getmtime(p) * 1000)
fp.close()
all += x,
fp = open("../modules/plugins.json", "w")
fp.write(dump(all))
|
0b77033563ab85c98ca5ea9c190bcee4da5c6aef
|
sanic_sentry.py
|
sanic_sentry.py
|
import logging
import sanic
import raven
import raven_aiohttp
from raven.handlers.logging import SentryHandler
try:
from sanic.log import logger
except ImportError:
logger = logging.getLogger('sanic')
class SanicSentry:
def __init__(self, app=None):
self.app = None
self.handler = None
self.client = None
if app is not None:
self.init_app(app)
def init_app(self, app: sanic.Sanic):
self.client = raven.Client(
dsn=app.config['SENTRY_DSN'],
transport=raven_aiohttp.AioHttpTransport,
**app.config.get('SENTRY_PARAMS')
)
self.handler = SentryHandler(client=self.client, level=app.config.get('SENTRY_LEVEL', logging.ERROR))
logger.addHandler(self.handler)
self.app = app
self.app.sentry = self
|
import logging
import sanic
import raven
import raven_aiohttp
from raven.handlers.logging import SentryHandler
try:
from sanic.log import logger
except ImportError:
logger = logging.getLogger('sanic')
class SanicSentry:
def __init__(self, app=None):
self.app = None
self.handler = None
self.client = None
if app is not None:
self.init_app(app)
def init_app(self, app: sanic.Sanic):
self.client = raven.Client(
dsn=app.config['SENTRY_DSN'],
transport=raven_aiohttp.AioHttpTransport,
**app.config.get('SENTRY_PARAMS', {})
)
self.handler = SentryHandler(client=self.client, level=app.config.get('SENTRY_LEVEL', logging.ERROR))
logger.addHandler(self.handler)
self.app = app
self.app.sentry = self
|
Add a default value for SENTRY_PARAMS
|
Add a default value for SENTRY_PARAMS
|
Python
|
mit
|
serathius/sanic-sentry
|
import logging
import sanic
import raven
import raven_aiohttp
from raven.handlers.logging import SentryHandler
try:
from sanic.log import logger
except ImportError:
logger = logging.getLogger('sanic')
class SanicSentry:
def __init__(self, app=None):
self.app = None
self.handler = None
self.client = None
if app is not None:
self.init_app(app)
def init_app(self, app: sanic.Sanic):
self.client = raven.Client(
dsn=app.config['SENTRY_DSN'],
transport=raven_aiohttp.AioHttpTransport,
**app.config.get('SENTRY_PARAMS')
)
self.handler = SentryHandler(client=self.client, level=app.config.get('SENTRY_LEVEL', logging.ERROR))
logger.addHandler(self.handler)
self.app = app
self.app.sentry = self
Add a default value for SENTRY_PARAMS
|
import logging
import sanic
import raven
import raven_aiohttp
from raven.handlers.logging import SentryHandler
try:
from sanic.log import logger
except ImportError:
logger = logging.getLogger('sanic')
class SanicSentry:
def __init__(self, app=None):
self.app = None
self.handler = None
self.client = None
if app is not None:
self.init_app(app)
def init_app(self, app: sanic.Sanic):
self.client = raven.Client(
dsn=app.config['SENTRY_DSN'],
transport=raven_aiohttp.AioHttpTransport,
**app.config.get('SENTRY_PARAMS', {})
)
self.handler = SentryHandler(client=self.client, level=app.config.get('SENTRY_LEVEL', logging.ERROR))
logger.addHandler(self.handler)
self.app = app
self.app.sentry = self
|
<commit_before>import logging
import sanic
import raven
import raven_aiohttp
from raven.handlers.logging import SentryHandler
try:
from sanic.log import logger
except ImportError:
logger = logging.getLogger('sanic')
class SanicSentry:
def __init__(self, app=None):
self.app = None
self.handler = None
self.client = None
if app is not None:
self.init_app(app)
def init_app(self, app: sanic.Sanic):
self.client = raven.Client(
dsn=app.config['SENTRY_DSN'],
transport=raven_aiohttp.AioHttpTransport,
**app.config.get('SENTRY_PARAMS')
)
self.handler = SentryHandler(client=self.client, level=app.config.get('SENTRY_LEVEL', logging.ERROR))
logger.addHandler(self.handler)
self.app = app
self.app.sentry = self
<commit_msg>Add a default value for SENTRY_PARAMS<commit_after>
|
import logging
import sanic
import raven
import raven_aiohttp
from raven.handlers.logging import SentryHandler
try:
from sanic.log import logger
except ImportError:
logger = logging.getLogger('sanic')
class SanicSentry:
def __init__(self, app=None):
self.app = None
self.handler = None
self.client = None
if app is not None:
self.init_app(app)
def init_app(self, app: sanic.Sanic):
self.client = raven.Client(
dsn=app.config['SENTRY_DSN'],
transport=raven_aiohttp.AioHttpTransport,
**app.config.get('SENTRY_PARAMS', {})
)
self.handler = SentryHandler(client=self.client, level=app.config.get('SENTRY_LEVEL', logging.ERROR))
logger.addHandler(self.handler)
self.app = app
self.app.sentry = self
|
import logging
import sanic
import raven
import raven_aiohttp
from raven.handlers.logging import SentryHandler
try:
from sanic.log import logger
except ImportError:
logger = logging.getLogger('sanic')
class SanicSentry:
def __init__(self, app=None):
self.app = None
self.handler = None
self.client = None
if app is not None:
self.init_app(app)
def init_app(self, app: sanic.Sanic):
self.client = raven.Client(
dsn=app.config['SENTRY_DSN'],
transport=raven_aiohttp.AioHttpTransport,
**app.config.get('SENTRY_PARAMS')
)
self.handler = SentryHandler(client=self.client, level=app.config.get('SENTRY_LEVEL', logging.ERROR))
logger.addHandler(self.handler)
self.app = app
self.app.sentry = self
Add a default value for SENTRY_PARAMSimport logging
import sanic
import raven
import raven_aiohttp
from raven.handlers.logging import SentryHandler
try:
from sanic.log import logger
except ImportError:
logger = logging.getLogger('sanic')
class SanicSentry:
def __init__(self, app=None):
self.app = None
self.handler = None
self.client = None
if app is not None:
self.init_app(app)
def init_app(self, app: sanic.Sanic):
self.client = raven.Client(
dsn=app.config['SENTRY_DSN'],
transport=raven_aiohttp.AioHttpTransport,
**app.config.get('SENTRY_PARAMS', {})
)
self.handler = SentryHandler(client=self.client, level=app.config.get('SENTRY_LEVEL', logging.ERROR))
logger.addHandler(self.handler)
self.app = app
self.app.sentry = self
|
<commit_before>import logging
import sanic
import raven
import raven_aiohttp
from raven.handlers.logging import SentryHandler
try:
from sanic.log import logger
except ImportError:
logger = logging.getLogger('sanic')
class SanicSentry:
def __init__(self, app=None):
self.app = None
self.handler = None
self.client = None
if app is not None:
self.init_app(app)
def init_app(self, app: sanic.Sanic):
self.client = raven.Client(
dsn=app.config['SENTRY_DSN'],
transport=raven_aiohttp.AioHttpTransport,
**app.config.get('SENTRY_PARAMS')
)
self.handler = SentryHandler(client=self.client, level=app.config.get('SENTRY_LEVEL', logging.ERROR))
logger.addHandler(self.handler)
self.app = app
self.app.sentry = self
<commit_msg>Add a default value for SENTRY_PARAMS<commit_after>import logging
import sanic
import raven
import raven_aiohttp
from raven.handlers.logging import SentryHandler
try:
from sanic.log import logger
except ImportError:
logger = logging.getLogger('sanic')
class SanicSentry:
def __init__(self, app=None):
self.app = None
self.handler = None
self.client = None
if app is not None:
self.init_app(app)
def init_app(self, app: sanic.Sanic):
self.client = raven.Client(
dsn=app.config['SENTRY_DSN'],
transport=raven_aiohttp.AioHttpTransport,
**app.config.get('SENTRY_PARAMS', {})
)
self.handler = SentryHandler(client=self.client, level=app.config.get('SENTRY_LEVEL', logging.ERROR))
logger.addHandler(self.handler)
self.app = app
self.app.sentry = self
|
aaa7da2b43ab08758456c972cd2bd727082c835d
|
build/release.py
|
build/release.py
|
#!/usr/bin/env python
import os
import sys
import shutil
import subprocess
from zipfile import ZipFile
if len(sys.argv) != 2:
print 'Usage: release.py version-number'
sys.exit(1)
version = sys.argv[1]
work_dir = 'minified'
name = 'goo-' + version
# Root directory inside zip file
zip_root = name + '/'
print 'Creating release', name
if os.path.isdir(work_dir):
shutil.rmtree(work_dir)
if os.name == 'nt':
command = 'cake.cmd'
else:
command = 'cake'
subprocess.check_call([command, 'minify'])
zipfile = ZipFile(name + '.zip', 'w')
zipfile.write('COPYING', zip_root + 'COPYING')
goo_root = work_dir + '/goo'
for root, dirs, files in os.walk(goo_root):
for f in files:
filename = root[len(goo_root) + 1:] + '/' + f
zipfile.write(root + '/' + f, zip_root + filename)
zipfile.close()
|
#!/usr/bin/env python
import os
import sys
import shutil
import subprocess
from zipfile import ZipFile
def prepend(filename, to_prepend):
"""Prepends a string to a file
"""
with open(filename, 'r') as stream:
content = stream.read()
with open(filename, 'w') as stream:
stream.write(to_prepend)
stream.write(content)
if len(sys.argv) != 2:
print 'Usage: release.py version-number'
sys.exit(1)
version = sys.argv[1]
work_dir = 'minified'
name = 'goo-' + version
# Root directory inside zip file
zip_root = name + '/'
print 'Creating release', name
if os.path.isdir(work_dir):
shutil.rmtree(work_dir)
if os.name == 'nt':
command = 'cake.cmd'
else:
command = 'cake'
subprocess.check_call([command, 'minify'])
zipfile = ZipFile(name + '.zip', 'w')
zipfile.write('COPYING', zip_root + 'COPYING')
goo_root = work_dir + '/goo'
prepend(goo_root + '/goo.js',
'/*\n' +
' * Goo Engine ' + version + '\n' +
' * Copyright 2013 Goo Technologies AB\n' +
' */\n'
)
for root, dirs, files in os.walk(goo_root):
for f in files:
filename = root[len(goo_root) + 1:] + '/' + f
zipfile.write(root + '/' + f, zip_root + filename)
zipfile.close()
|
Add version number and copyright to goo.js
|
Add version number and copyright to goo.js
This is useful to keep track of which engine version the tool uses, story #294
|
Python
|
mit
|
GooTechnologies/goojs,GooTechnologies/goojs,GooTechnologies/goojs
|
#!/usr/bin/env python
import os
import sys
import shutil
import subprocess
from zipfile import ZipFile
if len(sys.argv) != 2:
print 'Usage: release.py version-number'
sys.exit(1)
version = sys.argv[1]
work_dir = 'minified'
name = 'goo-' + version
# Root directory inside zip file
zip_root = name + '/'
print 'Creating release', name
if os.path.isdir(work_dir):
shutil.rmtree(work_dir)
if os.name == 'nt':
command = 'cake.cmd'
else:
command = 'cake'
subprocess.check_call([command, 'minify'])
zipfile = ZipFile(name + '.zip', 'w')
zipfile.write('COPYING', zip_root + 'COPYING')
goo_root = work_dir + '/goo'
for root, dirs, files in os.walk(goo_root):
for f in files:
filename = root[len(goo_root) + 1:] + '/' + f
zipfile.write(root + '/' + f, zip_root + filename)
zipfile.close()Add version number and copyright to goo.js
This is useful to keep track of which engine version the tool uses, story #294
|
#!/usr/bin/env python
import os
import sys
import shutil
import subprocess
from zipfile import ZipFile
def prepend(filename, to_prepend):
"""Prepends a string to a file
"""
with open(filename, 'r') as stream:
content = stream.read()
with open(filename, 'w') as stream:
stream.write(to_prepend)
stream.write(content)
if len(sys.argv) != 2:
print 'Usage: release.py version-number'
sys.exit(1)
version = sys.argv[1]
work_dir = 'minified'
name = 'goo-' + version
# Root directory inside zip file
zip_root = name + '/'
print 'Creating release', name
if os.path.isdir(work_dir):
shutil.rmtree(work_dir)
if os.name == 'nt':
command = 'cake.cmd'
else:
command = 'cake'
subprocess.check_call([command, 'minify'])
zipfile = ZipFile(name + '.zip', 'w')
zipfile.write('COPYING', zip_root + 'COPYING')
goo_root = work_dir + '/goo'
prepend(goo_root + '/goo.js',
'/*\n' +
' * Goo Engine ' + version + '\n' +
' * Copyright 2013 Goo Technologies AB\n' +
' */\n'
)
for root, dirs, files in os.walk(goo_root):
for f in files:
filename = root[len(goo_root) + 1:] + '/' + f
zipfile.write(root + '/' + f, zip_root + filename)
zipfile.close()
|
<commit_before>#!/usr/bin/env python
import os
import sys
import shutil
import subprocess
from zipfile import ZipFile
if len(sys.argv) != 2:
print 'Usage: release.py version-number'
sys.exit(1)
version = sys.argv[1]
work_dir = 'minified'
name = 'goo-' + version
# Root directory inside zip file
zip_root = name + '/'
print 'Creating release', name
if os.path.isdir(work_dir):
shutil.rmtree(work_dir)
if os.name == 'nt':
command = 'cake.cmd'
else:
command = 'cake'
subprocess.check_call([command, 'minify'])
zipfile = ZipFile(name + '.zip', 'w')
zipfile.write('COPYING', zip_root + 'COPYING')
goo_root = work_dir + '/goo'
for root, dirs, files in os.walk(goo_root):
for f in files:
filename = root[len(goo_root) + 1:] + '/' + f
zipfile.write(root + '/' + f, zip_root + filename)
zipfile.close()<commit_msg>Add version number and copyright to goo.js
This is useful to keep track of which engine version the tool uses, story #294<commit_after>
|
#!/usr/bin/env python
import os
import sys
import shutil
import subprocess
from zipfile import ZipFile
def prepend(filename, to_prepend):
"""Prepends a string to a file
"""
with open(filename, 'r') as stream:
content = stream.read()
with open(filename, 'w') as stream:
stream.write(to_prepend)
stream.write(content)
if len(sys.argv) != 2:
print 'Usage: release.py version-number'
sys.exit(1)
version = sys.argv[1]
work_dir = 'minified'
name = 'goo-' + version
# Root directory inside zip file
zip_root = name + '/'
print 'Creating release', name
if os.path.isdir(work_dir):
shutil.rmtree(work_dir)
if os.name == 'nt':
command = 'cake.cmd'
else:
command = 'cake'
subprocess.check_call([command, 'minify'])
zipfile = ZipFile(name + '.zip', 'w')
zipfile.write('COPYING', zip_root + 'COPYING')
goo_root = work_dir + '/goo'
prepend(goo_root + '/goo.js',
'/*\n' +
' * Goo Engine ' + version + '\n' +
' * Copyright 2013 Goo Technologies AB\n' +
' */\n'
)
for root, dirs, files in os.walk(goo_root):
for f in files:
filename = root[len(goo_root) + 1:] + '/' + f
zipfile.write(root + '/' + f, zip_root + filename)
zipfile.close()
|
#!/usr/bin/env python
import os
import sys
import shutil
import subprocess
from zipfile import ZipFile
if len(sys.argv) != 2:
print 'Usage: release.py version-number'
sys.exit(1)
version = sys.argv[1]
work_dir = 'minified'
name = 'goo-' + version
# Root directory inside zip file
zip_root = name + '/'
print 'Creating release', name
if os.path.isdir(work_dir):
shutil.rmtree(work_dir)
if os.name == 'nt':
command = 'cake.cmd'
else:
command = 'cake'
subprocess.check_call([command, 'minify'])
zipfile = ZipFile(name + '.zip', 'w')
zipfile.write('COPYING', zip_root + 'COPYING')
goo_root = work_dir + '/goo'
for root, dirs, files in os.walk(goo_root):
for f in files:
filename = root[len(goo_root) + 1:] + '/' + f
zipfile.write(root + '/' + f, zip_root + filename)
zipfile.close()Add version number and copyright to goo.js
This is useful to keep track of which engine version the tool uses, story #294#!/usr/bin/env python
import os
import sys
import shutil
import subprocess
from zipfile import ZipFile
def prepend(filename, to_prepend):
"""Prepends a string to a file
"""
with open(filename, 'r') as stream:
content = stream.read()
with open(filename, 'w') as stream:
stream.write(to_prepend)
stream.write(content)
if len(sys.argv) != 2:
print 'Usage: release.py version-number'
sys.exit(1)
version = sys.argv[1]
work_dir = 'minified'
name = 'goo-' + version
# Root directory inside zip file
zip_root = name + '/'
print 'Creating release', name
if os.path.isdir(work_dir):
shutil.rmtree(work_dir)
if os.name == 'nt':
command = 'cake.cmd'
else:
command = 'cake'
subprocess.check_call([command, 'minify'])
zipfile = ZipFile(name + '.zip', 'w')
zipfile.write('COPYING', zip_root + 'COPYING')
goo_root = work_dir + '/goo'
prepend(goo_root + '/goo.js',
'/*\n' +
' * Goo Engine ' + version + '\n' +
' * Copyright 2013 Goo Technologies AB\n' +
' */\n'
)
for root, dirs, files in os.walk(goo_root):
for f in files:
filename = root[len(goo_root) + 1:] + '/' + f
zipfile.write(root + '/' + f, zip_root + filename)
zipfile.close()
|
<commit_before>#!/usr/bin/env python
import os
import sys
import shutil
import subprocess
from zipfile import ZipFile
if len(sys.argv) != 2:
print 'Usage: release.py version-number'
sys.exit(1)
version = sys.argv[1]
work_dir = 'minified'
name = 'goo-' + version
# Root directory inside zip file
zip_root = name + '/'
print 'Creating release', name
if os.path.isdir(work_dir):
shutil.rmtree(work_dir)
if os.name == 'nt':
command = 'cake.cmd'
else:
command = 'cake'
subprocess.check_call([command, 'minify'])
zipfile = ZipFile(name + '.zip', 'w')
zipfile.write('COPYING', zip_root + 'COPYING')
goo_root = work_dir + '/goo'
for root, dirs, files in os.walk(goo_root):
for f in files:
filename = root[len(goo_root) + 1:] + '/' + f
zipfile.write(root + '/' + f, zip_root + filename)
zipfile.close()<commit_msg>Add version number and copyright to goo.js
This is useful to keep track of which engine version the tool uses, story #294<commit_after>#!/usr/bin/env python
import os
import sys
import shutil
import subprocess
from zipfile import ZipFile
def prepend(filename, to_prepend):
"""Prepends a string to a file
"""
with open(filename, 'r') as stream:
content = stream.read()
with open(filename, 'w') as stream:
stream.write(to_prepend)
stream.write(content)
if len(sys.argv) != 2:
print 'Usage: release.py version-number'
sys.exit(1)
version = sys.argv[1]
work_dir = 'minified'
name = 'goo-' + version
# Root directory inside zip file
zip_root = name + '/'
print 'Creating release', name
if os.path.isdir(work_dir):
shutil.rmtree(work_dir)
if os.name == 'nt':
command = 'cake.cmd'
else:
command = 'cake'
subprocess.check_call([command, 'minify'])
zipfile = ZipFile(name + '.zip', 'w')
zipfile.write('COPYING', zip_root + 'COPYING')
goo_root = work_dir + '/goo'
prepend(goo_root + '/goo.js',
'/*\n' +
' * Goo Engine ' + version + '\n' +
' * Copyright 2013 Goo Technologies AB\n' +
' */\n'
)
for root, dirs, files in os.walk(goo_root):
for f in files:
filename = root[len(goo_root) + 1:] + '/' + f
zipfile.write(root + '/' + f, zip_root + filename)
zipfile.close()
|
1c3f89110ede8998b63831c181c44e92709481b6
|
demo/widgy.py
|
demo/widgy.py
|
from __future__ import absolute_import
from widgy.site import WidgySite
class DemoWidgySite(WidgySite):
def valid_parent_of(self, parent, child_class, obj=None):
if isinstance(parent, I18NLayout):
return True
else:
return super(DemoWidgySite, self).valid_parent_of(parent, child_class, obj)
widgy_site = DemoWidgySite()
from widgy.contrib.widgy_i18n.models import I18NLayout
|
from __future__ import absolute_import
from widgy.site import ReviewedWidgySite
class DemoWidgySite(ReviewedWidgySite):
def valid_parent_of(self, parent, child_class, obj=None):
if isinstance(parent, I18NLayout):
return True
else:
return super(DemoWidgySite, self).valid_parent_of(parent, child_class, obj)
widgy_site = DemoWidgySite()
from widgy.contrib.widgy_i18n.models import I18NLayout
|
Enable the review queue on the demo site
|
Enable the review queue on the demo site
|
Python
|
apache-2.0
|
j00bar/django-widgy,j00bar/django-widgy,j00bar/django-widgy
|
from __future__ import absolute_import
from widgy.site import WidgySite
class DemoWidgySite(WidgySite):
def valid_parent_of(self, parent, child_class, obj=None):
if isinstance(parent, I18NLayout):
return True
else:
return super(DemoWidgySite, self).valid_parent_of(parent, child_class, obj)
widgy_site = DemoWidgySite()
from widgy.contrib.widgy_i18n.models import I18NLayout
Enable the review queue on the demo site
|
from __future__ import absolute_import
from widgy.site import ReviewedWidgySite
class DemoWidgySite(ReviewedWidgySite):
def valid_parent_of(self, parent, child_class, obj=None):
if isinstance(parent, I18NLayout):
return True
else:
return super(DemoWidgySite, self).valid_parent_of(parent, child_class, obj)
widgy_site = DemoWidgySite()
from widgy.contrib.widgy_i18n.models import I18NLayout
|
<commit_before>from __future__ import absolute_import
from widgy.site import WidgySite
class DemoWidgySite(WidgySite):
def valid_parent_of(self, parent, child_class, obj=None):
if isinstance(parent, I18NLayout):
return True
else:
return super(DemoWidgySite, self).valid_parent_of(parent, child_class, obj)
widgy_site = DemoWidgySite()
from widgy.contrib.widgy_i18n.models import I18NLayout
<commit_msg>Enable the review queue on the demo site<commit_after>
|
from __future__ import absolute_import
from widgy.site import ReviewedWidgySite
class DemoWidgySite(ReviewedWidgySite):
def valid_parent_of(self, parent, child_class, obj=None):
if isinstance(parent, I18NLayout):
return True
else:
return super(DemoWidgySite, self).valid_parent_of(parent, child_class, obj)
widgy_site = DemoWidgySite()
from widgy.contrib.widgy_i18n.models import I18NLayout
|
from __future__ import absolute_import
from widgy.site import WidgySite
class DemoWidgySite(WidgySite):
def valid_parent_of(self, parent, child_class, obj=None):
if isinstance(parent, I18NLayout):
return True
else:
return super(DemoWidgySite, self).valid_parent_of(parent, child_class, obj)
widgy_site = DemoWidgySite()
from widgy.contrib.widgy_i18n.models import I18NLayout
Enable the review queue on the demo sitefrom __future__ import absolute_import
from widgy.site import ReviewedWidgySite
class DemoWidgySite(ReviewedWidgySite):
def valid_parent_of(self, parent, child_class, obj=None):
if isinstance(parent, I18NLayout):
return True
else:
return super(DemoWidgySite, self).valid_parent_of(parent, child_class, obj)
widgy_site = DemoWidgySite()
from widgy.contrib.widgy_i18n.models import I18NLayout
|
<commit_before>from __future__ import absolute_import
from widgy.site import WidgySite
class DemoWidgySite(WidgySite):
def valid_parent_of(self, parent, child_class, obj=None):
if isinstance(parent, I18NLayout):
return True
else:
return super(DemoWidgySite, self).valid_parent_of(parent, child_class, obj)
widgy_site = DemoWidgySite()
from widgy.contrib.widgy_i18n.models import I18NLayout
<commit_msg>Enable the review queue on the demo site<commit_after>from __future__ import absolute_import
from widgy.site import ReviewedWidgySite
class DemoWidgySite(ReviewedWidgySite):
def valid_parent_of(self, parent, child_class, obj=None):
if isinstance(parent, I18NLayout):
return True
else:
return super(DemoWidgySite, self).valid_parent_of(parent, child_class, obj)
widgy_site = DemoWidgySite()
from widgy.contrib.widgy_i18n.models import I18NLayout
|
ea09470ebdd69af2fa1d7d07d7b04fe3ff857987
|
raffle.py
|
raffle.py
|
"""
St. George Game
raffle.py
Sage Berg
Created: 9 Dec 2014
"""
from random import randint
class Raffle(object):
"""
Raffle contains a list of action objects, one of which
will be chosen and shown to the player.
"""
def __init__(self):
self.options = dict() # Maps options to weights
def add(self, option, weight=1):
"""
"""
if option in self.options:
self.options[option] += weight
else:
self.options[option] = weight
def get(self):
"""
chooses one action from the bag and returns it.
"""
total_weights = 0
for weight in self.options.values():
total_weights += weight
roll = randint(0, total_weights)
for option, weight in self.options.items():
if roll <= weight:
return option
else:
roll -= weight
def merge(self, other):
"""
Merge the contents of another Raffle with this Raffle.
"""
for option, weight in other.options.items():
self.add(option, weight)
|
"""
St. George Game
raffle.py
Sage Berg
Created: 9 Dec 2014
"""
from random import randint
class Raffle(object):
"""
Raffle contains a list of action objects, one of which
will be chosen and shown to the player.
"""
def __init__(self):
self.options = dict() # Maps options to weights
def add(self, option, weight=1):
"""
"""
if option in self.options:
self.options[option] += weight
else:
self.options[option] = weight
def get(self):
"""
chooses one action from the bag and returns it.
"""
total_weights = 0
for weight in self.options.values():
total_weights += weight
roll = randint(0, total_weights)
for option, weight in self.options.items():
if roll <= weight:
return option
else:
roll -= weight
def merge(self, other):
"""
Merge the contents of another Raffle with this Raffle.
"""
for option, weight in other.options.items():
self.add(option, weight)
def __len__(self):
"""
Return the combined weight of all options in the Raffle.
"""
total = 0
for _, weight in self.options.items():
total += weight
return total
|
Add length method to Raffle
|
Add length method to Raffle
|
Python
|
apache-2.0
|
SageBerg/St.GeorgeGame,SageBerg/St.GeorgeGame,SageBerg/St.GeorgeGame,SageBerg/St.GeorgeGame
|
"""
St. George Game
raffle.py
Sage Berg
Created: 9 Dec 2014
"""
from random import randint
class Raffle(object):
"""
Raffle contains a list of action objects, one of which
will be chosen and shown to the player.
"""
def __init__(self):
self.options = dict() # Maps options to weights
def add(self, option, weight=1):
"""
"""
if option in self.options:
self.options[option] += weight
else:
self.options[option] = weight
def get(self):
"""
chooses one action from the bag and returns it.
"""
total_weights = 0
for weight in self.options.values():
total_weights += weight
roll = randint(0, total_weights)
for option, weight in self.options.items():
if roll <= weight:
return option
else:
roll -= weight
def merge(self, other):
"""
Merge the contents of another Raffle with this Raffle.
"""
for option, weight in other.options.items():
self.add(option, weight)
Add length method to Raffle
|
"""
St. George Game
raffle.py
Sage Berg
Created: 9 Dec 2014
"""
from random import randint
class Raffle(object):
"""
Raffle contains a list of action objects, one of which
will be chosen and shown to the player.
"""
def __init__(self):
self.options = dict() # Maps options to weights
def add(self, option, weight=1):
"""
"""
if option in self.options:
self.options[option] += weight
else:
self.options[option] = weight
def get(self):
"""
chooses one action from the bag and returns it.
"""
total_weights = 0
for weight in self.options.values():
total_weights += weight
roll = randint(0, total_weights)
for option, weight in self.options.items():
if roll <= weight:
return option
else:
roll -= weight
def merge(self, other):
"""
Merge the contents of another Raffle with this Raffle.
"""
for option, weight in other.options.items():
self.add(option, weight)
def __len__(self):
"""
Return the combined weight of all options in the Raffle.
"""
total = 0
for _, weight in self.options.items():
total += weight
return total
|
<commit_before>"""
St. George Game
raffle.py
Sage Berg
Created: 9 Dec 2014
"""
from random import randint
class Raffle(object):
"""
Raffle contains a list of action objects, one of which
will be chosen and shown to the player.
"""
def __init__(self):
self.options = dict() # Maps options to weights
def add(self, option, weight=1):
"""
"""
if option in self.options:
self.options[option] += weight
else:
self.options[option] = weight
def get(self):
"""
chooses one action from the bag and returns it.
"""
total_weights = 0
for weight in self.options.values():
total_weights += weight
roll = randint(0, total_weights)
for option, weight in self.options.items():
if roll <= weight:
return option
else:
roll -= weight
def merge(self, other):
"""
Merge the contents of another Raffle with this Raffle.
"""
for option, weight in other.options.items():
self.add(option, weight)
<commit_msg>Add length method to Raffle<commit_after>
|
"""
St. George Game
raffle.py
Sage Berg
Created: 9 Dec 2014
"""
from random import randint
class Raffle(object):
"""
Raffle contains a list of action objects, one of which
will be chosen and shown to the player.
"""
def __init__(self):
self.options = dict() # Maps options to weights
def add(self, option, weight=1):
"""
"""
if option in self.options:
self.options[option] += weight
else:
self.options[option] = weight
def get(self):
"""
chooses one action from the bag and returns it.
"""
total_weights = 0
for weight in self.options.values():
total_weights += weight
roll = randint(0, total_weights)
for option, weight in self.options.items():
if roll <= weight:
return option
else:
roll -= weight
def merge(self, other):
"""
Merge the contents of another Raffle with this Raffle.
"""
for option, weight in other.options.items():
self.add(option, weight)
def __len__(self):
"""
Return the combined weight of all options in the Raffle.
"""
total = 0
for _, weight in self.options.items():
total += weight
return total
|
"""
St. George Game
raffle.py
Sage Berg
Created: 9 Dec 2014
"""
from random import randint
class Raffle(object):
"""
Raffle contains a list of action objects, one of which
will be chosen and shown to the player.
"""
def __init__(self):
self.options = dict() # Maps options to weights
def add(self, option, weight=1):
"""
"""
if option in self.options:
self.options[option] += weight
else:
self.options[option] = weight
def get(self):
"""
chooses one action from the bag and returns it.
"""
total_weights = 0
for weight in self.options.values():
total_weights += weight
roll = randint(0, total_weights)
for option, weight in self.options.items():
if roll <= weight:
return option
else:
roll -= weight
def merge(self, other):
"""
Merge the contents of another Raffle with this Raffle.
"""
for option, weight in other.options.items():
self.add(option, weight)
Add length method to Raffle"""
St. George Game
raffle.py
Sage Berg
Created: 9 Dec 2014
"""
from random import randint
class Raffle(object):
"""
Raffle contains a list of action objects, one of which
will be chosen and shown to the player.
"""
def __init__(self):
self.options = dict() # Maps options to weights
def add(self, option, weight=1):
"""
"""
if option in self.options:
self.options[option] += weight
else:
self.options[option] = weight
def get(self):
"""
chooses one action from the bag and returns it.
"""
total_weights = 0
for weight in self.options.values():
total_weights += weight
roll = randint(0, total_weights)
for option, weight in self.options.items():
if roll <= weight:
return option
else:
roll -= weight
def merge(self, other):
"""
Merge the contents of another Raffle with this Raffle.
"""
for option, weight in other.options.items():
self.add(option, weight)
def __len__(self):
"""
Return the combined weight of all options in the Raffle.
"""
total = 0
for _, weight in self.options.items():
total += weight
return total
|
<commit_before>"""
St. George Game
raffle.py
Sage Berg
Created: 9 Dec 2014
"""
from random import randint
class Raffle(object):
"""
Raffle contains a list of action objects, one of which
will be chosen and shown to the player.
"""
def __init__(self):
self.options = dict() # Maps options to weights
def add(self, option, weight=1):
"""
"""
if option in self.options:
self.options[option] += weight
else:
self.options[option] = weight
def get(self):
"""
chooses one action from the bag and returns it.
"""
total_weights = 0
for weight in self.options.values():
total_weights += weight
roll = randint(0, total_weights)
for option, weight in self.options.items():
if roll <= weight:
return option
else:
roll -= weight
def merge(self, other):
"""
Merge the contents of another Raffle with this Raffle.
"""
for option, weight in other.options.items():
self.add(option, weight)
<commit_msg>Add length method to Raffle<commit_after>"""
St. George Game
raffle.py
Sage Berg
Created: 9 Dec 2014
"""
from random import randint
class Raffle(object):
"""
Raffle contains a list of action objects, one of which
will be chosen and shown to the player.
"""
def __init__(self):
self.options = dict() # Maps options to weights
def add(self, option, weight=1):
"""
"""
if option in self.options:
self.options[option] += weight
else:
self.options[option] = weight
def get(self):
"""
chooses one action from the bag and returns it.
"""
total_weights = 0
for weight in self.options.values():
total_weights += weight
roll = randint(0, total_weights)
for option, weight in self.options.items():
if roll <= weight:
return option
else:
roll -= weight
def merge(self, other):
"""
Merge the contents of another Raffle with this Raffle.
"""
for option, weight in other.options.items():
self.add(option, weight)
def __len__(self):
"""
Return the combined weight of all options in the Raffle.
"""
total = 0
for _, weight in self.options.items():
total += weight
return total
|
616e9727397853e8d8f8de5b2c040c99c91e4a50
|
gen_settings.py
|
gen_settings.py
|
import os
settings = os.path.join(os.path.dirname(__file__),'lib','mapnik_settings.js')
# this goes into a mapnik_settings.js file beside the C++ _mapnik.node
settings_template = """
module.exports.paths = {
'fonts': %s,
'input_plugins': %s
};
"""
def write_mapnik_settings(fonts='undefined',input_plugins='undefined'):
global settings_template
if '__dirname' in fonts or '__dirname' in input_plugins:
settings_template = "var path = require('path');\n" + settings_template
open(settings,'w').write(settings_template % (fonts,input_plugins))
if __name__ == '__main__':
settings_dict = {}
# settings for fonts and input plugins
settings_dict['input_plugins'] = '\'%s\'' % os.popen("mapnik-config --input-plugins").readline().strip()
settings_dict['fonts'] = '\'%s\'' % os.popen("mapnik-config --fonts").readline().strip()
write_mapnik_settings(**settings_dict)
|
import os
settings = os.path.join(os.path.dirname(__file__),'lib','mapnik_settings.js')
# this goes into a mapnik_settings.js file beside the C++ _mapnik.node
settings_template = """
module.exports.paths = {
'fonts': %s,
'input_plugins': %s
};
"""
def write_mapnik_settings(fonts='undefined',input_plugins='undefined'):
global settings_template
if '__dirname' in fonts or '__dirname' in input_plugins:
settings_template = "var path = require('path');\n" + settings_template
open(settings,'w').write(settings_template % (fonts,input_plugins))
if __name__ == '__main__':
settings_dict = {}
# settings for fonts and input plugins
if os.environ.has_key('MAPNIK_INPUT_PLUGINS_DIRECTORY'):
settings_dict['input_plugins'] = os.environ['MAPNIK_INPUT_PLUGINS_DIRECTORY']
else:
settings_dict['input_plugins'] = '\'%s\'' % os.popen("mapnik-config --input-plugins").readline().strip()
if os.environ.has_key('MAPNIK_FONT_DIRECTORY'):
settings_dict['fonts'] = os.environ['MAPNIK_FONT_DIRECTORY']
else:
settings_dict['fonts'] = '\'%s\'' % os.popen("mapnik-config --fonts").readline().strip()
write_mapnik_settings(**settings_dict)
|
Revert "stop reading fonts/input plugins from environ as we now have a working mapnik-config.bat on windows"
|
Revert "stop reading fonts/input plugins from environ as we now have a working mapnik-config.bat on windows"
This reverts commit d87c71142ba7bcc0d99d84886f3534dea7617b0c.
|
Python
|
bsd-3-clause
|
mapnik/node-mapnik,langateam/node-mapnik,mojodna/node-mapnik,CartoDB/node-mapnik,CartoDB/node-mapnik,MaxSem/node-mapnik,gravitystorm/node-mapnik,tomhughes/node-mapnik,mojodna/node-mapnik,tomhughes/node-mapnik,CartoDB/node-mapnik,Uli1/node-mapnik,mojodna/node-mapnik,stefanklug/node-mapnik,CartoDB/node-mapnik,langateam/node-mapnik,gravitystorm/node-mapnik,Uli1/node-mapnik,Uli1/node-mapnik,MaxSem/node-mapnik,CartoDB/node-mapnik,mapnik/node-mapnik,tomhughes/node-mapnik,stefanklug/node-mapnik,mapnik/node-mapnik,mojodna/node-mapnik,langateam/node-mapnik,MaxSem/node-mapnik,tomhughes/node-mapnik,gravitystorm/node-mapnik,langateam/node-mapnik,gravitystorm/node-mapnik,Uli1/node-mapnik,stefanklug/node-mapnik,tomhughes/node-mapnik,mapnik/node-mapnik,stefanklug/node-mapnik,MaxSem/node-mapnik,mapnik/node-mapnik,langateam/node-mapnik
|
import os
settings = os.path.join(os.path.dirname(__file__),'lib','mapnik_settings.js')
# this goes into a mapnik_settings.js file beside the C++ _mapnik.node
settings_template = """
module.exports.paths = {
'fonts': %s,
'input_plugins': %s
};
"""
def write_mapnik_settings(fonts='undefined',input_plugins='undefined'):
global settings_template
if '__dirname' in fonts or '__dirname' in input_plugins:
settings_template = "var path = require('path');\n" + settings_template
open(settings,'w').write(settings_template % (fonts,input_plugins))
if __name__ == '__main__':
settings_dict = {}
# settings for fonts and input plugins
settings_dict['input_plugins'] = '\'%s\'' % os.popen("mapnik-config --input-plugins").readline().strip()
settings_dict['fonts'] = '\'%s\'' % os.popen("mapnik-config --fonts").readline().strip()
write_mapnik_settings(**settings_dict)Revert "stop reading fonts/input plugins from environ as we now have a working mapnik-config.bat on windows"
This reverts commit d87c71142ba7bcc0d99d84886f3534dea7617b0c.
|
import os
settings = os.path.join(os.path.dirname(__file__),'lib','mapnik_settings.js')
# this goes into a mapnik_settings.js file beside the C++ _mapnik.node
settings_template = """
module.exports.paths = {
'fonts': %s,
'input_plugins': %s
};
"""
def write_mapnik_settings(fonts='undefined',input_plugins='undefined'):
global settings_template
if '__dirname' in fonts or '__dirname' in input_plugins:
settings_template = "var path = require('path');\n" + settings_template
open(settings,'w').write(settings_template % (fonts,input_plugins))
if __name__ == '__main__':
settings_dict = {}
# settings for fonts and input plugins
if os.environ.has_key('MAPNIK_INPUT_PLUGINS_DIRECTORY'):
settings_dict['input_plugins'] = os.environ['MAPNIK_INPUT_PLUGINS_DIRECTORY']
else:
settings_dict['input_plugins'] = '\'%s\'' % os.popen("mapnik-config --input-plugins").readline().strip()
if os.environ.has_key('MAPNIK_FONT_DIRECTORY'):
settings_dict['fonts'] = os.environ['MAPNIK_FONT_DIRECTORY']
else:
settings_dict['fonts'] = '\'%s\'' % os.popen("mapnik-config --fonts").readline().strip()
write_mapnik_settings(**settings_dict)
|
<commit_before>import os
settings = os.path.join(os.path.dirname(__file__),'lib','mapnik_settings.js')
# this goes into a mapnik_settings.js file beside the C++ _mapnik.node
settings_template = """
module.exports.paths = {
'fonts': %s,
'input_plugins': %s
};
"""
def write_mapnik_settings(fonts='undefined',input_plugins='undefined'):
global settings_template
if '__dirname' in fonts or '__dirname' in input_plugins:
settings_template = "var path = require('path');\n" + settings_template
open(settings,'w').write(settings_template % (fonts,input_plugins))
if __name__ == '__main__':
settings_dict = {}
# settings for fonts and input plugins
settings_dict['input_plugins'] = '\'%s\'' % os.popen("mapnik-config --input-plugins").readline().strip()
settings_dict['fonts'] = '\'%s\'' % os.popen("mapnik-config --fonts").readline().strip()
write_mapnik_settings(**settings_dict)<commit_msg>Revert "stop reading fonts/input plugins from environ as we now have a working mapnik-config.bat on windows"
This reverts commit d87c71142ba7bcc0d99d84886f3534dea7617b0c.<commit_after>
|
import os
settings = os.path.join(os.path.dirname(__file__),'lib','mapnik_settings.js')
# this goes into a mapnik_settings.js file beside the C++ _mapnik.node
settings_template = """
module.exports.paths = {
'fonts': %s,
'input_plugins': %s
};
"""
def write_mapnik_settings(fonts='undefined',input_plugins='undefined'):
global settings_template
if '__dirname' in fonts or '__dirname' in input_plugins:
settings_template = "var path = require('path');\n" + settings_template
open(settings,'w').write(settings_template % (fonts,input_plugins))
if __name__ == '__main__':
settings_dict = {}
# settings for fonts and input plugins
if os.environ.has_key('MAPNIK_INPUT_PLUGINS_DIRECTORY'):
settings_dict['input_plugins'] = os.environ['MAPNIK_INPUT_PLUGINS_DIRECTORY']
else:
settings_dict['input_plugins'] = '\'%s\'' % os.popen("mapnik-config --input-plugins").readline().strip()
if os.environ.has_key('MAPNIK_FONT_DIRECTORY'):
settings_dict['fonts'] = os.environ['MAPNIK_FONT_DIRECTORY']
else:
settings_dict['fonts'] = '\'%s\'' % os.popen("mapnik-config --fonts").readline().strip()
write_mapnik_settings(**settings_dict)
|
import os
settings = os.path.join(os.path.dirname(__file__),'lib','mapnik_settings.js')
# this goes into a mapnik_settings.js file beside the C++ _mapnik.node
settings_template = """
module.exports.paths = {
'fonts': %s,
'input_plugins': %s
};
"""
def write_mapnik_settings(fonts='undefined',input_plugins='undefined'):
global settings_template
if '__dirname' in fonts or '__dirname' in input_plugins:
settings_template = "var path = require('path');\n" + settings_template
open(settings,'w').write(settings_template % (fonts,input_plugins))
if __name__ == '__main__':
settings_dict = {}
# settings for fonts and input plugins
settings_dict['input_plugins'] = '\'%s\'' % os.popen("mapnik-config --input-plugins").readline().strip()
settings_dict['fonts'] = '\'%s\'' % os.popen("mapnik-config --fonts").readline().strip()
write_mapnik_settings(**settings_dict)Revert "stop reading fonts/input plugins from environ as we now have a working mapnik-config.bat on windows"
This reverts commit d87c71142ba7bcc0d99d84886f3534dea7617b0c.import os
settings = os.path.join(os.path.dirname(__file__),'lib','mapnik_settings.js')
# this goes into a mapnik_settings.js file beside the C++ _mapnik.node
settings_template = """
module.exports.paths = {
'fonts': %s,
'input_plugins': %s
};
"""
def write_mapnik_settings(fonts='undefined',input_plugins='undefined'):
global settings_template
if '__dirname' in fonts or '__dirname' in input_plugins:
settings_template = "var path = require('path');\n" + settings_template
open(settings,'w').write(settings_template % (fonts,input_plugins))
if __name__ == '__main__':
settings_dict = {}
# settings for fonts and input plugins
if os.environ.has_key('MAPNIK_INPUT_PLUGINS_DIRECTORY'):
settings_dict['input_plugins'] = os.environ['MAPNIK_INPUT_PLUGINS_DIRECTORY']
else:
settings_dict['input_plugins'] = '\'%s\'' % os.popen("mapnik-config --input-plugins").readline().strip()
if os.environ.has_key('MAPNIK_FONT_DIRECTORY'):
settings_dict['fonts'] = os.environ['MAPNIK_FONT_DIRECTORY']
else:
settings_dict['fonts'] = '\'%s\'' % os.popen("mapnik-config --fonts").readline().strip()
write_mapnik_settings(**settings_dict)
|
<commit_before>import os
settings = os.path.join(os.path.dirname(__file__),'lib','mapnik_settings.js')
# this goes into a mapnik_settings.js file beside the C++ _mapnik.node
settings_template = """
module.exports.paths = {
'fonts': %s,
'input_plugins': %s
};
"""
def write_mapnik_settings(fonts='undefined',input_plugins='undefined'):
global settings_template
if '__dirname' in fonts or '__dirname' in input_plugins:
settings_template = "var path = require('path');\n" + settings_template
open(settings,'w').write(settings_template % (fonts,input_plugins))
if __name__ == '__main__':
settings_dict = {}
# settings for fonts and input plugins
settings_dict['input_plugins'] = '\'%s\'' % os.popen("mapnik-config --input-plugins").readline().strip()
settings_dict['fonts'] = '\'%s\'' % os.popen("mapnik-config --fonts").readline().strip()
write_mapnik_settings(**settings_dict)<commit_msg>Revert "stop reading fonts/input plugins from environ as we now have a working mapnik-config.bat on windows"
This reverts commit d87c71142ba7bcc0d99d84886f3534dea7617b0c.<commit_after>import os
settings = os.path.join(os.path.dirname(__file__),'lib','mapnik_settings.js')
# this goes into a mapnik_settings.js file beside the C++ _mapnik.node
settings_template = """
module.exports.paths = {
'fonts': %s,
'input_plugins': %s
};
"""
def write_mapnik_settings(fonts='undefined',input_plugins='undefined'):
global settings_template
if '__dirname' in fonts or '__dirname' in input_plugins:
settings_template = "var path = require('path');\n" + settings_template
open(settings,'w').write(settings_template % (fonts,input_plugins))
if __name__ == '__main__':
settings_dict = {}
# settings for fonts and input plugins
if os.environ.has_key('MAPNIK_INPUT_PLUGINS_DIRECTORY'):
settings_dict['input_plugins'] = os.environ['MAPNIK_INPUT_PLUGINS_DIRECTORY']
else:
settings_dict['input_plugins'] = '\'%s\'' % os.popen("mapnik-config --input-plugins").readline().strip()
if os.environ.has_key('MAPNIK_FONT_DIRECTORY'):
settings_dict['fonts'] = os.environ['MAPNIK_FONT_DIRECTORY']
else:
settings_dict['fonts'] = '\'%s\'' % os.popen("mapnik-config --fonts").readline().strip()
write_mapnik_settings(**settings_dict)
|
bd3473a8514e6d323dd03174ce65ecf278fa3772
|
groups/admin.py
|
groups/admin.py
|
from django.contrib import admin
from .models import Discussion, Group
class GroupAdmin(admin.ModelAdmin):
filter_horizontal = ('moderators', 'watchers', 'members_if_private')
class Meta:
model = Group
class DiscussionAdmin(admin.ModelAdmin):
filter_horizontal = ('subscribers', 'ignorers')
class Meta:
model = Discussion
admin.site.register(Group, GroupAdmin)
admin.site.register(Discussion, DiscussionAdmin)
|
from django.contrib import admin
from .models import Discussion, Group
@admin.register(Group)
class GroupAdmin(admin.ModelAdmin):
filter_horizontal = ('moderators', 'watchers', 'members_if_private')
class Meta:
model = Group
@admin.register(Discussion)
class DiscussionAdmin(admin.ModelAdmin):
filter_horizontal = ('subscribers', 'ignorers')
class Meta:
model = Discussion
|
Use a decorator for slickness.
|
Use a decorator for slickness.
|
Python
|
bsd-2-clause
|
incuna/incuna-groups,incuna/incuna-groups
|
from django.contrib import admin
from .models import Discussion, Group
class GroupAdmin(admin.ModelAdmin):
filter_horizontal = ('moderators', 'watchers', 'members_if_private')
class Meta:
model = Group
class DiscussionAdmin(admin.ModelAdmin):
filter_horizontal = ('subscribers', 'ignorers')
class Meta:
model = Discussion
admin.site.register(Group, GroupAdmin)
admin.site.register(Discussion, DiscussionAdmin)
Use a decorator for slickness.
|
from django.contrib import admin
from .models import Discussion, Group
@admin.register(Group)
class GroupAdmin(admin.ModelAdmin):
filter_horizontal = ('moderators', 'watchers', 'members_if_private')
class Meta:
model = Group
@admin.register(Discussion)
class DiscussionAdmin(admin.ModelAdmin):
filter_horizontal = ('subscribers', 'ignorers')
class Meta:
model = Discussion
|
<commit_before>from django.contrib import admin
from .models import Discussion, Group
class GroupAdmin(admin.ModelAdmin):
filter_horizontal = ('moderators', 'watchers', 'members_if_private')
class Meta:
model = Group
class DiscussionAdmin(admin.ModelAdmin):
filter_horizontal = ('subscribers', 'ignorers')
class Meta:
model = Discussion
admin.site.register(Group, GroupAdmin)
admin.site.register(Discussion, DiscussionAdmin)
<commit_msg>Use a decorator for slickness.<commit_after>
|
from django.contrib import admin
from .models import Discussion, Group
@admin.register(Group)
class GroupAdmin(admin.ModelAdmin):
filter_horizontal = ('moderators', 'watchers', 'members_if_private')
class Meta:
model = Group
@admin.register(Discussion)
class DiscussionAdmin(admin.ModelAdmin):
filter_horizontal = ('subscribers', 'ignorers')
class Meta:
model = Discussion
|
from django.contrib import admin
from .models import Discussion, Group
class GroupAdmin(admin.ModelAdmin):
filter_horizontal = ('moderators', 'watchers', 'members_if_private')
class Meta:
model = Group
class DiscussionAdmin(admin.ModelAdmin):
filter_horizontal = ('subscribers', 'ignorers')
class Meta:
model = Discussion
admin.site.register(Group, GroupAdmin)
admin.site.register(Discussion, DiscussionAdmin)
Use a decorator for slickness.from django.contrib import admin
from .models import Discussion, Group
@admin.register(Group)
class GroupAdmin(admin.ModelAdmin):
filter_horizontal = ('moderators', 'watchers', 'members_if_private')
class Meta:
model = Group
@admin.register(Discussion)
class DiscussionAdmin(admin.ModelAdmin):
filter_horizontal = ('subscribers', 'ignorers')
class Meta:
model = Discussion
|
<commit_before>from django.contrib import admin
from .models import Discussion, Group
class GroupAdmin(admin.ModelAdmin):
filter_horizontal = ('moderators', 'watchers', 'members_if_private')
class Meta:
model = Group
class DiscussionAdmin(admin.ModelAdmin):
filter_horizontal = ('subscribers', 'ignorers')
class Meta:
model = Discussion
admin.site.register(Group, GroupAdmin)
admin.site.register(Discussion, DiscussionAdmin)
<commit_msg>Use a decorator for slickness.<commit_after>from django.contrib import admin
from .models import Discussion, Group
@admin.register(Group)
class GroupAdmin(admin.ModelAdmin):
filter_horizontal = ('moderators', 'watchers', 'members_if_private')
class Meta:
model = Group
@admin.register(Discussion)
class DiscussionAdmin(admin.ModelAdmin):
filter_horizontal = ('subscribers', 'ignorers')
class Meta:
model = Discussion
|
b9cf2145097f8d1c702183a09bf2d54f669e2218
|
skimage/filter/__init__.py
|
skimage/filter/__init__.py
|
from .lpi_filter import inverse, wiener, LPIFilter2D
from .ctmf import median_filter
from ._canny import canny
from .edges import (sobel, hsobel, vsobel, scharr, hscharr, vscharr, prewitt,
hprewitt, vprewitt, roberts , roberts_positive_diagonal,
roberts_negative_diagonal)
from ._denoise import denoise_tv_chambolle, tv_denoise
from ._denoise_cy import denoise_bilateral, denoise_tv_bregman
from ._rank_order import rank_order
from ._gabor import gabor_kernel, gabor_filter
from .thresholding import threshold_otsu, threshold_adaptive
__all__ = ['inverse',
'wiener',
'LPIFilter2D',
'median_filter',
'canny',
'sobel',
'hsobel',
'vsobel',
'scharr',
'hscharr',
'vscharr',
'prewitt',
'hprewitt',
'vprewitt',
'roberts',
'roberts_positive_diagonal',
'roberts_negative_diagonal',
'denoise_tv_chambolle',
'tv_denoise',
'denoise_bilateral',
'denoise_tv_bregman',
'rank_order',
'gabor_kernel',
'gabor_filter',
'threshold_otsu',
'threshold_adaptive']
|
from .lpi_filter import inverse, wiener, LPIFilter2D
from .ctmf import median_filter
from ._canny import canny
from .edges import (sobel, hsobel, vsobel, scharr, hscharr, vscharr, prewitt,
hprewitt, vprewitt, roberts , roberts_positive_diagonal,
roberts_negative_diagonal)
from ._denoise import denoise_tv_chambolle, tv_denoise
from ._denoise_cy import denoise_bilateral, denoise_tv_bregman
from ._rank_order import rank_order
from ._gabor import gabor_kernel, gabor_filter
from .thresholding import threshold_otsu, threshold_adaptive
from . import rank
__all__ = ['inverse',
'wiener',
'LPIFilter2D',
'median_filter',
'canny',
'sobel',
'hsobel',
'vsobel',
'scharr',
'hscharr',
'vscharr',
'prewitt',
'hprewitt',
'vprewitt',
'roberts',
'roberts_positive_diagonal',
'roberts_negative_diagonal',
'denoise_tv_chambolle',
'tv_denoise',
'denoise_bilateral',
'denoise_tv_bregman',
'rank_order',
'gabor_kernel',
'gabor_filter',
'threshold_otsu',
'threshold_adaptive',
'rank']
|
Add filter.rank to __all__ of filter package
|
Add filter.rank to __all__ of filter package
|
Python
|
bsd-3-clause
|
michaelpacer/scikit-image,oew1v07/scikit-image,vighneshbirodkar/scikit-image,michaelpacer/scikit-image,chriscrosscutler/scikit-image,juliusbierk/scikit-image,chintak/scikit-image,GaZ3ll3/scikit-image,warmspringwinds/scikit-image,ajaybhat/scikit-image,robintw/scikit-image,keflavich/scikit-image,chintak/scikit-image,jwiggins/scikit-image,rjeli/scikit-image,Britefury/scikit-image,bennlich/scikit-image,ofgulban/scikit-image,dpshelio/scikit-image,dpshelio/scikit-image,almarklein/scikit-image,keflavich/scikit-image,Midafi/scikit-image,pratapvardhan/scikit-image,emon10005/scikit-image,juliusbierk/scikit-image,chintak/scikit-image,pratapvardhan/scikit-image,bsipocz/scikit-image,jwiggins/scikit-image,almarklein/scikit-image,youprofit/scikit-image,SamHames/scikit-image,GaZ3ll3/scikit-image,Midafi/scikit-image,oew1v07/scikit-image,bennlich/scikit-image,SamHames/scikit-image,blink1073/scikit-image,warmspringwinds/scikit-image,Hiyorimi/scikit-image,chintak/scikit-image,newville/scikit-image,youprofit/scikit-image,Hiyorimi/scikit-image,SamHames/scikit-image,blink1073/scikit-image,rjeli/scikit-image,newville/scikit-image,paalge/scikit-image,paalge/scikit-image,almarklein/scikit-image,bsipocz/scikit-image,paalge/scikit-image,emon10005/scikit-image,ofgulban/scikit-image,michaelaye/scikit-image,ajaybhat/scikit-image,WarrenWeckesser/scikits-image,ClinicalGraphics/scikit-image,michaelaye/scikit-image,Britefury/scikit-image,vighneshbirodkar/scikit-image,SamHames/scikit-image,robintw/scikit-image,ClinicalGraphics/scikit-image,vighneshbirodkar/scikit-image,rjeli/scikit-image,almarklein/scikit-image,ofgulban/scikit-image,WarrenWeckesser/scikits-image,chriscrosscutler/scikit-image
|
from .lpi_filter import inverse, wiener, LPIFilter2D
from .ctmf import median_filter
from ._canny import canny
from .edges import (sobel, hsobel, vsobel, scharr, hscharr, vscharr, prewitt,
hprewitt, vprewitt, roberts , roberts_positive_diagonal,
roberts_negative_diagonal)
from ._denoise import denoise_tv_chambolle, tv_denoise
from ._denoise_cy import denoise_bilateral, denoise_tv_bregman
from ._rank_order import rank_order
from ._gabor import gabor_kernel, gabor_filter
from .thresholding import threshold_otsu, threshold_adaptive
__all__ = ['inverse',
'wiener',
'LPIFilter2D',
'median_filter',
'canny',
'sobel',
'hsobel',
'vsobel',
'scharr',
'hscharr',
'vscharr',
'prewitt',
'hprewitt',
'vprewitt',
'roberts',
'roberts_positive_diagonal',
'roberts_negative_diagonal',
'denoise_tv_chambolle',
'tv_denoise',
'denoise_bilateral',
'denoise_tv_bregman',
'rank_order',
'gabor_kernel',
'gabor_filter',
'threshold_otsu',
'threshold_adaptive']
Add filter.rank to __all__ of filter package
|
from .lpi_filter import inverse, wiener, LPIFilter2D
from .ctmf import median_filter
from ._canny import canny
from .edges import (sobel, hsobel, vsobel, scharr, hscharr, vscharr, prewitt,
hprewitt, vprewitt, roberts , roberts_positive_diagonal,
roberts_negative_diagonal)
from ._denoise import denoise_tv_chambolle, tv_denoise
from ._denoise_cy import denoise_bilateral, denoise_tv_bregman
from ._rank_order import rank_order
from ._gabor import gabor_kernel, gabor_filter
from .thresholding import threshold_otsu, threshold_adaptive
from . import rank
__all__ = ['inverse',
'wiener',
'LPIFilter2D',
'median_filter',
'canny',
'sobel',
'hsobel',
'vsobel',
'scharr',
'hscharr',
'vscharr',
'prewitt',
'hprewitt',
'vprewitt',
'roberts',
'roberts_positive_diagonal',
'roberts_negative_diagonal',
'denoise_tv_chambolle',
'tv_denoise',
'denoise_bilateral',
'denoise_tv_bregman',
'rank_order',
'gabor_kernel',
'gabor_filter',
'threshold_otsu',
'threshold_adaptive',
'rank']
|
<commit_before>from .lpi_filter import inverse, wiener, LPIFilter2D
from .ctmf import median_filter
from ._canny import canny
from .edges import (sobel, hsobel, vsobel, scharr, hscharr, vscharr, prewitt,
hprewitt, vprewitt, roberts , roberts_positive_diagonal,
roberts_negative_diagonal)
from ._denoise import denoise_tv_chambolle, tv_denoise
from ._denoise_cy import denoise_bilateral, denoise_tv_bregman
from ._rank_order import rank_order
from ._gabor import gabor_kernel, gabor_filter
from .thresholding import threshold_otsu, threshold_adaptive
__all__ = ['inverse',
'wiener',
'LPIFilter2D',
'median_filter',
'canny',
'sobel',
'hsobel',
'vsobel',
'scharr',
'hscharr',
'vscharr',
'prewitt',
'hprewitt',
'vprewitt',
'roberts',
'roberts_positive_diagonal',
'roberts_negative_diagonal',
'denoise_tv_chambolle',
'tv_denoise',
'denoise_bilateral',
'denoise_tv_bregman',
'rank_order',
'gabor_kernel',
'gabor_filter',
'threshold_otsu',
'threshold_adaptive']
<commit_msg>Add filter.rank to __all__ of filter package<commit_after>
|
from .lpi_filter import inverse, wiener, LPIFilter2D
from .ctmf import median_filter
from ._canny import canny
from .edges import (sobel, hsobel, vsobel, scharr, hscharr, vscharr, prewitt,
hprewitt, vprewitt, roberts , roberts_positive_diagonal,
roberts_negative_diagonal)
from ._denoise import denoise_tv_chambolle, tv_denoise
from ._denoise_cy import denoise_bilateral, denoise_tv_bregman
from ._rank_order import rank_order
from ._gabor import gabor_kernel, gabor_filter
from .thresholding import threshold_otsu, threshold_adaptive
from . import rank
__all__ = ['inverse',
'wiener',
'LPIFilter2D',
'median_filter',
'canny',
'sobel',
'hsobel',
'vsobel',
'scharr',
'hscharr',
'vscharr',
'prewitt',
'hprewitt',
'vprewitt',
'roberts',
'roberts_positive_diagonal',
'roberts_negative_diagonal',
'denoise_tv_chambolle',
'tv_denoise',
'denoise_bilateral',
'denoise_tv_bregman',
'rank_order',
'gabor_kernel',
'gabor_filter',
'threshold_otsu',
'threshold_adaptive',
'rank']
|
from .lpi_filter import inverse, wiener, LPIFilter2D
from .ctmf import median_filter
from ._canny import canny
from .edges import (sobel, hsobel, vsobel, scharr, hscharr, vscharr, prewitt,
hprewitt, vprewitt, roberts , roberts_positive_diagonal,
roberts_negative_diagonal)
from ._denoise import denoise_tv_chambolle, tv_denoise
from ._denoise_cy import denoise_bilateral, denoise_tv_bregman
from ._rank_order import rank_order
from ._gabor import gabor_kernel, gabor_filter
from .thresholding import threshold_otsu, threshold_adaptive
__all__ = ['inverse',
'wiener',
'LPIFilter2D',
'median_filter',
'canny',
'sobel',
'hsobel',
'vsobel',
'scharr',
'hscharr',
'vscharr',
'prewitt',
'hprewitt',
'vprewitt',
'roberts',
'roberts_positive_diagonal',
'roberts_negative_diagonal',
'denoise_tv_chambolle',
'tv_denoise',
'denoise_bilateral',
'denoise_tv_bregman',
'rank_order',
'gabor_kernel',
'gabor_filter',
'threshold_otsu',
'threshold_adaptive']
Add filter.rank to __all__ of filter packagefrom .lpi_filter import inverse, wiener, LPIFilter2D
from .ctmf import median_filter
from ._canny import canny
from .edges import (sobel, hsobel, vsobel, scharr, hscharr, vscharr, prewitt,
hprewitt, vprewitt, roberts , roberts_positive_diagonal,
roberts_negative_diagonal)
from ._denoise import denoise_tv_chambolle, tv_denoise
from ._denoise_cy import denoise_bilateral, denoise_tv_bregman
from ._rank_order import rank_order
from ._gabor import gabor_kernel, gabor_filter
from .thresholding import threshold_otsu, threshold_adaptive
from . import rank
__all__ = ['inverse',
'wiener',
'LPIFilter2D',
'median_filter',
'canny',
'sobel',
'hsobel',
'vsobel',
'scharr',
'hscharr',
'vscharr',
'prewitt',
'hprewitt',
'vprewitt',
'roberts',
'roberts_positive_diagonal',
'roberts_negative_diagonal',
'denoise_tv_chambolle',
'tv_denoise',
'denoise_bilateral',
'denoise_tv_bregman',
'rank_order',
'gabor_kernel',
'gabor_filter',
'threshold_otsu',
'threshold_adaptive',
'rank']
|
<commit_before>from .lpi_filter import inverse, wiener, LPIFilter2D
from .ctmf import median_filter
from ._canny import canny
from .edges import (sobel, hsobel, vsobel, scharr, hscharr, vscharr, prewitt,
hprewitt, vprewitt, roberts , roberts_positive_diagonal,
roberts_negative_diagonal)
from ._denoise import denoise_tv_chambolle, tv_denoise
from ._denoise_cy import denoise_bilateral, denoise_tv_bregman
from ._rank_order import rank_order
from ._gabor import gabor_kernel, gabor_filter
from .thresholding import threshold_otsu, threshold_adaptive
__all__ = ['inverse',
'wiener',
'LPIFilter2D',
'median_filter',
'canny',
'sobel',
'hsobel',
'vsobel',
'scharr',
'hscharr',
'vscharr',
'prewitt',
'hprewitt',
'vprewitt',
'roberts',
'roberts_positive_diagonal',
'roberts_negative_diagonal',
'denoise_tv_chambolle',
'tv_denoise',
'denoise_bilateral',
'denoise_tv_bregman',
'rank_order',
'gabor_kernel',
'gabor_filter',
'threshold_otsu',
'threshold_adaptive']
<commit_msg>Add filter.rank to __all__ of filter package<commit_after>from .lpi_filter import inverse, wiener, LPIFilter2D
from .ctmf import median_filter
from ._canny import canny
from .edges import (sobel, hsobel, vsobel, scharr, hscharr, vscharr, prewitt,
hprewitt, vprewitt, roberts , roberts_positive_diagonal,
roberts_negative_diagonal)
from ._denoise import denoise_tv_chambolle, tv_denoise
from ._denoise_cy import denoise_bilateral, denoise_tv_bregman
from ._rank_order import rank_order
from ._gabor import gabor_kernel, gabor_filter
from .thresholding import threshold_otsu, threshold_adaptive
from . import rank
__all__ = ['inverse',
'wiener',
'LPIFilter2D',
'median_filter',
'canny',
'sobel',
'hsobel',
'vsobel',
'scharr',
'hscharr',
'vscharr',
'prewitt',
'hprewitt',
'vprewitt',
'roberts',
'roberts_positive_diagonal',
'roberts_negative_diagonal',
'denoise_tv_chambolle',
'tv_denoise',
'denoise_bilateral',
'denoise_tv_bregman',
'rank_order',
'gabor_kernel',
'gabor_filter',
'threshold_otsu',
'threshold_adaptive',
'rank']
|
98a2b7e11eb3e0d5ddc89a4d40c3d10586e400ab
|
website/filters/__init__.py
|
website/filters/__init__.py
|
import hashlib
import urllib
# Adapted from https://github.com/zzzsochi/Flask-Gravatar/blob/master/flaskext/gravatar.py
def gravatar(user, use_ssl=False, d=None, r=None, size=None):
if use_ssl:
base_url = 'https://secure.gravatar.com/avatar/'
else:
base_url = 'http://www.gravatar.com/avatar/'
# user can be a User instance or a username string
username = user.username if hasattr(user, 'username') else user
hash_code = hashlib.md5(unicode(username).encode('utf-8')).hexdigest()
url = base_url + '?'
# Order of query params matters, due to a quirk with gravatar
params = [
('s', size),
('d', 'identicon'),
]
if r:
params.append(('r', r))
url = base_url + hash_code + '?' + urllib.urlencode(params)
return url
|
import hashlib
import urllib
# Adapted from https://github.com/zzzsochi/Flask-Gravatar/blob/master/flaskext/gravatar.py
def gravatar(user, use_ssl=False, d=None, r=None, size=None):
if use_ssl:
base_url = 'https://secure.gravatar.com/avatar/'
else:
base_url = 'http://www.gravatar.com/avatar/'
# user can be a User instance or a username string
username = user.username if hasattr(user, 'username') else user
hash_code = hashlib.md5(unicode(username).encode('utf-8')).hexdigest()
url = base_url + '?'
# Order of query params matters, due to a quirk with gravatar
params = [
('d', 'identicon'),
('s', size),
]
if r:
params.append(('r', r))
url = base_url + hash_code + '?' + urllib.urlencode(params)
return url
|
Fix ordering of query params
|
Fix ordering of query params
3rd time's a charm
|
Python
|
apache-2.0
|
mluke93/osf.io,binoculars/osf.io,leb2dg/osf.io,caseyrygt/osf.io,samanehsan/osf.io,cwisecarver/osf.io,petermalcolm/osf.io,kwierman/osf.io,emetsger/osf.io,mluo613/osf.io,wearpants/osf.io,samchrisinger/osf.io,amyshi188/osf.io,jnayak1/osf.io,TomHeatwole/osf.io,petermalcolm/osf.io,TomBaxter/osf.io,amyshi188/osf.io,TomHeatwole/osf.io,doublebits/osf.io,billyhunt/osf.io,crcresearch/osf.io,Ghalko/osf.io,saradbowman/osf.io,monikagrabowska/osf.io,mluo613/osf.io,alexschiller/osf.io,ticklemepierce/osf.io,njantrania/osf.io,caseyrollins/osf.io,zachjanicki/osf.io,brianjgeiger/osf.io,laurenrevere/osf.io,HalcyonChimera/osf.io,laurenrevere/osf.io,doublebits/osf.io,njantrania/osf.io,acshi/osf.io,danielneis/osf.io,mfraezz/osf.io,cslzchen/osf.io,Johnetordoff/osf.io,alexschiller/osf.io,ZobairAlijan/osf.io,caseyrollins/osf.io,KAsante95/osf.io,brandonPurvis/osf.io,kwierman/osf.io,zamattiac/osf.io,mluke93/osf.io,adlius/osf.io,Ghalko/osf.io,cslzchen/osf.io,kwierman/osf.io,DanielSBrown/osf.io,kch8qx/osf.io,mfraezz/osf.io,kwierman/osf.io,SSJohns/osf.io,asanfilippo7/osf.io,icereval/osf.io,RomanZWang/osf.io,alexschiller/osf.io,chennan47/osf.io,GageGaskins/osf.io,caneruguz/osf.io,baylee-d/osf.io,GageGaskins/osf.io,chennan47/osf.io,samanehsan/osf.io,petermalcolm/osf.io,RomanZWang/osf.io,DanielSBrown/osf.io,brianjgeiger/osf.io,kch8qx/osf.io,laurenrevere/osf.io,caseyrygt/osf.io,billyhunt/osf.io,billyhunt/osf.io,jnayak1/osf.io,chrisseto/osf.io,adlius/osf.io,wearpants/osf.io,cwisecarver/osf.io,samchrisinger/osf.io,cwisecarver/osf.io,alexschiller/osf.io,Nesiehr/osf.io,Johnetordoff/osf.io,acshi/osf.io,abought/osf.io,TomBaxter/osf.io,RomanZWang/osf.io,monikagrabowska/osf.io,wearpants/osf.io,binoculars/osf.io,caseyrollins/osf.io,zamattiac/osf.io,KAsante95/osf.io,mattclark/osf.io,felliott/osf.io,ticklemepierce/osf.io,SSJohns/osf.io,brandonPurvis/osf.io,brianjgeiger/osf.io,baylee-d/osf.io,HalcyonChimera/osf.io,HalcyonChimera/osf.io,mluo613/osf.io,zamattiac/osf.io,GageGaskins/osf.io,rdhyee/osf.io,Johnetordoff/osf.io,CenterForOpenScience/osf.io,cslzchen/osf.io,mattclark/osf.io,zachjanicki/osf.io,kch8qx/osf.io,mluke93/osf.io,KAsante95/osf.io,GageGaskins/osf.io,felliott/osf.io,haoyuchen1992/osf.io,erinspace/osf.io,HalcyonChimera/osf.io,hmoco/osf.io,felliott/osf.io,brandonPurvis/osf.io,zachjanicki/osf.io,ZobairAlijan/osf.io,mfraezz/osf.io,DanielSBrown/osf.io,cslzchen/osf.io,amyshi188/osf.io,saradbowman/osf.io,emetsger/osf.io,pattisdr/osf.io,rdhyee/osf.io,monikagrabowska/osf.io,zamattiac/osf.io,rdhyee/osf.io,erinspace/osf.io,haoyuchen1992/osf.io,chrisseto/osf.io,chrisseto/osf.io,Johnetordoff/osf.io,leb2dg/osf.io,samanehsan/osf.io,mluke93/osf.io,abought/osf.io,abought/osf.io,adlius/osf.io,RomanZWang/osf.io,caneruguz/osf.io,njantrania/osf.io,billyhunt/osf.io,erinspace/osf.io,brianjgeiger/osf.io,monikagrabowska/osf.io,SSJohns/osf.io,Nesiehr/osf.io,petermalcolm/osf.io,KAsante95/osf.io,hmoco/osf.io,caneruguz/osf.io,caseyrygt/osf.io,emetsger/osf.io,doublebits/osf.io,jnayak1/osf.io,danielneis/osf.io,danielneis/osf.io,alexschiller/osf.io,pattisdr/osf.io,CenterForOpenScience/osf.io,mattclark/osf.io,kch8qx/osf.io,cosenal/osf.io,caseyrygt/osf.io,kch8qx/osf.io,icereval/osf.io,aaxelb/osf.io,aaxelb/osf.io,crcresearch/osf.io,felliott/osf.io,mluo613/osf.io,haoyuchen1992/osf.io,sloria/osf.io,Ghalko/osf.io,hmoco/osf.io,aaxelb/osf.io,billyhunt/osf.io,cwisecarver/osf.io,RomanZWang/osf.io,samchrisinger/osf.io,GageGaskins/osf.io,jnayak1/osf.io,cosenal/osf.io,rdhyee/osf.io,Nesiehr/osf.io,acshi/osf.io,KAsante95/osf.io,samanehsan/osf.io,emetsger/osf.io,haoyuchen1992/osf.io,danielneis/osf.io,asanfilippo7/osf.io,TomHeatwole/osf.io,leb2dg/osf.io,adlius/osf.io,hmoco/osf.io,CenterForOpenScience/osf.io,cosenal/osf.io,brandonPurvis/osf.io,crcresearch/osf.io,wearpants/osf.io,mluo613/osf.io,ticklemepierce/osf.io,ticklemepierce/osf.io,aaxelb/osf.io,njantrania/osf.io,acshi/osf.io,mfraezz/osf.io,asanfilippo7/osf.io,brandonPurvis/osf.io,ZobairAlijan/osf.io,sloria/osf.io,binoculars/osf.io,CenterForOpenScience/osf.io,acshi/osf.io,asanfilippo7/osf.io,chrisseto/osf.io,DanielSBrown/osf.io,abought/osf.io,leb2dg/osf.io,TomBaxter/osf.io,samchrisinger/osf.io,amyshi188/osf.io,SSJohns/osf.io,doublebits/osf.io,cosenal/osf.io,caneruguz/osf.io,zachjanicki/osf.io,TomHeatwole/osf.io,pattisdr/osf.io,Nesiehr/osf.io,icereval/osf.io,chennan47/osf.io,monikagrabowska/osf.io,baylee-d/osf.io,doublebits/osf.io,sloria/osf.io,Ghalko/osf.io,ZobairAlijan/osf.io
|
import hashlib
import urllib
# Adapted from https://github.com/zzzsochi/Flask-Gravatar/blob/master/flaskext/gravatar.py
def gravatar(user, use_ssl=False, d=None, r=None, size=None):
if use_ssl:
base_url = 'https://secure.gravatar.com/avatar/'
else:
base_url = 'http://www.gravatar.com/avatar/'
# user can be a User instance or a username string
username = user.username if hasattr(user, 'username') else user
hash_code = hashlib.md5(unicode(username).encode('utf-8')).hexdigest()
url = base_url + '?'
# Order of query params matters, due to a quirk with gravatar
params = [
('s', size),
('d', 'identicon'),
]
if r:
params.append(('r', r))
url = base_url + hash_code + '?' + urllib.urlencode(params)
return url
Fix ordering of query params
3rd time's a charm
|
import hashlib
import urllib
# Adapted from https://github.com/zzzsochi/Flask-Gravatar/blob/master/flaskext/gravatar.py
def gravatar(user, use_ssl=False, d=None, r=None, size=None):
if use_ssl:
base_url = 'https://secure.gravatar.com/avatar/'
else:
base_url = 'http://www.gravatar.com/avatar/'
# user can be a User instance or a username string
username = user.username if hasattr(user, 'username') else user
hash_code = hashlib.md5(unicode(username).encode('utf-8')).hexdigest()
url = base_url + '?'
# Order of query params matters, due to a quirk with gravatar
params = [
('d', 'identicon'),
('s', size),
]
if r:
params.append(('r', r))
url = base_url + hash_code + '?' + urllib.urlencode(params)
return url
|
<commit_before>import hashlib
import urllib
# Adapted from https://github.com/zzzsochi/Flask-Gravatar/blob/master/flaskext/gravatar.py
def gravatar(user, use_ssl=False, d=None, r=None, size=None):
if use_ssl:
base_url = 'https://secure.gravatar.com/avatar/'
else:
base_url = 'http://www.gravatar.com/avatar/'
# user can be a User instance or a username string
username = user.username if hasattr(user, 'username') else user
hash_code = hashlib.md5(unicode(username).encode('utf-8')).hexdigest()
url = base_url + '?'
# Order of query params matters, due to a quirk with gravatar
params = [
('s', size),
('d', 'identicon'),
]
if r:
params.append(('r', r))
url = base_url + hash_code + '?' + urllib.urlencode(params)
return url
<commit_msg>Fix ordering of query params
3rd time's a charm<commit_after>
|
import hashlib
import urllib
# Adapted from https://github.com/zzzsochi/Flask-Gravatar/blob/master/flaskext/gravatar.py
def gravatar(user, use_ssl=False, d=None, r=None, size=None):
if use_ssl:
base_url = 'https://secure.gravatar.com/avatar/'
else:
base_url = 'http://www.gravatar.com/avatar/'
# user can be a User instance or a username string
username = user.username if hasattr(user, 'username') else user
hash_code = hashlib.md5(unicode(username).encode('utf-8')).hexdigest()
url = base_url + '?'
# Order of query params matters, due to a quirk with gravatar
params = [
('d', 'identicon'),
('s', size),
]
if r:
params.append(('r', r))
url = base_url + hash_code + '?' + urllib.urlencode(params)
return url
|
import hashlib
import urllib
# Adapted from https://github.com/zzzsochi/Flask-Gravatar/blob/master/flaskext/gravatar.py
def gravatar(user, use_ssl=False, d=None, r=None, size=None):
if use_ssl:
base_url = 'https://secure.gravatar.com/avatar/'
else:
base_url = 'http://www.gravatar.com/avatar/'
# user can be a User instance or a username string
username = user.username if hasattr(user, 'username') else user
hash_code = hashlib.md5(unicode(username).encode('utf-8')).hexdigest()
url = base_url + '?'
# Order of query params matters, due to a quirk with gravatar
params = [
('s', size),
('d', 'identicon'),
]
if r:
params.append(('r', r))
url = base_url + hash_code + '?' + urllib.urlencode(params)
return url
Fix ordering of query params
3rd time's a charmimport hashlib
import urllib
# Adapted from https://github.com/zzzsochi/Flask-Gravatar/blob/master/flaskext/gravatar.py
def gravatar(user, use_ssl=False, d=None, r=None, size=None):
if use_ssl:
base_url = 'https://secure.gravatar.com/avatar/'
else:
base_url = 'http://www.gravatar.com/avatar/'
# user can be a User instance or a username string
username = user.username if hasattr(user, 'username') else user
hash_code = hashlib.md5(unicode(username).encode('utf-8')).hexdigest()
url = base_url + '?'
# Order of query params matters, due to a quirk with gravatar
params = [
('d', 'identicon'),
('s', size),
]
if r:
params.append(('r', r))
url = base_url + hash_code + '?' + urllib.urlencode(params)
return url
|
<commit_before>import hashlib
import urllib
# Adapted from https://github.com/zzzsochi/Flask-Gravatar/blob/master/flaskext/gravatar.py
def gravatar(user, use_ssl=False, d=None, r=None, size=None):
if use_ssl:
base_url = 'https://secure.gravatar.com/avatar/'
else:
base_url = 'http://www.gravatar.com/avatar/'
# user can be a User instance or a username string
username = user.username if hasattr(user, 'username') else user
hash_code = hashlib.md5(unicode(username).encode('utf-8')).hexdigest()
url = base_url + '?'
# Order of query params matters, due to a quirk with gravatar
params = [
('s', size),
('d', 'identicon'),
]
if r:
params.append(('r', r))
url = base_url + hash_code + '?' + urllib.urlencode(params)
return url
<commit_msg>Fix ordering of query params
3rd time's a charm<commit_after>import hashlib
import urllib
# Adapted from https://github.com/zzzsochi/Flask-Gravatar/blob/master/flaskext/gravatar.py
def gravatar(user, use_ssl=False, d=None, r=None, size=None):
if use_ssl:
base_url = 'https://secure.gravatar.com/avatar/'
else:
base_url = 'http://www.gravatar.com/avatar/'
# user can be a User instance or a username string
username = user.username if hasattr(user, 'username') else user
hash_code = hashlib.md5(unicode(username).encode('utf-8')).hexdigest()
url = base_url + '?'
# Order of query params matters, due to a quirk with gravatar
params = [
('d', 'identicon'),
('s', size),
]
if r:
params.append(('r', r))
url = base_url + hash_code + '?' + urllib.urlencode(params)
return url
|
774b64779b18ff0d8fba048ab4c4cae53662628a
|
ummeli/vlive/auth/middleware.py
|
ummeli/vlive/auth/middleware.py
|
from django.contrib.auth.middleware import RemoteUserMiddleware
class VodafoneLiveUserMiddleware(RemoteUserMiddleware):
header = 'HTTP_X_UP_CALLING_LINE_ID'
class VodafoneLiveInfo(object):
pass
class VodafoneLiveInfoMiddleware(object):
"""
Friendlier access to device / request info that Vodafone Live makes
available to us via HTTP Headers
"""
def process_request(self, request):
vlive = VodafoneLiveInfo()
vlive.msisdn = request.META.get('HTTP_X_UP_CALLING_LINE_ID', 'unknown')
vlive.area = request.META.get('HTTP_X_VODAFONE_AREA', 'unknown')
request.vlive = vlive
print request.META
|
from django.contrib.auth.middleware import RemoteUserMiddleware
class VodafoneLiveUserMiddleware(RemoteUserMiddleware):
header = 'HTTP_X_UP_CALLING_LINE_ID'
class VodafoneLiveInfo(object):
pass
class VodafoneLiveInfoMiddleware(object):
"""
Friendlier access to device / request info that Vodafone Live makes
available to us via HTTP Headers
"""
def process_request(self, request):
vlive = VodafoneLiveInfo()
vlive.msisdn = request.META.get('HTTP_X_UP_CALLING_LINE_ID', 'unknown')
vlive.area = request.META.get('HTTP_X_VODAFONE_AREA', 'unknown')
request.vlive = vlive
|
Revert "printing META for troubleshooting"
|
Revert "printing META for troubleshooting"
This reverts commit 42d15d528da14866f2f0479da6462c17a02d8c84.
|
Python
|
bsd-3-clause
|
praekelt/ummeli,praekelt/ummeli,praekelt/ummeli
|
from django.contrib.auth.middleware import RemoteUserMiddleware
class VodafoneLiveUserMiddleware(RemoteUserMiddleware):
header = 'HTTP_X_UP_CALLING_LINE_ID'
class VodafoneLiveInfo(object):
pass
class VodafoneLiveInfoMiddleware(object):
"""
Friendlier access to device / request info that Vodafone Live makes
available to us via HTTP Headers
"""
def process_request(self, request):
vlive = VodafoneLiveInfo()
vlive.msisdn = request.META.get('HTTP_X_UP_CALLING_LINE_ID', 'unknown')
vlive.area = request.META.get('HTTP_X_VODAFONE_AREA', 'unknown')
request.vlive = vlive
print request.META
Revert "printing META for troubleshooting"
This reverts commit 42d15d528da14866f2f0479da6462c17a02d8c84.
|
from django.contrib.auth.middleware import RemoteUserMiddleware
class VodafoneLiveUserMiddleware(RemoteUserMiddleware):
header = 'HTTP_X_UP_CALLING_LINE_ID'
class VodafoneLiveInfo(object):
pass
class VodafoneLiveInfoMiddleware(object):
"""
Friendlier access to device / request info that Vodafone Live makes
available to us via HTTP Headers
"""
def process_request(self, request):
vlive = VodafoneLiveInfo()
vlive.msisdn = request.META.get('HTTP_X_UP_CALLING_LINE_ID', 'unknown')
vlive.area = request.META.get('HTTP_X_VODAFONE_AREA', 'unknown')
request.vlive = vlive
|
<commit_before>from django.contrib.auth.middleware import RemoteUserMiddleware
class VodafoneLiveUserMiddleware(RemoteUserMiddleware):
header = 'HTTP_X_UP_CALLING_LINE_ID'
class VodafoneLiveInfo(object):
pass
class VodafoneLiveInfoMiddleware(object):
"""
Friendlier access to device / request info that Vodafone Live makes
available to us via HTTP Headers
"""
def process_request(self, request):
vlive = VodafoneLiveInfo()
vlive.msisdn = request.META.get('HTTP_X_UP_CALLING_LINE_ID', 'unknown')
vlive.area = request.META.get('HTTP_X_VODAFONE_AREA', 'unknown')
request.vlive = vlive
print request.META
<commit_msg>Revert "printing META for troubleshooting"
This reverts commit 42d15d528da14866f2f0479da6462c17a02d8c84.<commit_after>
|
from django.contrib.auth.middleware import RemoteUserMiddleware
class VodafoneLiveUserMiddleware(RemoteUserMiddleware):
header = 'HTTP_X_UP_CALLING_LINE_ID'
class VodafoneLiveInfo(object):
pass
class VodafoneLiveInfoMiddleware(object):
"""
Friendlier access to device / request info that Vodafone Live makes
available to us via HTTP Headers
"""
def process_request(self, request):
vlive = VodafoneLiveInfo()
vlive.msisdn = request.META.get('HTTP_X_UP_CALLING_LINE_ID', 'unknown')
vlive.area = request.META.get('HTTP_X_VODAFONE_AREA', 'unknown')
request.vlive = vlive
|
from django.contrib.auth.middleware import RemoteUserMiddleware
class VodafoneLiveUserMiddleware(RemoteUserMiddleware):
header = 'HTTP_X_UP_CALLING_LINE_ID'
class VodafoneLiveInfo(object):
pass
class VodafoneLiveInfoMiddleware(object):
"""
Friendlier access to device / request info that Vodafone Live makes
available to us via HTTP Headers
"""
def process_request(self, request):
vlive = VodafoneLiveInfo()
vlive.msisdn = request.META.get('HTTP_X_UP_CALLING_LINE_ID', 'unknown')
vlive.area = request.META.get('HTTP_X_VODAFONE_AREA', 'unknown')
request.vlive = vlive
print request.META
Revert "printing META for troubleshooting"
This reverts commit 42d15d528da14866f2f0479da6462c17a02d8c84.from django.contrib.auth.middleware import RemoteUserMiddleware
class VodafoneLiveUserMiddleware(RemoteUserMiddleware):
header = 'HTTP_X_UP_CALLING_LINE_ID'
class VodafoneLiveInfo(object):
pass
class VodafoneLiveInfoMiddleware(object):
"""
Friendlier access to device / request info that Vodafone Live makes
available to us via HTTP Headers
"""
def process_request(self, request):
vlive = VodafoneLiveInfo()
vlive.msisdn = request.META.get('HTTP_X_UP_CALLING_LINE_ID', 'unknown')
vlive.area = request.META.get('HTTP_X_VODAFONE_AREA', 'unknown')
request.vlive = vlive
|
<commit_before>from django.contrib.auth.middleware import RemoteUserMiddleware
class VodafoneLiveUserMiddleware(RemoteUserMiddleware):
header = 'HTTP_X_UP_CALLING_LINE_ID'
class VodafoneLiveInfo(object):
pass
class VodafoneLiveInfoMiddleware(object):
"""
Friendlier access to device / request info that Vodafone Live makes
available to us via HTTP Headers
"""
def process_request(self, request):
vlive = VodafoneLiveInfo()
vlive.msisdn = request.META.get('HTTP_X_UP_CALLING_LINE_ID', 'unknown')
vlive.area = request.META.get('HTTP_X_VODAFONE_AREA', 'unknown')
request.vlive = vlive
print request.META
<commit_msg>Revert "printing META for troubleshooting"
This reverts commit 42d15d528da14866f2f0479da6462c17a02d8c84.<commit_after>from django.contrib.auth.middleware import RemoteUserMiddleware
class VodafoneLiveUserMiddleware(RemoteUserMiddleware):
header = 'HTTP_X_UP_CALLING_LINE_ID'
class VodafoneLiveInfo(object):
pass
class VodafoneLiveInfoMiddleware(object):
"""
Friendlier access to device / request info that Vodafone Live makes
available to us via HTTP Headers
"""
def process_request(self, request):
vlive = VodafoneLiveInfo()
vlive.msisdn = request.META.get('HTTP_X_UP_CALLING_LINE_ID', 'unknown')
vlive.area = request.META.get('HTTP_X_VODAFONE_AREA', 'unknown')
request.vlive = vlive
|
de1baa49fc34f8ecf4f7df4c723456348281df69
|
splunk_handler/__init__.py
|
splunk_handler/__init__.py
|
import logging
import socket
import traceback
from threading import Thread
import requests
class SplunkHandler(logging.Handler):
"""
A logging handler to send events to a Splunk Enterprise instance
"""
def __init__(self, host, port, username, password, index):
logging.Handler.__init__(self)
self.host = host
self.port = port
self.username = username
self.password = password
self.index = index
def emit(self, record):
thread = Thread(target=self._async_emit, args=(record, ))
thread.start()
def _async_emit(self, record):
try:
params = {
'host': socket.gethostname(),
'index': self.index,
'source': record.pathname,
'sourcetype': 'json'
}
url = 'https://%s:%s/services/receivers/simple' % (self.host, self.port)
payload = self.format(record)
auth = (self.username, self.password)
r = requests.post(
url,
auth=auth,
data=payload,
params=params
)
r.close()
except Exception, e:
print "Traceback:\n" + traceback.format_exc()
print "Exception in Splunk logging handler: %s" % str(e)
|
import logging
import socket
import traceback
from threading import Thread
import requests
class SplunkHandler(logging.Handler):
"""
A logging handler to send events to a Splunk Enterprise instance
"""
def __init__(self, host, port, username, password, index):
logging.Handler.__init__(self)
self.host = host
self.port = port
self.username = username
self.password = password
self.index = index
requests_log = logging.getLogger('requests')
requests_log.propagate = False
def emit(self, record):
thread = Thread(target=self._async_emit, args=(record, ))
thread.start()
def _async_emit(self, record):
try:
params = {
'host': socket.gethostname(),
'index': self.index,
'source': record.pathname,
'sourcetype': 'json'
}
url = 'https://%s:%s/services/receivers/simple' % (self.host, self.port)
payload = self.format(record)
auth = (self.username, self.password)
r = requests.post(
url,
auth=auth,
data=payload,
params=params
)
r.close()
except Exception, e:
print "Traceback:\n" + traceback.format_exc()
print "Exception in Splunk logging handler: %s" % str(e)
|
Add code to silence requests logger in the handler
|
Add code to silence requests logger in the handler
|
Python
|
mit
|
zach-taylor/splunk_handler,sullivanmatt/splunk_handler
|
import logging
import socket
import traceback
from threading import Thread
import requests
class SplunkHandler(logging.Handler):
"""
A logging handler to send events to a Splunk Enterprise instance
"""
def __init__(self, host, port, username, password, index):
logging.Handler.__init__(self)
self.host = host
self.port = port
self.username = username
self.password = password
self.index = index
def emit(self, record):
thread = Thread(target=self._async_emit, args=(record, ))
thread.start()
def _async_emit(self, record):
try:
params = {
'host': socket.gethostname(),
'index': self.index,
'source': record.pathname,
'sourcetype': 'json'
}
url = 'https://%s:%s/services/receivers/simple' % (self.host, self.port)
payload = self.format(record)
auth = (self.username, self.password)
r = requests.post(
url,
auth=auth,
data=payload,
params=params
)
r.close()
except Exception, e:
print "Traceback:\n" + traceback.format_exc()
print "Exception in Splunk logging handler: %s" % str(e)
Add code to silence requests logger in the handler
|
import logging
import socket
import traceback
from threading import Thread
import requests
class SplunkHandler(logging.Handler):
"""
A logging handler to send events to a Splunk Enterprise instance
"""
def __init__(self, host, port, username, password, index):
logging.Handler.__init__(self)
self.host = host
self.port = port
self.username = username
self.password = password
self.index = index
requests_log = logging.getLogger('requests')
requests_log.propagate = False
def emit(self, record):
thread = Thread(target=self._async_emit, args=(record, ))
thread.start()
def _async_emit(self, record):
try:
params = {
'host': socket.gethostname(),
'index': self.index,
'source': record.pathname,
'sourcetype': 'json'
}
url = 'https://%s:%s/services/receivers/simple' % (self.host, self.port)
payload = self.format(record)
auth = (self.username, self.password)
r = requests.post(
url,
auth=auth,
data=payload,
params=params
)
r.close()
except Exception, e:
print "Traceback:\n" + traceback.format_exc()
print "Exception in Splunk logging handler: %s" % str(e)
|
<commit_before>import logging
import socket
import traceback
from threading import Thread
import requests
class SplunkHandler(logging.Handler):
"""
A logging handler to send events to a Splunk Enterprise instance
"""
def __init__(self, host, port, username, password, index):
logging.Handler.__init__(self)
self.host = host
self.port = port
self.username = username
self.password = password
self.index = index
def emit(self, record):
thread = Thread(target=self._async_emit, args=(record, ))
thread.start()
def _async_emit(self, record):
try:
params = {
'host': socket.gethostname(),
'index': self.index,
'source': record.pathname,
'sourcetype': 'json'
}
url = 'https://%s:%s/services/receivers/simple' % (self.host, self.port)
payload = self.format(record)
auth = (self.username, self.password)
r = requests.post(
url,
auth=auth,
data=payload,
params=params
)
r.close()
except Exception, e:
print "Traceback:\n" + traceback.format_exc()
print "Exception in Splunk logging handler: %s" % str(e)
<commit_msg>Add code to silence requests logger in the handler<commit_after>
|
import logging
import socket
import traceback
from threading import Thread
import requests
class SplunkHandler(logging.Handler):
"""
A logging handler to send events to a Splunk Enterprise instance
"""
def __init__(self, host, port, username, password, index):
logging.Handler.__init__(self)
self.host = host
self.port = port
self.username = username
self.password = password
self.index = index
requests_log = logging.getLogger('requests')
requests_log.propagate = False
def emit(self, record):
thread = Thread(target=self._async_emit, args=(record, ))
thread.start()
def _async_emit(self, record):
try:
params = {
'host': socket.gethostname(),
'index': self.index,
'source': record.pathname,
'sourcetype': 'json'
}
url = 'https://%s:%s/services/receivers/simple' % (self.host, self.port)
payload = self.format(record)
auth = (self.username, self.password)
r = requests.post(
url,
auth=auth,
data=payload,
params=params
)
r.close()
except Exception, e:
print "Traceback:\n" + traceback.format_exc()
print "Exception in Splunk logging handler: %s" % str(e)
|
import logging
import socket
import traceback
from threading import Thread
import requests
class SplunkHandler(logging.Handler):
"""
A logging handler to send events to a Splunk Enterprise instance
"""
def __init__(self, host, port, username, password, index):
logging.Handler.__init__(self)
self.host = host
self.port = port
self.username = username
self.password = password
self.index = index
def emit(self, record):
thread = Thread(target=self._async_emit, args=(record, ))
thread.start()
def _async_emit(self, record):
try:
params = {
'host': socket.gethostname(),
'index': self.index,
'source': record.pathname,
'sourcetype': 'json'
}
url = 'https://%s:%s/services/receivers/simple' % (self.host, self.port)
payload = self.format(record)
auth = (self.username, self.password)
r = requests.post(
url,
auth=auth,
data=payload,
params=params
)
r.close()
except Exception, e:
print "Traceback:\n" + traceback.format_exc()
print "Exception in Splunk logging handler: %s" % str(e)
Add code to silence requests logger in the handlerimport logging
import socket
import traceback
from threading import Thread
import requests
class SplunkHandler(logging.Handler):
"""
A logging handler to send events to a Splunk Enterprise instance
"""
def __init__(self, host, port, username, password, index):
logging.Handler.__init__(self)
self.host = host
self.port = port
self.username = username
self.password = password
self.index = index
requests_log = logging.getLogger('requests')
requests_log.propagate = False
def emit(self, record):
thread = Thread(target=self._async_emit, args=(record, ))
thread.start()
def _async_emit(self, record):
try:
params = {
'host': socket.gethostname(),
'index': self.index,
'source': record.pathname,
'sourcetype': 'json'
}
url = 'https://%s:%s/services/receivers/simple' % (self.host, self.port)
payload = self.format(record)
auth = (self.username, self.password)
r = requests.post(
url,
auth=auth,
data=payload,
params=params
)
r.close()
except Exception, e:
print "Traceback:\n" + traceback.format_exc()
print "Exception in Splunk logging handler: %s" % str(e)
|
<commit_before>import logging
import socket
import traceback
from threading import Thread
import requests
class SplunkHandler(logging.Handler):
"""
A logging handler to send events to a Splunk Enterprise instance
"""
def __init__(self, host, port, username, password, index):
logging.Handler.__init__(self)
self.host = host
self.port = port
self.username = username
self.password = password
self.index = index
def emit(self, record):
thread = Thread(target=self._async_emit, args=(record, ))
thread.start()
def _async_emit(self, record):
try:
params = {
'host': socket.gethostname(),
'index': self.index,
'source': record.pathname,
'sourcetype': 'json'
}
url = 'https://%s:%s/services/receivers/simple' % (self.host, self.port)
payload = self.format(record)
auth = (self.username, self.password)
r = requests.post(
url,
auth=auth,
data=payload,
params=params
)
r.close()
except Exception, e:
print "Traceback:\n" + traceback.format_exc()
print "Exception in Splunk logging handler: %s" % str(e)
<commit_msg>Add code to silence requests logger in the handler<commit_after>import logging
import socket
import traceback
from threading import Thread
import requests
class SplunkHandler(logging.Handler):
"""
A logging handler to send events to a Splunk Enterprise instance
"""
def __init__(self, host, port, username, password, index):
logging.Handler.__init__(self)
self.host = host
self.port = port
self.username = username
self.password = password
self.index = index
requests_log = logging.getLogger('requests')
requests_log.propagate = False
def emit(self, record):
thread = Thread(target=self._async_emit, args=(record, ))
thread.start()
def _async_emit(self, record):
try:
params = {
'host': socket.gethostname(),
'index': self.index,
'source': record.pathname,
'sourcetype': 'json'
}
url = 'https://%s:%s/services/receivers/simple' % (self.host, self.port)
payload = self.format(record)
auth = (self.username, self.password)
r = requests.post(
url,
auth=auth,
data=payload,
params=params
)
r.close()
except Exception, e:
print "Traceback:\n" + traceback.format_exc()
print "Exception in Splunk logging handler: %s" % str(e)
|
d9abb2f56720480169d394a2cadd3cb9a77ac4f6
|
app/main/views/frameworks.py
|
app/main/views/frameworks.py
|
from flask import jsonify
from sqlalchemy.types import String
from sqlalchemy import func
import datetime
from .. import main
from ...models import db, Framework, DraftService, Service, User, Supplier, SelectionAnswers, AuditEvent
@main.route('/frameworks', methods=['GET'])
def list_frameworks():
frameworks = Framework.query.all()
return jsonify(
frameworks=[f.serialize() for f in frameworks]
)
@main.route('/frameworks/g-cloud-7/stats', methods=['GET'])
def get_framework_stats():
seven_days_ago = datetime.datetime.utcnow() + datetime.timedelta(-7)
lot_column = DraftService.data['lot'].cast(String).label('lot')
return str({
'services_drafts': DraftService.query.filter(
DraftService.status == "not-submitted"
).count(),
'services_complete': DraftService.query.filter(
DraftService.status == "submitted"
).count(),
'services_by_lot': dict(db.session.query(
lot_column, func.count(lot_column)
).group_by(lot_column).all()),
'users': User.query.count(),
'active_users': User.query.filter(User.logged_in_at > seven_days_ago).count(),
'suppliers': Supplier.query.count(),
'suppliers_interested': AuditEvent.query.filter(AuditEvent.type == 'register_framework_interest').count(),
'suppliers_with_complete_declaration': SelectionAnswers.find_by_framework('g-cloud-7').count()
})
|
from flask import jsonify
from sqlalchemy.types import String
from sqlalchemy import func
import datetime
from .. import main
from ...models import db, Framework, DraftService, Service, User, Supplier, SelectionAnswers, AuditEvent
@main.route('/frameworks', methods=['GET'])
def list_frameworks():
frameworks = Framework.query.all()
return jsonify(
frameworks=[f.serialize() for f in frameworks]
)
@main.route('/frameworks/g-cloud-7/stats', methods=['GET'])
def get_framework_stats():
seven_days_ago = datetime.datetime.utcnow() + datetime.timedelta(-7)
lot_column = DraftService.data['lot'].cast(String).label('lot')
return str({
'services_by_status': dict(db.session.query(
DraftService.status, func.count(DraftService.status)
).group_by(DraftService.status)),
'services_by_lot': dict(db.session.query(
lot_column, func.count(lot_column)
).group_by(lot_column).all()),
'users': User.query.count(),
'active_users': User.query.filter(User.logged_in_at > seven_days_ago).count(),
'suppliers': Supplier.query.count(),
'suppliers_interested': AuditEvent.query.filter(AuditEvent.type == 'register_framework_interest').count(),
'suppliers_with_complete_declaration': SelectionAnswers.find_by_framework('g-cloud-7').count()
})
|
Use one query with group_by for service status
|
Use one query with group_by for service status
|
Python
|
mit
|
alphagov/digitalmarketplace-api,alphagov/digitalmarketplace-api,alphagov/digitalmarketplace-api
|
from flask import jsonify
from sqlalchemy.types import String
from sqlalchemy import func
import datetime
from .. import main
from ...models import db, Framework, DraftService, Service, User, Supplier, SelectionAnswers, AuditEvent
@main.route('/frameworks', methods=['GET'])
def list_frameworks():
frameworks = Framework.query.all()
return jsonify(
frameworks=[f.serialize() for f in frameworks]
)
@main.route('/frameworks/g-cloud-7/stats', methods=['GET'])
def get_framework_stats():
seven_days_ago = datetime.datetime.utcnow() + datetime.timedelta(-7)
lot_column = DraftService.data['lot'].cast(String).label('lot')
return str({
'services_drafts': DraftService.query.filter(
DraftService.status == "not-submitted"
).count(),
'services_complete': DraftService.query.filter(
DraftService.status == "submitted"
).count(),
'services_by_lot': dict(db.session.query(
lot_column, func.count(lot_column)
).group_by(lot_column).all()),
'users': User.query.count(),
'active_users': User.query.filter(User.logged_in_at > seven_days_ago).count(),
'suppliers': Supplier.query.count(),
'suppliers_interested': AuditEvent.query.filter(AuditEvent.type == 'register_framework_interest').count(),
'suppliers_with_complete_declaration': SelectionAnswers.find_by_framework('g-cloud-7').count()
})
Use one query with group_by for service status
|
from flask import jsonify
from sqlalchemy.types import String
from sqlalchemy import func
import datetime
from .. import main
from ...models import db, Framework, DraftService, Service, User, Supplier, SelectionAnswers, AuditEvent
@main.route('/frameworks', methods=['GET'])
def list_frameworks():
frameworks = Framework.query.all()
return jsonify(
frameworks=[f.serialize() for f in frameworks]
)
@main.route('/frameworks/g-cloud-7/stats', methods=['GET'])
def get_framework_stats():
seven_days_ago = datetime.datetime.utcnow() + datetime.timedelta(-7)
lot_column = DraftService.data['lot'].cast(String).label('lot')
return str({
'services_by_status': dict(db.session.query(
DraftService.status, func.count(DraftService.status)
).group_by(DraftService.status)),
'services_by_lot': dict(db.session.query(
lot_column, func.count(lot_column)
).group_by(lot_column).all()),
'users': User.query.count(),
'active_users': User.query.filter(User.logged_in_at > seven_days_ago).count(),
'suppliers': Supplier.query.count(),
'suppliers_interested': AuditEvent.query.filter(AuditEvent.type == 'register_framework_interest').count(),
'suppliers_with_complete_declaration': SelectionAnswers.find_by_framework('g-cloud-7').count()
})
|
<commit_before>from flask import jsonify
from sqlalchemy.types import String
from sqlalchemy import func
import datetime
from .. import main
from ...models import db, Framework, DraftService, Service, User, Supplier, SelectionAnswers, AuditEvent
@main.route('/frameworks', methods=['GET'])
def list_frameworks():
frameworks = Framework.query.all()
return jsonify(
frameworks=[f.serialize() for f in frameworks]
)
@main.route('/frameworks/g-cloud-7/stats', methods=['GET'])
def get_framework_stats():
seven_days_ago = datetime.datetime.utcnow() + datetime.timedelta(-7)
lot_column = DraftService.data['lot'].cast(String).label('lot')
return str({
'services_drafts': DraftService.query.filter(
DraftService.status == "not-submitted"
).count(),
'services_complete': DraftService.query.filter(
DraftService.status == "submitted"
).count(),
'services_by_lot': dict(db.session.query(
lot_column, func.count(lot_column)
).group_by(lot_column).all()),
'users': User.query.count(),
'active_users': User.query.filter(User.logged_in_at > seven_days_ago).count(),
'suppliers': Supplier.query.count(),
'suppliers_interested': AuditEvent.query.filter(AuditEvent.type == 'register_framework_interest').count(),
'suppliers_with_complete_declaration': SelectionAnswers.find_by_framework('g-cloud-7').count()
})
<commit_msg>Use one query with group_by for service status<commit_after>
|
from flask import jsonify
from sqlalchemy.types import String
from sqlalchemy import func
import datetime
from .. import main
from ...models import db, Framework, DraftService, Service, User, Supplier, SelectionAnswers, AuditEvent
@main.route('/frameworks', methods=['GET'])
def list_frameworks():
frameworks = Framework.query.all()
return jsonify(
frameworks=[f.serialize() for f in frameworks]
)
@main.route('/frameworks/g-cloud-7/stats', methods=['GET'])
def get_framework_stats():
seven_days_ago = datetime.datetime.utcnow() + datetime.timedelta(-7)
lot_column = DraftService.data['lot'].cast(String).label('lot')
return str({
'services_by_status': dict(db.session.query(
DraftService.status, func.count(DraftService.status)
).group_by(DraftService.status)),
'services_by_lot': dict(db.session.query(
lot_column, func.count(lot_column)
).group_by(lot_column).all()),
'users': User.query.count(),
'active_users': User.query.filter(User.logged_in_at > seven_days_ago).count(),
'suppliers': Supplier.query.count(),
'suppliers_interested': AuditEvent.query.filter(AuditEvent.type == 'register_framework_interest').count(),
'suppliers_with_complete_declaration': SelectionAnswers.find_by_framework('g-cloud-7').count()
})
|
from flask import jsonify
from sqlalchemy.types import String
from sqlalchemy import func
import datetime
from .. import main
from ...models import db, Framework, DraftService, Service, User, Supplier, SelectionAnswers, AuditEvent
@main.route('/frameworks', methods=['GET'])
def list_frameworks():
frameworks = Framework.query.all()
return jsonify(
frameworks=[f.serialize() for f in frameworks]
)
@main.route('/frameworks/g-cloud-7/stats', methods=['GET'])
def get_framework_stats():
seven_days_ago = datetime.datetime.utcnow() + datetime.timedelta(-7)
lot_column = DraftService.data['lot'].cast(String).label('lot')
return str({
'services_drafts': DraftService.query.filter(
DraftService.status == "not-submitted"
).count(),
'services_complete': DraftService.query.filter(
DraftService.status == "submitted"
).count(),
'services_by_lot': dict(db.session.query(
lot_column, func.count(lot_column)
).group_by(lot_column).all()),
'users': User.query.count(),
'active_users': User.query.filter(User.logged_in_at > seven_days_ago).count(),
'suppliers': Supplier.query.count(),
'suppliers_interested': AuditEvent.query.filter(AuditEvent.type == 'register_framework_interest').count(),
'suppliers_with_complete_declaration': SelectionAnswers.find_by_framework('g-cloud-7').count()
})
Use one query with group_by for service statusfrom flask import jsonify
from sqlalchemy.types import String
from sqlalchemy import func
import datetime
from .. import main
from ...models import db, Framework, DraftService, Service, User, Supplier, SelectionAnswers, AuditEvent
@main.route('/frameworks', methods=['GET'])
def list_frameworks():
frameworks = Framework.query.all()
return jsonify(
frameworks=[f.serialize() for f in frameworks]
)
@main.route('/frameworks/g-cloud-7/stats', methods=['GET'])
def get_framework_stats():
seven_days_ago = datetime.datetime.utcnow() + datetime.timedelta(-7)
lot_column = DraftService.data['lot'].cast(String).label('lot')
return str({
'services_by_status': dict(db.session.query(
DraftService.status, func.count(DraftService.status)
).group_by(DraftService.status)),
'services_by_lot': dict(db.session.query(
lot_column, func.count(lot_column)
).group_by(lot_column).all()),
'users': User.query.count(),
'active_users': User.query.filter(User.logged_in_at > seven_days_ago).count(),
'suppliers': Supplier.query.count(),
'suppliers_interested': AuditEvent.query.filter(AuditEvent.type == 'register_framework_interest').count(),
'suppliers_with_complete_declaration': SelectionAnswers.find_by_framework('g-cloud-7').count()
})
|
<commit_before>from flask import jsonify
from sqlalchemy.types import String
from sqlalchemy import func
import datetime
from .. import main
from ...models import db, Framework, DraftService, Service, User, Supplier, SelectionAnswers, AuditEvent
@main.route('/frameworks', methods=['GET'])
def list_frameworks():
frameworks = Framework.query.all()
return jsonify(
frameworks=[f.serialize() for f in frameworks]
)
@main.route('/frameworks/g-cloud-7/stats', methods=['GET'])
def get_framework_stats():
seven_days_ago = datetime.datetime.utcnow() + datetime.timedelta(-7)
lot_column = DraftService.data['lot'].cast(String).label('lot')
return str({
'services_drafts': DraftService.query.filter(
DraftService.status == "not-submitted"
).count(),
'services_complete': DraftService.query.filter(
DraftService.status == "submitted"
).count(),
'services_by_lot': dict(db.session.query(
lot_column, func.count(lot_column)
).group_by(lot_column).all()),
'users': User.query.count(),
'active_users': User.query.filter(User.logged_in_at > seven_days_ago).count(),
'suppliers': Supplier.query.count(),
'suppliers_interested': AuditEvent.query.filter(AuditEvent.type == 'register_framework_interest').count(),
'suppliers_with_complete_declaration': SelectionAnswers.find_by_framework('g-cloud-7').count()
})
<commit_msg>Use one query with group_by for service status<commit_after>from flask import jsonify
from sqlalchemy.types import String
from sqlalchemy import func
import datetime
from .. import main
from ...models import db, Framework, DraftService, Service, User, Supplier, SelectionAnswers, AuditEvent
@main.route('/frameworks', methods=['GET'])
def list_frameworks():
frameworks = Framework.query.all()
return jsonify(
frameworks=[f.serialize() for f in frameworks]
)
@main.route('/frameworks/g-cloud-7/stats', methods=['GET'])
def get_framework_stats():
seven_days_ago = datetime.datetime.utcnow() + datetime.timedelta(-7)
lot_column = DraftService.data['lot'].cast(String).label('lot')
return str({
'services_by_status': dict(db.session.query(
DraftService.status, func.count(DraftService.status)
).group_by(DraftService.status)),
'services_by_lot': dict(db.session.query(
lot_column, func.count(lot_column)
).group_by(lot_column).all()),
'users': User.query.count(),
'active_users': User.query.filter(User.logged_in_at > seven_days_ago).count(),
'suppliers': Supplier.query.count(),
'suppliers_interested': AuditEvent.query.filter(AuditEvent.type == 'register_framework_interest').count(),
'suppliers_with_complete_declaration': SelectionAnswers.find_by_framework('g-cloud-7').count()
})
|
a0903bb9fd988662269e9f2ef7e38acd877a63d5
|
src/nodeconductor_saltstack/saltstack/handlers.py
|
src/nodeconductor_saltstack/saltstack/handlers.py
|
from __future__ import unicode_literals
import logging
from .log import event_logger
logger = logging.getLogger(__name__)
def log_saltstack_property_created(sender, instance, created=False, **kwargs):
if created:
event_logger.saltstack_property.info(
'%s {property_name} has been created.' % instance.get_type_display_name(),
event_type='saltstack_property_creation_succeeded',
event_context={
'property': instance,
})
else:
event_logger.saltstack_property.info(
'%s {property_name} has been updated.' % instance.get_type_display_name(),
event_type='saltstack_property_update_succeeded',
event_context={
'property': instance,
})
def log_saltstack_property_deleted(sender, instance, **kwargs):
event_logger.saltstack_property.info(
'%s {property_name} has been deleted.' % instance.get_type_display_name(),
event_type='saltstack_property_deletion_succeeded',
event_context={
'property': instance,
})
|
from __future__ import unicode_literals
import logging
from .log import event_logger
logger = logging.getLogger(__name__)
def log_saltstack_property_created(sender, instance, created=False, **kwargs):
if created:
event_logger.saltstack_property.info(
'%s {property_name} has been created in {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_creation_succeeded',
event_context={
'property': instance,
})
else:
event_logger.saltstack_property.info(
'%s {property_name} has been updated in {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_update_succeeded',
event_context={
'property': instance,
})
def log_saltstack_property_deleted(sender, instance, **kwargs):
event_logger.saltstack_property.info(
'%s {property_name} has been deleted from {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_deletion_succeeded',
event_context={
'property': instance,
})
|
Add more details to event logs for property CRUD
|
Add more details to event logs for property CRUD
|
Python
|
mit
|
opennode/nodeconductor-saltstack
|
from __future__ import unicode_literals
import logging
from .log import event_logger
logger = logging.getLogger(__name__)
def log_saltstack_property_created(sender, instance, created=False, **kwargs):
if created:
event_logger.saltstack_property.info(
'%s {property_name} has been created.' % instance.get_type_display_name(),
event_type='saltstack_property_creation_succeeded',
event_context={
'property': instance,
})
else:
event_logger.saltstack_property.info(
'%s {property_name} has been updated.' % instance.get_type_display_name(),
event_type='saltstack_property_update_succeeded',
event_context={
'property': instance,
})
def log_saltstack_property_deleted(sender, instance, **kwargs):
event_logger.saltstack_property.info(
'%s {property_name} has been deleted.' % instance.get_type_display_name(),
event_type='saltstack_property_deletion_succeeded',
event_context={
'property': instance,
})
Add more details to event logs for property CRUD
|
from __future__ import unicode_literals
import logging
from .log import event_logger
logger = logging.getLogger(__name__)
def log_saltstack_property_created(sender, instance, created=False, **kwargs):
if created:
event_logger.saltstack_property.info(
'%s {property_name} has been created in {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_creation_succeeded',
event_context={
'property': instance,
})
else:
event_logger.saltstack_property.info(
'%s {property_name} has been updated in {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_update_succeeded',
event_context={
'property': instance,
})
def log_saltstack_property_deleted(sender, instance, **kwargs):
event_logger.saltstack_property.info(
'%s {property_name} has been deleted from {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_deletion_succeeded',
event_context={
'property': instance,
})
|
<commit_before>from __future__ import unicode_literals
import logging
from .log import event_logger
logger = logging.getLogger(__name__)
def log_saltstack_property_created(sender, instance, created=False, **kwargs):
if created:
event_logger.saltstack_property.info(
'%s {property_name} has been created.' % instance.get_type_display_name(),
event_type='saltstack_property_creation_succeeded',
event_context={
'property': instance,
})
else:
event_logger.saltstack_property.info(
'%s {property_name} has been updated.' % instance.get_type_display_name(),
event_type='saltstack_property_update_succeeded',
event_context={
'property': instance,
})
def log_saltstack_property_deleted(sender, instance, **kwargs):
event_logger.saltstack_property.info(
'%s {property_name} has been deleted.' % instance.get_type_display_name(),
event_type='saltstack_property_deletion_succeeded',
event_context={
'property': instance,
})
<commit_msg>Add more details to event logs for property CRUD<commit_after>
|
from __future__ import unicode_literals
import logging
from .log import event_logger
logger = logging.getLogger(__name__)
def log_saltstack_property_created(sender, instance, created=False, **kwargs):
if created:
event_logger.saltstack_property.info(
'%s {property_name} has been created in {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_creation_succeeded',
event_context={
'property': instance,
})
else:
event_logger.saltstack_property.info(
'%s {property_name} has been updated in {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_update_succeeded',
event_context={
'property': instance,
})
def log_saltstack_property_deleted(sender, instance, **kwargs):
event_logger.saltstack_property.info(
'%s {property_name} has been deleted from {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_deletion_succeeded',
event_context={
'property': instance,
})
|
from __future__ import unicode_literals
import logging
from .log import event_logger
logger = logging.getLogger(__name__)
def log_saltstack_property_created(sender, instance, created=False, **kwargs):
if created:
event_logger.saltstack_property.info(
'%s {property_name} has been created.' % instance.get_type_display_name(),
event_type='saltstack_property_creation_succeeded',
event_context={
'property': instance,
})
else:
event_logger.saltstack_property.info(
'%s {property_name} has been updated.' % instance.get_type_display_name(),
event_type='saltstack_property_update_succeeded',
event_context={
'property': instance,
})
def log_saltstack_property_deleted(sender, instance, **kwargs):
event_logger.saltstack_property.info(
'%s {property_name} has been deleted.' % instance.get_type_display_name(),
event_type='saltstack_property_deletion_succeeded',
event_context={
'property': instance,
})
Add more details to event logs for property CRUDfrom __future__ import unicode_literals
import logging
from .log import event_logger
logger = logging.getLogger(__name__)
def log_saltstack_property_created(sender, instance, created=False, **kwargs):
if created:
event_logger.saltstack_property.info(
'%s {property_name} has been created in {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_creation_succeeded',
event_context={
'property': instance,
})
else:
event_logger.saltstack_property.info(
'%s {property_name} has been updated in {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_update_succeeded',
event_context={
'property': instance,
})
def log_saltstack_property_deleted(sender, instance, **kwargs):
event_logger.saltstack_property.info(
'%s {property_name} has been deleted from {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_deletion_succeeded',
event_context={
'property': instance,
})
|
<commit_before>from __future__ import unicode_literals
import logging
from .log import event_logger
logger = logging.getLogger(__name__)
def log_saltstack_property_created(sender, instance, created=False, **kwargs):
if created:
event_logger.saltstack_property.info(
'%s {property_name} has been created.' % instance.get_type_display_name(),
event_type='saltstack_property_creation_succeeded',
event_context={
'property': instance,
})
else:
event_logger.saltstack_property.info(
'%s {property_name} has been updated.' % instance.get_type_display_name(),
event_type='saltstack_property_update_succeeded',
event_context={
'property': instance,
})
def log_saltstack_property_deleted(sender, instance, **kwargs):
event_logger.saltstack_property.info(
'%s {property_name} has been deleted.' % instance.get_type_display_name(),
event_type='saltstack_property_deletion_succeeded',
event_context={
'property': instance,
})
<commit_msg>Add more details to event logs for property CRUD<commit_after>from __future__ import unicode_literals
import logging
from .log import event_logger
logger = logging.getLogger(__name__)
def log_saltstack_property_created(sender, instance, created=False, **kwargs):
if created:
event_logger.saltstack_property.info(
'%s {property_name} has been created in {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_creation_succeeded',
event_context={
'property': instance,
})
else:
event_logger.saltstack_property.info(
'%s {property_name} has been updated in {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_update_succeeded',
event_context={
'property': instance,
})
def log_saltstack_property_deleted(sender, instance, **kwargs):
event_logger.saltstack_property.info(
'%s {property_name} has been deleted from {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_deletion_succeeded',
event_context={
'property': instance,
})
|
18318b3bb431c8a5ec9261d6dd190997613cf1ed
|
src/pytest_django_casperjs/tests/test_fixtures.py
|
src/pytest_django_casperjs/tests/test_fixtures.py
|
from __future__ import with_statement
import django
import pytest
from django.conf import settings as real_settings
from django.utils.encoding import force_text
from django.test.client import Client, RequestFactory
from .app.models import Item
from pytest_django_casperjs.compat import urlopen
django # Avoid pyflakes complaints
class TestCasperJSLiveServer:
# Partially based on the LiveServer test case from pytest_django'
pytestmark = [
pytest.mark.django_db()
]
def test_url(self, casper_js):
assert casper_js.url == force_text(casper_js)
def test_db_changes_visibility(self, casper_js):
response_data = urlopen(casper_js + '/item_count/').read()
assert force_text(response_data) == 'Item count: 0'
Item.objects.create(name='foo')
response_data = urlopen(casper_js + '/item_count/').read()
assert force_text(response_data) == 'Item count: 1'
def test_fixture_db(self, db, casper_js):
Item.objects.create(name='foo')
response_data = urlopen(casper_js + '/item_count/').read()
assert force_text(response_data) == 'Item count: 1'
@pytest.fixture
def item(self):
# This has not requested database access so should fail.
# Unfortunately the _live_server_helper autouse fixture makes this
# test work.
with pytest.raises(pytest.fail.Exception):
Item.objects.create(name='foo')
@pytest.mark.xfail
def test_item(self, item, casper_js):
# test should fail/pass in setup
pass
@pytest.fixture
def item_db(self, db):
return Item.objects.create(name='foo')
def test_item_db(self, item_db, casper_js):
response_data = urlopen(casper_js + '/item_count/').read()
assert force_text(response_data) == 'Item count: 1'
|
from __future__ import with_statement
import django
import pytest
from django.conf import settings as real_settings
from django.utils.encoding import force_text
from django.test.client import Client, RequestFactory
from .app.models import Item
from pytest_django_casperjs.compat import urlopen
django # Avoid pyflakes complaints
@pytest.mark.django_db
class TestCasperJSLiveServer:
# Partially based on the LiveServer test case from pytest_django'
def test_url(self, casper_js):
assert casper_js.url == force_text(casper_js)
|
Remove more irrelevant tests, those will be replaced with proper casperjs tests
|
Remove more irrelevant tests, those will be replaced with proper casperjs tests
|
Python
|
bsd-3-clause
|
EnTeQuAk/pytest-django-casperjs
|
from __future__ import with_statement
import django
import pytest
from django.conf import settings as real_settings
from django.utils.encoding import force_text
from django.test.client import Client, RequestFactory
from .app.models import Item
from pytest_django_casperjs.compat import urlopen
django # Avoid pyflakes complaints
class TestCasperJSLiveServer:
# Partially based on the LiveServer test case from pytest_django'
pytestmark = [
pytest.mark.django_db()
]
def test_url(self, casper_js):
assert casper_js.url == force_text(casper_js)
def test_db_changes_visibility(self, casper_js):
response_data = urlopen(casper_js + '/item_count/').read()
assert force_text(response_data) == 'Item count: 0'
Item.objects.create(name='foo')
response_data = urlopen(casper_js + '/item_count/').read()
assert force_text(response_data) == 'Item count: 1'
def test_fixture_db(self, db, casper_js):
Item.objects.create(name='foo')
response_data = urlopen(casper_js + '/item_count/').read()
assert force_text(response_data) == 'Item count: 1'
@pytest.fixture
def item(self):
# This has not requested database access so should fail.
# Unfortunately the _live_server_helper autouse fixture makes this
# test work.
with pytest.raises(pytest.fail.Exception):
Item.objects.create(name='foo')
@pytest.mark.xfail
def test_item(self, item, casper_js):
# test should fail/pass in setup
pass
@pytest.fixture
def item_db(self, db):
return Item.objects.create(name='foo')
def test_item_db(self, item_db, casper_js):
response_data = urlopen(casper_js + '/item_count/').read()
assert force_text(response_data) == 'Item count: 1'
Remove more irrelevant tests, those will be replaced with proper casperjs tests
|
from __future__ import with_statement
import django
import pytest
from django.conf import settings as real_settings
from django.utils.encoding import force_text
from django.test.client import Client, RequestFactory
from .app.models import Item
from pytest_django_casperjs.compat import urlopen
django # Avoid pyflakes complaints
@pytest.mark.django_db
class TestCasperJSLiveServer:
# Partially based on the LiveServer test case from pytest_django'
def test_url(self, casper_js):
assert casper_js.url == force_text(casper_js)
|
<commit_before>from __future__ import with_statement
import django
import pytest
from django.conf import settings as real_settings
from django.utils.encoding import force_text
from django.test.client import Client, RequestFactory
from .app.models import Item
from pytest_django_casperjs.compat import urlopen
django # Avoid pyflakes complaints
class TestCasperJSLiveServer:
# Partially based on the LiveServer test case from pytest_django'
pytestmark = [
pytest.mark.django_db()
]
def test_url(self, casper_js):
assert casper_js.url == force_text(casper_js)
def test_db_changes_visibility(self, casper_js):
response_data = urlopen(casper_js + '/item_count/').read()
assert force_text(response_data) == 'Item count: 0'
Item.objects.create(name='foo')
response_data = urlopen(casper_js + '/item_count/').read()
assert force_text(response_data) == 'Item count: 1'
def test_fixture_db(self, db, casper_js):
Item.objects.create(name='foo')
response_data = urlopen(casper_js + '/item_count/').read()
assert force_text(response_data) == 'Item count: 1'
@pytest.fixture
def item(self):
# This has not requested database access so should fail.
# Unfortunately the _live_server_helper autouse fixture makes this
# test work.
with pytest.raises(pytest.fail.Exception):
Item.objects.create(name='foo')
@pytest.mark.xfail
def test_item(self, item, casper_js):
# test should fail/pass in setup
pass
@pytest.fixture
def item_db(self, db):
return Item.objects.create(name='foo')
def test_item_db(self, item_db, casper_js):
response_data = urlopen(casper_js + '/item_count/').read()
assert force_text(response_data) == 'Item count: 1'
<commit_msg>Remove more irrelevant tests, those will be replaced with proper casperjs tests<commit_after>
|
from __future__ import with_statement
import django
import pytest
from django.conf import settings as real_settings
from django.utils.encoding import force_text
from django.test.client import Client, RequestFactory
from .app.models import Item
from pytest_django_casperjs.compat import urlopen
django # Avoid pyflakes complaints
@pytest.mark.django_db
class TestCasperJSLiveServer:
# Partially based on the LiveServer test case from pytest_django'
def test_url(self, casper_js):
assert casper_js.url == force_text(casper_js)
|
from __future__ import with_statement
import django
import pytest
from django.conf import settings as real_settings
from django.utils.encoding import force_text
from django.test.client import Client, RequestFactory
from .app.models import Item
from pytest_django_casperjs.compat import urlopen
django # Avoid pyflakes complaints
class TestCasperJSLiveServer:
# Partially based on the LiveServer test case from pytest_django'
pytestmark = [
pytest.mark.django_db()
]
def test_url(self, casper_js):
assert casper_js.url == force_text(casper_js)
def test_db_changes_visibility(self, casper_js):
response_data = urlopen(casper_js + '/item_count/').read()
assert force_text(response_data) == 'Item count: 0'
Item.objects.create(name='foo')
response_data = urlopen(casper_js + '/item_count/').read()
assert force_text(response_data) == 'Item count: 1'
def test_fixture_db(self, db, casper_js):
Item.objects.create(name='foo')
response_data = urlopen(casper_js + '/item_count/').read()
assert force_text(response_data) == 'Item count: 1'
@pytest.fixture
def item(self):
# This has not requested database access so should fail.
# Unfortunately the _live_server_helper autouse fixture makes this
# test work.
with pytest.raises(pytest.fail.Exception):
Item.objects.create(name='foo')
@pytest.mark.xfail
def test_item(self, item, casper_js):
# test should fail/pass in setup
pass
@pytest.fixture
def item_db(self, db):
return Item.objects.create(name='foo')
def test_item_db(self, item_db, casper_js):
response_data = urlopen(casper_js + '/item_count/').read()
assert force_text(response_data) == 'Item count: 1'
Remove more irrelevant tests, those will be replaced with proper casperjs testsfrom __future__ import with_statement
import django
import pytest
from django.conf import settings as real_settings
from django.utils.encoding import force_text
from django.test.client import Client, RequestFactory
from .app.models import Item
from pytest_django_casperjs.compat import urlopen
django # Avoid pyflakes complaints
@pytest.mark.django_db
class TestCasperJSLiveServer:
# Partially based on the LiveServer test case from pytest_django'
def test_url(self, casper_js):
assert casper_js.url == force_text(casper_js)
|
<commit_before>from __future__ import with_statement
import django
import pytest
from django.conf import settings as real_settings
from django.utils.encoding import force_text
from django.test.client import Client, RequestFactory
from .app.models import Item
from pytest_django_casperjs.compat import urlopen
django # Avoid pyflakes complaints
class TestCasperJSLiveServer:
# Partially based on the LiveServer test case from pytest_django'
pytestmark = [
pytest.mark.django_db()
]
def test_url(self, casper_js):
assert casper_js.url == force_text(casper_js)
def test_db_changes_visibility(self, casper_js):
response_data = urlopen(casper_js + '/item_count/').read()
assert force_text(response_data) == 'Item count: 0'
Item.objects.create(name='foo')
response_data = urlopen(casper_js + '/item_count/').read()
assert force_text(response_data) == 'Item count: 1'
def test_fixture_db(self, db, casper_js):
Item.objects.create(name='foo')
response_data = urlopen(casper_js + '/item_count/').read()
assert force_text(response_data) == 'Item count: 1'
@pytest.fixture
def item(self):
# This has not requested database access so should fail.
# Unfortunately the _live_server_helper autouse fixture makes this
# test work.
with pytest.raises(pytest.fail.Exception):
Item.objects.create(name='foo')
@pytest.mark.xfail
def test_item(self, item, casper_js):
# test should fail/pass in setup
pass
@pytest.fixture
def item_db(self, db):
return Item.objects.create(name='foo')
def test_item_db(self, item_db, casper_js):
response_data = urlopen(casper_js + '/item_count/').read()
assert force_text(response_data) == 'Item count: 1'
<commit_msg>Remove more irrelevant tests, those will be replaced with proper casperjs tests<commit_after>from __future__ import with_statement
import django
import pytest
from django.conf import settings as real_settings
from django.utils.encoding import force_text
from django.test.client import Client, RequestFactory
from .app.models import Item
from pytest_django_casperjs.compat import urlopen
django # Avoid pyflakes complaints
@pytest.mark.django_db
class TestCasperJSLiveServer:
# Partially based on the LiveServer test case from pytest_django'
def test_url(self, casper_js):
assert casper_js.url == force_text(casper_js)
|
8696885e9f1535bdfb8dbc0e285c67d1e6d41a95
|
datasets/admin.py
|
datasets/admin.py
|
from django.contrib import admin
from datasets.models import Dataset, Sound, Annotation, Vote, Taxonomy, DatasetRelease, TaxonomyNode
admin.site.register(Dataset)
admin.site.register(Sound)
admin.site.register(Annotation)
admin.site.register(Vote)
admin.site.register(Taxonomy)
admin.site.register(DatasetRelease)
admin.site.register(TaxonomyNode)
|
from django.contrib import admin
from datasets.models import Dataset, Sound, Annotation, Vote, Taxonomy, DatasetRelease, TaxonomyNode
class TaxonomyNodeAdmin(admin.ModelAdmin):
fields = ('node_id', 'name', 'description', 'citation_uri', 'faq')
admin.site.register(Dataset)
admin.site.register(Sound)
admin.site.register(Annotation)
admin.site.register(Vote)
admin.site.register(Taxonomy)
admin.site.register(DatasetRelease)
admin.site.register(TaxonomyNode, TaxonomyNodeAdmin)
|
Add custom Admin model TaxonomyNode, hide freesound ex
|
Add custom Admin model TaxonomyNode, hide freesound ex
|
Python
|
agpl-3.0
|
MTG/freesound-datasets,MTG/freesound-datasets,MTG/freesound-datasets,MTG/freesound-datasets
|
from django.contrib import admin
from datasets.models import Dataset, Sound, Annotation, Vote, Taxonomy, DatasetRelease, TaxonomyNode
admin.site.register(Dataset)
admin.site.register(Sound)
admin.site.register(Annotation)
admin.site.register(Vote)
admin.site.register(Taxonomy)
admin.site.register(DatasetRelease)
admin.site.register(TaxonomyNode)
Add custom Admin model TaxonomyNode, hide freesound ex
|
from django.contrib import admin
from datasets.models import Dataset, Sound, Annotation, Vote, Taxonomy, DatasetRelease, TaxonomyNode
class TaxonomyNodeAdmin(admin.ModelAdmin):
fields = ('node_id', 'name', 'description', 'citation_uri', 'faq')
admin.site.register(Dataset)
admin.site.register(Sound)
admin.site.register(Annotation)
admin.site.register(Vote)
admin.site.register(Taxonomy)
admin.site.register(DatasetRelease)
admin.site.register(TaxonomyNode, TaxonomyNodeAdmin)
|
<commit_before>from django.contrib import admin
from datasets.models import Dataset, Sound, Annotation, Vote, Taxonomy, DatasetRelease, TaxonomyNode
admin.site.register(Dataset)
admin.site.register(Sound)
admin.site.register(Annotation)
admin.site.register(Vote)
admin.site.register(Taxonomy)
admin.site.register(DatasetRelease)
admin.site.register(TaxonomyNode)
<commit_msg>Add custom Admin model TaxonomyNode, hide freesound ex<commit_after>
|
from django.contrib import admin
from datasets.models import Dataset, Sound, Annotation, Vote, Taxonomy, DatasetRelease, TaxonomyNode
class TaxonomyNodeAdmin(admin.ModelAdmin):
fields = ('node_id', 'name', 'description', 'citation_uri', 'faq')
admin.site.register(Dataset)
admin.site.register(Sound)
admin.site.register(Annotation)
admin.site.register(Vote)
admin.site.register(Taxonomy)
admin.site.register(DatasetRelease)
admin.site.register(TaxonomyNode, TaxonomyNodeAdmin)
|
from django.contrib import admin
from datasets.models import Dataset, Sound, Annotation, Vote, Taxonomy, DatasetRelease, TaxonomyNode
admin.site.register(Dataset)
admin.site.register(Sound)
admin.site.register(Annotation)
admin.site.register(Vote)
admin.site.register(Taxonomy)
admin.site.register(DatasetRelease)
admin.site.register(TaxonomyNode)
Add custom Admin model TaxonomyNode, hide freesound exfrom django.contrib import admin
from datasets.models import Dataset, Sound, Annotation, Vote, Taxonomy, DatasetRelease, TaxonomyNode
class TaxonomyNodeAdmin(admin.ModelAdmin):
fields = ('node_id', 'name', 'description', 'citation_uri', 'faq')
admin.site.register(Dataset)
admin.site.register(Sound)
admin.site.register(Annotation)
admin.site.register(Vote)
admin.site.register(Taxonomy)
admin.site.register(DatasetRelease)
admin.site.register(TaxonomyNode, TaxonomyNodeAdmin)
|
<commit_before>from django.contrib import admin
from datasets.models import Dataset, Sound, Annotation, Vote, Taxonomy, DatasetRelease, TaxonomyNode
admin.site.register(Dataset)
admin.site.register(Sound)
admin.site.register(Annotation)
admin.site.register(Vote)
admin.site.register(Taxonomy)
admin.site.register(DatasetRelease)
admin.site.register(TaxonomyNode)
<commit_msg>Add custom Admin model TaxonomyNode, hide freesound ex<commit_after>from django.contrib import admin
from datasets.models import Dataset, Sound, Annotation, Vote, Taxonomy, DatasetRelease, TaxonomyNode
class TaxonomyNodeAdmin(admin.ModelAdmin):
fields = ('node_id', 'name', 'description', 'citation_uri', 'faq')
admin.site.register(Dataset)
admin.site.register(Sound)
admin.site.register(Annotation)
admin.site.register(Vote)
admin.site.register(Taxonomy)
admin.site.register(DatasetRelease)
admin.site.register(TaxonomyNode, TaxonomyNodeAdmin)
|
a14256e715d51728ad4c2bde7ec52f13def6b2a6
|
director/views.py
|
director/views.py
|
from django.shortcuts import redirect
from django.urls import reverse
from django.views.generic import View
class HomeView(View):
def get(self, *args, **kwargs):
if self.request.user.is_authenticated:
return redirect(reverse('project_list'))
else:
return redirect(reverse('beta_token'))
|
from django.shortcuts import redirect
from django.urls import reverse
from accounts.views import BetaTokenView
class HomeView(BetaTokenView):
"""
Home page view.
Care needs to be taken that this view returns a 200 response (not a redirect)
for unauthenticated users. This is because GCP load balancers ping the / path
as a health check and will fail if anything other than a 200 is returned.
"""
def get(self, *args, **kwargs):
if self.request.user.is_authenticated:
return redirect(reverse('project_list'))
else:
return super().get(*args, **kwargs)
|
Fix home view so it returns 200 for unauthenticated health check
|
Fix home view so it returns 200 for unauthenticated health check
|
Python
|
apache-2.0
|
stencila/hub,stencila/hub,stencila/hub,stencila/hub,stencila/hub
|
from django.shortcuts import redirect
from django.urls import reverse
from django.views.generic import View
class HomeView(View):
def get(self, *args, **kwargs):
if self.request.user.is_authenticated:
return redirect(reverse('project_list'))
else:
return redirect(reverse('beta_token'))
Fix home view so it returns 200 for unauthenticated health check
|
from django.shortcuts import redirect
from django.urls import reverse
from accounts.views import BetaTokenView
class HomeView(BetaTokenView):
"""
Home page view.
Care needs to be taken that this view returns a 200 response (not a redirect)
for unauthenticated users. This is because GCP load balancers ping the / path
as a health check and will fail if anything other than a 200 is returned.
"""
def get(self, *args, **kwargs):
if self.request.user.is_authenticated:
return redirect(reverse('project_list'))
else:
return super().get(*args, **kwargs)
|
<commit_before>from django.shortcuts import redirect
from django.urls import reverse
from django.views.generic import View
class HomeView(View):
def get(self, *args, **kwargs):
if self.request.user.is_authenticated:
return redirect(reverse('project_list'))
else:
return redirect(reverse('beta_token'))
<commit_msg>Fix home view so it returns 200 for unauthenticated health check<commit_after>
|
from django.shortcuts import redirect
from django.urls import reverse
from accounts.views import BetaTokenView
class HomeView(BetaTokenView):
"""
Home page view.
Care needs to be taken that this view returns a 200 response (not a redirect)
for unauthenticated users. This is because GCP load balancers ping the / path
as a health check and will fail if anything other than a 200 is returned.
"""
def get(self, *args, **kwargs):
if self.request.user.is_authenticated:
return redirect(reverse('project_list'))
else:
return super().get(*args, **kwargs)
|
from django.shortcuts import redirect
from django.urls import reverse
from django.views.generic import View
class HomeView(View):
def get(self, *args, **kwargs):
if self.request.user.is_authenticated:
return redirect(reverse('project_list'))
else:
return redirect(reverse('beta_token'))
Fix home view so it returns 200 for unauthenticated health checkfrom django.shortcuts import redirect
from django.urls import reverse
from accounts.views import BetaTokenView
class HomeView(BetaTokenView):
"""
Home page view.
Care needs to be taken that this view returns a 200 response (not a redirect)
for unauthenticated users. This is because GCP load balancers ping the / path
as a health check and will fail if anything other than a 200 is returned.
"""
def get(self, *args, **kwargs):
if self.request.user.is_authenticated:
return redirect(reverse('project_list'))
else:
return super().get(*args, **kwargs)
|
<commit_before>from django.shortcuts import redirect
from django.urls import reverse
from django.views.generic import View
class HomeView(View):
def get(self, *args, **kwargs):
if self.request.user.is_authenticated:
return redirect(reverse('project_list'))
else:
return redirect(reverse('beta_token'))
<commit_msg>Fix home view so it returns 200 for unauthenticated health check<commit_after>from django.shortcuts import redirect
from django.urls import reverse
from accounts.views import BetaTokenView
class HomeView(BetaTokenView):
"""
Home page view.
Care needs to be taken that this view returns a 200 response (not a redirect)
for unauthenticated users. This is because GCP load balancers ping the / path
as a health check and will fail if anything other than a 200 is returned.
"""
def get(self, *args, **kwargs):
if self.request.user.is_authenticated:
return redirect(reverse('project_list'))
else:
return super().get(*args, **kwargs)
|
95ceea4ce45d531c277c00456639a42cfd18f129
|
djangae/patches/json.py
|
djangae/patches/json.py
|
from functools import wraps
def additional_type_handler(func):
@wraps(func)
def _wrapper(self, o):
if isinstance(o, set):
# Return a string representing a set
return "{" + ",".join([repr(x) for x in o]) + "}"
else:
return func(self, o)
return _wrapper
def patch():
"""
This patches Djangos JSON encoder so it can deal with
set(). This is necessary because otherwise we can't
serialize SetFields
"""
from django.core.serializers.json import DjangoJSONEncoder
DjangoJSONEncoder.default = additional_type_handler(DjangoJSONEncoder.default)
|
from functools import wraps
def additional_type_handler(func):
@wraps(func)
def _wrapper(self, o):
if isinstance(o, set):
# Return a string representing a set
return "{" + ",".join([str(x) for x in o]) + "}"
else:
return func(self, o)
return _wrapper
def patch():
"""
This patches Djangos JSON encoder so it can deal with
set(). This is necessary because otherwise we can't
serialize SetFields
"""
from django.core.serializers.json import DjangoJSONEncoder
DjangoJSONEncoder.default = additional_type_handler(DjangoJSONEncoder.default)
|
Use str() not repr() to avoid trailing L on longs
|
Use str() not repr() to avoid trailing L on longs
|
Python
|
bsd-3-clause
|
grzes/djangae,leekchan/djangae,SiPiggles/djangae,potatolondon/djangae,martinogden/djangae,potatolondon/djangae,jscissr/djangae,chargrizzle/djangae,trik/djangae,leekchan/djangae,kirberich/djangae,SiPiggles/djangae,chargrizzle/djangae,wangjun/djangae,jscissr/djangae,martinogden/djangae,pablorecio/djangae,jscissr/djangae,pablorecio/djangae,kirberich/djangae,asendecka/djangae,grzes/djangae,kirberich/djangae,armirusco/djangae,grzes/djangae,wangjun/djangae,armirusco/djangae,asendecka/djangae,trik/djangae,leekchan/djangae,asendecka/djangae,chargrizzle/djangae,martinogden/djangae,pablorecio/djangae,SiPiggles/djangae,trik/djangae,armirusco/djangae,wangjun/djangae
|
from functools import wraps
def additional_type_handler(func):
@wraps(func)
def _wrapper(self, o):
if isinstance(o, set):
# Return a string representing a set
return "{" + ",".join([repr(x) for x in o]) + "}"
else:
return func(self, o)
return _wrapper
def patch():
"""
This patches Djangos JSON encoder so it can deal with
set(). This is necessary because otherwise we can't
serialize SetFields
"""
from django.core.serializers.json import DjangoJSONEncoder
DjangoJSONEncoder.default = additional_type_handler(DjangoJSONEncoder.default)
Use str() not repr() to avoid trailing L on longs
|
from functools import wraps
def additional_type_handler(func):
@wraps(func)
def _wrapper(self, o):
if isinstance(o, set):
# Return a string representing a set
return "{" + ",".join([str(x) for x in o]) + "}"
else:
return func(self, o)
return _wrapper
def patch():
"""
This patches Djangos JSON encoder so it can deal with
set(). This is necessary because otherwise we can't
serialize SetFields
"""
from django.core.serializers.json import DjangoJSONEncoder
DjangoJSONEncoder.default = additional_type_handler(DjangoJSONEncoder.default)
|
<commit_before>from functools import wraps
def additional_type_handler(func):
@wraps(func)
def _wrapper(self, o):
if isinstance(o, set):
# Return a string representing a set
return "{" + ",".join([repr(x) for x in o]) + "}"
else:
return func(self, o)
return _wrapper
def patch():
"""
This patches Djangos JSON encoder so it can deal with
set(). This is necessary because otherwise we can't
serialize SetFields
"""
from django.core.serializers.json import DjangoJSONEncoder
DjangoJSONEncoder.default = additional_type_handler(DjangoJSONEncoder.default)
<commit_msg>Use str() not repr() to avoid trailing L on longs<commit_after>
|
from functools import wraps
def additional_type_handler(func):
@wraps(func)
def _wrapper(self, o):
if isinstance(o, set):
# Return a string representing a set
return "{" + ",".join([str(x) for x in o]) + "}"
else:
return func(self, o)
return _wrapper
def patch():
"""
This patches Djangos JSON encoder so it can deal with
set(). This is necessary because otherwise we can't
serialize SetFields
"""
from django.core.serializers.json import DjangoJSONEncoder
DjangoJSONEncoder.default = additional_type_handler(DjangoJSONEncoder.default)
|
from functools import wraps
def additional_type_handler(func):
@wraps(func)
def _wrapper(self, o):
if isinstance(o, set):
# Return a string representing a set
return "{" + ",".join([repr(x) for x in o]) + "}"
else:
return func(self, o)
return _wrapper
def patch():
"""
This patches Djangos JSON encoder so it can deal with
set(). This is necessary because otherwise we can't
serialize SetFields
"""
from django.core.serializers.json import DjangoJSONEncoder
DjangoJSONEncoder.default = additional_type_handler(DjangoJSONEncoder.default)
Use str() not repr() to avoid trailing L on longsfrom functools import wraps
def additional_type_handler(func):
@wraps(func)
def _wrapper(self, o):
if isinstance(o, set):
# Return a string representing a set
return "{" + ",".join([str(x) for x in o]) + "}"
else:
return func(self, o)
return _wrapper
def patch():
"""
This patches Djangos JSON encoder so it can deal with
set(). This is necessary because otherwise we can't
serialize SetFields
"""
from django.core.serializers.json import DjangoJSONEncoder
DjangoJSONEncoder.default = additional_type_handler(DjangoJSONEncoder.default)
|
<commit_before>from functools import wraps
def additional_type_handler(func):
@wraps(func)
def _wrapper(self, o):
if isinstance(o, set):
# Return a string representing a set
return "{" + ",".join([repr(x) for x in o]) + "}"
else:
return func(self, o)
return _wrapper
def patch():
"""
This patches Djangos JSON encoder so it can deal with
set(). This is necessary because otherwise we can't
serialize SetFields
"""
from django.core.serializers.json import DjangoJSONEncoder
DjangoJSONEncoder.default = additional_type_handler(DjangoJSONEncoder.default)
<commit_msg>Use str() not repr() to avoid trailing L on longs<commit_after>from functools import wraps
def additional_type_handler(func):
@wraps(func)
def _wrapper(self, o):
if isinstance(o, set):
# Return a string representing a set
return "{" + ",".join([str(x) for x in o]) + "}"
else:
return func(self, o)
return _wrapper
def patch():
"""
This patches Djangos JSON encoder so it can deal with
set(). This is necessary because otherwise we can't
serialize SetFields
"""
from django.core.serializers.json import DjangoJSONEncoder
DjangoJSONEncoder.default = additional_type_handler(DjangoJSONEncoder.default)
|
7bd19241e0502789bed482291554e8341034d377
|
bpmodule/testing/__init__.py
|
bpmodule/testing/__init__.py
|
from .modinfo import *
# SO file
from .testing import *
# For output
from bpmodule import output
def PrintHeader(s):
output.Output(output.Line("="))
output.Output("%1%\n", s)
output.Output(output.Line("="))
def PrintResults(nfailed):
output.Output("\n\n")
if nfailed > 0:
output.Output(output.Line("="))
output.Error("RESULTS: %1% failed\n", nfailed)
output.Output(output.Line("="))
else:
output.Output(output.Line("="))
output.Success("RESULTS: %1% failed\n", nfailed)
output.Output(output.Line("="))
output.Output("\n\n")
|
from .modinfo import *
# SO file
from .testing import *
# For output
from bpmodule.output import Output, Error, Warning, Success, Debug, Line
##################
# For testing on the python side
##################
def PyTestFunc(itest, desc, expected, func, *args):
fmt = "%|1$5| : %|2$-5| %|3$-5| %|4$-9| : %5%\n"
try:
func(*args)
except Exception as e:
if expected == False:
Output(fmt, itest, "False", "False", "Success", desc)
return 0
else:
Debug(str(e) + "\n")
Error(fmt, itest, "True", "False", "FAILED", desc)
return 1
except:
if expected == False:
Output(fmt, itest, "False", "False", "Success", desc)
return 0
else:
Error(fmt, itest, "True", "False", "FAILED", desc)
return 1
if expected == True:
Output(fmt, itest, "True", "True", "Success", desc)
return 0
else:
Error(fmt, itest, "False", "True", "FAILED", desc)
return 1
def PrintHeader(s):
Output(Line("="))
Output("%1%\n", s)
Output(Line("="))
def PrintResults(nfailed):
Output("\n\n")
if nfailed > 0:
Output(Line("="))
Error("RESULTS: %1% failed\n", nfailed)
Output(Line("="))
else:
Output(Line("="))
Success("RESULTS: %1% failed\n", nfailed)
Output(Line("="))
Output("\n\n")
|
Add testing function from python
|
Add testing function from python
|
Python
|
bsd-3-clause
|
pulsar-chem/Pulsar-Core,pulsar-chem/Pulsar-Core,pulsar-chem/Pulsar-Core,pulsar-chem/Pulsar-Core
|
from .modinfo import *
# SO file
from .testing import *
# For output
from bpmodule import output
def PrintHeader(s):
output.Output(output.Line("="))
output.Output("%1%\n", s)
output.Output(output.Line("="))
def PrintResults(nfailed):
output.Output("\n\n")
if nfailed > 0:
output.Output(output.Line("="))
output.Error("RESULTS: %1% failed\n", nfailed)
output.Output(output.Line("="))
else:
output.Output(output.Line("="))
output.Success("RESULTS: %1% failed\n", nfailed)
output.Output(output.Line("="))
output.Output("\n\n")
Add testing function from python
|
from .modinfo import *
# SO file
from .testing import *
# For output
from bpmodule.output import Output, Error, Warning, Success, Debug, Line
##################
# For testing on the python side
##################
def PyTestFunc(itest, desc, expected, func, *args):
fmt = "%|1$5| : %|2$-5| %|3$-5| %|4$-9| : %5%\n"
try:
func(*args)
except Exception as e:
if expected == False:
Output(fmt, itest, "False", "False", "Success", desc)
return 0
else:
Debug(str(e) + "\n")
Error(fmt, itest, "True", "False", "FAILED", desc)
return 1
except:
if expected == False:
Output(fmt, itest, "False", "False", "Success", desc)
return 0
else:
Error(fmt, itest, "True", "False", "FAILED", desc)
return 1
if expected == True:
Output(fmt, itest, "True", "True", "Success", desc)
return 0
else:
Error(fmt, itest, "False", "True", "FAILED", desc)
return 1
def PrintHeader(s):
Output(Line("="))
Output("%1%\n", s)
Output(Line("="))
def PrintResults(nfailed):
Output("\n\n")
if nfailed > 0:
Output(Line("="))
Error("RESULTS: %1% failed\n", nfailed)
Output(Line("="))
else:
Output(Line("="))
Success("RESULTS: %1% failed\n", nfailed)
Output(Line("="))
Output("\n\n")
|
<commit_before>from .modinfo import *
# SO file
from .testing import *
# For output
from bpmodule import output
def PrintHeader(s):
output.Output(output.Line("="))
output.Output("%1%\n", s)
output.Output(output.Line("="))
def PrintResults(nfailed):
output.Output("\n\n")
if nfailed > 0:
output.Output(output.Line("="))
output.Error("RESULTS: %1% failed\n", nfailed)
output.Output(output.Line("="))
else:
output.Output(output.Line("="))
output.Success("RESULTS: %1% failed\n", nfailed)
output.Output(output.Line("="))
output.Output("\n\n")
<commit_msg>Add testing function from python<commit_after>
|
from .modinfo import *
# SO file
from .testing import *
# For output
from bpmodule.output import Output, Error, Warning, Success, Debug, Line
##################
# For testing on the python side
##################
def PyTestFunc(itest, desc, expected, func, *args):
fmt = "%|1$5| : %|2$-5| %|3$-5| %|4$-9| : %5%\n"
try:
func(*args)
except Exception as e:
if expected == False:
Output(fmt, itest, "False", "False", "Success", desc)
return 0
else:
Debug(str(e) + "\n")
Error(fmt, itest, "True", "False", "FAILED", desc)
return 1
except:
if expected == False:
Output(fmt, itest, "False", "False", "Success", desc)
return 0
else:
Error(fmt, itest, "True", "False", "FAILED", desc)
return 1
if expected == True:
Output(fmt, itest, "True", "True", "Success", desc)
return 0
else:
Error(fmt, itest, "False", "True", "FAILED", desc)
return 1
def PrintHeader(s):
Output(Line("="))
Output("%1%\n", s)
Output(Line("="))
def PrintResults(nfailed):
Output("\n\n")
if nfailed > 0:
Output(Line("="))
Error("RESULTS: %1% failed\n", nfailed)
Output(Line("="))
else:
Output(Line("="))
Success("RESULTS: %1% failed\n", nfailed)
Output(Line("="))
Output("\n\n")
|
from .modinfo import *
# SO file
from .testing import *
# For output
from bpmodule import output
def PrintHeader(s):
output.Output(output.Line("="))
output.Output("%1%\n", s)
output.Output(output.Line("="))
def PrintResults(nfailed):
output.Output("\n\n")
if nfailed > 0:
output.Output(output.Line("="))
output.Error("RESULTS: %1% failed\n", nfailed)
output.Output(output.Line("="))
else:
output.Output(output.Line("="))
output.Success("RESULTS: %1% failed\n", nfailed)
output.Output(output.Line("="))
output.Output("\n\n")
Add testing function from pythonfrom .modinfo import *
# SO file
from .testing import *
# For output
from bpmodule.output import Output, Error, Warning, Success, Debug, Line
##################
# For testing on the python side
##################
def PyTestFunc(itest, desc, expected, func, *args):
fmt = "%|1$5| : %|2$-5| %|3$-5| %|4$-9| : %5%\n"
try:
func(*args)
except Exception as e:
if expected == False:
Output(fmt, itest, "False", "False", "Success", desc)
return 0
else:
Debug(str(e) + "\n")
Error(fmt, itest, "True", "False", "FAILED", desc)
return 1
except:
if expected == False:
Output(fmt, itest, "False", "False", "Success", desc)
return 0
else:
Error(fmt, itest, "True", "False", "FAILED", desc)
return 1
if expected == True:
Output(fmt, itest, "True", "True", "Success", desc)
return 0
else:
Error(fmt, itest, "False", "True", "FAILED", desc)
return 1
def PrintHeader(s):
Output(Line("="))
Output("%1%\n", s)
Output(Line("="))
def PrintResults(nfailed):
Output("\n\n")
if nfailed > 0:
Output(Line("="))
Error("RESULTS: %1% failed\n", nfailed)
Output(Line("="))
else:
Output(Line("="))
Success("RESULTS: %1% failed\n", nfailed)
Output(Line("="))
Output("\n\n")
|
<commit_before>from .modinfo import *
# SO file
from .testing import *
# For output
from bpmodule import output
def PrintHeader(s):
output.Output(output.Line("="))
output.Output("%1%\n", s)
output.Output(output.Line("="))
def PrintResults(nfailed):
output.Output("\n\n")
if nfailed > 0:
output.Output(output.Line("="))
output.Error("RESULTS: %1% failed\n", nfailed)
output.Output(output.Line("="))
else:
output.Output(output.Line("="))
output.Success("RESULTS: %1% failed\n", nfailed)
output.Output(output.Line("="))
output.Output("\n\n")
<commit_msg>Add testing function from python<commit_after>from .modinfo import *
# SO file
from .testing import *
# For output
from bpmodule.output import Output, Error, Warning, Success, Debug, Line
##################
# For testing on the python side
##################
def PyTestFunc(itest, desc, expected, func, *args):
fmt = "%|1$5| : %|2$-5| %|3$-5| %|4$-9| : %5%\n"
try:
func(*args)
except Exception as e:
if expected == False:
Output(fmt, itest, "False", "False", "Success", desc)
return 0
else:
Debug(str(e) + "\n")
Error(fmt, itest, "True", "False", "FAILED", desc)
return 1
except:
if expected == False:
Output(fmt, itest, "False", "False", "Success", desc)
return 0
else:
Error(fmt, itest, "True", "False", "FAILED", desc)
return 1
if expected == True:
Output(fmt, itest, "True", "True", "Success", desc)
return 0
else:
Error(fmt, itest, "False", "True", "FAILED", desc)
return 1
def PrintHeader(s):
Output(Line("="))
Output("%1%\n", s)
Output(Line("="))
def PrintResults(nfailed):
Output("\n\n")
if nfailed > 0:
Output(Line("="))
Error("RESULTS: %1% failed\n", nfailed)
Output(Line("="))
else:
Output(Line("="))
Success("RESULTS: %1% failed\n", nfailed)
Output(Line("="))
Output("\n\n")
|
a4b475120fd58f135695e071424a3fa1024ae649
|
lib/__init__.py
|
lib/__init__.py
|
"""Package providing an implementation of the TOPKAPI model and some utilities.
The interface isn't stable yet so be prepared to update your code
on a regular basis...
"""
__author__ = 'Theo Vischel'
__version__ = '0.1'
import model
|
"""Package providing an implementation of the TOPKAPI model and some utilities.
The interface isn't stable yet so be prepared to update your code
on a regular basis...
"""
__author__ = 'Theo Vischel'
__version__ = '0.2'
import model
|
Update version number to 0.2.
|
Update version number to 0.2.
|
Python
|
bsd-3-clause
|
sahg/PyTOPKAPI,scottza/PyTOPKAPI
|
"""Package providing an implementation of the TOPKAPI model and some utilities.
The interface isn't stable yet so be prepared to update your code
on a regular basis...
"""
__author__ = 'Theo Vischel'
__version__ = '0.1'
import model
Update version number to 0.2.
|
"""Package providing an implementation of the TOPKAPI model and some utilities.
The interface isn't stable yet so be prepared to update your code
on a regular basis...
"""
__author__ = 'Theo Vischel'
__version__ = '0.2'
import model
|
<commit_before>"""Package providing an implementation of the TOPKAPI model and some utilities.
The interface isn't stable yet so be prepared to update your code
on a regular basis...
"""
__author__ = 'Theo Vischel'
__version__ = '0.1'
import model
<commit_msg>Update version number to 0.2.<commit_after>
|
"""Package providing an implementation of the TOPKAPI model and some utilities.
The interface isn't stable yet so be prepared to update your code
on a regular basis...
"""
__author__ = 'Theo Vischel'
__version__ = '0.2'
import model
|
"""Package providing an implementation of the TOPKAPI model and some utilities.
The interface isn't stable yet so be prepared to update your code
on a regular basis...
"""
__author__ = 'Theo Vischel'
__version__ = '0.1'
import model
Update version number to 0.2."""Package providing an implementation of the TOPKAPI model and some utilities.
The interface isn't stable yet so be prepared to update your code
on a regular basis...
"""
__author__ = 'Theo Vischel'
__version__ = '0.2'
import model
|
<commit_before>"""Package providing an implementation of the TOPKAPI model and some utilities.
The interface isn't stable yet so be prepared to update your code
on a regular basis...
"""
__author__ = 'Theo Vischel'
__version__ = '0.1'
import model
<commit_msg>Update version number to 0.2.<commit_after>"""Package providing an implementation of the TOPKAPI model and some utilities.
The interface isn't stable yet so be prepared to update your code
on a regular basis...
"""
__author__ = 'Theo Vischel'
__version__ = '0.2'
import model
|
bb88b1d2e2c4d3eb482c3cf32d1a53c9e89f94cf
|
conftest.py
|
conftest.py
|
# -*- coding:utf-8 -*-
from __future__ import unicode_literals
from django.db import connection
def pytest_report_header(config):
with connection.cursor() as cursor:
cursor.execute("SELECT VERSION()")
version = cursor.fetchone()[0]
return "MySQL version: {}".format(version)
|
# -*- coding:utf-8 -*-
from __future__ import unicode_literals
import django
from django.db import connection
def pytest_report_header(config):
dot_version = '.'.join(str(x) for x in django.VERSION)
header = "Django version: " + dot_version
if hasattr(connection, '_nodb_connection'):
with connection._nodb_connection.cursor() as cursor:
cursor.execute("SELECT VERSION()")
version = cursor.fetchone()[0]
header += "\nMySQL version: {}".format(version)
return header
|
Fix pytest version report when database does not exist, add Django version header
|
Fix pytest version report when database does not exist, add Django version header
|
Python
|
mit
|
nickmeharry/django-mysql,arnau126/django-mysql,arnau126/django-mysql,nickmeharry/django-mysql,adamchainz/django-mysql
|
# -*- coding:utf-8 -*-
from __future__ import unicode_literals
from django.db import connection
def pytest_report_header(config):
with connection.cursor() as cursor:
cursor.execute("SELECT VERSION()")
version = cursor.fetchone()[0]
return "MySQL version: {}".format(version)
Fix pytest version report when database does not exist, add Django version header
|
# -*- coding:utf-8 -*-
from __future__ import unicode_literals
import django
from django.db import connection
def pytest_report_header(config):
dot_version = '.'.join(str(x) for x in django.VERSION)
header = "Django version: " + dot_version
if hasattr(connection, '_nodb_connection'):
with connection._nodb_connection.cursor() as cursor:
cursor.execute("SELECT VERSION()")
version = cursor.fetchone()[0]
header += "\nMySQL version: {}".format(version)
return header
|
<commit_before># -*- coding:utf-8 -*-
from __future__ import unicode_literals
from django.db import connection
def pytest_report_header(config):
with connection.cursor() as cursor:
cursor.execute("SELECT VERSION()")
version = cursor.fetchone()[0]
return "MySQL version: {}".format(version)
<commit_msg>Fix pytest version report when database does not exist, add Django version header<commit_after>
|
# -*- coding:utf-8 -*-
from __future__ import unicode_literals
import django
from django.db import connection
def pytest_report_header(config):
dot_version = '.'.join(str(x) for x in django.VERSION)
header = "Django version: " + dot_version
if hasattr(connection, '_nodb_connection'):
with connection._nodb_connection.cursor() as cursor:
cursor.execute("SELECT VERSION()")
version = cursor.fetchone()[0]
header += "\nMySQL version: {}".format(version)
return header
|
# -*- coding:utf-8 -*-
from __future__ import unicode_literals
from django.db import connection
def pytest_report_header(config):
with connection.cursor() as cursor:
cursor.execute("SELECT VERSION()")
version = cursor.fetchone()[0]
return "MySQL version: {}".format(version)
Fix pytest version report when database does not exist, add Django version header# -*- coding:utf-8 -*-
from __future__ import unicode_literals
import django
from django.db import connection
def pytest_report_header(config):
dot_version = '.'.join(str(x) for x in django.VERSION)
header = "Django version: " + dot_version
if hasattr(connection, '_nodb_connection'):
with connection._nodb_connection.cursor() as cursor:
cursor.execute("SELECT VERSION()")
version = cursor.fetchone()[0]
header += "\nMySQL version: {}".format(version)
return header
|
<commit_before># -*- coding:utf-8 -*-
from __future__ import unicode_literals
from django.db import connection
def pytest_report_header(config):
with connection.cursor() as cursor:
cursor.execute("SELECT VERSION()")
version = cursor.fetchone()[0]
return "MySQL version: {}".format(version)
<commit_msg>Fix pytest version report when database does not exist, add Django version header<commit_after># -*- coding:utf-8 -*-
from __future__ import unicode_literals
import django
from django.db import connection
def pytest_report_header(config):
dot_version = '.'.join(str(x) for x in django.VERSION)
header = "Django version: " + dot_version
if hasattr(connection, '_nodb_connection'):
with connection._nodb_connection.cursor() as cursor:
cursor.execute("SELECT VERSION()")
version = cursor.fetchone()[0]
header += "\nMySQL version: {}".format(version)
return header
|
22a852a9ad0521496e8b0be52b37d111c3402bb4
|
conftest.py
|
conftest.py
|
import pytest
from models.broadcasted_models import two_comp_model
from utilities.phoenix_utils import load_starfish_spectrum
@pytest.fixture
def host():
"""Host spectrum fixture."""
mod_spec = load_starfish_spectrum([5200, 4.50, 0.0], limits=[2110, 2170], normalize=True)
return mod_spec
@pytest.fixture
def comp():
"""Noramlized Companion spectrum fixture."""
mod_spec = load_starfish_spectrum([2600, 4.50, 0.0], limits=[2110, 2170], normalize=True)
return mod_spec
@pytest.fixture(params=["scalar", "linear", "quadratic", "exponential"])
def norm_method(request):
return request.param
@pytest.fixture()
def tcm_model(host, comp):
return two_comp_model(host.xaxis, host.flux, comp.xaxis, alphas=[0.1, 0.2, 0.3],
rvs=[-0.25, 0.25], gammas=[1, 2, 3, 4])
|
import pytest
from models.broadcasted_models import two_comp_model
from utilities.phoenix_utils import load_starfish_spectrum
@pytest.fixture
def host():
"""Host spectrum fixture."""
mod_spec = load_starfish_spectrum([5200, 4.50, 0.0], limits=[2110, 2165], normalize=True)
return mod_spec
@pytest.fixture
def comp():
"""Noramlized Companion spectrum fixture."""
mod_spec = load_starfish_spectrum([2600, 4.50, 0.0], limits=[2110, 2165], normalize=True)
return mod_spec
@pytest.fixture(params=["scalar", "linear", "quadratic", "exponential"])
def norm_method(request):
return request.param
@pytest.fixture()
def tcm_model(host, comp):
return two_comp_model(host.xaxis, host.flux, comp.xaxis, alphas=[0.1, 0.2, 0.3],
rvs=[-0.25, 0.25], gammas=[1, 2, 3, 4])
|
Revert "tweak host fixture limits"
|
Revert "tweak host fixture limits"
This reverts commit 06e9a964dec8392007e3af87d1a41bbe119158ca.
|
Python
|
mit
|
jason-neal/companion_simulations,jason-neal/companion_simulations
|
import pytest
from models.broadcasted_models import two_comp_model
from utilities.phoenix_utils import load_starfish_spectrum
@pytest.fixture
def host():
"""Host spectrum fixture."""
mod_spec = load_starfish_spectrum([5200, 4.50, 0.0], limits=[2110, 2170], normalize=True)
return mod_spec
@pytest.fixture
def comp():
"""Noramlized Companion spectrum fixture."""
mod_spec = load_starfish_spectrum([2600, 4.50, 0.0], limits=[2110, 2170], normalize=True)
return mod_spec
@pytest.fixture(params=["scalar", "linear", "quadratic", "exponential"])
def norm_method(request):
return request.param
@pytest.fixture()
def tcm_model(host, comp):
return two_comp_model(host.xaxis, host.flux, comp.xaxis, alphas=[0.1, 0.2, 0.3],
rvs=[-0.25, 0.25], gammas=[1, 2, 3, 4])
Revert "tweak host fixture limits"
This reverts commit 06e9a964dec8392007e3af87d1a41bbe119158ca.
|
import pytest
from models.broadcasted_models import two_comp_model
from utilities.phoenix_utils import load_starfish_spectrum
@pytest.fixture
def host():
"""Host spectrum fixture."""
mod_spec = load_starfish_spectrum([5200, 4.50, 0.0], limits=[2110, 2165], normalize=True)
return mod_spec
@pytest.fixture
def comp():
"""Noramlized Companion spectrum fixture."""
mod_spec = load_starfish_spectrum([2600, 4.50, 0.0], limits=[2110, 2165], normalize=True)
return mod_spec
@pytest.fixture(params=["scalar", "linear", "quadratic", "exponential"])
def norm_method(request):
return request.param
@pytest.fixture()
def tcm_model(host, comp):
return two_comp_model(host.xaxis, host.flux, comp.xaxis, alphas=[0.1, 0.2, 0.3],
rvs=[-0.25, 0.25], gammas=[1, 2, 3, 4])
|
<commit_before>import pytest
from models.broadcasted_models import two_comp_model
from utilities.phoenix_utils import load_starfish_spectrum
@pytest.fixture
def host():
"""Host spectrum fixture."""
mod_spec = load_starfish_spectrum([5200, 4.50, 0.0], limits=[2110, 2170], normalize=True)
return mod_spec
@pytest.fixture
def comp():
"""Noramlized Companion spectrum fixture."""
mod_spec = load_starfish_spectrum([2600, 4.50, 0.0], limits=[2110, 2170], normalize=True)
return mod_spec
@pytest.fixture(params=["scalar", "linear", "quadratic", "exponential"])
def norm_method(request):
return request.param
@pytest.fixture()
def tcm_model(host, comp):
return two_comp_model(host.xaxis, host.flux, comp.xaxis, alphas=[0.1, 0.2, 0.3],
rvs=[-0.25, 0.25], gammas=[1, 2, 3, 4])
<commit_msg>Revert "tweak host fixture limits"
This reverts commit 06e9a964dec8392007e3af87d1a41bbe119158ca.<commit_after>
|
import pytest
from models.broadcasted_models import two_comp_model
from utilities.phoenix_utils import load_starfish_spectrum
@pytest.fixture
def host():
"""Host spectrum fixture."""
mod_spec = load_starfish_spectrum([5200, 4.50, 0.0], limits=[2110, 2165], normalize=True)
return mod_spec
@pytest.fixture
def comp():
"""Noramlized Companion spectrum fixture."""
mod_spec = load_starfish_spectrum([2600, 4.50, 0.0], limits=[2110, 2165], normalize=True)
return mod_spec
@pytest.fixture(params=["scalar", "linear", "quadratic", "exponential"])
def norm_method(request):
return request.param
@pytest.fixture()
def tcm_model(host, comp):
return two_comp_model(host.xaxis, host.flux, comp.xaxis, alphas=[0.1, 0.2, 0.3],
rvs=[-0.25, 0.25], gammas=[1, 2, 3, 4])
|
import pytest
from models.broadcasted_models import two_comp_model
from utilities.phoenix_utils import load_starfish_spectrum
@pytest.fixture
def host():
"""Host spectrum fixture."""
mod_spec = load_starfish_spectrum([5200, 4.50, 0.0], limits=[2110, 2170], normalize=True)
return mod_spec
@pytest.fixture
def comp():
"""Noramlized Companion spectrum fixture."""
mod_spec = load_starfish_spectrum([2600, 4.50, 0.0], limits=[2110, 2170], normalize=True)
return mod_spec
@pytest.fixture(params=["scalar", "linear", "quadratic", "exponential"])
def norm_method(request):
return request.param
@pytest.fixture()
def tcm_model(host, comp):
return two_comp_model(host.xaxis, host.flux, comp.xaxis, alphas=[0.1, 0.2, 0.3],
rvs=[-0.25, 0.25], gammas=[1, 2, 3, 4])
Revert "tweak host fixture limits"
This reverts commit 06e9a964dec8392007e3af87d1a41bbe119158ca.import pytest
from models.broadcasted_models import two_comp_model
from utilities.phoenix_utils import load_starfish_spectrum
@pytest.fixture
def host():
"""Host spectrum fixture."""
mod_spec = load_starfish_spectrum([5200, 4.50, 0.0], limits=[2110, 2165], normalize=True)
return mod_spec
@pytest.fixture
def comp():
"""Noramlized Companion spectrum fixture."""
mod_spec = load_starfish_spectrum([2600, 4.50, 0.0], limits=[2110, 2165], normalize=True)
return mod_spec
@pytest.fixture(params=["scalar", "linear", "quadratic", "exponential"])
def norm_method(request):
return request.param
@pytest.fixture()
def tcm_model(host, comp):
return two_comp_model(host.xaxis, host.flux, comp.xaxis, alphas=[0.1, 0.2, 0.3],
rvs=[-0.25, 0.25], gammas=[1, 2, 3, 4])
|
<commit_before>import pytest
from models.broadcasted_models import two_comp_model
from utilities.phoenix_utils import load_starfish_spectrum
@pytest.fixture
def host():
"""Host spectrum fixture."""
mod_spec = load_starfish_spectrum([5200, 4.50, 0.0], limits=[2110, 2170], normalize=True)
return mod_spec
@pytest.fixture
def comp():
"""Noramlized Companion spectrum fixture."""
mod_spec = load_starfish_spectrum([2600, 4.50, 0.0], limits=[2110, 2170], normalize=True)
return mod_spec
@pytest.fixture(params=["scalar", "linear", "quadratic", "exponential"])
def norm_method(request):
return request.param
@pytest.fixture()
def tcm_model(host, comp):
return two_comp_model(host.xaxis, host.flux, comp.xaxis, alphas=[0.1, 0.2, 0.3],
rvs=[-0.25, 0.25], gammas=[1, 2, 3, 4])
<commit_msg>Revert "tweak host fixture limits"
This reverts commit 06e9a964dec8392007e3af87d1a41bbe119158ca.<commit_after>import pytest
from models.broadcasted_models import two_comp_model
from utilities.phoenix_utils import load_starfish_spectrum
@pytest.fixture
def host():
"""Host spectrum fixture."""
mod_spec = load_starfish_spectrum([5200, 4.50, 0.0], limits=[2110, 2165], normalize=True)
return mod_spec
@pytest.fixture
def comp():
"""Noramlized Companion spectrum fixture."""
mod_spec = load_starfish_spectrum([2600, 4.50, 0.0], limits=[2110, 2165], normalize=True)
return mod_spec
@pytest.fixture(params=["scalar", "linear", "quadratic", "exponential"])
def norm_method(request):
return request.param
@pytest.fixture()
def tcm_model(host, comp):
return two_comp_model(host.xaxis, host.flux, comp.xaxis, alphas=[0.1, 0.2, 0.3],
rvs=[-0.25, 0.25], gammas=[1, 2, 3, 4])
|
aaf7cb7ecc1a74fb2b222fd21aea7116dac2ca98
|
contribs.py
|
contribs.py
|
# Get Contribution Count
import urllib
import datetime
import HTMLParser
class ContribParser(HTMLParser.HTMLParser):
today = datetime.date.today().isoformat()
def handle_starttag(self, tag, attrs):
if tag == 'rect' and self.is_today(attrs):
self.count = self.get_count(attrs)
def is_today(self, attrs):
for name, value in attrs:
if name == 'data-date' and value == self.today:
return True
return False
def get_count(self, attrs):
for name, value in attrs:
if name == 'data-count':
return value
return None
def getContribs(username):
url = 'https://github.com/users/:user/contributions'
req = urllib.urlopen(url.replace(':user', username))
parser = ContribParser()
parser.feed(req.read())
return parser.count
|
# Get Contribution Count
import urllib
import datetime
import HTMLParser
class ContribParser(HTMLParser.HTMLParser):
def __init__(self):
self.today = datetime.date.today().isoformat()
HTMLParser.HTMLParser.__init__(self)
def handle_starttag(self, tag, attrs):
if tag == 'rect' and self.is_today(attrs):
self.count = self.get_count(attrs)
def is_today(self, attrs):
for name, value in attrs:
if name == 'data-date' and value == self.today:
return True
return False
def get_count(self, attrs):
for name, value in attrs:
if name == 'data-count':
return value
return None
def getContribs(username):
url = 'https://github.com/users/:user/contributions'
req = urllib.urlopen(url.replace(':user', username))
parser = ContribParser()
parser.feed(req.read())
return parser.count
|
Fix date issue (I think)
|
Fix date issue (I think)
|
Python
|
mit
|
chrisfosterelli/commitwatch
|
# Get Contribution Count
import urllib
import datetime
import HTMLParser
class ContribParser(HTMLParser.HTMLParser):
today = datetime.date.today().isoformat()
def handle_starttag(self, tag, attrs):
if tag == 'rect' and self.is_today(attrs):
self.count = self.get_count(attrs)
def is_today(self, attrs):
for name, value in attrs:
if name == 'data-date' and value == self.today:
return True
return False
def get_count(self, attrs):
for name, value in attrs:
if name == 'data-count':
return value
return None
def getContribs(username):
url = 'https://github.com/users/:user/contributions'
req = urllib.urlopen(url.replace(':user', username))
parser = ContribParser()
parser.feed(req.read())
return parser.count
Fix date issue (I think)
|
# Get Contribution Count
import urllib
import datetime
import HTMLParser
class ContribParser(HTMLParser.HTMLParser):
def __init__(self):
self.today = datetime.date.today().isoformat()
HTMLParser.HTMLParser.__init__(self)
def handle_starttag(self, tag, attrs):
if tag == 'rect' and self.is_today(attrs):
self.count = self.get_count(attrs)
def is_today(self, attrs):
for name, value in attrs:
if name == 'data-date' and value == self.today:
return True
return False
def get_count(self, attrs):
for name, value in attrs:
if name == 'data-count':
return value
return None
def getContribs(username):
url = 'https://github.com/users/:user/contributions'
req = urllib.urlopen(url.replace(':user', username))
parser = ContribParser()
parser.feed(req.read())
return parser.count
|
<commit_before>
# Get Contribution Count
import urllib
import datetime
import HTMLParser
class ContribParser(HTMLParser.HTMLParser):
today = datetime.date.today().isoformat()
def handle_starttag(self, tag, attrs):
if tag == 'rect' and self.is_today(attrs):
self.count = self.get_count(attrs)
def is_today(self, attrs):
for name, value in attrs:
if name == 'data-date' and value == self.today:
return True
return False
def get_count(self, attrs):
for name, value in attrs:
if name == 'data-count':
return value
return None
def getContribs(username):
url = 'https://github.com/users/:user/contributions'
req = urllib.urlopen(url.replace(':user', username))
parser = ContribParser()
parser.feed(req.read())
return parser.count
<commit_msg>Fix date issue (I think)<commit_after>
|
# Get Contribution Count
import urllib
import datetime
import HTMLParser
class ContribParser(HTMLParser.HTMLParser):
def __init__(self):
self.today = datetime.date.today().isoformat()
HTMLParser.HTMLParser.__init__(self)
def handle_starttag(self, tag, attrs):
if tag == 'rect' and self.is_today(attrs):
self.count = self.get_count(attrs)
def is_today(self, attrs):
for name, value in attrs:
if name == 'data-date' and value == self.today:
return True
return False
def get_count(self, attrs):
for name, value in attrs:
if name == 'data-count':
return value
return None
def getContribs(username):
url = 'https://github.com/users/:user/contributions'
req = urllib.urlopen(url.replace(':user', username))
parser = ContribParser()
parser.feed(req.read())
return parser.count
|
# Get Contribution Count
import urllib
import datetime
import HTMLParser
class ContribParser(HTMLParser.HTMLParser):
today = datetime.date.today().isoformat()
def handle_starttag(self, tag, attrs):
if tag == 'rect' and self.is_today(attrs):
self.count = self.get_count(attrs)
def is_today(self, attrs):
for name, value in attrs:
if name == 'data-date' and value == self.today:
return True
return False
def get_count(self, attrs):
for name, value in attrs:
if name == 'data-count':
return value
return None
def getContribs(username):
url = 'https://github.com/users/:user/contributions'
req = urllib.urlopen(url.replace(':user', username))
parser = ContribParser()
parser.feed(req.read())
return parser.count
Fix date issue (I think)
# Get Contribution Count
import urllib
import datetime
import HTMLParser
class ContribParser(HTMLParser.HTMLParser):
def __init__(self):
self.today = datetime.date.today().isoformat()
HTMLParser.HTMLParser.__init__(self)
def handle_starttag(self, tag, attrs):
if tag == 'rect' and self.is_today(attrs):
self.count = self.get_count(attrs)
def is_today(self, attrs):
for name, value in attrs:
if name == 'data-date' and value == self.today:
return True
return False
def get_count(self, attrs):
for name, value in attrs:
if name == 'data-count':
return value
return None
def getContribs(username):
url = 'https://github.com/users/:user/contributions'
req = urllib.urlopen(url.replace(':user', username))
parser = ContribParser()
parser.feed(req.read())
return parser.count
|
<commit_before>
# Get Contribution Count
import urllib
import datetime
import HTMLParser
class ContribParser(HTMLParser.HTMLParser):
today = datetime.date.today().isoformat()
def handle_starttag(self, tag, attrs):
if tag == 'rect' and self.is_today(attrs):
self.count = self.get_count(attrs)
def is_today(self, attrs):
for name, value in attrs:
if name == 'data-date' and value == self.today:
return True
return False
def get_count(self, attrs):
for name, value in attrs:
if name == 'data-count':
return value
return None
def getContribs(username):
url = 'https://github.com/users/:user/contributions'
req = urllib.urlopen(url.replace(':user', username))
parser = ContribParser()
parser.feed(req.read())
return parser.count
<commit_msg>Fix date issue (I think)<commit_after>
# Get Contribution Count
import urllib
import datetime
import HTMLParser
class ContribParser(HTMLParser.HTMLParser):
def __init__(self):
self.today = datetime.date.today().isoformat()
HTMLParser.HTMLParser.__init__(self)
def handle_starttag(self, tag, attrs):
if tag == 'rect' and self.is_today(attrs):
self.count = self.get_count(attrs)
def is_today(self, attrs):
for name, value in attrs:
if name == 'data-date' and value == self.today:
return True
return False
def get_count(self, attrs):
for name, value in attrs:
if name == 'data-count':
return value
return None
def getContribs(username):
url = 'https://github.com/users/:user/contributions'
req = urllib.urlopen(url.replace(':user', username))
parser = ContribParser()
parser.feed(req.read())
return parser.count
|
22be6bb3593f948893ab3f797d34e20e66fff841
|
example.py
|
example.py
|
import discord
import asyncio
client = discord.Client()
@client.event
async def on_ready():
print('Logged in as')
print(client.user.name)
print(client.user.id)
print('------')
@client.event
async def on_message(message):
if message.content.startswith('!test'):
counter = 0
tmp = await client.send_message(message.channel, 'Calculating messages...')
async for log in client.logs_from(message.channel, limit=100):
if log.author == message.author:
counter += 1
await client.edit_message(tmp, 'You have {} messages.'.format(counter))
elif message.content.startswith('!sleep'):
await asyncio.sleep(5)
await client.send_message(message.channel, 'Done sleeping')
client.run('token')
|
import discord
import asyncio
import os
#Set up Client State
CLIENT_TOKEN=os.environ['TOKEN']
client = discord.Client()
@client.event
async def on_ready():
print('Logged in as')
print(client.user.name)
print(client.user.id)
print('------')
@client.event
async def on_message(message):
if message.content.startswith('!test'):
counter = 0
tmp = await client.send_message(message.channel, 'Calculating messages...')
async for log in client.logs_from(message.channel, limit=100):
if log.author == message.author:
counter += 1
await client.edit_message(tmp, 'You have {} messages.'.format(counter))
elif message.content.startswith('!sleep'):
await asyncio.sleep(5)
await client.send_message(message.channel, 'Done sleeping')
client.run(CLIENT_TOKEN)
|
Use env value for client token
|
Use env value for client token
|
Python
|
mit
|
gryffon/SusumuTakuan,gryffon/SusumuTakuan
|
import discord
import asyncio
client = discord.Client()
@client.event
async def on_ready():
print('Logged in as')
print(client.user.name)
print(client.user.id)
print('------')
@client.event
async def on_message(message):
if message.content.startswith('!test'):
counter = 0
tmp = await client.send_message(message.channel, 'Calculating messages...')
async for log in client.logs_from(message.channel, limit=100):
if log.author == message.author:
counter += 1
await client.edit_message(tmp, 'You have {} messages.'.format(counter))
elif message.content.startswith('!sleep'):
await asyncio.sleep(5)
await client.send_message(message.channel, 'Done sleeping')
client.run('token')Use env value for client token
|
import discord
import asyncio
import os
#Set up Client State
CLIENT_TOKEN=os.environ['TOKEN']
client = discord.Client()
@client.event
async def on_ready():
print('Logged in as')
print(client.user.name)
print(client.user.id)
print('------')
@client.event
async def on_message(message):
if message.content.startswith('!test'):
counter = 0
tmp = await client.send_message(message.channel, 'Calculating messages...')
async for log in client.logs_from(message.channel, limit=100):
if log.author == message.author:
counter += 1
await client.edit_message(tmp, 'You have {} messages.'.format(counter))
elif message.content.startswith('!sleep'):
await asyncio.sleep(5)
await client.send_message(message.channel, 'Done sleeping')
client.run(CLIENT_TOKEN)
|
<commit_before>import discord
import asyncio
client = discord.Client()
@client.event
async def on_ready():
print('Logged in as')
print(client.user.name)
print(client.user.id)
print('------')
@client.event
async def on_message(message):
if message.content.startswith('!test'):
counter = 0
tmp = await client.send_message(message.channel, 'Calculating messages...')
async for log in client.logs_from(message.channel, limit=100):
if log.author == message.author:
counter += 1
await client.edit_message(tmp, 'You have {} messages.'.format(counter))
elif message.content.startswith('!sleep'):
await asyncio.sleep(5)
await client.send_message(message.channel, 'Done sleeping')
client.run('token')<commit_msg>Use env value for client token<commit_after>
|
import discord
import asyncio
import os
#Set up Client State
CLIENT_TOKEN=os.environ['TOKEN']
client = discord.Client()
@client.event
async def on_ready():
print('Logged in as')
print(client.user.name)
print(client.user.id)
print('------')
@client.event
async def on_message(message):
if message.content.startswith('!test'):
counter = 0
tmp = await client.send_message(message.channel, 'Calculating messages...')
async for log in client.logs_from(message.channel, limit=100):
if log.author == message.author:
counter += 1
await client.edit_message(tmp, 'You have {} messages.'.format(counter))
elif message.content.startswith('!sleep'):
await asyncio.sleep(5)
await client.send_message(message.channel, 'Done sleeping')
client.run(CLIENT_TOKEN)
|
import discord
import asyncio
client = discord.Client()
@client.event
async def on_ready():
print('Logged in as')
print(client.user.name)
print(client.user.id)
print('------')
@client.event
async def on_message(message):
if message.content.startswith('!test'):
counter = 0
tmp = await client.send_message(message.channel, 'Calculating messages...')
async for log in client.logs_from(message.channel, limit=100):
if log.author == message.author:
counter += 1
await client.edit_message(tmp, 'You have {} messages.'.format(counter))
elif message.content.startswith('!sleep'):
await asyncio.sleep(5)
await client.send_message(message.channel, 'Done sleeping')
client.run('token')Use env value for client tokenimport discord
import asyncio
import os
#Set up Client State
CLIENT_TOKEN=os.environ['TOKEN']
client = discord.Client()
@client.event
async def on_ready():
print('Logged in as')
print(client.user.name)
print(client.user.id)
print('------')
@client.event
async def on_message(message):
if message.content.startswith('!test'):
counter = 0
tmp = await client.send_message(message.channel, 'Calculating messages...')
async for log in client.logs_from(message.channel, limit=100):
if log.author == message.author:
counter += 1
await client.edit_message(tmp, 'You have {} messages.'.format(counter))
elif message.content.startswith('!sleep'):
await asyncio.sleep(5)
await client.send_message(message.channel, 'Done sleeping')
client.run(CLIENT_TOKEN)
|
<commit_before>import discord
import asyncio
client = discord.Client()
@client.event
async def on_ready():
print('Logged in as')
print(client.user.name)
print(client.user.id)
print('------')
@client.event
async def on_message(message):
if message.content.startswith('!test'):
counter = 0
tmp = await client.send_message(message.channel, 'Calculating messages...')
async for log in client.logs_from(message.channel, limit=100):
if log.author == message.author:
counter += 1
await client.edit_message(tmp, 'You have {} messages.'.format(counter))
elif message.content.startswith('!sleep'):
await asyncio.sleep(5)
await client.send_message(message.channel, 'Done sleeping')
client.run('token')<commit_msg>Use env value for client token<commit_after>import discord
import asyncio
import os
#Set up Client State
CLIENT_TOKEN=os.environ['TOKEN']
client = discord.Client()
@client.event
async def on_ready():
print('Logged in as')
print(client.user.name)
print(client.user.id)
print('------')
@client.event
async def on_message(message):
if message.content.startswith('!test'):
counter = 0
tmp = await client.send_message(message.channel, 'Calculating messages...')
async for log in client.logs_from(message.channel, limit=100):
if log.author == message.author:
counter += 1
await client.edit_message(tmp, 'You have {} messages.'.format(counter))
elif message.content.startswith('!sleep'):
await asyncio.sleep(5)
await client.send_message(message.channel, 'Done sleeping')
client.run(CLIENT_TOKEN)
|
e3d8b836681a0cb4795d317c7a23defd6004c967
|
pytest_run.py
|
pytest_run.py
|
# coding=utf-8
"""This is a script for running pytest from the command line.
This script exists so that the project directory gets added to sys.path, which
prevents us from accidentally testing the globally installed willie version.
pytest_run.py
Copyright 2013, Ari Koivula, <ari@koivu.la>
Licensed under the Eiffel Forum License 2.
http://willie.dfbta.net
"""
if __name__ == "__main__":
import sys
import pytest
returncode = pytest.main()
sys.exit(returncode)
|
#!/usr/bin/env python
# coding=utf-8
"""This is a script for running pytest from the command line.
This script exists so that the project directory gets added to sys.path, which
prevents us from accidentally testing the globally installed willie version.
pytest_run.py
Copyright 2013, Ari Koivula, <ari@koivu.la>
Licensed under the Eiffel Forum License 2.
http://willie.dfbta.net
"""
if __name__ == "__main__":
import sys
import pytest
returncode = pytest.main()
sys.exit(returncode)
|
Add shebang to testing script
|
Add shebang to testing script
|
Python
|
mit
|
Uname-a/knife_scraper,Uname-a/knife_scraper,Uname-a/knife_scraper
|
# coding=utf-8
"""This is a script for running pytest from the command line.
This script exists so that the project directory gets added to sys.path, which
prevents us from accidentally testing the globally installed willie version.
pytest_run.py
Copyright 2013, Ari Koivula, <ari@koivu.la>
Licensed under the Eiffel Forum License 2.
http://willie.dfbta.net
"""
if __name__ == "__main__":
import sys
import pytest
returncode = pytest.main()
sys.exit(returncode)
Add shebang to testing script
|
#!/usr/bin/env python
# coding=utf-8
"""This is a script for running pytest from the command line.
This script exists so that the project directory gets added to sys.path, which
prevents us from accidentally testing the globally installed willie version.
pytest_run.py
Copyright 2013, Ari Koivula, <ari@koivu.la>
Licensed under the Eiffel Forum License 2.
http://willie.dfbta.net
"""
if __name__ == "__main__":
import sys
import pytest
returncode = pytest.main()
sys.exit(returncode)
|
<commit_before># coding=utf-8
"""This is a script for running pytest from the command line.
This script exists so that the project directory gets added to sys.path, which
prevents us from accidentally testing the globally installed willie version.
pytest_run.py
Copyright 2013, Ari Koivula, <ari@koivu.la>
Licensed under the Eiffel Forum License 2.
http://willie.dfbta.net
"""
if __name__ == "__main__":
import sys
import pytest
returncode = pytest.main()
sys.exit(returncode)
<commit_msg>Add shebang to testing script<commit_after>
|
#!/usr/bin/env python
# coding=utf-8
"""This is a script for running pytest from the command line.
This script exists so that the project directory gets added to sys.path, which
prevents us from accidentally testing the globally installed willie version.
pytest_run.py
Copyright 2013, Ari Koivula, <ari@koivu.la>
Licensed under the Eiffel Forum License 2.
http://willie.dfbta.net
"""
if __name__ == "__main__":
import sys
import pytest
returncode = pytest.main()
sys.exit(returncode)
|
# coding=utf-8
"""This is a script for running pytest from the command line.
This script exists so that the project directory gets added to sys.path, which
prevents us from accidentally testing the globally installed willie version.
pytest_run.py
Copyright 2013, Ari Koivula, <ari@koivu.la>
Licensed under the Eiffel Forum License 2.
http://willie.dfbta.net
"""
if __name__ == "__main__":
import sys
import pytest
returncode = pytest.main()
sys.exit(returncode)
Add shebang to testing script#!/usr/bin/env python
# coding=utf-8
"""This is a script for running pytest from the command line.
This script exists so that the project directory gets added to sys.path, which
prevents us from accidentally testing the globally installed willie version.
pytest_run.py
Copyright 2013, Ari Koivula, <ari@koivu.la>
Licensed under the Eiffel Forum License 2.
http://willie.dfbta.net
"""
if __name__ == "__main__":
import sys
import pytest
returncode = pytest.main()
sys.exit(returncode)
|
<commit_before># coding=utf-8
"""This is a script for running pytest from the command line.
This script exists so that the project directory gets added to sys.path, which
prevents us from accidentally testing the globally installed willie version.
pytest_run.py
Copyright 2013, Ari Koivula, <ari@koivu.la>
Licensed under the Eiffel Forum License 2.
http://willie.dfbta.net
"""
if __name__ == "__main__":
import sys
import pytest
returncode = pytest.main()
sys.exit(returncode)
<commit_msg>Add shebang to testing script<commit_after>#!/usr/bin/env python
# coding=utf-8
"""This is a script for running pytest from the command line.
This script exists so that the project directory gets added to sys.path, which
prevents us from accidentally testing the globally installed willie version.
pytest_run.py
Copyright 2013, Ari Koivula, <ari@koivu.la>
Licensed under the Eiffel Forum License 2.
http://willie.dfbta.net
"""
if __name__ == "__main__":
import sys
import pytest
returncode = pytest.main()
sys.exit(returncode)
|
d5a5e46b2fbc9284213aef3ec45f0605b002b7b1
|
axes/management/commands/axes_reset.py
|
axes/management/commands/axes_reset.py
|
from django.core.management.base import BaseCommand
from axes.utils import reset
class Command(BaseCommand):
help = ("resets any lockouts or failed login records. If called with an "
"IP, resets only for that IP")
def add_arguments(self, parser):
parser.add_argument('ip', nargs='*')
def handle(self, *args, **kwargs):
count = 0
if kwargs:
for ip in kwargs['ip']:
count += reset(ip=ip)
else:
count = reset()
if count:
print('{0} attempts removed.'.format(count))
else:
print('No attempts found.')
|
from django.core.management.base import BaseCommand
from axes.utils import reset
class Command(BaseCommand):
help = ("resets any lockouts or failed login records. If called with an "
"IP, resets only for that IP")
def add_arguments(self, parser):
parser.add_argument('ip', nargs='*')
def handle(self, *args, **kwargs):
count = 0
if kwargs and kwargs.get('ip'):
for ip in kwargs['ip']:
count += reset(ip=ip)
else:
count = reset()
if count:
print('{0} attempts removed.'.format(count))
else:
print('No attempts found.')
|
Reset all attempts when ip not specified
|
Reset all attempts when ip not specified
When no ip address positional arguments are specified, reset all attempts, as with reset() and per documentation.
|
Python
|
mit
|
svenhertle/django-axes,django-pci/django-axes,jazzband/django-axes
|
from django.core.management.base import BaseCommand
from axes.utils import reset
class Command(BaseCommand):
help = ("resets any lockouts or failed login records. If called with an "
"IP, resets only for that IP")
def add_arguments(self, parser):
parser.add_argument('ip', nargs='*')
def handle(self, *args, **kwargs):
count = 0
if kwargs:
for ip in kwargs['ip']:
count += reset(ip=ip)
else:
count = reset()
if count:
print('{0} attempts removed.'.format(count))
else:
print('No attempts found.')
Reset all attempts when ip not specified
When no ip address positional arguments are specified, reset all attempts, as with reset() and per documentation.
|
from django.core.management.base import BaseCommand
from axes.utils import reset
class Command(BaseCommand):
help = ("resets any lockouts or failed login records. If called with an "
"IP, resets only for that IP")
def add_arguments(self, parser):
parser.add_argument('ip', nargs='*')
def handle(self, *args, **kwargs):
count = 0
if kwargs and kwargs.get('ip'):
for ip in kwargs['ip']:
count += reset(ip=ip)
else:
count = reset()
if count:
print('{0} attempts removed.'.format(count))
else:
print('No attempts found.')
|
<commit_before>from django.core.management.base import BaseCommand
from axes.utils import reset
class Command(BaseCommand):
help = ("resets any lockouts or failed login records. If called with an "
"IP, resets only for that IP")
def add_arguments(self, parser):
parser.add_argument('ip', nargs='*')
def handle(self, *args, **kwargs):
count = 0
if kwargs:
for ip in kwargs['ip']:
count += reset(ip=ip)
else:
count = reset()
if count:
print('{0} attempts removed.'.format(count))
else:
print('No attempts found.')
<commit_msg>Reset all attempts when ip not specified
When no ip address positional arguments are specified, reset all attempts, as with reset() and per documentation.<commit_after>
|
from django.core.management.base import BaseCommand
from axes.utils import reset
class Command(BaseCommand):
help = ("resets any lockouts or failed login records. If called with an "
"IP, resets only for that IP")
def add_arguments(self, parser):
parser.add_argument('ip', nargs='*')
def handle(self, *args, **kwargs):
count = 0
if kwargs and kwargs.get('ip'):
for ip in kwargs['ip']:
count += reset(ip=ip)
else:
count = reset()
if count:
print('{0} attempts removed.'.format(count))
else:
print('No attempts found.')
|
from django.core.management.base import BaseCommand
from axes.utils import reset
class Command(BaseCommand):
help = ("resets any lockouts or failed login records. If called with an "
"IP, resets only for that IP")
def add_arguments(self, parser):
parser.add_argument('ip', nargs='*')
def handle(self, *args, **kwargs):
count = 0
if kwargs:
for ip in kwargs['ip']:
count += reset(ip=ip)
else:
count = reset()
if count:
print('{0} attempts removed.'.format(count))
else:
print('No attempts found.')
Reset all attempts when ip not specified
When no ip address positional arguments are specified, reset all attempts, as with reset() and per documentation.from django.core.management.base import BaseCommand
from axes.utils import reset
class Command(BaseCommand):
help = ("resets any lockouts or failed login records. If called with an "
"IP, resets only for that IP")
def add_arguments(self, parser):
parser.add_argument('ip', nargs='*')
def handle(self, *args, **kwargs):
count = 0
if kwargs and kwargs.get('ip'):
for ip in kwargs['ip']:
count += reset(ip=ip)
else:
count = reset()
if count:
print('{0} attempts removed.'.format(count))
else:
print('No attempts found.')
|
<commit_before>from django.core.management.base import BaseCommand
from axes.utils import reset
class Command(BaseCommand):
help = ("resets any lockouts or failed login records. If called with an "
"IP, resets only for that IP")
def add_arguments(self, parser):
parser.add_argument('ip', nargs='*')
def handle(self, *args, **kwargs):
count = 0
if kwargs:
for ip in kwargs['ip']:
count += reset(ip=ip)
else:
count = reset()
if count:
print('{0} attempts removed.'.format(count))
else:
print('No attempts found.')
<commit_msg>Reset all attempts when ip not specified
When no ip address positional arguments are specified, reset all attempts, as with reset() and per documentation.<commit_after>from django.core.management.base import BaseCommand
from axes.utils import reset
class Command(BaseCommand):
help = ("resets any lockouts or failed login records. If called with an "
"IP, resets only for that IP")
def add_arguments(self, parser):
parser.add_argument('ip', nargs='*')
def handle(self, *args, **kwargs):
count = 0
if kwargs and kwargs.get('ip'):
for ip in kwargs['ip']:
count += reset(ip=ip)
else:
count = reset()
if count:
print('{0} attempts removed.'.format(count))
else:
print('No attempts found.')
|
132932747a1f7da67413b9c0cf7916707c1e3d19
|
src/python/services/CVMFSAppVersions.py
|
src/python/services/CVMFSAppVersions.py
|
"""CVMFS Servcice."""
import os
import re
import cherrypy
import html
from natsort import natsorted
version_re = re.compile(r"^release-(\d{1,3}\.\d{1,3}\.\d{1,3})$")
@cherrypy.popargs('appid')
class CVMFSAppVersions(object):
"""
CVMFS App Version checking service.
CVMFS Service to get the list of versions available
on CVMFS for a given app.
"""
def __init__(self, cvmfs_root, valid_apps):
"""Initialise."""
self.cvmfs_root = cvmfs_root
self.valid_apps = valid_apps
@cherrypy.expose
def index(self, appid=None):
"""Return the index page."""
print "IN CVMFSAppVersion: appid=(%s)" % appid
if appid not in self.valid_apps:
print "Invalid app type %s" % appid
return ''
html_ = html.HTML()
_, dirs, _ = os.walk(os.path.join(self.cvmfs_root, appid)).next()
for dir_ in natsorted(dirs):
for version in version_re.findall(dir_):
html_.option(version)
return str(html_)
|
"""CVMFS Servcice."""
import os
import re
import cherrypy
import html
from natsort import natsorted
VERSION_RE = re.compile(r"^release-(\d{1,3}\.\d{1,3}\.\d{1,3})$")
@cherrypy.popargs('appid')
class CVMFSAppVersions(object):
"""
CVMFS App Version checking service.
CVMFS Service to get the list of versions available
on CVMFS for a given app.
"""
def __init__(self, cvmfs_root, valid_apps):
"""Initialise."""
self.cvmfs_root = cvmfs_root
self.valid_apps = valid_apps
@cherrypy.expose
def index(self, appid=None):
"""Return the index page."""
print "IN CVMFSAppVersion: appid=(%s)" % appid
if appid not in self.valid_apps:
print "Invalid app type %s" % appid
return ''
html_ = html.HTML()
_, dirs, _ = os.walk(os.path.join(self.cvmfs_root, appid)).next()
for dir_ in natsorted(dirs):
for version in VERSION_RE.findall(dir_):
html_.option(version)
return str(html_)
|
Change the re const name to uppercase.
|
Change the re const name to uppercase.
|
Python
|
mit
|
alexanderrichards/LZProduction,alexanderrichards/LZProduction,alexanderrichards/LZProduction,alexanderrichards/LZProduction
|
"""CVMFS Servcice."""
import os
import re
import cherrypy
import html
from natsort import natsorted
version_re = re.compile(r"^release-(\d{1,3}\.\d{1,3}\.\d{1,3})$")
@cherrypy.popargs('appid')
class CVMFSAppVersions(object):
"""
CVMFS App Version checking service.
CVMFS Service to get the list of versions available
on CVMFS for a given app.
"""
def __init__(self, cvmfs_root, valid_apps):
"""Initialise."""
self.cvmfs_root = cvmfs_root
self.valid_apps = valid_apps
@cherrypy.expose
def index(self, appid=None):
"""Return the index page."""
print "IN CVMFSAppVersion: appid=(%s)" % appid
if appid not in self.valid_apps:
print "Invalid app type %s" % appid
return ''
html_ = html.HTML()
_, dirs, _ = os.walk(os.path.join(self.cvmfs_root, appid)).next()
for dir_ in natsorted(dirs):
for version in version_re.findall(dir_):
html_.option(version)
return str(html_)
Change the re const name to uppercase.
|
"""CVMFS Servcice."""
import os
import re
import cherrypy
import html
from natsort import natsorted
VERSION_RE = re.compile(r"^release-(\d{1,3}\.\d{1,3}\.\d{1,3})$")
@cherrypy.popargs('appid')
class CVMFSAppVersions(object):
"""
CVMFS App Version checking service.
CVMFS Service to get the list of versions available
on CVMFS for a given app.
"""
def __init__(self, cvmfs_root, valid_apps):
"""Initialise."""
self.cvmfs_root = cvmfs_root
self.valid_apps = valid_apps
@cherrypy.expose
def index(self, appid=None):
"""Return the index page."""
print "IN CVMFSAppVersion: appid=(%s)" % appid
if appid not in self.valid_apps:
print "Invalid app type %s" % appid
return ''
html_ = html.HTML()
_, dirs, _ = os.walk(os.path.join(self.cvmfs_root, appid)).next()
for dir_ in natsorted(dirs):
for version in VERSION_RE.findall(dir_):
html_.option(version)
return str(html_)
|
<commit_before>"""CVMFS Servcice."""
import os
import re
import cherrypy
import html
from natsort import natsorted
version_re = re.compile(r"^release-(\d{1,3}\.\d{1,3}\.\d{1,3})$")
@cherrypy.popargs('appid')
class CVMFSAppVersions(object):
"""
CVMFS App Version checking service.
CVMFS Service to get the list of versions available
on CVMFS for a given app.
"""
def __init__(self, cvmfs_root, valid_apps):
"""Initialise."""
self.cvmfs_root = cvmfs_root
self.valid_apps = valid_apps
@cherrypy.expose
def index(self, appid=None):
"""Return the index page."""
print "IN CVMFSAppVersion: appid=(%s)" % appid
if appid not in self.valid_apps:
print "Invalid app type %s" % appid
return ''
html_ = html.HTML()
_, dirs, _ = os.walk(os.path.join(self.cvmfs_root, appid)).next()
for dir_ in natsorted(dirs):
for version in version_re.findall(dir_):
html_.option(version)
return str(html_)
<commit_msg>Change the re const name to uppercase.<commit_after>
|
"""CVMFS Servcice."""
import os
import re
import cherrypy
import html
from natsort import natsorted
VERSION_RE = re.compile(r"^release-(\d{1,3}\.\d{1,3}\.\d{1,3})$")
@cherrypy.popargs('appid')
class CVMFSAppVersions(object):
"""
CVMFS App Version checking service.
CVMFS Service to get the list of versions available
on CVMFS for a given app.
"""
def __init__(self, cvmfs_root, valid_apps):
"""Initialise."""
self.cvmfs_root = cvmfs_root
self.valid_apps = valid_apps
@cherrypy.expose
def index(self, appid=None):
"""Return the index page."""
print "IN CVMFSAppVersion: appid=(%s)" % appid
if appid not in self.valid_apps:
print "Invalid app type %s" % appid
return ''
html_ = html.HTML()
_, dirs, _ = os.walk(os.path.join(self.cvmfs_root, appid)).next()
for dir_ in natsorted(dirs):
for version in VERSION_RE.findall(dir_):
html_.option(version)
return str(html_)
|
"""CVMFS Servcice."""
import os
import re
import cherrypy
import html
from natsort import natsorted
version_re = re.compile(r"^release-(\d{1,3}\.\d{1,3}\.\d{1,3})$")
@cherrypy.popargs('appid')
class CVMFSAppVersions(object):
"""
CVMFS App Version checking service.
CVMFS Service to get the list of versions available
on CVMFS for a given app.
"""
def __init__(self, cvmfs_root, valid_apps):
"""Initialise."""
self.cvmfs_root = cvmfs_root
self.valid_apps = valid_apps
@cherrypy.expose
def index(self, appid=None):
"""Return the index page."""
print "IN CVMFSAppVersion: appid=(%s)" % appid
if appid not in self.valid_apps:
print "Invalid app type %s" % appid
return ''
html_ = html.HTML()
_, dirs, _ = os.walk(os.path.join(self.cvmfs_root, appid)).next()
for dir_ in natsorted(dirs):
for version in version_re.findall(dir_):
html_.option(version)
return str(html_)
Change the re const name to uppercase."""CVMFS Servcice."""
import os
import re
import cherrypy
import html
from natsort import natsorted
VERSION_RE = re.compile(r"^release-(\d{1,3}\.\d{1,3}\.\d{1,3})$")
@cherrypy.popargs('appid')
class CVMFSAppVersions(object):
"""
CVMFS App Version checking service.
CVMFS Service to get the list of versions available
on CVMFS for a given app.
"""
def __init__(self, cvmfs_root, valid_apps):
"""Initialise."""
self.cvmfs_root = cvmfs_root
self.valid_apps = valid_apps
@cherrypy.expose
def index(self, appid=None):
"""Return the index page."""
print "IN CVMFSAppVersion: appid=(%s)" % appid
if appid not in self.valid_apps:
print "Invalid app type %s" % appid
return ''
html_ = html.HTML()
_, dirs, _ = os.walk(os.path.join(self.cvmfs_root, appid)).next()
for dir_ in natsorted(dirs):
for version in VERSION_RE.findall(dir_):
html_.option(version)
return str(html_)
|
<commit_before>"""CVMFS Servcice."""
import os
import re
import cherrypy
import html
from natsort import natsorted
version_re = re.compile(r"^release-(\d{1,3}\.\d{1,3}\.\d{1,3})$")
@cherrypy.popargs('appid')
class CVMFSAppVersions(object):
"""
CVMFS App Version checking service.
CVMFS Service to get the list of versions available
on CVMFS for a given app.
"""
def __init__(self, cvmfs_root, valid_apps):
"""Initialise."""
self.cvmfs_root = cvmfs_root
self.valid_apps = valid_apps
@cherrypy.expose
def index(self, appid=None):
"""Return the index page."""
print "IN CVMFSAppVersion: appid=(%s)" % appid
if appid not in self.valid_apps:
print "Invalid app type %s" % appid
return ''
html_ = html.HTML()
_, dirs, _ = os.walk(os.path.join(self.cvmfs_root, appid)).next()
for dir_ in natsorted(dirs):
for version in version_re.findall(dir_):
html_.option(version)
return str(html_)
<commit_msg>Change the re const name to uppercase.<commit_after>"""CVMFS Servcice."""
import os
import re
import cherrypy
import html
from natsort import natsorted
VERSION_RE = re.compile(r"^release-(\d{1,3}\.\d{1,3}\.\d{1,3})$")
@cherrypy.popargs('appid')
class CVMFSAppVersions(object):
"""
CVMFS App Version checking service.
CVMFS Service to get the list of versions available
on CVMFS for a given app.
"""
def __init__(self, cvmfs_root, valid_apps):
"""Initialise."""
self.cvmfs_root = cvmfs_root
self.valid_apps = valid_apps
@cherrypy.expose
def index(self, appid=None):
"""Return the index page."""
print "IN CVMFSAppVersion: appid=(%s)" % appid
if appid not in self.valid_apps:
print "Invalid app type %s" % appid
return ''
html_ = html.HTML()
_, dirs, _ = os.walk(os.path.join(self.cvmfs_root, appid)).next()
for dir_ in natsorted(dirs):
for version in VERSION_RE.findall(dir_):
html_.option(version)
return str(html_)
|
b26047600202a9776c99323813cf17b0aa951dcd
|
app/routes.py
|
app/routes.py
|
from flask import jsonify
from . import app
import mapper
import utils
from predict import predictor
@app.route("/", methods=["GET"])
def index():
firebase_dump = mapper.get_dump_firebase()
response = firebase_dump.get_all()
response = response or {}
return jsonify(response)
@app.route("/build", methods=["POST"])
def build_model():
predictor.preprocess_airports()
if not predictor.model:
predictor.build_model()
return jsonify({"message:" : "OK"})
@app.route("/predict", methods=["GET"])
def predict_all_delays():
results = None
try:
results = predictor.predict_all()
except Exception as e:
return jsonify({"message" : e.message})
return jsonify(results)
@app.route("/predict/<airport_code>", methods=["GET"])
def predict_delay(airport_code):
firebase_source = mapper.get_source_firebase()
airport_status = firebase_source.get_airport(airport_code)
cleaned_data = utils.get_clean_data(airport_status)
res = predictor.predict(airport_code)
cleaned_data["prediction"] = bool(res[0])
return jsonify(cleaned_data)
@app.route("/status", methods=["GET"])
def get_airport_statuses():
firebase_source = mapper.get_source_firebase()
airports = firebase_source.get_all()
for airport_code, status in airports.items():
if "status" in status:
del status["status"]
return jsonify(airports)
|
from flask import jsonify
from . import app
import mapper
import utils
from predict import predictor
@app.route("/", methods=["GET"])
def index():
return app.send_static_file("index.html")
@app.route("/build", methods=["POST"])
def build_model():
predictor.preprocess_airports()
if not predictor.model:
predictor.build_model()
return jsonify({"message:" : "OK"})
@app.route("/predict", methods=["GET"])
def predict_all_delays():
results = None
try:
results = predictor.predict_all()
except Exception as e:
print "ERROR", e.message
return jsonify({"message" : e.message})
return jsonify(results)
@app.route("/predict/<airport_code>", methods=["GET"])
def predict_delay(airport_code):
firebase_source = mapper.get_source_firebase()
airport_status = firebase_source.get_airport(airport_code)
cleaned_data = utils.get_clean_data(airport_status)
res = predictor.predict(airport_code)
cleaned_data["prediction"] = bool(res[0])
return jsonify(cleaned_data)
@app.route("/status", methods=["GET"])
def get_airport_statuses():
firebase_source = mapper.get_source_firebase()
airports = firebase_source.get_all()
results = []
for airport_code, status in airports.items():
try:
results.append(utils.get_clean_data(status))
except:
pass
results = {"items":results}
return jsonify(results)
|
Return index.html in root and transform /status results
|
Return index.html in root and transform /status results
|
Python
|
mit
|
MachineLearningProject/flight-delay-prediction,MachineLearningProject/flight-delay-prediction,MachineLearningProject/flight-delay-prediction
|
from flask import jsonify
from . import app
import mapper
import utils
from predict import predictor
@app.route("/", methods=["GET"])
def index():
firebase_dump = mapper.get_dump_firebase()
response = firebase_dump.get_all()
response = response or {}
return jsonify(response)
@app.route("/build", methods=["POST"])
def build_model():
predictor.preprocess_airports()
if not predictor.model:
predictor.build_model()
return jsonify({"message:" : "OK"})
@app.route("/predict", methods=["GET"])
def predict_all_delays():
results = None
try:
results = predictor.predict_all()
except Exception as e:
return jsonify({"message" : e.message})
return jsonify(results)
@app.route("/predict/<airport_code>", methods=["GET"])
def predict_delay(airport_code):
firebase_source = mapper.get_source_firebase()
airport_status = firebase_source.get_airport(airport_code)
cleaned_data = utils.get_clean_data(airport_status)
res = predictor.predict(airport_code)
cleaned_data["prediction"] = bool(res[0])
return jsonify(cleaned_data)
@app.route("/status", methods=["GET"])
def get_airport_statuses():
firebase_source = mapper.get_source_firebase()
airports = firebase_source.get_all()
for airport_code, status in airports.items():
if "status" in status:
del status["status"]
return jsonify(airports)
Return index.html in root and transform /status results
|
from flask import jsonify
from . import app
import mapper
import utils
from predict import predictor
@app.route("/", methods=["GET"])
def index():
return app.send_static_file("index.html")
@app.route("/build", methods=["POST"])
def build_model():
predictor.preprocess_airports()
if not predictor.model:
predictor.build_model()
return jsonify({"message:" : "OK"})
@app.route("/predict", methods=["GET"])
def predict_all_delays():
results = None
try:
results = predictor.predict_all()
except Exception as e:
print "ERROR", e.message
return jsonify({"message" : e.message})
return jsonify(results)
@app.route("/predict/<airport_code>", methods=["GET"])
def predict_delay(airport_code):
firebase_source = mapper.get_source_firebase()
airport_status = firebase_source.get_airport(airport_code)
cleaned_data = utils.get_clean_data(airport_status)
res = predictor.predict(airport_code)
cleaned_data["prediction"] = bool(res[0])
return jsonify(cleaned_data)
@app.route("/status", methods=["GET"])
def get_airport_statuses():
firebase_source = mapper.get_source_firebase()
airports = firebase_source.get_all()
results = []
for airport_code, status in airports.items():
try:
results.append(utils.get_clean_data(status))
except:
pass
results = {"items":results}
return jsonify(results)
|
<commit_before>from flask import jsonify
from . import app
import mapper
import utils
from predict import predictor
@app.route("/", methods=["GET"])
def index():
firebase_dump = mapper.get_dump_firebase()
response = firebase_dump.get_all()
response = response or {}
return jsonify(response)
@app.route("/build", methods=["POST"])
def build_model():
predictor.preprocess_airports()
if not predictor.model:
predictor.build_model()
return jsonify({"message:" : "OK"})
@app.route("/predict", methods=["GET"])
def predict_all_delays():
results = None
try:
results = predictor.predict_all()
except Exception as e:
return jsonify({"message" : e.message})
return jsonify(results)
@app.route("/predict/<airport_code>", methods=["GET"])
def predict_delay(airport_code):
firebase_source = mapper.get_source_firebase()
airport_status = firebase_source.get_airport(airport_code)
cleaned_data = utils.get_clean_data(airport_status)
res = predictor.predict(airport_code)
cleaned_data["prediction"] = bool(res[0])
return jsonify(cleaned_data)
@app.route("/status", methods=["GET"])
def get_airport_statuses():
firebase_source = mapper.get_source_firebase()
airports = firebase_source.get_all()
for airport_code, status in airports.items():
if "status" in status:
del status["status"]
return jsonify(airports)
<commit_msg>Return index.html in root and transform /status results<commit_after>
|
from flask import jsonify
from . import app
import mapper
import utils
from predict import predictor
@app.route("/", methods=["GET"])
def index():
return app.send_static_file("index.html")
@app.route("/build", methods=["POST"])
def build_model():
predictor.preprocess_airports()
if not predictor.model:
predictor.build_model()
return jsonify({"message:" : "OK"})
@app.route("/predict", methods=["GET"])
def predict_all_delays():
results = None
try:
results = predictor.predict_all()
except Exception as e:
print "ERROR", e.message
return jsonify({"message" : e.message})
return jsonify(results)
@app.route("/predict/<airport_code>", methods=["GET"])
def predict_delay(airport_code):
firebase_source = mapper.get_source_firebase()
airport_status = firebase_source.get_airport(airport_code)
cleaned_data = utils.get_clean_data(airport_status)
res = predictor.predict(airport_code)
cleaned_data["prediction"] = bool(res[0])
return jsonify(cleaned_data)
@app.route("/status", methods=["GET"])
def get_airport_statuses():
firebase_source = mapper.get_source_firebase()
airports = firebase_source.get_all()
results = []
for airport_code, status in airports.items():
try:
results.append(utils.get_clean_data(status))
except:
pass
results = {"items":results}
return jsonify(results)
|
from flask import jsonify
from . import app
import mapper
import utils
from predict import predictor
@app.route("/", methods=["GET"])
def index():
firebase_dump = mapper.get_dump_firebase()
response = firebase_dump.get_all()
response = response or {}
return jsonify(response)
@app.route("/build", methods=["POST"])
def build_model():
predictor.preprocess_airports()
if not predictor.model:
predictor.build_model()
return jsonify({"message:" : "OK"})
@app.route("/predict", methods=["GET"])
def predict_all_delays():
results = None
try:
results = predictor.predict_all()
except Exception as e:
return jsonify({"message" : e.message})
return jsonify(results)
@app.route("/predict/<airport_code>", methods=["GET"])
def predict_delay(airport_code):
firebase_source = mapper.get_source_firebase()
airport_status = firebase_source.get_airport(airport_code)
cleaned_data = utils.get_clean_data(airport_status)
res = predictor.predict(airport_code)
cleaned_data["prediction"] = bool(res[0])
return jsonify(cleaned_data)
@app.route("/status", methods=["GET"])
def get_airport_statuses():
firebase_source = mapper.get_source_firebase()
airports = firebase_source.get_all()
for airport_code, status in airports.items():
if "status" in status:
del status["status"]
return jsonify(airports)
Return index.html in root and transform /status resultsfrom flask import jsonify
from . import app
import mapper
import utils
from predict import predictor
@app.route("/", methods=["GET"])
def index():
return app.send_static_file("index.html")
@app.route("/build", methods=["POST"])
def build_model():
predictor.preprocess_airports()
if not predictor.model:
predictor.build_model()
return jsonify({"message:" : "OK"})
@app.route("/predict", methods=["GET"])
def predict_all_delays():
results = None
try:
results = predictor.predict_all()
except Exception as e:
print "ERROR", e.message
return jsonify({"message" : e.message})
return jsonify(results)
@app.route("/predict/<airport_code>", methods=["GET"])
def predict_delay(airport_code):
firebase_source = mapper.get_source_firebase()
airport_status = firebase_source.get_airport(airport_code)
cleaned_data = utils.get_clean_data(airport_status)
res = predictor.predict(airport_code)
cleaned_data["prediction"] = bool(res[0])
return jsonify(cleaned_data)
@app.route("/status", methods=["GET"])
def get_airport_statuses():
firebase_source = mapper.get_source_firebase()
airports = firebase_source.get_all()
results = []
for airport_code, status in airports.items():
try:
results.append(utils.get_clean_data(status))
except:
pass
results = {"items":results}
return jsonify(results)
|
<commit_before>from flask import jsonify
from . import app
import mapper
import utils
from predict import predictor
@app.route("/", methods=["GET"])
def index():
firebase_dump = mapper.get_dump_firebase()
response = firebase_dump.get_all()
response = response or {}
return jsonify(response)
@app.route("/build", methods=["POST"])
def build_model():
predictor.preprocess_airports()
if not predictor.model:
predictor.build_model()
return jsonify({"message:" : "OK"})
@app.route("/predict", methods=["GET"])
def predict_all_delays():
results = None
try:
results = predictor.predict_all()
except Exception as e:
return jsonify({"message" : e.message})
return jsonify(results)
@app.route("/predict/<airport_code>", methods=["GET"])
def predict_delay(airport_code):
firebase_source = mapper.get_source_firebase()
airport_status = firebase_source.get_airport(airport_code)
cleaned_data = utils.get_clean_data(airport_status)
res = predictor.predict(airport_code)
cleaned_data["prediction"] = bool(res[0])
return jsonify(cleaned_data)
@app.route("/status", methods=["GET"])
def get_airport_statuses():
firebase_source = mapper.get_source_firebase()
airports = firebase_source.get_all()
for airport_code, status in airports.items():
if "status" in status:
del status["status"]
return jsonify(airports)
<commit_msg>Return index.html in root and transform /status results<commit_after>from flask import jsonify
from . import app
import mapper
import utils
from predict import predictor
@app.route("/", methods=["GET"])
def index():
return app.send_static_file("index.html")
@app.route("/build", methods=["POST"])
def build_model():
predictor.preprocess_airports()
if not predictor.model:
predictor.build_model()
return jsonify({"message:" : "OK"})
@app.route("/predict", methods=["GET"])
def predict_all_delays():
results = None
try:
results = predictor.predict_all()
except Exception as e:
print "ERROR", e.message
return jsonify({"message" : e.message})
return jsonify(results)
@app.route("/predict/<airport_code>", methods=["GET"])
def predict_delay(airport_code):
firebase_source = mapper.get_source_firebase()
airport_status = firebase_source.get_airport(airport_code)
cleaned_data = utils.get_clean_data(airport_status)
res = predictor.predict(airport_code)
cleaned_data["prediction"] = bool(res[0])
return jsonify(cleaned_data)
@app.route("/status", methods=["GET"])
def get_airport_statuses():
firebase_source = mapper.get_source_firebase()
airports = firebase_source.get_all()
results = []
for airport_code, status in airports.items():
try:
results.append(utils.get_clean_data(status))
except:
pass
results = {"items":results}
return jsonify(results)
|
9bb14514a523484af6313008baef3b7cfd987951
|
tests/__init__.py
|
tests/__init__.py
|
import sys
import doctest
def fix_doctests(suite):
if sys.version_info.major >= 3:
return
for case in suite._tests:
# Add some more flags.
case._dt_optionflags = (
(case._dt_optionflags or 0) |
doctest.IGNORE_EXCEPTION_DETAIL |
doctest.ELLIPSIS |
doctest.NORMALIZE_WHITESPACE
)
test = case._dt_test
for example in test.examples:
# Remove b prefix from strings.
if example.want.startswith("b'"):
example.want = example.want[1:]
def get_doctests(mod):
suite = doctest.DocTestSuite(mod)
fix_doctests(suite)
return suite
|
import sys
import doctest
def fix_doctests(suite):
if sys.version_info[0] >= 3:
return
for case in suite._tests:
# Add some more flags.
case._dt_optionflags = (
(case._dt_optionflags or 0) |
doctest.IGNORE_EXCEPTION_DETAIL |
doctest.ELLIPSIS |
doctest.NORMALIZE_WHITESPACE
)
test = case._dt_test
for example in test.examples:
# Remove b prefix from strings.
if example.want.startswith("b'"):
example.want = example.want[1:]
def get_doctests(mod):
suite = doctest.DocTestSuite(mod)
fix_doctests(suite)
return suite
|
Fix version_info check for Python2.6
|
Tests: Fix version_info check for Python2.6
|
Python
|
bsd-3-clause
|
mikeboers/PyTomCrypt,mikeboers/PyTomCrypt,mikeboers/PyTomCrypt
|
import sys
import doctest
def fix_doctests(suite):
if sys.version_info.major >= 3:
return
for case in suite._tests:
# Add some more flags.
case._dt_optionflags = (
(case._dt_optionflags or 0) |
doctest.IGNORE_EXCEPTION_DETAIL |
doctest.ELLIPSIS |
doctest.NORMALIZE_WHITESPACE
)
test = case._dt_test
for example in test.examples:
# Remove b prefix from strings.
if example.want.startswith("b'"):
example.want = example.want[1:]
def get_doctests(mod):
suite = doctest.DocTestSuite(mod)
fix_doctests(suite)
return suite
Tests: Fix version_info check for Python2.6
|
import sys
import doctest
def fix_doctests(suite):
if sys.version_info[0] >= 3:
return
for case in suite._tests:
# Add some more flags.
case._dt_optionflags = (
(case._dt_optionflags or 0) |
doctest.IGNORE_EXCEPTION_DETAIL |
doctest.ELLIPSIS |
doctest.NORMALIZE_WHITESPACE
)
test = case._dt_test
for example in test.examples:
# Remove b prefix from strings.
if example.want.startswith("b'"):
example.want = example.want[1:]
def get_doctests(mod):
suite = doctest.DocTestSuite(mod)
fix_doctests(suite)
return suite
|
<commit_before>import sys
import doctest
def fix_doctests(suite):
if sys.version_info.major >= 3:
return
for case in suite._tests:
# Add some more flags.
case._dt_optionflags = (
(case._dt_optionflags or 0) |
doctest.IGNORE_EXCEPTION_DETAIL |
doctest.ELLIPSIS |
doctest.NORMALIZE_WHITESPACE
)
test = case._dt_test
for example in test.examples:
# Remove b prefix from strings.
if example.want.startswith("b'"):
example.want = example.want[1:]
def get_doctests(mod):
suite = doctest.DocTestSuite(mod)
fix_doctests(suite)
return suite
<commit_msg>Tests: Fix version_info check for Python2.6<commit_after>
|
import sys
import doctest
def fix_doctests(suite):
if sys.version_info[0] >= 3:
return
for case in suite._tests:
# Add some more flags.
case._dt_optionflags = (
(case._dt_optionflags or 0) |
doctest.IGNORE_EXCEPTION_DETAIL |
doctest.ELLIPSIS |
doctest.NORMALIZE_WHITESPACE
)
test = case._dt_test
for example in test.examples:
# Remove b prefix from strings.
if example.want.startswith("b'"):
example.want = example.want[1:]
def get_doctests(mod):
suite = doctest.DocTestSuite(mod)
fix_doctests(suite)
return suite
|
import sys
import doctest
def fix_doctests(suite):
if sys.version_info.major >= 3:
return
for case in suite._tests:
# Add some more flags.
case._dt_optionflags = (
(case._dt_optionflags or 0) |
doctest.IGNORE_EXCEPTION_DETAIL |
doctest.ELLIPSIS |
doctest.NORMALIZE_WHITESPACE
)
test = case._dt_test
for example in test.examples:
# Remove b prefix from strings.
if example.want.startswith("b'"):
example.want = example.want[1:]
def get_doctests(mod):
suite = doctest.DocTestSuite(mod)
fix_doctests(suite)
return suite
Tests: Fix version_info check for Python2.6import sys
import doctest
def fix_doctests(suite):
if sys.version_info[0] >= 3:
return
for case in suite._tests:
# Add some more flags.
case._dt_optionflags = (
(case._dt_optionflags or 0) |
doctest.IGNORE_EXCEPTION_DETAIL |
doctest.ELLIPSIS |
doctest.NORMALIZE_WHITESPACE
)
test = case._dt_test
for example in test.examples:
# Remove b prefix from strings.
if example.want.startswith("b'"):
example.want = example.want[1:]
def get_doctests(mod):
suite = doctest.DocTestSuite(mod)
fix_doctests(suite)
return suite
|
<commit_before>import sys
import doctest
def fix_doctests(suite):
if sys.version_info.major >= 3:
return
for case in suite._tests:
# Add some more flags.
case._dt_optionflags = (
(case._dt_optionflags or 0) |
doctest.IGNORE_EXCEPTION_DETAIL |
doctest.ELLIPSIS |
doctest.NORMALIZE_WHITESPACE
)
test = case._dt_test
for example in test.examples:
# Remove b prefix from strings.
if example.want.startswith("b'"):
example.want = example.want[1:]
def get_doctests(mod):
suite = doctest.DocTestSuite(mod)
fix_doctests(suite)
return suite
<commit_msg>Tests: Fix version_info check for Python2.6<commit_after>import sys
import doctest
def fix_doctests(suite):
if sys.version_info[0] >= 3:
return
for case in suite._tests:
# Add some more flags.
case._dt_optionflags = (
(case._dt_optionflags or 0) |
doctest.IGNORE_EXCEPTION_DETAIL |
doctest.ELLIPSIS |
doctest.NORMALIZE_WHITESPACE
)
test = case._dt_test
for example in test.examples:
# Remove b prefix from strings.
if example.want.startswith("b'"):
example.want = example.want[1:]
def get_doctests(mod):
suite = doctest.DocTestSuite(mod)
fix_doctests(suite)
return suite
|
600839e3c51d2091a6c434ac31ea11dc9ed2db85
|
foialist/forms.py
|
foialist/forms.py
|
from django import forms
from foialist.models import *
class FileForm(forms.ModelForm):
class Meta:
model = File
exclude = ('entry', 'size')
class EntryForm(forms.ModelForm):
govt_entity = forms.CharField(label="Gov't. entity")
class Meta:
model = Entry
# exclude = ('slug', 'poster_slug', 'show', 'date_posted', 'entity')
fields = ('title', 'narrative', 'government_entity', 'date_requested', 'date_filed', 'poster', 'email')
class CommentForm(forms.ModelForm):
poster = forms.CharField()
class Meta:
model = Comment
|
from django import forms
from foialist.models import *
class FileForm(forms.ModelForm):
class Meta:
model = File
exclude = ('entry', 'size')
class EntryForm(forms.ModelForm):
govt_entity = forms.CharField(label="Gov't. entity")
class Meta:
model = Entry
fields = ('title', 'narrative', 'govt_entity', 'date_requested',
'date_filed', 'poster', 'email')
class CommentForm(forms.ModelForm):
poster = forms.CharField()
class Meta:
model = Comment
|
Correct mismatched field names in EntryForm.
|
Correct mismatched field names in EntryForm.
|
Python
|
bsd-3-clause
|
a2civictech/a2docs-sources,a2civictech/a2docs-sources,a2civictech/a2docs-sources
|
from django import forms
from foialist.models import *
class FileForm(forms.ModelForm):
class Meta:
model = File
exclude = ('entry', 'size')
class EntryForm(forms.ModelForm):
govt_entity = forms.CharField(label="Gov't. entity")
class Meta:
model = Entry
# exclude = ('slug', 'poster_slug', 'show', 'date_posted', 'entity')
fields = ('title', 'narrative', 'government_entity', 'date_requested', 'date_filed', 'poster', 'email')
class CommentForm(forms.ModelForm):
poster = forms.CharField()
class Meta:
model = CommentCorrect mismatched field names in EntryForm.
|
from django import forms
from foialist.models import *
class FileForm(forms.ModelForm):
class Meta:
model = File
exclude = ('entry', 'size')
class EntryForm(forms.ModelForm):
govt_entity = forms.CharField(label="Gov't. entity")
class Meta:
model = Entry
fields = ('title', 'narrative', 'govt_entity', 'date_requested',
'date_filed', 'poster', 'email')
class CommentForm(forms.ModelForm):
poster = forms.CharField()
class Meta:
model = Comment
|
<commit_before>from django import forms
from foialist.models import *
class FileForm(forms.ModelForm):
class Meta:
model = File
exclude = ('entry', 'size')
class EntryForm(forms.ModelForm):
govt_entity = forms.CharField(label="Gov't. entity")
class Meta:
model = Entry
# exclude = ('slug', 'poster_slug', 'show', 'date_posted', 'entity')
fields = ('title', 'narrative', 'government_entity', 'date_requested', 'date_filed', 'poster', 'email')
class CommentForm(forms.ModelForm):
poster = forms.CharField()
class Meta:
model = Comment<commit_msg>Correct mismatched field names in EntryForm.<commit_after>
|
from django import forms
from foialist.models import *
class FileForm(forms.ModelForm):
class Meta:
model = File
exclude = ('entry', 'size')
class EntryForm(forms.ModelForm):
govt_entity = forms.CharField(label="Gov't. entity")
class Meta:
model = Entry
fields = ('title', 'narrative', 'govt_entity', 'date_requested',
'date_filed', 'poster', 'email')
class CommentForm(forms.ModelForm):
poster = forms.CharField()
class Meta:
model = Comment
|
from django import forms
from foialist.models import *
class FileForm(forms.ModelForm):
class Meta:
model = File
exclude = ('entry', 'size')
class EntryForm(forms.ModelForm):
govt_entity = forms.CharField(label="Gov't. entity")
class Meta:
model = Entry
# exclude = ('slug', 'poster_slug', 'show', 'date_posted', 'entity')
fields = ('title', 'narrative', 'government_entity', 'date_requested', 'date_filed', 'poster', 'email')
class CommentForm(forms.ModelForm):
poster = forms.CharField()
class Meta:
model = CommentCorrect mismatched field names in EntryForm.from django import forms
from foialist.models import *
class FileForm(forms.ModelForm):
class Meta:
model = File
exclude = ('entry', 'size')
class EntryForm(forms.ModelForm):
govt_entity = forms.CharField(label="Gov't. entity")
class Meta:
model = Entry
fields = ('title', 'narrative', 'govt_entity', 'date_requested',
'date_filed', 'poster', 'email')
class CommentForm(forms.ModelForm):
poster = forms.CharField()
class Meta:
model = Comment
|
<commit_before>from django import forms
from foialist.models import *
class FileForm(forms.ModelForm):
class Meta:
model = File
exclude = ('entry', 'size')
class EntryForm(forms.ModelForm):
govt_entity = forms.CharField(label="Gov't. entity")
class Meta:
model = Entry
# exclude = ('slug', 'poster_slug', 'show', 'date_posted', 'entity')
fields = ('title', 'narrative', 'government_entity', 'date_requested', 'date_filed', 'poster', 'email')
class CommentForm(forms.ModelForm):
poster = forms.CharField()
class Meta:
model = Comment<commit_msg>Correct mismatched field names in EntryForm.<commit_after>from django import forms
from foialist.models import *
class FileForm(forms.ModelForm):
class Meta:
model = File
exclude = ('entry', 'size')
class EntryForm(forms.ModelForm):
govt_entity = forms.CharField(label="Gov't. entity")
class Meta:
model = Entry
fields = ('title', 'narrative', 'govt_entity', 'date_requested',
'date_filed', 'poster', 'email')
class CommentForm(forms.ModelForm):
poster = forms.CharField()
class Meta:
model = Comment
|
5d7a179e99632e2b8ca30bfa444497636492ca5a
|
catsnap/web/middleware/exception_logger.py
|
catsnap/web/middleware/exception_logger.py
|
import sys
import traceback
class ExceptionLogger(object):
def __init__(self, app):
self.app = app
def __call__(self, environ, start_response):
try:
return self.app(environ, start_response)
except Exception:
(exc_type, exc_value, trace) = sys.exc_info()
traceback.print_exception(exc_type,
exc_value,
trace,
sys.stdout)
raise
|
import sys
import traceback
class ExceptionLogger(object):
def __init__(self, app):
self.app = app
def __call__(self, environ, start_response):
try:
return self.app(environ, start_response)
except Exception:
(exc_type, exc_value, trace) = sys.exc_info()
traceback.print_exception(exc_type,
exc_value,
trace,
file=sys.stdout)
raise
|
Fix the exception logger to actually log to stdout
|
Fix the exception logger to actually log to stdout
|
Python
|
mit
|
ErinCall/catsnap,ErinCall/catsnap,ErinCall/catsnap
|
import sys
import traceback
class ExceptionLogger(object):
def __init__(self, app):
self.app = app
def __call__(self, environ, start_response):
try:
return self.app(environ, start_response)
except Exception:
(exc_type, exc_value, trace) = sys.exc_info()
traceback.print_exception(exc_type,
exc_value,
trace,
sys.stdout)
raise
Fix the exception logger to actually log to stdout
|
import sys
import traceback
class ExceptionLogger(object):
def __init__(self, app):
self.app = app
def __call__(self, environ, start_response):
try:
return self.app(environ, start_response)
except Exception:
(exc_type, exc_value, trace) = sys.exc_info()
traceback.print_exception(exc_type,
exc_value,
trace,
file=sys.stdout)
raise
|
<commit_before>import sys
import traceback
class ExceptionLogger(object):
def __init__(self, app):
self.app = app
def __call__(self, environ, start_response):
try:
return self.app(environ, start_response)
except Exception:
(exc_type, exc_value, trace) = sys.exc_info()
traceback.print_exception(exc_type,
exc_value,
trace,
sys.stdout)
raise
<commit_msg>Fix the exception logger to actually log to stdout<commit_after>
|
import sys
import traceback
class ExceptionLogger(object):
def __init__(self, app):
self.app = app
def __call__(self, environ, start_response):
try:
return self.app(environ, start_response)
except Exception:
(exc_type, exc_value, trace) = sys.exc_info()
traceback.print_exception(exc_type,
exc_value,
trace,
file=sys.stdout)
raise
|
import sys
import traceback
class ExceptionLogger(object):
def __init__(self, app):
self.app = app
def __call__(self, environ, start_response):
try:
return self.app(environ, start_response)
except Exception:
(exc_type, exc_value, trace) = sys.exc_info()
traceback.print_exception(exc_type,
exc_value,
trace,
sys.stdout)
raise
Fix the exception logger to actually log to stdoutimport sys
import traceback
class ExceptionLogger(object):
def __init__(self, app):
self.app = app
def __call__(self, environ, start_response):
try:
return self.app(environ, start_response)
except Exception:
(exc_type, exc_value, trace) = sys.exc_info()
traceback.print_exception(exc_type,
exc_value,
trace,
file=sys.stdout)
raise
|
<commit_before>import sys
import traceback
class ExceptionLogger(object):
def __init__(self, app):
self.app = app
def __call__(self, environ, start_response):
try:
return self.app(environ, start_response)
except Exception:
(exc_type, exc_value, trace) = sys.exc_info()
traceback.print_exception(exc_type,
exc_value,
trace,
sys.stdout)
raise
<commit_msg>Fix the exception logger to actually log to stdout<commit_after>import sys
import traceback
class ExceptionLogger(object):
def __init__(self, app):
self.app = app
def __call__(self, environ, start_response):
try:
return self.app(environ, start_response)
except Exception:
(exc_type, exc_value, trace) = sys.exc_info()
traceback.print_exception(exc_type,
exc_value,
trace,
file=sys.stdout)
raise
|
6a7302bed399aba98b01490f78728d3daa57e092
|
opps/images/generate.py
|
opps/images/generate.py
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from libthumbor import CryptoURL
from django.conf import settings
crypto = CryptoURL(key=settings.THUMBOR_SECURITY_KEY)
def _remove_prefix(url, prefix):
if url.startswith(prefix):
return url[len(prefix):]
return url
def _remove_schema(url):
return _remove_prefix(url, 'http://')
def _prepend_media_url(url):
if url.startswith(settings.MEDIA_URL):
url = _remove_prefix(url, settings.MEDIA_URL)
url.lstrip('/')
return u'{}/{}'.format(settings.THUMBOR_MEDIA_URL, url)
return url
def image_url(image_url, **kwargs):
if not settings.THUMBOR_ENABLED:
# TODO: alternative ways for generating the thumbnails
return image_url
image_url = _prepend_media_url(image_url)
image_url = _remove_schema(image_url)
encrypted_url = crypto.generate(image_url=image_url,
**dict(settings.THUMBOR_ARGUMENTS,
**kwargs)).strip('/')
return u'{}/{}'.format(settings.THUMBOR_SERVER, encrypted_url)
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from libthumbor import CryptoURL
from django.conf import settings
crypto = CryptoURL(key=settings.THUMBOR_SECURITY_KEY)
def _remove_prefix(url, prefix):
if url.startswith(prefix):
return url[len(prefix):]
return url
def _remove_schema(url):
return _remove_prefix(url, 'http://')
def _prepend_media_url(url):
if url.startswith(settings.MEDIA_URL):
url = _remove_prefix(url, settings.MEDIA_URL)
url.lstrip('/')
return u'{}/{}'.format(settings.THUMBOR_MEDIA_URL, url)
return url
def image_url(image_url, **kwargs):
if not settings.THUMBOR_ENABLED or image_url.split('.')[-1] == 'git':
# TODO: alternative ways for generating the thumbnails
return image_url
image_url = _prepend_media_url(image_url)
image_url = _remove_schema(image_url)
encrypted_url = crypto.generate(image_url=image_url,
**dict(settings.THUMBOR_ARGUMENTS,
**kwargs)).strip('/')
return u'{}/{}'.format(settings.THUMBOR_SERVER, encrypted_url)
|
Fix render git, if gif file not render via thumbor
|
Fix render git, if gif file not render via thumbor
|
Python
|
mit
|
opps/opps,YACOWS/opps,williamroot/opps,opps/opps,williamroot/opps,jeanmask/opps,jeanmask/opps,williamroot/opps,opps/opps,YACOWS/opps,jeanmask/opps,YACOWS/opps,williamroot/opps,jeanmask/opps,opps/opps,YACOWS/opps
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from libthumbor import CryptoURL
from django.conf import settings
crypto = CryptoURL(key=settings.THUMBOR_SECURITY_KEY)
def _remove_prefix(url, prefix):
if url.startswith(prefix):
return url[len(prefix):]
return url
def _remove_schema(url):
return _remove_prefix(url, 'http://')
def _prepend_media_url(url):
if url.startswith(settings.MEDIA_URL):
url = _remove_prefix(url, settings.MEDIA_URL)
url.lstrip('/')
return u'{}/{}'.format(settings.THUMBOR_MEDIA_URL, url)
return url
def image_url(image_url, **kwargs):
if not settings.THUMBOR_ENABLED:
# TODO: alternative ways for generating the thumbnails
return image_url
image_url = _prepend_media_url(image_url)
image_url = _remove_schema(image_url)
encrypted_url = crypto.generate(image_url=image_url,
**dict(settings.THUMBOR_ARGUMENTS,
**kwargs)).strip('/')
return u'{}/{}'.format(settings.THUMBOR_SERVER, encrypted_url)
Fix render git, if gif file not render via thumbor
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from libthumbor import CryptoURL
from django.conf import settings
crypto = CryptoURL(key=settings.THUMBOR_SECURITY_KEY)
def _remove_prefix(url, prefix):
if url.startswith(prefix):
return url[len(prefix):]
return url
def _remove_schema(url):
return _remove_prefix(url, 'http://')
def _prepend_media_url(url):
if url.startswith(settings.MEDIA_URL):
url = _remove_prefix(url, settings.MEDIA_URL)
url.lstrip('/')
return u'{}/{}'.format(settings.THUMBOR_MEDIA_URL, url)
return url
def image_url(image_url, **kwargs):
if not settings.THUMBOR_ENABLED or image_url.split('.')[-1] == 'git':
# TODO: alternative ways for generating the thumbnails
return image_url
image_url = _prepend_media_url(image_url)
image_url = _remove_schema(image_url)
encrypted_url = crypto.generate(image_url=image_url,
**dict(settings.THUMBOR_ARGUMENTS,
**kwargs)).strip('/')
return u'{}/{}'.format(settings.THUMBOR_SERVER, encrypted_url)
|
<commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from libthumbor import CryptoURL
from django.conf import settings
crypto = CryptoURL(key=settings.THUMBOR_SECURITY_KEY)
def _remove_prefix(url, prefix):
if url.startswith(prefix):
return url[len(prefix):]
return url
def _remove_schema(url):
return _remove_prefix(url, 'http://')
def _prepend_media_url(url):
if url.startswith(settings.MEDIA_URL):
url = _remove_prefix(url, settings.MEDIA_URL)
url.lstrip('/')
return u'{}/{}'.format(settings.THUMBOR_MEDIA_URL, url)
return url
def image_url(image_url, **kwargs):
if not settings.THUMBOR_ENABLED:
# TODO: alternative ways for generating the thumbnails
return image_url
image_url = _prepend_media_url(image_url)
image_url = _remove_schema(image_url)
encrypted_url = crypto.generate(image_url=image_url,
**dict(settings.THUMBOR_ARGUMENTS,
**kwargs)).strip('/')
return u'{}/{}'.format(settings.THUMBOR_SERVER, encrypted_url)
<commit_msg>Fix render git, if gif file not render via thumbor<commit_after>
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from libthumbor import CryptoURL
from django.conf import settings
crypto = CryptoURL(key=settings.THUMBOR_SECURITY_KEY)
def _remove_prefix(url, prefix):
if url.startswith(prefix):
return url[len(prefix):]
return url
def _remove_schema(url):
return _remove_prefix(url, 'http://')
def _prepend_media_url(url):
if url.startswith(settings.MEDIA_URL):
url = _remove_prefix(url, settings.MEDIA_URL)
url.lstrip('/')
return u'{}/{}'.format(settings.THUMBOR_MEDIA_URL, url)
return url
def image_url(image_url, **kwargs):
if not settings.THUMBOR_ENABLED or image_url.split('.')[-1] == 'git':
# TODO: alternative ways for generating the thumbnails
return image_url
image_url = _prepend_media_url(image_url)
image_url = _remove_schema(image_url)
encrypted_url = crypto.generate(image_url=image_url,
**dict(settings.THUMBOR_ARGUMENTS,
**kwargs)).strip('/')
return u'{}/{}'.format(settings.THUMBOR_SERVER, encrypted_url)
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from libthumbor import CryptoURL
from django.conf import settings
crypto = CryptoURL(key=settings.THUMBOR_SECURITY_KEY)
def _remove_prefix(url, prefix):
if url.startswith(prefix):
return url[len(prefix):]
return url
def _remove_schema(url):
return _remove_prefix(url, 'http://')
def _prepend_media_url(url):
if url.startswith(settings.MEDIA_URL):
url = _remove_prefix(url, settings.MEDIA_URL)
url.lstrip('/')
return u'{}/{}'.format(settings.THUMBOR_MEDIA_URL, url)
return url
def image_url(image_url, **kwargs):
if not settings.THUMBOR_ENABLED:
# TODO: alternative ways for generating the thumbnails
return image_url
image_url = _prepend_media_url(image_url)
image_url = _remove_schema(image_url)
encrypted_url = crypto.generate(image_url=image_url,
**dict(settings.THUMBOR_ARGUMENTS,
**kwargs)).strip('/')
return u'{}/{}'.format(settings.THUMBOR_SERVER, encrypted_url)
Fix render git, if gif file not render via thumbor#!/usr/bin/env python
# -*- coding: utf-8 -*-
from libthumbor import CryptoURL
from django.conf import settings
crypto = CryptoURL(key=settings.THUMBOR_SECURITY_KEY)
def _remove_prefix(url, prefix):
if url.startswith(prefix):
return url[len(prefix):]
return url
def _remove_schema(url):
return _remove_prefix(url, 'http://')
def _prepend_media_url(url):
if url.startswith(settings.MEDIA_URL):
url = _remove_prefix(url, settings.MEDIA_URL)
url.lstrip('/')
return u'{}/{}'.format(settings.THUMBOR_MEDIA_URL, url)
return url
def image_url(image_url, **kwargs):
if not settings.THUMBOR_ENABLED or image_url.split('.')[-1] == 'git':
# TODO: alternative ways for generating the thumbnails
return image_url
image_url = _prepend_media_url(image_url)
image_url = _remove_schema(image_url)
encrypted_url = crypto.generate(image_url=image_url,
**dict(settings.THUMBOR_ARGUMENTS,
**kwargs)).strip('/')
return u'{}/{}'.format(settings.THUMBOR_SERVER, encrypted_url)
|
<commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from libthumbor import CryptoURL
from django.conf import settings
crypto = CryptoURL(key=settings.THUMBOR_SECURITY_KEY)
def _remove_prefix(url, prefix):
if url.startswith(prefix):
return url[len(prefix):]
return url
def _remove_schema(url):
return _remove_prefix(url, 'http://')
def _prepend_media_url(url):
if url.startswith(settings.MEDIA_URL):
url = _remove_prefix(url, settings.MEDIA_URL)
url.lstrip('/')
return u'{}/{}'.format(settings.THUMBOR_MEDIA_URL, url)
return url
def image_url(image_url, **kwargs):
if not settings.THUMBOR_ENABLED:
# TODO: alternative ways for generating the thumbnails
return image_url
image_url = _prepend_media_url(image_url)
image_url = _remove_schema(image_url)
encrypted_url = crypto.generate(image_url=image_url,
**dict(settings.THUMBOR_ARGUMENTS,
**kwargs)).strip('/')
return u'{}/{}'.format(settings.THUMBOR_SERVER, encrypted_url)
<commit_msg>Fix render git, if gif file not render via thumbor<commit_after>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from libthumbor import CryptoURL
from django.conf import settings
crypto = CryptoURL(key=settings.THUMBOR_SECURITY_KEY)
def _remove_prefix(url, prefix):
if url.startswith(prefix):
return url[len(prefix):]
return url
def _remove_schema(url):
return _remove_prefix(url, 'http://')
def _prepend_media_url(url):
if url.startswith(settings.MEDIA_URL):
url = _remove_prefix(url, settings.MEDIA_URL)
url.lstrip('/')
return u'{}/{}'.format(settings.THUMBOR_MEDIA_URL, url)
return url
def image_url(image_url, **kwargs):
if not settings.THUMBOR_ENABLED or image_url.split('.')[-1] == 'git':
# TODO: alternative ways for generating the thumbnails
return image_url
image_url = _prepend_media_url(image_url)
image_url = _remove_schema(image_url)
encrypted_url = crypto.generate(image_url=image_url,
**dict(settings.THUMBOR_ARGUMENTS,
**kwargs)).strip('/')
return u'{}/{}'.format(settings.THUMBOR_SERVER, encrypted_url)
|
059230327fcebb35c881f8a6bc2ee12fed29d442
|
mcp/config.py
|
mcp/config.py
|
#
# Copyright (c) 2005-2006 rPath, Inc.
#
# All rights reserved
#
import os
from conary import conarycfg
from conary.lib import cfgtypes
class MCPConfig(conarycfg.ConfigFile):
basePath = os.path.join(os.path.sep, 'srv', 'rbuilder', 'mcp')
logPath = os.path.join(basePath, 'logs')
queueHost = '127.0.0.1'
queuePort = (cfgtypes.CfgInt, 61613)
namespace = 'mcp'
slaveTroveName = 'group-jobslave'
slaveTroveLabel = 'products.rpath.com@rpath:js'
|
#
# Copyright (c) 2005-2006 rPath, Inc.
#
# All rights reserved
#
import os
from conary import conarycfg
from conary.lib import cfgtypes
class MCPConfig(conarycfg.ConfigFile):
basePath = os.path.join(os.path.sep, 'srv', 'rbuilder', 'mcp')
logPath = os.path.join(os.path.sep, 'var', 'log', 'rbuilder')
queueHost = '127.0.0.1'
queuePort = (cfgtypes.CfgInt, 61613)
namespace = 'mcp'
slaveTroveName = 'group-jobslave'
slaveTroveLabel = 'products.rpath.com@rpath:js'
|
Move default location for MCP logs into /var/log/rbuilder/
|
Move default location for MCP logs into /var/log/rbuilder/
|
Python
|
apache-2.0
|
sassoftware/mcp,sassoftware/mcp
|
#
# Copyright (c) 2005-2006 rPath, Inc.
#
# All rights reserved
#
import os
from conary import conarycfg
from conary.lib import cfgtypes
class MCPConfig(conarycfg.ConfigFile):
basePath = os.path.join(os.path.sep, 'srv', 'rbuilder', 'mcp')
logPath = os.path.join(basePath, 'logs')
queueHost = '127.0.0.1'
queuePort = (cfgtypes.CfgInt, 61613)
namespace = 'mcp'
slaveTroveName = 'group-jobslave'
slaveTroveLabel = 'products.rpath.com@rpath:js'
Move default location for MCP logs into /var/log/rbuilder/
|
#
# Copyright (c) 2005-2006 rPath, Inc.
#
# All rights reserved
#
import os
from conary import conarycfg
from conary.lib import cfgtypes
class MCPConfig(conarycfg.ConfigFile):
basePath = os.path.join(os.path.sep, 'srv', 'rbuilder', 'mcp')
logPath = os.path.join(os.path.sep, 'var', 'log', 'rbuilder')
queueHost = '127.0.0.1'
queuePort = (cfgtypes.CfgInt, 61613)
namespace = 'mcp'
slaveTroveName = 'group-jobslave'
slaveTroveLabel = 'products.rpath.com@rpath:js'
|
<commit_before>#
# Copyright (c) 2005-2006 rPath, Inc.
#
# All rights reserved
#
import os
from conary import conarycfg
from conary.lib import cfgtypes
class MCPConfig(conarycfg.ConfigFile):
basePath = os.path.join(os.path.sep, 'srv', 'rbuilder', 'mcp')
logPath = os.path.join(basePath, 'logs')
queueHost = '127.0.0.1'
queuePort = (cfgtypes.CfgInt, 61613)
namespace = 'mcp'
slaveTroveName = 'group-jobslave'
slaveTroveLabel = 'products.rpath.com@rpath:js'
<commit_msg>Move default location for MCP logs into /var/log/rbuilder/<commit_after>
|
#
# Copyright (c) 2005-2006 rPath, Inc.
#
# All rights reserved
#
import os
from conary import conarycfg
from conary.lib import cfgtypes
class MCPConfig(conarycfg.ConfigFile):
basePath = os.path.join(os.path.sep, 'srv', 'rbuilder', 'mcp')
logPath = os.path.join(os.path.sep, 'var', 'log', 'rbuilder')
queueHost = '127.0.0.1'
queuePort = (cfgtypes.CfgInt, 61613)
namespace = 'mcp'
slaveTroveName = 'group-jobslave'
slaveTroveLabel = 'products.rpath.com@rpath:js'
|
#
# Copyright (c) 2005-2006 rPath, Inc.
#
# All rights reserved
#
import os
from conary import conarycfg
from conary.lib import cfgtypes
class MCPConfig(conarycfg.ConfigFile):
basePath = os.path.join(os.path.sep, 'srv', 'rbuilder', 'mcp')
logPath = os.path.join(basePath, 'logs')
queueHost = '127.0.0.1'
queuePort = (cfgtypes.CfgInt, 61613)
namespace = 'mcp'
slaveTroveName = 'group-jobslave'
slaveTroveLabel = 'products.rpath.com@rpath:js'
Move default location for MCP logs into /var/log/rbuilder/#
# Copyright (c) 2005-2006 rPath, Inc.
#
# All rights reserved
#
import os
from conary import conarycfg
from conary.lib import cfgtypes
class MCPConfig(conarycfg.ConfigFile):
basePath = os.path.join(os.path.sep, 'srv', 'rbuilder', 'mcp')
logPath = os.path.join(os.path.sep, 'var', 'log', 'rbuilder')
queueHost = '127.0.0.1'
queuePort = (cfgtypes.CfgInt, 61613)
namespace = 'mcp'
slaveTroveName = 'group-jobslave'
slaveTroveLabel = 'products.rpath.com@rpath:js'
|
<commit_before>#
# Copyright (c) 2005-2006 rPath, Inc.
#
# All rights reserved
#
import os
from conary import conarycfg
from conary.lib import cfgtypes
class MCPConfig(conarycfg.ConfigFile):
basePath = os.path.join(os.path.sep, 'srv', 'rbuilder', 'mcp')
logPath = os.path.join(basePath, 'logs')
queueHost = '127.0.0.1'
queuePort = (cfgtypes.CfgInt, 61613)
namespace = 'mcp'
slaveTroveName = 'group-jobslave'
slaveTroveLabel = 'products.rpath.com@rpath:js'
<commit_msg>Move default location for MCP logs into /var/log/rbuilder/<commit_after>#
# Copyright (c) 2005-2006 rPath, Inc.
#
# All rights reserved
#
import os
from conary import conarycfg
from conary.lib import cfgtypes
class MCPConfig(conarycfg.ConfigFile):
basePath = os.path.join(os.path.sep, 'srv', 'rbuilder', 'mcp')
logPath = os.path.join(os.path.sep, 'var', 'log', 'rbuilder')
queueHost = '127.0.0.1'
queuePort = (cfgtypes.CfgInt, 61613)
namespace = 'mcp'
slaveTroveName = 'group-jobslave'
slaveTroveLabel = 'products.rpath.com@rpath:js'
|
520ad6a456cbd94e176bb54373669baf5e8cfbd9
|
sprockets/mixins/correlation/__init__.py
|
sprockets/mixins/correlation/__init__.py
|
from .mixins import HandlerMixin
version_info = (1, 0, 2)
__version__ = '.'.join(str(v) for v in version_info[:3])
|
try:
from .mixins import HandlerMixin
except ImportError as error:
class HandlerMixin(object):
def __init__(self, *args, **kwargs):
raise error
version_info = (1, 0, 2)
__version__ = '.'.join(str(v) for v in version_info[:3])
|
Fix retrieving __version__ without Tornado installed.
|
Fix retrieving __version__ without Tornado installed.
|
Python
|
bsd-3-clause
|
sprockets/sprockets.mixins.correlation
|
from .mixins import HandlerMixin
version_info = (1, 0, 2)
__version__ = '.'.join(str(v) for v in version_info[:3])
Fix retrieving __version__ without Tornado installed.
|
try:
from .mixins import HandlerMixin
except ImportError as error:
class HandlerMixin(object):
def __init__(self, *args, **kwargs):
raise error
version_info = (1, 0, 2)
__version__ = '.'.join(str(v) for v in version_info[:3])
|
<commit_before>from .mixins import HandlerMixin
version_info = (1, 0, 2)
__version__ = '.'.join(str(v) for v in version_info[:3])
<commit_msg>Fix retrieving __version__ without Tornado installed.<commit_after>
|
try:
from .mixins import HandlerMixin
except ImportError as error:
class HandlerMixin(object):
def __init__(self, *args, **kwargs):
raise error
version_info = (1, 0, 2)
__version__ = '.'.join(str(v) for v in version_info[:3])
|
from .mixins import HandlerMixin
version_info = (1, 0, 2)
__version__ = '.'.join(str(v) for v in version_info[:3])
Fix retrieving __version__ without Tornado installed.try:
from .mixins import HandlerMixin
except ImportError as error:
class HandlerMixin(object):
def __init__(self, *args, **kwargs):
raise error
version_info = (1, 0, 2)
__version__ = '.'.join(str(v) for v in version_info[:3])
|
<commit_before>from .mixins import HandlerMixin
version_info = (1, 0, 2)
__version__ = '.'.join(str(v) for v in version_info[:3])
<commit_msg>Fix retrieving __version__ without Tornado installed.<commit_after>try:
from .mixins import HandlerMixin
except ImportError as error:
class HandlerMixin(object):
def __init__(self, *args, **kwargs):
raise error
version_info = (1, 0, 2)
__version__ = '.'.join(str(v) for v in version_info[:3])
|
99eafe1fb8ed3edce0d8d025b74ffdffa3bf8ae6
|
fabfile.py
|
fabfile.py
|
import sys
import sh
from fabric import api as fab
sed = sh.sed.bake('-i bak -e')
TRAVIS_YAML = '.travis.yml'
REPLACE_LANGUAGE = 's/language: .*/language: {}/'
def is_dirty():
return "" != sh.git.status(porcelain=True).strip()
def release(language, message):
if is_dirty():
sys.exit("Repo must be in clean state before deploying. Please commit changes.")
sed(REPLACE_LANGUAGE.format(language), TRAVIS_YAML)
if is_dirty():
sh.git.add(TRAVIS_YAML)
sh.git.commit(m=message, allow_empty=True)
sh.git.pull(rebase=True)
sh.git.push()
@fab.task
def update():
if is_dirty():
sys.exit("Repo must be in clean state before deploying. Please commit changes.")
sh.git.submodule.update(remote=True, rebase=True)
if is_dirty():
sh.git.add(all=True)
sh.git.commit(m="Update submodules to origin")
@fab.task
def release_osx():
release('objective-c', "Release OS X")
@fab.task
def release_linux():
release('python', "Release Linux")
|
import sys
import sh
from fabric import api as fab
sed = sh.sed.bake('-i bak -e')
TRAVIS_YAML = '.travis.yml'
REPLACE_LANGUAGE = 's/language: .*/language: {}/'
def is_dirty():
return "" != sh.git.status(porcelain=True).strip()
def release(language, message):
if is_dirty():
sys.exit("Repo must be in clean state before deploying. Please commit changes.")
sed(REPLACE_LANGUAGE.format(language), TRAVIS_YAML)
if is_dirty():
sh.git.add(TRAVIS_YAML)
sh.git.commit(m=message, allow_empty=True)
sh.git.pull(rebase=True)
sh.git.push()
@fab.task
def update():
if is_dirty():
sys.exit("Repo must be in clean state before deploying. Please commit changes.")
sh.git.submodule.update(remote=True, rebase=True)
if is_dirty():
sh.git.add(all=True)
sh.git.commit(m="Update submodules to origin")
else:
sys.exit('Nothing to update.')
@fab.task
def release_osx():
release('objective-c', "Release OS X")
@fab.task
def release_linux():
release('python', "Release Linux")
|
Print if nothing to update
|
Print if nothing to update
|
Python
|
bsd-3-clause
|
datamicroscopes/release,jzf2101/release,jzf2101/release,datamicroscopes/release
|
import sys
import sh
from fabric import api as fab
sed = sh.sed.bake('-i bak -e')
TRAVIS_YAML = '.travis.yml'
REPLACE_LANGUAGE = 's/language: .*/language: {}/'
def is_dirty():
return "" != sh.git.status(porcelain=True).strip()
def release(language, message):
if is_dirty():
sys.exit("Repo must be in clean state before deploying. Please commit changes.")
sed(REPLACE_LANGUAGE.format(language), TRAVIS_YAML)
if is_dirty():
sh.git.add(TRAVIS_YAML)
sh.git.commit(m=message, allow_empty=True)
sh.git.pull(rebase=True)
sh.git.push()
@fab.task
def update():
if is_dirty():
sys.exit("Repo must be in clean state before deploying. Please commit changes.")
sh.git.submodule.update(remote=True, rebase=True)
if is_dirty():
sh.git.add(all=True)
sh.git.commit(m="Update submodules to origin")
@fab.task
def release_osx():
release('objective-c', "Release OS X")
@fab.task
def release_linux():
release('python', "Release Linux")
Print if nothing to update
|
import sys
import sh
from fabric import api as fab
sed = sh.sed.bake('-i bak -e')
TRAVIS_YAML = '.travis.yml'
REPLACE_LANGUAGE = 's/language: .*/language: {}/'
def is_dirty():
return "" != sh.git.status(porcelain=True).strip()
def release(language, message):
if is_dirty():
sys.exit("Repo must be in clean state before deploying. Please commit changes.")
sed(REPLACE_LANGUAGE.format(language), TRAVIS_YAML)
if is_dirty():
sh.git.add(TRAVIS_YAML)
sh.git.commit(m=message, allow_empty=True)
sh.git.pull(rebase=True)
sh.git.push()
@fab.task
def update():
if is_dirty():
sys.exit("Repo must be in clean state before deploying. Please commit changes.")
sh.git.submodule.update(remote=True, rebase=True)
if is_dirty():
sh.git.add(all=True)
sh.git.commit(m="Update submodules to origin")
else:
sys.exit('Nothing to update.')
@fab.task
def release_osx():
release('objective-c', "Release OS X")
@fab.task
def release_linux():
release('python', "Release Linux")
|
<commit_before>import sys
import sh
from fabric import api as fab
sed = sh.sed.bake('-i bak -e')
TRAVIS_YAML = '.travis.yml'
REPLACE_LANGUAGE = 's/language: .*/language: {}/'
def is_dirty():
return "" != sh.git.status(porcelain=True).strip()
def release(language, message):
if is_dirty():
sys.exit("Repo must be in clean state before deploying. Please commit changes.")
sed(REPLACE_LANGUAGE.format(language), TRAVIS_YAML)
if is_dirty():
sh.git.add(TRAVIS_YAML)
sh.git.commit(m=message, allow_empty=True)
sh.git.pull(rebase=True)
sh.git.push()
@fab.task
def update():
if is_dirty():
sys.exit("Repo must be in clean state before deploying. Please commit changes.")
sh.git.submodule.update(remote=True, rebase=True)
if is_dirty():
sh.git.add(all=True)
sh.git.commit(m="Update submodules to origin")
@fab.task
def release_osx():
release('objective-c', "Release OS X")
@fab.task
def release_linux():
release('python', "Release Linux")
<commit_msg>Print if nothing to update<commit_after>
|
import sys
import sh
from fabric import api as fab
sed = sh.sed.bake('-i bak -e')
TRAVIS_YAML = '.travis.yml'
REPLACE_LANGUAGE = 's/language: .*/language: {}/'
def is_dirty():
return "" != sh.git.status(porcelain=True).strip()
def release(language, message):
if is_dirty():
sys.exit("Repo must be in clean state before deploying. Please commit changes.")
sed(REPLACE_LANGUAGE.format(language), TRAVIS_YAML)
if is_dirty():
sh.git.add(TRAVIS_YAML)
sh.git.commit(m=message, allow_empty=True)
sh.git.pull(rebase=True)
sh.git.push()
@fab.task
def update():
if is_dirty():
sys.exit("Repo must be in clean state before deploying. Please commit changes.")
sh.git.submodule.update(remote=True, rebase=True)
if is_dirty():
sh.git.add(all=True)
sh.git.commit(m="Update submodules to origin")
else:
sys.exit('Nothing to update.')
@fab.task
def release_osx():
release('objective-c', "Release OS X")
@fab.task
def release_linux():
release('python', "Release Linux")
|
import sys
import sh
from fabric import api as fab
sed = sh.sed.bake('-i bak -e')
TRAVIS_YAML = '.travis.yml'
REPLACE_LANGUAGE = 's/language: .*/language: {}/'
def is_dirty():
return "" != sh.git.status(porcelain=True).strip()
def release(language, message):
if is_dirty():
sys.exit("Repo must be in clean state before deploying. Please commit changes.")
sed(REPLACE_LANGUAGE.format(language), TRAVIS_YAML)
if is_dirty():
sh.git.add(TRAVIS_YAML)
sh.git.commit(m=message, allow_empty=True)
sh.git.pull(rebase=True)
sh.git.push()
@fab.task
def update():
if is_dirty():
sys.exit("Repo must be in clean state before deploying. Please commit changes.")
sh.git.submodule.update(remote=True, rebase=True)
if is_dirty():
sh.git.add(all=True)
sh.git.commit(m="Update submodules to origin")
@fab.task
def release_osx():
release('objective-c', "Release OS X")
@fab.task
def release_linux():
release('python', "Release Linux")
Print if nothing to updateimport sys
import sh
from fabric import api as fab
sed = sh.sed.bake('-i bak -e')
TRAVIS_YAML = '.travis.yml'
REPLACE_LANGUAGE = 's/language: .*/language: {}/'
def is_dirty():
return "" != sh.git.status(porcelain=True).strip()
def release(language, message):
if is_dirty():
sys.exit("Repo must be in clean state before deploying. Please commit changes.")
sed(REPLACE_LANGUAGE.format(language), TRAVIS_YAML)
if is_dirty():
sh.git.add(TRAVIS_YAML)
sh.git.commit(m=message, allow_empty=True)
sh.git.pull(rebase=True)
sh.git.push()
@fab.task
def update():
if is_dirty():
sys.exit("Repo must be in clean state before deploying. Please commit changes.")
sh.git.submodule.update(remote=True, rebase=True)
if is_dirty():
sh.git.add(all=True)
sh.git.commit(m="Update submodules to origin")
else:
sys.exit('Nothing to update.')
@fab.task
def release_osx():
release('objective-c', "Release OS X")
@fab.task
def release_linux():
release('python', "Release Linux")
|
<commit_before>import sys
import sh
from fabric import api as fab
sed = sh.sed.bake('-i bak -e')
TRAVIS_YAML = '.travis.yml'
REPLACE_LANGUAGE = 's/language: .*/language: {}/'
def is_dirty():
return "" != sh.git.status(porcelain=True).strip()
def release(language, message):
if is_dirty():
sys.exit("Repo must be in clean state before deploying. Please commit changes.")
sed(REPLACE_LANGUAGE.format(language), TRAVIS_YAML)
if is_dirty():
sh.git.add(TRAVIS_YAML)
sh.git.commit(m=message, allow_empty=True)
sh.git.pull(rebase=True)
sh.git.push()
@fab.task
def update():
if is_dirty():
sys.exit("Repo must be in clean state before deploying. Please commit changes.")
sh.git.submodule.update(remote=True, rebase=True)
if is_dirty():
sh.git.add(all=True)
sh.git.commit(m="Update submodules to origin")
@fab.task
def release_osx():
release('objective-c', "Release OS X")
@fab.task
def release_linux():
release('python', "Release Linux")
<commit_msg>Print if nothing to update<commit_after>import sys
import sh
from fabric import api as fab
sed = sh.sed.bake('-i bak -e')
TRAVIS_YAML = '.travis.yml'
REPLACE_LANGUAGE = 's/language: .*/language: {}/'
def is_dirty():
return "" != sh.git.status(porcelain=True).strip()
def release(language, message):
if is_dirty():
sys.exit("Repo must be in clean state before deploying. Please commit changes.")
sed(REPLACE_LANGUAGE.format(language), TRAVIS_YAML)
if is_dirty():
sh.git.add(TRAVIS_YAML)
sh.git.commit(m=message, allow_empty=True)
sh.git.pull(rebase=True)
sh.git.push()
@fab.task
def update():
if is_dirty():
sys.exit("Repo must be in clean state before deploying. Please commit changes.")
sh.git.submodule.update(remote=True, rebase=True)
if is_dirty():
sh.git.add(all=True)
sh.git.commit(m="Update submodules to origin")
else:
sys.exit('Nothing to update.')
@fab.task
def release_osx():
release('objective-c', "Release OS X")
@fab.task
def release_linux():
release('python', "Release Linux")
|
70f0be172801ee5fd205a90c78e2bf66f8e4ae07
|
playserver/webserver.py
|
playserver/webserver.py
|
import flask
from . import track
app = flask.Flask(__name__)
@app.route("/")
def root():
song = track.getCurrentSong()
artist = track.getCurrentArtist()
album = track.getCurrentAlbum()
return "{} by {} - {}".format(song, artist, album)
|
import flask
import json
from . import track
app = flask.Flask(__name__)
@app.route("/")
def root():
song = track.getCurrentSong()
artist = track.getCurrentArtist()
album = track.getCurrentAlbum()
return "{} by {} - {}".format(song, artist, album)
@app.route("/get_song_info")
def getSongInfo():
return json.dumps({
"name": track.getCurrentSong(),
"artist": track.getCurrentAritst(),
"album": track.getCrrentAlbum()
})
@app.route("/play_pause", methods = ["POST"])
def playPause():
track.playPause()
return ""
@app.route("/next", methods = ["POST"])
def next():
track.next()
return ""
@app.route("/previous", methods = ["POST"])
def previous():
track.previous()
return ""
|
Add basic routes for controls and song info
|
Add basic routes for controls and song info
|
Python
|
mit
|
ollien/playserver,ollien/playserver,ollien/playserver
|
import flask
from . import track
app = flask.Flask(__name__)
@app.route("/")
def root():
song = track.getCurrentSong()
artist = track.getCurrentArtist()
album = track.getCurrentAlbum()
return "{} by {} - {}".format(song, artist, album)
Add basic routes for controls and song info
|
import flask
import json
from . import track
app = flask.Flask(__name__)
@app.route("/")
def root():
song = track.getCurrentSong()
artist = track.getCurrentArtist()
album = track.getCurrentAlbum()
return "{} by {} - {}".format(song, artist, album)
@app.route("/get_song_info")
def getSongInfo():
return json.dumps({
"name": track.getCurrentSong(),
"artist": track.getCurrentAritst(),
"album": track.getCrrentAlbum()
})
@app.route("/play_pause", methods = ["POST"])
def playPause():
track.playPause()
return ""
@app.route("/next", methods = ["POST"])
def next():
track.next()
return ""
@app.route("/previous", methods = ["POST"])
def previous():
track.previous()
return ""
|
<commit_before>import flask
from . import track
app = flask.Flask(__name__)
@app.route("/")
def root():
song = track.getCurrentSong()
artist = track.getCurrentArtist()
album = track.getCurrentAlbum()
return "{} by {} - {}".format(song, artist, album)
<commit_msg>Add basic routes for controls and song info<commit_after>
|
import flask
import json
from . import track
app = flask.Flask(__name__)
@app.route("/")
def root():
song = track.getCurrentSong()
artist = track.getCurrentArtist()
album = track.getCurrentAlbum()
return "{} by {} - {}".format(song, artist, album)
@app.route("/get_song_info")
def getSongInfo():
return json.dumps({
"name": track.getCurrentSong(),
"artist": track.getCurrentAritst(),
"album": track.getCrrentAlbum()
})
@app.route("/play_pause", methods = ["POST"])
def playPause():
track.playPause()
return ""
@app.route("/next", methods = ["POST"])
def next():
track.next()
return ""
@app.route("/previous", methods = ["POST"])
def previous():
track.previous()
return ""
|
import flask
from . import track
app = flask.Flask(__name__)
@app.route("/")
def root():
song = track.getCurrentSong()
artist = track.getCurrentArtist()
album = track.getCurrentAlbum()
return "{} by {} - {}".format(song, artist, album)
Add basic routes for controls and song infoimport flask
import json
from . import track
app = flask.Flask(__name__)
@app.route("/")
def root():
song = track.getCurrentSong()
artist = track.getCurrentArtist()
album = track.getCurrentAlbum()
return "{} by {} - {}".format(song, artist, album)
@app.route("/get_song_info")
def getSongInfo():
return json.dumps({
"name": track.getCurrentSong(),
"artist": track.getCurrentAritst(),
"album": track.getCrrentAlbum()
})
@app.route("/play_pause", methods = ["POST"])
def playPause():
track.playPause()
return ""
@app.route("/next", methods = ["POST"])
def next():
track.next()
return ""
@app.route("/previous", methods = ["POST"])
def previous():
track.previous()
return ""
|
<commit_before>import flask
from . import track
app = flask.Flask(__name__)
@app.route("/")
def root():
song = track.getCurrentSong()
artist = track.getCurrentArtist()
album = track.getCurrentAlbum()
return "{} by {} - {}".format(song, artist, album)
<commit_msg>Add basic routes for controls and song info<commit_after>import flask
import json
from . import track
app = flask.Flask(__name__)
@app.route("/")
def root():
song = track.getCurrentSong()
artist = track.getCurrentArtist()
album = track.getCurrentAlbum()
return "{} by {} - {}".format(song, artist, album)
@app.route("/get_song_info")
def getSongInfo():
return json.dumps({
"name": track.getCurrentSong(),
"artist": track.getCurrentAritst(),
"album": track.getCrrentAlbum()
})
@app.route("/play_pause", methods = ["POST"])
def playPause():
track.playPause()
return ""
@app.route("/next", methods = ["POST"])
def next():
track.next()
return ""
@app.route("/previous", methods = ["POST"])
def previous():
track.previous()
return ""
|
76d60adabc44fd3bbd432ee2cdad011b542a2fee
|
nel/features/mapping.py
|
nel/features/mapping.py
|
import math
import numpy
class FeatureMapper(object):
def __init__(self, features):
self.features = features
def map(self, fv):
raise NotImplementedError
def __call__(self, doc):
for chain in doc.chains:
for c in chain.candidates:
c.fv = self.map(numpy.array([c.features[f] for f in self.features]))
return doc
class ZeroMeanUnitVarianceMapper(FeatureMapper):
def __init__(self, features, means, stds):
super(ZeroMeanUnitVarianceMapper,self).__init__(features)
self.mean = means
self.std = stds
def map(self, fv):
return (fv - self.mean) / self.std
class PolynomialMapper(ZeroMeanUnitVarianceMapper):
def __init__(self, features, means, stds):
super(PolynomialMapper,self).__init__(features, means, stds)
def map(self, fv):
fv = list(super(PolynomialMapper, self).map(fv))
sz = len(fv)
for i in xrange(0, sz):
for j in xrange(i, sz):
weight = 1.0 if i != j else math.sqrt(2.0)
fv.append(weight * fv[i]*fv[j])
return numpy.array(fv)
FEATURE_MAPPERS = {cls.__name__:cls for cls in [ZeroMeanUnitVarianceMapper,PolynomialMapper]}
|
import math
import numpy
class FeatureMapper(object):
def __init__(self, features):
self.features = features
def map(self, fv):
raise NotImplementedError
def __call__(self, doc):
for chain in doc.chains:
for c in chain.candidates:
c.fv = self.map(numpy.array([c.features[f] for f in self.features]))
return doc
def feature_vector_length(self):
raise NotImplementedError
class ZeroMeanUnitVarianceMapper(FeatureMapper):
def __init__(self, features, means, stds):
super(ZeroMeanUnitVarianceMapper,self).__init__(features)
self.mean = means
self.std = stds
def map(self, fv):
return (fv - self.mean) / self.std
def feature_vector_length(self):
return len(self.features)
class PolynomialMapper(ZeroMeanUnitVarianceMapper):
def __init__(self, features, means, stds):
super(PolynomialMapper,self).__init__(features, means, stds)
def map(self, fv):
fv = list(super(PolynomialMapper, self).map(fv))
sz = len(fv)
for i in xrange(0, sz):
for j in xrange(i, sz):
weight = 1.0 if i != j else math.sqrt(2.0)
fv.append(weight * fv[i]*fv[j])
return numpy.array(fv)
def feature_vector_length(self):
n = len(self.features)
return n + n*(n+1)/2
FEATURE_MAPPERS = {cls.__name__:cls for cls in [ZeroMeanUnitVarianceMapper,PolynomialMapper]}
|
Add feature vector size calculation method to mapper interface
|
Add feature vector size calculation method to mapper interface
|
Python
|
mit
|
wikilinks/nel,wikilinks/nel
|
import math
import numpy
class FeatureMapper(object):
def __init__(self, features):
self.features = features
def map(self, fv):
raise NotImplementedError
def __call__(self, doc):
for chain in doc.chains:
for c in chain.candidates:
c.fv = self.map(numpy.array([c.features[f] for f in self.features]))
return doc
class ZeroMeanUnitVarianceMapper(FeatureMapper):
def __init__(self, features, means, stds):
super(ZeroMeanUnitVarianceMapper,self).__init__(features)
self.mean = means
self.std = stds
def map(self, fv):
return (fv - self.mean) / self.std
class PolynomialMapper(ZeroMeanUnitVarianceMapper):
def __init__(self, features, means, stds):
super(PolynomialMapper,self).__init__(features, means, stds)
def map(self, fv):
fv = list(super(PolynomialMapper, self).map(fv))
sz = len(fv)
for i in xrange(0, sz):
for j in xrange(i, sz):
weight = 1.0 if i != j else math.sqrt(2.0)
fv.append(weight * fv[i]*fv[j])
return numpy.array(fv)
FEATURE_MAPPERS = {cls.__name__:cls for cls in [ZeroMeanUnitVarianceMapper,PolynomialMapper]}
Add feature vector size calculation method to mapper interface
|
import math
import numpy
class FeatureMapper(object):
def __init__(self, features):
self.features = features
def map(self, fv):
raise NotImplementedError
def __call__(self, doc):
for chain in doc.chains:
for c in chain.candidates:
c.fv = self.map(numpy.array([c.features[f] for f in self.features]))
return doc
def feature_vector_length(self):
raise NotImplementedError
class ZeroMeanUnitVarianceMapper(FeatureMapper):
def __init__(self, features, means, stds):
super(ZeroMeanUnitVarianceMapper,self).__init__(features)
self.mean = means
self.std = stds
def map(self, fv):
return (fv - self.mean) / self.std
def feature_vector_length(self):
return len(self.features)
class PolynomialMapper(ZeroMeanUnitVarianceMapper):
def __init__(self, features, means, stds):
super(PolynomialMapper,self).__init__(features, means, stds)
def map(self, fv):
fv = list(super(PolynomialMapper, self).map(fv))
sz = len(fv)
for i in xrange(0, sz):
for j in xrange(i, sz):
weight = 1.0 if i != j else math.sqrt(2.0)
fv.append(weight * fv[i]*fv[j])
return numpy.array(fv)
def feature_vector_length(self):
n = len(self.features)
return n + n*(n+1)/2
FEATURE_MAPPERS = {cls.__name__:cls for cls in [ZeroMeanUnitVarianceMapper,PolynomialMapper]}
|
<commit_before>import math
import numpy
class FeatureMapper(object):
def __init__(self, features):
self.features = features
def map(self, fv):
raise NotImplementedError
def __call__(self, doc):
for chain in doc.chains:
for c in chain.candidates:
c.fv = self.map(numpy.array([c.features[f] for f in self.features]))
return doc
class ZeroMeanUnitVarianceMapper(FeatureMapper):
def __init__(self, features, means, stds):
super(ZeroMeanUnitVarianceMapper,self).__init__(features)
self.mean = means
self.std = stds
def map(self, fv):
return (fv - self.mean) / self.std
class PolynomialMapper(ZeroMeanUnitVarianceMapper):
def __init__(self, features, means, stds):
super(PolynomialMapper,self).__init__(features, means, stds)
def map(self, fv):
fv = list(super(PolynomialMapper, self).map(fv))
sz = len(fv)
for i in xrange(0, sz):
for j in xrange(i, sz):
weight = 1.0 if i != j else math.sqrt(2.0)
fv.append(weight * fv[i]*fv[j])
return numpy.array(fv)
FEATURE_MAPPERS = {cls.__name__:cls for cls in [ZeroMeanUnitVarianceMapper,PolynomialMapper]}
<commit_msg>Add feature vector size calculation method to mapper interface<commit_after>
|
import math
import numpy
class FeatureMapper(object):
def __init__(self, features):
self.features = features
def map(self, fv):
raise NotImplementedError
def __call__(self, doc):
for chain in doc.chains:
for c in chain.candidates:
c.fv = self.map(numpy.array([c.features[f] for f in self.features]))
return doc
def feature_vector_length(self):
raise NotImplementedError
class ZeroMeanUnitVarianceMapper(FeatureMapper):
def __init__(self, features, means, stds):
super(ZeroMeanUnitVarianceMapper,self).__init__(features)
self.mean = means
self.std = stds
def map(self, fv):
return (fv - self.mean) / self.std
def feature_vector_length(self):
return len(self.features)
class PolynomialMapper(ZeroMeanUnitVarianceMapper):
def __init__(self, features, means, stds):
super(PolynomialMapper,self).__init__(features, means, stds)
def map(self, fv):
fv = list(super(PolynomialMapper, self).map(fv))
sz = len(fv)
for i in xrange(0, sz):
for j in xrange(i, sz):
weight = 1.0 if i != j else math.sqrt(2.0)
fv.append(weight * fv[i]*fv[j])
return numpy.array(fv)
def feature_vector_length(self):
n = len(self.features)
return n + n*(n+1)/2
FEATURE_MAPPERS = {cls.__name__:cls for cls in [ZeroMeanUnitVarianceMapper,PolynomialMapper]}
|
import math
import numpy
class FeatureMapper(object):
def __init__(self, features):
self.features = features
def map(self, fv):
raise NotImplementedError
def __call__(self, doc):
for chain in doc.chains:
for c in chain.candidates:
c.fv = self.map(numpy.array([c.features[f] for f in self.features]))
return doc
class ZeroMeanUnitVarianceMapper(FeatureMapper):
def __init__(self, features, means, stds):
super(ZeroMeanUnitVarianceMapper,self).__init__(features)
self.mean = means
self.std = stds
def map(self, fv):
return (fv - self.mean) / self.std
class PolynomialMapper(ZeroMeanUnitVarianceMapper):
def __init__(self, features, means, stds):
super(PolynomialMapper,self).__init__(features, means, stds)
def map(self, fv):
fv = list(super(PolynomialMapper, self).map(fv))
sz = len(fv)
for i in xrange(0, sz):
for j in xrange(i, sz):
weight = 1.0 if i != j else math.sqrt(2.0)
fv.append(weight * fv[i]*fv[j])
return numpy.array(fv)
FEATURE_MAPPERS = {cls.__name__:cls for cls in [ZeroMeanUnitVarianceMapper,PolynomialMapper]}
Add feature vector size calculation method to mapper interfaceimport math
import numpy
class FeatureMapper(object):
def __init__(self, features):
self.features = features
def map(self, fv):
raise NotImplementedError
def __call__(self, doc):
for chain in doc.chains:
for c in chain.candidates:
c.fv = self.map(numpy.array([c.features[f] for f in self.features]))
return doc
def feature_vector_length(self):
raise NotImplementedError
class ZeroMeanUnitVarianceMapper(FeatureMapper):
def __init__(self, features, means, stds):
super(ZeroMeanUnitVarianceMapper,self).__init__(features)
self.mean = means
self.std = stds
def map(self, fv):
return (fv - self.mean) / self.std
def feature_vector_length(self):
return len(self.features)
class PolynomialMapper(ZeroMeanUnitVarianceMapper):
def __init__(self, features, means, stds):
super(PolynomialMapper,self).__init__(features, means, stds)
def map(self, fv):
fv = list(super(PolynomialMapper, self).map(fv))
sz = len(fv)
for i in xrange(0, sz):
for j in xrange(i, sz):
weight = 1.0 if i != j else math.sqrt(2.0)
fv.append(weight * fv[i]*fv[j])
return numpy.array(fv)
def feature_vector_length(self):
n = len(self.features)
return n + n*(n+1)/2
FEATURE_MAPPERS = {cls.__name__:cls for cls in [ZeroMeanUnitVarianceMapper,PolynomialMapper]}
|
<commit_before>import math
import numpy
class FeatureMapper(object):
def __init__(self, features):
self.features = features
def map(self, fv):
raise NotImplementedError
def __call__(self, doc):
for chain in doc.chains:
for c in chain.candidates:
c.fv = self.map(numpy.array([c.features[f] for f in self.features]))
return doc
class ZeroMeanUnitVarianceMapper(FeatureMapper):
def __init__(self, features, means, stds):
super(ZeroMeanUnitVarianceMapper,self).__init__(features)
self.mean = means
self.std = stds
def map(self, fv):
return (fv - self.mean) / self.std
class PolynomialMapper(ZeroMeanUnitVarianceMapper):
def __init__(self, features, means, stds):
super(PolynomialMapper,self).__init__(features, means, stds)
def map(self, fv):
fv = list(super(PolynomialMapper, self).map(fv))
sz = len(fv)
for i in xrange(0, sz):
for j in xrange(i, sz):
weight = 1.0 if i != j else math.sqrt(2.0)
fv.append(weight * fv[i]*fv[j])
return numpy.array(fv)
FEATURE_MAPPERS = {cls.__name__:cls for cls in [ZeroMeanUnitVarianceMapper,PolynomialMapper]}
<commit_msg>Add feature vector size calculation method to mapper interface<commit_after>import math
import numpy
class FeatureMapper(object):
def __init__(self, features):
self.features = features
def map(self, fv):
raise NotImplementedError
def __call__(self, doc):
for chain in doc.chains:
for c in chain.candidates:
c.fv = self.map(numpy.array([c.features[f] for f in self.features]))
return doc
def feature_vector_length(self):
raise NotImplementedError
class ZeroMeanUnitVarianceMapper(FeatureMapper):
def __init__(self, features, means, stds):
super(ZeroMeanUnitVarianceMapper,self).__init__(features)
self.mean = means
self.std = stds
def map(self, fv):
return (fv - self.mean) / self.std
def feature_vector_length(self):
return len(self.features)
class PolynomialMapper(ZeroMeanUnitVarianceMapper):
def __init__(self, features, means, stds):
super(PolynomialMapper,self).__init__(features, means, stds)
def map(self, fv):
fv = list(super(PolynomialMapper, self).map(fv))
sz = len(fv)
for i in xrange(0, sz):
for j in xrange(i, sz):
weight = 1.0 if i != j else math.sqrt(2.0)
fv.append(weight * fv[i]*fv[j])
return numpy.array(fv)
def feature_vector_length(self):
n = len(self.features)
return n + n*(n+1)/2
FEATURE_MAPPERS = {cls.__name__:cls for cls in [ZeroMeanUnitVarianceMapper,PolynomialMapper]}
|
141ad63b28eff5c7a034d479b98c83334ff1f0a3
|
provokator/site/util.py
|
provokator/site/util.py
|
#!/usr/bin/python3 -tt
# -*- coding: utf-8 -*-
__all__ = ['internal_origin_only']
from urllib.parse import urlparse
from functools import wraps
from werkzeug.exceptions import Forbidden
import flask
import re
def internal_origin_only(fn):
@wraps(fn)
def wrapper(*args, **kwargs):
h = urlparse('http://' + flask.request.headers.get('Host', ''))
host = '%s:%i' % (h.hostname, h.port or 80)
if 'Origin' in flask.request.headers:
o = urlparse(flask.request.headers.get('Origin'))
origin = '%s:%i' % (o.hostname, o.port or 80)
elif 'Referer' in flask.request.headers:
r = urlparse(flask.request.headers.get('Referer'))
origin = '%s:%i' % (r.hostname, r.port or 80)
else:
origin = host
if host != origin:
raise Forbidden('Cross-Site Request Forbidden')
return fn(*args, **kwargs)
return wrapper
# vim:set sw=4 ts=4 et:
|
#!/usr/bin/python3 -tt
# -*- coding: utf-8 -*-
__all__ = ['internal_origin_only']
from urllib.parse import urlparse
from functools import wraps
from werkzeug.exceptions import Forbidden
import flask
import re
def internal_origin_only(fn):
@wraps(fn)
def wrapper(*args, **kwargs):
host = flask.request.headers.get('X-Forwarded-Host') or \
flask.request.headers.get('Host', '')
h = urlparse('http://' + host)
origin = flask.request.headers.get('Origin') or \
flask.request.headers.get('Referer') or \
host
o = urlparse(origin)
if h.hostname != o.hostname:
raise Forbidden('Cross-Site Request Forbidden')
return fn(*args, **kwargs)
return wrapper
# vim:set sw=4 ts=4 et:
|
Change cross-origin check to work behind proxies
|
Change cross-origin check to work behind proxies
Signed-off-by: Jan Dvořák <86df5a4870880bf501c926309e3bcfbe57789f3f@anilinux.org>
|
Python
|
mit
|
techlib/provokator,techlib/provokator
|
#!/usr/bin/python3 -tt
# -*- coding: utf-8 -*-
__all__ = ['internal_origin_only']
from urllib.parse import urlparse
from functools import wraps
from werkzeug.exceptions import Forbidden
import flask
import re
def internal_origin_only(fn):
@wraps(fn)
def wrapper(*args, **kwargs):
h = urlparse('http://' + flask.request.headers.get('Host', ''))
host = '%s:%i' % (h.hostname, h.port or 80)
if 'Origin' in flask.request.headers:
o = urlparse(flask.request.headers.get('Origin'))
origin = '%s:%i' % (o.hostname, o.port or 80)
elif 'Referer' in flask.request.headers:
r = urlparse(flask.request.headers.get('Referer'))
origin = '%s:%i' % (r.hostname, r.port or 80)
else:
origin = host
if host != origin:
raise Forbidden('Cross-Site Request Forbidden')
return fn(*args, **kwargs)
return wrapper
# vim:set sw=4 ts=4 et:
Change cross-origin check to work behind proxies
Signed-off-by: Jan Dvořák <86df5a4870880bf501c926309e3bcfbe57789f3f@anilinux.org>
|
#!/usr/bin/python3 -tt
# -*- coding: utf-8 -*-
__all__ = ['internal_origin_only']
from urllib.parse import urlparse
from functools import wraps
from werkzeug.exceptions import Forbidden
import flask
import re
def internal_origin_only(fn):
@wraps(fn)
def wrapper(*args, **kwargs):
host = flask.request.headers.get('X-Forwarded-Host') or \
flask.request.headers.get('Host', '')
h = urlparse('http://' + host)
origin = flask.request.headers.get('Origin') or \
flask.request.headers.get('Referer') or \
host
o = urlparse(origin)
if h.hostname != o.hostname:
raise Forbidden('Cross-Site Request Forbidden')
return fn(*args, **kwargs)
return wrapper
# vim:set sw=4 ts=4 et:
|
<commit_before>#!/usr/bin/python3 -tt
# -*- coding: utf-8 -*-
__all__ = ['internal_origin_only']
from urllib.parse import urlparse
from functools import wraps
from werkzeug.exceptions import Forbidden
import flask
import re
def internal_origin_only(fn):
@wraps(fn)
def wrapper(*args, **kwargs):
h = urlparse('http://' + flask.request.headers.get('Host', ''))
host = '%s:%i' % (h.hostname, h.port or 80)
if 'Origin' in flask.request.headers:
o = urlparse(flask.request.headers.get('Origin'))
origin = '%s:%i' % (o.hostname, o.port or 80)
elif 'Referer' in flask.request.headers:
r = urlparse(flask.request.headers.get('Referer'))
origin = '%s:%i' % (r.hostname, r.port or 80)
else:
origin = host
if host != origin:
raise Forbidden('Cross-Site Request Forbidden')
return fn(*args, **kwargs)
return wrapper
# vim:set sw=4 ts=4 et:
<commit_msg>Change cross-origin check to work behind proxies
Signed-off-by: Jan Dvořák <86df5a4870880bf501c926309e3bcfbe57789f3f@anilinux.org><commit_after>
|
#!/usr/bin/python3 -tt
# -*- coding: utf-8 -*-
__all__ = ['internal_origin_only']
from urllib.parse import urlparse
from functools import wraps
from werkzeug.exceptions import Forbidden
import flask
import re
def internal_origin_only(fn):
@wraps(fn)
def wrapper(*args, **kwargs):
host = flask.request.headers.get('X-Forwarded-Host') or \
flask.request.headers.get('Host', '')
h = urlparse('http://' + host)
origin = flask.request.headers.get('Origin') or \
flask.request.headers.get('Referer') or \
host
o = urlparse(origin)
if h.hostname != o.hostname:
raise Forbidden('Cross-Site Request Forbidden')
return fn(*args, **kwargs)
return wrapper
# vim:set sw=4 ts=4 et:
|
#!/usr/bin/python3 -tt
# -*- coding: utf-8 -*-
__all__ = ['internal_origin_only']
from urllib.parse import urlparse
from functools import wraps
from werkzeug.exceptions import Forbidden
import flask
import re
def internal_origin_only(fn):
@wraps(fn)
def wrapper(*args, **kwargs):
h = urlparse('http://' + flask.request.headers.get('Host', ''))
host = '%s:%i' % (h.hostname, h.port or 80)
if 'Origin' in flask.request.headers:
o = urlparse(flask.request.headers.get('Origin'))
origin = '%s:%i' % (o.hostname, o.port or 80)
elif 'Referer' in flask.request.headers:
r = urlparse(flask.request.headers.get('Referer'))
origin = '%s:%i' % (r.hostname, r.port or 80)
else:
origin = host
if host != origin:
raise Forbidden('Cross-Site Request Forbidden')
return fn(*args, **kwargs)
return wrapper
# vim:set sw=4 ts=4 et:
Change cross-origin check to work behind proxies
Signed-off-by: Jan Dvořák <86df5a4870880bf501c926309e3bcfbe57789f3f@anilinux.org>#!/usr/bin/python3 -tt
# -*- coding: utf-8 -*-
__all__ = ['internal_origin_only']
from urllib.parse import urlparse
from functools import wraps
from werkzeug.exceptions import Forbidden
import flask
import re
def internal_origin_only(fn):
@wraps(fn)
def wrapper(*args, **kwargs):
host = flask.request.headers.get('X-Forwarded-Host') or \
flask.request.headers.get('Host', '')
h = urlparse('http://' + host)
origin = flask.request.headers.get('Origin') or \
flask.request.headers.get('Referer') or \
host
o = urlparse(origin)
if h.hostname != o.hostname:
raise Forbidden('Cross-Site Request Forbidden')
return fn(*args, **kwargs)
return wrapper
# vim:set sw=4 ts=4 et:
|
<commit_before>#!/usr/bin/python3 -tt
# -*- coding: utf-8 -*-
__all__ = ['internal_origin_only']
from urllib.parse import urlparse
from functools import wraps
from werkzeug.exceptions import Forbidden
import flask
import re
def internal_origin_only(fn):
@wraps(fn)
def wrapper(*args, **kwargs):
h = urlparse('http://' + flask.request.headers.get('Host', ''))
host = '%s:%i' % (h.hostname, h.port or 80)
if 'Origin' in flask.request.headers:
o = urlparse(flask.request.headers.get('Origin'))
origin = '%s:%i' % (o.hostname, o.port or 80)
elif 'Referer' in flask.request.headers:
r = urlparse(flask.request.headers.get('Referer'))
origin = '%s:%i' % (r.hostname, r.port or 80)
else:
origin = host
if host != origin:
raise Forbidden('Cross-Site Request Forbidden')
return fn(*args, **kwargs)
return wrapper
# vim:set sw=4 ts=4 et:
<commit_msg>Change cross-origin check to work behind proxies
Signed-off-by: Jan Dvořák <86df5a4870880bf501c926309e3bcfbe57789f3f@anilinux.org><commit_after>#!/usr/bin/python3 -tt
# -*- coding: utf-8 -*-
__all__ = ['internal_origin_only']
from urllib.parse import urlparse
from functools import wraps
from werkzeug.exceptions import Forbidden
import flask
import re
def internal_origin_only(fn):
@wraps(fn)
def wrapper(*args, **kwargs):
host = flask.request.headers.get('X-Forwarded-Host') or \
flask.request.headers.get('Host', '')
h = urlparse('http://' + host)
origin = flask.request.headers.get('Origin') or \
flask.request.headers.get('Referer') or \
host
o = urlparse(origin)
if h.hostname != o.hostname:
raise Forbidden('Cross-Site Request Forbidden')
return fn(*args, **kwargs)
return wrapper
# vim:set sw=4 ts=4 et:
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.